Datasets:
bitext-innovations
commited on
Commit
•
edf2358
1
Parent(s):
9d51ad5
Update README.md
Browse files
README.md
CHANGED
@@ -25,7 +25,7 @@ size_categories:
|
|
25 |
|
26 |
## Overview
|
27 |
|
28 |
-
This hybrid synthetic dataset is designed to be used to fine-tune Large Language Models such as GPT, Mistral and OpenELM, and has been generated using our NLP/NLG technology and our automated Data Labeling (DAL) tools. The goal is to demonstrate how Verticalization/Domain Adaptation for the
|
29 |
|
30 |
The dataset has the following specs:
|
31 |
|
|
|
25 |
|
26 |
## Overview
|
27 |
|
28 |
+
This hybrid synthetic dataset is designed to be used to fine-tune Large Language Models such as GPT, Mistral and OpenELM, and has been generated using our NLP/NLG technology and our automated Data Labeling (DAL) tools. The goal is to demonstrate how Verticalization/Domain Adaptation for the Customer Support sector can be easily achieved using our two-step approach to LLM Fine-Tuning. For example, if you are [ACME Company], you can create your own customized LLM by first training a fine-tuned model using this dataset, and then further fine-tuning it with a small amount of your own data. An overview of this approach can be found at: [From General-Purpose LLMs to Verticalized Enterprise Models](https://www.bitext.com/blog/general-purpose-models-verticalized-enterprise-genai/)
|
29 |
|
30 |
The dataset has the following specs:
|
31 |
|