instruction-pretrain commited on
Commit
f5021bc
•
1 Parent(s): 92758fa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -4
README.md CHANGED
@@ -22,15 +22,17 @@ We explore supervised multitask pre-training by proposing ***Instruction Pre-Tra
22
  </p>
23
 
24
  **************************** **Updates** ****************************
 
25
  * 2024/7/15: We scaled up the pre-trained tokens from 100B to 250B, with the number of synthesized instruction-response pairs reaching 500M! Below, we show the performance trend on downstream tasks throughout the pre-training process:
26
- <p align='center'>
27
- <img src="https://cdn-uploads.huggingface.co/production/uploads/66711d2ee12fa6cc5f5dfc89/0okCfRkC6uALTfuNxt0Fa.png" width="700">
28
  </p>
29
  * 2024/6/21: Released the [paper](https://huggingface.co/papers/2406.14491), [code](https://github.com/microsoft/LMOps), and [resources](https://huggingface.co/instruction-pretrain)
30
 
31
  ## Resources
32
- **🤗 We share our data and models with example usages, feel free to open any issues or discussions! 🤗**
33
 
 
34
  - Context-Based Instruction Synthesizer: [instruction-synthesizer](https://huggingface.co/instruction-pretrain/instruction-synthesizer)
35
  - Fine-Tuning Data for the Synthesizer: [ft-instruction-synthesizer-collection](https://huggingface.co/datasets/instruction-pretrain/ft-instruction-synthesizer-collection)
36
  - General Models Pre-Trained from Scratch (on 100B tokes):
@@ -110,7 +112,7 @@ Instruction Pre-Training
110
  }
111
  ```
112
 
113
- [AdaptLLM](https://huggingface.co/papers/2309.09530)
114
  ```bibtex
115
  @inproceedings{
116
  cheng2024adapting,
 
22
  </p>
23
 
24
  **************************** **Updates** ****************************
25
+ * 2024/7/31: Updated pre-training suggestions in the `Advanced Usage` section of [instruction-synthesizer](https://huggingface.co/instruction-pretrain/instruction-synthesizer)
26
  * 2024/7/15: We scaled up the pre-trained tokens from 100B to 250B, with the number of synthesized instruction-response pairs reaching 500M! Below, we show the performance trend on downstream tasks throughout the pre-training process:
27
+ <p align='left'>
28
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/66711d2ee12fa6cc5f5dfc89/0okCfRkC6uALTfuNxt0Fa.png" width="500">
29
  </p>
30
  * 2024/6/21: Released the [paper](https://huggingface.co/papers/2406.14491), [code](https://github.com/microsoft/LMOps), and [resources](https://huggingface.co/instruction-pretrain)
31
 
32
  ## Resources
33
+ **🤗 We share our data and models with example usages, feel free to open any discussions at [this page](https://huggingface.co/papers/2406.14491)! 🤗**
34
 
35
+ - Thanks to the demo [davanstrien/instruction-synthesizer](https://huggingface.co/spaces/davanstrien/instruction-synthesizer) for implementing our approach
36
  - Context-Based Instruction Synthesizer: [instruction-synthesizer](https://huggingface.co/instruction-pretrain/instruction-synthesizer)
37
  - Fine-Tuning Data for the Synthesizer: [ft-instruction-synthesizer-collection](https://huggingface.co/datasets/instruction-pretrain/ft-instruction-synthesizer-collection)
38
  - General Models Pre-Trained from Scratch (on 100B tokes):
 
112
  }
113
  ```
114
 
115
+ [Adapt LLM to Domains](https://huggingface.co/papers/2309.09530)
116
  ```bibtex
117
  @inproceedings{
118
  cheng2024adapting,