matthieumeeus97 commited on
Commit
a859a39
·
verified ·
1 Parent(s): 60295dc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -77,7 +77,7 @@ We provide 6 variants (of which 3 base and 3 instruction-tuned models):
77
  - **Llama-3-ChocoLlama-8B-base** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-base)): A language-adapted version of Meta's Llama-8-8B, fine-tuned on the same Dutch dataset as ChocoLlama-2-7B-base, again using LoRa.
78
  - **Llama-3-ChocoLlama-instruct** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-instruct)): An instruction-tuned version of Llama-3-ChocoLlama-8B-base, fine-tuned on the same dataset as ChocoLlama-2-7B-instruct, again using SFT followed by DPO.
79
 
80
- For benchmark results for all models, including compared to their base models and other Dutch LLMs, we refer to our paper [here](some_url).
81
 
82
  ### Model Description
83
 
@@ -89,8 +89,8 @@ For benchmark results for all models, including compared to their base models an
89
 
90
  ### Model Sources
91
 
92
- - **Repository:** Will be released soon.
93
- - **Paper:** Will be released soon.
94
 
95
  ## Uses
96
 
 
77
  - **Llama-3-ChocoLlama-8B-base** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-base)): A language-adapted version of Meta's Llama-8-8B, fine-tuned on the same Dutch dataset as ChocoLlama-2-7B-base, again using LoRa.
78
  - **Llama-3-ChocoLlama-instruct** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-instruct)): An instruction-tuned version of Llama-3-ChocoLlama-8B-base, fine-tuned on the same dataset as ChocoLlama-2-7B-instruct, again using SFT followed by DPO.
79
 
80
+ For benchmark results for all models, including compared to their base models and other Dutch LLMs, we refer to our paper [here](https://arxiv.org/pdf/2412.07633).
81
 
82
  ### Model Description
83
 
 
89
 
90
  ### Model Sources
91
 
92
+ - **Repository:** [on Github here](https://github.com/ChocoLlamaModel/ChocoLlama).
93
+ - **Paper:** [on ArXiv here](https://arxiv.org/pdf/2412.07633).
94
 
95
  ## Uses
96