Update README.md
Browse files
README.md
CHANGED
@@ -77,7 +77,7 @@ We provide 6 variants (of which 3 base and 3 instruction-tuned models):
|
|
77 |
- **Llama-3-ChocoLlama-8B-base** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-base)): A language-adapted version of Meta's Llama-8-8B, fine-tuned on the same Dutch dataset as ChocoLlama-2-7B-base, again using LoRa.
|
78 |
- **Llama-3-ChocoLlama-instruct** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-instruct)): An instruction-tuned version of Llama-3-ChocoLlama-8B-base, fine-tuned on the same dataset as ChocoLlama-2-7B-instruct, again using SFT followed by DPO.
|
79 |
|
80 |
-
For benchmark results for all models, including compared to their base models and other Dutch LLMs, we refer to our paper [here](
|
81 |
|
82 |
### Model Description
|
83 |
|
@@ -89,8 +89,8 @@ For benchmark results for all models, including compared to their base models an
|
|
89 |
|
90 |
### Model Sources
|
91 |
|
92 |
-
- **Repository:**
|
93 |
-
- **Paper:**
|
94 |
|
95 |
## Uses
|
96 |
|
|
|
77 |
- **Llama-3-ChocoLlama-8B-base** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-base)): A language-adapted version of Meta's Llama-8-8B, fine-tuned on the same Dutch dataset as ChocoLlama-2-7B-base, again using LoRa.
|
78 |
- **Llama-3-ChocoLlama-instruct** ([link](https://huggingface.co/ChocoLlama/Llama-3-ChocoLlama-8B-instruct)): An instruction-tuned version of Llama-3-ChocoLlama-8B-base, fine-tuned on the same dataset as ChocoLlama-2-7B-instruct, again using SFT followed by DPO.
|
79 |
|
80 |
+
For benchmark results for all models, including compared to their base models and other Dutch LLMs, we refer to our paper [here](https://arxiv.org/pdf/2412.07633).
|
81 |
|
82 |
### Model Description
|
83 |
|
|
|
89 |
|
90 |
### Model Sources
|
91 |
|
92 |
+
- **Repository:** [on Github here](https://github.com/ChocoLlamaModel/ChocoLlama).
|
93 |
+
- **Paper:** [on ArXiv here](https://arxiv.org/pdf/2412.07633).
|
94 |
|
95 |
## Uses
|
96 |
|