robinsmits
commited on
Commit
•
5afafd5
1
Parent(s):
9959dd7
Update README.md
Browse files
README.md
CHANGED
@@ -127,6 +127,35 @@ DPO Finetuning was performed on the Dutch [BramVanroy/ultra_feedback_dutch_clean
|
|
127 |
See [Qwen/Qwen1.5-7B-Chat](https://huggingface.co/Qwen/Qwen1.5-7B-Chat) for all information about the base model.
|
128 |
|
129 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
130 |
## Model usage
|
131 |
|
132 |
A basic example of how to use the finetuned model.
|
@@ -231,7 +260,7 @@ The following hyperparameters were used during training:
|
|
231 |
- Tokenizers 0.15.2
|
232 |
|
233 |
## Citation
|
234 |
-
Thanks to the creators of Qwen1.5 for
|
235 |
```
|
236 |
@article{qwen,
|
237 |
title={Qwen Technical Report},
|
@@ -240,16 +269,3 @@ Thanks to the creators of Qwen1.5 for there great work!
|
|
240 |
year={2023}
|
241 |
}
|
242 |
```
|
243 |
-
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
244 |
-
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_robinsmits__Qwen1.5-7B-Dutch-Chat)
|
245 |
-
|
246 |
-
| Metric |Value|
|
247 |
-
|---------------------------------|----:|
|
248 |
-
|Avg. |53.66|
|
249 |
-
|AI2 Reasoning Challenge (25-Shot)|53.92|
|
250 |
-
|HellaSwag (10-Shot) |76.03|
|
251 |
-
|MMLU (5-Shot) |62.38|
|
252 |
-
|TruthfulQA (0-shot) |45.34|
|
253 |
-
|Winogrande (5-shot) |68.82|
|
254 |
-
|GSM8k (5-shot) |15.47|
|
255 |
-
|
|
|
127 |
See [Qwen/Qwen1.5-7B-Chat](https://huggingface.co/Qwen/Qwen1.5-7B-Chat) for all information about the base model.
|
128 |
|
129 |
|
130 |
+
|
131 |
+
## ScandEval Dutch Leaderboard Evaluation Results
|
132 |
+
|
133 |
+
For evaluation results based on the Dutch language you can take a look at the site of ScandEval.
|
134 |
+
|
135 |
+
This model achieves a score which is very close to the performance of GPT-3.5.
|
136 |
+
|
137 |
+
[Dutch Natural Language Understanding](https://scandeval.com/dutch-nlu/)
|
138 |
+
|
139 |
+
[Dutch Natural Language Generation](https://scandeval.com/dutch-nlg/)
|
140 |
+
|
141 |
+
|
142 |
+
|
143 |
+
## [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
144 |
+
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_robinsmits__Qwen1.5-7B-Dutch-Chat)
|
145 |
+
|
146 |
+
Note that these Evaluation Results are for the English language.
|
147 |
+
|
148 |
+
| Metric |Value|
|
149 |
+
|---------------------------------|----:|
|
150 |
+
|Avg. |53.66|
|
151 |
+
|AI2 Reasoning Challenge (25-Shot)|53.92|
|
152 |
+
|HellaSwag (10-Shot) |76.03|
|
153 |
+
|MMLU (5-Shot) |62.38|
|
154 |
+
|TruthfulQA (0-shot) |45.34|
|
155 |
+
|Winogrande (5-shot) |68.82|
|
156 |
+
|GSM8k (5-shot) |15.47|
|
157 |
+
|
158 |
+
|
159 |
## Model usage
|
160 |
|
161 |
A basic example of how to use the finetuned model.
|
|
|
260 |
- Tokenizers 0.15.2
|
261 |
|
262 |
## Citation
|
263 |
+
Thanks to the creators of Qwen1.5 for their great work!
|
264 |
```
|
265 |
@article{qwen,
|
266 |
title={Qwen Technical Report},
|
|
|
269 |
year={2023}
|
270 |
}
|
271 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|