Update README.md
Browse files
README.md
CHANGED
@@ -4,26 +4,48 @@ tags:
|
|
4 |
- merge
|
5 |
- mergekit
|
6 |
- vilm/vinallama-7b-chat
|
7 |
-
- vilm/vinallama-7b-chat
|
8 |
-
- vilm/vinallama-7b-chat
|
9 |
-
- vilm/vinallama-7b-chat
|
10 |
-
- vilm/vinallama-7b-chat
|
11 |
-
- vilm/vinallama-7b-chat
|
12 |
-
- vilm/vinallama-7b-chat
|
13 |
-
- vilm/vinallama-7b-chat
|
14 |
---
|
15 |
|
16 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
17 |
|
18 |
This model is a merge of the following models made with [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
|
19 |
* [vilm/vinallama-7b-chat](https://huggingface.co/vilm/vinallama-7b-chat)
|
20 |
-
|
21 |
-
* [vilm/vinallama-7b-chat](https://huggingface.co/vilm/vinallama-7b-chat)
|
22 |
-
* [vilm/vinallama-7b-chat](https://huggingface.co/vilm/vinallama-7b-chat)
|
23 |
-
* [vilm/vinallama-7b-chat](https://huggingface.co/vilm/vinallama-7b-chat)
|
24 |
-
* [vilm/vinallama-7b-chat](https://huggingface.co/vilm/vinallama-7b-chat)
|
25 |
-
* [vilm/vinallama-7b-chat](https://huggingface.co/vilm/vinallama-7b-chat)
|
26 |
-
* [vilm/vinallama-7b-chat](https://huggingface.co/vilm/vinallama-7b-chat)
|
27 |
|
28 |
## 🧩 Configuration
|
29 |
|
|
|
4 |
- merge
|
5 |
- mergekit
|
6 |
- vilm/vinallama-7b-chat
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
---
|
8 |
|
9 |
+
# VinaLLaMA - State-of-the-art Vietnamese LLMs
|
10 |
+
|
11 |
+
![image](https://i.ibb.co/W0dq12n/vinallama.png)
|
12 |
+
|
13 |
+
Read our [Paper](https://huggingface.co/papers/2312.11011)
|
14 |
+
|
15 |
+
Prompt Format (ChatML):
|
16 |
+
|
17 |
+
```
|
18 |
+
<|im_start|>system
|
19 |
+
Bạn là một trợ lí AI hữu ích. Hãy trả lời người dùng một cách chính xác.
|
20 |
+
<|im_end|>
|
21 |
+
<|im_start|>user
|
22 |
+
Hello world!<|im_end|>
|
23 |
+
<|im_start|>assistant
|
24 |
+
```
|
25 |
+
|
26 |
+
## Evaluation
|
27 |
+
|
28 |
+
We evaluated our model via peer comparison on multiple publicly available dataset using
|
29 |
+
<a href="https://github.com/hieunguyen1053/lm-evaluation-harness"> @hieunguyen1053 fork of lm-evaluation-harness </a>
|
30 |
+
, and combine the results with that provided by the authors of VinaLLaMA. The results are bellow:
|
31 |
+
| Model | Model size | arc_vi (acc) | hellaswag_vi (acc) | mmlu_vi (acc) | truthfulqa_vi (acc) | Average |
|
32 |
+
| ------------------------ | ---------- | ------------ | ------------------ | ------------- | ------------------- | ------- |
|
33 |
+
| URA-LLaMA-13B | | 0,3752 | 0,4830 | 0,3973 | 0,4574 | 0,4282 |
|
34 |
+
| BLOOMZ-7B | | 0,3205 | 0,4930 | 0,3975 | 0,4523 | 0,4158 |
|
35 |
+
| PhoGPT-7B5-Instruct | | 0,2470 | 0,2578 | 0,2413 | 0,4759 | 0,3055 |
|
36 |
+
| SeaLLM-7B-chat | | 0,3607 | 0,5112 | 0,3339 | 0,4948 | 0,4252 |
|
37 |
+
| Vietcuna-7b-v3 | | 0,3419 | 0,4939 | 0,3354 | 0,4807 | 0,4130 |
|
38 |
+
| VinaLLaMA-2.7B-chat | | 0,3273 | 0,4814 | 0,3051 | 0,4972 | 0,4028 |
|
39 |
+
| VinaLLaMA-7B-chat | | 0,4239 | 0,5407 | 0,3932 | 0,5251 | 0,4707 |
|
40 |
+
| VBD-LLaMA2-7B-50b | | 0,3222 | 0,5195 | 0,2964 | 0,4614 | 0,3999 |
|
41 |
+
| VBD-LLaMA2-7B-50b-Chat | | 0,3585 | 0,5207 | 0,3444 | 0,5179 | 0,4354 |
|
42 |
+
| VinaLLaMA-12.5B-chat-DUS | | 0,4325 | 0,5816 | 0,3875 | 0,5850 | 0,4967 |
|
43 |
+
|
44 |
+
# Merging Methods
|
45 |
|
46 |
This model is a merge of the following models made with [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
|
47 |
* [vilm/vinallama-7b-chat](https://huggingface.co/vilm/vinallama-7b-chat)
|
48 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
49 |
|
50 |
## 🧩 Configuration
|
51 |
|