Update README.md
Browse files
README.md
CHANGED
@@ -15,7 +15,7 @@ Alpha-Instruct has achieved outstanding performance on the **LogicKor, scoring a
|
|
15 |
|
16 |
---
|
17 |
## Overview
|
18 |
-
Alpha-Instruct is our latest language model, developed using 'Evolutionary Model Merging' technique. This method employs a 1:1 ratio of task-specific datasets from KoBEST and Haerae, resulting in a model
|
19 |
- [Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) (Base)
|
20 |
- [Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) (Instruct)
|
21 |
- [Llama-3-Open-Ko-8B](beomi/Llama-3-Open-Ko-8B) (Continual Pretrained)
|
@@ -36,7 +36,7 @@ Results in [LogicKor](https://github.com/StableFluffy/LogicKor)* are as follows:
|
|
36 |
|:------------------------------:|:------------:|:-----------:|:--------:|
|
37 |
| MLP-KTLim/llama-3-Korean-Bllossom-8B | 4.238 | 3.404 | 3.821 |
|
38 |
| Alpha-Ko-Evo | 5.143 | 5.238 | 5.190 |
|
39 |
-
| Alpha-Ko-Instruct (alt) | 7.095 | 6.571 | **6.833** |
|
40 |
| Alpha-Ko-Instruct | **7.143** | 6.065 | 6.620 |
|
41 |
| Alpha-Ko-Instruct-marlin (4bit) | 6.857 | 5.738 | 6.298 |
|
42 |
|
@@ -44,7 +44,7 @@ Results in [LogicKor](https://github.com/StableFluffy/LogicKor)* are as follows:
|
|
44 |
|
45 |
Result in KoBEST(acc, num_shot=5) are as follows:
|
46 |
|
47 |
-
| Task | beomi/Llama-3-Open-Ko-8B-Instruct | maywell/Llama-3-Ko-8B-Instruct | Alpha-Ko-Evo | Alpha-Ko-Instruct |
|
48 |
| --- | --- | --- | --- | --- |
|
49 |
| kobest overall | 0.6220 | 0.6852 |0.7229|0.7055
|
50 |
| kobest_boolq| 0.6254 | 0.7208 | 0.8547 | 0.8369
|
|
|
15 |
|
16 |
---
|
17 |
## Overview
|
18 |
+
Alpha-Instruct is our latest language model, developed using 'Evolutionary Model Merging' technique. This method employs a 1:1 ratio of task-specific datasets from KoBEST and Haerae, resulting in a model with named 'Alpha-Ko-8B-Evo'. The following models were used for merging:
|
19 |
- [Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) (Base)
|
20 |
- [Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) (Instruct)
|
21 |
- [Llama-3-Open-Ko-8B](beomi/Llama-3-Open-Ko-8B) (Continual Pretrained)
|
|
|
36 |
|:------------------------------:|:------------:|:-----------:|:--------:|
|
37 |
| MLP-KTLim/llama-3-Korean-Bllossom-8B | 4.238 | 3.404 | 3.821 |
|
38 |
| Alpha-Ko-Evo | 5.143 | 5.238 | 5.190 |
|
39 |
+
| Alpha-Ko-Instruct (alt) | 7.095 | **6.571** | **6.833** |
|
40 |
| Alpha-Ko-Instruct | **7.143** | 6.065 | 6.620 |
|
41 |
| Alpha-Ko-Instruct-marlin (4bit) | 6.857 | 5.738 | 6.298 |
|
42 |
|
|
|
44 |
|
45 |
Result in KoBEST(acc, num_shot=5) are as follows:
|
46 |
|
47 |
+
| Task | beomi/Llama-3-Open-Ko-8B-Instruct | maywell/Llama-3-Ko-8B-Instruct | **Alpha-Ko-Evo** | **Alpha-Ko-Instruct** |
|
48 |
| --- | --- | --- | --- | --- |
|
49 |
| kobest overall | 0.6220 | 0.6852 |0.7229|0.7055
|
50 |
| kobest_boolq| 0.6254 | 0.7208 | 0.8547 | 0.8369
|