Update README.md
Browse files
README.md
CHANGED
@@ -97,13 +97,13 @@ while True:
|
|
97 |
## Evaluations
|
98 |
The following data has been re-evaluated and calculated as the average for each test.
|
99 |
|
100 |
-
| Benchmark | Qwen2.5-7B-Instruct |
|
101 |
-
|
102 |
-
| IF_Eval | 76.44 | **76.49**
|
103 |
-
| MMLU Pro | 43.12 | 41.71
|
104 |
-
| TruthfulQA | 62.46 | **64.92**
|
105 |
-
| BBH | 53.92 | 52.77
|
106 |
-
| GPQA | 31.91 | **31.97**
|
107 |
|
108 |
The script used for evaluation can be found inside this repository under /eval.sh, or click [here](https://huggingface.co/huihui-ai/Qwen2.5-7B-Instruct-abliterated/blob/main/eval.sh)
|
109 |
|
|
|
97 |
## Evaluations
|
98 |
The following data has been re-evaluated and calculated as the average for each test.
|
99 |
|
100 |
+
| Benchmark | Qwen2.5-7B-Instruct | Qwen2.5-7B-Instruct-abliterated |
|
101 |
+
|-------------|---------------------|---------------------------------|
|
102 |
+
| IF_Eval | 76.44 | **76.49** |
|
103 |
+
| MMLU Pro | 43.12 | 41.71 |
|
104 |
+
| TruthfulQA | 62.46 | **64.92** |
|
105 |
+
| BBH | 53.92 | 52.77 |
|
106 |
+
| GPQA | 31.91 | **31.97** |
|
107 |
|
108 |
The script used for evaluation can be found inside this repository under /eval.sh, or click [here](https://huggingface.co/huihui-ai/Qwen2.5-7B-Instruct-abliterated/blob/main/eval.sh)
|
109 |
|