leaderboard-pr-bot commited on
Commit
0976d03
1 Parent(s): de6fa19

Adding Evaluation Results

Browse files

This is an automated PR created with https://huggingface.co/spaces/Weyaxi/open-llm-leaderboard-results-pr

The purpose of this PR is to add evaluation results from the Open LLM Leaderboard to your model card.

If you encounter any issues, please report them to https://huggingface.co/spaces/Weyaxi/open-llm-leaderboard-results-pr/discussions

Files changed (1) hide show
  1. README.md +21 -8
README.md CHANGED
@@ -1,6 +1,8 @@
1
  ---
 
 
2
  license: cc-by-nc-4.0
3
- base_model: mlabonne/NeuralMonarch-7B
4
  tags:
5
  - generated_from_trainer
6
  - mistral
@@ -10,15 +12,13 @@ tags:
10
  - gpt4
11
  - synthetic data
12
  - distillation
13
- model-index:
14
- - name: AlphaMonarch-dora
15
- results: []
16
  datasets:
17
  - argilla/OpenHermes2.5-dpo-binarized-alpha
18
- language:
19
- - en
20
- library_name: transformers
21
  pipeline_tag: text-generation
 
 
 
22
  ---
23
  # AlphaMonarch-dora
24
 
@@ -93,4 +93,17 @@ The following hyperparameters were used during training:
93
  - Peft 0.9.1.dev0
94
  - Datasets 2.18.0
95
  - torch 2.2.0
96
- - accelerate 0.27.2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
  license: cc-by-nc-4.0
5
+ library_name: transformers
6
  tags:
7
  - generated_from_trainer
8
  - mistral
 
12
  - gpt4
13
  - synthetic data
14
  - distillation
 
 
 
15
  datasets:
16
  - argilla/OpenHermes2.5-dpo-binarized-alpha
17
+ base_model: mlabonne/NeuralMonarch-7B
 
 
18
  pipeline_tag: text-generation
19
+ model-index:
20
+ - name: AlphaMonarch-dora
21
+ results: []
22
  ---
23
  # AlphaMonarch-dora
24
 
 
93
  - Peft 0.9.1.dev0
94
  - Datasets 2.18.0
95
  - torch 2.2.0
96
+ - accelerate 0.27.2
97
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
98
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_abideen__AlphaMonarch-dora)
99
+
100
+ | Metric |Value|
101
+ |---------------------------------|----:|
102
+ |Avg. |75.86|
103
+ |AI2 Reasoning Challenge (25-Shot)|73.21|
104
+ |HellaSwag (10-Shot) |89.26|
105
+ |MMLU (5-Shot) |64.47|
106
+ |TruthfulQA (0-shot) |78.02|
107
+ |Winogrande (5-shot) |84.45|
108
+ |GSM8k (5-shot) |65.73|
109
+