Llama-2-7b-orca-v1 / README.md
leaderboard-pr-bot's picture
Adding Evaluation Results
c1ac949
|
raw
history blame
864 Bytes
metadata
license: mit
datasets:
  - Open-Orca/OpenOrca
language:
  - en
library_name: transformers
pipeline_tag: text-generation

img

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 48.17
ARC (25-shot) 56.31
HellaSwag (10-shot) 79.14
MMLU (5-shot) 52.71
TruthfulQA (0-shot) 50.19
Winogrande (5-shot) 75.22
GSM8K (5-shot) 7.81
DROP (3-shot) 15.81