Breeze-Petro-7B-Instruct-v1
- Model creator: MediaTek Research
- Original model: MediaTek-Research/Breeze-7B-Instruct-v1_0
Model Description
This is a model fine-tuned based on Breeze-7B-Instruct-v1_0. The training set is mainly based on chemical knowledge and procedural knowledge. Supplement knowledge about the petroleum industry.
- Developed by: RebeccaChou
- License: apache-2.0
- Finetuned from model : [MediaTek-Research/Breeze-7B-Instruct-v1_0
- Language(s) (NLP): [English.ηΉι«δΈζ]
π Table of Contents
1.Open LLM Leaderboard - ARC - HellaSwag - MMLU - TruthfulQA - Winogrande - GSM8K 3. EvalPlus Leaderboard - HumanEval - HumanEval_Plus - MBPP - MBPP_Plus 4. Prompt Format 5. Quantized Models 6. Gratitude
π Open LLM Leaderboard
WestSeverus-7B-DPO-v2 is one of the top 7B model in Open LLM Leaderboard and it outperforms on TruthfulQA and GSM8K.
Metric | Value |
---|---|
Avg. | 59.32 |
AI2 Reasoning Challenge (25-Shot) | 58.87 |
HellaSwag (10-Shot) | 79.17 |
MMLU (5-Shot) | 56.62 |
TruthfulQA (0-shot) | 46.36 |
Winogrande (5-shot) | 73.64 |
GSM8k (5-shot) | 41.24 |
Detailed results can be found here
β‘ EvalPlus Leaderboard
Model | HumanEval | HumanEval_Plus | MBPP | MBPP_Plus |
---|---|---|---|---|
phi-2-2.7B | 48.2 | 43.3 | 61.9 | 51.4 |
SOLAR-10.7B-Instruct-v1.0 | 42.1 | 34.3 | 42.9 | 34.6 |
CodeLlama-7B | 37.8 | 34.1 | 57.6 | 45.4 |
π οΈ Quantized Models
Training Details
Training Data
- Dataset: Rebecca19990101/petro-dataset-v2
- Downloads last month
- 12
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.