MiquMaid-v2-70B / README.md
leaderboard-pr-bot's picture
Adding Evaluation Results
115ad9f verified
|
raw
history blame
5.12 kB
---
license: cc-by-nc-4.0
tags:
- not-for-all-audiences
- nsfw
model-index:
- name: MiquMaid-v2-70B
results:
- task:
type: text-generation
name: Text Generation
dataset:
name: AI2 Reasoning Challenge (25-Shot)
type: ai2_arc
config: ARC-Challenge
split: test
args:
num_few_shot: 25
metrics:
- type: acc_norm
value: 70.48
name: normalized accuracy
source:
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=NeverSleep/MiquMaid-v2-70B
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: HellaSwag (10-Shot)
type: hellaswag
split: validation
args:
num_few_shot: 10
metrics:
- type: acc_norm
value: 87.49
name: normalized accuracy
source:
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=NeverSleep/MiquMaid-v2-70B
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: MMLU (5-Shot)
type: cais/mmlu
config: all
split: test
args:
num_few_shot: 5
metrics:
- type: acc
value: 75.18
name: accuracy
source:
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=NeverSleep/MiquMaid-v2-70B
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: TruthfulQA (0-shot)
type: truthful_qa
config: multiple_choice
split: validation
args:
num_few_shot: 0
metrics:
- type: mc2
value: 57.62
source:
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=NeverSleep/MiquMaid-v2-70B
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: Winogrande (5-shot)
type: winogrande
config: winogrande_xl
split: validation
args:
num_few_shot: 5
metrics:
- type: acc
value: 84.77
name: accuracy
source:
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=NeverSleep/MiquMaid-v2-70B
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: GSM8k (5-shot)
type: gsm8k
config: main
split: test
args:
num_few_shot: 5
metrics:
- type: acc
value: 15.62
name: accuracy
source:
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=NeverSleep/MiquMaid-v2-70B
name: Open LLM Leaderboard
---
## MiquMaid v2
Check out our blogpost about this model series [Here!](https://ikaridevgit.github.io/index.html?blog=blogid-6&bo=true#Miqu-base) - Join our Discord server [Here!](https://discord.gg/Bb8pRUXy3Z)
<center>[<a href="https://huggingface.co/NeverSleep/MiquMaid-v2-70B">V2-70B</a> - <a href="https://huggingface.co/NeverSleep/MiquMaid-v2-70B-DPO">V2-70B-DPO</a> - <a href="https://huggingface.co/NeverSleep/MiquMaid-v2-2x70B">V2-2x70B</a> - <a href="https://huggingface.co/NeverSleep/MiquMaid-v2-2x70B-DPO">V2-2x70B-DPO</a>]
</br>
<div style="width: 100%;">
<img src="https://cdn-uploads.huggingface.co/production/uploads/63ab1241ad514ca8d1430003/1ziZ_5Rl7bsQKFrgH6ZCf.png" style="display: block; margin: auto;">
</div></center>
This model uses the Alpaca **prompting format**
Model trained for RP conversation on Miqu-70B with our magic sauce.
## Credits:
- Undi
- IkariDev
## Description
This repo contains FP16 files of MiquMaid-v2-70B.
Switch: [FP16](https://huggingface.co/NeverSleep/MiquMaid-v2-70B) - [GGUF](https://huggingface.co/NeverSleep/MiquMaid-v2-70B-GGUF)
## Training data used:
- [Aesir datasets](https://huggingface.co/MinervaAI)
- [NoRobots](https://huggingface.co/datasets/Doctor-Shotgun/no-robots-sharegpt)
- [limarp](https://huggingface.co/datasets/lemonilia/LimaRP)
- [toxic-dpo-v0.1-sharegpt](https://huggingface.co/datasets/Undi95/toxic-dpo-v0.1-sharegpt)
- [ToxicQAFinal](https://huggingface.co/datasets/NobodyExistsOnTheInternet/ToxicQAFinal)
### Custom format:
```
### Instruction:
{system prompt}
### Input:
{input}
### Response:
{reply}
```
## Others
Undi: If you want to support us, you can [here](https://ko-fi.com/undiai).
IkariDev: Visit my [retro/neocities style website](https://ikaridevgit.github.io/) please kek
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_NeverSleep__MiquMaid-v2-70B)
| Metric |Value|
|---------------------------------|----:|
|Avg. |65.19|
|AI2 Reasoning Challenge (25-Shot)|70.48|
|HellaSwag (10-Shot) |87.49|
|MMLU (5-Shot) |75.18|
|TruthfulQA (0-shot) |57.62|
|Winogrande (5-shot) |84.77|
|GSM8k (5-shot) |15.62|