Update README.md
Browse files
README.md
CHANGED
@@ -10,7 +10,7 @@ tags:
|
|
10 |
|
11 |
# 🦜 EmertonMonarch-7B
|
12 |
|
13 |
-
EmertonOmniBeagle-7B-dpo is a DPO fine-tune of [mlabonne/Monarch-7B](https://huggingface.co/mlabonne/OmniBeagle-7B) using the [yleo/emerton_dpo_pairs_judge](https://huggingface.co/datasets/yleo/emerton_dpo_pairs_judge) preference dataset created from [Intel/orca_dpo_pairs](https://huggingface.co/datasets/Intel/orca_dpo_pairs) by replacing gpt 3.5 answer by a gpt4 Turbo answer. Then,
|
14 |
|
15 |
## 🔍 Applications
|
16 |
|
|
|
10 |
|
11 |
# 🦜 EmertonMonarch-7B
|
12 |
|
13 |
+
EmertonOmniBeagle-7B-dpo is a DPO fine-tune of [mlabonne/Monarch-7B](https://huggingface.co/mlabonne/OmniBeagle-7B) using the [yleo/emerton_dpo_pairs_judge](https://huggingface.co/datasets/yleo/emerton_dpo_pairs_judge) preference dataset created from [Intel/orca_dpo_pairs](https://huggingface.co/datasets/Intel/orca_dpo_pairs) by replacing gpt 3.5 answer by a gpt4 Turbo answer. Then, LLM-Blender is used to judge between GPT4 and GPT4 Turbo.
|
14 |
|
15 |
## 🔍 Applications
|
16 |
|