Add official github repo
Browse files
README.md
CHANGED
@@ -162,6 +162,8 @@ model-index:
|
|
162 |
|
163 |
**Mistral-ORPO** is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) using the *odds ratio preference optimization (ORPO)*. With ORPO, the model directly learns the preference without the supervised fine-tuning warmup phase. **Mistral-ORPO-β** is fine-tuned exclusively on the 61k instances of the cleaned version of UltraFeedback, [argilla/ultrafeedback-binarized-preferences-cleaned](https://huggingface.co/datasets/argilla/ultrafeedback-binarized-preferences-cleaned), by [Argilla](https://huggingface.co/argilla).
|
164 |
|
|
|
|
|
165 |
## 👍 **Model Performance**
|
166 |
|
167 |
### 1) AlpacaEval & MT-Bench
|
|
|
162 |
|
163 |
**Mistral-ORPO** is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) using the *odds ratio preference optimization (ORPO)*. With ORPO, the model directly learns the preference without the supervised fine-tuning warmup phase. **Mistral-ORPO-β** is fine-tuned exclusively on the 61k instances of the cleaned version of UltraFeedback, [argilla/ultrafeedback-binarized-preferences-cleaned](https://huggingface.co/datasets/argilla/ultrafeedback-binarized-preferences-cleaned), by [Argilla](https://huggingface.co/argilla).
|
164 |
|
165 |
+
- **Github Repository**: https://github.com/xfactlab/orpo
|
166 |
+
|
167 |
## 👍 **Model Performance**
|
168 |
|
169 |
### 1) AlpacaEval & MT-Bench
|