JW17 commited on
Commit
e74a696
1 Parent(s): 382418f

Add official github repo

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -162,6 +162,8 @@ model-index:
162
 
163
  **Mistral-ORPO** is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) using the *odds ratio preference optimization (ORPO)*. With ORPO, the model directly learns the preference without the supervised fine-tuning warmup phase. **Mistral-ORPO-β** is fine-tuned exclusively on the 61k instances of the cleaned version of UltraFeedback, [argilla/ultrafeedback-binarized-preferences-cleaned](https://huggingface.co/datasets/argilla/ultrafeedback-binarized-preferences-cleaned), by [Argilla](https://huggingface.co/argilla).
164
 
 
 
165
  ## 👍 **Model Performance**
166
 
167
  ### 1) AlpacaEval & MT-Bench
 
162
 
163
  **Mistral-ORPO** is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) using the *odds ratio preference optimization (ORPO)*. With ORPO, the model directly learns the preference without the supervised fine-tuning warmup phase. **Mistral-ORPO-β** is fine-tuned exclusively on the 61k instances of the cleaned version of UltraFeedback, [argilla/ultrafeedback-binarized-preferences-cleaned](https://huggingface.co/datasets/argilla/ultrafeedback-binarized-preferences-cleaned), by [Argilla](https://huggingface.co/argilla).
164
 
165
+ - **Github Repository**: https://github.com/xfactlab/orpo
166
+
167
  ## 👍 **Model Performance**
168
 
169
  ### 1) AlpacaEval & MT-Bench