haoranxu commited on
Commit
ec7e8c8
•
1 Parent(s): a72f26f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -0
README.md CHANGED
@@ -3,6 +3,14 @@ license: mit
3
  ---
4
  **[ALMA-R](https://arxiv.org/abs/2401.08417)** builds upon [ALMA models](https://arxiv.org/abs/2309.11674), with further LoRA fine-tuning with our proposed **Contrastive Preference Optimization (CPO)** as opposed to the Supervised Fine-tuning used in ALMA. CPO fine-tuning requires our [triplet preference data](https://huggingface.co/datasets/haoranxu/ALMA-R-Preference) for preference learning. ALMA-R now can matches or even exceeds GPT-4 or WMT winners!
5
 
 
 
 
 
 
 
 
 
6
 
7
  # Download ALMA(-R) Models and Dataset 🚀
8
 
 
3
  ---
4
  **[ALMA-R](https://arxiv.org/abs/2401.08417)** builds upon [ALMA models](https://arxiv.org/abs/2309.11674), with further LoRA fine-tuning with our proposed **Contrastive Preference Optimization (CPO)** as opposed to the Supervised Fine-tuning used in ALMA. CPO fine-tuning requires our [triplet preference data](https://huggingface.co/datasets/haoranxu/ALMA-R-Preference) for preference learning. ALMA-R now can matches or even exceeds GPT-4 or WMT winners!
5
 
6
+ @misc{xu2024contrastive,
7
+ title={Contrastive Preference Optimization: Pushing the Boundaries of LLM Performance in Machine Translation},
8
+ author={Haoran Xu and Amr Sharaf and Yunmo Chen and Weiting Tan and Lingfeng Shen and Benjamin Van Durme and Kenton Murray and Young Jin Kim},
9
+ year={2024},
10
+ eprint={2401.08417},
11
+ archivePrefix={arXiv},
12
+ primaryClass={cs.CL}
13
+ }
14
 
15
  # Download ALMA(-R) Models and Dataset 🚀
16