Update README.md
Browse files
README.md
CHANGED
@@ -3,6 +3,14 @@ license: mit
|
|
3 |
---
|
4 |
**[ALMA-R](https://arxiv.org/abs/2401.08417)** builds upon [ALMA models](https://arxiv.org/abs/2309.11674), with further LoRA fine-tuning with our proposed **Contrastive Preference Optimization (CPO)** as opposed to the Supervised Fine-tuning used in ALMA. CPO fine-tuning requires our [triplet preference data](https://huggingface.co/datasets/haoranxu/ALMA-R-Preference) for preference learning. ALMA-R now can matches or even exceeds GPT-4 or WMT winners!
|
5 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6 |
|
7 |
# Download ALMA(-R) Models and Dataset 🚀
|
8 |
|
|
|
3 |
---
|
4 |
**[ALMA-R](https://arxiv.org/abs/2401.08417)** builds upon [ALMA models](https://arxiv.org/abs/2309.11674), with further LoRA fine-tuning with our proposed **Contrastive Preference Optimization (CPO)** as opposed to the Supervised Fine-tuning used in ALMA. CPO fine-tuning requires our [triplet preference data](https://huggingface.co/datasets/haoranxu/ALMA-R-Preference) for preference learning. ALMA-R now can matches or even exceeds GPT-4 or WMT winners!
|
5 |
|
6 |
+
@misc{xu2024contrastive,
|
7 |
+
title={Contrastive Preference Optimization: Pushing the Boundaries of LLM Performance in Machine Translation},
|
8 |
+
author={Haoran Xu and Amr Sharaf and Yunmo Chen and Weiting Tan and Lingfeng Shen and Benjamin Van Durme and Kenton Murray and Young Jin Kim},
|
9 |
+
year={2024},
|
10 |
+
eprint={2401.08417},
|
11 |
+
archivePrefix={arXiv},
|
12 |
+
primaryClass={cs.CL}
|
13 |
+
}
|
14 |
|
15 |
# Download ALMA(-R) Models and Dataset 🚀
|
16 |
|