Update README.md
Browse files
README.md
CHANGED
@@ -38,4 +38,18 @@ RLAIF-V maximally exploits the open-source feedback from two key perspectives, i
|
|
38 |
- **Trained on data:** [RLAIF-V-Dataset](https://huggingface.co/datasets/HaoyeZhang/RLAIF-V-Dataset)
|
39 |
|
40 |
## Usage
|
41 |
-
Please look at [GitHub](https://github.com/RLHF-V/RLAIF-V) for more details about usage.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
38 |
- **Trained on data:** [RLAIF-V-Dataset](https://huggingface.co/datasets/HaoyeZhang/RLAIF-V-Dataset)
|
39 |
|
40 |
## Usage
|
41 |
+
Please look at [GitHub](https://github.com/RLHF-V/RLAIF-V) for more details about usage.
|
42 |
+
|
43 |
+
|
44 |
+
## Citation
|
45 |
+
|
46 |
+
If you find our model/code/paper helpful, please consider cite our papers 📝 and star us ⭐️!
|
47 |
+
|
48 |
+
```bibtex
|
49 |
+
@article{yu2024rlaifv,
|
50 |
+
title={RLAIF-V: Aligning MLLMs through Open-Source AI Feedback for Super GPT-4V Trustworthiness},
|
51 |
+
author={Tianyu Yu and Haoye Zhang and Yuan Yao and Yunkai Dang and Da Chen and Xiaoman Lu and Ganqu Cui and Taiwen He and Zhiyuan Liu and Tat-Seng Chua and Maosong Sun},
|
52 |
+
journal={arXiv preprint arXiv:2405.17220},
|
53 |
+
year={2024},
|
54 |
+
}
|
55 |
+
```
|