Commit
•
e90de7f
1
Parent(s):
897c938
Update README.md (#9)
Browse files- Update README.md (612c109000bdcc2ac2f704ebed714de7a655a45e)
Co-authored-by: Gloria Lee <[email protected]>
README.md
CHANGED
@@ -54,6 +54,7 @@ license_link: LICENSE
|
|
54 |
- [Training](#training)
|
55 |
- [Limitations](#limitations)
|
56 |
- [Why Yi-VL?](#why-yi-vl)
|
|
|
57 |
- [Benchmarks](#benchmarks)
|
58 |
- [Showcases](#showcases)
|
59 |
- [How to use Yi-VL?](#how-to-use-yi-vl)
|
@@ -159,7 +160,23 @@ This is the initial release of the Yi-VL, which comes with some known limitation
|
|
159 |
- Other limitations of the Yi LLM.
|
160 |
|
161 |
# Why Yi-VL?
|
162 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
163 |
## Benchmarks
|
164 |
|
165 |
Yi-VL outperforms all existing open-source models in [MMMU](https://mmmu-benchmark.github.io) and [CMMMU](https://cmmmu-benchmark.github.io/), two advanced benchmarks that include massive multi-discipline multimodal questions (based on data available up to January 2024).
|
|
|
54 |
- [Training](#training)
|
55 |
- [Limitations](#limitations)
|
56 |
- [Why Yi-VL?](#why-yi-vl)
|
57 |
+
- [Tech report](#tech-report)
|
58 |
- [Benchmarks](#benchmarks)
|
59 |
- [Showcases](#showcases)
|
60 |
- [How to use Yi-VL?](#how-to-use-yi-vl)
|
|
|
160 |
- Other limitations of the Yi LLM.
|
161 |
|
162 |
# Why Yi-VL?
|
163 |
+
|
164 |
+
## Tech report
|
165 |
+
|
166 |
+
For detailed capabilities of the Yi series model, see [Yi: Open Foundation Models by 01.AI](https://arxiv.org/abs/2403.04652).
|
167 |
+
|
168 |
+
### Citation
|
169 |
+
```
|
170 |
+
@misc{ai2024yi,
|
171 |
+
title={Yi: Open Foundation Models by 01.AI},
|
172 |
+
author={01. AI and : and Alex Young and Bei Chen and Chao Li and Chengen Huang and Ge Zhang and Guanwei Zhang and Heng Li and Jiangcheng Zhu and Jianqun Chen and Jing Chang and Kaidong Yu and Peng Liu and Qiang Liu and Shawn Yue and Senbin Yang and Shiming Yang and Tao Yu and Wen Xie and Wenhao Huang and Xiaohui Hu and Xiaoyi Ren and Xinyao Niu and Pengcheng Nie and Yuchi Xu and Yudong Liu and Yue Wang and Yuxuan Cai and Zhenyu Gu and Zhiyuan Liu and Zonghong Dai},
|
173 |
+
year={2024},
|
174 |
+
eprint={2403.04652},
|
175 |
+
archivePrefix={arXiv},
|
176 |
+
primaryClass={cs.CL}
|
177 |
+
}
|
178 |
+
```
|
179 |
+
|
180 |
## Benchmarks
|
181 |
|
182 |
Yi-VL outperforms all existing open-source models in [MMMU](https://mmmu-benchmark.github.io) and [CMMMU](https://cmmmu-benchmark.github.io/), two advanced benchmarks that include massive multi-discipline multimodal questions (based on data available up to January 2024).
|