shikiw commited on
Commit
f0aaa5f
1 Parent(s): 55435ed

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -1
README.md CHANGED
@@ -3,4 +3,24 @@ license: llama2
3
  language:
4
  - en
5
  - zh
6
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  language:
4
  - en
5
  - zh
6
+ tags:
7
+ - multimodal
8
+ datasets:
9
+ - liuhaotian/LLaVA-Pretrain
10
+ base_model:
11
+ - lmsys/vicuna-7b-v1.5
12
+ pipeline_tag: image-text-to-text
13
+ library_name: transformers
14
+ ---
15
+
16
+
17
+ ## **Citation**
18
+ If you find this model useful, please cite the following paper
19
+ ```
20
+ @article{huang2024deciphering,
21
+ title={Deciphering Cross-Modal Alignment in Large Vision-Language Models with Modality Integration Rate},
22
+ author={Huang, Qidong and Dong, Xiaoyi and Zhang, Pan and Zang, Yuhang and Cao, Yuhang and Wang, Jiaqi and Lin, Dahua and Zhang, Weiming and Yu, Nenghai},
23
+ journal={arXiv preprint arXiv:2410.07167},
24
+ year={2024}
25
+ }
26
+ ```