Image-Text-to-Text
Transformers
Safetensors
English
MLLM
Inference Endpoints
liyang commited on
Commit
549533d
1 Parent(s): 699a22a

update readme

Browse files
Files changed (1) hide show
  1. README.md +37 -3
README.md CHANGED
@@ -1,3 +1,37 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Model Card
2
+ <!-- Provide a quick summary of what the model is/does. -->
3
+ Parrot is a multi-language and multi-modal large language model capable of achieving excellent performance.
4
+ For a comprehensive introduction, please refer to [Parrot Paper](https://arxiv.org/abs/2406.02539) and [Parrot GitHub](https://github.com/AIDC-AI/Parrot).
5
+
6
+ # Model Details
7
+ ![](https://github.com/AIDC-AI/Parrot/images/teaser.png)
8
+
9
+ # Performance
10
+ ![](https://github.com/AIDC-AI/Parrot/images/teaser.png)
11
+
12
+ # Usage
13
+
14
+ Below is a code snippet to run Parrot with multimodal inputs. For additional usage instructions, including inference wrapper and Gradio UI, please refer to [Parrot GitHub](https://github.com/AIDC-AI/Parrot).
15
+ ```markdown
16
+ pip install torch==2.1.2 transformers==4.43.2 pillow==10.3.0
17
+ ```
18
+ ```python
19
+ import torch
20
+ from PIL import Image
21
+ from transformers import AutoModelForCausalLM
22
+ ```
23
+
24
+ # Citation
25
+ If you find Parrot useful, please cite the paper
26
+
27
+ ```markdown
28
+ @article{sun2024parrot,
29
+ title={Parrot: Multilingual Visual Instruction Tuning},
30
+ author={Sun, Hai-Long and Zhou, Da-Wei and Li, Yang and Lu, Shiyin and Yi, Chao and Chen, Qing-Guo and Xu, Zhao and Luo, Weihua and Zhang, Kaifu and Zhan, De-Chuan and others},
31
+ journal={arXiv preprint arXiv:2406.02539},
32
+ year={2024}
33
+ }
34
+ ```
35
+
36
+ # License
37
+ The project is licensed under Apache License Version 2.0 and is restricted to uses that comply with the license agreements of Qwen and Clip.