Commit
•
976b0b8
1
Parent(s):
c767ad8
Update README.md
Browse files
README.md
CHANGED
@@ -3,6 +3,8 @@ language:
|
|
3 |
- en
|
4 |
license: llama2
|
5 |
pipeline_tag: image-text-to-text
|
|
|
|
|
6 |
---
|
7 |
|
8 |
# LLaVA-NeXT-Video Model Card
|
@@ -18,7 +20,7 @@ LLaVA-Next-Video is an open-source chatbot trained by fine-tuning LLM on multimo
|
|
18 |
The model is a current SOTA among open-source models on [VideoMME bench](https://arxiv.org/abs/2405.21075).
|
19 |
Base LLM: [lmsys/vicuna-7b-v1.5](https://huggingface.co/lmsys/vicuna-13b-v1.5)
|
20 |
|
21 |
-
|
22 |
|
23 |
|
24 |
**Model date:**
|
@@ -231,5 +233,4 @@ If you find our paper and code useful in your research:
|
|
231 |
month={January},
|
232 |
year={2024}
|
233 |
}
|
234 |
-
```
|
235 |
-
|
|
|
3 |
- en
|
4 |
license: llama2
|
5 |
pipeline_tag: image-text-to-text
|
6 |
+
datasets:
|
7 |
+
- lmms-lab/VideoChatGPT
|
8 |
---
|
9 |
|
10 |
# LLaVA-NeXT-Video Model Card
|
|
|
20 |
The model is a current SOTA among open-source models on [VideoMME bench](https://arxiv.org/abs/2405.21075).
|
21 |
Base LLM: [lmsys/vicuna-7b-v1.5](https://huggingface.co/lmsys/vicuna-13b-v1.5)
|
22 |
|
23 |
+
![llava_next_video_arch](demo.png)
|
24 |
|
25 |
|
26 |
**Model date:**
|
|
|
233 |
month={January},
|
234 |
year={2024}
|
235 |
}
|
236 |
+
```
|
|