Zery
/

Image-Text-to-Text
Transformers
PyTorch
English
share4v
text-generation
Zery commited on
Commit
2c87ec5
·
verified ·
1 Parent(s): d6316c5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -7
README.md CHANGED
@@ -1,9 +1,34 @@
1
  ---
2
- license: apache-2.0
3
- datasets:
4
- - Zery/BS-Objaverse
5
- - Lin-Chen/ShareGPT4V
6
- language:
7
- - en
8
  pipeline_tag: image-text-to-text
9
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ inference: false
 
 
 
 
 
3
  pipeline_tag: image-text-to-text
4
+ ---
5
+ <br>
6
+ <br>
7
+
8
+ # ShareGPT4V-7B Model Card
9
+
10
+ ## Model details
11
+
12
+ **Model type:**
13
+ MV-LLaVA-7B is an open-source chatbot for 3D multi-view images trained by fine-tuning CLIP vision tower and LLaMA/Vicuna on GPT4-Vision-assisted [BS-Objaverse](https://huggingface.co/datasets/Zery/BS-Objaverse) data and [ShareGPT4V](https://huggingface.co/datasets/Lin-Chen/ShareGPT4V) data.
14
+
15
+ **Model date:**
16
+ MV-LLaVA-7B was trained in Apr, 2024.
17
+
18
+ **Paper or resources for more information:**
19
+ [[Project](https://sunzey.github.io/Bootstrap3D/)] [[Paper](https://huggingface.co/papers/2311.12793)] [[Code](https://github.com/SunzeY/Bootstrap3D)]
20
+
21
+ ## Usage
22
+ You can directly utilize this model as we provide in our [[repository](https://github.com/SunzeY/Bootstrap3D/tree/main/MV_LLaVA)].
23
+ ## License
24
+ Llama 2 is licensed under the LLAMA 2 Community License,
25
+ Copyright (c) Meta Platforms, Inc. All Rights Reserved.
26
+ ## Intended use
27
+ **Primary intended uses:**
28
+ The primary use of ShareGPT4V-7B is research on large multimodal models and chatbots for 3D content.
29
+ **Primary intended users:**
30
+ The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.
31
+ ## Training dataset
32
+ - 1.2M ShareGPT4V-PT data
33
+ - 30K GPT4-Vision-generated multi-view image-text pairs
34
+ - LLaVA instruction-tuning data