Update README.md
Browse files
README.md
CHANGED
@@ -40,7 +40,7 @@ https://github.com/NVLabs/VILA
|
|
40 |
- [Dataset Licenses](https://github.com/Efficient-Large-Model/VILA/blob/main/data_prepare/LICENSE) for each one used during training.
|
41 |
|
42 |
**Where to send questions or comments about the model:**
|
43 |
-
https://github.com/
|
44 |
|
45 |
## Intended use
|
46 |
**Primary intended uses:**
|
@@ -49,6 +49,40 @@ The primary use of VILA is research on large multimodal models and chatbots.
|
|
49 |
**Primary intended users:**
|
50 |
The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.
|
51 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
52 |
## Training dataset
|
53 |
See [Dataset Preparation](https://github.com/Efficient-Large-Model/VILA/blob/main/data_prepare/README.md) for more details.
|
54 |
|
|
|
40 |
- [Dataset Licenses](https://github.com/Efficient-Large-Model/VILA/blob/main/data_prepare/LICENSE) for each one used during training.
|
41 |
|
42 |
**Where to send questions or comments about the model:**
|
43 |
+
https://github.com/NVLabs/VILA/issues
|
44 |
|
45 |
## Intended use
|
46 |
**Primary intended uses:**
|
|
|
49 |
**Primary intended users:**
|
50 |
The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.
|
51 |
|
52 |
+
## Model Architecture:
|
53 |
+
**Architecture Type:** Transformer
|
54 |
+
**Network Architecture:** InternViT, Yi
|
55 |
+
|
56 |
+
## Input:
|
57 |
+
**Input Type:** Image, Video, Text
|
58 |
+
**Input Format:** Red, Green, Blue; MP4 ;String
|
59 |
+
**Input Parameters:** 2D, 3D
|
60 |
+
|
61 |
+
## Output:
|
62 |
+
**Output Type:** Text
|
63 |
+
**Output Format:** String
|
64 |
+
|
65 |
+
**Supported Hardware Microarchitecture Compatibility:**
|
66 |
+
* Ampere
|
67 |
+
* Jetson
|
68 |
+
* Hopper
|
69 |
+
* Lovelace
|
70 |
+
|
71 |
+
**[Preferred/Supported] Operating System(s):** <br>
|
72 |
+
Linux
|
73 |
+
|
74 |
+
## Model Version(s):
|
75 |
+
VILA1.5-3B
|
76 |
+
VILA1.5-3B-s2
|
77 |
+
Llama-3-VILA1.5-8B
|
78 |
+
VILA1.5-13B
|
79 |
+
VILA1.5-40B
|
80 |
+
VILA1.5-3B-AWQ
|
81 |
+
VILA1.5-3B-s2-AWQ
|
82 |
+
Llama-3-VILA1.5-8B-AWQ
|
83 |
+
VILA1.5-13B-AWQ
|
84 |
+
VILA1.5-40B-AWQ
|
85 |
+
|
86 |
## Training dataset
|
87 |
See [Dataset Preparation](https://github.com/Efficient-Large-Model/VILA/blob/main/data_prepare/README.md) for more details.
|
88 |
|