klldmofashi commited on
Commit
428e95e
·
verified ·
1 Parent(s): c8c45c0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -1
README.md CHANGED
@@ -40,7 +40,7 @@ https://github.com/NVLabs/VILA
40
  - [Dataset Licenses](https://github.com/Efficient-Large-Model/VILA/blob/main/data_prepare/LICENSE) for each one used during training.
41
 
42
  **Where to send questions or comments about the model:**
43
- https://github.com/Efficient-Large-Model/VILA/issues
44
 
45
  ## Intended use
46
  **Primary intended uses:**
@@ -49,6 +49,40 @@ The primary use of VILA is research on large multimodal models and chatbots.
49
  **Primary intended users:**
50
  The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.
51
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
52
  ## Training dataset
53
  See [Dataset Preparation](https://github.com/Efficient-Large-Model/VILA/blob/main/data_prepare/README.md) for more details.
54
 
 
40
  - [Dataset Licenses](https://github.com/Efficient-Large-Model/VILA/blob/main/data_prepare/LICENSE) for each one used during training.
41
 
42
  **Where to send questions or comments about the model:**
43
+ https://github.com/NVLabs/VILA/issues
44
 
45
  ## Intended use
46
  **Primary intended uses:**
 
49
  **Primary intended users:**
50
  The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.
51
 
52
+ ## Model Architecture:
53
+ **Architecture Type:** Transformer
54
+ **Network Architecture:** InternViT, Yi
55
+
56
+ ## Input:
57
+ **Input Type:** Image, Video, Text
58
+ **Input Format:** Red, Green, Blue; MP4 ;String
59
+ **Input Parameters:** 2D, 3D
60
+
61
+ ## Output:
62
+ **Output Type:** Text
63
+ **Output Format:** String
64
+
65
+ **Supported Hardware Microarchitecture Compatibility:**
66
+ * Ampere
67
+ * Jetson
68
+ * Hopper
69
+ * Lovelace
70
+
71
+ **[Preferred/Supported] Operating System(s):** <br>
72
+ Linux
73
+
74
+ ## Model Version(s):
75
+ VILA1.5-3B
76
+ VILA1.5-3B-s2
77
+ Llama-3-VILA1.5-8B
78
+ VILA1.5-13B
79
+ VILA1.5-40B
80
+ VILA1.5-3B-AWQ
81
+ VILA1.5-3B-s2-AWQ
82
+ Llama-3-VILA1.5-8B-AWQ
83
+ VILA1.5-13B-AWQ
84
+ VILA1.5-40B-AWQ
85
+
86
  ## Training dataset
87
  See [Dataset Preparation](https://github.com/Efficient-Large-Model/VILA/blob/main/data_prepare/README.md) for more details.
88