Update README.md
Browse files
README.md
CHANGED
@@ -26,7 +26,7 @@ It is _**the largest open-source vision/vision-language foundation model (14B)**
|
|
26 |
- **Model Stats:**
|
27 |
- Params (M): 5903
|
28 |
- Image size: 448 x 448
|
29 |
-
- **Pretrain Dataset:** LAION-en, LAION-COCO, COYO, CC12M, CC3M, SBU, Wukong, LAION-multi
|
30 |
- **Note:** This model has 48 blocks, and we found that using the output after the fourth-to-last block worked best for VLLM. Therefore, **please set mm_vision_select_layer=-4 when using this model to build VLLM.**
|
31 |
|
32 |
## Model Usage (Image Embeddings)
|
|
|
26 |
- **Model Stats:**
|
27 |
- Params (M): 5903
|
28 |
- Image size: 448 x 448
|
29 |
+
- **Pretrain Dataset:** LAION-en, LAION-COCO, COYO, CC12M, CC3M, SBU, Wukong, LAION-multi, OCR data
|
30 |
- **Note:** This model has 48 blocks, and we found that using the output after the fourth-to-last block worked best for VLLM. Therefore, **please set mm_vision_select_layer=-4 when using this model to build VLLM.**
|
31 |
|
32 |
## Model Usage (Image Embeddings)
|