Update README.md
Browse files
README.md
CHANGED
@@ -26,25 +26,6 @@ We release InternViT-6B-448px-V1-0, which is integrated into [InternVL-Chat-V1-1
|
|
26 |
- **Pretrain Dataset:** LAION-en, LAION-COCO, COYO, CC12M, CC3M, SBU, Wukong, LAION-multi, OCR-related datasets.
|
27 |
- **Note:** This model has 48 blocks, and we found that using the output after the fourth-to-last block worked best for MLLM. Therefore, when building a MLLM with this model, **please use the features from the fourth-to-last layer.**
|
28 |
|
29 |
-
## Released Models
|
30 |
-
### Vision Foundation model
|
31 |
-
| Model | Date | Download | Note |
|
32 |
-
| ----------------------- | ---------- | ---------------------------------------------------------------------- | -------------------------------- |
|
33 |
-
| InternViT-6B-448px-V1-5 | 2024.04.20 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternViT-6B-448px-V1-5) | support dynamic resolution, super strong OCR (🔥new) |
|
34 |
-
| InternViT-6B-448px-V1-2 | 2024.02.11 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternViT-6B-448px-V1-2) | 448 resolution |
|
35 |
-
| InternViT-6B-448px-V1-0 | 2024.01.30 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternViT-6B-448px-V1-0) | 448 resolution |
|
36 |
-
| InternViT-6B-224px | 2023.12.22 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternViT-6B-224px) | vision foundation model |
|
37 |
-
| InternVL-14B-224px | 2023.12.22 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-14B-224px) | vision-language foundation model |
|
38 |
-
|
39 |
-
### Multimodal Large Language Model (MLLM)
|
40 |
-
| Model | Date | Download | Note |
|
41 |
-
| ----------------------- | ---------- | --------------------------------------------------------------------------- | ---------------------------------- |
|
42 |
-
| InternVL-Chat-V1-5 | 2024.04.18 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-5) | support 4K image; super strong OCR; Approaching the performance of GPT-4V and Gemini Pro on various benchmarks like MMMU, DocVQA, ChartQA, MathVista, etc. (🔥new)|
|
43 |
-
| InternVL-Chat-V1-2-Plus | 2024.02.21 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-2-Plus) | more SFT data and stronger |
|
44 |
-
| InternVL-Chat-V1-2 | 2024.02.11 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-2) | scaling up LLM to 34B |
|
45 |
-
| InternVL-Chat-V1-1 | 2024.01.24 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-1) | support Chinese and stronger OCR |
|
46 |
-
|
47 |
-
|
48 |
## Model Usage (Image Embeddings)
|
49 |
|
50 |
```python
|
|
|
26 |
- **Pretrain Dataset:** LAION-en, LAION-COCO, COYO, CC12M, CC3M, SBU, Wukong, LAION-multi, OCR-related datasets.
|
27 |
- **Note:** This model has 48 blocks, and we found that using the output after the fourth-to-last block worked best for MLLM. Therefore, when building a MLLM with this model, **please use the features from the fourth-to-last layer.**
|
28 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
29 |
## Model Usage (Image Embeddings)
|
30 |
|
31 |
```python
|