czczup commited on
Commit
101bc9b
·
verified ·
1 Parent(s): 24c2ed6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -19
README.md CHANGED
@@ -26,25 +26,6 @@ We release InternViT-6B-448px-V1-0, which is integrated into [InternVL-Chat-V1-1
26
  - **Pretrain Dataset:** LAION-en, LAION-COCO, COYO, CC12M, CC3M, SBU, Wukong, LAION-multi, OCR-related datasets.
27
  - **Note:** This model has 48 blocks, and we found that using the output after the fourth-to-last block worked best for MLLM. Therefore, when building a MLLM with this model, **please use the features from the fourth-to-last layer.**
28
 
29
- ## Released Models
30
- ### Vision Foundation model
31
- | Model | Date | Download | Note |
32
- | ----------------------- | ---------- | ---------------------------------------------------------------------- | -------------------------------- |
33
- | InternViT-6B-448px-V1-5 | 2024.04.20 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternViT-6B-448px-V1-5) | support dynamic resolution, super strong OCR (🔥new) |
34
- | InternViT-6B-448px-V1-2 | 2024.02.11 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternViT-6B-448px-V1-2) | 448 resolution |
35
- | InternViT-6B-448px-V1-0 | 2024.01.30 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternViT-6B-448px-V1-0) | 448 resolution |
36
- | InternViT-6B-224px | 2023.12.22 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternViT-6B-224px) | vision foundation model |
37
- | InternVL-14B-224px | 2023.12.22 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-14B-224px) | vision-language foundation model |
38
-
39
- ### Multimodal Large Language Model (MLLM)
40
- | Model | Date | Download | Note |
41
- | ----------------------- | ---------- | --------------------------------------------------------------------------- | ---------------------------------- |
42
- | InternVL-Chat-V1-5 | 2024.04.18 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-5) | support 4K image; super strong OCR; Approaching the performance of GPT-4V and Gemini Pro on various benchmarks like MMMU, DocVQA, ChartQA, MathVista, etc. (🔥new)|
43
- | InternVL-Chat-V1-2-Plus | 2024.02.21 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-2-Plus) | more SFT data and stronger |
44
- | InternVL-Chat-V1-2 | 2024.02.11 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-2) | scaling up LLM to 34B |
45
- | InternVL-Chat-V1-1 | 2024.01.24 | 🤗 [HF link](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-1) | support Chinese and stronger OCR |
46
-
47
-
48
  ## Model Usage (Image Embeddings)
49
 
50
  ```python
 
26
  - **Pretrain Dataset:** LAION-en, LAION-COCO, COYO, CC12M, CC3M, SBU, Wukong, LAION-multi, OCR-related datasets.
27
  - **Note:** This model has 48 blocks, and we found that using the output after the fourth-to-last block worked best for MLLM. Therefore, when building a MLLM with this model, **please use the features from the fourth-to-last layer.**
28
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  ## Model Usage (Image Embeddings)
30
 
31
  ```python