Update README.md
#2
by
srikanthreddykasam
- opened
README.md
CHANGED
@@ -8,7 +8,7 @@ library_name: transformers
|
|
8 |
|
9 |
## 1. Introduction
|
10 |
|
11 |
-
Introducing DeepSeek-VL2, an advanced series of large Mixture-of-Experts (MoE) Vision-Language Models that significantly improves upon its predecessor, DeepSeek-VL. DeepSeek-VL2 demonstrates superior capabilities across various tasks, including but not limited to visual question answering, optical character recognition, document/table/chart understanding, and visual grounding. Our model series is composed of three variants: DeepSeek-VL2-Tiny, DeepSeek-VL2-Small and DeepSeek-VL2, with
|
12 |
DeepSeek-VL2 achieves competitive or state-of-the-art performance with similar or fewer activated parameters compared to existing open-source dense and MoE-based models.
|
13 |
|
14 |
|
@@ -53,7 +53,7 @@ from deepseek_vl.utils.io import load_pil_images
|
|
53 |
|
54 |
|
55 |
# specify the path to the model
|
56 |
-
model_path = "deepseek-ai/deepseek-vl2-
|
57 |
vl_chat_processor: DeepseekVLV2Processor = DeepseekVLV2Processor.from_pretrained(model_path)
|
58 |
tokenizer = vl_chat_processor.tokenizer
|
59 |
|
|
|
8 |
|
9 |
## 1. Introduction
|
10 |
|
11 |
+
Introducing DeepSeek-VL2, an advanced series of large Mixture-of-Experts (MoE) Vision-Language Models that significantly improves upon its predecessor, DeepSeek-VL. DeepSeek-VL2 demonstrates superior capabilities across various tasks, including but not limited to visual question answering, optical character recognition, document/table/chart understanding, and visual grounding. Our model series is composed of three variants: DeepSeek-VL2-Tiny, DeepSeek-VL2-Small and DeepSeek-VL2, with 3.37B, 16.1B and 27.5B activated parameters respectively.
|
12 |
DeepSeek-VL2 achieves competitive or state-of-the-art performance with similar or fewer activated parameters compared to existing open-source dense and MoE-based models.
|
13 |
|
14 |
|
|
|
53 |
|
54 |
|
55 |
# specify the path to the model
|
56 |
+
model_path = "deepseek-ai/deepseek-vl2-tiny"
|
57 |
vl_chat_processor: DeepseekVLV2Processor = DeepseekVLV2Processor.from_pretrained(model_path)
|
58 |
tokenizer = vl_chat_processor.tokenizer
|
59 |
|