Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -8,7 +8,7 @@ library_name: transformers
8
 
9
  ## 1. Introduction
10
 
11
- Introducing DeepSeek-VL2, an advanced series of large Mixture-of-Experts (MoE) Vision-Language Models that significantly improves upon its predecessor, DeepSeek-VL. DeepSeek-VL2 demonstrates superior capabilities across various tasks, including but not limited to visual question answering, optical character recognition, document/table/chart understanding, and visual grounding. Our model series is composed of three variants: DeepSeek-VL2-Tiny, DeepSeek-VL2-Small and DeepSeek-VL2, with 1.0B, 2.8B and 4.5B activated parameters respectively.
12
  DeepSeek-VL2 achieves competitive or state-of-the-art performance with similar or fewer activated parameters compared to existing open-source dense and MoE-based models.
13
 
14
 
@@ -53,7 +53,7 @@ from deepseek_vl.utils.io import load_pil_images
53
 
54
 
55
  # specify the path to the model
56
- model_path = "deepseek-ai/deepseek-vl2-small"
57
  vl_chat_processor: DeepseekVLV2Processor = DeepseekVLV2Processor.from_pretrained(model_path)
58
  tokenizer = vl_chat_processor.tokenizer
59
 
 
8
 
9
  ## 1. Introduction
10
 
11
+ Introducing DeepSeek-VL2, an advanced series of large Mixture-of-Experts (MoE) Vision-Language Models that significantly improves upon its predecessor, DeepSeek-VL. DeepSeek-VL2 demonstrates superior capabilities across various tasks, including but not limited to visual question answering, optical character recognition, document/table/chart understanding, and visual grounding. Our model series is composed of three variants: DeepSeek-VL2-Tiny, DeepSeek-VL2-Small and DeepSeek-VL2, with 3.37B, 16.1B and 27.5B activated parameters respectively.
12
  DeepSeek-VL2 achieves competitive or state-of-the-art performance with similar or fewer activated parameters compared to existing open-source dense and MoE-based models.
13
 
14
 
 
53
 
54
 
55
  # specify the path to the model
56
+ model_path = "deepseek-ai/deepseek-vl2-tiny"
57
  vl_chat_processor: DeepseekVLV2Processor = DeepseekVLV2Processor.from_pretrained(model_path)
58
  tokenizer = vl_chat_processor.tokenizer
59