sflindrs commited on
Commit
d404fc4
·
verified ·
1 Parent(s): f1a5355

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +9 -7
app.py CHANGED
@@ -10,14 +10,16 @@ TRENDING_MODELS = [
10
  "Salesforce/blip2-opt-2.7b", # Uses Blip2Config
11
  "Salesforce/blip2-flan-t5-xl", # Uses Blip2Config
12
  "Salesforce/instructblip-vicuna-7b", # Uses InstructBlipConfig
13
- "LLaVA/LLaVA-7B", # Uses LlavaConfig
14
- "LLaVA/LLaVA-13B", # Uses LlavaConfig
15
- "LlavaNext/LlavaNext-7B", # Uses LlavaNextConfig
16
- "Qwen2-VL-7B", # Uses Qwen2VLConfig
17
- "google/pix2struct-base", # Uses Pix2StructConfig
18
  "nlpconnect/vit-gpt2-image-captioning", # Uses VisionEncoderDecoderConfig
19
- "LlavaOneVision/LlavaOneVision-7B", # Uses LlavaOnevisionConfig
20
- "MosaicML/mllama-7b" # Uses MllamaConfig
 
 
21
  ]
22
 
23
  # --- Helper: if the user selects "Custom", then they can enter any model identifier ---
 
10
  "Salesforce/blip2-opt-2.7b", # Uses Blip2Config
11
  "Salesforce/blip2-flan-t5-xl", # Uses Blip2Config
12
  "Salesforce/instructblip-vicuna-7b", # Uses InstructBlipConfig
13
+ "llava-hf/llava-1.5-7b-hf", # Uses LlavaConfig
14
+ "liuhaotian/llava-v1.5-13b", # Uses LlavaConfig
15
+ "llava-hf/llava-v1.6-mistral-7b-hf", # Uses LlavaNextConfig
16
+ "Qwen/Qwen2-VL-7B-Instruct", # Uses Qwen2VLConfig
17
+ "google/pix2struct-ai2d-base", # Uses Pix2StructConfig
18
  "nlpconnect/vit-gpt2-image-captioning", # Uses VisionEncoderDecoderConfig
19
+ "llava-hf/llava-onevision-qwen2-0.5b-ov-hf", # Uses LlavaOnevisionConfig
20
+ "mosaicml/mpt-7b-chat", # Uses MllamaConfig
21
+ "ibm-granite/granite-vision-3.1-2b-preview",
22
+ "allenai/Molmo-7B-D-0924"
23
  ]
24
 
25
  # --- Helper: if the user selects "Custom", then they can enter any model identifier ---