wjm55 commited on
Commit
2b2bd20
·
1 Parent(s): 2f86d0f

Update model references to Qwen2.5-VL-7B-Instruct

Browse files
Files changed (1) hide show
  1. app.py +4 -4
app.py CHANGED
@@ -49,12 +49,12 @@ def array_to_image_path(image_array):
49
  return full_path
50
 
51
  models = {
52
- "Qwen2.5-VL-7B-Instruct": Qwen2_5_VLForConditionalGeneration.from_pretrained("Qwen2.5-VL-7B-Instruct", trust_remote_code=True, torch_dtype="auto").cuda().eval()
53
 
54
  }
55
 
56
  processors = {
57
- "Qwen2.5-VL-7B-Instruct": AutoProcessor.from_pretrained("Qwen2.5-VL-7B-Instruct", trust_remote_code=True)
58
  }
59
 
60
  DESCRIPTION = "This demo uses[Qwen2.5-VL-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-7B-Instruct)"
@@ -67,7 +67,7 @@ assistant_prompt = '<|assistant|>\n'
67
  prompt_suffix = "<|end|>\n"
68
 
69
  @spaces.GPU
70
- def run_example(image, model_id="Qwen/Qwen2-VL-7B-Instruct", run_ner=False, ner_labels=DEFAULT_NER_LABELS):
71
  # First get the OCR text
72
  text_input = "Convert the image to text."
73
  image_path = array_to_image_path(image)
@@ -235,7 +235,7 @@ with gr.Blocks(css=css) as demo:
235
  with gr.Row():
236
  with gr.Column(elem_classes="input-container"):
237
  input_img = gr.Image(label="Input Picture", elem_classes="gr-image-input")
238
- model_selector = gr.Dropdown(choices=list(models.keys()), label="Model", value="Qwen/Qwen2-VL-7B-Instruct", elem_classes="gr-dropdown")
239
 
240
  # Add NER controls
241
  with gr.Row():
 
49
  return full_path
50
 
51
  models = {
52
+ "Qwen/Qwen2.5-VL-7B-Instruct": Qwen2_5_VLForConditionalGeneration.from_pretrained("Qwen/Qwen2.5-VL-7B-Instruct", trust_remote_code=True, torch_dtype="auto").cuda().eval()
53
 
54
  }
55
 
56
  processors = {
57
+ "Qwen/Qwen2.5-VL-7B-Instruct": AutoProcessor.from_pretrained("Qwen/Qwen2.5-VL-7B-Instruct", trust_remote_code=True)
58
  }
59
 
60
  DESCRIPTION = "This demo uses[Qwen2.5-VL-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-7B-Instruct)"
 
67
  prompt_suffix = "<|end|>\n"
68
 
69
  @spaces.GPU
70
+ def run_example(image, model_id="Qwen/Qwen2.5-VL-7B-Instruct", run_ner=False, ner_labels=DEFAULT_NER_LABELS):
71
  # First get the OCR text
72
  text_input = "Convert the image to text."
73
  image_path = array_to_image_path(image)
 
235
  with gr.Row():
236
  with gr.Column(elem_classes="input-container"):
237
  input_img = gr.Image(label="Input Picture", elem_classes="gr-image-input")
238
+ model_selector = gr.Dropdown(choices=list(models.keys()), label="Model", value="Qwen/Qwen2.5-VL-7B-Instruct", elem_classes="gr-dropdown")
239
 
240
  # Add NER controls
241
  with gr.Row():