walaa2022 commited on
Commit
2186a64
·
verified ·
1 Parent(s): 3aad013

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -6,7 +6,7 @@ Optimized for Hugging Face deployment with efficient resource usage
6
 
7
  import gradio as gr
8
  import torch
9
- from transformers import Qwen2VLForConditionalGeneration, AutoProcessor
10
  from qwen_vl_utils import process_vision_info
11
  from PIL import Image
12
  import json
@@ -39,9 +39,9 @@ def load_model_cached(model_name: str = "Qwen/Qwen2.5-VL-3B-Instruct"):
39
 
40
  try:
41
  # Load with memory optimization for HF Spaces
42
- MODEL = Qwen2VLForConditionalGeneration.from_pretrained(
43
  model_name,
44
- torch_dtype=torch.bfloat16 if DEVICE == "cuda" else torch.float32,
45
  device_map="auto" if DEVICE == "cuda" else None,
46
  trust_remote_code=True,
47
  attn_implementation="eager", # More stable for HF Spaces
 
6
 
7
  import gradio as gr
8
  import torch
9
+ from transformers import Qwen2_5_VLForConditionalGeneration, AutoTokenizer, AutoProcessor
10
  from qwen_vl_utils import process_vision_info
11
  from PIL import Image
12
  import json
 
39
 
40
  try:
41
  # Load with memory optimization for HF Spaces
42
+ MODEL = Qwen2_5_VLForConditionalGeneration.from_pretrained(
43
  model_name,
44
+ torch_dtype='auto',
45
  device_map="auto" if DEVICE == "cuda" else None,
46
  trust_remote_code=True,
47
  attn_implementation="eager", # More stable for HF Spaces