Update app.py
Browse files
app.py
CHANGED
@@ -25,12 +25,6 @@ warnings.filterwarnings('ignore')
|
|
25 |
model = None
|
26 |
tokenizer = None
|
27 |
|
28 |
-
if 'HUGGING_FACE_HUB_TOKEN' in os.environ:
|
29 |
-
print("กำลังเข้าสู่ระบบ Hugging Face Hub...")
|
30 |
-
login(token=os.environ['HUGGING_FACE_HUB_TOKEN'])
|
31 |
-
else:
|
32 |
-
print("คำเตือน: ไม่พบ HUGGING_FACE_HUB_TOKEN")
|
33 |
-
|
34 |
###@spaces.GPU
|
35 |
def load_model():
|
36 |
global model
|
@@ -42,7 +36,7 @@ def load_model():
|
|
42 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
43 |
model = AutoModelForVision2Seq.from_pretrained(
|
44 |
"0llheaven/Llama-3.2-11B-Vision-Radiology-mini",
|
45 |
-
load_in_4bit = True,
|
46 |
device_map=device,
|
47 |
torch_dtype = torch.float16
|
48 |
)
|
|
|
25 |
model = None
|
26 |
tokenizer = None
|
27 |
|
|
|
|
|
|
|
|
|
|
|
|
|
28 |
###@spaces.GPU
|
29 |
def load_model():
|
30 |
global model
|
|
|
36 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
37 |
model = AutoModelForVision2Seq.from_pretrained(
|
38 |
"0llheaven/Llama-3.2-11B-Vision-Radiology-mini",
|
39 |
+
# load_in_4bit = True,
|
40 |
device_map=device,
|
41 |
torch_dtype = torch.float16
|
42 |
)
|