Spaces:
Runtime error
Runtime error
rev app
Browse files
app.py
CHANGED
@@ -82,7 +82,7 @@ else:
|
|
82 |
|
83 |
###@spaces.GPU
|
84 |
def load_model():
|
85 |
-
global model
|
86 |
print("กำลังโหลดโมเดล...")
|
87 |
try:
|
88 |
# โหลด tokenizer จาก base model
|
@@ -115,7 +115,7 @@ def load_model():
|
|
115 |
# FastVisionModel.for_inference(model)
|
116 |
# print("โหลดโมเดลสำเร็จ!")
|
117 |
# return True
|
118 |
-
from transformers import AutoModelForVision2Seq
|
119 |
import torch
|
120 |
|
121 |
# print("กำลังโหลด tokenizer...")
|
@@ -129,19 +129,25 @@ def load_model():
|
|
129 |
# use_auth_token=True
|
130 |
# )
|
131 |
|
132 |
-
print(f"2. ประเภทของ tokenizer: {type(tokenizer)}")
|
133 |
-
print(f"3. เมธอดที่มีใน tokenizer: {dir(tokenizer)}")
|
134 |
-
print("4. Global tokenizer after assignment:", type(tokenizer))
|
135 |
|
136 |
-
print("โหลด tokenizer สำเร็จ กำลังโหลดโมเดลที่ fine-tune...")
|
137 |
|
138 |
# โหลดโมเดล fine-tuned
|
139 |
print("กำลังโหลดโมเดล fine-tuned...")
|
|
|
|
|
|
|
|
|
|
|
140 |
model = AutoModelForVision2Seq.from_pretrained(
|
141 |
"Aekanun/Llama-3.2-11B-Vision-Instruct-XRay",
|
|
|
142 |
device_map="auto",
|
143 |
-
torch_dtype=torch.float16
|
144 |
-
)
|
145 |
|
146 |
print("โหลดโมเดลสำเร็จ!")
|
147 |
return True
|
@@ -158,7 +164,7 @@ def process_image(image):
|
|
158 |
# import os
|
159 |
# os.environ['TORCH_LOGS'] = '+dynamo'
|
160 |
# os.environ['TORCHDYNAMO_VERBOSE'] = '1'
|
161 |
-
global model
|
162 |
|
163 |
### โหลด tokenizer จาก base model
|
164 |
from unsloth import FastVisionModel
|
@@ -174,8 +180,9 @@ def process_image(image):
|
|
174 |
)
|
175 |
###
|
176 |
|
177 |
-
|
178 |
print("\nใน process_image():")
|
|
|
179 |
print("A. Type of tokenizer:", type(tokenizer))
|
180 |
if tokenizer is not None:
|
181 |
print("B. Available methods:", dir(tokenizer))
|
|
|
82 |
|
83 |
###@spaces.GPU
|
84 |
def load_model():
|
85 |
+
global model
|
86 |
print("กำลังโหลดโมเดล...")
|
87 |
try:
|
88 |
# โหลด tokenizer จาก base model
|
|
|
115 |
# FastVisionModel.for_inference(model)
|
116 |
# print("โหลดโมเดลสำเร็จ!")
|
117 |
# return True
|
118 |
+
from transformers import AutoModelForVision2Seq
|
119 |
import torch
|
120 |
|
121 |
# print("กำลังโหลด tokenizer...")
|
|
|
129 |
# use_auth_token=True
|
130 |
# )
|
131 |
|
132 |
+
# print(f"2. ประเภทของ tokenizer: {type(tokenizer)}")
|
133 |
+
# print(f"3. เมธอดที่มีใน tokenizer: {dir(tokenizer)}")
|
134 |
+
# print("4. Global tokenizer after assignment:", type(tokenizer))
|
135 |
|
136 |
+
# print("โหลด tokenizer สำเร็จ กำลังโหลดโมเดลที่ fine-tune...")
|
137 |
|
138 |
# โหลดโมเดล fine-tuned
|
139 |
print("กำลังโหลดโมเดล fine-tuned...")
|
140 |
+
# model = AutoModelForVision2Seq.from_pretrained(
|
141 |
+
# "Aekanun/Llama-3.2-11B-Vision-Instruct-XRay",
|
142 |
+
# device_map="auto",
|
143 |
+
# torch_dtype=torch.float16
|
144 |
+
# ).to('cuda')
|
145 |
model = AutoModelForVision2Seq.from_pretrained(
|
146 |
"Aekanun/Llama-3.2-11B-Vision-Instruct-XRay",
|
147 |
+
load_in_4bit = True,
|
148 |
device_map="auto",
|
149 |
+
torch_dtype = torch.float16
|
150 |
+
)
|
151 |
|
152 |
print("โหลดโมเดลสำเร็จ!")
|
153 |
return True
|
|
|
164 |
# import os
|
165 |
# os.environ['TORCH_LOGS'] = '+dynamo'
|
166 |
# os.environ['TORCHDYNAMO_VERBOSE'] = '1'
|
167 |
+
global model
|
168 |
|
169 |
### โหลด tokenizer จาก base model
|
170 |
from unsloth import FastVisionModel
|
|
|
180 |
)
|
181 |
###
|
182 |
|
183 |
+
|
184 |
print("\nใน process_image():")
|
185 |
+
print("Type of model:", type(model))
|
186 |
print("A. Type of tokenizer:", type(tokenizer))
|
187 |
if tokenizer is not None:
|
188 |
print("B. Available methods:", dir(tokenizer))
|