Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -104,11 +104,12 @@ def build_prompt(chat_history, current_input_text, video_frames=None, image_file
|
|
104 |
MODEL_ID = "mistralai/Mistral-Small-3.1-24B-Instruct-2503"
|
105 |
SYSTEM_PROMPT = load_system_prompt(MODEL_ID, "SYSTEM_PROMPT.txt")
|
106 |
|
107 |
-
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
|
108 |
model = AutoModelForCausalLM.from_pretrained(
|
109 |
MODEL_ID,
|
110 |
torch_dtype=torch.float16,
|
111 |
-
device_map="auto"
|
|
|
112 |
).to(device)
|
113 |
model.eval()
|
114 |
|
@@ -201,4 +202,4 @@ demo = gr.ChatInterface(
|
|
201 |
)
|
202 |
|
203 |
if __name__ == "__main__":
|
204 |
-
demo.queue(max_size=20).launch(share=True)
|
|
|
104 |
MODEL_ID = "mistralai/Mistral-Small-3.1-24B-Instruct-2503"
|
105 |
SYSTEM_PROMPT = load_system_prompt(MODEL_ID, "SYSTEM_PROMPT.txt")
|
106 |
|
107 |
+
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, trust_remote_code=True)
|
108 |
model = AutoModelForCausalLM.from_pretrained(
|
109 |
MODEL_ID,
|
110 |
torch_dtype=torch.float16,
|
111 |
+
device_map="auto",
|
112 |
+
trust_remote_code=True
|
113 |
).to(device)
|
114 |
model.eval()
|
115 |
|
|
|
202 |
)
|
203 |
|
204 |
if __name__ == "__main__":
|
205 |
+
demo.queue(max_size=20).launch(share=True)
|