Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -1,12 +1,13 @@
|
|
1 |
import gradio as gr
|
2 |
from transformers import pipeline, AutoModelForCausalLM, AutoTokenizer
|
3 |
import spaces
|
|
|
4 |
|
5 |
# Load model with 8-bit precision
|
6 |
-
model_name = "yasserrmd/SmolLM2-
|
7 |
model = AutoModelForCausalLM.from_pretrained(
|
8 |
model_name,
|
9 |
-
|
10 |
device_map ="cuda"
|
11 |
)
|
12 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
@@ -29,7 +30,7 @@ def chat_assistant(chat_history, user_input):
|
|
29 |
|
30 |
response = generator(
|
31 |
[{"role": "system", "content": "You are a Data Loss Prevention (DLP) assistant designed to help users with questions and tasks related to data security, compliance, and policy enforcement. Respond concisely and professionally, offering practical guidance while ensuring clarity. If additional context or follow-up questions are required, ask the user to refine their input or provide specific examples."},
|
32 |
-
{"role": "user", "content": user_input}], max_new_tokens=
|
33 |
)[0]["generated_text"]
|
34 |
|
35 |
# Append to chat history
|
|
|
1 |
import gradio as gr
|
2 |
from transformers import pipeline, AutoModelForCausalLM, AutoTokenizer
|
3 |
import spaces
|
4 |
+
import torch
|
5 |
|
6 |
# Load model with 8-bit precision
|
7 |
+
model_name = "yasserrmd/SmolLM2-135M-synthetic-dlp"
|
8 |
model = AutoModelForCausalLM.from_pretrained(
|
9 |
model_name,
|
10 |
+
torch_dtype=torch.bfloat16
|
11 |
device_map ="cuda"
|
12 |
)
|
13 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
|
|
30 |
|
31 |
response = generator(
|
32 |
[{"role": "system", "content": "You are a Data Loss Prevention (DLP) assistant designed to help users with questions and tasks related to data security, compliance, and policy enforcement. Respond concisely and professionally, offering practical guidance while ensuring clarity. If additional context or follow-up questions are required, ask the user to refine their input or provide specific examples."},
|
33 |
+
{"role": "user", "content": user_input}], max_new_tokens=512, return_full_text=True
|
34 |
)[0]["generated_text"]
|
35 |
|
36 |
# Append to chat history
|