Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -27,7 +27,7 @@ if not torch.cuda.is_available():
|
|
27 |
|
28 |
|
29 |
if torch.cuda.is_available():
|
30 |
-
model_id = "wangzhang/ChatSDB-
|
31 |
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.float16, device_map="auto")
|
32 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
33 |
tokenizer.use_default_system_prompt = False
|
@@ -49,9 +49,9 @@ def generate(
|
|
49 |
conversation.append({"role": "system", "content": system_prompt})
|
50 |
for user, assistant in chat_history:
|
51 |
conversation.extend([{"role": "user", "content": user}, {"role": "assistant", "content": assistant}])
|
52 |
-
conversation.append({"role": "user", "content":
|
53 |
|
54 |
-
chat = tokenizer.apply_chat_template(
|
55 |
inputs = tokenizer(chat, return_tensors="pt", add_special_tokens=False).to("cuda")
|
56 |
if len(inputs) > MAX_INPUT_TOKEN_LENGTH:
|
57 |
inputs = inputs[-MAX_INPUT_TOKEN_LENGTH:]
|
|
|
27 |
|
28 |
|
29 |
if torch.cuda.is_available():
|
30 |
+
model_id = "wangzhang/ChatSDB-tb-testing"
|
31 |
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.float16, device_map="auto")
|
32 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
33 |
tokenizer.use_default_system_prompt = False
|
|
|
49 |
conversation.append({"role": "system", "content": system_prompt})
|
50 |
for user, assistant in chat_history:
|
51 |
conversation.extend([{"role": "user", "content": user}, {"role": "assistant", "content": assistant}])
|
52 |
+
conversation.append({"role": "user", "content": message)
|
53 |
|
54 |
+
chat = tokenizer.apply_chat_template(conversation, tokenize=False)
|
55 |
inputs = tokenizer(chat, return_tensors="pt", add_special_tokens=False).to("cuda")
|
56 |
if len(inputs) > MAX_INPUT_TOKEN_LENGTH:
|
57 |
inputs = inputs[-MAX_INPUT_TOKEN_LENGTH:]
|