ruslanmv commited on
Commit
a02afc4
·
verified ·
1 Parent(s): c3c8527

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -29,10 +29,10 @@ def askme(symptoms, question):
29
  prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
30
  inputs = tokenizer(prompt, return_tensors="pt").to(device) # Ensure inputs are on CUDA device
31
  outputs = model.generate(**inputs, max_new_tokens=200, use_cache=True)
32
- response_text = tokenizer.batch_decode(outputs)[0].strip() #skip_special_tokens=True
33
  # Remove system messages and content
34
  # Extract only the assistant's response
35
- assistant_response = response_text.split("<|im_start|>assistant")[1].strip().replace('<|im_end', '')
36
  # Extract only the assistant's response
37
  return assistant_response
38
 
 
29
  prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
30
  inputs = tokenizer(prompt, return_tensors="pt").to(device) # Ensure inputs are on CUDA device
31
  outputs = model.generate(**inputs, max_new_tokens=200, use_cache=True)
32
+ response_text = tokenizer.batch_decode(outputs, skip_special_tokens=False)[0].strip() #skip_special_tokens=True
33
  # Remove system messages and content
34
  # Extract only the assistant's response
35
+ assistant_response = response_text.split("<|im_start|>assistant")[1].strip().replace('<|im_end|>', '')
36
  # Extract only the assistant's response
37
  return assistant_response
38