mlabonne commited on
Commit
044264a
·
1 Parent(s): cb8a8d5

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -4
app.py CHANGED
@@ -34,12 +34,12 @@ def predict(message, history):
34
  stop = StopOnTokens()
35
 
36
  # Formatting the input for the model.
37
- messages = "</s>".join(["</s>".join(["\n<|user|>:" + item[0], "\n<|assistant|>:" + item[1]])
38
- for item in history_transformer_format])
39
- model_inputs = tokenizer([messages], return_tensors="pt")
40
  streamer = TextIteratorStreamer(tokenizer, timeout=10., skip_prompt=True, skip_special_tokens=True)
41
  generate_kwargs = dict(
42
- model_inputs,
43
  streamer=streamer,
44
  max_new_tokens=1024,
45
  do_sample=True,
 
34
  stop = StopOnTokens()
35
 
36
  # Formatting the input for the model.
37
+ system_prompt = "<|im_start|>system\nYou are Phixtral, a helpful AI assistant.<|im_end|>"
38
+ messages = system_prompt + "".join(["".join(["\n<|im_start|>user\n" + item[0], "<|im_end|>\n<|im_start|>assistant\n" + item[1]]) for item in history_transformer_format])
39
+ input_ids = tokenizer([messages], return_tensors="pt").to('cuda')
40
  streamer = TextIteratorStreamer(tokenizer, timeout=10., skip_prompt=True, skip_special_tokens=True)
41
  generate_kwargs = dict(
42
+ input_ids,
43
  streamer=streamer,
44
  max_new_tokens=1024,
45
  do_sample=True,