Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -28,6 +28,8 @@ def predict(message, history):
|
|
28 |
|
29 |
generation_config = GenerationConfig(
|
30 |
penalty_alpha=0.6,
|
|
|
|
|
31 |
do_sample=True,
|
32 |
top_k=5,
|
33 |
temperature=0.5,
|
@@ -39,7 +41,7 @@ def predict(message, history):
|
|
39 |
|
40 |
outputs = model.generate(**inputs,streamer=streamer, generation_config=generation_config)
|
41 |
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
42 |
-
return response[len(formatted_prompt(message))
|
43 |
|
44 |
# partial_message = ""
|
45 |
# for i in response:
|
|
|
28 |
|
29 |
generation_config = GenerationConfig(
|
30 |
penalty_alpha=0.6,
|
31 |
+
early_stopping=True,
|
32 |
+
num_beams=4,
|
33 |
do_sample=True,
|
34 |
top_k=5,
|
35 |
temperature=0.5,
|
|
|
41 |
|
42 |
outputs = model.generate(**inputs,streamer=streamer, generation_config=generation_config)
|
43 |
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
44 |
+
return response[len(formatted_prompt(message)):]
|
45 |
|
46 |
# partial_message = ""
|
47 |
# for i in response:
|