Update app.py
Browse files
app.py
CHANGED
@@ -75,12 +75,14 @@ def generate(prompt, history=[], temperature=0.1, max_new_tokens=10000, top_p=0.
|
|
75 |
return
|
76 |
|
77 |
formatted_prompt = format_prompt(prompt, history)
|
|
|
78 |
try:
|
79 |
stream = client.text_generation(formatted_prompt, temperature=temperature, max_new_tokens=min(max_new_tokens, available_tokens),
|
80 |
top_p=top_p, repetition_penalty=repetition_penalty, do_sample=True, seed=42, stream=True)
|
81 |
for response in stream:
|
82 |
-
|
83 |
-
|
|
|
84 |
except Exception as e:
|
85 |
yield f"Error: {str(e)}\nTotal tokens used: {total_tokens_used}"
|
86 |
|
|
|
75 |
return
|
76 |
|
77 |
formatted_prompt = format_prompt(prompt, history)
|
78 |
+
output_accumulated = ""
|
79 |
try:
|
80 |
stream = client.text_generation(formatted_prompt, temperature=temperature, max_new_tokens=min(max_new_tokens, available_tokens),
|
81 |
top_p=top_p, repetition_penalty=repetition_penalty, do_sample=True, seed=42, stream=True)
|
82 |
for response in stream:
|
83 |
+
output_part = response['generated_text'] if 'generated_text' in response else str(response)
|
84 |
+
output_accumulated += output_part
|
85 |
+
yield output_accumulated + f"\n\n---\nTotal tokens used: {total_tokens_used}"
|
86 |
except Exception as e:
|
87 |
yield f"Error: {str(e)}\nTotal tokens used: {total_tokens_used}"
|
88 |
|