Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -14,8 +14,7 @@ def format_prompt(message, history):
|
|
14 |
prompt += f"[INST] {message} [/INST]"
|
15 |
return prompt
|
16 |
|
17 |
-
def generate(
|
18 |
-
prompt, history, temperature=0.9, max_new_tokens=16000, top_p=0.95, repetition_penalty=1.0,):
|
19 |
generate_kwargs = dict(
|
20 |
temperature=temperature,
|
21 |
max_new_tokens=max_new_tokens,
|
@@ -27,10 +26,10 @@ def generate(
|
|
27 |
stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
|
28 |
output = ""
|
29 |
|
30 |
-
|
31 |
-
|
32 |
-
|
33 |
-
|
34 |
|
35 |
css = """
|
36 |
#mkd {
|
|
|
14 |
prompt += f"[INST] {message} [/INST]"
|
15 |
return prompt
|
16 |
|
17 |
+
def generate(prompt, history, temperature=0.9, max_new_tokens=16000, top_p=0.95, repetition_penalty=1.0,):
|
|
|
18 |
generate_kwargs = dict(
|
19 |
temperature=temperature,
|
20 |
max_new_tokens=max_new_tokens,
|
|
|
26 |
stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
|
27 |
output = ""
|
28 |
|
29 |
+
for response in stream:
|
30 |
+
output += response.token.text
|
31 |
+
yield output
|
32 |
+
return output
|
33 |
|
34 |
css = """
|
35 |
#mkd {
|