Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -56,12 +56,15 @@ def format_prompt_default(message, history,cust_p):
|
|
56 |
if history:
|
57 |
#<start_of_turn>userHow does the brain work?<end_of_turn><start_of_turn>model
|
58 |
for user_prompt, bot_response in history:
|
59 |
-
prompt += f"{user_prompt}
|
60 |
print(prompt)
|
61 |
-
prompt += f"{bot_response}
|
62 |
print(prompt)
|
63 |
-
|
64 |
prompt+=cust_p.replace("USER_INPUT",message)
|
|
|
|
|
|
|
65 |
return prompt
|
66 |
|
67 |
|
@@ -72,7 +75,7 @@ def load_models(inp):
|
|
72 |
print(models[inp])
|
73 |
model_state= InferenceClient(models[inp])
|
74 |
out_box=gr.update(label=models[inp])
|
75 |
-
prompt_out="USER_INPUT
|
76 |
return out_box,prompt_out, model_state
|
77 |
|
78 |
|
@@ -167,9 +170,9 @@ with gr.Blocks() as app:
|
|
167 |
rand = gr.Checkbox(label="Random Seed", value=True)
|
168 |
seed=gr.Slider(label="Seed", minimum=1, maximum=1111111111111111,step=1, value=rand_val)
|
169 |
tokens = gr.Slider(label="Max new tokens",value=1600,minimum=0,maximum=8000,step=64,interactive=True, visible=True,info="The maximum number of tokens")
|
170 |
-
temp=gr.Slider(label="Temperature",step=0.01, minimum=0.01, maximum=1.0, value=0.
|
171 |
-
top_p=gr.Slider(label="Top-P",step=0.01, minimum=0.01, maximum=1.0, value=0.
|
172 |
-
rep_p=gr.Slider(label="Repetition Penalty",step=0.01, minimum=0.1, maximum=2.0, value=
|
173 |
chat_mem=gr.Number(label="Chat Memory", info="Number of previous chats to retain",value=4)
|
174 |
with gr.Accordion(label="Screenshot",open=False):
|
175 |
with gr.Row():
|
|
|
56 |
if history:
|
57 |
#<start_of_turn>userHow does the brain work?<end_of_turn><start_of_turn>model
|
58 |
for user_prompt, bot_response in history:
|
59 |
+
prompt += f"<|startoftext|>{user_prompt}<|endoftext|>"
|
60 |
print(prompt)
|
61 |
+
prompt += f"<|startoftext|>{bot_response}<|endoftext|>"
|
62 |
print(prompt)
|
63 |
+
prompt += "<|startoftext|>"
|
64 |
prompt+=cust_p.replace("USER_INPUT",message)
|
65 |
+
prompt += "<|endoftext|>"
|
66 |
+
prompt += "<|startoftext|>"
|
67 |
+
|
68 |
return prompt
|
69 |
|
70 |
|
|
|
75 |
print(models[inp])
|
76 |
model_state= InferenceClient(models[inp])
|
77 |
out_box=gr.update(label=models[inp])
|
78 |
+
prompt_out="<|startoftext|>USER_INPUT<|endoftext|>"
|
79 |
return out_box,prompt_out, model_state
|
80 |
|
81 |
|
|
|
170 |
rand = gr.Checkbox(label="Random Seed", value=True)
|
171 |
seed=gr.Slider(label="Seed", minimum=1, maximum=1111111111111111,step=1, value=rand_val)
|
172 |
tokens = gr.Slider(label="Max new tokens",value=1600,minimum=0,maximum=8000,step=64,interactive=True, visible=True,info="The maximum number of tokens")
|
173 |
+
temp=gr.Slider(label="Temperature",step=0.01, minimum=0.01, maximum=1.0, value=0.99)
|
174 |
+
top_p=gr.Slider(label="Top-P",step=0.01, minimum=0.01, maximum=1.0, value=0.99)
|
175 |
+
rep_p=gr.Slider(label="Repetition Penalty",step=0.01, minimum=0.1, maximum=2.0, value=1.2)
|
176 |
chat_mem=gr.Number(label="Chat Memory", info="Number of previous chats to retain",value=4)
|
177 |
with gr.Accordion(label="Screenshot",open=False):
|
178 |
with gr.Row():
|