Omnibus commited on
Commit
eec8c3a
·
verified ·
1 Parent(s): 92fbdae

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +10 -7
app.py CHANGED
@@ -56,12 +56,15 @@ def format_prompt_default(message, history,cust_p):
56
  if history:
57
  #<start_of_turn>userHow does the brain work?<end_of_turn><start_of_turn>model
58
  for user_prompt, bot_response in history:
59
- prompt += f"{user_prompt}\n"
60
  print(prompt)
61
- prompt += f"{bot_response}\n"
62
  print(prompt)
63
- #prompt += f"{message}\n"
64
  prompt+=cust_p.replace("USER_INPUT",message)
 
 
 
65
  return prompt
66
 
67
 
@@ -72,7 +75,7 @@ def load_models(inp):
72
  print(models[inp])
73
  model_state= InferenceClient(models[inp])
74
  out_box=gr.update(label=models[inp])
75
- prompt_out="USER_INPUT\n"
76
  return out_box,prompt_out, model_state
77
 
78
 
@@ -167,9 +170,9 @@ with gr.Blocks() as app:
167
  rand = gr.Checkbox(label="Random Seed", value=True)
168
  seed=gr.Slider(label="Seed", minimum=1, maximum=1111111111111111,step=1, value=rand_val)
169
  tokens = gr.Slider(label="Max new tokens",value=1600,minimum=0,maximum=8000,step=64,interactive=True, visible=True,info="The maximum number of tokens")
170
- temp=gr.Slider(label="Temperature",step=0.01, minimum=0.01, maximum=1.0, value=0.49)
171
- top_p=gr.Slider(label="Top-P",step=0.01, minimum=0.01, maximum=1.0, value=0.49)
172
- rep_p=gr.Slider(label="Repetition Penalty",step=0.01, minimum=0.1, maximum=2.0, value=0.99)
173
  chat_mem=gr.Number(label="Chat Memory", info="Number of previous chats to retain",value=4)
174
  with gr.Accordion(label="Screenshot",open=False):
175
  with gr.Row():
 
56
  if history:
57
  #<start_of_turn>userHow does the brain work?<end_of_turn><start_of_turn>model
58
  for user_prompt, bot_response in history:
59
+ prompt += f"<|startoftext|>{user_prompt}<|endoftext|>"
60
  print(prompt)
61
+ prompt += f"<|startoftext|>{bot_response}<|endoftext|>"
62
  print(prompt)
63
+ prompt += "<|startoftext|>"
64
  prompt+=cust_p.replace("USER_INPUT",message)
65
+ prompt += "<|endoftext|>"
66
+ prompt += "<|startoftext|>"
67
+
68
  return prompt
69
 
70
 
 
75
  print(models[inp])
76
  model_state= InferenceClient(models[inp])
77
  out_box=gr.update(label=models[inp])
78
+ prompt_out="<|startoftext|>USER_INPUT<|endoftext|>"
79
  return out_box,prompt_out, model_state
80
 
81
 
 
170
  rand = gr.Checkbox(label="Random Seed", value=True)
171
  seed=gr.Slider(label="Seed", minimum=1, maximum=1111111111111111,step=1, value=rand_val)
172
  tokens = gr.Slider(label="Max new tokens",value=1600,minimum=0,maximum=8000,step=64,interactive=True, visible=True,info="The maximum number of tokens")
173
+ temp=gr.Slider(label="Temperature",step=0.01, minimum=0.01, maximum=1.0, value=0.99)
174
+ top_p=gr.Slider(label="Top-P",step=0.01, minimum=0.01, maximum=1.0, value=0.99)
175
+ rep_p=gr.Slider(label="Repetition Penalty",step=0.01, minimum=0.1, maximum=2.0, value=1.2)
176
  chat_mem=gr.Number(label="Chat Memory", info="Number of previous chats to retain",value=4)
177
  with gr.Accordion(label="Screenshot",open=False):
178
  with gr.Row():