Tonic commited on
Commit
40c0747
·
verified ·
1 Parent(s): ebbfcba

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -7
app.py CHANGED
@@ -24,7 +24,7 @@ model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloa
24
  @torch.inference_mode()
25
  @spaces.GPU
26
  def predict_math_bot(user_message, system_message="", max_new_tokens=125, temperature=0.1, top_p=0.9, repetition_penalty=1.9, do_sample=False):
27
- prompt = f"<|system|>\n{system_message}\n<|user|>\n{user_message}<|assistant|>" if system_message else user_message
28
  inputs = tokenizer(prompt, return_tensors='pt', add_special_tokens=False)
29
  input_ids = inputs["input_ids"].to(model.device)
30
 
@@ -53,16 +53,16 @@ def main():
53
  top_p = gr.Slider(label="Top-p (nucleus sampling)", value=0.90, minimum=0.01, maximum=0.99)
54
  repetition_penalty = gr.Slider(label="Repetition penalty", value=1.9, minimum=1.0, maximum=2.0)
55
  do_sample = gr.Checkbox(label="Uncheck for faster inference", value=False)
56
-
57
- with gr.Row():
58
- user_message = gr.Textbox(label="🫡Your Message", lines=3, placeholder="Enter your math query here...")
59
- system_message = gr.Textbox(label="📉System Prompt", lines=2, placeholder="Optional: Set a scene or introduce a character...")
60
-
61
  gr.Button("Try🫡📉MetaMath").click(
62
  predict_math_bot,
63
  inputs=[user_message, system_message, max_new_tokens, temperature, top_p, repetition_penalty, do_sample],
64
  outputs=output_text
65
- )
 
 
 
 
 
66
 
67
  # Example placeholders
68
  # demo.add_examples([
 
24
  @torch.inference_mode()
25
  @spaces.GPU
26
  def predict_math_bot(user_message, system_message="", max_new_tokens=125, temperature=0.1, top_p=0.9, repetition_penalty=1.9, do_sample=False):
27
+ prompt = f"<s><INST>{user_message}{system_message}<|\INST|>" if system_message else user_message
28
  inputs = tokenizer(prompt, return_tensors='pt', add_special_tokens=False)
29
  input_ids = inputs["input_ids"].to(model.device)
30
 
 
53
  top_p = gr.Slider(label="Top-p (nucleus sampling)", value=0.90, minimum=0.01, maximum=0.99)
54
  repetition_penalty = gr.Slider(label="Repetition penalty", value=1.9, minimum=1.0, maximum=2.0)
55
  do_sample = gr.Checkbox(label="Uncheck for faster inference", value=False)
 
 
 
 
 
56
  gr.Button("Try🫡📉MetaMath").click(
57
  predict_math_bot,
58
  inputs=[user_message, system_message, max_new_tokens, temperature, top_p, repetition_penalty, do_sample],
59
  outputs=output_text
60
+ )
61
+ with gr.Row():
62
+ user_message = gr.Textbox(label="🫡Your Message", lines=3, placeholder="Enter your math query here...")
63
+ system_message = gr.Textbox(label="📉System Prompt", lines=2, placeholder="Optional: Set a scene or introduce a character...")
64
+
65
+
66
 
67
  # Example placeholders
68
  # demo.add_examples([