as-cle-bert commited on
Commit
8e63a33
·
verified ·
1 Parent(s): 5bb3730

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -5
app.py CHANGED
@@ -58,7 +58,7 @@ def replyA(prompt, history, modelA):
58
  df = pd.read_csv("models.csv")
59
  df.loc[df["MODEL"] == modelA, "MATCHES_PLAYED"] += 1
60
  df.to_csv("models.csv", index=False)
61
- pipeA = pipeline("text-generation", model=models_and_tokenizers[modelA][0], tokenizer=models_and_tokenizers[modelA][1], max_new_tokens=512, repetition_penalty=1.5, temperature=0.5, device="cuda")
62
  responseA = run_inference(pipeA, prompt)
63
  r = ''
64
  for c in responseA:
@@ -72,7 +72,7 @@ def replyB(prompt, history, modelB):
72
  df = pd.read_csv("models.csv")
73
  df.loc[df["MODEL"] == modelB, "MATCHES_PLAYED"] += 1
74
  df.to_csv("models.csv", index=False)
75
- pipeB = pipeline("text-generation", model=models_and_tokenizers[modelB][0], tokenizer=models_and_tokenizers[modelB][1], max_new_tokens=512, repetition_penalty=1.5, temperature=0.5, device="cuda")
76
  responseB = run_inference(pipeB, prompt)
77
  r = ''
78
  for c in responseB:
@@ -106,9 +106,7 @@ with gr.Blocks() as demo1:
106
  btnA.click(modelA_button, inputs=None, outputs=None)
107
  btnB.click(modelB_button, inputs=None, outputs=None)
108
 
109
- demo = gr.TabbedInterface([demo1, demo2], ["Chat Arena", "Leaderboard"], title="""<h1 align='center'>SmolLM Arena</h1>
110
- <h2 align='center'>Cast your vote to choose the best Small Language Model (100M-1.7B)!🚀</h2>
111
- <h3 align='center'>[<a href="https://github.com/AstraBert/smollm-arena">GitHub</a>] [<a href="https://github.com/AstraBert/smollm-arena?tab=readme-ov-file#usage">Usage Guide</a>]""")
112
 
113
  if __name__ == "__main__":
114
  demo.launch(server_name="0.0.0.0", server_port=7860)
 
58
  df = pd.read_csv("models.csv")
59
  df.loc[df["MODEL"] == modelA, "MATCHES_PLAYED"] += 1
60
  df.to_csv("models.csv", index=False)
61
+ pipeA = pipeline("text-generation", model=models_and_tokenizers[modelA][0], tokenizer=models_and_tokenizers[modelA][1], max_new_tokens=512, repetition_penalty=1.5, temperature=0.5, device_map="cuda:0")
62
  responseA = run_inference(pipeA, prompt)
63
  r = ''
64
  for c in responseA:
 
72
  df = pd.read_csv("models.csv")
73
  df.loc[df["MODEL"] == modelB, "MATCHES_PLAYED"] += 1
74
  df.to_csv("models.csv", index=False)
75
+ pipeB = pipeline("text-generation", model=models_and_tokenizers[modelB][0], tokenizer=models_and_tokenizers[modelB][1], max_new_tokens=512, repetition_penalty=1.5, temperature=0.5, device_map="cuda:0")
76
  responseB = run_inference(pipeB, prompt)
77
  r = ''
78
  for c in responseB:
 
106
  btnA.click(modelA_button, inputs=None, outputs=None)
107
  btnB.click(modelB_button, inputs=None, outputs=None)
108
 
109
+ demo = gr.TabbedInterface([demo1, demo2], ["Chat Arena", "Leaderboard"])
 
 
110
 
111
  if __name__ == "__main__":
112
  demo.launch(server_name="0.0.0.0", server_port=7860)