PierreBrunelle commited on
Commit
d6f2992
·
verified ·
1 Parent(s): f519aed

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -9
app.py CHANGED
@@ -39,7 +39,7 @@ def calculate_readability(text: str) -> float:
39
  return 206.835 - 1.015 * average_words_per_sentence
40
 
41
  # Function to run inference and analysis
42
- def run_inference_and_analysis(task, system_prompt, input_text, temperature, top_p, max_tokens, min_tokens, stop, random_seed, safe_prompt):
43
  # Initialize Pixeltable
44
  pxt.drop_table('mistral_prompts', ignore_errors=True)
45
  t = pxt.create_table('mistral_prompts', {
@@ -50,7 +50,6 @@ def run_inference_and_analysis(task, system_prompt, input_text, temperature, top
50
  'temperature': pxt.Float,
51
  'top_p': pxt.Float,
52
  'max_tokens': pxt.Int,
53
- 'min_tokens': pxt.Int,
54
  'stop': pxt.String,
55
  'random_seed': pxt.Int,
56
  'safe_prompt': pxt.Bool
@@ -65,7 +64,6 @@ def run_inference_and_analysis(task, system_prompt, input_text, temperature, top
65
  'temperature': temperature,
66
  'top_p': top_p,
67
  'max_tokens': max_tokens,
68
- 'min_tokens': min_tokens,
69
  'stop': stop,
70
  'random_seed': random_seed,
71
  'safe_prompt': safe_prompt
@@ -82,7 +80,6 @@ def run_inference_and_analysis(task, system_prompt, input_text, temperature, top
82
  'temperature': temperature,
83
  'top_p': top_p,
84
  'max_tokens': max_tokens if max_tokens is not None else 300,
85
- 'min_tokens': min_tokens,
86
  'stop': stop.split(',') if stop else None,
87
  'random_seed': random_seed,
88
  'safe_prompt': safe_prompt
@@ -128,7 +125,6 @@ def run_inference_and_analysis(task, system_prompt, input_text, temperature, top
128
  t.temperature,
129
  t.top_p,
130
  t.max_tokens,
131
- t.min_tokens,
132
  t.stop,
133
  t.random_seed,
134
  t.safe_prompt
@@ -208,7 +204,6 @@ def gradio_interface():
208
  temperature = gr.Slider(minimum=0, maximum=1, value=0.7, step=0.1, label="Temperature")
209
  top_p = gr.Slider(minimum=0, maximum=1, value=0.9, step=0.1, label="Top P")
210
  max_tokens = gr.Number(label="Max Tokens", value=300)
211
- min_tokens = gr.Number(label="Min Tokens", value=None)
212
  stop = gr.Textbox(label="Stop Sequences (comma-separated)")
213
  random_seed = gr.Number(label="Random Seed", value=None)
214
  safe_prompt = gr.Checkbox(label="Safe Prompt", value=False)
@@ -249,7 +244,6 @@ def gradio_interface():
249
  "Temperature",
250
  "Top P",
251
  "Max Tokens",
252
- "Min Tokens",
253
  "Stop Sequences",
254
  "Random Seed",
255
  "Safe Prompt"
@@ -290,7 +284,7 @@ def gradio_interface():
290
 
291
  gr.Examples(
292
  examples=examples,
293
- inputs=[task, system_prompt, input_text, temperature, top_p, max_tokens, min_tokens, stop, random_seed, safe_prompt],
294
  outputs=[omn_response, ml_response, large_sentiment, open_sentiment, large_keywords, open_keywords, large_readability, open_readability],
295
  fn=run_inference_and_analysis,
296
  cache_examples=True,
@@ -304,7 +298,7 @@ def gradio_interface():
304
 
305
  submit_btn.click(
306
  run_inference_and_analysis,
307
- inputs=[task, system_prompt, input_text, temperature, top_p, max_tokens, min_tokens, stop, random_seed, safe_prompt],
308
  outputs=[omn_response, ml_response, large_sentiment, open_sentiment, large_keywords, open_keywords, large_readability, open_readability, history, responses, analysis, params]
309
  )
310
 
 
39
  return 206.835 - 1.015 * average_words_per_sentence
40
 
41
  # Function to run inference and analysis
42
+ def run_inference_and_analysis(task, system_prompt, input_text, temperature, top_p, max_tokens, stop, random_seed, safe_prompt):
43
  # Initialize Pixeltable
44
  pxt.drop_table('mistral_prompts', ignore_errors=True)
45
  t = pxt.create_table('mistral_prompts', {
 
50
  'temperature': pxt.Float,
51
  'top_p': pxt.Float,
52
  'max_tokens': pxt.Int,
 
53
  'stop': pxt.String,
54
  'random_seed': pxt.Int,
55
  'safe_prompt': pxt.Bool
 
64
  'temperature': temperature,
65
  'top_p': top_p,
66
  'max_tokens': max_tokens,
 
67
  'stop': stop,
68
  'random_seed': random_seed,
69
  'safe_prompt': safe_prompt
 
80
  'temperature': temperature,
81
  'top_p': top_p,
82
  'max_tokens': max_tokens if max_tokens is not None else 300,
 
83
  'stop': stop.split(',') if stop else None,
84
  'random_seed': random_seed,
85
  'safe_prompt': safe_prompt
 
125
  t.temperature,
126
  t.top_p,
127
  t.max_tokens,
 
128
  t.stop,
129
  t.random_seed,
130
  t.safe_prompt
 
204
  temperature = gr.Slider(minimum=0, maximum=1, value=0.7, step=0.1, label="Temperature")
205
  top_p = gr.Slider(minimum=0, maximum=1, value=0.9, step=0.1, label="Top P")
206
  max_tokens = gr.Number(label="Max Tokens", value=300)
 
207
  stop = gr.Textbox(label="Stop Sequences (comma-separated)")
208
  random_seed = gr.Number(label="Random Seed", value=None)
209
  safe_prompt = gr.Checkbox(label="Safe Prompt", value=False)
 
244
  "Temperature",
245
  "Top P",
246
  "Max Tokens",
 
247
  "Stop Sequences",
248
  "Random Seed",
249
  "Safe Prompt"
 
284
 
285
  gr.Examples(
286
  examples=examples,
287
+ inputs=[task, system_prompt, input_text, temperature, top_p, max_tokens, stop, random_seed, safe_prompt],
288
  outputs=[omn_response, ml_response, large_sentiment, open_sentiment, large_keywords, open_keywords, large_readability, open_readability],
289
  fn=run_inference_and_analysis,
290
  cache_examples=True,
 
298
 
299
  submit_btn.click(
300
  run_inference_and_analysis,
301
+ inputs=[task, system_prompt, input_text, temperature, top_p, max_tokens, stop, random_seed, safe_prompt],
302
  outputs=[omn_response, ml_response, large_sentiment, open_sentiment, large_keywords, open_keywords, large_readability, open_readability, history, responses, analysis, params]
303
  )
304