Update app.py
Browse files
app.py
CHANGED
@@ -39,7 +39,7 @@ def calculate_readability(text: str) -> float:
|
|
39 |
return 206.835 - 1.015 * average_words_per_sentence
|
40 |
|
41 |
# Function to run inference and analysis
|
42 |
-
def run_inference_and_analysis(task, system_prompt, input_text, temperature, top_p, max_tokens,
|
43 |
# Initialize Pixeltable
|
44 |
pxt.drop_table('mistral_prompts', ignore_errors=True)
|
45 |
t = pxt.create_table('mistral_prompts', {
|
@@ -50,7 +50,6 @@ def run_inference_and_analysis(task, system_prompt, input_text, temperature, top
|
|
50 |
'temperature': pxt.Float,
|
51 |
'top_p': pxt.Float,
|
52 |
'max_tokens': pxt.Int,
|
53 |
-
'min_tokens': pxt.Int,
|
54 |
'stop': pxt.String,
|
55 |
'random_seed': pxt.Int,
|
56 |
'safe_prompt': pxt.Bool
|
@@ -65,7 +64,6 @@ def run_inference_and_analysis(task, system_prompt, input_text, temperature, top
|
|
65 |
'temperature': temperature,
|
66 |
'top_p': top_p,
|
67 |
'max_tokens': max_tokens,
|
68 |
-
'min_tokens': min_tokens,
|
69 |
'stop': stop,
|
70 |
'random_seed': random_seed,
|
71 |
'safe_prompt': safe_prompt
|
@@ -82,7 +80,6 @@ def run_inference_and_analysis(task, system_prompt, input_text, temperature, top
|
|
82 |
'temperature': temperature,
|
83 |
'top_p': top_p,
|
84 |
'max_tokens': max_tokens if max_tokens is not None else 300,
|
85 |
-
'min_tokens': min_tokens,
|
86 |
'stop': stop.split(',') if stop else None,
|
87 |
'random_seed': random_seed,
|
88 |
'safe_prompt': safe_prompt
|
@@ -128,7 +125,6 @@ def run_inference_and_analysis(task, system_prompt, input_text, temperature, top
|
|
128 |
t.temperature,
|
129 |
t.top_p,
|
130 |
t.max_tokens,
|
131 |
-
t.min_tokens,
|
132 |
t.stop,
|
133 |
t.random_seed,
|
134 |
t.safe_prompt
|
@@ -208,7 +204,6 @@ def gradio_interface():
|
|
208 |
temperature = gr.Slider(minimum=0, maximum=1, value=0.7, step=0.1, label="Temperature")
|
209 |
top_p = gr.Slider(minimum=0, maximum=1, value=0.9, step=0.1, label="Top P")
|
210 |
max_tokens = gr.Number(label="Max Tokens", value=300)
|
211 |
-
min_tokens = gr.Number(label="Min Tokens", value=None)
|
212 |
stop = gr.Textbox(label="Stop Sequences (comma-separated)")
|
213 |
random_seed = gr.Number(label="Random Seed", value=None)
|
214 |
safe_prompt = gr.Checkbox(label="Safe Prompt", value=False)
|
@@ -249,7 +244,6 @@ def gradio_interface():
|
|
249 |
"Temperature",
|
250 |
"Top P",
|
251 |
"Max Tokens",
|
252 |
-
"Min Tokens",
|
253 |
"Stop Sequences",
|
254 |
"Random Seed",
|
255 |
"Safe Prompt"
|
@@ -290,7 +284,7 @@ def gradio_interface():
|
|
290 |
|
291 |
gr.Examples(
|
292 |
examples=examples,
|
293 |
-
inputs=[task, system_prompt, input_text, temperature, top_p, max_tokens,
|
294 |
outputs=[omn_response, ml_response, large_sentiment, open_sentiment, large_keywords, open_keywords, large_readability, open_readability],
|
295 |
fn=run_inference_and_analysis,
|
296 |
cache_examples=True,
|
@@ -304,7 +298,7 @@ def gradio_interface():
|
|
304 |
|
305 |
submit_btn.click(
|
306 |
run_inference_and_analysis,
|
307 |
-
inputs=[task, system_prompt, input_text, temperature, top_p, max_tokens,
|
308 |
outputs=[omn_response, ml_response, large_sentiment, open_sentiment, large_keywords, open_keywords, large_readability, open_readability, history, responses, analysis, params]
|
309 |
)
|
310 |
|
|
|
39 |
return 206.835 - 1.015 * average_words_per_sentence
|
40 |
|
41 |
# Function to run inference and analysis
|
42 |
+
def run_inference_and_analysis(task, system_prompt, input_text, temperature, top_p, max_tokens, stop, random_seed, safe_prompt):
|
43 |
# Initialize Pixeltable
|
44 |
pxt.drop_table('mistral_prompts', ignore_errors=True)
|
45 |
t = pxt.create_table('mistral_prompts', {
|
|
|
50 |
'temperature': pxt.Float,
|
51 |
'top_p': pxt.Float,
|
52 |
'max_tokens': pxt.Int,
|
|
|
53 |
'stop': pxt.String,
|
54 |
'random_seed': pxt.Int,
|
55 |
'safe_prompt': pxt.Bool
|
|
|
64 |
'temperature': temperature,
|
65 |
'top_p': top_p,
|
66 |
'max_tokens': max_tokens,
|
|
|
67 |
'stop': stop,
|
68 |
'random_seed': random_seed,
|
69 |
'safe_prompt': safe_prompt
|
|
|
80 |
'temperature': temperature,
|
81 |
'top_p': top_p,
|
82 |
'max_tokens': max_tokens if max_tokens is not None else 300,
|
|
|
83 |
'stop': stop.split(',') if stop else None,
|
84 |
'random_seed': random_seed,
|
85 |
'safe_prompt': safe_prompt
|
|
|
125 |
t.temperature,
|
126 |
t.top_p,
|
127 |
t.max_tokens,
|
|
|
128 |
t.stop,
|
129 |
t.random_seed,
|
130 |
t.safe_prompt
|
|
|
204 |
temperature = gr.Slider(minimum=0, maximum=1, value=0.7, step=0.1, label="Temperature")
|
205 |
top_p = gr.Slider(minimum=0, maximum=1, value=0.9, step=0.1, label="Top P")
|
206 |
max_tokens = gr.Number(label="Max Tokens", value=300)
|
|
|
207 |
stop = gr.Textbox(label="Stop Sequences (comma-separated)")
|
208 |
random_seed = gr.Number(label="Random Seed", value=None)
|
209 |
safe_prompt = gr.Checkbox(label="Safe Prompt", value=False)
|
|
|
244 |
"Temperature",
|
245 |
"Top P",
|
246 |
"Max Tokens",
|
|
|
247 |
"Stop Sequences",
|
248 |
"Random Seed",
|
249 |
"Safe Prompt"
|
|
|
284 |
|
285 |
gr.Examples(
|
286 |
examples=examples,
|
287 |
+
inputs=[task, system_prompt, input_text, temperature, top_p, max_tokens, stop, random_seed, safe_prompt],
|
288 |
outputs=[omn_response, ml_response, large_sentiment, open_sentiment, large_keywords, open_keywords, large_readability, open_readability],
|
289 |
fn=run_inference_and_analysis,
|
290 |
cache_examples=True,
|
|
|
298 |
|
299 |
submit_btn.click(
|
300 |
run_inference_and_analysis,
|
301 |
+
inputs=[task, system_prompt, input_text, temperature, top_p, max_tokens, stop, random_seed, safe_prompt],
|
302 |
outputs=[omn_response, ml_response, large_sentiment, open_sentiment, large_keywords, open_keywords, large_readability, open_readability, history, responses, analysis, params]
|
303 |
)
|
304 |
|