CedricZ commited on
Commit
beb0f21
·
verified ·
1 Parent(s): c775e8a

Upload 2 files

Browse files
Files changed (2) hide show
  1. app.py +45 -0
  2. requirements.txt +4 -0
app.py ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ from transformers import pipeline, GenerationConfig
3
+
4
+ #Load pipeline
5
+ generator = pipeline("text-generation", model="gpt2")
6
+ config = GenerationConfig.from_pretrained("gpt2")
7
+
8
+ def greet(prompt, temperature, max_lenth, top_p, samples):
9
+ #Enable greedy decoding if temperature is 0
10
+ config.do_sample = True if temperature != 0 else False
11
+
12
+ #Takes temperature and top_p value
13
+ config.temperature = temperature
14
+ config.top_p = top_p
15
+
16
+ #Samples
17
+ sample_input = {'Sample 1': 'Hi, this is a demo prompt for you',
18
+ 'Sample 2': 'Alber Einstein is a famous physicist graduated from',
19
+ 'Sample 3': 'University of Zurich locate at'}
20
+
21
+ #Choose between samples
22
+ if samples and prompt == '':
23
+ prompt = sample_input[samples]
24
+
25
+ a = generator(prompt, max_new_tokens=max_lenth, generation_config=config)
26
+ return a[0]['generated_text']
27
+
28
+ demo = gr.Interface(
29
+ fn=greet,
30
+ inputs=[gr.Textbox(placeholder = "Write a tagline for an ice cream shop.", label="prompt", lines=5),
31
+ gr.Slider(value=1, minimum=0, maximum=2, label='temperature',
32
+ info='''Temperature controls the randomness of the text generation.
33
+ 1.0 makes the model more likely to generate diverse and sometimes more unexpected outputs.
34
+ 0.0 makes the model's responses more deterministic and predictable.'''),
35
+ gr.Slider(value=16, minimum=1, maximum=256, step=1, label='max_lenth',
36
+ info='''Maximum number of tokens that the model will generate in the output.'''),
37
+ gr.Slider(value=1, minimum=0, maximum=1, label='top_p',
38
+ info='''Top-p controls the model's focus during text generation.
39
+ It allows only the most probable tokens to be considered for generation, where the cumulative probability of these tokens must exceed this value.'''),
40
+ gr.Dropdown(['Sample 1', 'Sample 2', 'Sample 3'], label="Sample Prompts",
41
+ info='''Some sample Prompts for you!''')],
42
+ outputs=[gr.Textbox(label='Output texts')],
43
+ )
44
+
45
+ demo.launch()
requirements.txt ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ gradio==4.28.3
2
+ gradio_client==0.16.0
3
+ huggingface-hub==0.22.2
4
+ transformers==4.39.3