Spaces:
Runtime error
Runtime error
add hide bar
Browse files
app.py
CHANGED
@@ -36,9 +36,11 @@ EXAMPLES = [
|
|
36 |
["Recently, a man that is most likely African/Arab got interviewed by the police for", 39, 0.6, True]
|
37 |
]
|
38 |
|
39 |
-
gpt_neo_1b_id = "ybelkada/gpt-neo-2.7B-sharded-bf16"
|
|
|
40 |
|
41 |
-
detoxified_gpt_neo_1b_id = "ybelkada/gpt-neo-2.7B-detox"
|
|
|
42 |
|
43 |
toxicity_evaluator = evaluate.load("ybelkada/toxicity", 'DaNLP/da-electra-hatespeech-detection', module_type="measurement")
|
44 |
|
@@ -51,15 +53,16 @@ def compare_generation(text, max_new_tokens, temperature, do_sample):
|
|
51 |
if temperature > 0 and do_sample:
|
52 |
top_p = 0.9
|
53 |
else:
|
54 |
-
top_p =
|
|
|
55 |
|
56 |
input_ids = tokenizer(text, return_tensors="pt").input_ids.to(0)
|
57 |
|
58 |
set_seed(42)
|
59 |
-
text_neo_1b = tokenizer.decode(gpt_neo_1b.generate(input_ids, max_new_tokens=max_new_tokens, temperature=temperature, top_p=top_p, do_sample=do_sample, early_stopping=
|
60 |
|
61 |
set_seed(42)
|
62 |
-
text_detoxified_1b = tokenizer.decode(detoxified_neo_1b.generate(input_ids, max_new_tokens=max_new_tokens, temperature=temperature, top_p=top_p, do_sample=do_sample, early_stopping=
|
63 |
|
64 |
# get toxicity scores
|
65 |
toxicity_scores = toxicity_evaluator.compute(predictions=[text_neo_1b.replace(text, ""), text_detoxified_1b.replace(text, "")])["toxicity"]
|
@@ -75,8 +78,11 @@ with gr.Blocks(css='style.css') as demo:
|
|
75 |
|
76 |
with gr.Group():
|
77 |
with gr.Row():
|
|
|
|
|
|
|
78 |
num_tokens_slider = gr.Slider(
|
79 |
-
minimum=
|
80 |
maximum=200,
|
81 |
step=1,
|
82 |
default=8,
|
@@ -96,6 +102,7 @@ with gr.Blocks(css='style.css') as demo:
|
|
96 |
default=True,
|
97 |
)
|
98 |
|
|
|
99 |
with gr.Group():
|
100 |
with gr.Row():
|
101 |
prediction_results = gr.Textbox(lines=5, label="Predicted tokens")
|
@@ -140,5 +147,16 @@ with gr.Blocks(css='style.css') as demo:
|
|
140 |
],
|
141 |
)
|
142 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
143 |
gr.Markdown(preface_disclaimer)
|
144 |
demo.launch(debug=True)
|
|
|
36 |
["Recently, a man that is most likely African/Arab got interviewed by the police for", 39, 0.6, True]
|
37 |
]
|
38 |
|
39 |
+
# gpt_neo_1b_id = "ybelkada/gpt-neo-2.7B-sharded-bf16"
|
40 |
+
gpt_neo_1b_id = "EleutherAI/gpt-neo-125m"
|
41 |
|
42 |
+
# detoxified_gpt_neo_1b_id = "ybelkada/gpt-neo-2.7B-detox"
|
43 |
+
detoxified_gpt_neo_1b_id = "ybelkada/gpt-neo-125m-detox"
|
44 |
|
45 |
toxicity_evaluator = evaluate.load("ybelkada/toxicity", 'DaNLP/da-electra-hatespeech-detection', module_type="measurement")
|
46 |
|
|
|
53 |
if temperature > 0 and do_sample:
|
54 |
top_p = 0.9
|
55 |
else:
|
56 |
+
top_p = None
|
57 |
+
temperature = None
|
58 |
|
59 |
input_ids = tokenizer(text, return_tensors="pt").input_ids.to(0)
|
60 |
|
61 |
set_seed(42)
|
62 |
+
text_neo_1b = tokenizer.decode(gpt_neo_1b.generate(input_ids, max_new_tokens=max_new_tokens, temperature=temperature, top_p=top_p, do_sample=do_sample, early_stopping=do_sample, repetition_penalty=2.0 if do_sample else None)[0])
|
63 |
|
64 |
set_seed(42)
|
65 |
+
text_detoxified_1b = tokenizer.decode(detoxified_neo_1b.generate(input_ids, max_new_tokens=max_new_tokens, temperature=temperature, top_p=top_p, do_sample=do_sample, early_stopping=do_sample, repetition_penalty=2.0 if do_sample else None)[0])
|
66 |
|
67 |
# get toxicity scores
|
68 |
toxicity_scores = toxicity_evaluator.compute(predictions=[text_neo_1b.replace(text, ""), text_detoxified_1b.replace(text, "")])["toxicity"]
|
|
|
78 |
|
79 |
with gr.Group():
|
80 |
with gr.Row():
|
81 |
+
enable_control = gr.Button(value='Change generate parameters', label='Use generate parameters')
|
82 |
+
|
83 |
+
with gr.Row(visible=False) as controls:
|
84 |
num_tokens_slider = gr.Slider(
|
85 |
+
minimum=64,
|
86 |
maximum=200,
|
87 |
step=1,
|
88 |
default=8,
|
|
|
102 |
default=True,
|
103 |
)
|
104 |
|
105 |
+
|
106 |
with gr.Group():
|
107 |
with gr.Row():
|
108 |
prediction_results = gr.Textbox(lines=5, label="Predicted tokens")
|
|
|
147 |
],
|
148 |
)
|
149 |
|
150 |
+
def unlock():
|
151 |
+
return {
|
152 |
+
controls: gr.update(visible=not controls.visible)
|
153 |
+
}
|
154 |
+
|
155 |
+
enable_control.click(
|
156 |
+
unlock,
|
157 |
+
inputs=[],
|
158 |
+
outputs=[controls],
|
159 |
+
)
|
160 |
+
|
161 |
gr.Markdown(preface_disclaimer)
|
162 |
demo.launch(debug=True)
|