Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -9,10 +9,10 @@ tokenizer = AutoTokenizer.from_pretrained("distilgpt2")
|
|
9 |
def generate_text(inputs):
|
10 |
responses = []
|
11 |
for input_text in inputs:
|
12 |
-
input_tensor = tokenizer(input_text, return_tensors="pt")
|
13 |
output = model.generate(**input_tensor)
|
14 |
responses.append(tokenizer.decode(output[0], skip_special_tokens=True))
|
15 |
return responses
|
16 |
|
17 |
iface = gr.Interface(fn=generate_text, inputs="text", outputs="text", allow_flagging="never")
|
18 |
-
iface.launch(share=True)
|
|
|
9 |
def generate_text(inputs):
|
10 |
responses = []
|
11 |
for input_text in inputs:
|
12 |
+
input_tensor = tokenizer(input_text, return_tensors="pt", clean_up_tokenization_spaces=True)
|
13 |
output = model.generate(**input_tensor)
|
14 |
responses.append(tokenizer.decode(output[0], skip_special_tokens=True))
|
15 |
return responses
|
16 |
|
17 |
iface = gr.Interface(fn=generate_text, inputs="text", outputs="text", allow_flagging="never")
|
18 |
+
iface.launch(share=True) # Adjust share=True based on your usage context
|