Spaces:
Sleeping
Sleeping
going back
Browse files
app.py
CHANGED
@@ -39,15 +39,7 @@ def create_response_original(input_str,
|
|
39 |
# num_return_sequences=num_return_sequences)[0])
|
40 |
|
41 |
input_ids = tokenizer.encode(input_str + tokenizer.eos_token, return_tensors="pt")
|
42 |
-
output_ids = original_model.generate(**tokenizer(input_str+tokenizer.eos_token,return_tensors="pt",max_length=200)
|
43 |
-
do_sample=do_sample,
|
44 |
-
max_length=100,
|
45 |
-
temperature=temperature,
|
46 |
-
top_p=top_p,
|
47 |
-
top_k=top_k,
|
48 |
-
repetition_penalty=repetition_penalty,
|
49 |
-
num_return_sequences=num_return_sequences,
|
50 |
-
num_beams=num_beams)
|
51 |
outputs = []
|
52 |
for output_id in output_ids:
|
53 |
output = tokenizer.decode(output_id, skip_special_tokens=True)
|
|
|
39 |
# num_return_sequences=num_return_sequences)[0])
|
40 |
|
41 |
input_ids = tokenizer.encode(input_str + tokenizer.eos_token, return_tensors="pt")
|
42 |
+
output_ids = original_model.generate(**tokenizer(input_str+tokenizer.eos_token,return_tensors="pt",max_length=200))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
43 |
outputs = []
|
44 |
for output_id in output_ids:
|
45 |
output = tokenizer.decode(output_id, skip_special_tokens=True)
|