Spaces:
Runtime error
Runtime error
Alberto Carmona
commited on
Commit
·
95b13f9
1
Parent(s):
af90128
Increase the output tokens
Browse files- functions.py +1 -1
functions.py
CHANGED
@@ -32,7 +32,7 @@ def summarize_text(text: str):
|
|
32 |
print(['summarize_text', 'generating'])
|
33 |
with torch.cuda.amp.autocast():
|
34 |
output_tokens = model.generate(**batch,
|
35 |
-
max_new_tokens=
|
36 |
generation_config=generation_config
|
37 |
)
|
38 |
output = tokenizer.decode(output_tokens[0], skip_special_tokens=True)
|
|
|
32 |
print(['summarize_text', 'generating'])
|
33 |
with torch.cuda.amp.autocast():
|
34 |
output_tokens = model.generate(**batch,
|
35 |
+
max_new_tokens=512,
|
36 |
generation_config=generation_config
|
37 |
)
|
38 |
output = tokenizer.decode(output_tokens[0], skip_special_tokens=True)
|