Spaces:
Build error
Build error
Update app.py
Browse files
app.py
CHANGED
@@ -29,7 +29,7 @@ bos_token = tokenizer.bos_token
|
|
29 |
model = AutoModelForCausalLM.from_pretrained(model_id).to(device)
|
30 |
model.generation_config.pad_token_id = tokenizer.pad_token_id
|
31 |
|
32 |
-
torch.manual_seed(
|
33 |
|
34 |
@spaces.GPU
|
35 |
def generate_fakypedia(article_title: str):
|
@@ -37,7 +37,7 @@ def generate_fakypedia(article_title: str):
|
|
37 |
result = ""
|
38 |
string_to_tokenize= f"{bos_token}\\%{article_title}"
|
39 |
input_ids = tokenizer(string_to_tokenize, return_tensors="pt").input_ids.to(device)
|
40 |
-
sample_outputs = model.generate(input_ids, do_sample=True,repetition_penalty=1.
|
41 |
if article_title == None or len(article_title) == 0:
|
42 |
result += f"# תוצאות פייקיפדיה עם כותרות אקראיות \n"
|
43 |
article_title = ""
|
|
|
29 |
model = AutoModelForCausalLM.from_pretrained(model_id).to(device)
|
30 |
model.generation_config.pad_token_id = tokenizer.pad_token_id
|
31 |
|
32 |
+
torch.manual_seed(1234)
|
33 |
|
34 |
@spaces.GPU
|
35 |
def generate_fakypedia(article_title: str):
|
|
|
37 |
result = ""
|
38 |
string_to_tokenize= f"{bos_token}\\%{article_title}"
|
39 |
input_ids = tokenizer(string_to_tokenize, return_tensors="pt").input_ids.to(device)
|
40 |
+
sample_outputs = model.generate(input_ids, do_sample=True,repetition_penalty=1.2, temperature=0.50, max_length=128, num_return_sequences=3)
|
41 |
if article_title == None or len(article_title) == 0:
|
42 |
result += f"# תוצאות פייקיפדיה עם כותרות אקראיות \n"
|
43 |
article_title = ""
|