Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -7,12 +7,12 @@ import spaces
|
|
7 |
import torch
|
8 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
9 |
|
10 |
-
tokenizer = AutoTokenizer.from_pretrained("
|
11 |
-
model = AutoModelForCausalLM.from_pretrained("
|
12 |
|
13 |
@spaces.GPU
|
14 |
def text_generation(input_text, seed):
|
15 |
-
input_ids = tokenizer(input_text, return_tensors="pt")
|
16 |
torch.manual_seed(seed) # Max value: 18446744073709551615
|
17 |
outputs = model.generate(input_ids, do_sample=True, max_length=100)
|
18 |
generated_text = tokenizer.batch_decode(outputs, skip_special_tokens=True)
|
|
|
7 |
import torch
|
8 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
9 |
|
10 |
+
tokenizer = AutoTokenizer.from_pretrained("appvoid/palmer-004")
|
11 |
+
model = AutoModelForCausalLM.from_pretrained("appvoid/palmer-004")
|
12 |
|
13 |
@spaces.GPU
|
14 |
def text_generation(input_text, seed):
|
15 |
+
input_ids = tokenizer(input_text, return_tensors="pt")
|
16 |
torch.manual_seed(seed) # Max value: 18446744073709551615
|
17 |
outputs = model.generate(input_ids, do_sample=True, max_length=100)
|
18 |
generated_text = tokenizer.batch_decode(outputs, skip_special_tokens=True)
|