Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -7,11 +7,11 @@ import random
|
|
7 |
# import torch
|
8 |
|
9 |
# device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
|
10 |
-
model_name = "bloom-560m"
|
11 |
-
model = AutoModelForCausalLM.from_pretrained(f'jslin09/{model_name}-finetuned-fraud').to('cuda')
|
12 |
-
tokenizer = BloomTokenizerFast.from_pretrained(f'bigscience/{model_name}', bos_token = '<s>', eos_token = '</s>', pad_token = '<pad>')
|
13 |
-
|
14 |
-
|
15 |
|
16 |
@spaces.GPU
|
17 |
def rnd_generate(prompt):
|
@@ -45,6 +45,7 @@ def generate(prompt):
|
|
45 |
top_p=0.9
|
46 |
)
|
47 |
try:
|
|
|
48 |
result = tokenizer.decode(results[0]).split("<bos>")[1].split("<eos>")[0]
|
49 |
except IndexError:
|
50 |
result = tokenizer.decode(results[0]).split("<bos>")[1]
|
|
|
7 |
# import torch
|
8 |
|
9 |
# device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
|
10 |
+
#model_name = "bloom-560m"
|
11 |
+
#model = AutoModelForCausalLM.from_pretrained(f'jslin09/{model_name}-finetuned-fraud').to('cuda')
|
12 |
+
#tokenizer = BloomTokenizerFast.from_pretrained(f'bigscience/{model_name}', bos_token = '<s>', eos_token = '</s>', pad_token = '<pad>')
|
13 |
+
model = AutoModelForCausalLM.from_pretrained("jslin09/gemma2-2b-fraud").to('cuda')
|
14 |
+
tokenizer = AutoTokenizer.from_pretrained("jslin09/gemma2-2b-fraud", bos_token = '<bos>', eos_token = '<eos>', pad_token = '<pad>')
|
15 |
|
16 |
@spaces.GPU
|
17 |
def rnd_generate(prompt):
|
|
|
45 |
top_p=0.9
|
46 |
)
|
47 |
try:
|
48 |
+
print(results)
|
49 |
result = tokenizer.decode(results[0]).split("<bos>")[1].split("<eos>")[0]
|
50 |
except IndexError:
|
51 |
result = tokenizer.decode(results[0]).split("<bos>")[1]
|