Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -4,11 +4,11 @@ from transformers import AutoModelForCausalLM
|
|
4 |
from transformers import BloomTokenizerFast
|
5 |
from transformers import pipeline, set_seed
|
6 |
import random
|
7 |
-
import torch
|
8 |
|
9 |
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
|
10 |
model_name = "bloom-560m"
|
11 |
-
model = AutoModelForCausalLM.from_pretrained(f'jslin09/{model_name}-finetuned-fraud').to(
|
12 |
tokenizer = BloomTokenizerFast.from_pretrained(f'bigscience/{model_name}', bos_token = '<s>', eos_token = '</s>', pad_token = '<pad>')
|
13 |
|
14 |
def rnd_generate(prompt):
|
|
|
4 |
from transformers import BloomTokenizerFast
|
5 |
from transformers import pipeline, set_seed
|
6 |
import random
|
7 |
+
# import torch
|
8 |
|
9 |
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
|
10 |
model_name = "bloom-560m"
|
11 |
+
model = AutoModelForCausalLM.from_pretrained(f'jslin09/{model_name}-finetuned-fraud').to('cuda')
|
12 |
tokenizer = BloomTokenizerFast.from_pretrained(f'bigscience/{model_name}', bos_token = '<s>', eos_token = '</s>', pad_token = '<pad>')
|
13 |
|
14 |
def rnd_generate(prompt):
|