jslin09 commited on
Commit
ff977d0
·
verified ·
1 Parent(s): 1eef8af

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -5
app.py CHANGED
@@ -7,11 +7,11 @@ import random
7
  # import torch
8
 
9
  # device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
10
- #model_name = "bloom-560m"
11
- #model = AutoModelForCausalLM.from_pretrained(f'jslin09/{model_name}-finetuned-fraud').to('cuda')
12
- #tokenizer = BloomTokenizerFast.from_pretrained(f'bigscience/{model_name}', bos_token = '<s>', eos_token = '</s>', pad_token = '<pad>')
13
- model = AutoModelForCausalLM.from_pretrained("jslin09/gemma2-2b-fraud").to('cuda')
14
- tokenizer = AutoTokenizer.from_pretrained("jslin09/gemma2-2b-fraud", bos_token = '<bos>', eos_token = '<eos>', pad_token = '<pad>')
15
 
16
  @spaces.GPU
17
  def rnd_generate(prompt):
 
7
  # import torch
8
 
9
  # device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
10
+ model_name = "bloom-560m"
11
+ model = AutoModelForCausalLM.from_pretrained(f'jslin09/{model_name}-finetuned-fraud').to('cuda')
12
+ tokenizer = BloomTokenizerFast.from_pretrained(f'bigscience/{model_name}', bos_token = '<s>', eos_token = '</s>', pad_token = '<pad>')
13
+ #model = AutoModelForCausalLM.from_pretrained("jslin09/gemma2-2b-fraud").to('cuda')
14
+ #tokenizer = AutoTokenizer.from_pretrained("jslin09/gemma2-2b-fraud", bos_token = '<bos>', eos_token = '<eos>', pad_token = '<pad>')
15
 
16
  @spaces.GPU
17
  def rnd_generate(prompt):