jslin09 commited on
Commit
a892b5e
·
verified ·
1 Parent(s): ff977d0

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -5
app.py CHANGED
@@ -7,11 +7,11 @@ import random
7
  # import torch
8
 
9
  # device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
10
- model_name = "bloom-560m"
11
- model = AutoModelForCausalLM.from_pretrained(f'jslin09/{model_name}-finetuned-fraud').to('cuda')
12
- tokenizer = BloomTokenizerFast.from_pretrained(f'bigscience/{model_name}', bos_token = '<s>', eos_token = '</s>', pad_token = '<pad>')
13
- #model = AutoModelForCausalLM.from_pretrained("jslin09/gemma2-2b-fraud").to('cuda')
14
- #tokenizer = AutoTokenizer.from_pretrained("jslin09/gemma2-2b-fraud", bos_token = '<bos>', eos_token = '<eos>', pad_token = '<pad>')
15
 
16
  @spaces.GPU
17
  def rnd_generate(prompt):
@@ -45,6 +45,7 @@ def generate(prompt):
45
  top_p=0.9
46
  )
47
  try:
 
48
  result = tokenizer.decode(results[0]).split("<bos>")[1].split("<eos>")[0]
49
  except IndexError:
50
  result = tokenizer.decode(results[0]).split("<bos>")[1]
 
7
  # import torch
8
 
9
  # device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
10
+ #model_name = "bloom-560m"
11
+ #model = AutoModelForCausalLM.from_pretrained(f'jslin09/{model_name}-finetuned-fraud').to('cuda')
12
+ #tokenizer = BloomTokenizerFast.from_pretrained(f'bigscience/{model_name}', bos_token = '<s>', eos_token = '</s>', pad_token = '<pad>')
13
+ model = AutoModelForCausalLM.from_pretrained("jslin09/gemma2-2b-fraud").to('cuda')
14
+ tokenizer = AutoTokenizer.from_pretrained("jslin09/gemma2-2b-fraud", bos_token = '<bos>', eos_token = '<eos>', pad_token = '<pad>')
15
 
16
  @spaces.GPU
17
  def rnd_generate(prompt):
 
45
  top_p=0.9
46
  )
47
  try:
48
+ print(results)
49
  result = tokenizer.decode(results[0]).split("<bos>")[1].split("<eos>")[0]
50
  except IndexError:
51
  result = tokenizer.decode(results[0]).split("<bos>")[1]