jslin09 commited on
Commit
6ba7279
·
verified ·
1 Parent(s): 34ea07c

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -0
app.py CHANGED
@@ -11,6 +11,7 @@ model_name = "bloom-560m"
11
  model = AutoModelForCausalLM.from_pretrained(f'jslin09/{model_name}-finetuned-fraud').to('cuda')
12
  tokenizer = BloomTokenizerFast.from_pretrained(f'bigscience/{model_name}', bos_token = '<s>', eos_token = '</s>', pad_token = '<pad>')
13
 
 
14
  def rnd_generate(prompt):
15
  rnd_seed = random.randint(10, 500)
16
  set_seed(rnd_seed)
@@ -24,6 +25,7 @@ def rnd_generate(prompt):
24
  top_p=0.9)
25
  return tokenizer.decode(results[0])
26
 
 
27
  def generate(prompt):
28
  result_length = len(prompt) + 4
29
  inputs = tokenizer(prompt, return_tensors="pt") # 回傳的張量使用 Pytorch的格式。如果是 Tensorflow 格式的話,則指定為 "tf"。
 
11
  model = AutoModelForCausalLM.from_pretrained(f'jslin09/{model_name}-finetuned-fraud').to('cuda')
12
  tokenizer = BloomTokenizerFast.from_pretrained(f'bigscience/{model_name}', bos_token = '<s>', eos_token = '</s>', pad_token = '<pad>')
13
 
14
+ @spaces.GPU
15
  def rnd_generate(prompt):
16
  rnd_seed = random.randint(10, 500)
17
  set_seed(rnd_seed)
 
25
  top_p=0.9)
26
  return tokenizer.decode(results[0])
27
 
28
+ @spaces.GPU
29
  def generate(prompt):
30
  result_length = len(prompt) + 4
31
  inputs = tokenizer(prompt, return_tensors="pt") # 回傳的張量使用 Pytorch的格式。如果是 Tensorflow 格式的話,則指定為 "tf"。