nroggendorff commited on
Commit
a896396
·
verified ·
1 Parent(s): 7a51941

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -0
app.py CHANGED
@@ -1,5 +1,6 @@
1
  import gradio as gr
2
  import os
 
3
  import torch
4
  from transformers import AutoTokenizer, AutoModelForCausalLM
5
 
@@ -7,6 +8,7 @@ model_path = "cognitivecomputations/dolphin-2.8-mistral-7b-v02"
7
  tokenizer = AutoTokenizer.from_pretrained(model_path)
8
  model = AutoModelForCausalLM.from_pretrained(model_path)
9
 
 
10
  def chat(prompt):
11
  input_ids = tokenizer.encode(prompt, return_tensors="pt")
12
  output = model.generate(input_ids, max_length=1024, num_return_sequences=1, top_p=0.9, top_k=50, num_beams=2, early_stopping=True)
 
1
  import gradio as gr
2
  import os
3
+ import spaces
4
  import torch
5
  from transformers import AutoTokenizer, AutoModelForCausalLM
6
 
 
8
  tokenizer = AutoTokenizer.from_pretrained(model_path)
9
  model = AutoModelForCausalLM.from_pretrained(model_path)
10
 
11
+ @spaces.GPU
12
  def chat(prompt):
13
  input_ids = tokenizer.encode(prompt, return_tensors="pt")
14
  output = model.generate(input_ids, max_length=1024, num_return_sequences=1, top_p=0.9, top_k=50, num_beams=2, early_stopping=True)