hanzla javaid commited on
Commit
97e6085
·
1 Parent(s): 638e465
Files changed (1) hide show
  1. app.py +4 -5
app.py CHANGED
@@ -7,9 +7,9 @@ import spaces
7
  loaded_models = {}
8
 
9
  # List of available models (update with your preferred models)
10
- models = ["gpt2", "gpt2-medium", "gpt2-large", "EleutherAI/gpt-neo-1.3B"]
11
-
12
 
 
13
  def load_model(model_name):
14
  if model_name not in loaded_models:
15
  print(f"Loading model: {model_name}")
@@ -23,10 +23,9 @@ def get_model_response(model_name, message):
23
  model, tokenizer = load_model(model_name)
24
  inputs = tokenizer(message, return_tensors="pt").to(model.device)
25
 
26
- with torch.no_grad():
27
- outputs = model.generate(**inputs, max_length=100, num_return_sequences=1, temperature=0.7)
28
 
29
- response = tokenizer.decode(outputs[0], skip_special_tokens=True)
30
  return response
31
 
32
 
 
7
  loaded_models = {}
8
 
9
  # List of available models (update with your preferred models)
10
+ models = ["hanzla/gemma-2b-datascience-instruct-v5", "hanzla/gemma-2b-datascience-instruct-v4.5"]
 
11
 
12
+ @spaces.GPU
13
  def load_model(model_name):
14
  if model_name not in loaded_models:
15
  print(f"Loading model: {model_name}")
 
23
  model, tokenizer = load_model(model_name)
24
  inputs = tokenizer(message, return_tensors="pt").to(model.device)
25
 
26
+ outputs = model.generate(**inputs)
 
27
 
28
+ response = tokenizer.decode(outputs[0])
29
  return response
30
 
31