Spaces:
Runtime error
Runtime error
hanzla javaid
commited on
Commit
·
97e6085
1
Parent(s):
638e465
test
Browse files
app.py
CHANGED
@@ -7,9 +7,9 @@ import spaces
|
|
7 |
loaded_models = {}
|
8 |
|
9 |
# List of available models (update with your preferred models)
|
10 |
-
models = ["
|
11 |
-
|
12 |
|
|
|
13 |
def load_model(model_name):
|
14 |
if model_name not in loaded_models:
|
15 |
print(f"Loading model: {model_name}")
|
@@ -23,10 +23,9 @@ def get_model_response(model_name, message):
|
|
23 |
model, tokenizer = load_model(model_name)
|
24 |
inputs = tokenizer(message, return_tensors="pt").to(model.device)
|
25 |
|
26 |
-
|
27 |
-
outputs = model.generate(**inputs, max_length=100, num_return_sequences=1, temperature=0.7)
|
28 |
|
29 |
-
response = tokenizer.decode(outputs[0]
|
30 |
return response
|
31 |
|
32 |
|
|
|
7 |
loaded_models = {}
|
8 |
|
9 |
# List of available models (update with your preferred models)
|
10 |
+
models = ["hanzla/gemma-2b-datascience-instruct-v5", "hanzla/gemma-2b-datascience-instruct-v4.5"]
|
|
|
11 |
|
12 |
+
@spaces.GPU
|
13 |
def load_model(model_name):
|
14 |
if model_name not in loaded_models:
|
15 |
print(f"Loading model: {model_name}")
|
|
|
23 |
model, tokenizer = load_model(model_name)
|
24 |
inputs = tokenizer(message, return_tensors="pt").to(model.device)
|
25 |
|
26 |
+
outputs = model.generate(**inputs)
|
|
|
27 |
|
28 |
+
response = tokenizer.decode(outputs[0])
|
29 |
return response
|
30 |
|
31 |
|