Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -27,7 +27,8 @@ AVAILABLE_MODELS = {
|
|
27 |
"opt-350m": "facebook/opt-350m",
|
28 |
"pythia-160m": "EleutherAI/pythia-160m",
|
29 |
"mistral-7b": "mistralai/Mistral-7B-Instruct-v0.3",
|
30 |
-
"deepseek-small": "deepseek-ai/DeepSeek-V2-Lite"
|
|
|
31 |
}
|
32 |
|
33 |
# Initialize model and tokenizer
|
@@ -45,7 +46,7 @@ def load_model(model_name):
|
|
45 |
current_model_name = model_name
|
46 |
|
47 |
# Load the default model at startup
|
48 |
-
load_model("
|
49 |
|
50 |
@spaces.GPU()
|
51 |
def get_next_token_predictions(text, model_name, top_k=10):
|
|
|
27 |
"opt-350m": "facebook/opt-350m",
|
28 |
"pythia-160m": "EleutherAI/pythia-160m",
|
29 |
"mistral-7b": "mistralai/Mistral-7B-Instruct-v0.3",
|
30 |
+
"deepseek-small": "deepseek-ai/DeepSeek-V2-Lite",
|
31 |
+
"llama": "meta-llama/Llama-3.2-1B"
|
32 |
}
|
33 |
|
34 |
# Initialize model and tokenizer
|
|
|
46 |
current_model_name = model_name
|
47 |
|
48 |
# Load the default model at startup
|
49 |
+
load_model("mistral-7b")
|
50 |
|
51 |
@spaces.GPU()
|
52 |
def get_next_token_predictions(text, model_name, top_k=10):
|