Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -5,7 +5,7 @@ from llama_cpp import Llama # GGUF CPU backend
|
|
5 |
# ---------- model loading (one‑time) ----------
|
6 |
MODEL_REPO = "MaziyarPanahi/gemma-2b-it-GGUF"
|
7 |
MODEL_FILE = "gemma-2b-it.Q4_K_M.gguf" # 1.6 GB 4‑bit
|
8 |
-
CTX_SIZE =
|
9 |
|
10 |
model_path = hf_hub_download(repo_id=MODEL_REPO, filename=MODEL_FILE)
|
11 |
llm = Llama(
|
|
|
5 |
# ---------- model loading (one‑time) ----------
|
6 |
MODEL_REPO = "MaziyarPanahi/gemma-2b-it-GGUF"
|
7 |
MODEL_FILE = "gemma-2b-it.Q4_K_M.gguf" # 1.6 GB 4‑bit
|
8 |
+
CTX_SIZE = 2048
|
9 |
|
10 |
model_path = hf_hub_download(repo_id=MODEL_REPO, filename=MODEL_FILE)
|
11 |
llm = Llama(
|