Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -5,7 +5,7 @@ from llama_cpp import Llama # GGUF CPU backend
|
|
5 |
# ---------- model loading (one‑time) ----------
|
6 |
MODEL_REPO = "MaziyarPanahi/gemma-2b-it-GGUF"
|
7 |
MODEL_FILE = "gemma-2b-it.Q4_K_M.gguf" # 1.6 GB 4‑bit
|
8 |
-
CTX_SIZE =
|
9 |
|
10 |
model_path = hf_hub_download(repo_id=MODEL_REPO, filename=MODEL_FILE)
|
11 |
llm = Llama(
|
@@ -63,8 +63,8 @@ def analyze_ads(file):
|
|
63 |
# stream=True → токены сразу в логах, ответ ~25‑30 с
|
64 |
stream = llm(
|
65 |
prompt,
|
66 |
-
max_tokens=
|
67 |
-
temperature=0.
|
68 |
top_p=0.8,
|
69 |
stream=True
|
70 |
)
|
|
|
5 |
# ---------- model loading (one‑time) ----------
|
6 |
MODEL_REPO = "MaziyarPanahi/gemma-2b-it-GGUF"
|
7 |
MODEL_FILE = "gemma-2b-it.Q4_K_M.gguf" # 1.6 GB 4‑bit
|
8 |
+
CTX_SIZE = 4096
|
9 |
|
10 |
model_path = hf_hub_download(repo_id=MODEL_REPO, filename=MODEL_FILE)
|
11 |
llm = Llama(
|
|
|
63 |
# stream=True → токены сразу в логах, ответ ~25‑30 с
|
64 |
stream = llm(
|
65 |
prompt,
|
66 |
+
max_tokens=800,
|
67 |
+
temperature=0.4,
|
68 |
top_p=0.8,
|
69 |
stream=True
|
70 |
)
|