astonn commited on
Commit
1c6417a
·
verified ·
1 Parent(s): 706cbfb

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -5,7 +5,7 @@ from llama_cpp import Llama # GGUF CPU backend
5
  # ---------- model loading (one‑time) ----------
6
  MODEL_REPO = "MaziyarPanahi/gemma-2b-it-GGUF"
7
  MODEL_FILE = "gemma-2b-it.Q4_K_M.gguf" # 1.6 GB 4‑bit
8
- CTX_SIZE = 2048
9
 
10
  model_path = hf_hub_download(repo_id=MODEL_REPO, filename=MODEL_FILE)
11
  llm = Llama(
@@ -63,8 +63,8 @@ def analyze_ads(file):
63
  # stream=True → токены сразу в логах, ответ ~25‑30 с
64
  stream = llm(
65
  prompt,
66
- max_tokens=250,
67
- temperature=0.6,
68
  top_p=0.8,
69
  stream=True
70
  )
 
5
  # ---------- model loading (one‑time) ----------
6
  MODEL_REPO = "MaziyarPanahi/gemma-2b-it-GGUF"
7
  MODEL_FILE = "gemma-2b-it.Q4_K_M.gguf" # 1.6 GB 4‑bit
8
+ CTX_SIZE = 4096
9
 
10
  model_path = hf_hub_download(repo_id=MODEL_REPO, filename=MODEL_FILE)
11
  llm = Llama(
 
63
  # stream=True → токены сразу в логах, ответ ~25‑30 с
64
  stream = llm(
65
  prompt,
66
+ max_tokens=800,
67
+ temperature=0.4,
68
  top_p=0.8,
69
  stream=True
70
  )