Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -17,7 +17,7 @@ model_name = "unsloth.Q4_K_M.gguf"
|
|
17 |
model_path = os.path.join(model_dir, model_name)
|
18 |
|
19 |
hf_hub_download(
|
20 |
-
repo_id="vislupus/bulgarian-joke-master-
|
21 |
filename=model_name,
|
22 |
local_dir=model_dir
|
23 |
)
|
@@ -27,7 +27,7 @@ if not os.path.exists(model_path):
|
|
27 |
|
28 |
llm = Llama(model_path=model_path)
|
29 |
|
30 |
-
def generate_response(message, history, temperature=0.7, top_p=1.0, max_tokens=
|
31 |
try:
|
32 |
response = llm(message, max_tokens=max_tokens, temperature=temperature, top_p=top_p)
|
33 |
return response["choices"][0]["text"].strip()
|
|
|
17 |
model_path = os.path.join(model_dir, model_name)
|
18 |
|
19 |
hf_hub_download(
|
20 |
+
repo_id="vislupus/bulgarian-joke-master-gemma-2-2b-it-bnb-4bit-gguf",
|
21 |
filename=model_name,
|
22 |
local_dir=model_dir
|
23 |
)
|
|
|
27 |
|
28 |
llm = Llama(model_path=model_path)
|
29 |
|
30 |
+
def generate_response(message, history, temperature=0.7, top_p=1.0, max_tokens=1280):
|
31 |
try:
|
32 |
response = llm(message, max_tokens=max_tokens, temperature=temperature, top_p=top_p)
|
33 |
return response["choices"][0]["text"].strip()
|