Update app.py
Browse files
app.py
CHANGED
@@ -23,7 +23,7 @@ def load_model():
|
|
23 |
|
24 |
return Llama(
|
25 |
model_path=model_path,
|
26 |
-
n_ctx=1024,
|
27 |
n_threads=4,
|
28 |
verbose=False
|
29 |
)
|
@@ -31,9 +31,9 @@ def load_model():
|
|
31 |
# Load model at startup
|
32 |
llm = load_model()
|
33 |
|
34 |
-
# Generation function with
|
35 |
def generate_text(prompt, max_tokens=MAX_TOKENS, temp=0.7, top_p=0.95):
|
36 |
-
"""Generate text with repetition
|
37 |
if not prompt.strip():
|
38 |
return "⚠️ Please enter a valid prompt."
|
39 |
|
@@ -47,8 +47,8 @@ def generate_text(prompt, max_tokens=MAX_TOKENS, temp=0.7, top_p=0.95):
|
|
47 |
temperature=temp,
|
48 |
top_p=top_p,
|
49 |
echo=False,
|
50 |
-
|
51 |
-
|
52 |
)
|
53 |
return output["choices"][0]["text"].strip()
|
54 |
except Exception as e:
|
|
|
23 |
|
24 |
return Llama(
|
25 |
model_path=model_path,
|
26 |
+
n_ctx=1024,
|
27 |
n_threads=4,
|
28 |
verbose=False
|
29 |
)
|
|
|
31 |
# Load model at startup
|
32 |
llm = load_model()
|
33 |
|
34 |
+
# Generation function with anti-repetition
|
35 |
def generate_text(prompt, max_tokens=MAX_TOKENS, temp=0.7, top_p=0.95):
|
36 |
+
"""Generate text with improved repetition handling"""
|
37 |
if not prompt.strip():
|
38 |
return "⚠️ Please enter a valid prompt."
|
39 |
|
|
|
47 |
temperature=temp,
|
48 |
top_p=top_p,
|
49 |
echo=False,
|
50 |
+
# Only use supported parameters
|
51 |
+
repeat_penalty=1.5 # Increased from 1.2
|
52 |
)
|
53 |
return output["choices"][0]["text"].strip()
|
54 |
except Exception as e:
|