Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -5,7 +5,7 @@ from huggingface_hub import InferenceClient
|
|
5 |
|
6 |
# --- Configuration Constants ---
|
7 |
# Define the maximum number of tokens the model should generate in a single response
|
8 |
-
FIXED_MAX_TOKENS =
|
9 |
|
10 |
|
11 |
# --- Initialize the InferenceClient ---
|
|
|
5 |
|
6 |
# --- Configuration Constants ---
|
7 |
# Define the maximum number of tokens the model should generate in a single response
|
8 |
+
FIXED_MAX_TOKENS = 1000 # Note: This is a very high value, typical values are much lower (e.g., 512, 1024, 2048, 4096 for many models)
|
9 |
|
10 |
|
11 |
# --- Initialize the InferenceClient ---
|