Spaces:
Running
Running
Update server.js
Browse files
server.js
CHANGED
@@ -240,13 +240,6 @@ app.post("/api/ask-ai", async (req, res) => {
|
|
240 |
? DEFAULT_PROVIDER
|
241 |
: PROVIDERS[provider] ?? DEFAULT_PROVIDER;
|
242 |
|
243 |
-
if (provider !== "auto" && TOKENS_USED >= selectedProvider.max_tokens) {
|
244 |
-
return res.status(400).send({
|
245 |
-
ok: false,
|
246 |
-
openSelectProvider: true,
|
247 |
-
message: `Context is too long. ${selectedProvider.name} allow ${selectedProvider.max_tokens} max tokens.`,
|
248 |
-
});
|
249 |
-
}
|
250 |
|
251 |
try {
|
252 |
let completeResponse = "";
|
@@ -289,7 +282,7 @@ app.post("/api/ask-ai", async (req, res) => {
|
|
289 |
model: MODEL_ID, // You can adjust this model as needed
|
290 |
messages: messages,
|
291 |
stream: true,
|
292 |
-
max_tokens:
|
293 |
})
|
294 |
});
|
295 |
|
|
|
240 |
? DEFAULT_PROVIDER
|
241 |
: PROVIDERS[provider] ?? DEFAULT_PROVIDER;
|
242 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
243 |
|
244 |
try {
|
245 |
let completeResponse = "";
|
|
|
282 |
model: MODEL_ID, // You can adjust this model as needed
|
283 |
messages: messages,
|
284 |
stream: true,
|
285 |
+
max_tokens: 1048576
|
286 |
})
|
287 |
});
|
288 |
|