Spaces:
Building
Building
Update chat_handler.py
Browse files- chat_handler.py +7 -7
chat_handler.py
CHANGED
@@ -89,7 +89,7 @@ async def llm_generate(s: Session, prompt: str, user_msg: str) -> str:
|
|
89 |
log_info(f"🚀 Calling LLM for session {s.session_id[:8]}...")
|
90 |
log_info(f"📋 Prompt preview (first 200 chars): {prompt[:200]}...")
|
91 |
|
92 |
-
history = s.
|
93 |
|
94 |
# Call the configured LLM provider
|
95 |
raw = await llm_provider.generate(
|
@@ -347,7 +347,7 @@ async def chat(req: ChatRequest, x_session_id: str = Header(...)):
|
|
347 |
# Process based on current state - MEVCUT KODU KORUYORUZ
|
348 |
if session.state == "idle":
|
349 |
# Build intent detection prompt
|
350 |
-
prompt = build_intent_prompt(version, session.
|
351 |
raw = await llm_generate(session, prompt, req.message)
|
352 |
|
353 |
# Check for intent
|
@@ -383,7 +383,7 @@ async def chat(req: ChatRequest, x_session_id: str = Header(...)):
|
|
383 |
param_prompt = build_parameter_prompt(
|
384 |
intent_config,
|
385 |
session.variables,
|
386 |
-
session.
|
387 |
project.default_locale
|
388 |
)
|
389 |
param_question = await llm_generate(session, param_prompt, req.message)
|
@@ -436,7 +436,7 @@ async def chat(req: ChatRequest, x_session_id: str = Header(...)):
|
|
436 |
param_prompt = build_parameter_prompt(
|
437 |
intent_config,
|
438 |
session.variables,
|
439 |
-
session.
|
440 |
project.default_locale
|
441 |
)
|
442 |
param_question = await llm_generate(session, param_prompt, req.message)
|
@@ -485,7 +485,7 @@ async def handle_new_message(session: Session, user_input: str) -> str:
|
|
485 |
return "Proje konfigürasyonu bulunamadı."
|
486 |
|
487 |
# Build intent detection prompt
|
488 |
-
prompt = build_intent_prompt(version, session.
|
489 |
|
490 |
# Get LLM response
|
491 |
raw = await llm_generate(session, prompt, user_input)
|
@@ -526,7 +526,7 @@ async def handle_new_message(session: Session, user_input: str) -> str:
|
|
526 |
param_prompt = build_parameter_prompt(
|
527 |
intent_config,
|
528 |
session.variables,
|
529 |
-
session.
|
530 |
project.default_locale
|
531 |
)
|
532 |
param_question = await llm_generate(session, param_prompt, user_input)
|
@@ -586,7 +586,7 @@ async def handle_parameter_followup(session: Session, user_input: str) -> str:
|
|
586 |
param_prompt = build_parameter_prompt(
|
587 |
intent_config,
|
588 |
session.variables,
|
589 |
-
session.
|
590 |
project.default_locale
|
591 |
)
|
592 |
param_question = await llm_generate(session, param_prompt, user_input)
|
|
|
89 |
log_info(f"🚀 Calling LLM for session {s.session_id[:8]}...")
|
90 |
log_info(f"📋 Prompt preview (first 200 chars): {prompt[:200]}...")
|
91 |
|
92 |
+
history = s.chat_history
|
93 |
|
94 |
# Call the configured LLM provider
|
95 |
raw = await llm_provider.generate(
|
|
|
347 |
# Process based on current state - MEVCUT KODU KORUYORUZ
|
348 |
if session.state == "idle":
|
349 |
# Build intent detection prompt
|
350 |
+
prompt = build_intent_prompt(version, session.chat_history, project.default_locale)
|
351 |
raw = await llm_generate(session, prompt, req.message)
|
352 |
|
353 |
# Check for intent
|
|
|
383 |
param_prompt = build_parameter_prompt(
|
384 |
intent_config,
|
385 |
session.variables,
|
386 |
+
session.chat_history,
|
387 |
project.default_locale
|
388 |
)
|
389 |
param_question = await llm_generate(session, param_prompt, req.message)
|
|
|
436 |
param_prompt = build_parameter_prompt(
|
437 |
intent_config,
|
438 |
session.variables,
|
439 |
+
session.chat_history,
|
440 |
project.default_locale
|
441 |
)
|
442 |
param_question = await llm_generate(session, param_prompt, req.message)
|
|
|
485 |
return "Proje konfigürasyonu bulunamadı."
|
486 |
|
487 |
# Build intent detection prompt
|
488 |
+
prompt = build_intent_prompt(version, session.chat_history, project.default_locale)
|
489 |
|
490 |
# Get LLM response
|
491 |
raw = await llm_generate(session, prompt, user_input)
|
|
|
526 |
param_prompt = build_parameter_prompt(
|
527 |
intent_config,
|
528 |
session.variables,
|
529 |
+
session.chat_history,
|
530 |
project.default_locale
|
531 |
)
|
532 |
param_question = await llm_generate(session, param_prompt, user_input)
|
|
|
586 |
param_prompt = build_parameter_prompt(
|
587 |
intent_config,
|
588 |
session.variables,
|
589 |
+
session.chat_history,
|
590 |
project.default_locale
|
591 |
)
|
592 |
param_question = await llm_generate(session, param_prompt, user_input)
|