Update app.py
Browse files
app.py
CHANGED
@@ -1,45 +1,24 @@
|
|
1 |
import gradio as gr
|
2 |
from my_memory_logic import run_with_session_memory
|
3 |
|
4 |
-
def chat_interface_fn(message, history
|
5 |
"""
|
6 |
Multi-turn chat function for Gradio's ChatInterface.
|
7 |
-
|
8 |
-
Deduplicates consecutive repeated Q&A pairs to avoid repetition.
|
9 |
"""
|
10 |
# Initialize history if None
|
11 |
-
|
12 |
-
history = []
|
13 |
-
|
14 |
-
# Ensure we're working with the correct history format
|
15 |
-
# Gradio 3.x sends history as a list of tuples (user, assistant)
|
16 |
-
if isinstance(history, list):
|
17 |
-
if len(history) > 0 and isinstance(history[0], tuple):
|
18 |
-
history = [
|
19 |
-
{"role": "user" if i % 2 == 0 else "assistant", "content": msg}
|
20 |
-
for tup in history
|
21 |
-
for i, msg in enumerate([tup[0], tup[1]])
|
22 |
-
]
|
23 |
|
24 |
# Get answer from the session-based memory pipeline
|
25 |
try:
|
26 |
-
answer = run_with_session_memory(message,
|
27 |
except Exception as e:
|
28 |
print(f"Error in run_with_session_memory: {str(e)}")
|
29 |
answer = "I apologize, but I encountered an error processing your request."
|
30 |
|
31 |
-
#
|
32 |
-
|
33 |
-
|
34 |
-
{"role": "user", "content": message},
|
35 |
-
{"role": "assistant", "content": answer}
|
36 |
-
]
|
37 |
-
|
38 |
-
# Convert history to format expected by Gradio
|
39 |
-
chat_history = [(msg["content"], hist["content"])
|
40 |
-
for msg, hist in zip(new_history[::2], new_history[1::2])]
|
41 |
-
|
42 |
-
return chat_history, new_history
|
43 |
|
44 |
# Custom CSS for chat interface
|
45 |
my_chat_css = """
|
@@ -60,7 +39,7 @@ with gr.Blocks(css=my_chat_css) as demo:
|
|
60 |
session_id_box = gr.Textbox(label="Session ID", value="abc123", interactive=True)
|
61 |
|
62 |
chat_interface = gr.ChatInterface(
|
63 |
-
fn=
|
64 |
title="DailyWellnessAI (Session-based Memory)",
|
65 |
description="Ask your questions. The session_id determines your stored memory."
|
66 |
)
|
|
|
1 |
import gradio as gr
|
2 |
from my_memory_logic import run_with_session_memory
|
3 |
|
4 |
+
def chat_interface_fn(message, history):
|
5 |
"""
|
6 |
Multi-turn chat function for Gradio's ChatInterface.
|
7 |
+
Returns response and history in the format expected by Gradio.
|
|
|
8 |
"""
|
9 |
# Initialize history if None
|
10 |
+
history = history or []
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
|
12 |
# Get answer from the session-based memory pipeline
|
13 |
try:
|
14 |
+
answer = run_with_session_memory(message, session_id_box.value)
|
15 |
except Exception as e:
|
16 |
print(f"Error in run_with_session_memory: {str(e)}")
|
17 |
answer = "I apologize, but I encountered an error processing your request."
|
18 |
|
19 |
+
# Return the answer and append to history
|
20 |
+
history.append((message, answer))
|
21 |
+
return history
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
22 |
|
23 |
# Custom CSS for chat interface
|
24 |
my_chat_css = """
|
|
|
39 |
session_id_box = gr.Textbox(label="Session ID", value="abc123", interactive=True)
|
40 |
|
41 |
chat_interface = gr.ChatInterface(
|
42 |
+
fn=chat_interface_fn,
|
43 |
title="DailyWellnessAI (Session-based Memory)",
|
44 |
description="Ask your questions. The session_id determines your stored memory."
|
45 |
)
|