Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -64,9 +64,14 @@ def setup_multiple_chains():
|
|
64 |
)
|
65 |
|
66 |
conversation_memory = ConversationBufferMemory(
|
67 |
-
memory_key="chat_history",
|
|
|
|
|
|
|
|
|
68 |
)
|
69 |
|
|
|
70 |
llm_chain = LLMChain(
|
71 |
llm=llm,
|
72 |
prompt=daysoff_assistant_prompt,
|
@@ -118,9 +123,19 @@ async def handle_message(message: cl.Message):
|
|
118 |
except requests.exceptions.RequestException as e:
|
119 |
await cl.Message(content=f"Request failed: {str(e)}").send()
|
120 |
else:
|
121 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
122 |
|
123 |
|
124 |
-
response_key = "output" if "output" in response else "text"
|
125 |
-
await cl.Message(response.get(response_key, "")).send()
|
126 |
-
return message.content
|
|
|
64 |
)
|
65 |
|
66 |
conversation_memory = ConversationBufferMemory(
|
67 |
+
memory_key="chat_history",
|
68 |
+
input_key="question", # ?
|
69 |
+
output_key="text", # ?
|
70 |
+
max_len=30,
|
71 |
+
return_messages=True
|
72 |
)
|
73 |
|
74 |
+
|
75 |
llm_chain = LLMChain(
|
76 |
llm=llm,
|
77 |
prompt=daysoff_assistant_prompt,
|
|
|
123 |
except requests.exceptions.RequestException as e:
|
124 |
await cl.Message(content=f"Request failed: {str(e)}").send()
|
125 |
else:
|
126 |
+
try:
|
127 |
+
response = await llm_chain.ainvoke({
|
128 |
+
"question": user_message,
|
129 |
+
"chat_history": ""
|
130 |
+
}, callbacks=[cl.AsyncLangchainCallbackHandler()])
|
131 |
+
|
132 |
+
await cl.Message(content=response["text"]).send()
|
133 |
+
|
134 |
+
except Exception as e:
|
135 |
+
await cl.Message(content=f"Error: {str(e)}").send()
|
136 |
+
|
137 |
|
138 |
|
139 |
+
#response_key = "output" if "output" in response else "text"
|
140 |
+
#await cl.Message(response.get(response_key, "")).send()
|
141 |
+
#return message.content
|