Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -29,7 +29,6 @@ Conversation Context:
|
|
29 |
Chat History: {chat_history}
|
30 |
Provided Documents: {provided_docs}
|
31 |
Extracted Text from Links: {extracted_text}
|
32 |
-
User Input: {user_input}
|
33 |
|
34 |
User: {user_question}
|
35 |
Chatto:
|
@@ -37,7 +36,7 @@ Chatto:
|
|
37 |
|
38 |
|
39 |
prompt = PromptTemplate(
|
40 |
-
input_variables=["chat_history", "provided_docs", "extracted_text", "
|
41 |
template=template
|
42 |
)
|
43 |
|
@@ -62,7 +61,7 @@ def conversational_chat(query):
|
|
62 |
|
63 |
result = llm_chain.predict(
|
64 |
chat_history=previous_response,
|
65 |
-
|
66 |
provided_docs=provided_docs,
|
67 |
extracted_text=extracted_text
|
68 |
)
|
|
|
29 |
Chat History: {chat_history}
|
30 |
Provided Documents: {provided_docs}
|
31 |
Extracted Text from Links: {extracted_text}
|
|
|
32 |
|
33 |
User: {user_question}
|
34 |
Chatto:
|
|
|
36 |
|
37 |
|
38 |
prompt = PromptTemplate(
|
39 |
+
input_variables=["chat_history", "provided_docs", "extracted_text", "user_question"],
|
40 |
template=template
|
41 |
)
|
42 |
|
|
|
61 |
|
62 |
result = llm_chain.predict(
|
63 |
chat_history=previous_response,
|
64 |
+
user_question=query,
|
65 |
provided_docs=provided_docs,
|
66 |
extracted_text=extracted_text
|
67 |
)
|