Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -48,7 +48,7 @@ def retrieve_documents(query, k=5):
|
|
48 |
distances, indices = index.search(query_embedding, k)
|
49 |
return [documents[i] for i in indices[0]]
|
50 |
|
51 |
-
def respond(message, history, system_message, max_tokens, temperature, top_p):
|
52 |
# Retrieve relevant documents
|
53 |
relevant_docs = retrieve_documents(message)
|
54 |
context = "\n\n".join(relevant_docs[:3]) # Limit context to top 3 documents
|
@@ -67,9 +67,7 @@ def respond(message, history, system_message, max_tokens, temperature, top_p):
|
|
67 |
|
68 |
messages.append({"role": "user", "content": message})
|
69 |
|
70 |
-
|
71 |
-
|
72 |
-
for message in client.chat_completion(
|
73 |
messages,
|
74 |
max_tokens=max_tokens,
|
75 |
stream=True,
|
@@ -78,8 +76,7 @@ def respond(message, history, system_message, max_tokens, temperature, top_p):
|
|
78 |
):
|
79 |
if message.choices and message.choices[0].delta and message.choices[0].delta.content:
|
80 |
token = message.choices[0].delta.content
|
81 |
-
|
82 |
-
yield response
|
83 |
|
84 |
demo = gr.ChatInterface(
|
85 |
respond,
|
|
|
48 |
distances, indices = index.search(query_embedding, k)
|
49 |
return [documents[i] for i in indices[0]]
|
50 |
|
51 |
+
async def respond(message, history, system_message, max_tokens, temperature, top_p):
|
52 |
# Retrieve relevant documents
|
53 |
relevant_docs = retrieve_documents(message)
|
54 |
context = "\n\n".join(relevant_docs[:3]) # Limit context to top 3 documents
|
|
|
67 |
|
68 |
messages.append({"role": "user", "content": message})
|
69 |
|
70 |
+
async for message in client.chat_completion(
|
|
|
|
|
71 |
messages,
|
72 |
max_tokens=max_tokens,
|
73 |
stream=True,
|
|
|
76 |
):
|
77 |
if message.choices and message.choices[0].delta and message.choices[0].delta.content:
|
78 |
token = message.choices[0].delta.content
|
79 |
+
yield token
|
|
|
80 |
|
81 |
demo = gr.ChatInterface(
|
82 |
respond,
|