Spaces:
Sleeping
Sleeping
πwπ
Browse files
app.py
CHANGED
|
@@ -22,6 +22,7 @@ device = torch.device("cuda")
|
|
| 22 |
model = model.to(device)
|
| 23 |
RAG = SentenceTransformer("mixedbread-ai/mxbai-embed-large-v1")
|
| 24 |
TOP_K = 3
|
|
|
|
| 25 |
# prepare data
|
| 26 |
# since data is too big we will only select the first 3K lines
|
| 27 |
|
|
@@ -57,15 +58,14 @@ def prepare_prompt(query, retrieved_examples):
|
|
| 57 |
def talk(message, history):
|
| 58 |
retrieved_examples = search(message)
|
| 59 |
message, metadata = prepare_prompt(message, retrieved_examples)
|
| 60 |
-
resources =
|
| 61 |
for title, url in metadata:
|
| 62 |
resources += f"[{title}]({url}), "
|
| 63 |
chat = []
|
| 64 |
for item in history:
|
| 65 |
chat.append({"role": "user", "content": item[0]})
|
| 66 |
-
|
| 67 |
-
|
| 68 |
-
chat.append({"role": "assistant", "content": cleaned_past})
|
| 69 |
chat.append({"role": "user", "content": message})
|
| 70 |
messages = tokenizer.apply_chat_template(
|
| 71 |
chat, tokenize=False, add_generation_prompt=True
|
|
|
|
| 22 |
model = model.to(device)
|
| 23 |
RAG = SentenceTransformer("mixedbread-ai/mxbai-embed-large-v1")
|
| 24 |
TOP_K = 3
|
| 25 |
+
HEADER = "\n# RESOURCES:\n"
|
| 26 |
# prepare data
|
| 27 |
# since data is too big we will only select the first 3K lines
|
| 28 |
|
|
|
|
| 58 |
def talk(message, history):
|
| 59 |
retrieved_examples = search(message)
|
| 60 |
message, metadata = prepare_prompt(message, retrieved_examples)
|
| 61 |
+
resources = HEADER
|
| 62 |
for title, url in metadata:
|
| 63 |
resources += f"[{title}]({url}), "
|
| 64 |
chat = []
|
| 65 |
for item in history:
|
| 66 |
chat.append({"role": "user", "content": item[0]})
|
| 67 |
+
cleaned_past = item[1].split(HEADER)[0]
|
| 68 |
+
chat.append({"role": "assistant", "content": cleaned_past})
|
|
|
|
| 69 |
chat.append({"role": "user", "content": message})
|
| 70 |
messages = tokenizer.apply_chat_template(
|
| 71 |
chat, tokenize=False, add_generation_prompt=True
|