Spaces:
Sleeping
Sleeping
gizemsarsinlar
commited on
Commit
•
6f7c8df
1
Parent(s):
f7fafa9
Update README.md
Browse files
README.md
CHANGED
@@ -10,4 +10,35 @@ pinned: false
|
|
10 |
license: apache-2.0
|
11 |
---
|
12 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
An example chatbot using [Gradio](https://gradio.app), [`huggingface_hub`](https://huggingface.co/docs/huggingface_hub/v0.22.2/en/index), and the [Hugging Face Inference API](https://huggingface.co/docs/api-inference/index).
|
|
|
10 |
license: apache-2.0
|
11 |
---
|
12 |
|
13 |
+
# NVIDIA AI Document Chatbot
|
14 |
+
|
15 |
+
This project is a document-based chatbot application. It helps users ask questions about specific documents and receive accurate responses based on those documents.
|
16 |
+
|
17 |
+
## Models and Components Used:
|
18 |
+
|
19 |
+
1. **NVIDIAEmbeddings (NV-Embed-QA)**:
|
20 |
+
- This model extracts vector representations of texts to better understand documents.
|
21 |
+
- The **NV-Embed-QA** model is used to find relevant information in documents to answer questions.
|
22 |
+
|
23 |
+
2. **ChatDocument (mistralai/mixtral-8x7b-instruct-v0.1)**:
|
24 |
+
- The **Mistral-8x7B Instruct** model is responsible for answering user questions about documents. It specializes in extracting information from documents and responding conversationally.
|
25 |
+
|
26 |
+
## Application Workflow:
|
27 |
+
|
28 |
+
1. **Loading Documents**:
|
29 |
+
- Specific academic papers are loaded using `ArxivLoader`. These documents are split into text chunks and filtered based on predefined rules.
|
30 |
+
- The documents are then added to a FAISS vector store, which allows for efficient and fast document chunk retrieval.
|
31 |
+
|
32 |
+
2. **Chat and Document Querying**:
|
33 |
+
- The user's questions are processed according to a predefined chat template. The response is generated based on both the conversation history and information retrieved from the documents.
|
34 |
+
- The `chat_gen` function takes the user's input and generates responses using NVIDIA models, pulling relevant information from the documents.
|
35 |
+
|
36 |
+
3. **Remembering Document Content and Conversation History**:
|
37 |
+
- Previous user messages and responses are stored in a conversation memory and used for answering future questions more effectively.
|
38 |
+
|
39 |
+
### Conclusion:
|
40 |
+
|
41 |
+
This application leverages NVIDIA's powerful language models and embedding tools to generate intelligent, document-driven conversational responses.
|
42 |
+
|
43 |
+
|
44 |
An example chatbot using [Gradio](https://gradio.app), [`huggingface_hub`](https://huggingface.co/docs/huggingface_hub/v0.22.2/en/index), and the [Hugging Face Inference API](https://huggingface.co/docs/api-inference/index).
|