olenkap commited on
Commit
a1fe12d
·
verified ·
1 Parent(s): df1ba52

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -42
README.md CHANGED
@@ -1,42 +1,48 @@
1
- # RAG. Question answering bot.
2
- ![](https://i.giphy.com/media/v1.Y2lkPTc5MGI3NjExNXZyaTM1anczcGE0cDliYWZkNXhvY3ZrOGRzeTJ5a3EwcXl3aGVnZCZlcD12MV9pbnRlcm5hbF9naWZfYnlfaWQmY3Q9Zw/12xDxBbj7CPAOI/giphy.gif)
3
-
4
- ### Topics
5
- - [Data source](#data-source) ✔️
6
- - [Chunking](#chunking) ✔️
7
- - [LLM](#llm) ✔️
8
- - [Retriever](#retriever) ✔️
9
- - [Reranker](#reranker) ✔️
10
- - [Citation](#citation) ❌
11
- - [Web UI and deployment](#web-ui-and-deployment) ✔️
12
-
13
-
14
- ## Data source
15
-
16
- I used documents found on the Internet. You can take a look at them in **docs** directory, and you can ask questions based on that context. There also is possibility to upload your txt file and use it as a context.
17
-
18
- ## Chunking
19
- Chunking was performed using the same method explained in live-coding session. No other libraries were involved.
20
-
21
- ## LLM
22
- As LLM I used pretrained model [llama3-70b-8192](https://huggingface.co/Groq/Llama-3-Groq-70B-Tool-Use).
23
-
24
- ## Retriever
25
- Retrieving can be performed in three different ways. You can either use BM25 retriever or a dense retriever by calculating semantic scores. Using both of them in hybrid approach is also an option.
26
-
27
- Dense retriever used in this lab - sentence-transformers/all-distilroberta-v1.
28
-
29
- #### Here's an example when dense retriever works better than BM25:
30
- ![image_2024-12-08_21-59-16](https://github.com/user-attachments/assets/50e2b14a-dc28-4e4b-8752-90ab46d2d883)
31
- ![image_2024-12-08_22-00-02](https://github.com/user-attachments/assets/eec3f618-99f0-4b43-b6a5-573ef612ae3e)
32
-
33
-
34
- ## Reranker
35
-
36
- As a reranker there was used cross encoder cross-encoder/stsb-roberta-base. It may be not efficient in my case, as far as amount of documents is quite small, so it takes time to process the data, but does not improve the process of extracting context.
37
- ## Citation
38
-
39
- Isn't implemented
40
- ## Web UI and deployment
41
- I used gradio lib for demo and hosting.
42
-
 
 
 
 
 
 
 
1
+ ---
2
+ title: RAG
3
+ app_file: app.py
4
+ sdk: gradio
5
+ sdk_version: 5.8.0
6
+ ---
7
+ # RAG. Question answering bot.
8
+ ![](https://i.giphy.com/media/v1.Y2lkPTc5MGI3NjExNXZyaTM1anczcGE0cDliYWZkNXhvY3ZrOGRzeTJ5a3EwcXl3aGVnZCZlcD12MV9pbnRlcm5hbF9naWZfYnlfaWQmY3Q9Zw/12xDxBbj7CPAOI/giphy.gif)
9
+
10
+ ### Topics
11
+ - [Data source](#data-source) ✔️
12
+ - [Chunking](#chunking) ✔️
13
+ - [LLM](#llm) ✔️
14
+ - [Retriever](#retriever) ✔️
15
+ - [Reranker](#reranker) ✔️
16
+ - [Citation](#citation)
17
+ - [Web UI and deployment](#web-ui-and-deployment) ✔️
18
+
19
+
20
+ ## Data source
21
+
22
+ I used documents found on the Internet. You can take a look at them in **docs** directory, and you can ask questions based on that context. There also is possibility to upload your txt file and use it as a context.
23
+
24
+ ## Chunking
25
+ Chunking was performed using the same method explained in live-coding session. No other libraries were involved.
26
+
27
+ ## LLM
28
+ As LLM I used pretrained model [llama3-70b-8192](https://huggingface.co/Groq/Llama-3-Groq-70B-Tool-Use).
29
+
30
+ ## Retriever
31
+ Retrieving can be performed in three different ways. You can either use BM25 retriever or a dense retriever by calculating semantic scores. Using both of them in hybrid approach is also an option.
32
+
33
+ Dense retriever used in this lab - sentence-transformers/all-distilroberta-v1.
34
+
35
+ #### Here's an example when dense retriever works better than BM25:
36
+ ![image_2024-12-08_21-59-16](https://github.com/user-attachments/assets/50e2b14a-dc28-4e4b-8752-90ab46d2d883)
37
+ ![image_2024-12-08_22-00-02](https://github.com/user-attachments/assets/eec3f618-99f0-4b43-b6a5-573ef612ae3e)
38
+
39
+
40
+ ## Reranker
41
+
42
+ As a reranker there was used cross encoder cross-encoder/stsb-roberta-base. It may be not efficient in my case, as far as amount of documents is quite small, so it takes time to process the data, but does not improve the process of extracting context.
43
+ ## Citation
44
+
45
+ Isn't implemented
46
+ ## Web UI and deployment
47
+ I used gradio lib for demo and hosting.
48
+