Spaces:
No application file
No application file
Update README.md
Browse files
README.md
CHANGED
@@ -8,99 +8,12 @@ pinned: false
|
|
8 |
license: apache-2.0
|
9 |
---
|
10 |
# fun dialogues
|
11 |
-
A library of
|
12 |
-
|
13 |
-
repository link: https://github.com/eduand-alvarez/fun-dialogues
|
14 |
-
|
15 |
-
### Key components of fun dialogues:
|
16 |
-
- **JudgyRAG**: RAG eval tool
|
17 |
-
- **Benchmark Datasets**: custom benchmark datasets
|
18 |
-
- **Dialogues**: fictitious dialogue datasets for fun, experimentation, and testing.
|
19 |
-
|
20 |
-
You can install fun dialogues using pip: `pip install fundialogues`
|
21 |
-
|
22 |
-
<div align="center">
|
23 |
-
<img src="https://github.com/eduand-alvarez/fun-dialogues/assets/57263404/1d8ce401-b595-442f-980c-8ae06ed9d4b2" alt="Fun Dialogues Logo" width="750"/>
|
24 |
-
</div>
|
25 |
-
|
26 |
-
# JudgyRAG
|
27 |
-
|
28 |
-
JudgyRAG is a component of the FunDialogues Python library focused on evaluating the performance of Retrieval-Augmented Generation (RAG) systems. It facilitates this by creating synthetic datasets based on custom datasets, enabling a unique assessment of a RAG system's question-answering capabilities in a zero-shot Q&A context. Initially, JudgyRAG's primary functionality is the automatic generation of custom multiple-choice Q&A datasets. Future iterations will introduce further automation to seamlessly integrate with popular frameworks, enhancing testing and benchmarking processes.
|
29 |
-
|
30 |
-
### Workflow
|
31 |
-
|
32 |
-
The workflow for JudgyRAG includes:
|
33 |
-
|
34 |
-
1. **Scraping PDFs**: Information is extracted from PDFs into structured text formats.
|
35 |
-
2. **Chunking Data**: Extracted data is chunked similarly to vector database embeddings for RAG, simulating data breakdown and storage.
|
36 |
-
3. **Question Generation**: Each chunk acts as a knowledge base, with custom prompts instructing supported models (currently LLaMA 7B and 13B chat) to generate multiple-choice questions.
|
37 |
-
4. **Iterative Parsing**: Chunks are processed iteratively, generating a multiple-choice question for each.
|
38 |
-
5. **Quality Checks**: Poor-quality chunks leading to failed question generation are flagged for user review.
|
39 |
-
6. **Benchmark Compilation**: The final document includes multiple-choice questions, correct answers, and source knowledge chunks.
|
40 |
-
7. **RAG System Evaluation**: The synthetic benchmark dataset can be used to assess a RAG system, with automation for this process planned for future updates.
|
41 |
-
|
42 |
-
### Environment Setup
|
43 |
-
|
44 |
-
Follow these steps to set up your environment for JudgyRAG:
|
45 |
-
|
46 |
-
#### Step 1
|
47 |
-
Install Visual Studio 2022 Community Edition with the “Desktop development with C++” workload.
|
48 |
-
|
49 |
-
#### Step 2
|
50 |
-
Update to the latest GPU driver.
|
51 |
-
|
52 |
-
#### Step 3
|
53 |
-
Install the Intel® oneAPI Base Toolkit 2024.0.
|
54 |
-
|
55 |
-
#### Step 4
|
56 |
-
Download the necessary wheels:
|
57 |
-
|
58 |
-
```bash
|
59 |
-
wget https://intel-extension-for-pytorch.s3.amazonaws.com/ipex_stable/xpu/torch-2.1.0a0%2Bcxx11.abi-cp39-cp39-win_amd64.whl
|
60 |
-
wget https://intel-extension-for-pytorch.s3.amazonaws.com/ipex_stable/xpu/torchvision-0.16.0a0%2Bcxx11.abi-cp39-cp39-win_amd64.whl
|
61 |
-
wget https://intel-extension-for-pytorch.s3.amazonaws.com/ipex_stable/xpu/intel_extension_for_pytorch-2.1.10%2Bxpu-cp39-cp39-win_amd64.whl
|
62 |
-
```
|
63 |
-
|
64 |
-
#### Step 5
|
65 |
-
|
66 |
-
Install the downloaded packages and BigDL LLM:
|
67 |
-
|
68 |
-
```bash
|
69 |
-
pip install torch-2.1.0a0+cxx11.abi-cp39-cp39-win_amd64.whl
|
70 |
-
pip install torchvision-0.16.0a0+cxx11.abi-cp39-cp39-win_amd64.whl
|
71 |
-
pip install intel_extension_for_pytorch-2.1.10+xpu-cp39-cp39-win_amd64.whl
|
72 |
-
pip install --pre --upgrade bigdl-llm[xpu]
|
73 |
-
conda install libuv
|
74 |
-
```
|
75 |
-
|
76 |
-
#### Step 6
|
77 |
-
Activate the Intel oneAPI environment:
|
78 |
-
|
79 |
-
```bash
|
80 |
-
call "C:\Program Files (x86)\Intel\oneAPI\setvars.bat"
|
81 |
-
```
|
82 |
-
|
83 |
-
For the latest setup instructions for BigDL LLM inference, visit [BigDL Documentation](https://bigdl.readthedocs.io/en/latest/doc/LLM/Overview/install_gpu.html)
|
84 |
-
|
85 |
-
### Example Usage of JudgyRAG
|
86 |
-
|
87 |
-
```python
|
88 |
-
from fundialogues import benchgen, judgypdf
|
89 |
-
|
90 |
-
folder_path = ""
|
91 |
-
output_directory = ""
|
92 |
-
chunk_file = ""
|
93 |
-
benchmark_output_directory = ""
|
94 |
-
|
95 |
-
judgypdf(folder_path, output_directory)
|
96 |
-
benchgen(chunk_file, benchmark_output_directory)
|
97 |
-
```
|
98 |
# Benchmark Datasets
|
99 |
|
100 |
### opengeoquery-v1
|
101 |
OpenGeoQuery-v1 is the first edition of a benchmark dataset composed of statements associated with the geosciences. The content of the dataset touches on topics like geophysics, petrology, minerology, seismology, geomorphology, etc. The purpose of this dataset is to use as a benchmark and for fine-tuning small geoscience LLMs (coming soon).
|
102 |
|
103 |
-
|
104 |
# Dialogues
|
105 |
- Customer Service
|
106 |
- Grocery Cashier: 100 fictitious examples of dialogues between a customer at a grocery store and the cashier.
|
@@ -114,22 +27,6 @@ OpenGeoQuery-v1 is the first edition of a benchmark dataset composed of statemen
|
|
114 |
- Basketball Coach: 100 fictitious examples of dialogues between a basketball coach and the players on the court during a game.
|
115 |
|
116 |
|
117 |
-
### How to Load Dialogues
|
118 |
-
|
119 |
-
Loading dialogues can be accomplished using the fun dialogues library or Hugging Face datasets library.
|
120 |
-
|
121 |
-
### Load using fun dialogues
|
122 |
-
|
123 |
-
Assuming you've already installed fundialogues.
|
124 |
-
|
125 |
-
Use loader utility to load dataset as pandas dataframe. Further processing might be required for use.
|
126 |
-
```
|
127 |
-
from fundialogues import dialoader
|
128 |
-
|
129 |
-
# load as pandas dataframe
|
130 |
-
physics_office_hours = dialoader("FunDialogues/academia-physics-office-hours")
|
131 |
-
```
|
132 |
-
|
133 |
### Loading using Hugging Face datasets
|
134 |
|
135 |
1. Install datasets package `pip install datasets`
|
|
|
8 |
license: apache-2.0
|
9 |
---
|
10 |
# fun dialogues
|
11 |
+
A library of datasetsand open source models that can be used for training and inference for prototyping purposes.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
12 |
# Benchmark Datasets
|
13 |
|
14 |
### opengeoquery-v1
|
15 |
OpenGeoQuery-v1 is the first edition of a benchmark dataset composed of statements associated with the geosciences. The content of the dataset touches on topics like geophysics, petrology, minerology, seismology, geomorphology, etc. The purpose of this dataset is to use as a benchmark and for fine-tuning small geoscience LLMs (coming soon).
|
16 |
|
|
|
17 |
# Dialogues
|
18 |
- Customer Service
|
19 |
- Grocery Cashier: 100 fictitious examples of dialogues between a customer at a grocery store and the cashier.
|
|
|
27 |
- Basketball Coach: 100 fictitious examples of dialogues between a basketball coach and the players on the court during a game.
|
28 |
|
29 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
30 |
### Loading using Hugging Face datasets
|
31 |
|
32 |
1. Install datasets package `pip install datasets`
|