eduardo-alvarez commited on
Commit
4cd237d
·
verified ·
1 Parent(s): 8fd271f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -104
README.md CHANGED
@@ -8,99 +8,12 @@ pinned: false
8
  license: apache-2.0
9
  ---
10
  # fun dialogues
11
- A library of datasets, LLM tooling, and open source models that can be used for training and inference for prototyping purposes. The project began as a collection of fictitious dialogues that can be used to train language models or augment prompts for prototyping and educational purposes. It has now grown to include tooling for LLM application development purposes and open source models.
12
-
13
- repository link: https://github.com/eduand-alvarez/fun-dialogues
14
-
15
- ### Key components of fun dialogues:
16
- - **JudgyRAG**: RAG eval tool
17
- - **Benchmark Datasets**: custom benchmark datasets
18
- - **Dialogues**: fictitious dialogue datasets for fun, experimentation, and testing.
19
-
20
- You can install fun dialogues using pip: `pip install fundialogues`
21
-
22
- <div align="center">
23
- <img src="https://github.com/eduand-alvarez/fun-dialogues/assets/57263404/1d8ce401-b595-442f-980c-8ae06ed9d4b2" alt="Fun Dialogues Logo" width="750"/>
24
- </div>
25
-
26
- # JudgyRAG
27
-
28
- JudgyRAG is a component of the FunDialogues Python library focused on evaluating the performance of Retrieval-Augmented Generation (RAG) systems. It facilitates this by creating synthetic datasets based on custom datasets, enabling a unique assessment of a RAG system's question-answering capabilities in a zero-shot Q&A context. Initially, JudgyRAG's primary functionality is the automatic generation of custom multiple-choice Q&A datasets. Future iterations will introduce further automation to seamlessly integrate with popular frameworks, enhancing testing and benchmarking processes.
29
-
30
- ### Workflow
31
-
32
- The workflow for JudgyRAG includes:
33
-
34
- 1. **Scraping PDFs**: Information is extracted from PDFs into structured text formats.
35
- 2. **Chunking Data**: Extracted data is chunked similarly to vector database embeddings for RAG, simulating data breakdown and storage.
36
- 3. **Question Generation**: Each chunk acts as a knowledge base, with custom prompts instructing supported models (currently LLaMA 7B and 13B chat) to generate multiple-choice questions.
37
- 4. **Iterative Parsing**: Chunks are processed iteratively, generating a multiple-choice question for each.
38
- 5. **Quality Checks**: Poor-quality chunks leading to failed question generation are flagged for user review.
39
- 6. **Benchmark Compilation**: The final document includes multiple-choice questions, correct answers, and source knowledge chunks.
40
- 7. **RAG System Evaluation**: The synthetic benchmark dataset can be used to assess a RAG system, with automation for this process planned for future updates.
41
-
42
- ### Environment Setup
43
-
44
- Follow these steps to set up your environment for JudgyRAG:
45
-
46
- #### Step 1
47
- Install Visual Studio 2022 Community Edition with the “Desktop development with C++” workload.
48
-
49
- #### Step 2
50
- Update to the latest GPU driver.
51
-
52
- #### Step 3
53
- Install the Intel® oneAPI Base Toolkit 2024.0.
54
-
55
- #### Step 4
56
- Download the necessary wheels:
57
-
58
- ```bash
59
- wget https://intel-extension-for-pytorch.s3.amazonaws.com/ipex_stable/xpu/torch-2.1.0a0%2Bcxx11.abi-cp39-cp39-win_amd64.whl
60
- wget https://intel-extension-for-pytorch.s3.amazonaws.com/ipex_stable/xpu/torchvision-0.16.0a0%2Bcxx11.abi-cp39-cp39-win_amd64.whl
61
- wget https://intel-extension-for-pytorch.s3.amazonaws.com/ipex_stable/xpu/intel_extension_for_pytorch-2.1.10%2Bxpu-cp39-cp39-win_amd64.whl
62
- ```
63
-
64
- #### Step 5
65
-
66
- Install the downloaded packages and BigDL LLM:
67
-
68
- ```bash
69
- pip install torch-2.1.0a0+cxx11.abi-cp39-cp39-win_amd64.whl
70
- pip install torchvision-0.16.0a0+cxx11.abi-cp39-cp39-win_amd64.whl
71
- pip install intel_extension_for_pytorch-2.1.10+xpu-cp39-cp39-win_amd64.whl
72
- pip install --pre --upgrade bigdl-llm[xpu]
73
- conda install libuv
74
- ```
75
-
76
- #### Step 6
77
- Activate the Intel oneAPI environment:
78
-
79
- ```bash
80
- call "C:\Program Files (x86)\Intel\oneAPI\setvars.bat"
81
- ```
82
-
83
- For the latest setup instructions for BigDL LLM inference, visit [BigDL Documentation](https://bigdl.readthedocs.io/en/latest/doc/LLM/Overview/install_gpu.html)
84
-
85
- ### Example Usage of JudgyRAG
86
-
87
- ```python
88
- from fundialogues import benchgen, judgypdf
89
-
90
- folder_path = ""
91
- output_directory = ""
92
- chunk_file = ""
93
- benchmark_output_directory = ""
94
-
95
- judgypdf(folder_path, output_directory)
96
- benchgen(chunk_file, benchmark_output_directory)
97
- ```
98
  # Benchmark Datasets
99
 
100
  ### opengeoquery-v1
101
  OpenGeoQuery-v1 is the first edition of a benchmark dataset composed of statements associated with the geosciences. The content of the dataset touches on topics like geophysics, petrology, minerology, seismology, geomorphology, etc. The purpose of this dataset is to use as a benchmark and for fine-tuning small geoscience LLMs (coming soon).
102
 
103
-
104
  # Dialogues
105
  - Customer Service
106
  - Grocery Cashier: 100 fictitious examples of dialogues between a customer at a grocery store and the cashier.
@@ -114,22 +27,6 @@ OpenGeoQuery-v1 is the first edition of a benchmark dataset composed of statemen
114
  - Basketball Coach: 100 fictitious examples of dialogues between a basketball coach and the players on the court during a game.
115
 
116
 
117
- ### How to Load Dialogues
118
-
119
- Loading dialogues can be accomplished using the fun dialogues library or Hugging Face datasets library.
120
-
121
- ### Load using fun dialogues
122
-
123
- Assuming you've already installed fundialogues.
124
-
125
- Use loader utility to load dataset as pandas dataframe. Further processing might be required for use.
126
- ```
127
- from fundialogues import dialoader
128
-
129
- # load as pandas dataframe
130
- physics_office_hours = dialoader("FunDialogues/academia-physics-office-hours")
131
- ```
132
-
133
  ### Loading using Hugging Face datasets
134
 
135
  1. Install datasets package `pip install datasets`
 
8
  license: apache-2.0
9
  ---
10
  # fun dialogues
11
+ A library of datasetsand open source models that can be used for training and inference for prototyping purposes.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
  # Benchmark Datasets
13
 
14
  ### opengeoquery-v1
15
  OpenGeoQuery-v1 is the first edition of a benchmark dataset composed of statements associated with the geosciences. The content of the dataset touches on topics like geophysics, petrology, minerology, seismology, geomorphology, etc. The purpose of this dataset is to use as a benchmark and for fine-tuning small geoscience LLMs (coming soon).
16
 
 
17
  # Dialogues
18
  - Customer Service
19
  - Grocery Cashier: 100 fictitious examples of dialogues between a customer at a grocery store and the cashier.
 
27
  - Basketball Coach: 100 fictitious examples of dialogues between a basketball coach and the players on the court during a game.
28
 
29
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30
  ### Loading using Hugging Face datasets
31
 
32
  1. Install datasets package `pip install datasets`