|
--- |
|
language: |
|
- fr |
|
- en |
|
license: cc-by-sa-3.0 |
|
configs: |
|
- config_name: default |
|
data_files: |
|
- split: train |
|
path: data/train-* |
|
- split: eval |
|
path: data/eval-* |
|
- split: test |
|
path: data/test-* |
|
dataset_info: |
|
features: |
|
- name: instruction |
|
dtype: string |
|
- name: context |
|
dtype: string |
|
- name: category |
|
dtype: string |
|
- name: fr_context |
|
dtype: string |
|
- name: fr_response |
|
dtype: string |
|
- name: fr_instruction |
|
dtype: string |
|
- name: response |
|
dtype: string |
|
- name: qid |
|
dtype: int64 |
|
splits: |
|
- name: train |
|
num_bytes: 9785049.205202311 |
|
num_examples: 3300 |
|
- name: eval |
|
num_bytes: 1340255.2244701348 |
|
num_examples: 452 |
|
- name: test |
|
num_bytes: 1186066.570327553 |
|
num_examples: 400 |
|
download_size: 7746263 |
|
dataset_size: 12311371.0 |
|
--- |
|
# Dataset Card for "dolly_context_enfr" |
|
|
|
This is a filtered version of [databricks-dolly-15k](https://huggingface.co/datasets/databricks/databricks-dolly-15k), |
|
then traduced to french with Deepl pro API, the best translation solution available on the market. |
|
|
|
Our goal is to gather french data on question answering on context, where the model should not bring new information |
|
not present in the context given. Our goal is to limit hallucination. |
|
The filtering have been done in three parts: |
|
- We keep only the data with a not empty context (we are not interested in random chat or not sourced information) |
|
- We don't take data where the answer is more than 1,5 times longer than the context, our study of the data showed |
|
that in those cases the information come from other sources than the context, and/or concist of a copy past of the context |
|
- For long context data (>1000 characters), we don't take data where the answer is longer than context (character wize) |
|
- We also filter around 30 data with too long context (10k character), answer (5k character) and instruction (5k character) |
|
as ther were showed to have a wrong format |
|
|
|
Our filtered version of dolly dataset only contain 3 of the 7 categories, |
|
the annotation guidelines for each of the categories were as follows: |
|
|
|
- **Closed QA**: Write a question or instruction that requires factually correct response based on a passage of text from Wikipedia. The question can be complex and can involve human-level reasoning capabilities, but should not require special knowledge. To create a question for this task include both the text of the question as well as the reference text in the form. |
|
- **Summarization**: Give a summary of a paragraph from Wikipedia. Please don't ask questions that will require more than 3-5 minutes to answer. To create a question for this task include both the text of the question as well as the reference text in the form. |
|
- **Information Extraction**: These questions involve reading a paragraph from Wikipedia and extracting information from the passage. Everything required to produce an answer (e.g. a list, keywords etc) should be included in the passages. To create a question for this task include both the text of the question as well as the reference text in the form. |
|
|
|
| Category | Samples | |
|
| - | - | |
|
| closed_qa | 1711 | |
|
| information_extraction | 1377 | |
|
| summarization | 1064 | |
|
|
|
Note that we considered 'brainstorming' and 'classification' data, |
|
but there are not suited for our LLM project, and very subjective (as not based on a context), so we decided to not use them. |
|
|
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/62ce7972a1006f883519d88a/h_qjY7Tt5INoylK3oOvFA.png) |
|
|