Datasets:

Modalities:
Tabular
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
pandas
License:
min-wikisplit-pp / README.md
hpprc's picture
Update README.md
d5749f7 verified
---
dataset_info:
features:
- name: id
dtype: int64
- name: complex
dtype: string
- name: simple_reversed
dtype: string
- name: simple_tokenized
sequence: string
- name: simple_original
dtype: string
- name: entailment_prob
dtype: float64
splits:
- name: train
num_bytes: 115032683
num_examples: 139241
- name: validation
num_bytes: 14334442
num_examples: 17424
- name: test
num_bytes: 14285722
num_examples: 17412
download_size: 91848881
dataset_size: 143652847
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
- split: test
path: data/test-*
license: cc-by-sa-4.0
task_categories:
- text2text-generation
language:
- en
pretty_name: MinWikiSplit++
---
# MinWikiSplit++
This dataset is the HuggingFace version of MinWikiSplit++.
MinWikiSplit++ enhances the original [MinWikiSplit](https://aclanthology.org/W19-8615/) by applying two techniques: filtering through NLI classification and sentence-order reversing, which help to remove noise and reduce hallucinations compared to the original MinWikiSplit.
The preprocessed MinWikiSplit dataset that formed the basis for this can be found [here](https://huggingface.co/datasets/cl-nagoya/min-wikisplit).
## Dataset Description
- **Repository:** https://github.com/nttcslab-nlp/wikisplit-pp
- **Paper:** https://arxiv.org/abs/2404.09002
- **Point of Contact:** [Hayato Tsukagoshi](mailto:[email protected])
## Usage
```python
import datasets as ds
dataset: ds.DatasetDict = ds.load_dataset("cl-nagoya/min-wikisplit-pp")
print(dataset)
# DatasetDict({
# train: Dataset({
# features: ['id', 'complex', 'simple_reversed', 'simple_tokenized', 'simple_original', 'entailment_prob'],
# num_rows: 139241
# })
# validation: Dataset({
# features: ['id', 'complex', 'simple_reversed', 'simple_tokenized', 'simple_original', 'entailment_prob'],
# num_rows: 17424
# })
# test: Dataset({
# features: ['id', 'complex', 'simple_reversed', 'simple_tokenized', 'simple_original', 'entailment_prob'],
# num_rows: 17412
# })
# })
```
### Data Fields
- id: The ID of the data (note that it is not compatible with the existing MinWikiSplit)
- complex: A complex sentence
- simple_reversed: Simple sentences with their order reversed
- simple_tokenized: A list of simple sentences split by [PySBD](https://github.com/nipunsadvilkar/pySBD), not reversed in order
- simple_original: Simple sentences in their original order
- entailment_prob: The average probability that each simple sentence is classified as an entailment according to the complex sentence. [DeBERTa-xxl](https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli) is used for the NLI classification.
## Paper
Tsukagoshi et al., [WikiSplit++: Easy Data Refinement for Split and Rephrase](https://arxiv.org/abs/2404.09002), LREC-COLING 2024.
## License
MinWikiSplit is build upon the [WikiSplit](https://github.com/google-research-datasets/wiki-split) dataset, which is distributed under the CC-BY-SA 4.0 license.
Therefore, this dataset follows suit and is distributed under the CC-BY-SA 4.0 license.