albertvillanova
HF staff
Convert dataset sizes from base 2 to base 10 in the dataset card (#1)
3a54beb
annotations_creators: | |
- crowdsourced | |
language: | |
- az | |
- be | |
- en | |
- es | |
- fr | |
- gl | |
- he | |
- it | |
- pt | |
- ru | |
- tr | |
language_creators: | |
- expert-generated | |
license: | |
- cc-by-nc-nd-4.0 | |
multilinguality: | |
- translation | |
pretty_name: TEDHrlr | |
size_categories: | |
- 1M<n<10M | |
source_datasets: | |
- extended|ted_talks_iwslt | |
task_categories: | |
- translation | |
task_ids: [] | |
paperswithcode_id: null | |
dataset_info: | |
- config_name: az_to_en | |
features: | |
- name: translation | |
dtype: | |
translation: | |
languages: | |
- az | |
- en | |
splits: | |
- name: test | |
num_bytes: 186540 | |
num_examples: 904 | |
- name: train | |
num_bytes: 1226853 | |
num_examples: 5947 | |
- name: validation | |
num_bytes: 122709 | |
num_examples: 672 | |
download_size: 131005909 | |
dataset_size: 1536102 | |
- config_name: aztr_to_en | |
features: | |
- name: translation | |
dtype: | |
translation: | |
languages: | |
- az_tr | |
- en | |
splits: | |
- name: test | |
num_bytes: 186540 | |
num_examples: 904 | |
- name: train | |
num_bytes: 39834469 | |
num_examples: 188397 | |
- name: validation | |
num_bytes: 122709 | |
num_examples: 672 | |
download_size: 131005909 | |
dataset_size: 40143718 | |
- config_name: be_to_en | |
features: | |
- name: translation | |
dtype: | |
translation: | |
languages: | |
- be | |
- en | |
splits: | |
- name: test | |
num_bytes: 186606 | |
num_examples: 665 | |
- name: train | |
num_bytes: 1176899 | |
num_examples: 4510 | |
- name: validation | |
num_bytes: 59328 | |
num_examples: 249 | |
download_size: 131005909 | |
dataset_size: 1422833 | |
- config_name: beru_to_en | |
features: | |
- name: translation | |
dtype: | |
translation: | |
languages: | |
- be_ru | |
- en | |
splits: | |
- name: test | |
num_bytes: 186606 | |
num_examples: 665 | |
- name: train | |
num_bytes: 59953616 | |
num_examples: 212615 | |
- name: validation | |
num_bytes: 59328 | |
num_examples: 249 | |
download_size: 131005909 | |
dataset_size: 60199550 | |
- config_name: es_to_pt | |
features: | |
- name: translation | |
dtype: | |
translation: | |
languages: | |
- es | |
- pt | |
splits: | |
- name: test | |
num_bytes: 343640 | |
num_examples: 1764 | |
- name: train | |
num_bytes: 8611393 | |
num_examples: 44939 | |
- name: validation | |
num_bytes: 181535 | |
num_examples: 1017 | |
download_size: 131005909 | |
dataset_size: 9136568 | |
- config_name: fr_to_pt | |
features: | |
- name: translation | |
dtype: | |
translation: | |
languages: | |
- fr | |
- pt | |
splits: | |
- name: test | |
num_bytes: 311650 | |
num_examples: 1495 | |
- name: train | |
num_bytes: 8755387 | |
num_examples: 43874 | |
- name: validation | |
num_bytes: 212317 | |
num_examples: 1132 | |
download_size: 131005909 | |
dataset_size: 9279354 | |
- config_name: gl_to_en | |
features: | |
- name: translation | |
dtype: | |
translation: | |
languages: | |
- gl | |
- en | |
splits: | |
- name: test | |
num_bytes: 193213 | |
num_examples: 1008 | |
- name: train | |
num_bytes: 1961363 | |
num_examples: 10018 | |
- name: validation | |
num_bytes: 137929 | |
num_examples: 683 | |
download_size: 131005909 | |
dataset_size: 2292505 | |
- config_name: glpt_to_en | |
features: | |
- name: translation | |
dtype: | |
translation: | |
languages: | |
- gl_pt | |
- en | |
splits: | |
- name: test | |
num_bytes: 193213 | |
num_examples: 1008 | |
- name: train | |
num_bytes: 11734254 | |
num_examples: 61803 | |
- name: validation | |
num_bytes: 137929 | |
num_examples: 683 | |
download_size: 131005909 | |
dataset_size: 12065396 | |
- config_name: he_to_pt | |
features: | |
- name: translation | |
dtype: | |
translation: | |
languages: | |
- he | |
- pt | |
splits: | |
- name: test | |
num_bytes: 361378 | |
num_examples: 1624 | |
- name: train | |
num_bytes: 10627615 | |
num_examples: 48512 | |
- name: validation | |
num_bytes: 230725 | |
num_examples: 1146 | |
download_size: 131005909 | |
dataset_size: 11219718 | |
- config_name: it_to_pt | |
features: | |
- name: translation | |
dtype: | |
translation: | |
languages: | |
- it | |
- pt | |
splits: | |
- name: test | |
num_bytes: 324726 | |
num_examples: 1670 | |
- name: train | |
num_bytes: 8905825 | |
num_examples: 46260 | |
- name: validation | |
num_bytes: 210375 | |
num_examples: 1163 | |
download_size: 131005909 | |
dataset_size: 9440926 | |
- config_name: pt_to_en | |
features: | |
- name: translation | |
dtype: | |
translation: | |
languages: | |
- pt | |
- en | |
splits: | |
- name: test | |
num_bytes: 347803 | |
num_examples: 1804 | |
- name: train | |
num_bytes: 9772911 | |
num_examples: 51786 | |
- name: validation | |
num_bytes: 207960 | |
num_examples: 1194 | |
download_size: 131005909 | |
dataset_size: 10328674 | |
- config_name: ru_to_en | |
features: | |
- name: translation | |
dtype: | |
translation: | |
languages: | |
- ru | |
- en | |
splits: | |
- name: test | |
num_bytes: 1459576 | |
num_examples: 5477 | |
- name: train | |
num_bytes: 58778442 | |
num_examples: 208107 | |
- name: validation | |
num_bytes: 1318357 | |
num_examples: 4806 | |
download_size: 131005909 | |
dataset_size: 61556375 | |
- config_name: ru_to_pt | |
features: | |
- name: translation | |
dtype: | |
translation: | |
languages: | |
- ru | |
- pt | |
splits: | |
- name: test | |
num_bytes: 409062 | |
num_examples: 1589 | |
- name: train | |
num_bytes: 11882860 | |
num_examples: 47279 | |
- name: validation | |
num_bytes: 276866 | |
num_examples: 1185 | |
download_size: 131005909 | |
dataset_size: 12568788 | |
- config_name: tr_to_en | |
features: | |
- name: translation | |
dtype: | |
translation: | |
languages: | |
- tr | |
- en | |
splits: | |
- name: test | |
num_bytes: 1026406 | |
num_examples: 5030 | |
- name: train | |
num_bytes: 38607636 | |
num_examples: 182451 | |
- name: validation | |
num_bytes: 832358 | |
num_examples: 4046 | |
download_size: 131005909 | |
dataset_size: 40466400 | |
# Dataset Card for "ted_hrlr" | |
## Table of Contents | |
- [Dataset Description](#dataset-description) | |
- [Dataset Summary](#dataset-summary) | |
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) | |
- [Languages](#languages) | |
- [Dataset Structure](#dataset-structure) | |
- [Data Instances](#data-instances) | |
- [Data Fields](#data-fields) | |
- [Data Splits](#data-splits) | |
- [Dataset Creation](#dataset-creation) | |
- [Curation Rationale](#curation-rationale) | |
- [Source Data](#source-data) | |
- [Annotations](#annotations) | |
- [Personal and Sensitive Information](#personal-and-sensitive-information) | |
- [Considerations for Using the Data](#considerations-for-using-the-data) | |
- [Social Impact of Dataset](#social-impact-of-dataset) | |
- [Discussion of Biases](#discussion-of-biases) | |
- [Other Known Limitations](#other-known-limitations) | |
- [Additional Information](#additional-information) | |
- [Dataset Curators](#dataset-curators) | |
- [Licensing Information](#licensing-information) | |
- [Citation Information](#citation-information) | |
- [Contributions](#contributions) | |
## Dataset Description | |
- **Homepage:** | |
- **Repository:** https://github.com/neulab/word-embeddings-for-nmt | |
- **Paper:** [When and Why Are Pre-Trained Word Embeddings Useful for Neural Machine Translation?](https://aclanthology.org/N18-2084/) | |
- **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | |
- **Size of downloaded dataset files:** 1.83 GB | |
- **Size of the generated dataset:** 281.66 MB | |
- **Total amount of disk used:** 2.12 GB | |
### Dataset Summary | |
Data sets derived from TED talk transcripts for comparing similar language pairs | |
where one is high resource and the other is low resource. | |
### Supported Tasks and Leaderboards | |
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | |
### Languages | |
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | |
## Dataset Structure | |
### Data Instances | |
#### az_to_en | |
- **Size of downloaded dataset files:** 131.01 MB | |
- **Size of the generated dataset:** 1.53 MB | |
- **Total amount of disk used:** 132.54 MB | |
An example of 'train' looks as follows. | |
``` | |
{ | |
"translation": { | |
"az": "zəhmət olmasa , sizə xitab edən sözlər eşidəndə əlinizi qaldırın .", | |
"en": "please raise your hand if something applies to you ." | |
} | |
} | |
``` | |
#### aztr_to_en | |
- **Size of downloaded dataset files:** 131.01 MB | |
- **Size of the generated dataset:** 40.14 MB | |
- **Total amount of disk used:** 171.15 MB | |
An example of 'train' looks as follows. | |
``` | |
{ | |
"translation": { | |
"az_tr": "zəhmət olmasa , sizə xitab edən sözlər eşidəndə əlinizi qaldırın .", | |
"en": "please raise your hand if something applies to you ." | |
} | |
} | |
``` | |
#### be_to_en | |
- **Size of downloaded dataset files:** 131.01 MB | |
- **Size of the generated dataset:** 1.43 MB | |
- **Total amount of disk used:** 132.42 MB | |
An example of 'train' looks as follows. | |
``` | |
{ | |
"translation": { | |
"be": "zəhmət olmasa , sizə xitab edən sözlər eşidəndə əlinizi qaldırın .", | |
"en": "please raise your hand if something applies to you ." | |
} | |
} | |
``` | |
#### beru_to_en | |
- **Size of downloaded dataset files:** 131.01 MB | |
- **Size of the generated dataset:** 60.20 MB | |
- **Total amount of disk used:** 191.21 MB | |
An example of 'validation' looks as follows. | |
``` | |
This example was too long and was cropped: | |
{ | |
"translation": "{\"be_ru\": \"11 yaşımdaydım . səhərin birində , evimizdəki sevinc səslərinə oyandığım indiki kimi yadımdadır .\", \"en\": \"when i was..." | |
} | |
``` | |
#### es_to_pt | |
- **Size of downloaded dataset files:** 131.01 MB | |
- **Size of the generated dataset:** 9.13 MB | |
- **Total amount of disk used:** 140.14 MB | |
An example of 'validation' looks as follows. | |
``` | |
This example was too long and was cropped: | |
{ | |
"translation": "{\"es\": \"11 yaşımdaydım . səhərin birində , evimizdəki sevinc səslərinə oyandığım indiki kimi yadımdadır .\", \"pt\": \"when i was 11..." | |
} | |
``` | |
### Data Fields | |
The data fields are the same among all splits. | |
#### az_to_en | |
- `translation`: a multilingual `string` variable, with possible languages including `az`, `en`. | |
#### aztr_to_en | |
- `translation`: a multilingual `string` variable, with possible languages including `az_tr`, `en`. | |
#### be_to_en | |
- `translation`: a multilingual `string` variable, with possible languages including `be`, `en`. | |
#### beru_to_en | |
- `translation`: a multilingual `string` variable, with possible languages including `be_ru`, `en`. | |
#### es_to_pt | |
- `translation`: a multilingual `string` variable, with possible languages including `es`, `pt`. | |
### Data Splits | |
| name |train |validation|test| | |
|----------|-----:|---------:|---:| | |
|az_to_en | 5947| 672| 904| | |
|aztr_to_en|188397| 672| 904| | |
|be_to_en | 4510| 249| 665| | |
|beru_to_en|212615| 249| 665| | |
|es_to_pt | 44939| 1017|1764| | |
## Dataset Creation | |
### Curation Rationale | |
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | |
### Source Data | |
#### Initial Data Collection and Normalization | |
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | |
#### Who are the source language producers? | |
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | |
### Annotations | |
#### Annotation process | |
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | |
#### Who are the annotators? | |
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | |
### Personal and Sensitive Information | |
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | |
## Considerations for Using the Data | |
### Social Impact of Dataset | |
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | |
### Discussion of Biases | |
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | |
### Other Known Limitations | |
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | |
## Additional Information | |
### Dataset Curators | |
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | |
### Licensing Information | |
[More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards) | |
### Citation Information | |
``` | |
@inproceedings{qi-etal-2018-pre, | |
title = "When and Why Are Pre-Trained Word Embeddings Useful for Neural Machine Translation?", | |
author = "Qi, Ye and | |
Sachan, Devendra and | |
Felix, Matthieu and | |
Padmanabhan, Sarguna and | |
Neubig, Graham", | |
booktitle = "Proceedings of the 2018 Conference of the North {A}merican Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers)", | |
month = jun, | |
year = "2018", | |
address = "New Orleans, Louisiana", | |
publisher = "Association for Computational Linguistics", | |
url = "https://aclanthology.org/N18-2084", | |
doi = "10.18653/v1/N18-2084", | |
pages = "529--535", | |
} | |
``` | |
### Contributions | |
Thanks to [@thomwolf](https://github.com/thomwolf), [@lewtun](https://github.com/lewtun), [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset. |