parquet-converter commited on
Commit
85a62cc
·
1 Parent(s): 61a5b55

Update parquet files

Browse files
README.md DELETED
@@ -1,20 +0,0 @@
1
- ---
2
- annotations_creators: []
3
- language:
4
- - en
5
- - ro
6
- language_creators: []
7
- license: []
8
- multilinguality:
9
- - translation
10
- pretty_name: CCMatrix (en-ro)
11
- size_categories:
12
- - 100K<n<1M
13
- source_datasets: []
14
- tags: []
15
- task_categories:
16
- - translation
17
- task_ids: []
18
- ---
19
- A sampled version of the [CCMatrix](https://huggingface.co/datasets/yhavinga/ccmatrix) dataset for the English-Romanian pair, containing 1M train entries.
20
- Please refer to the original for more info.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
dataset_infos.json DELETED
@@ -1,51 +0,0 @@
1
- {"din0s--ccmatrix_en-ro": {
2
- "description": "CCMatrix: Mining Billions of High-Quality Parallel Sentences on the WEB\n\nWe show that margin-based bitext mining in LASER's multilingual sentence space can be applied to\nmonolingual corpora of billions of sentences to produce high quality aligned translation data.\nWe use thirty-two snapshots of a curated common crawl corpus [1] totaling 69 billion unique sentences.\nUsing one unified approach for 80 languages, we were able to mine 10.8 billion parallel sentences,\nout of which only 2.9 billion are aligned with English.\n\nIMPORTANT: Please cite reference [2][3] if you use this data.\n\n[1] Guillaume Wenzek, Marie-Anne Lachaux, Alexis Conneau, Vishrav Chaudhary, Francisco Guzm\u00e1n, Armand Jouli\n and Edouard Grave, CCNet: Extracting High Quality Monolingual Datasets from Web Crawl Data\n\n[2] Holger Schwenk, Guillaume Wenzek, Sergey Edunov, Edouard Grave and Armand Joulin,\n CCMatrix: Mining Billions of High-Quality Parallel Sentences on the WEB\n\n[3] Angela Fan, Shruti Bhosale, Holger Schwenk, Zhiyi Ma, Ahmed El-Kishky, Siddharth Goyal, Mandeep Baines,\n Onur Celebi, Guillaume Wenzek, Vishrav Chaudhary, Naman Goyal, Tom Birch, Vitaliy Liptchinsky,\n Sergey Edunov, Edouard Grave, Michael Auli, and Armand Joulin.\n Beyond English-Centric Multilingual Machine Translation\n \n90 languages, 1,197 bitexts\ntotal number of files: 90\ntotal number of tokens: 112.14G\ntotal number of sentence fragments: 7.37G\n",
3
- "citation": " Guillaume Wenzek, Marie-Anne Lachaux, Alexis Conneau, Vishrav Chaudhary, Francisco Guzm\u00e1n, Armand Jouli and Edouard Grave, CCNet: Extracting High Quality Monolingual Datasets from Web Crawl Data\n",
4
- "homepage": "https://opus.nlpl.eu/CCMatrix.php",
5
- "license": "",
6
- "features": {
7
- "id": {
8
- "dtype": "int32",
9
- "id": null,
10
- "_type": "Value"
11
- },
12
- "score": {
13
- "dtype": "float32",
14
- "id": null,
15
- "_type": "Value"
16
- },
17
- "translation": {
18
- "languages": [
19
- "en",
20
- "ro"
21
- ],
22
- "id": null,
23
- "_type": "Translation"
24
- }
25
- },
26
- "post_processed": null,
27
- "supervised_keys": null,
28
- "task_templates": null,
29
- "builder_name": "ccmatrix",
30
- "config_name": "en-ro",
31
- "version": {
32
- "version_str": "1.0.0",
33
- "description": null,
34
- "major": 1,
35
- "minor": 0,
36
- "patch": 0
37
- },
38
- "splits": {
39
- "train": {
40
- "name": "train",
41
- "num_bytes": 225154422.65269262,
42
- "num_examples": 1000000,
43
- "dataset_name": "ccmatrix_en-ro"
44
- }
45
- },
46
- "download_checksums": null,
47
- "download_size": 167665274,
48
- "post_processing_size": null,
49
- "dataset_size": 225154422.65269262,
50
- "size_in_bytes": 392819696.6526926
51
- }}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
data/train-00000-of-00001-7dea4209c77eafbe.parquet → din0s--ccmatrix_en-ro/parquet-train.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:31097d6be9d88dd6c5bf9bd6570da5060825ae9b99fa07ca358e69183ec1e72c
3
- size 167665274
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:134851655c2783dd945ce127e3fe18337baca7c779938883f77a204ab2f0dcba
3
+ size 168902964