parquet-converter commited on
Commit
8809cf0
·
1 Parent(s): 69ba4e2

Update parquet files

Browse files
.gitattributes DELETED
@@ -1,37 +0,0 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ftz filter=lfs diff=lfs merge=lfs -text
6
- *.gz filter=lfs diff=lfs merge=lfs -text
7
- *.h5 filter=lfs diff=lfs merge=lfs -text
8
- *.joblib filter=lfs diff=lfs merge=lfs -text
9
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
10
- *.model filter=lfs diff=lfs merge=lfs -text
11
- *.msgpack filter=lfs diff=lfs merge=lfs -text
12
- *.onnx filter=lfs diff=lfs merge=lfs -text
13
- *.ot filter=lfs diff=lfs merge=lfs -text
14
- *.parquet filter=lfs diff=lfs merge=lfs -text
15
- *.pb filter=lfs diff=lfs merge=lfs -text
16
- *.pt filter=lfs diff=lfs merge=lfs -text
17
- *.pth filter=lfs diff=lfs merge=lfs -text
18
- *.rar filter=lfs diff=lfs merge=lfs -text
19
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
20
- *.tar.* filter=lfs diff=lfs merge=lfs -text
21
- *.tflite filter=lfs diff=lfs merge=lfs -text
22
- *.tgz filter=lfs diff=lfs merge=lfs -text
23
- *.wasm filter=lfs diff=lfs merge=lfs -text
24
- *.xz filter=lfs diff=lfs merge=lfs -text
25
- *.zip filter=lfs diff=lfs merge=lfs -text
26
- *.zstandard filter=lfs diff=lfs merge=lfs -text
27
- *tfevents* filter=lfs diff=lfs merge=lfs -text
28
- # Audio files - uncompressed
29
- *.pcm filter=lfs diff=lfs merge=lfs -text
30
- *.sam filter=lfs diff=lfs merge=lfs -text
31
- *.raw filter=lfs diff=lfs merge=lfs -text
32
- # Audio files - compressed
33
- *.aac filter=lfs diff=lfs merge=lfs -text
34
- *.flac filter=lfs diff=lfs merge=lfs -text
35
- *.mp3 filter=lfs diff=lfs merge=lfs -text
36
- *.ogg filter=lfs diff=lfs merge=lfs -text
37
- *.wav filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
.gitignore DELETED
@@ -1 +0,0 @@
1
- .ipynb_checkpoints/
 
 
Generator.ipynb DELETED
@@ -1,186 +0,0 @@
1
- {
2
- "cells": [
3
- {
4
- "cell_type": "code",
5
- "execution_count": 2,
6
- "id": "bbd1b7a1-dbb7-4243-99e0-70a6cd47d573",
7
- "metadata": {},
8
- "outputs": [
9
- {
10
- "data": {
11
- "application/vnd.jupyter.widget-view+json": {
12
- "model_id": "bcc2f5482d8342a7915cecf9e7855531",
13
- "version_major": 2,
14
- "version_minor": 0
15
- },
16
- "text/plain": [
17
- "VBox(children=(HTML(value='<center> <img\\nsrc=https://huggingface.co/front/assets/huggingface_logo-noborder.sv…"
18
- ]
19
- },
20
- "metadata": {},
21
- "output_type": "display_data"
22
- }
23
- ],
24
- "source": [
25
- "from huggingface_hub import notebook_login\n",
26
- "notebook_login()"
27
- ]
28
- },
29
- {
30
- "cell_type": "code",
31
- "execution_count": 13,
32
- "id": "306958c8-4603-4b9b-b941-6a824777164d",
33
- "metadata": {},
34
- "outputs": [],
35
- "source": [
36
- "import librosa\n",
37
- "import math\n",
38
- "import pyarrow as pa\n",
39
- "import pandas as pd\n",
40
- "from datasets import load_dataset_builder, SplitGenerator, Split, Dataset, table"
41
- ]
42
- },
43
- {
44
- "cell_type": "code",
45
- "execution_count": 14,
46
- "id": "4ac69d3b-38c6-49af-aefe-63755bf3f0e9",
47
- "metadata": {},
48
- "outputs": [],
49
- "source": [
50
- "SAMPLE_RATE = 16_000\n",
51
- "MAX_LENGTH_IN_SECONDS = 20.0\n",
52
- "\n",
53
- "def add_audio(file, words):\n",
54
- " audio, _ = librosa.load(file, sr=SAMPLE_RATE)\n",
55
- " return {\n",
56
- " \"audio\": audio,\n",
57
- " }"
58
- ]
59
- },
60
- {
61
- "cell_type": "code",
62
- "execution_count": 15,
63
- "id": "9192b631-388f-4306-b975-9ba770b9dc4d",
64
- "metadata": {},
65
- "outputs": [],
66
- "source": [
67
- "audio, _ = librosa.load('clips/1.wav', sr=SAMPLE_RATE)\n",
68
- " \n",
69
- "df = pd.DataFrame({\n",
70
- " 'audio': [audio],\n",
71
- " 'text': ['bjorn.'],\n",
72
- "})\n",
73
- "tbl = table.InMemoryTable(\n",
74
- " pa.Table.from_pandas(df)\n",
75
- ")\n",
76
- "ds = Dataset(tbl, split=[\"test\", \"training\"])"
77
- ]
78
- },
79
- {
80
- "cell_type": "code",
81
- "execution_count": 17,
82
- "id": "f37d68ea-cbe7-4dd1-8215-f9449fe047f4",
83
- "metadata": {},
84
- "outputs": [],
85
- "source": [
86
- "ds.save_to_disk(\"data/test/\")\n",
87
- "ds.save_to_disk(\"data/training/\")"
88
- ]
89
- },
90
- {
91
- "cell_type": "code",
92
- "execution_count": 12,
93
- "id": "bac1a601-a7a1-434e-917d-0e372684f56b",
94
- "metadata": {},
95
- "outputs": [
96
- {
97
- "name": "stderr",
98
- "output_type": "stream",
99
- "text": [
100
- "Resuming upload of the dataset shards.\n"
101
- ]
102
- },
103
- {
104
- "data": {
105
- "application/vnd.jupyter.widget-view+json": {
106
- "model_id": "61cfa14ccb514ff4961072752bc3d4da",
107
- "version_major": 2,
108
- "version_minor": 0
109
- },
110
- "text/plain": [
111
- "Pushing dataset shards to the dataset hub: 0%| | 0/1 [00:00<?, ?it/s]"
112
- ]
113
- },
114
- "metadata": {},
115
- "output_type": "display_data"
116
- },
117
- {
118
- "data": {
119
- "application/vnd.jupyter.widget-view+json": {
120
- "model_id": "5803c7d37ce1426794af8ad65f618275",
121
- "version_major": 2,
122
- "version_minor": 0
123
- },
124
- "text/plain": [
125
- "Downloading metadata: 0%| | 0.00/1.20k [00:00<?, ?B/s]"
126
- ]
127
- },
128
- "metadata": {},
129
- "output_type": "display_data"
130
- },
131
- {
132
- "name": "stderr",
133
- "output_type": "stream",
134
- "text": [
135
- "Updating downloaded metadata with the new split.\n"
136
- ]
137
- },
138
- {
139
- "ename": "ValueError",
140
- "evalue": "Split ['test', 'training'] already present",
141
- "output_type": "error",
142
- "traceback": [
143
- "\u001b[0;31m---------------------------------------------------------------------------\u001b[0m",
144
- "\u001b[0;31mValueError\u001b[0m Traceback (most recent call last)",
145
- "Input \u001b[0;32mIn [12]\u001b[0m, in \u001b[0;36m<cell line: 1>\u001b[0;34m()\u001b[0m\n\u001b[0;32m----> 1\u001b[0m \u001b[43mds\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mpush_to_hub\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[38;5;124;43msharpcoder/bjorn_training\u001b[39;49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[43m)\u001b[49m\n",
146
- "File \u001b[0;32m~/.local/lib/python3.10/site-packages/datasets/arrow_dataset.py:4342\u001b[0m, in \u001b[0;36mDataset.push_to_hub\u001b[0;34m(self, repo_id, split, private, token, branch, max_shard_size, shard_size, embed_external_files)\u001b[0m\n\u001b[1;32m 4340\u001b[0m repo_info\u001b[38;5;241m.\u001b[39mdataset_size \u001b[38;5;241m+\u001b[39m\u001b[38;5;241m=\u001b[39m dataset_nbytes\n\u001b[1;32m 4341\u001b[0m repo_info\u001b[38;5;241m.\u001b[39msize_in_bytes \u001b[38;5;241m=\u001b[39m repo_info\u001b[38;5;241m.\u001b[39mdownload_size \u001b[38;5;241m+\u001b[39m repo_info\u001b[38;5;241m.\u001b[39mdataset_size\n\u001b[0;32m-> 4342\u001b[0m repo_info\u001b[38;5;241m.\u001b[39msplits[split] \u001b[38;5;241m=\u001b[39m SplitInfo(\n\u001b[1;32m 4343\u001b[0m split, num_bytes\u001b[38;5;241m=\u001b[39mdataset_nbytes, num_examples\u001b[38;5;241m=\u001b[39m\u001b[38;5;28mlen\u001b[39m(\u001b[38;5;28mself\u001b[39m), dataset_name\u001b[38;5;241m=\u001b[39mdataset_name\n\u001b[1;32m 4344\u001b[0m )\n\u001b[1;32m 4345\u001b[0m info_to_dump \u001b[38;5;241m=\u001b[39m repo_info\n\u001b[1;32m 4346\u001b[0m buffer \u001b[38;5;241m=\u001b[39m BytesIO()\n",
147
- "File \u001b[0;32m~/.local/lib/python3.10/site-packages/datasets/splits.py:523\u001b[0m, in \u001b[0;36mSplitDict.__setitem__\u001b[0;34m(self, key, value)\u001b[0m\n\u001b[1;32m 521\u001b[0m \u001b[38;5;28;01mraise\u001b[39;00m \u001b[38;5;167;01mValueError\u001b[39;00m(\u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mCannot add elem. (key mismatch: \u001b[39m\u001b[38;5;124m'\u001b[39m\u001b[38;5;132;01m{\u001b[39;00mkey\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m'\u001b[39m\u001b[38;5;124m != \u001b[39m\u001b[38;5;124m'\u001b[39m\u001b[38;5;132;01m{\u001b[39;00mvalue\u001b[38;5;241m.\u001b[39mname\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m'\u001b[39m\u001b[38;5;124m)\u001b[39m\u001b[38;5;124m\"\u001b[39m)\n\u001b[1;32m 522\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m key \u001b[38;5;129;01min\u001b[39;00m \u001b[38;5;28mself\u001b[39m:\n\u001b[0;32m--> 523\u001b[0m \u001b[38;5;28;01mraise\u001b[39;00m \u001b[38;5;167;01mValueError\u001b[39;00m(\u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mSplit \u001b[39m\u001b[38;5;132;01m{\u001b[39;00mkey\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m already present\u001b[39m\u001b[38;5;124m\"\u001b[39m)\n\u001b[1;32m 524\u001b[0m \u001b[38;5;28msuper\u001b[39m()\u001b[38;5;241m.\u001b[39m\u001b[38;5;21m__setitem__\u001b[39m(key, value)\n",
148
- "\u001b[0;31mValueError\u001b[0m: Split ['test', 'training'] already present"
149
- ]
150
- }
151
- ],
152
- "source": [
153
- "# ds.push_to_hub(\"sharpcoder/bjorn_training\")"
154
- ]
155
- },
156
- {
157
- "cell_type": "code",
158
- "execution_count": null,
159
- "id": "b070517c-2dfc-4f1b-baed-1748a9d5f088",
160
- "metadata": {},
161
- "outputs": [],
162
- "source": []
163
- }
164
- ],
165
- "metadata": {
166
- "kernelspec": {
167
- "display_name": "Python 3 (ipykernel)",
168
- "language": "python",
169
- "name": "python3"
170
- },
171
- "language_info": {
172
- "codemirror_mode": {
173
- "name": "ipython",
174
- "version": 3
175
- },
176
- "file_extension": ".py",
177
- "mimetype": "text/x-python",
178
- "name": "python",
179
- "nbconvert_exporter": "python",
180
- "pygments_lexer": "ipython3",
181
- "version": "3.10.4"
182
- }
183
- },
184
- "nbformat": 4,
185
- "nbformat_minor": 5
186
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
clips/1.wav DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:410b562cc93b4093b3df7cffd1b04f72fef8667b14d9e10052bf432fbd5e1eaa
3
- size 279144
 
 
 
 
dataset_infos.json DELETED
@@ -1,56 +0,0 @@
1
- {"sharpcoder--bjorn_training": {
2
- "description": "",
3
- "citation": "",
4
- "homepage": "",
5
- "license": "",
6
- "features": {
7
- "audio": [
8
- {
9
- "array": {
10
- "feature": {
11
- "dtype": "float32",
12
- "id": null,
13
- "_type": "Value"
14
- },
15
- "length": -1,
16
- "id": null,
17
- "_type": "Sequence"
18
- },
19
- "file": {
20
- "dtype": "string",
21
- "id": null,
22
- "_type": "Value"
23
- },
24
- "sample_rate": {
25
- "dtype": "int64",
26
- "id": null,
27
- "_type": "Value"
28
- }
29
- }
30
- ],
31
- "text": {
32
- "dtype": "string",
33
- "id": null,
34
- "_type": "Value"
35
- }
36
- },
37
- "post_processed": null,
38
- "supervised_keys": null,
39
- "task_templates": null,
40
- "builder_name": null,
41
- "config_name": null,
42
- "version": null,
43
- "splits": {
44
- "train": {
45
- "name": "train",
46
- "num_bytes": 889041,
47
- "num_examples": 4,
48
- "dataset_name": "bjorn_training"
49
- }
50
- },
51
- "download_checksums": null,
52
- "download_size": 669477,
53
- "post_processing_size": null,
54
- "dataset_size": 889041,
55
- "size_in_bytes": 1558518
56
- }}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
data/train-00000-of-00001-0556263be39146c2.parquet → sharpcoder--bjorn_training/parquet-train.parquet RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ccc74c69354ffc1d314a7b103f0e16a3b3ecb205513a0faaa0d37a972060cafd
3
- size 669477
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:662cae7e3aa663661825f2e5ae61701b7b27e7bbc28100931cadacd84ddad842
3
+ size 670189