add dataset_info in dataset metadata
Browse files
README.md
CHANGED
@@ -18,6 +18,179 @@ source_datasets:
|
|
18 |
task_categories:
|
19 |
- automatic-speech-recognition
|
20 |
task_ids: []
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
21 |
---
|
22 |
|
23 |
# Dataset Card for AMI Corpus
|
@@ -323,4 +496,4 @@ CC BY 4.0
|
|
323 |
### Contributions
|
324 |
|
325 |
Thanks to [@cahya-wirawan](https://github.com/cahya-wirawan) and [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset.
|
326 |
-
#### TODO
|
|
|
18 |
task_categories:
|
19 |
- automatic-speech-recognition
|
20 |
task_ids: []
|
21 |
+
dataset_info:
|
22 |
+
- config_name: microphone-single
|
23 |
+
features:
|
24 |
+
- name: word_ids
|
25 |
+
sequence: string
|
26 |
+
- name: word_start_times
|
27 |
+
sequence: float32
|
28 |
+
- name: word_end_times
|
29 |
+
sequence: float32
|
30 |
+
- name: word_speakers
|
31 |
+
sequence: string
|
32 |
+
- name: segment_ids
|
33 |
+
sequence: string
|
34 |
+
- name: segment_start_times
|
35 |
+
sequence: float32
|
36 |
+
- name: segment_end_times
|
37 |
+
sequence: float32
|
38 |
+
- name: segment_speakers
|
39 |
+
sequence: string
|
40 |
+
- name: words
|
41 |
+
sequence: string
|
42 |
+
- name: channels
|
43 |
+
sequence: string
|
44 |
+
- name: file
|
45 |
+
dtype: string
|
46 |
+
- name: audio
|
47 |
+
dtype:
|
48 |
+
audio:
|
49 |
+
sampling_rate: 16000
|
50 |
+
splits:
|
51 |
+
- name: test
|
52 |
+
num_bytes: 4821283
|
53 |
+
num_examples: 16
|
54 |
+
- name: train
|
55 |
+
num_bytes: 42013753
|
56 |
+
num_examples: 134
|
57 |
+
- name: validation
|
58 |
+
num_bytes: 5110497
|
59 |
+
num_examples: 18
|
60 |
+
download_size: 11387715153
|
61 |
+
dataset_size: 51945533
|
62 |
+
- config_name: microphone-multi
|
63 |
+
features:
|
64 |
+
- name: word_ids
|
65 |
+
sequence: string
|
66 |
+
- name: word_start_times
|
67 |
+
sequence: float32
|
68 |
+
- name: word_end_times
|
69 |
+
sequence: float32
|
70 |
+
- name: word_speakers
|
71 |
+
sequence: string
|
72 |
+
- name: segment_ids
|
73 |
+
sequence: string
|
74 |
+
- name: segment_start_times
|
75 |
+
sequence: float32
|
76 |
+
- name: segment_end_times
|
77 |
+
sequence: float32
|
78 |
+
- name: segment_speakers
|
79 |
+
sequence: string
|
80 |
+
- name: words
|
81 |
+
sequence: string
|
82 |
+
- name: channels
|
83 |
+
sequence: string
|
84 |
+
- name: file-1-1
|
85 |
+
dtype: string
|
86 |
+
- name: file-1-2
|
87 |
+
dtype: string
|
88 |
+
- name: file-1-3
|
89 |
+
dtype: string
|
90 |
+
- name: file-1-4
|
91 |
+
dtype: string
|
92 |
+
- name: file-1-5
|
93 |
+
dtype: string
|
94 |
+
- name: file-1-6
|
95 |
+
dtype: string
|
96 |
+
- name: file-1-7
|
97 |
+
dtype: string
|
98 |
+
- name: file-1-8
|
99 |
+
dtype: string
|
100 |
+
splits:
|
101 |
+
- name: test
|
102 |
+
num_bytes: 4834751
|
103 |
+
num_examples: 16
|
104 |
+
- name: train
|
105 |
+
num_bytes: 42126341
|
106 |
+
num_examples: 134
|
107 |
+
- name: validation
|
108 |
+
num_bytes: 5125645
|
109 |
+
num_examples: 18
|
110 |
+
download_size: 90941506169
|
111 |
+
dataset_size: 52086737
|
112 |
+
- config_name: headset-single
|
113 |
+
features:
|
114 |
+
- name: word_ids
|
115 |
+
sequence: string
|
116 |
+
- name: word_start_times
|
117 |
+
sequence: float32
|
118 |
+
- name: word_end_times
|
119 |
+
sequence: float32
|
120 |
+
- name: word_speakers
|
121 |
+
sequence: string
|
122 |
+
- name: segment_ids
|
123 |
+
sequence: string
|
124 |
+
- name: segment_start_times
|
125 |
+
sequence: float32
|
126 |
+
- name: segment_end_times
|
127 |
+
sequence: float32
|
128 |
+
- name: segment_speakers
|
129 |
+
sequence: string
|
130 |
+
- name: words
|
131 |
+
sequence: string
|
132 |
+
- name: channels
|
133 |
+
sequence: string
|
134 |
+
- name: file
|
135 |
+
dtype: string
|
136 |
+
- name: audio
|
137 |
+
dtype:
|
138 |
+
audio:
|
139 |
+
sampling_rate: 16000
|
140 |
+
splits:
|
141 |
+
- name: test
|
142 |
+
num_bytes: 4821283
|
143 |
+
num_examples: 16
|
144 |
+
- name: train
|
145 |
+
num_bytes: 42491091
|
146 |
+
num_examples: 136
|
147 |
+
- name: validation
|
148 |
+
num_bytes: 5110497
|
149 |
+
num_examples: 18
|
150 |
+
download_size: 11505070978
|
151 |
+
dataset_size: 52422871
|
152 |
+
- config_name: headset-multi
|
153 |
+
features:
|
154 |
+
- name: word_ids
|
155 |
+
sequence: string
|
156 |
+
- name: word_start_times
|
157 |
+
sequence: float32
|
158 |
+
- name: word_end_times
|
159 |
+
sequence: float32
|
160 |
+
- name: word_speakers
|
161 |
+
sequence: string
|
162 |
+
- name: segment_ids
|
163 |
+
sequence: string
|
164 |
+
- name: segment_start_times
|
165 |
+
sequence: float32
|
166 |
+
- name: segment_end_times
|
167 |
+
sequence: float32
|
168 |
+
- name: segment_speakers
|
169 |
+
sequence: string
|
170 |
+
- name: words
|
171 |
+
sequence: string
|
172 |
+
- name: channels
|
173 |
+
sequence: string
|
174 |
+
- name: file-0
|
175 |
+
dtype: string
|
176 |
+
- name: file-1
|
177 |
+
dtype: string
|
178 |
+
- name: file-2
|
179 |
+
dtype: string
|
180 |
+
- name: file-3
|
181 |
+
dtype: string
|
182 |
+
splits:
|
183 |
+
- name: test
|
184 |
+
num_bytes: 4827055
|
185 |
+
num_examples: 16
|
186 |
+
- name: train
|
187 |
+
num_bytes: 42540063
|
188 |
+
num_examples: 136
|
189 |
+
- name: validation
|
190 |
+
num_bytes: 5116989
|
191 |
+
num_examples: 18
|
192 |
+
download_size: 45951596391
|
193 |
+
dataset_size: 52484107
|
194 |
---
|
195 |
|
196 |
# Dataset Card for AMI Corpus
|
|
|
496 |
### Contributions
|
497 |
|
498 |
Thanks to [@cahya-wirawan](https://github.com/cahya-wirawan) and [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset.
|
499 |
+
#### TODO
|