parquet-converter commited on
Commit
f0008d4
1 Parent(s): f33c72a

Update parquet files

Browse files
README.md DELETED
@@ -1,271 +0,0 @@
1
- ---
2
- annotations_creators:
3
- - expert-generated
4
- - crowdsourced
5
- - machine-generated
6
- language:
7
- - en
8
- language_creators:
9
- - crowdsourced
10
- - expert-generated
11
- license:
12
- - cc-by-4.0
13
- - apache-2.0
14
- - cc0-1.0
15
- - cc-by-nc-3.0
16
- - other
17
- multilinguality:
18
- - monolingual
19
- pretty_name: esc-datasets
20
- size_categories:
21
- - 100K<n<1M
22
- - 1M<n<10M
23
- source_datasets:
24
- - original
25
- - extended|librispeech_asr
26
- - extended|common_voice
27
- tags:
28
- - asr
29
- - benchmark
30
- - speech
31
- - esc
32
- task_categories:
33
- - automatic-speech-recognition
34
- task_ids: []
35
- extra_gated_prompt: |-
36
- Three of the ESC datasets have specific terms of usage that must be agreed to before using the data.
37
- To do so, fill in the access forms on the specific datasets' pages:
38
- * Common Voice: https://huggingface.co/datasets/mozilla-foundation/common_voice_9_0
39
- * GigaSpeech: https://huggingface.co/datasets/speechcolab/gigaspeech
40
- * SPGISpeech: https://huggingface.co/datasets/kensho/spgispeech
41
- extra_gated_fields:
42
- I hereby confirm that I have registered on the original Common Voice page and agree to not attempt to determine the identity of speakers in the Common Voice dataset: checkbox
43
- I hereby confirm that I have accepted the terms of usages on GigaSpeech page: checkbox
44
- I hereby confirm that I have accepted the terms of usages on SPGISpeech page: checkbox
45
- ---
46
-
47
- All eight of datasets in ESC can be downloaded and prepared in just a single line of code through the Hugging Face Datasets library:
48
-
49
- ```python
50
- from datasets import load_dataset
51
-
52
- librispeech = load_dataset("esc-benchmark/esc-datasets", "librispeech", split="train")
53
- ```
54
-
55
- - `"esc-benchmark"`: the repository namespace. This is fixed for all ESC datasets.
56
-
57
- - `"librispeech"`: the dataset name. This can be changed to any of any one of the eight datasets in ESC to download that dataset.
58
-
59
- - `split="train"`: the split. Set this to one of train/validation/test to generate a specific split. Omit the `split` argument to generate all splits for a dataset.
60
-
61
-
62
- The datasets are full prepared, such that the audio and transcription files can be used directly in training/evaluation scripts.
63
-
64
-
65
- ## Dataset Information
66
-
67
- A data point can be accessed by indexing the dataset object loaded through `load_dataset`:
68
-
69
- ```python
70
- print(librispeech[0])
71
- ```
72
-
73
- A typical data point comprises the path to the audio file and its transcription. Also included is information of the dataset from which the sample derives and a unique identifier name:
74
-
75
- ```python
76
- {
77
- 'dataset': 'librispeech',
78
- 'audio': {'path': '/home/esc-bencher/.cache/huggingface/datasets/downloads/extracted/d2da1969fe9e7d06661b5dc370cf2e3c119a14c35950045bcb76243b264e4f01/374-180298-0000.flac',
79
- 'array': array([ 7.01904297e-04, 7.32421875e-04, 7.32421875e-04, ...,
80
- -2.74658203e-04, -1.83105469e-04, -3.05175781e-05]),
81
- 'sampling_rate': 16000},
82
- 'text': 'chapter sixteen i might have told you of the beginning of this liaison in a few lines but i wanted you to see every step by which we came i to agree to whatever marguerite wished',
83
- 'id': '374-180298-0000'
84
- }
85
- ```
86
-
87
- ### Data Fields
88
-
89
- - `dataset`: name of the ESC dataset from which the sample is taken.
90
-
91
- - `audio`: a dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate.
92
-
93
- - `text`: the transcription of the audio file.
94
-
95
- - `id`: unique id of the data sample.
96
-
97
- ### Data Preparation
98
- #### Audio
99
- The audio for all ESC datasets is segmented into sample lengths suitable for training ASR systems. The Hugging Face datasets library decodes audio files on the fly, reading the segments and converting them to a Python arrays. Consequently, no further preparation of the audio is required to be used in training/evaluation scripts.
100
-
101
- Note that when accessing the audio column: `dataset[0]["audio"]` the audio file is automatically decoded and resampled to `dataset.features["audio"].sampling_rate`. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the `"audio"` column, i.e. `dataset[0]["audio"]` should always be preferred over `dataset["audio"][0]`.
102
-
103
- #### Transcriptions
104
- The transcriptions corresponding to each audio file are provided in their 'error corrected' format. No transcription pre-processing is applied to the text, only necessary 'error correction' steps such as removing junk tokens (_&lt;unk>_) or converting symbolic punctuation to spelled out form (_&lt;comma>_ to _,_). As such, no further preparation of the transcriptions is required to be used in training/evaluation scripts.
105
-
106
- Transcriptions are provided for training and validation splits. The transcriptions are **not** provided for the test splits. The ESC benchmark requires you to generate predictions for the test sets and upload them to https://huggingface.co/spaces/esc-benchmark/esc for scoring.
107
-
108
- ### Access
109
- All eight of the datasets in ESC are accessible and licensing is freely available. Three of the ESC datasets have specific terms of usage that must be agreed to before using the data. To do so, fill in the access forms on the specific datasets' pages:
110
- * Common Voice: https://huggingface.co/datasets/mozilla-foundation/common_voice_9_0
111
- * GigaSpeech: https://huggingface.co/datasets/speechcolab/gigaspeech
112
- * SPGISpeech: https://huggingface.co/datasets/kensho/spgispeech
113
-
114
- ## LibriSpeech
115
-
116
- The LibriSpeech corpus is a standard large-scale corpus for assessing ASR systems. It consists of approximately 1,000 hours of narrated audiobooks from the [LibriVox](https://librivox.org) project. It is licensed under CC-BY-4.0.
117
-
118
- Example Usage:
119
-
120
- ```python
121
- librispeech = load_dataset("esc-benchmark/esc-datasets", "librispeech")
122
- ```
123
-
124
- Train/validation splits:
125
- - `train` (combination of `train.clean.100`, `train.clean.360` and `train.other.500`)
126
- - `validation.clean`
127
- - `validation.other`
128
-
129
- Test splits:
130
- - `test.clean`
131
- - `test.other`
132
-
133
- Also available are subsets of the train split, which can be accessed by setting the `subconfig` argument:
134
- ```python
135
- librispeech = load_dataset("esc-benchmark/esc-datasets", "librispeech", subconfig="clean.100")
136
- ```
137
-
138
- - `clean.100`: 100 hours of training data from the 'clean' subset
139
- - `clean.360`: 360 hours of training data from the 'clean' subset
140
- - `other.500`: 500 hours of training data from the 'other' subset
141
-
142
- ## Common Voice
143
- Common Voice is a series of crowd-sourced open-licensed speech datasets where speakers record text from Wikipedia in various languages. The English subset of contains approximately 1,400 hours of audio data from speakers of various nationalities, accents and different recording conditions. It is licensed under CC0-1.0.
144
-
145
- Example usage:
146
-
147
- ```python
148
- common_voice = load_dataset("esc-benchmark/esc-datasets", "common_voice", use_auth_token=True)
149
- ```
150
-
151
- Training/validation splits:
152
- - `train`
153
- - `validation`
154
-
155
- Test splits:
156
- - `test`
157
-
158
- ## VoxPopuli
159
- VoxPopuli s a large-scale multilingual speech corpus consisting of political data sourced from 2009-2020 European Parliament event recordings. The English subset contains approximately 550 hours of speech largely from non-native English speakers. It is licensed under CC0.
160
-
161
- Example usage:
162
-
163
- ```python
164
- voxpopuli = load_dataset("esc-benchmark/esc-datasets", "voxpopuli")
165
- ```
166
-
167
- Training/validation splits:
168
- - `train`
169
- - `validation`
170
-
171
- Test splits:
172
- - `test`
173
-
174
- ## TED-LIUM
175
- TED-LIUM consists of English-language TED Talk conference videos covering a range of different cultural, political, and academic topics. It contains approximately 450 hours of transcribed speech data. It is licensed under CC-BY-NC-ND 3.0.
176
-
177
- Example usage:
178
-
179
- ```python
180
- tedlium = load_dataset("esc-benchmark/esc-datasets", "tedlium")
181
- ```
182
-
183
- Training/validation splits:
184
- - `train`
185
- - `validation`
186
-
187
- Test splits:
188
- - `test`
189
-
190
- ## GigaSpeech
191
- GigaSpeech is a multi-domain English speech recognition corpus created from audiobooks, podcasts and YouTube. We provide the large train set (2,500 hours) and the standard validation and test splits. It is licensed under apache-2.0.
192
-
193
- Example usage:
194
-
195
- ```python
196
- gigaspeech = load_dataset("esc-benchmark/esc-datasets", "gigaspeech", use_auth_token=True)
197
- ```
198
-
199
- Training/validation splits:
200
- - `train` (`l` subset of training data (2,500 h))
201
- - `validation`
202
-
203
- Test splits:
204
- - `test`
205
-
206
- Also available are subsets of the train split, which can be accessed by setting the `subconfig` argument:
207
- ```python
208
- gigaspeech = load_dataset("esc-benchmark/esc-datasets", "spgispeech", subconfig="xs", use_auth_token=True)
209
- ```
210
- - `xs`: extra-small subset of training data (10 h)
211
- - `s`: small subset of training data (250 h)
212
- - `m`: medium subset of training data (1,000 h)
213
- - `xl`: extra-large subset of training data (10,000 h)
214
-
215
- ## SPGISpeech
216
- SPGISpeech consists of company earnings calls that have been manually transcribed by S&P Global, Inc according to a professional style guide. We provide the large train set (5,000 hours) and the standard validation and test splits. It is licensed under a Kensho user agreement.
217
-
218
- Loading the dataset requires authorization.
219
-
220
- Example usage:
221
-
222
- ```python
223
- spgispeech = load_dataset("esc-benchmark/esc-datasets", "spgispeech", use_auth_token=True)
224
- ```
225
-
226
- Training/validation splits:
227
- - `train` (`l` subset of training data (~5,000 h))
228
- - `validation`
229
-
230
- Test splits:
231
- - `test`
232
-
233
- Also available are subsets of the train split, which can be accessed by setting the `subconfig` argument:
234
- ```python
235
- spgispeech = load_dataset("esc-benchmark/esc-datasets", "spgispeech", subconfig="s", use_auth_token=True)
236
- ```
237
- - `s`: small subset of training data (~200 h)
238
- - `m`: medium subset of training data (~1,000 h)
239
-
240
-
241
- ## Earnings-22
242
- Earnings-22 is a 119-hour corpus of English-language earnings calls collected from global companies, with speakers of many different nationalities and accents. It is licensed under CC-BY-SA-4.0.
243
-
244
- Example usage:
245
-
246
- ```python
247
- earnings22 = load_dataset("esc-benchmark/esc-datasets", "earnings22")
248
- ```
249
-
250
- Training/validation splits:
251
- - `train`
252
- - `validation`
253
-
254
- Test splits:
255
- - `test`
256
-
257
- ## AMI
258
- The AMI Meeting Corpus consists of 100 hours of meeting recordings from multiple recording devices synced to a common timeline. It is licensed under CC-BY-4.0.
259
-
260
- Example usage:
261
-
262
- ```python
263
- ami = load_dataset("esc-benchmark/esc-datasets", "ami")
264
- ```
265
-
266
- Training/validation splits:
267
- - `train`
268
- - `validation`
269
-
270
- Test splits:
271
- - `test`
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cv_release_stats.py DELETED
@@ -1 +0,0 @@
1
- STATS = {'bundleURLTemplate': 'cv-corpus-9.0-2022-04-27/cv-corpus-9.0-2022-04-27-{locale}.tar.gz', 'date': '2022-04-27', 'name': 'Common Voice Corpus 9.0', 'multilingual': True, 'locales': {'en': {'duration': 10632910099, 'buckets': {'dev': 16335, 'invalidated': 242945, 'other': 267228, 'reported': 3903, 'test': 16335, 'train': 890116, 'validated': 1556254}, 'reportedSentences': 3842, 'clips': 2066427, 'splits': {'accent': {'': 1}, 'age': {'': 0.37, 'twenties': 0.24, 'sixties': 0.04, 'thirties': 0.13, 'teens': 0.06, 'seventies': 0.01, 'fourties': 0.1, 'fifties': 0.05, 'eighties': 0, 'nineties': 0}, 'gender': {'': 0.37, 'male': 0.46, 'female': 0.15, 'other': 0.02}}, 'users': 81085, 'size': 76768058135, 'checksum': '7c789a325e1d9f16137c1aeab9f45a1fde049a39b8fbf49e9ef4a8827ae67eae', 'avgDurationSecs': 5.146, 'validDurationSecs': 8007787.777, 'totalHrs': 2953.58, 'validHrs': 2224.38}, 'eu': {'reportedSentences': 44, 'buckets': {'dev': 6558, 'invalidated': 5785, 'other': 26668, 'reported': 43, 'test': 6558, 'train': 10825, 'validated': 69111}, 'duration': 527089551, 'clips': 101564, 'splits': {'accent': {'': 1}, 'age': {'fourties': 0.13, 'thirties': 0.07, 'fifties': 0.14, 'twenties': 0.35, '': 0.25, 'teens': 0.03, 'sixties': 0.02, 'seventies': 0}, 'gender': {'male': 0.47, 'female': 0.26, '': 0.25, 'other': 0.02}}, 'users': 1201, 'size': 3975713611, 'checksum': 'd28b239b95019469c3fa3a64587906efac9e152f2cb4a19d3175d9fa7f4f2e18', 'avgDurationSecs': 5.19, 'validDurationSecs': 358667.303, 'totalHrs': 146.41, 'validHrs': 99.62}, 'tr': {'reportedSentences': 306, 'buckets': {'dev': 8313, 'invalidated': 3262, 'other': 151, 'reported': 305, 'test': 8362, 'train': 16862, 'validated': 66078}, 'duration': 256673727, 'clips': 69491, 'splits': {'accent': {'': 1}, 'age': {'': 0.33, 'thirties': 0.1, 'twenties': 0.28, 'teens': 0.02, 'fourties': 0.04, 'fifties': 0.09, 'sixties': 0.09, 'eighties': 0, 'seventies': 0.04}, 'gender': {'': 0.33, 'male': 0.46, 'female': 0.2, 'other': 0}}, 'users': 1264, 'size': 1608944247, 'checksum': '7064e9f756fb93b4d0c1b0dae619280d3470c5e95cb7e5406953ac4888dbd8d7', 'avgDurationSecs': 3.694, 'validDurationSecs': 244067.383, 'totalHrs': 71.29, 'validHrs': 67.79}, 'ar': {'reportedSentences': 2033, 'buckets': {'dev': 10220, 'invalidated': 14873, 'other': 30302, 'reported': 2040, 'test': 10421, 'train': 28139, 'validated': 75322}, 'duration': 502349703, 'clips': 120497, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.12, '': 0.58, 'twenties': 0.26, 'fourties': 0.01, 'teens': 0.03, 'fifties': 0, 'sixties': 0, 'nineties': 0}, 'gender': {'female': 0.17, '': 0.57, 'male': 0.25, 'other': 0}}, 'users': 1237, 'size': 2995697303, 'checksum': 'd29c81a1801ba01af97b7a90548dfd0409b9efa9b9d1291e28deee90267a746b', 'avgDurationSecs': 4.169, 'validDurationSecs': 314017, 'totalHrs': 139.54, 'validHrs': 87.23}, 'zh-TW': {'reportedSentences': 139, 'buckets': {'dev': 4618, 'invalidated': 4482, 'other': 37593, 'reported': 138, 'test': 4618, 'train': 6393, 'validated': 74894}, 'duration': 380039242, 'clips': 116969, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.2, 'twenties': 0.33, 'teens': 0.05, '': 0.28, 'fifties': 0.04, 'seventies': 0, 'fourties': 0.1, 'sixties': 0}, 'gender': {'male': 0.46, '': 0.27, 'female': 0.25, 'other': 0.02}}, 'users': 2002, 'size': 2719746719, 'checksum': 'b40023ca8f5113a3766ba2d0844c3e846ff48eb1d4408d9ae912d4f9caa183eb', 'avgDurationSecs': 3.249, 'validDurationSecs': 243335.063, 'totalHrs': 105.56, 'validHrs': 67.59}, 'br': {'duration': 71324827, 'reportedSentences': 183, 'buckets': {'dev': 2161, 'invalidated': 756, 'other': 11283, 'reported': 182, 'test': 2150, 'train': 2561, 'validated': 11152}, 'clips': 23191, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.25, '': 0.33, 'fifties': 0.06, 'fourties': 0.07, 'thirties': 0.08, 'sixties': 0.18, 'seventies': 0.02, 'teens': 0.01}, 'gender': {'male': 0.65, '': 0.34, 'female': 0.02}}, 'users': 172, 'size': 525880271, 'checksum': '7a49666d12df795d95c5d64e827970e40863f26dfb87ceae61d14819788e6cfa', 'avgDurationSecs': 3.076, 'validDurationSecs': 34298.412, 'totalHrs': 19.81, 'validHrs': 9.52}, 'pt': {'duration': 484853047, 'reportedSentences': 2305, 'buckets': {'dev': 8528, 'invalidated': 4416, 'other': 11900, 'reported': 2310, 'test': 8539, 'train': 17559, 'validated': 99173}, 'clips': 115489, 'splits': {'accent': {'': 1}, 'age': {'': 0.2, 'twenties': 0.42, 'teens': 0.03, 'thirties': 0.22, 'fourties': 0.1, 'sixties': 0, 'fifties': 0.03, 'seventies': 0}, 'gender': {'': 0.2, 'male': 0.74, 'female': 0.04, 'other': 0.02}}, 'users': 2453, 'size': 3187978215, 'checksum': '1a748f21ba194d57ecd6418cce9692a198eeabcb44ba6e2b7a34435efc6625a4', 'avgDurationSecs': 4.198, 'validDurationSecs': 416354.209, 'totalHrs': 134.68, 'validHrs': 115.65}, 'eo': {'duration': 6712729612, 'reportedSentences': 2113, 'buckets': {'dev': 14911, 'invalidated': 127285, 'other': 130550, 'reported': 2115, 'test': 14915, 'train': 143973, 'validated': 848434}, 'clips': 1106269, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.57, 'thirties': 0.12, '': 0.19, 'fourties': 0.04, 'fifties': 0.02, 'seventies': 0, 'teens': 0.05, 'sixties': 0, 'eighties': 0}, 'gender': {'male': 0.69, '': 0.2, 'female': 0.11, 'other': 0}}, 'users': 1498, 'size': 40086922305, 'checksum': '5b341a961cf9f680d5c97d0be4f817fedbdfe695a3cf18825363c4c9104899b7', 'avgDurationSecs': 6.068, 'validDurationSecs': 5148297, 'totalHrs': 1864.65, 'validHrs': 1430.08}, 'zh-CN': {'reportedSentences': 445, 'buckets': {'dev': 9740, 'invalidated': 6689, 'other': 14066, 'reported': 458, 'test': 9747, 'train': 23539, 'validated': 48516}, 'duration': 358872900, 'clips': 69271, 'splits': {'accent': {'': 1}, 'age': {'': 0.39, 'teens': 0.09, 'twenties': 0.38, 'thirties': 0.11, 'fourties': 0.03, 'nineties': 0, 'fifties': 0, 'sixties': 0}, 'gender': {'': 0.39, 'male': 0.51, 'female': 0.1, 'other': 0.01}}, 'users': 4233, 'size': 2637465454, 'checksum': '36faeffbdbbdeab6a8994ceddc3df326503777754141f9135c24fd87d1a301a0', 'avgDurationSecs': 5.181, 'validDurationSecs': 251347.283, 'totalHrs': 99.68, 'validHrs': 69.81}, 'id': {'reportedSentences': 261, 'buckets': {'dev': 3218, 'invalidated': 2454, 'other': 22787, 'reported': 260, 'test': 3622, 'train': 5043, 'validated': 23132}, 'duration': 196639788, 'clips': 48373, 'splits': {'accent': {'': 1}, 'age': {'': 0.26, 'twenties': 0.39, 'thirties': 0.07, 'teens': 0.26, 'fifties': 0, 'fourties': 0.02}, 'gender': {'': 0.26, 'male': 0.41, 'female': 0.29, 'other': 0.04}}, 'users': 416, 'size': 1253048208, 'checksum': '874e959e2ca1aacc502ff969a3e54de792dd41e4f672ae1fd9d38213f4bf4139', 'avgDurationSecs': 4.065, 'validDurationSecs': 94033.274, 'totalHrs': 54.62, 'validHrs': 26.12}, 'ia': {'reportedSentences': 259, 'buckets': {'dev': 1790, 'invalidated': 328, 'other': 2707, 'reported': 263, 'test': 1722, 'train': 5056, 'validated': 11338}, 'duration': 60049596, 'clips': 14373, 'splits': {'accent': {'': 1}, 'age': {'seventies': 0.22, 'fourties': 0.3, '': 0.39, 'twenties': 0.05, 'thirties': 0.02, 'teens': 0, 'fifties': 0.03, 'sixties': 0}, 'gender': {'male': 0.61, '': 0.39, 'female': 0.01}}, 'users': 59, 'size': 407391806, 'checksum': '520a355dd5a0d8980cc5fc22d07e95db39eda4967d13d32aef77fd306a0c6979', 'avgDurationSecs': 4.178, 'validDurationSecs': 47369.535, 'totalHrs': 16.68, 'validHrs': 13.15}, 'lv': {'duration': 30488525, 'reportedSentences': 26, 'buckets': {'dev': 1841, 'invalidated': 166, 'other': 1204, 'reported': 25, 'test': 2152, 'train': 3123, 'validated': 7570}, 'clips': 8940, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.49, 'fourties': 0.03, '': 0.17, 'twenties': 0.28, 'teens': 0.03, 'fifties': 0}, 'gender': {'male': 0.7, 'female': 0.13, '': 0.17}}, 'users': 115, 'size': 224727245, 'checksum': '9b967d04e672fe34fd7b7d3356fd2224be03b8d25d516f962b66ce66f2be5ca6', 'avgDurationSecs': 3.41, 'validDurationSecs': 25816.346, 'totalHrs': 8.46, 'validHrs': 7.17}, 'ja': {'duration': 177935145, 'reportedSentences': 129, 'buckets': {'dev': 4445, 'invalidated': 2195, 'other': 310, 'reported': 129, 'test': 4495, 'train': 6212, 'validated': 34715}, 'clips': 37220, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.32, '': 0.24, 'teens': 0.04, 'fifties': 0.01, 'thirties': 0.1, 'fourties': 0.28, 'sixties': 0, 'seventies': 0}, 'gender': {'male': 0.52, '': 0.21, 'female': 0.26, 'other': 0}}, 'users': 619, 'size': 1081598283, 'checksum': '6d142a6969698a8505fa32e5fd4636e502116ded6d0e6b49e66340480cb70150', 'avgDurationSecs': 4.781, 'validDurationSecs': 165959.66, 'totalHrs': 49.42, 'validHrs': 46.09}, 'rw': {'duration': 8580385049, 'reportedSentences': 624, 'buckets': {'dev': 15988, 'invalidated': 227719, 'other': 47336, 'reported': 623, 'test': 16213, 'train': 1003811, 'validated': 1438343}, 'clips': 1713398, 'splits': {'accent': {'': 1}, 'age': {'': 0.05, 'twenties': 0.61, 'thirties': 0.12, 'teens': 0.2, 'fourties': 0.02, 'fifties': 0}, 'gender': {'': 0.1, 'male': 0.57, 'female': 0.33, 'other': 0}}, 'users': 1068, 'size': 60996878145, 'checksum': '9bfafd887f7a2a73ba0f819e932cf2836b8dd336ff0c0cd4999b904dda08df87', 'avgDurationSecs': 5.008, 'validDurationSecs': 7202959.717, 'totalHrs': 2383.44, 'validHrs': 2000.82}, 'sv-SE': {'reportedSentences': 571, 'buckets': {'dev': 5005, 'invalidated': 1335, 'other': 5357, 'reported': 570, 'test': 5045, 'train': 7302, 'validated': 38398}, 'duration': 177932535, 'clips': 45090, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.25, '': 0.18, 'teens': 0.03, 'fifties': 0.03, 'twenties': 0.12, 'fourties': 0.38, 'sixties': 0, 'seventies': 0}, 'gender': {'male': 0.47, '': 0.19, 'female': 0.34, 'other': 0}}, 'users': 738, 'size': 1131052430, 'checksum': '064c1aace5a33e08d035568573fc4f36426c3e7110d6a8482124ceca1ebc6c66', 'avgDurationSecs': 3.946, 'validDurationSecs': 151524.805, 'totalHrs': 49.42, 'validHrs': 42.09}, 'cnh': {'duration': 20673132, 'reportedSentences': 9, 'buckets': {'dev': 761, 'invalidated': 436, 'other': 2907, 'reported': 8, 'test': 763, 'train': 817, 'validated': 2458}, 'clips': 5801, 'splits': {'accent': {'': 1}, 'age': {'': 0.51, 'twenties': 0.36, 'fourties': 0.01, 'teens': 0.02, 'thirties': 0.08, 'fifties': 0.02}, 'gender': {'': 0.51, 'male': 0.33, 'female': 0.16}}, 'users': 298, 'size': 161375852, 'checksum': '2ee28e4c1e5c953b1ed85e42f80f96d71f5185a78a29123bdbaf2aed4a3685f0', 'avgDurationSecs': 3.564, 'validDurationSecs': 8759.62, 'totalHrs': 5.74, 'validHrs': 2.43}, 'et': {'duration': 177339854, 'reportedSentences': 461, 'buckets': {'dev': 2636, 'invalidated': 6094, 'other': 129, 'reported': 464, 'test': 2636, 'train': 3135, 'validated': 20013}, 'clips': 26236, 'splits': {'accent': {'': 1}, 'age': {'': 0.2, 'thirties': 0.08, 'twenties': 0.68, 'fourties': 0.04, 'fifties': 0, 'seventies': 0, 'teens': 0}, 'gender': {'': 0.2, 'male': 0.55, 'female': 0.24, 'other': 0}}, 'users': 793, 'size': 1227039248, 'checksum': '0d90ac5db32f4cdc1d228a7f28d1aded44fdd2155c6770bba177ddadb6290caa', 'avgDurationSecs': 6.759, 'validDurationSecs': 135276.052, 'totalHrs': 49.26, 'validHrs': 37.57}, 'ky': {'duration': 161279904, 'reportedSentences': 35, 'buckets': {'dev': 1613, 'invalidated': 5582, 'other': 262, 'reported': 34, 'test': 1613, 'train': 1787, 'validated': 29662}, 'clips': 35506, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.08, '': 0.07, 'fourties': 0.01, 'twenties': 0.66, 'teens': 0.18}, 'gender': {'male': 0.53, '': 0.11, 'female': 0.35, 'other': 0}}, 'users': 239, 'size': 1042617186, 'checksum': 'aa1f7ea3fcf417af988009de613a4185216793858f606c80f8cb4622a984a5a7', 'avgDurationSecs': 4.542, 'validDurationSecs': 134734.538, 'totalHrs': 44.79, 'validHrs': 37.42}, 'ro': {'duration': 134129767, 'reportedSentences': 298, 'buckets': {'dev': 3801, 'invalidated': 855, 'other': 18353, 'reported': 297, 'test': 3827, 'train': 5076, 'validated': 14653}, 'clips': 33861, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.14, 'teens': 0.02, '': 0.11, 'fourties': 0.06, 'sixties': 0, 'twenties': 0.66, 'fifties': 0.01, 'eighties': 0}, 'gender': {'male': 0.73, '': 0.11, 'female': 0.16, 'other': 0.01}}, 'users': 346, 'size': 844195813, 'checksum': '69ee370178e60bd581c38805b46abc5a9c0687fe5bd89b0b983181cb5fa931d1', 'avgDurationSecs': 3.961, 'validDurationSecs': 58043.279, 'totalHrs': 37.25, 'validHrs': 16.12}, 'hsb': {'duration': 10103328, 'reportedSentences': 57, 'buckets': {'dev': 172, 'invalidated': 228, 'other': 17, 'reported': 56, 'test': 431, 'train': 808, 'validated': 1411}, 'clips': 1656, 'splits': {'accent': {'': 1}, 'age': {'fourties': 0.55, '': 0.17, 'thirties': 0.11, 'sixties': 0, 'seventies': 0.03, 'twenties': 0.11, 'fifties': 0.03}, 'gender': {'male': 0.83, '': 0.17, 'other': 0}}, 'users': 19, 'size': 79356270, 'checksum': '8b82e5155d6f1d1b5bbf832db01212956b96fd3540b65f6cc564382b0478232c', 'avgDurationSecs': 6.101, 'validDurationSecs': 8608.572, 'totalHrs': 2.8, 'validHrs': 2.39}, 'el': {'duration': 95865010, 'reportedSentences': 65, 'buckets': {'dev': 1695, 'invalidated': 758, 'other': 8255, 'reported': 64, 'test': 1686, 'train': 1926, 'validated': 14223}, 'clips': 23236, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.4, 'fourties': 0.14, '': 0.28, 'twenties': 0.14, 'fifties': 0.03, 'teens': 0.01, 'sixties': 0}, 'gender': {'male': 0.67, '': 0.28, 'other': 0.02, 'female': 0.03}}, 'users': 325, 'size': 652538106, 'checksum': '2265f0030dbdf412d04fbe0c866f4417d804f636d2b7859ea0297f17d264223d', 'avgDurationSecs': 4.126, 'validDurationSecs': 58679.981, 'totalHrs': 26.62, 'validHrs': 16.29}, 'cs': {'duration': 250513527, 'reportedSentences': 693, 'buckets': {'dev': 7051, 'invalidated': 1231, 'other': 9110, 'reported': 696, 'test': 7536, 'train': 14289, 'validated': 47484}, 'clips': 57825, 'splits': {'accent': {'': 1}, 'age': {'fourties': 0.2, '': 0.36, 'thirties': 0.14, 'teens': 0.01, 'twenties': 0.27, 'fifties': 0.02, 'sixties': 0, 'seventies': 0}, 'gender': {'male': 0.62, '': 0.35, 'female': 0.03}}, 'users': 534, 'size': 1768374191, 'checksum': 'addd26310d181309a55dbc22f00ad3bfabcbbcbddc59d7661a0c5d9af6a6aa99', 'avgDurationSecs': 4.332, 'validDurationSecs': 205713.52, 'totalHrs': 69.58, 'validHrs': 57.14}, 'pl': {'duration': 593355386, 'reportedSentences': 528, 'buckets': {'dev': 8155, 'invalidated': 6004, 'other': 7115, 'reported': 528, 'test': 8155, 'train': 15997, 'validated': 119647}, 'clips': 132766, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.28, '': 0.24, 'teens': 0.02, 'thirties': 0.33, 'fourties': 0.12, 'fifties': 0.01, 'nineties': 0.01, 'sixties': 0}, 'gender': {'male': 0.6, '': 0.25, 'female': 0.14, 'other': 0.01}}, 'users': 3062, 'size': 4272080136, 'checksum': '0fe371030664d372c821312fa3baff27d90e553d4553a2f6b1d1566e4e56e8c3', 'avgDurationSecs': 4.469, 'validDurationSecs': 534724.19, 'totalHrs': 164.82, 'validHrs': 148.53}, 'rm-sursilv': {'duration': 38420393, 'reportedSentences': 10, 'buckets': {'dev': 1334, 'invalidated': 670, 'other': 2160, 'reported': 9, 'test': 1315, 'train': 1526, 'validated': 4176}, 'clips': 7006, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.03, 'twenties': 0.1, '': 0.64, 'teens': 0.06, 'fourties': 0.18}, 'gender': {'male': 0.17, 'female': 0.19, '': 0.64, 'other': 0}}, 'users': 85, 'size': 290857567, 'checksum': '67a9e201cb078d43832af37bdd3a2def823aecb8c74fc5f4045ff2f0e633ed7b', 'avgDurationSecs': 5.484, 'validDurationSecs': 22900.879, 'totalHrs': 10.67, 'validHrs': 6.36}, 'rm-vallader': {'duration': 15074402, 'reportedSentences': 29, 'buckets': {'dev': 375, 'invalidated': 391, 'other': 724, 'reported': 30, 'test': 435, 'train': 664, 'validated': 1481}, 'clips': 2596, 'splits': {'accent': {'': 1}, 'age': {'': 0.36, 'fourties': 0.41, 'twenties': 0.14, 'thirties': 0.06, 'fifties': 0, 'sixties': 0.03}, 'gender': {'': 0.36, 'male': 0.44, 'female': 0.19, 'other': 0.01}}, 'users': 51, 'size': 115232789, 'checksum': '0e05b05e19d8a6c7df38840cf9af511a15d105d2455ae0323f542a13d1bc51a7', 'avgDurationSecs': 5.807, 'validDurationSecs': 8599.842, 'totalHrs': 4.18, 'validHrs': 2.38}, 'mn': {'duration': 67682312, 'reportedSentences': 19, 'buckets': {'dev': 1852, 'invalidated': 742, 'other': 3411, 'reported': 18, 'test': 1867, 'train': 2177, 'validated': 8208}, 'clips': 12361, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.23, '': 0.27, 'twenties': 0.42, 'fourties': 0.01, 'teens': 0.02, 'nineties': 0.06, 'fifties': 0}, 'gender': {'male': 0.36, '': 0.27, 'female': 0.31, 'other': 0.06}}, 'users': 465, 'size': 516172228, 'checksum': 'ce3e8ce228c263b937d063ca7f77d1fd200139b0685d84f9947852c1c774a2ca', 'avgDurationSecs': 5.475, 'validDurationSecs': 44942.676, 'totalHrs': 18.8, 'validHrs': 12.48}, 'zh-HK': {'duration': 451153356, 'reportedSentences': 615, 'buckets': {'dev': 5578, 'invalidated': 4045, 'other': 15366, 'reported': 626, 'test': 5578, 'train': 8392, 'validated': 87390}, 'clips': 106801, 'splits': {'accent': {'': 1}, 'age': {'fourties': 0.13, 'thirties': 0.11, '': 0.41, 'teens': 0.02, 'fifties': 0.03, 'seventies': 0, 'sixties': 0, 'twenties': 0.31}, 'gender': {'male': 0.41, '': 0.37, 'female': 0.22, 'other': 0.01}}, 'users': 2869, 'size': 3287174684, 'checksum': '02132780a56946a6069129e8c15541616de2ebe789e58599215028dc604bb618', 'avgDurationSecs': 4.224, 'validDurationSecs': 369156.579, 'totalHrs': 125.32, 'validHrs': 102.54}, 'ab': {'duration': 301136940, 'reportedSentences': 218, 'buckets': {'dev': 9151, 'invalidated': 5268, 'other': 11583, 'reported': 219, 'test': 9115, 'train': 21027, 'validated': 41903}, 'clips': 58754, 'splits': {'accent': {'': 1}, 'age': {'seventies': 0.01, 'thirties': 0.13, '': 0.19, 'teens': 0.28, 'twenties': 0.19, 'fifties': 0.06, 'sixties': 0.05, 'fourties': 0.09, 'eighties': 0.01}, 'gender': {'male': 0.18, 'female': 0.64, '': 0.18}}, 'users': 397, 'size': 1720010158, 'checksum': 'bdd127169cd25a2c40d27e04a124b98a9962bcc70fdf31ed6877ba4050c8c1b6', 'avgDurationSecs': 5.125, 'validDurationSecs': 214769.057, 'totalHrs': 83.64, 'validHrs': 59.65}, 'cv': {'duration': 84820368, 'reportedSentences': 138, 'buckets': {'dev': 1282, 'invalidated': 1889, 'other': 1261, 'reported': 142, 'test': 1266, 'train': 1416, 'validated': 13793}, 'clips': 16943, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.47, '': 0.22, 'fourties': 0.07, 'thirties': 0.01, 'teens': 0.22, 'fifties': 0.01}, 'gender': {'male': 0.51, '': 0.19, 'female': 0.29}}, 'users': 102, 'size': 590791689, 'checksum': 'd98d872e530e3142105ae3c8e014c834f51bacce64d3bd5b40a5644933a90c40', 'avgDurationSecs': 5.006, 'validDurationSecs': 69050.778, 'totalHrs': 23.56, 'validHrs': 19.18}, 'uk': {'duration': 287407536, 'reportedSentences': 579, 'buckets': {'dev': 6316, 'invalidated': 2238, 'other': 7740, 'reported': 578, 'test': 6320, 'train': 10207, 'validated': 49567}, 'clips': 59545, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.27, 'teens': 0.07, '': 0.27, 'fourties': 0.13, 'thirties': 0.25, 'fifties': 0, 'sixties': 0}, 'gender': {'male': 0.56, 'female': 0.16, '': 0.27}}, 'users': 696, 'size': 1971028073, 'checksum': 'f806bb2600f1aa2b4314593854fe38fc3f36bab1308d0a27ec1db0eef7649b0c', 'avgDurationSecs': 4.827, 'validDurationSecs': 239246.441, 'totalHrs': 79.83, 'validHrs': 66.45}, 'mt': {'duration': 61045812, 'reportedSentences': 8, 'buckets': {'dev': 1596, 'invalidated': 319, 'other': 6237, 'reported': 7, 'test': 1627, 'train': 1951, 'validated': 6331}, 'clips': 12887, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.19, '': 0.26, 'fourties': 0.17, 'thirties': 0.09, 'teens': 0.03, 'fifties': 0.26, 'sixties': 0.01}, 'gender': {'male': 0.25, '': 0.26, 'female': 0.48, 'other': 0.01}}, 'users': 204, 'size': 454454984, 'checksum': 'cf1045f68501fecb4977d198909b2b09a9e11eed24796baff4464ca4105e8d3c', 'avgDurationSecs': 4.737, 'validDurationSecs': 29989.993, 'totalHrs': 16.95, 'validHrs': 8.33}, 'as': {'duration': 5201165, 'reportedSentences': 10, 'buckets': {'dev': 116, 'invalidated': 47, 'other': 10, 'reported': 9, 'test': 299, 'train': 508, 'validated': 923}, 'clips': 980, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.4, '': 0.55, 'thirties': 0.05, 'teens': 0.01}, 'gender': {'male': 0.46, '': 0.53, 'female': 0}}, 'users': 41, 'size': 35835442, 'checksum': '4a4710429af3e14f4f7d4dd58b8a2c18164bdf1538b598d6385f09bdccef8d3a', 'avgDurationSecs': 5.307, 'validDurationSecs': 4898.648, 'totalHrs': 1.44, 'validHrs': 1.36}, 'ka': {'duration': 29727684, 'reportedSentences': 37, 'buckets': {'dev': 1353, 'invalidated': 362, 'other': 0, 'reported': 36, 'test': 1363, 'train': 1686, 'validated': 5207}, 'clips': 5569, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.38, 'thirties': 0.24, '': 0.34, 'fourties': 0.02, 'fifties': 0, 'teens': 0.01}, 'gender': {'male': 0.46, 'female': 0.2, '': 0.35}}, 'users': 132, 'size': 196283323, 'checksum': 'f499863ec4f7697cfb15f46730b630bc2123e8e044950c30b41912d1e897aebb', 'avgDurationSecs': 5.338, 'validDurationSecs': 27795.304, 'totalHrs': 8.25, 'validHrs': 7.72}, 'fy-NL': {'duration': 456026103, 'reportedSentences': 400, 'buckets': {'dev': 3024, 'invalidated': 2912, 'other': 52677, 'reported': 402, 'test': 3024, 'train': 3699, 'validated': 36036}, 'clips': 91625, 'splits': {'accent': {'': 1}, 'age': {'': 0.59, 'fifties': 0.12, 'thirties': 0.04, 'twenties': 0.02, 'fourties': 0.07, 'sixties': 0.14, 'seventies': 0.01, 'teens': 0, 'eighties': 0}, 'gender': {'': 0.6, 'male': 0.1, 'female': 0.3}}, 'users': 1138, 'size': 2856459160, 'checksum': '300f2e5662aec1a3b5a402c6aa8a597c555be69b81fbe6d94a54358befc437b2', 'avgDurationSecs': 4.977, 'validDurationSecs': 179354.506, 'totalHrs': 126.67, 'validHrs': 49.82}, 'dv': {'duration': 210812751, 'reportedSentences': 50, 'buckets': {'dev': 2242, 'invalidated': 1539, 'other': 14415, 'reported': 49, 'test': 2248, 'train': 2623, 'validated': 25854}, 'clips': 41808, 'splits': {'accent': {'': 1}, 'age': {'': 0.22, 'twenties': 0.18, 'thirties': 0.36, 'fourties': 0.23, 'teens': 0.01, 'nineties': 0}, 'gender': {'': 0.22, 'male': 0.28, 'female': 0.5}}, 'users': 308, 'size': 1355645101, 'checksum': '4fc327c12b947f9a5e9fd9feebfcfa95a3896d79db98dc3bb227266eff04f56f', 'avgDurationSecs': 5.042, 'validDurationSecs': 130366.266, 'totalHrs': 58.55, 'validHrs': 36.21}, 'pa-IN': {'duration': 12812318, 'reportedSentences': 232, 'buckets': {'dev': 279, 'invalidated': 74, 'other': 1243, 'reported': 233, 'test': 388, 'train': 671, 'validated': 1338}, 'clips': 2655, 'splits': {'accent': {'': 1}, 'age': {'': 0.23, 'fourties': 0.04, 'fifties': 0.05, 'thirties': 0.41, 'twenties': 0.26, 'sixties': 0, 'teens': 0}, 'gender': {'': 0.23, 'male': 0.77, 'female': 0}}, 'users': 51, 'size': 92598301, 'checksum': 'a536ccde38a695f9a4d43b5f1a0b2f5a85c6a4d0801cc5d6341f136ab86131c5', 'avgDurationSecs': 4.826, 'validDurationSecs': 6456.829, 'totalHrs': 3.55, 'validHrs': 1.79}, 'vi': {'duration': 62852480, 'reportedSentences': 178, 'buckets': {'dev': 185, 'invalidated': 313, 'other': 11250, 'reported': 179, 'test': 1087, 'train': 2645, 'validated': 4267}, 'clips': 15830, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.02, 'twenties': 0.18, '': 0.24, 'teens': 0.22, 'seventies': 0, 'fourties': 0.02, 'sixties': 0.32}, 'gender': {'male': 0.53, '': 0.24, 'female': 0.21, 'other': 0.02}}, 'users': 219, 'size': 365213095, 'checksum': 'dd4553ac5cabeee244c71a1ad4909f4a0f723e4b33399bc852578b0e0958ff15', 'avgDurationSecs': 3.97, 'validDurationSecs': 16941.979, 'totalHrs': 17.45, 'validHrs': 4.7}, 'or': {'duration': 33971316, 'reportedSentences': 5, 'buckets': {'dev': 309, 'invalidated': 153, 'other': 5442, 'reported': 4, 'test': 218, 'train': 477, 'validated': 1136}, 'clips': 6731, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.17, '': 0.08, 'thirties': 0.74, 'fourties': 0, 'teens': 0}, 'gender': {'male': 0.92, '': 0.08, 'female': 0}}, 'users': 85, 'size': 247540742, 'checksum': '8d1833393a267c07db96cf403dffbd841caaf79d8152e2e7270742f6b0fd4cd7', 'avgDurationSecs': 5.047, 'validDurationSecs': 5733.385, 'totalHrs': 9.43, 'validHrs': 1.59}, 'ga-IE': {'duration': 32223447, 'reportedSentences': 10, 'buckets': {'dev': 512, 'invalidated': 797, 'other': 3813, 'reported': 9, 'test': 512, 'train': 532, 'validated': 4526}, 'clips': 9136, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.25, '': 0.37, 'thirties': 0.26, 'fourties': 0.04, 'sixties': 0.01, 'teens': 0.02, 'fifties': 0.05}, 'gender': {'male': 0.49, '': 0.37, 'female': 0.14, 'other': 0}}, 'users': 157, 'size': 224532834, 'checksum': 'e7dfcbff72b543dac3f3d330f656f441973360ce3813b917af4139616cc30aa0', 'avgDurationSecs': 3.527, 'validDurationSecs': 15963.586, 'totalHrs': 8.95, 'validHrs': 4.43}, 'fi': {'duration': 55337331, 'reportedSentences': 39, 'buckets': {'dev': 1433, 'invalidated': 189, 'other': 5046, 'reported': 38, 'test': 1761, 'train': 2314, 'validated': 6929}, 'clips': 12164, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.13, '': 0.38, 'twenties': 0.1, 'fourties': 0.36, 'teens': 0.01, 'fifties': 0.02, 'seventies': 0}, 'gender': {'male': 0.23, '': 0.38, 'female': 0.38, 'other': 0}}, 'users': 178, 'size': 330869041, 'checksum': '6de0884a012e3e3f329f558920cdbfd9cb4051de620b385d11f9eae0f109668d', 'avgDurationSecs': 4.549, 'validDurationSecs': 31521.898, 'totalHrs': 15.37, 'validHrs': 8.75}, 'hu': {'duration': 91660753, 'reportedSentences': 87, 'buckets': {'dev': 4139, 'invalidated': 794, 'other': 2286, 'reported': 86, 'test': 4582, 'train': 6835, 'validated': 15592}, 'clips': 18672, 'splits': {'accent': {'': 1}, 'age': {'teens': 0.08, '': 0.29, 'thirties': 0.15, 'twenties': 0.39, 'fifties': 0.06, 'fourties': 0.02, 'sixties': 0.01}, 'gender': {'male': 0.6, '': 0.29, 'female': 0.11}}, 'users': 209, 'size': 593607072, 'checksum': 'b6c20484d966f8127cb1630e92b4ad54bfb5a93e933332f20045853679f39796', 'avgDurationSecs': 4.909, 'validDurationSecs': 76541.049, 'totalHrs': 25.46, 'validHrs': 21.26}, 'th': {'duration': 1304114448, 'reportedSentences': 3861, 'buckets': {'dev': 10826, 'invalidated': 8226, 'other': 178341, 'reported': 3862, 'test': 10826, 'train': 30991, 'validated': 125227}, 'clips': 311794, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.2, '': 0.44, 'thirties': 0.05, 'fourties': 0.04, 'teens': 0.05, 'fifties': 0.22, 'eighties': 0, 'sixties': 0}, 'gender': {'male': 0.36, '': 0.44, 'female': 0.19, 'other': 0.01}}, 'users': 7460, 'size': 7590672263, 'checksum': '2a06cc0ebb4f988e83d434c3c6abeed26d983513fe9fd14db3499a65101a759c', 'avgDurationSecs': 4.183, 'validDurationSecs': 523776.404, 'totalHrs': 362.25, 'validHrs': 145.49}, 'lt': {'duration': 73483892, 'reportedSentences': 122, 'buckets': {'dev': 3377, 'invalidated': 551, 'other': 1512, 'reported': 122, 'test': 3666, 'train': 5143, 'validated': 12189}, 'clips': 14252, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.33, '': 0.24, 'thirties': 0.28, 'fifties': 0.05, 'sixties': 0.01, 'teens': 0.03, 'fourties': 0.05}, 'gender': {'male': 0.62, '': 0.24, 'female': 0.14}}, 'users': 252, 'size': 451652341, 'checksum': '4a1d8d6f0465c785f8acf4c2eb2b3120626895aa36bb187c034c8d9bb5188e26', 'avgDurationSecs': 5.156, 'validDurationSecs': 62846.98, 'totalHrs': 20.41, 'validHrs': 17.45}, 'lg': {'duration': 1720023723, 'reportedSentences': 6034, 'buckets': {'dev': 12660, 'invalidated': 38088, 'other': 5933, 'reported': 6039, 'test': 12717, 'train': 55005, 'validated': 252207}, 'clips': 296228, 'splits': {'accent': {'': 1}, 'age': {'': 0.27, 'thirties': 0.22, 'twenties': 0.41, 'fourties': 0.05, 'fifties': 0.03, 'teens': 0.01, 'nineties': 0, 'sixties': 0.01}, 'gender': {'': 0.27, 'female': 0.4, 'male': 0.34}}, 'users': 487, 'size': 10083037056, 'checksum': '249e7fd866620807d9567948abebdf268f445081f22ba25c4cb13a62c87812bd', 'avgDurationSecs': 5.806, 'validDurationSecs': 1464419.377, 'totalHrs': 477.78, 'validHrs': 406.78}, 'hi': {'duration': 62575584, 'reportedSentences': 104, 'buckets': {'dev': 2064, 'invalidated': 638, 'other': 3295, 'reported': 103, 'test': 2839, 'train': 4217, 'validated': 9140}, 'clips': 13073, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.34, 'fourties': 0.03, '': 0.33, 'thirties': 0.28, 'teens': 0.01, 'fifties': 0.01, 'sixties': 0}, 'gender': {'male': 0.63, 'female': 0.04, '': 0.33}}, 'users': 299, 'size': 369876544, 'checksum': '18a2826d53ced567c72cfddf98fd9ac27e79506af6d897a5339cfc46f0f19787', 'avgDurationSecs': 4.787, 'validDurationSecs': 43749.777, 'totalHrs': 17.38, 'validHrs': 12.15}, 'bas': {'duration': 9991980, 'reportedSentences': 8, 'buckets': {'dev': 457, 'invalidated': 483, 'other': 109, 'reported': 7, 'test': 444, 'train': 763, 'validated': 1664}, 'clips': 2256, 'splits': {'accent': {'': 1}, 'age': {'': 0.98, 'fourties': 0.01, 'teens': 0.01}, 'gender': {'': 0.98, 'female': 0.02}}, 'users': 32, 'size': 55579555, 'checksum': 'b289a4d3740c92cb17d5caa25bcbb3303f7c583f9019ead72553c2c044eaafad', 'avgDurationSecs': 4.429, 'validDurationSecs': 7369.971, 'totalHrs': 2.77, 'validHrs': 2.04}, 'sk': {'duration': 68353272, 'reportedSentences': 31, 'buckets': {'dev': 2258, 'invalidated': 691, 'other': 211, 'reported': 30, 'test': 2366, 'train': 2879, 'validated': 16180}, 'clips': 17082, 'splits': {'accent': {'': 1}, 'age': {'': 0.52, 'thirties': 0.23, 'twenties': 0.04, 'fourties': 0.11, 'teens': 0.11}, 'gender': {'': 0.52, 'male': 0.38, 'female': 0.09, 'other': 0.02}}, 'users': 136, 'size': 382625533, 'checksum': '7e603516b30793784726c7752b539282cc2c851d0e47b2dfe3c371ecdd751d0e', 'avgDurationSecs': 4.001, 'validDurationSecs': 64743.938, 'totalHrs': 18.98, 'validHrs': 17.98}, 'kmr': {'duration': 191911824, 'reportedSentences': 578, 'buckets': {'dev': 2396, 'invalidated': 1542, 'other': 3403, 'reported': 577, 'test': 2380, 'train': 2835, 'validated': 38448}, 'clips': 43393, 'splits': {'accent': {'': 1}, 'age': {'': 0.5, 'twenties': 0.31, 'thirties': 0.06, 'fourties': 0.04, 'fifties': 0.09, 'teens': 0.02, 'sixties': 0}, 'gender': {'': 0.5, 'male': 0.34, 'female': 0.16}}, 'users': 299, 'size': 1022772194, 'checksum': 'b469b04d1da7f47d5a9d1b32c7f533cc1921678c8e43bb22300e670e7201474a', 'avgDurationSecs': 4.423, 'validDurationSecs': 170041.846, 'totalHrs': 53.3, 'validHrs': 47.23}, 'bg': {'duration': 46011060, 'reportedSentences': 146, 'buckets': {'dev': 902, 'invalidated': 375, 'other': 2109, 'reported': 145, 'test': 1813, 'train': 3149, 'validated': 5875}, 'clips': 8359, 'splits': {'accent': {'': 1}, 'age': {'fourties': 0.33, 'thirties': 0.07, '': 0.41, 'twenties': 0.18, 'teens': 0.01, 'sixties': 0}, 'gender': {'male': 0.52, 'female': 0.06, '': 0.41}}, 'users': 65, 'size': 268561655, 'checksum': '93d3702a7a91e2e58bafd8c66fcb406be321d70504c5cba95611b34b13ca07df', 'avgDurationSecs': 5.504, 'validDurationSecs': 32338.196, 'totalHrs': 12.78, 'validHrs': 8.98}, 'kk': {'duration': 6555816, 'reportedSentences': 8, 'buckets': {'dev': 341, 'invalidated': 191, 'other': 7, 'reported': 7, 'test': 365, 'train': 401, 'validated': 1112}, 'clips': 1310, 'splits': {'accent': {'': 1}, 'age': {'': 0.53, 'thirties': 0.03, 'twenties': 0.28, 'teens': 0.06, 'fifties': 0.1}, 'gender': {'': 0.54, 'male': 0.43, 'female': 0.03}}, 'users': 79, 'size': 37614021, 'checksum': '028c7a95b3402f5b51df4edc086670593625d1ce26f8cbdbf6c14273c88c3b6a', 'avgDurationSecs': 5.004, 'validDurationSecs': 5564.937, 'totalHrs': 1.82, 'validHrs': 1.54}, 'ba': {'duration': 958205736, 'reportedSentences': 863, 'buckets': {'dev': 14560, 'invalidated': 7884, 'other': 60, 'reported': 866, 'test': 14509, 'train': 118993, 'validated': 208559}, 'clips': 216503, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.17, '': 0.3, 'fourties': 0.06, 'fifties': 0.05, 'twenties': 0.17, 'sixties': 0.2, 'seventies': 0, 'teens': 0.04}, 'gender': {'male': 0.3, '': 0.3, 'female': 0.4}}, 'users': 888, 'size': 5375041473, 'checksum': '12b85d10cfae0784937717b2213b667210f9e2bb7e37c38d06c62b34920a0d57', 'avgDurationSecs': 4.426, 'validDurationSecs': 923046.933, 'totalHrs': 266.16, 'validHrs': 256.4}, 'gl': {'duration': 56512476, 'reportedSentences': 166, 'buckets': {'dev': 2382, 'invalidated': 304, 'other': 3305, 'reported': 165, 'test': 2375, 'train': 3291, 'validated': 8106}, 'clips': 11715, 'splits': {'accent': {'': 1}, 'age': {'': 0.37, 'thirties': 0.41, 'fifties': 0.09, 'twenties': 0.07, 'fourties': 0.04, 'teens': 0, 'sixties': 0.01}, 'gender': {'': 0.38, 'male': 0.4, 'female': 0.21, 'other': 0.01}}, 'users': 140, 'size': 324986608, 'checksum': '4ab20faa2d8e9811bf6ab4135daefa61a998225db481a3918a8e117e2a1e5b0e', 'avgDurationSecs': 4.824, 'validDurationSecs': 39102.871, 'totalHrs': 15.69, 'validHrs': 10.86}, 'ug': {'duration': 235420236, 'reportedSentences': 181, 'buckets': {'dev': 2746, 'invalidated': 1932, 'other': 62, 'reported': 180, 'test': 2746, 'train': 3292, 'validated': 37038}, 'clips': 39032, 'splits': {'accent': {'': 1}, 'age': {'': 0.64, 'fifties': 0.02, 'twenties': 0.14, 'thirties': 0.12, 'fourties': 0.07, 'teens': 0, 'eighties': 0.01}, 'gender': {'': 0.64, 'male': 0.27, 'female': 0.09, 'other': 0}}, 'users': 385, 'size': 1372345697, 'checksum': '4d904c8244451208b4563a525413aae4d0d44a7e1643bfdc1e42bf8a12de66f4', 'avgDurationSecs': 6.031, 'validDurationSecs': 223393.49, 'totalHrs': 65.39, 'validHrs': 62.05}, 'hy-AM': {'duration': 16162056, 'reportedSentences': 26, 'buckets': {'dev': 344, 'invalidated': 89, 'other': 1246, 'reported': 25, 'test': 369, 'train': 595, 'validated': 1309}, 'clips': 2644, 'splits': {'accent': {'': 1}, 'age': {'': 0.37, 'thirties': 0.14, 'twenties': 0.36, 'fifties': 0.04, 'teens': 0.08}, 'gender': {'': 0.37, 'male': 0.23, 'female': 0.4}}, 'users': 60, 'size': 94866525, 'checksum': '7d9535bd096dd60bac0f692b5621a3995b3f6c26add8b1bd63e3e76bca730a30', 'avgDurationSecs': 6.113, 'validDurationSecs': 8001.563, 'totalHrs': 4.48, 'validHrs': 2.22}, 'be': {'duration': 3797989380, 'reportedSentences': 3097, 'buckets': {'dev': 15866, 'invalidated': 25367, 'other': 574, 'reported': 3098, 'test': 15877, 'train': 346139, 'validated': 771750}, 'clips': 797691, 'splits': {'accent': {'': 1}, 'age': {'': 0.78, 'fourties': 0.07, 'thirties': 0.08, 'twenties': 0.06, 'teens': 0.01, 'fifties': 0, 'sixties': 0, 'seventies': 0}, 'gender': {'': 0.78, 'male': 0.1, 'female': 0.13, 'other': 0}}, 'users': 6316, 'size': 21530210404, 'checksum': '6ec07dc257128cf65b80ce4a5de429bdb2e43d6052b589c4859783faada0d9e0', 'avgDurationSecs': 4.761, 'validDurationSecs': 3674478.343, 'totalHrs': 1054.99, 'validHrs': 1020.68}, 'ur': {'duration': 48370572, 'reportedSentences': 35, 'buckets': {'dev': 2864, 'invalidated': 903, 'other': 121, 'reported': 37, 'test': 2796, 'train': 3674, 'validated': 10411}, 'clips': 11435, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.71, '': 0.18, 'fourties': 0.07, 'thirties': 0.02, 'teens': 0.01, 'fifties': 0}, 'gender': {'male': 0.61, '': 0.18, 'female': 0.21}}, 'users': 108, 'size': 283099171, 'checksum': '53ee69b3a42ceabacbb88d5a7831d5e6f8748bb84fde4087b6a1e5877f653d4b', 'avgDurationSecs': 4.23, 'validDurationSecs': 44039.005, 'totalHrs': 13.43, 'validHrs': 12.23}, 'gn': {'duration': 8978220, 'reportedSentences': 21, 'buckets': {'dev': 100, 'invalidated': 62, 'other': 1430, 'reported': 20, 'test': 163, 'train': 295, 'validated': 558}, 'clips': 2050, 'splits': {'accent': {'': 1}, 'age': {'': 0.6, 'twenties': 0.37, 'thirties': 0.03}, 'gender': {'': 0.6, 'male': 0.29, 'female': 0.11}}, 'users': 62, 'size': 49435800, 'checksum': '9dc4b54c18770a50f6d0d56faec37cbe3e95ddf70890791d7277f15d8a002af6', 'avgDurationSecs': 4.38, 'validDurationSecs': 2443.828, 'totalHrs': 2.49, 'validHrs': 0.67}, 'sr': {'duration': 6616980, 'reportedSentences': 19, 'buckets': {'dev': 516, 'invalidated': 38, 'other': 0, 'reported': 18, 'test': 742, 'train': 1037, 'validated': 2297}, 'clips': 2335, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.69, '': 0.15, 'fifties': 0.01, 'fourties': 0.11, 'thirties': 0.04, 'teens': 0}, 'gender': {'male': 0.39, '': 0.15, 'female': 0.46}}, 'users': 53, 'size': 36628440, 'checksum': 'c80dbab34150c44621bcd2340f2aad60f7711b989b726f0aab6dc5655c532299', 'avgDurationSecs': 2.834, 'validDurationSecs': 6509.295, 'totalHrs': 1.83, 'validHrs': 1.8}, 'uz': {'duration': 833326524, 'reportedSentences': 1664, 'buckets': {'dev': 10819, 'invalidated': 12183, 'other': 115624, 'reported': 1681, 'test': 12007, 'train': 44479, 'validated': 78437}, 'clips': 206244, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.38, '': 0.42, 'thirties': 0.01, 'teens': 0.18, 'fifties': 0, 'fourties': 0.01, 'nineties': 0}, 'gender': {'male': 0.42, '': 0.42, 'female': 0.16}}, 'users': 1503, 'size': 4644040593, 'checksum': '7ed210d6c97d6fee3bd57c6d1f85e11a32d7a7237a89d09470ab04e8e98373e4', 'avgDurationSecs': 4.04, 'validDurationSecs': 316923.802, 'totalHrs': 231.47, 'validHrs': 88.03}, 'mr': {'duration': 88055640, 'reportedSentences': 38, 'buckets': {'dev': 1794, 'invalidated': 2094, 'other': 1896, 'reported': 37, 'test': 1806, 'train': 2125, 'validated': 10331}, 'clips': 14321, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.14, 'sixties': 0, 'twenties': 0.26, '': 0.05, 'teens': 0.55}, 'gender': {'male': 0.19, 'female': 0.76, '': 0.05}}, 'users': 70, 'size': 516203389, 'checksum': '3312d72c9a10f8762cbe0a5c7903faf3cb6b74a446c3028408c47b32d4f6b4cc', 'avgDurationSecs': 6.149, 'validDurationSecs': 63522.297, 'totalHrs': 24.45, 'validHrs': 17.64}, 'da': {'duration': 30950316, 'reportedSentences': 114, 'buckets': {'dev': 1602, 'invalidated': 264, 'other': 387, 'reported': 113, 'test': 1583, 'train': 2038, 'validated': 6509}, 'clips': 7160, 'splits': {'accent': {'': 1}, 'age': {'': 0.37, 'thirties': 0.3, 'twenties': 0.21, 'sixties': 0, 'fourties': 0.07, 'fifties': 0.04, 'teens': 0}, 'gender': {'': 0.37, 'female': 0.09, 'male': 0.54}}, 'users': 169, 'size': 177172753, 'checksum': 'a57080a05c26bfa96cd4c9c0d83b01961781930fcd1a1bd7c3bb47ea377bae37', 'avgDurationSecs': 4.323, 'validDurationSecs': 28136.258, 'totalHrs': 8.59, 'validHrs': 7.81}, 'myv': {'duration': 8582580, 'reportedSentences': 15, 'buckets': {'dev': 351, 'invalidated': 4, 'other': 324, 'reported': 14, 'test': 297, 'train': 522, 'validated': 1173}, 'clips': 1501, 'splits': {'accent': {'': 1}, 'age': {'sixties': 0.3, '': 0.34, 'thirties': 0.32, 'twenties': 0.03}, 'gender': {'male': 0.66, '': 0.34}}, 'users': 5, 'size': 50163913, 'checksum': '3b5c0c8f99951e6ae6c20064a7d1e9204f92b3a9fa10c54cc636931cbde0cf9a', 'avgDurationSecs': 5.718, 'validDurationSecs': 6707.106, 'totalHrs': 2.38, 'validHrs': 1.86}, 'nn-NO': {'duration': 2319192, 'reportedSentences': 5, 'buckets': {'dev': 96, 'invalidated': 15, 'other': 14, 'reported': 4, 'test': 150, 'train': 234, 'validated': 485}, 'clips': 514, 'splits': {'accent': {'': 1}, 'age': {'': 0.39, 'thirties': 0.34, 'twenties': 0.27}, 'gender': {'': 0.39, 'female': 0.29, 'male': 0.27, 'other': 0.05}}, 'users': 18, 'size': 13594849, 'checksum': '3efa85f8616fffda1e35a76d8a1fe8a93b6136e31fd1ae092ec5e6eb179061fe', 'avgDurationSecs': 4.512, 'validDurationSecs': 2188.343, 'totalHrs': 0.64, 'validHrs': 0.6}, 'ha': {'duration': 38472084, 'reportedSentences': 14, 'buckets': {'dev': 0, 'invalidated': 153, 'other': 5876, 'reported': 14, 'test': 896, 'train': 1941, 'validated': 2837}, 'clips': 8866, 'splits': {'accent': {'': 1}, 'age': {'': 0.18, 'thirties': 0.77, 'twenties': 0.02, 'fourties': 0, 'fifties': 0.03}, 'gender': {'': 0.18, 'male': 0.54, 'female': 0.28}}, 'users': 27, 'size': 225075429, 'checksum': '395c80ab6bacae21042ff7360976b7bde7c3cec805ca97dac1761138dfbac9e7', 'avgDurationSecs': 4.339, 'validDurationSecs': 12310.546, 'totalHrs': 10.68, 'validHrs': 3.41}, 'ckb': {'duration': 189151776, 'reportedSentences': 1750, 'buckets': {'dev': 4070, 'invalidated': 1414, 'other': 3434, 'reported': 1749, 'test': 4311, 'train': 6403, 'validated': 45048}, 'clips': 49896, 'splits': {'accent': {'': 1}, 'age': {'': 0.41, 'thirties': 0.11, 'twenties': 0.42, 'fourties': 0.01, 'teens': 0.01, 'fifties': 0.03}, 'gender': {'': 0.38, 'male': 0.56, 'female': 0.06, 'other': 0}}, 'users': 250, 'size': 1015400895, 'checksum': '6aa0b9d17f31a6f8fa0b23fb90e1134e75c23f104acc2b69c704f0adcdd9ea30', 'avgDurationSecs': 3.791, 'validDurationSecs': 170773.393, 'totalHrs': 52.54, 'validHrs': 47.43}, 'ml': {'duration': 9444672, 'reportedSentences': 107, 'buckets': {'dev': 0, 'invalidated': 1, 'other': 1914, 'reported': 106, 'test': 24, 'train': 366, 'validated': 390}, 'clips': 2305, 'splits': {'accent': {'': 1}, 'age': {'': 0.47, 'twenties': 0.53}, 'gender': {'': 0.47, 'male': 0.53}}, 'users': 16, 'size': 54752972, 'checksum': '57e0639b25530a0feb23530b857858782828833cd203a09018aadf99cf7376d8', 'avgDurationSecs': 4.097, 'validDurationSecs': 1598.014, 'totalHrs': 2.62, 'validHrs': 0.44}, 'mdf': {'duration': 1791720, 'reportedSentences': 6, 'buckets': {'dev': 47, 'invalidated': 5, 'other': 98, 'reported': 5, 'test': 73, 'train': 116, 'validated': 236}, 'clips': 339, 'splits': {'accent': {'': 1}, 'age': {'sixties': 0.06, '': 0.59, 'fourties': 0.35}, 'gender': {'male': 0.06, '': 0.59, 'female': 0.35}}, 'users': 10, 'size': 10525562, 'checksum': 'a2d5ad04d92814618b20fb80881afa151784fb893f94e020976d7618f3f1a6b5', 'avgDurationSecs': 5.285, 'validDurationSecs': 1247.333, 'totalHrs': 0.49, 'validHrs': 0.34}, 'sw': {'duration': 2589824160, 'reportedSentences': 363, 'buckets': {'dev': 9008, 'invalidated': 10070, 'other': 353864, 'reported': 368, 'test': 8934, 'train': 19861, 'validated': 120101}, 'clips': 484035, 'splits': {'accent': {'': 1}, 'age': {'': 0.29, 'twenties': 0.45, 'thirties': 0.14, 'teens': 0, 'fifties': 0.06, 'fourties': 0.05, 'sixties': 0.01}, 'gender': {'': 0.26, 'male': 0.39, 'female': 0.35, 'other': 0}}, 'users': 569, 'size': 15139558447, 'checksum': 'd124fdb0e12074adfa137f69f28f35b9d8d7cdcdf13f4feae0d39bd252c07a65', 'avgDurationSecs': 5.35, 'validDurationSecs': 642599.133, 'totalHrs': 719.39, 'validHrs': 178.49}, 'sat': {'duration': 2042928, 'reportedSentences': 7, 'buckets': {'dev': 0, 'invalidated': 12, 'other': 242, 'reported': 6, 'test': 47, 'train': 109, 'validated': 156}, 'clips': 410, 'splits': {'accent': {'': 1}, 'age': {'': 0.13, 'twenties': 0.62, 'fourties': 0.02, 'fifties': 0.02, 'teens': 0.02, 'thirties': 0.18}, 'gender': {'': 0.09, 'male': 0.89, 'female': 0.02}}, 'users': 8, 'size': 10811844, 'checksum': '4c00be164ac7ac124d3bed39ef87c82cec3f8bc911aefee451b802d65bd443e8', 'avgDurationSecs': 4.983, 'validDurationSecs': 777.309, 'totalHrs': 0.56, 'validHrs': 0.21}, 'tig': {'duration': 74016, 'reportedSentences': 1, 'buckets': {'dev': 0, 'invalidated': 8, 'other': 0, 'reported': 0, 'test': 0, 'train': 10, 'validated': 10}, 'clips': 18, 'splits': {'accent': {'': 1}, 'age': {'': 0.72, 'twenties': 0.28}, 'gender': {'': 0.72, 'male': 0.28}}, 'users': 4, 'size': 436155, 'checksum': '37fdd8b0cfded4c466b3c32d4d7cc670165383ef77c303b5e69b2c937b981729', 'avgDurationSecs': 4.112, 'validDurationSecs': 41.12, 'totalHrs': 0.02, 'validHrs': 0.01}, 'ig': {'duration': 29924424, 'reportedSentences': 5, 'buckets': {'dev': 0, 'invalidated': 1, 'other': 5481, 'reported': 5, 'test': 2, 'train': 5, 'validated': 7}, 'clips': 5489, 'splits': {'accent': {'': 1}, 'age': {'': 1, 'twenties': 0.33, 'teens': 0.05, 'eighties': 0, 'thirties': 0.04, 'sixties': 0.02}, 'gender': {'': 1, 'male': 0.14, 'female': 0.31}}, 'users': 96, 'size': 174792302, 'checksum': '921d62787743f854bf771f3f3b9492b167292992d9d3565a535b3f943bdc0c9c', 'avgDurationSecs': 5.452, 'validDurationSecs': 38.162, 'totalHrs': 8.31, 'validHrs': 0.01}, 'nan-tw': {'duration': 11365740, 'reportedSentences': 20, 'buckets': {'dev': 161, 'invalidated': 82, 'other': 3211, 'reported': 19, 'test': 285, 'train': 503, 'validated': 950}, 'clips': 4243, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.16, '': 0.2, 'twenties': 0.52, 'fourties': 0.1, 'teens': 0.01, 'fifties': 0}, 'gender': {'male': 0.43, '': 0.2, 'other': 0.27, 'female': 0.09}}, 'users': 55, 'size': 65348573, 'checksum': '275abb62dd72bfe1a6db9092851da16e308d8b172402bfee5c82e7094de48feb', 'avgDurationSecs': 2.679, 'validDurationSecs': 2544.769, 'totalHrs': 3.15, 'validHrs': 0.7}, 'mhr': {'duration': 236876904, 'reportedSentences': 32, 'buckets': {'dev': 9152, 'invalidated': 1405, 'other': 6723, 'reported': 31, 'test': 9074, 'train': 21828, 'validated': 41094}, 'clips': 49222, 'splits': {'accent': {'': 1}, 'age': {'fifties': 0.1, '': 0.11, 'sixties': 0.11, 'thirties': 0.19, 'fourties': 0.21, 'twenties': 0.24, 'teens': 0.05}, 'gender': {'male': 0.23, '': 0.11, 'female': 0.66}}, 'users': 180, 'size': 1341912846, 'checksum': '39621e9ca4d12e2872546131f1eca3197324943ee29bba657f10489f098cc67a', 'avgDurationSecs': 4.812, 'validDurationSecs': 197761.56, 'totalHrs': 65.79, 'validHrs': 54.93}, 'bn': {'duration': 1438112808, 'reportedSentences': 693, 'buckets': {'dev': 7748, 'invalidated': 5844, 'other': 192522, 'reported': 717, 'test': 7748, 'train': 14503, 'validated': 32754}, 'clips': 231120, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.02, 'twenties': 0.22, '': 0.72, 'teens': 0.04, 'fourties': 0}, 'gender': {'male': 0.24, '': 0.72, 'female': 0.04, 'other': 0}}, 'users': 19863, 'size': 8262390506, 'checksum': '599a5f7c9e55a297928da390345a19180b279a1f013081e7255a657fc99f98d5', 'avgDurationSecs': 6.222, 'validDurationSecs': 203807.316, 'totalHrs': 399.47, 'validHrs': 56.61}, 'tok': {'duration': 25490268, 'reportedSentences': 88, 'buckets': {'dev': 1076, 'invalidated': 129, 'other': 1787, 'reported': 87, 'test': 1326, 'train': 2242, 'validated': 5250}, 'clips': 7166, 'splits': {'accent': {'': 1}, 'age': {'': 0.51, 'twenties': 0.23, 'teens': 0.16, 'thirties': 0.09}, 'gender': {'': 0.51, 'male': 0.37, 'other': 0.01, 'female': 0.11}}, 'users': 50, 'size': 148247234, 'checksum': '9e6d7dd7728199a2de17d97c79e87fe8c522b8c1575b595d2dcfca9a38e36d53', 'avgDurationSecs': 3.557, 'validDurationSecs': 18674.84, 'totalHrs': 7.08, 'validHrs': 5.18}, 'yue': {'duration': 58318992, 'reportedSentences': 96, 'buckets': {'dev': 1648, 'invalidated': 877, 'other': 5871, 'reported': 95, 'test': 1636, 'train': 1854, 'validated': 6882}, 'clips': 13630, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.35, '': 0.36, 'twenties': 0.14, 'fourties': 0.08, 'sixties': 0, 'fifties': 0, 'teens': 0.06}, 'gender': {'male': 0.15, '': 0.49, 'female': 0.31, 'other': 0.05}}, 'users': 332, 'size': 331489433, 'checksum': '6aec37d62a600b641d5377d25b956e5f6918cf9abe05debbdfdfe3ff8ff35624', 'avgDurationSecs': 4.279, 'validDurationSecs': 29446.17, 'totalHrs': 16.19, 'validHrs': 8.17}, 'sah': {'duration': 24207744, 'reportedSentences': 1, 'buckets': {'dev': 1083, 'invalidated': 99, 'other': 26, 'reported': 0, 'test': 1191, 'train': 1570, 'validated': 3912}, 'clips': 4037, 'splits': {'accent': {'': 1}, 'age': {'': 0.37, 'twenties': 0.03, 'fourties': 0.07, 'thirties': 0.43, 'teens': 0.1, 'fifties': 0}, 'gender': {'': 0.37, 'male': 0.54, 'female': 0.1}}, 'users': 52, 'size': 184519417, 'checksum': '3e6bfeba4c17c80eb7c2e5b5cd95ba958747001ed9a49d52d86333c5332ada5b', 'avgDurationSecs': 5.996, 'validDurationSecs': 23458.185, 'totalHrs': 6.72, 'validHrs': 6.51}, 'fa': {'buckets': {'dev': 9852, 'invalidated': 13220, 'other': 34785, 'reported': 2078, 'test': 9852, 'train': 24150, 'validated': 290215}, 'reportedSentences': 2070, 'duration': 1356155252, 'clips': 338220, 'splits': {'accent': {'': 1}, 'age': {'': 0.24, 'twenties': 0.31, 'thirties': 0.37, 'fifties': 0.02, 'fourties': 0.03, 'teens': 0.03, 'sixties': 0}, 'gender': {'': 0.21, 'male': 0.72, 'female': 0.07, 'other': 0}}, 'users': 4058, 'size': 10032969547, 'checksum': '62f07bed6cb1d28e51575c684bcb6eea1b50e553f50640d8c82576fb150c01fe', 'avgDurationSecs': 4.01, 'validDurationSecs': 1163670.382, 'totalHrs': 376.7, 'validHrs': 323.24}, 'fr': {'buckets': {'dev': 16041, 'invalidated': 54752, 'other': 1757, 'reported': 6205, 'test': 16041, 'train': 445537, 'validated': 612070}, 'duration': 3334982374, 'reportedSentences': 6129, 'clips': 668579, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.18, 'thirties': 0.17, '': 0.35, 'teens': 0.03, 'fourties': 0.13, 'fifties': 0.09, 'sixties': 0.03, 'seventies': 0.01, 'eighties': 0, 'nineties': 0}, 'gender': {'male': 0.62, '': 0.27, 'female': 0.1, 'other': 0.01}}, 'users': 16291, 'size': 24260798182, 'checksum': '6e5326e0cd1841ed3e5fceeb46e71d39c1cb653968f0e888c62d0369cc06cb9d', 'avgDurationSecs': 4.988, 'validDurationSecs': 3053106.157, 'totalHrs': 926.38, 'validHrs': 848.08}, 'es': {'buckets': {'dev': 15449, 'invalidated': 47995, 'other': 197465, 'reported': 1690, 'test': 15449, 'train': 215745, 'validated': 290972}, 'duration': 2704659930, 'reportedSentences': 1677, 'clips': 536432, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.11, '': 0.32, 'fifties': 0.08, 'twenties': 0.25, 'teens': 0.03, 'fourties': 0.06, 'sixties': 0.15, 'eighties': 0, 'seventies': 0, 'nineties': 0}, 'gender': {'male': 0.51, '': 0.32, 'other': 0.01, 'female': 0.17}}, 'users': 23013, 'size': 19800467043, 'checksum': '61b8bd40123962d5e98e7740b11a5f0b3abb8691c44632e0f24b30833d41e3f1', 'avgDurationSecs': 5.042, 'validDurationSecs': 1467064.435, 'totalHrs': 751.29, 'validHrs': 407.51}, 'sl': {'buckets': {'dev': 1223, 'invalidated': 225, 'other': 905, 'reported': 28, 'test': 1219, 'train': 1384, 'validated': 9291}, 'duration': 40129030, 'reportedSentences': 29, 'clips': 10421, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.52, 'teens': 0.09, '': 0.17, 'sixties': 0.08, 'fifties': 0.08, 'fourties': 0.02, 'thirties': 0.05}, 'gender': {'female': 0.15, 'male': 0.67, '': 0.17, 'other': 0}}, 'users': 128, 'size': 289537155, 'checksum': '394a3c8c1fe1c24ae7b63ccab3428a5d87e5f51730984038f09254a2793e88b5', 'avgDurationSecs': 3.851, 'validDurationSecs': 35777.643, 'totalHrs': 11.14, 'validHrs': 9.93}, 'kab': {'buckets': {'dev': 14875, 'invalidated': 19148, 'other': 101261, 'reported': 4807, 'test': 14875, 'train': 140213, 'validated': 596320}, 'duration': 2382988048, 'reportedSentences': 4802, 'clips': 716729, 'splits': {'accent': {'': 1}, 'age': {'fourties': 0.09, 'thirties': 0.3, '': 0.27, 'fifties': 0.19, 'twenties': 0.12, 'eighties': 0, 'teens': 0, 'sixties': 0.03, 'seventies': 0}, 'gender': {'male': 0.55, '': 0.25, 'female': 0.2, 'other': 0}}, 'users': 1441, 'size': 17975235335, 'checksum': '9eb1cbf3d9d83c9659b581cbe4088bc9f275443fcfeafbdc09f6dfc9ef875b0c', 'avgDurationSecs': 3.325, 'validDurationSecs': 1982650.95, 'totalHrs': 661.94, 'validHrs': 550.73}, 'cy': {'buckets': {'dev': 5192, 'invalidated': 4283, 'other': 17960, 'reported': 152, 'test': 5202, 'train': 7578, 'validated': 87106}, 'duration': 529187805, 'reportedSentences': 153, 'clips': 109349, 'splits': {'accent': {'': 1}, 'age': {'fourties': 0.16, 'twenties': 0.13, 'sixties': 0.07, 'fifties': 0.09, '': 0.42, 'thirties': 0.09, 'seventies': 0.01, 'eighties': 0, 'teens': 0.02}, 'gender': {'male': 0.33, 'female': 0.25, '': 0.41, 'other': 0.01}}, 'users': 1705, 'size': 3908225447, 'checksum': 'dcc085ccdf2790020c305db1f56d19f67035cdf784746fd3c658c8d90b8388fb', 'avgDurationSecs': 4.839, 'validDurationSecs': 421544.165, 'totalHrs': 146.99, 'validHrs': 117.09}, 'ca': {'buckets': {'dev': 16230, 'invalidated': 56440, 'other': 456885, 'reported': 3754, 'test': 16230, 'train': 657479, 'validated': 803439}, 'duration': 7362990642, 'reportedSentences': 3686, 'clips': 1316764, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.07, 'fifties': 0.17, 'fourties': 0.11, 'twenties': 0.05, '': 0.34, 'sixties': 0.22, 'teens': 0.01, 'seventies': 0.03, 'eighties': 0, 'nineties': 0}, 'gender': {'male': 0.45, '': 0.34, 'female': 0.21, 'other': 0}}, 'users': 26053, 'size': 47095083457, 'checksum': '30993c465a078f4144f1399553d92506d72cbb42ae3f00e34f69f704f0f5b39f', 'avgDurationSecs': 5.592, 'validDurationSecs': 4492615.107, 'totalHrs': 2045.27, 'validHrs': 1247.94}, 'de': {'buckets': {'dev': 16033, 'invalidated': 45660, 'other': 5246, 'reported': 7276, 'test': 16033, 'train': 439325, 'validated': 765790}, 'duration': 4197654857, 'reportedSentences': 7212, 'clips': 816696, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.2, 'fourties': 0.17, '': 0.31, 'thirties': 0.15, 'teens': 0.03, 'sixties': 0.03, 'fifties': 0.1, 'seventies': 0, 'eighties': 0, 'nineties': 0}, 'gender': {'male': 0.59, '': 0.31, 'female': 0.09, 'other': 0.01}}, 'users': 16645, 'size': 30220261910, 'checksum': 'a8f088b7a55f9754c165c218655b0a7131d3f2e17f56d8cbc9ba6eca32228a81', 'avgDurationSecs': 5.14, 'validDurationSecs': 3936008.151, 'totalHrs': 1166.01, 'validHrs': 1093.33}, 'tt': {'buckets': {'dev': 3069, 'invalidated': 378, 'other': 0, 'reported': 3, 'test': 5055, 'train': 9840, 'validated': 28452}, 'duration': 107943146, 'reportedSentences': 4, 'clips': 28830, 'splits': {'accent': {'': 1}, 'age': {'': 0.19, 'thirties': 0.73, 'twenties': 0.05, 'sixties': 0, 'fifties': 0.01, 'teens': 0, 'fourties': 0, 'seventies': 0.01}, 'gender': {'': 0.19, 'male': 0.79, 'female': 0.02}}, 'users': 215, 'size': 800243818, 'checksum': '14d199b9b94afb46db45f8db6e159b6cc63dab09d778db4fcfa15f0246498f52', 'avgDurationSecs': 3.744, 'validDurationSecs': 106527.866, 'totalHrs': 29.98, 'validHrs': 29.59}, 'ta': {'buckets': {'dev': 11563, 'invalidated': 5515, 'other': 76526, 'reported': 3229, 'test': 11712, 'train': 39758, 'validated': 128031}, 'duration': 1302133236, 'reportedSentences': 3229, 'clips': 210072, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.08, 'thirties': 0.09, '': 0.71, 'fourties': 0.03, 'seventies': 0.02, 'fifties': 0.03, 'teens': 0.04, 'sixties': 0, 'eighties': 0}, 'gender': {'male': 0.16, '': 0.7, 'other': 0, 'female': 0.13}}, 'users': 730, 'size': 7812683439, 'checksum': 'c8ac4adfb3e1bc0f8cec4ed15cd61ec12bc0784036f5edb0cc7528a2277dcc77', 'avgDurationSecs': 6.199, 'validDurationSecs': 793601.338, 'totalHrs': 361.7, 'validHrs': 220.44}, 'ru': {'buckets': {'dev': 9457, 'invalidated': 6549, 'other': 18733, 'reported': 290, 'test': 9456, 'train': 21909, 'validated': 115007}, 'duration': 725508648, 'reportedSentences': 284, 'clips': 140289, 'splits': {'accent': {'': 1}, 'age': {'twenties': 0.41, 'teens': 0.08, '': 0.2, 'fourties': 0.14, 'thirties': 0.13, 'fifties': 0.03, 'sixties': 0, 'seventies': 0}, 'gender': {'male': 0.63, '': 0.21, 'other': 0, 'female': 0.16}}, 'users': 2585, 'size': 5132638847, 'checksum': 'f3e386715e55b5e293b77ccda252d1c4c5458fb5f1290f54e241e7fe45c0bd60', 'avgDurationSecs': 5.172, 'validDurationSecs': 594762.049, 'totalHrs': 201.53, 'validHrs': 165.21}, 'nl': {'buckets': {'dev': 10368, 'invalidated': 4995, 'other': 2233, 'reported': 310, 'test': 10552, 'train': 29031, 'validated': 82976}, 'duration': 388289017, 'reportedSentences': 311, 'clips': 90204, 'splits': {'accent': {'': 1}, 'age': {'': 0.41, 'twenties': 0.22, 'fourties': 0.14, 'thirties': 0.11, 'teens': 0.02, 'fifties': 0.08, 'sixties': 0.01, 'nineties': 0, 'eighties': 0, 'seventies': 0}, 'gender': {'': 0.42, 'male': 0.47, 'female': 0.11, 'other': 0}}, 'users': 1488, 'size': 2678633936, 'checksum': '23467ca7212bedde6f5b53689cb76e8a78d7502e937d990c64f304cf4f472dbc', 'avgDurationSecs': 4.305, 'validDurationSecs': 357175.618, 'totalHrs': 107.85, 'validHrs': 99.21}, 'it': {'buckets': {'dev': 14935, 'invalidated': 16988, 'other': 27, 'reported': 5204, 'test': 14934, 'train': 145968, 'validated': 212264}, 'duration': 1226597196, 'reportedSentences': 5200, 'clips': 229279, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.16, 'twenties': 0.2, '': 0.29, 'fifties': 0.16, 'fourties': 0.14, 'seventies': 0, 'sixties': 0.02, 'teens': 0.01, 'eighties': 0, 'nineties': 0}, 'gender': {'female': 0.12, 'male': 0.59, '': 0.29, 'other': 0}}, 'users': 6640, 'size': 8536806025, 'checksum': '3c97337d99684cda8c3c383b68555f1124fd8646df61e03b7578212580d71f1b', 'avgDurationSecs': 5.35, 'validDurationSecs': 1135570.319, 'totalHrs': 340.72, 'validHrs': 315.43}, 'vot': {'duration': 1025976, 'buckets': {'dev': 0, 'invalidated': 324, 'other': 0, 'test': 6, 'train': 96, 'validated': 102}, 'clips': 426, 'splits': {'accent': {'': 1}, 'age': {'': 0.25, 'twenties': 0.73, 'teens': 0.01}, 'gender': {'': 0.25, 'male': 0.75}}, 'users': 5, 'size': 7892306, 'checksum': '5ca8a934d111c6ea330df3c1f52c45fbec0dbbae0ff4fe83e97c32b2f6294b60', 'avgDurationSecs': 2.408, 'validDurationSecs': 245.656, 'totalHrs': 0.28, 'validHrs': 0.06}, 'az': {'duration': 604476, 'buckets': {'dev': 15, 'invalidated': 28, 'other': 0, 'test': 26, 'train': 39, 'validated': 80}, 'clips': 108, 'splits': {'accent': {'': 1}, 'age': {'': 0.41, 'twenties': 0.56, 'fourties': 0.03}, 'gender': {'': 0.41, 'male': 0.59}}, 'users': 13, 'size': 3529914, 'checksum': '8fabb36856ca872e56a5cb9ada9029086c05600a809524afda3e8883a687640a', 'avgDurationSecs': 5.597, 'validDurationSecs': 447.76, 'totalHrs': 0.16, 'validHrs': 0.12}, 'mk': {'duration': 729000, 'buckets': {'dev': 0, 'invalidated': 7, 'other': 9, 'test': 15, 'train': 114, 'validated': 129}, 'clips': 145, 'splits': {'accent': {'': 1}, 'age': {'thirties': 0.38, '': 1, 'twenties': 0.14, 'teens': 0.1}, 'gender': {'male': 0.62, '': 1}}, 'users': 5, 'size': 4286542, 'checksum': 'e49da4f1b244b00f67879b48877ad85b9659ff2b050c34d1427e220a9b67ebb9', 'avgDurationSecs': 5.028, 'validDurationSecs': 648.559, 'totalHrs': 0.2, 'validHrs': 0.18}}, 'totalDuration': 72782088097, 'totalValidDurationSecs': 53904443, 'totalHrs': 20217, 'totalValidHrs': 14973, 'version': '9.0.0'}
 
 
esc-datasets.py DELETED
@@ -1,1484 +0,0 @@
1
- # coding=utf-8
2
- #
3
- # Licensed under the Apache License, Version 2.0 (the "License");
4
- # you may not use this file except in compliance with the License.
5
- # You may obtain a copy of the License at
6
- #
7
- # http://www.apache.org/licenses/LICENSE-2.0
8
- #
9
- # Unless required by applicable law or agreed to in writing, software
10
- # distributed under the License is distributed on an "AS IS" BASIS,
11
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
- # See the License for the specific language governing permissions and
13
- # limitations under the License.
14
-
15
- # Lint as: python3
16
- """ESC benchmark datasets."""
17
-
18
- import csv
19
- from collections import defaultdict
20
- import os
21
- import json
22
- import urllib
23
- import re
24
- import logging
25
-
26
- import soundfile as sf
27
- import numpy as np
28
- from tqdm.auto import tqdm
29
- import requests
30
- from io import BytesIO
31
- from pathlib import Path
32
- from huggingface_hub import HfApi, HfFolder
33
- import datasets
34
-
35
-
36
- _DESCRIPTIONS = {
37
- "ami": """
38
- The AMI Meeting Corpus is a multi-modal data set consisting of 100 hours of meeting recordings.
39
- The AMI Meeting Corpus consists of 100 hours of meeting recordings. The recordings use a range of signals
40
- synchronized to a common timeline. These include close-talking and far-field microphones, individual and
41
- room-view video cameras, and output from a slide projector and an electronic whiteboard.
42
- """,
43
- "spgispeech": """
44
- The SPGISpeech corpus is derived from company earnings calls manually transcribed by S&P Global, Inc.
45
- according to a professional style guide detailing conventions for capitalization, punctuation, denormalization
46
- of non-standard words and tran- scription of disfluencies in spontaneous speech. The basic unit of SPGISpeech is a
47
- pair consisting of a 5 to 15 second long 16 bit, 16kHz mono wav audio file and its transcription.
48
- """,
49
- "voxpopuli": """
50
- A large-scale multilingual speech corpus for representation learning, semi-supervised learning and interpretation.
51
- The raw data is collected from 2009-2020 European Parliament event recordings.
52
- """,
53
- "tedlium": """
54
- The TED-LIUM corpus is English-language TED talks, with transcriptions, sampled at 16kHz.
55
- All talks and text are property of TED Conferences LLC.
56
- """,
57
- "gigaspeech": """
58
- GigaSpeech is an evolving, multi-domain English speech recognition corpus with 10,000 hours of high quality
59
- labeled audio suitable for supervised training, and 40,000 hours of total audio suitable for semi-supervised
60
- and unsupervised training. Around 40,000 hours of transcribed audio is first collected from audiobooks, podcasts
61
- and YouTube, covering both read and spontaneous speaking styles, and a variety of topics, such as arts, science,
62
- sports, etc. A new forced alignment and segmentation pipeline is proposed to create sentence segments suitable
63
- for speech recognition training, and to filter out segments with low-quality transcription. For system training,
64
- GigaSpeech provides five subsets of different sizes, 10h, 250h, 1000h, 2500h, and 10000h.
65
- """,
66
- "librispeech": """
67
- LibriSpeech is a corpus of approximately 1000 hours of read English speech with sampling rate of 16 kHz,
68
- prepared by Vassil Panayotov with the assistance of Daniel Povey. The data is derived from read
69
- audiobooks from the LibriVox project, and has been carefully segmented and aligned.
70
- """,
71
- "common_voice": """
72
- Common Voice is Mozilla's initiative to help teach machines how real people speak.
73
- The Common Voice dataset consists of a unique MP3 and corresponding text file.
74
- """,
75
- "earnings22": """
76
- The Earnings 22 dataset ( also referred to as earnings22 ) is a 119-hour corpus of English-language earnings calls
77
- collected from global companies. The primary purpose is to serve as a benchmark for industrial and academic
78
- automatic speech recognition (ASR) models on real-world accented speech.
79
- """
80
- }
81
-
82
- _CITATIONS = {
83
- "ami": """
84
- @inproceedings{10.1007/11677482_3,
85
- author = {Carletta, Jean and Ashby, Simone and Bourban, Sebastien and Flynn, Mike and Guillemot, Mael and Hain, Thomas
86
- and Kadlec, Jaroslav and Karaiskos, Vasilis and Kraaij, Wessel and Kronenthal, Melissa and Lathoud, Guillaume
87
- and Lincoln, Mike and Lisowska, Agnes and McCowan, Iain and Post, Wilfried and Reidsma, Dennis and Wellner, Pierre},
88
- title = {The AMI Meeting Corpus: A Pre-Announcement},
89
- year = {2005},
90
- isbn = {3540325492},
91
- publisher = {Springer-Verlag},
92
- address = {Berlin, Heidelberg},
93
- url = {https://doi.org/10.1007/11677482_3},
94
- doi = {10.1007/11677482_3},
95
- booktitle = {Proceedings of the Second International Conference on Machine Learning for Multimodal Interaction},
96
- pages = {28–39},
97
- numpages = {12},
98
- location = {Edinburgh, UK},
99
- series = {MLMI'05}
100
- }
101
- """,
102
- "spgispeech": """
103
- @article{2021arXiv210402014O,
104
- author = {{O'Neill}, Patrick K. and {Lavrukhin}, Vitaly and {Majumdar}, Somshubra and {Noroozi}, Vahid and {Zhang}, Yuekai
105
- and {Kuchaiev}, Oleksii and {Balam}, Jagadeesh and {Dovzhenko}, Yuliya and {Freyberg}, Keenan and {Shulman}, Michael D.
106
- and {Ginsburg}, Boris and {Watanabe}, Shinji and {Kucsko}, Georg},
107
- title = "{SPGISpeech: 5,000 hours of transcribed financial audio for fully formatted end-to-end speech recognition}",
108
- journal = {arXiv e-prints},
109
- keywords = {Computer Science - Computation and Language, Electrical Engineering and Systems Science - Audio and Speech Processing},
110
- year = 2021,
111
- month = apr,
112
- eid = {arXiv:2104.02014},
113
- pages = {arXiv:2104.02014},
114
- eprint = {2104.02014},
115
- primaryClass = {cs.CL},
116
- adsurl = {https://ui.adsabs.harvard.edu/abs/2021arXiv210402014O},
117
- adsnote = {Provided by the SAO/NASA Astrophysics Data System}
118
- }
119
- """,
120
- "voxpopuli": """
121
- @inproceedings{wang-etal-2021-voxpopuli,
122
- title = "{V}ox{P}opuli: A Large-Scale Multilingual Speech Corpus for Representation Learning,
123
- Semi-Supervised Learning and Interpretation",
124
- author = "Wang, Changhan and Riviere, Morgane and Lee, Ann and Wu, Anne and Talnikar, Chaitanya and Haziza,
125
- Daniel and Williamson, Mary and Pino, Juan and Dupoux, Emmanuel",
126
- booktitle = "Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th
127
- International Joint Conference on Natural Language Processing (Volume 1: Long Papers)",
128
- month = aug,
129
- year = "2021",
130
- publisher = "Association for Computational Linguistics",
131
- url = "https://aclanthology.org/2021.acl-long.80",
132
- doi = "10.18653/v1/2021.acl-long.80",
133
- pages = "993--1003",
134
- }
135
- """,
136
- "tedlium": """
137
- @inproceedings{hernandez2018tedlium3,
138
- title={TED-LIUM 3: twice as much data and corpus repartition for experiments on speaker adaptation},
139
- author={Hernandez, Fran{\\c{c}}ois and Nguyen, Vincent and Ghannay, Sahar and Tomashenko, Natalia and Est{\\`e}ve, Yannick},
140
- booktitle={International Conference on Speech and Computer},
141
- pages={198--208},
142
- year={2018},
143
- organization={Springer}
144
- }
145
- """,
146
- "gigaspeech": """
147
- @article{DBLP:journals/corr/abs-2106-06909,
148
- author = {Guoguo Chen and Shuzhou Chai and Guanbo Wang and Jiayu Du and Wei{-}Qiang Zhang and Chao Weng and Dan Su
149
- and Daniel Povey and Jan Trmal and Junbo Zhang and Mingjie Jin and Sanjeev Khudanpur and Shinji Watanabe and
150
- Shuaijiang Zhao and Wei Zou and Xiangang Li and Xuchen Yao and Yongqing Wang and Yujun Wang and Zhao You and Zhiyong Yan},
151
- title = {GigaSpeech: An Evolving, Multi-domain {ASR} Corpus with 10, 000 Hours
152
- of Transcribed Audio},
153
- journal = {CoRR},
154
- volume = {abs/2106.06909},
155
- year = {2021},
156
- url = {https://arxiv.org/abs/2106.06909},
157
- eprinttype = {arXiv},
158
- eprint = {2106.06909},
159
- timestamp = {Wed, 29 Dec 2021 14:29:26 +0100},
160
- biburl = {https://dblp.org/rec/journals/corr/abs-2106-06909.bib},
161
- bibsource = {dblp computer science bibliography, https://dblp.org}
162
- }
163
- """,
164
- "librispeech": """
165
- @inproceedings{panayotov2015librispeech,
166
- title={Librispeech: an ASR corpus based on public domain audio books},
167
- author={Panayotov, Vassil and Chen, Guoguo and Povey, Daniel and Khudanpur, Sanjeev},
168
- booktitle={Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on},
169
- pages={5206--5210},
170
- year={2015},
171
- organization={IEEE}
172
- }
173
- """,
174
- "common_voice": """
175
- @inproceedings{commonvoice:2020,
176
- author = {Ardila, R. and Branson, M. and Davis, K. and Henretty, M. and Kohler, M. and Meyer, J. and Morais, R. and Saunders, L. and Tyers, F. M. and Weber, G.},
177
- title = {Common Voice: A Massively-Multilingual Speech Corpus},
178
- booktitle = {Proceedings of the 12th Conference on Language Resources and Evaluation (LREC 2020)},
179
- pages = {4211--4215},
180
- year = 2020
181
- }
182
- """,
183
- "earnings22": """
184
- @misc{https://doi.org/10.48550/arxiv.2203.15591,
185
- doi = {10.48550/ARXIV.2203.15591},
186
- url = {https://arxiv.org/abs/2203.15591},
187
- author = {Del Rio, Miguel and Ha, Peter and McNamara, Quinten and Miller, Corey and Chandra, Shipra},
188
- keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
189
- title = {Earnings-22: A Practical Benchmark for Accents in the Wild},
190
- publisher = {arXiv},
191
- year = {2022},
192
- copyright = {Creative Commons Attribution Share Alike 4.0 International}
193
- }
194
- """,
195
- }
196
-
197
- _HOMEPAGE_URLS = {
198
- "ami": "https://groups.inf.ed.ac.uk/ami/corpus/",
199
- "spgispeech": "https://datasets.kensho.com/datasets/spgispeech",
200
- "voxpopuli": "https://github.com/facebookresearch/voxpopuli",
201
- "tedlium": "https://www.openslr.org/51/",
202
- "gigaspeech": "https://github.com/SpeechColab/GigaSpeech",
203
- "librispeech": "http://www.openslr.org/12",
204
- "common_voice": "https://commonvoice.mozilla.org/en/datasets",
205
- "earnings22": "https://github.com/revdotcom/speech-datasets/tree/main/earnings22",
206
- }
207
-
208
- _LICENSES = {
209
- "ami": "CC BY 4.0",
210
- "spgispeech": "Custom license (academic use only)",
211
- "voxpopuli": "CC0, also see https://www.europarl.europa.eu/legal-notice/en/",
212
- "tedlium": "Creative Commons BY-NC-ND 3.0 (http://creativecommons.org/licenses/by-nc-nd/3.0/deed.en)",
213
- "gigaspeech": "Apache License 2.0",
214
- "librispeech": "CC BY 4.0",
215
- "common_voice": "Mozilla Public License 2.0 (https://github.com/common-voice/common-voice/blob/main/LICENSE)",
216
- "earnings22": "CC BY-SA 4.0",
217
- }
218
-
219
- _DATASET_TO_CONFIGS = {
220
- "spgispeech": ["l", "s", "m"],
221
- "gigaspeech": ["l", "xs", "s", "m", "xl"],
222
- "librispeech": ["default", "clean.100", "clean.360", "other.500"],
223
- }
224
-
225
- _ALL_CONFIGS = list(_DATASET_TO_CONFIGS) + ["earnings22", "ami", "tedlium", "voxpopuli", "common_voice"]
226
-
227
-
228
- class ESCConfig(datasets.BuilderConfig):
229
- """BuilderConfig for ESC benchmark dataset. """
230
-
231
- def __init__(self, name, subconfig, description, citation, homepage, license, **kwargs):
232
- """
233
- Args:
234
- name: `string`, name of a dataset to be downloaded (for example, "gigaspeech")
235
- subconfig: `string`, specific configuration of a dataset, relevant for "spgispeech", "gigaspeech", and "librispeech"
236
- description: `string`: dataset decsription
237
- citation: `string`: dataset citation
238
- homepage: `string`: dataset homepage
239
- license: `string`: dataset license
240
- **kwargs: keyword arguments forwarded to super.
241
- """
242
- if name in _DATASET_TO_CONFIGS:
243
- # first config is the default one
244
- self.subconfig = _DATASET_TO_CONFIGS[name][0] if subconfig == "default" else subconfig
245
- else:
246
- self.subconfig = None
247
-
248
- super(ESCConfig, self).__init__(
249
- name=name,
250
- version=datasets.Version("1.0.0", ""),
251
- **kwargs
252
- )
253
- self.description = description
254
- self.citation = citation
255
- self.homepage = homepage
256
- self.license = license
257
-
258
-
259
- def _build_config(name, subconfig):
260
- return ESCConfig(
261
- name=name,
262
- subconfig=subconfig,
263
- description=_DESCRIPTIONS[name],
264
- citation=_CITATIONS[name],
265
- homepage=_HOMEPAGE_URLS[name],
266
- license=_LICENSES[name],
267
- )
268
-
269
-
270
- class ESCDatasets(datasets.GeneratorBasedBuilder):
271
- """ESC benchmark dataset dataset."""
272
-
273
- DEFAULT_WRITER_BATCH_SIZE = 256
274
- BUILDER_CONFIGS = [
275
- _build_config(name, subconfig="default") for name in _ALL_CONFIGS
276
- ]
277
-
278
- def _info(self):
279
- features = datasets.Features(
280
- {
281
- "audio": datasets.Audio(sampling_rate=16_000),
282
- "dataset": datasets.Value("string"),
283
- "text": datasets.Value("string"),
284
- "id": datasets.Value("string"),
285
- }
286
- )
287
- return datasets.DatasetInfo( # TODO: add benchmark's own license and description
288
- features=features,
289
- description=self.config.description,
290
- homepage=self.config.homepage,
291
- license=self.config.license,
292
- citation=self.config.citation,
293
- )
294
-
295
- def _split_generators(self, dl_manager):
296
- if self.config.name == "ami":
297
- return self._ami_split_generators(dl_manager)
298
- elif self.config.name == "spgispeech":
299
- return self._spgispeech_split_generators(dl_manager)
300
- elif self.config.name == "voxpopuli":
301
- return self._voxpopuli_split_generators(dl_manager)
302
- elif self.config.name == "tedlium":
303
- return self._tedlium_split_generators(dl_manager)
304
- elif self.config.name == "gigaspeech":
305
- return self._gigaspeech_split_generators(dl_manager)
306
- elif self.config.name == "librispeech":
307
- return self._librispeech_split_generators(dl_manager)
308
- elif self.config.name == "common_voice":
309
- return self._common_voice_split_generators(dl_manager)
310
- elif self.config.name == "earnings22":
311
- return self._earnings_split_generators(dl_manager)
312
-
313
- def _generate_examples(self, *args, **kwargs):
314
- if self.config.name == "ami":
315
- yield from self._ami_generate_examples(*args, **kwargs)
316
- elif self.config.name == "spgispeech":
317
- yield from self._spgispeech_generate_examples(*args, **kwargs)
318
- elif self.config.name == "voxpopuli":
319
- yield from self._voxpopuli_generate_examples(*args, **kwargs)
320
- elif self.config.name == "tedlium":
321
- yield from self._tedlium_generate_examples(*args, **kwargs)
322
- elif self.config.name == "gigaspeech":
323
- yield from self._gigaspeech_generate_examples(*args, **kwargs)
324
- elif self.config.name == "librispeech":
325
- yield from self._librispeech_generate_examples(*args, **kwargs)
326
- elif self.config.name == "common_voice":
327
- yield from self._common_voice_generate_examples(*args, **kwargs)
328
- elif self.config.name == "earnings22":
329
- yield from self._earnings_generate_examples(*args, **kwargs)
330
-
331
- def _ami_split_generators(self, dl_manager):
332
- splits = ["train", "dev", "eval"]
333
-
334
- audio_archives_urls = {}
335
- for split in splits:
336
- audio_archives_urls[split] = [
337
- _AMI_AUDIO_ARCHIVE_URL.format(split=split, _id=m) for m in _AMI_SAMPLE_IDS[split]
338
- ]
339
-
340
- audio_archives = dl_manager.download(audio_archives_urls)
341
- local_extracted_archives_paths = dl_manager.extract(audio_archives) if not dl_manager.is_streaming else {
342
- split: [None] * len(audio_archives[split]) for split in splits
343
- }
344
-
345
- annotations_urls = {split: _AMI_ANNOTATIONS_ARCHIVE_URL.format(split=split) for split in splits}
346
- annotations = dl_manager.download(annotations_urls)
347
-
348
- return [
349
- datasets.SplitGenerator(
350
- name=datasets.Split.TRAIN,
351
- gen_kwargs={
352
- "audio_archives": [dl_manager.iter_archive(archive) for archive in audio_archives["train"]],
353
- "local_extracted_archives_paths": local_extracted_archives_paths["train"],
354
- "annotation": annotations["train"],
355
- "split": "train"
356
- },
357
- ),
358
- datasets.SplitGenerator(
359
- name=datasets.Split.VALIDATION,
360
- gen_kwargs={
361
- "audio_archives": [dl_manager.iter_archive(archive) for archive in audio_archives["dev"]],
362
- "local_extracted_archives_paths": local_extracted_archives_paths["dev"],
363
- "annotation": annotations["dev"],
364
- "split": "dev"
365
- },
366
- ),
367
- datasets.SplitGenerator(
368
- name=datasets.Split.TEST,
369
- gen_kwargs={
370
- "audio_archives": [dl_manager.iter_archive(archive) for archive in audio_archives["eval"]],
371
- "local_extracted_archives_paths": local_extracted_archives_paths["eval"],
372
- "annotation": annotations["eval"],
373
- "split": "eval"
374
- },
375
- ),
376
- ]
377
-
378
- def _ami_generate_examples(self, audio_archives, local_extracted_archives_paths, annotation, split):
379
- assert len(audio_archives) == len(local_extracted_archives_paths)
380
-
381
- with open(annotation, "r", encoding="utf-8") as f:
382
- transcriptions = {}
383
- for line in f.readlines():
384
- line_items = line.strip().split()
385
- _id = line_items[0]
386
- text = " ".join(line_items[1:])
387
- _, meeting_id, microphone_id, speaker_id, begin_time, end_time = _id.split("_")
388
- audio_filename = "_".join([split, _id.lower()]) + ".wav"
389
-
390
- transcriptions[audio_filename] = {
391
- "id": _id,
392
- "text": text if split != "eval" else "",
393
- }
394
-
395
- features = ["id", "text"]
396
- for archive, local_archive_path in zip(audio_archives, local_extracted_archives_paths):
397
- for audio_path, audio_file in archive:
398
- # audio_path is like 'EN2001a/train_ami_en2001a_h00_mee068_0414915_0415078.wav'
399
- audio_meta = transcriptions[audio_path.split("/")[-1]]
400
-
401
- yield audio_path, {
402
- "audio": {
403
- "path": os.path.join(local_archive_path, audio_path) if local_archive_path else audio_path,
404
- "bytes": audio_file.read(),
405
- },
406
- "dataset": "ami",
407
- **{feature: audio_meta[feature] for feature in features}
408
- }
409
-
410
- def _spgispeech_split_generators(self, dl_manager):
411
- subconfig = self.config.subconfig
412
- subsets = [subconfig] + ["dev", "test"]
413
-
414
- meta_path = dl_manager.download_and_extract(
415
- {subset: os.path.join(_SPGISPEECH_META_BASE_URL, _SPGISPEECH_META_FILENAMES[subset]) for subset in subsets}
416
- )
417
-
418
- archive_urls = defaultdict(list)
419
- for subset in subsets:
420
- for subset_dir in _SPGISPEECH_SUBSET_TO_DIR[subset]:
421
- for archive_name in _SPGISPEECH_AUDIO_ARCHIVES_NAMES[subset_dir]:
422
- archive_urls[subset].append(os.path.join(_SPGISPEECH_AUDIO_BASE_URL, subset_dir, archive_name))
423
-
424
- archive_paths = dl_manager.download(archive_urls)
425
-
426
- local_extracted_archive_paths = (
427
- dl_manager.extract(archive_paths)
428
- if not dl_manager.is_streaming
429
- else {subset: [None] * len(archive_paths[subset]) for subset in subsets}
430
- )
431
-
432
- return [
433
- datasets.SplitGenerator(
434
- name=datasets.Split.TRAIN,
435
- gen_kwargs={
436
- "local_extracted_archive_paths": local_extracted_archive_paths[subconfig],
437
- "archives": [dl_manager.iter_archive(path) for path in archive_paths[subconfig]],
438
- "meta_path": meta_path[subconfig],
439
- "is_test": False,
440
- },
441
- ),
442
- datasets.SplitGenerator(
443
- name=datasets.Split.VALIDATION,
444
- gen_kwargs={
445
- "local_extracted_archive_paths": local_extracted_archive_paths["dev"],
446
- "archives": [dl_manager.iter_archive(path) for path in archive_paths["dev"]],
447
- "meta_path": meta_path["dev"],
448
- "is_test": False,
449
- },
450
- ),
451
- datasets.SplitGenerator(
452
- name=datasets.Split.TEST,
453
- gen_kwargs={
454
- "local_extracted_archive_paths": local_extracted_archive_paths["test"],
455
- "archives": [dl_manager.iter_archive(path) for path in archive_paths["test"]],
456
- "meta_path": meta_path["test"],
457
- "is_test": True,
458
- },
459
- ),
460
- ]
461
-
462
- def _spgispeech_generate_examples(self, local_extracted_archive_paths, archives, meta_path, is_test):
463
- # define the expected metadata dict keys,
464
- # some files have metadata with erroneous entries that we have to filter out
465
- dict_keys = {"id": "wav_filename", "text": "transcript"}
466
-
467
- logging.info("Reading spgispeech metadata")
468
- with open(meta_path, encoding="utf-8") as f:
469
- csvreader = csv.DictReader(f, delimiter="|")
470
- metadata = {x["wav_filename"]: dict((k, x[v]) for k, v in dict_keys.items()) for x in tqdm(csvreader, leave=False)}
471
-
472
- for local_extracted_archive_path, archive in zip(local_extracted_archive_paths, archives):
473
- # Here we iterate over all the files within the TAR archive:
474
- for audio_filename, audio_file in archive:
475
- audio_filename = audio_filename.lstrip("./")
476
- # if an audio file exists locally (i.e. in default, non-streaming mode) set the full path to it
477
- # joining path to directory that the archive was extracted to and audio filename.
478
- path = (
479
- os.path.join(local_extracted_archive_path, audio_filename)
480
- if local_extracted_archive_path
481
- else audio_filename
482
- )
483
- # get the .wav filename by removing the directory path from the audio filename
484
- wav_filename = "/".join(audio_filename.split("/")[-2:])
485
- example = dict(metadata[wav_filename])
486
- if is_test: example["text"] = ""
487
- example["audio"] = {"path": path, "bytes": audio_file.read()}
488
- example["dataset"] = "spgispeech"
489
- yield audio_filename, example
490
-
491
- def _voxpopuli_split_generators(self, dl_manager):
492
- n_shards_path = dl_manager.download_and_extract(_VOXPOPULI_N_SHARDS_FILE)
493
- with open(n_shards_path) as f:
494
- n_shards = json.load(f)["en"] # we use only English language in this benchmark
495
- splits = ["train", "dev", "test"]
496
-
497
- audio_urls = {}
498
- for split in splits:
499
- audio_urls[split] = [
500
- _VOXPOPULI_AUDIO_ARCHIVE_PATH.format(split=split, n_shard=i) for i in range(n_shards[split])
501
- ]
502
-
503
- meta_urls = {
504
- split: _VOXPOPULI_METADATA_PATH.format(split=split) for split in splits
505
- }
506
-
507
- dl_manager.download_config.num_proc = len(audio_urls["train"]) // 4
508
- meta_paths = dl_manager.download_and_extract(meta_urls)
509
- audio_paths = dl_manager.download(audio_urls)
510
-
511
- local_extracted_audio_paths = (
512
- dl_manager.extract(audio_paths) if not dl_manager.is_streaming else
513
- {
514
- split: [None] * len(audio_paths[split]) for split in splits
515
- }
516
- )
517
- return [
518
- datasets.SplitGenerator(
519
- name=datasets.Split.TRAIN,
520
- gen_kwargs={
521
- "audio_archives": [dl_manager.iter_archive(archive) for archive in audio_paths["train"]],
522
- "local_extracted_archives_paths": local_extracted_audio_paths["train"],
523
- "meta_path": meta_paths["train"],
524
- "is_test": False,
525
- }
526
- ),
527
- datasets.SplitGenerator(
528
- name=datasets.Split.VALIDATION,
529
- gen_kwargs={
530
- "audio_archives": [dl_manager.iter_archive(archive) for archive in audio_paths["dev"]],
531
- "local_extracted_archives_paths": local_extracted_audio_paths["dev"],
532
- "meta_path": meta_paths["dev"],
533
- "is_test": False,
534
- }
535
- ),
536
- datasets.SplitGenerator(
537
- name=datasets.Split.TEST,
538
- gen_kwargs={
539
- "audio_archives": [dl_manager.iter_archive(archive) for archive in audio_paths["test"]],
540
- "local_extracted_archives_paths": local_extracted_audio_paths["test"],
541
- "meta_path": meta_paths["test"],
542
- "is_test": True,
543
- }
544
- ),
545
- ]
546
-
547
- def _voxpopuli_generate_examples(self, audio_archives, local_extracted_archives_paths, meta_path, is_test):
548
- assert len(audio_archives) == len(local_extracted_archives_paths)
549
-
550
- logging.info("Reading voxpopuli metadata.")
551
- with open(meta_path) as f:
552
- metadata = {x["id"]: x for x in tqdm(csv.DictReader(f, delimiter="\t"), leave=False)}
553
-
554
- for audio_archive, local_extracted_archive_path in zip(audio_archives, local_extracted_archives_paths):
555
- for audio_filename, audio_file in audio_archive:
556
- audio_id = audio_filename.split(os.sep)[-1].split(".wav")[0]
557
- path = os.path.join(local_extracted_archive_path, audio_filename) if local_extracted_archive_path else audio_filename
558
-
559
- yield audio_id, {
560
- "id": audio_id,
561
- "text": metadata[audio_id]["normalized_text"].lower() if not is_test else "",
562
- "audio": {"path": path, "bytes": audio_file.read()},
563
- "dataset": "voxpopuli",
564
- }
565
-
566
- def _librispeech_split_generators(self, dl_manager):
567
- dev_splits, test_splits = ["dev.clean", "dev.other"], ["test.clean", "test.other"]
568
- train_splits = ["train.clean.100", "train.clean.360", "train.other.500"] \
569
- if self.config.subconfig == "default" else [f"train.{self.config.subconfig}"]
570
- dl_urls = {config_name: _LIBRISPEECH_DL_URLS[config_name] for config_name in train_splits + dev_splits + test_splits}
571
- archive_paths = dl_manager.download(dl_urls)
572
- # (Optional) In non-streaming mode, we can extract the archive locally to have actual local audio files:
573
- local_extracted_archives = dl_manager.extract(archive_paths) if not dl_manager.is_streaming else {}
574
- train_split = [
575
- datasets.SplitGenerator(
576
- name="train",
577
- gen_kwargs={
578
- "local_extracted_archives": [local_extracted_archives.get(train_name) for train_name in train_splits],
579
- "archives": [dl_manager.iter_archive(archive_paths[train_name]) for train_name in train_splits],
580
- "is_test": False,
581
- },
582
- )
583
- ]
584
- dev_splits = [
585
- datasets.SplitGenerator(
586
- name="validation.clean",
587
- gen_kwargs={
588
- "local_extracted_archives": [local_extracted_archives.get("dev.clean")],
589
- "archives": [dl_manager.iter_archive(archive_paths["dev.clean"])],
590
- "is_test": False,
591
- },
592
- ),
593
- datasets.SplitGenerator(
594
- name="validation.other",
595
- gen_kwargs={
596
- "local_extracted_archives": [local_extracted_archives.get("dev.other")],
597
- "archives": [dl_manager.iter_archive(archive_paths["dev.other"])],
598
- "is_test": False,
599
- },
600
- ),
601
- ]
602
- test_splits = [
603
- datasets.SplitGenerator(
604
- name="test.clean",
605
- gen_kwargs={
606
- "local_extracted_archives": [local_extracted_archives.get("test.clean")],
607
- "archives": [dl_manager.iter_archive(archive_paths["test.clean"])],
608
- "is_test": True,
609
- },
610
- ),
611
- datasets.SplitGenerator(
612
- name="test.other",
613
- gen_kwargs={
614
- "local_extracted_archives": [local_extracted_archives.get("test.other")],
615
- "archives": [dl_manager.iter_archive(archive_paths["test.other"])],
616
- "is_test": True,
617
- },
618
- ),
619
- ]
620
- return train_split + dev_splits + test_splits
621
-
622
- def _librispeech_generate_examples(self, archives, local_extracted_archives, is_test):
623
- key = 0
624
- audio_data = {}
625
- transcripts = []
626
- for archive, local_extracted_archive in zip(archives, local_extracted_archives):
627
- for path, f in archive:
628
- if path.endswith(".flac"):
629
- id_ = path.split("/")[-1][: -len(".flac")]
630
- audio_data[id_] = f.read()
631
- elif path.endswith(".trans.txt"):
632
- for line in f:
633
- if line:
634
- line = line.decode("utf-8").strip()
635
- id_, transcript = line.split(" ", 1)
636
-
637
- # Error correction
638
- transcript = transcript.lower()
639
-
640
- audio_file = f"{id_}.flac"
641
- audio_file = (
642
- os.path.join(local_extracted_archive, audio_file)
643
- if local_extracted_archive
644
- else audio_file
645
- )
646
- transcripts.append(
647
- {
648
- "id": id_,
649
- "file": audio_file,
650
- "text": transcript if not is_test else "",
651
- }
652
- )
653
- if audio_data and len(audio_data) == len(transcripts):
654
- for transcript in transcripts:
655
- audio = {"path": transcript["file"], "bytes": audio_data[transcript["id"]]}
656
- del transcript["file"]
657
- yield key, {"audio": audio, "dataset": "librispeech", **transcript}
658
- key += 1
659
- audio_data = {}
660
- transcripts = []
661
-
662
- def _common_voice_get_bundle_url(self, locale, url_template):
663
- # path = encodeURIComponent(path)
664
- path = url_template.replace("{locale}", locale)
665
- path = urllib.parse.quote(path.encode("utf-8"), safe="~()*!.'")
666
- # use_cdn = self.config.size_bytes < 20 * 1024 * 1024 * 1024
667
- # response = requests.get(f"{_API_URL}/bucket/dataset/{path}/{use_cdn}", timeout=10.0).json()
668
- response = requests.get(f"{_COMMON_VOICE_API_URL}/bucket/dataset/{path}", timeout=10.0).json()
669
- return response["url"]
670
-
671
- def _common_voice_log_download(self, locale, bundle_version, auth_token):
672
- if isinstance(auth_token, bool):
673
- auth_token = HfFolder().get_token()
674
- whoami = HfApi().whoami(auth_token)
675
- email = whoami["email"] if "email" in whoami else ""
676
- payload = {"email": email, "locale": locale, "dataset": bundle_version}
677
- requests.post(f"{_COMMON_VOICE_API_URL}/{locale}/downloaders", json=payload).json()
678
-
679
- def _common_voice_split_generators(self, dl_manager):
680
- """Returns SplitGenerators."""
681
- hf_auth_token = dl_manager.download_config.use_auth_token
682
- if hf_auth_token is None:
683
- raise ConnectionError(
684
- "Please set use_auth_token=True or use_auth_token='<TOKEN>' to download this dataset"
685
- )
686
-
687
- bundle_version = _COMMON_VOICE_BUNDLE_URL_TEMPLATE.split("/")[0]
688
- dl_manager.download_config.ignore_url_params = True
689
-
690
- self._common_voice_log_download("en", bundle_version, hf_auth_token)
691
- archive_path = dl_manager.download(self._common_voice_get_bundle_url("en", _COMMON_VOICE_BUNDLE_URL_TEMPLATE))
692
- local_extracted_archive = dl_manager.extract(archive_path) if not dl_manager.is_streaming else None
693
-
694
- path_to_data = "/".join([bundle_version, "en"])
695
- path_to_clips = "/".join([path_to_data, "clips"]) if path_to_data else "clips"
696
-
697
- return [
698
- datasets.SplitGenerator(
699
- name=datasets.Split.TRAIN,
700
- gen_kwargs={
701
- "local_extracted_archive": local_extracted_archive,
702
- "archive_iterator": dl_manager.iter_archive(archive_path),
703
- "metadata_filepath": "/".join([path_to_data, "train.tsv"]) if path_to_data else "train.tsv",
704
- "path_to_clips": path_to_clips,
705
- "is_test": False,
706
- },
707
- ),
708
- datasets.SplitGenerator(
709
- name=datasets.Split.VALIDATION,
710
- gen_kwargs={
711
- "local_extracted_archive": local_extracted_archive,
712
- "archive_iterator": dl_manager.iter_archive(archive_path),
713
- "metadata_filepath": "/".join([path_to_data, "dev.tsv"]) if path_to_data else "dev.tsv",
714
- "path_to_clips": path_to_clips,
715
- "is_test": False,
716
- },
717
- ),
718
- datasets.SplitGenerator(
719
- name=datasets.Split.TEST,
720
- gen_kwargs={
721
- "local_extracted_archive": local_extracted_archive,
722
- "archive_iterator": dl_manager.iter_archive(archive_path),
723
- "metadata_filepath": "/".join([path_to_data, "test.tsv"]) if path_to_data else "test.tsv",
724
- "path_to_clips": path_to_clips,
725
- "is_test": True,
726
- },
727
- ),
728
- ]
729
-
730
- def _common_voice_generate_examples(
731
- self,
732
- local_extracted_archive,
733
- archive_iterator,
734
- metadata_filepath,
735
- path_to_clips,
736
- is_test,
737
- ):
738
- """Yields examples."""
739
- data_fields = list(self._info().features.keys())
740
- metadata = {}
741
- metadata_found = False
742
- for path, f in archive_iterator:
743
- if path == metadata_filepath:
744
- metadata_found = True
745
- lines = (line.decode("utf-8") for line in f)
746
- reader = csv.DictReader(lines, delimiter="\t", quoting=csv.QUOTE_NONE)
747
- for row in reader:
748
- # set absolute path for mp3 audio file
749
- if not row["path"].endswith(".mp3"):
750
- row["path"] += ".mp3"
751
- row["path"] = os.path.join(path_to_clips, row["path"])
752
- # accent -> accents in CV 8.0
753
- if "accents" in row:
754
- row["accent"] = row["accents"]
755
- del row["accents"]
756
- # if data is incomplete, fill with empty values
757
- for field in data_fields:
758
- if field not in row:
759
- row[field] = ""
760
- metadata[row["path"]] = row
761
- elif path.startswith(path_to_clips):
762
- assert metadata_found, "Found audio clips before the metadata TSV file."
763
- if not metadata:
764
- break
765
- if path in metadata:
766
- dict_result = dict(metadata[path])
767
- # set the audio feature and the path to the extracted file
768
- path = os.path.join(local_extracted_archive, path) if local_extracted_archive else path
769
- result = {"id": dict_result["client_id"], "dataset": "common_voice",
770
- "audio": {"path": path, "bytes": f.read()}}
771
-
772
- # Error correction
773
- text = dict_result["sentence"]
774
- if text.startswith('"') and text.endswith('"'):
775
- # we can remove trailing quotation marks as they do not affect the transcription
776
- text = text[1:-1]
777
- # replace double quotation marks with single
778
- text = text.replace('""', '"')
779
- result["text"] = text if not is_test else ""
780
-
781
- yield path, result
782
-
783
- def _tedlium_split_generators(self, dl_manager):
784
- archive_path = dl_manager.download(_TEDLIUM_URLS)
785
- # (Optional) In non-streaming mode, we can extract the archive locally to have actual local audio files:
786
- local_extracted_archive = dl_manager.extract(archive_path) if not dl_manager.is_streaming else {}
787
- split_paths = [
788
- (datasets.Split.TRAIN, "train"),
789
- (datasets.Split.VALIDATION, "dev"),
790
- (datasets.Split.TEST, "test"),
791
- ]
792
- splits = []
793
- for split, split_name in split_paths:
794
- kwargs = {
795
- "filepath": [dl_manager.iter_archive(sharded_path) for sharded_path in archive_path[split_name]],
796
- "local_extracted_archive": local_extracted_archive.get(split_name),
797
- "split_path": split_name,
798
- }
799
- splits.append(datasets.SplitGenerator(name=split, gen_kwargs=kwargs))
800
- return splits
801
-
802
- def _tedlium_generate_examples(self, filepath, local_extracted_archive, split_path):
803
- """Generate examples from a TED-LIUM stm file."""
804
- if local_extracted_archive:
805
- for local_archive in local_extracted_archive:
806
- # The stm directory houses the speaker and transcription information in .stm format
807
- split_dir = os.path.join(local_archive, split_path)
808
- stm_files = [os.path.join(split_dir, f) for f in os.listdir(split_dir) if f.endswith(".stm")]
809
- for file in stm_files:
810
- # the .sph speaker file almost always has the same file name as the .stm file
811
- speaker_file = Path(file).stem
812
- audio_file = os.path.join(split_dir, speaker_file + ".sph")
813
- segment, sampling_rate = sf.read(audio_file, dtype=np.int16)
814
- with open(file) as f:
815
- for line in f:
816
- line = line.strip()
817
- fn, channel, speaker, start, end, label, transcript = line.split(" ", 6)
818
- transcript = _maybe_trim_suffix(transcript)
819
-
820
- # Error correction
821
- transcript = transcript.lower()
822
- if transcript in ignore_segments:
823
- continue
824
- # delete the <unk> token from the text
825
- transcript = transcript.replace("<unk>", "")
826
- # replace spaced apostrophes with un-spaced (it 's -> it's)
827
- for contraction in tedlium_contractions:
828
- transcript = transcript.replace(contraction, contraction[1:])
829
- # JIWER compliance (for WER/CER calc.)
830
- # remove multiple spaces
831
- transcript = re.sub(r"\s\s+", " ", transcript)
832
- # strip trailing spaces
833
- transcript = transcript.strip()
834
- if len(transcript) == 0:
835
- continue
836
-
837
- if speaker_file != fn:
838
- # handle the case where the stm file does not have the same file name as the transcript
839
- speaker_file = fn
840
- audio_file = os.path.join(split_dir, speaker_file + ".sph")
841
- segment, sampling_rate = sf.read(audio_file, dtype=np.int16)
842
- samples = _extract_audio_segment(segment, sampling_rate, float(start), float(end))
843
- key = "-".join([speaker, start, end, label])
844
- example = {
845
- "audio": {"path": audio_file, "array": samples, "sampling_rate": sampling_rate},
846
- "text": transcript if split_path != "test" else "",
847
- "id": key,
848
- "dataset": "tedlium",
849
- }
850
- yield key, example
851
-
852
- else:
853
- audio_data = {}
854
- transcripts = defaultdict(list)
855
- for file in filepath:
856
- for path, f in file:
857
- if path.endswith(".sph"):
858
- # get the speaker id
859
- fn = path.split("/")[-1].strip(".sph")
860
- # read the audio data from raw byte form and add key-value pair to dict
861
- audio_data[fn] = sf.read(BytesIO(f.read()), dtype=np.int16)
862
- elif path.endswith(".stm"):
863
- for line in f:
864
- if line:
865
- line = line.decode("utf-8").strip()
866
- fn, channel, speaker, start, end, label, transcript = line.split(" ", 6)
867
- transcript = _maybe_trim_suffix(transcript)
868
-
869
- # Error correction
870
- transcript = transcript.lower()
871
- if transcript in ignore_segments:
872
- continue
873
- # delete the <unk> token from the text
874
- transcript = transcript.replace("<unk>", "")
875
- # replace spaced apostrophes with un-spaced (it 's -> it's)
876
- for contraction in tedlium_contractions:
877
- transcript = transcript.replace(contraction, contraction[1:])
878
- # JIWER compliance (for WER/CER calc.)
879
- # remove multiple spaces
880
- transcript = re.sub(r"\s\s+", " ", transcript)
881
- # strip trailing spaces
882
- transcript = transcript.strip()
883
- if len(transcript) == 0:
884
- continue
885
-
886
- audio_file = path.replace("stm", "sph")
887
- key = "-".join([speaker, start, end, label])
888
- # append metadata information to the dict of transcripts for the associated speaker
889
- transcripts[fn].append(
890
- {
891
- "text": transcript,
892
- "file": audio_file,
893
- "id": key,
894
- "start": start,
895
- "end": end,
896
- "channel": channel,
897
- "fn": fn,
898
- }
899
- )
900
-
901
- if audio_data and audio_data.keys() == transcripts.keys():
902
- for fn, speaker in transcripts.items():
903
- for transcript in speaker:
904
- segment, sampling_rate = audio_data[transcript["fn"]]
905
- samples = _extract_audio_segment(
906
- segment,
907
- sampling_rate,
908
- float(transcript["start"]),
909
- float(transcript["end"]),
910
- )
911
- audio = {"path": transcript["file"], "array": samples,
912
- "sampling_rate": sampling_rate}
913
- key = transcript["id"]
914
- yield key, {
915
- "audio": audio,
916
- "text": transcript["text"] if split_path != "test" else "",
917
- "dataset": "tedlium",
918
- "id": transcript["id"],
919
- }
920
- audio_data = {}
921
- transcripts = defaultdict(list)
922
-
923
- def _gigaspeech_split_generators(self, dl_manager):
924
- splits_to_configs = {
925
- "train": _GIGASPEECH_CONFIGS_TO_ALL_CONFIGS[self.config.subconfig],
926
- "dev": ["dev"],
927
- "test": ["test"],
928
- }
929
-
930
- # 1. prepare sharded archives with audio files
931
- audio_archives_urls = defaultdict(list)
932
- for split, subsets in splits_to_configs.items():
933
- for subset in subsets:
934
- audio_archives_urls[split].extend(
935
- [
936
- _GIGASPEECH_AUDIO_ARCHIVE_URL.format(subset=subset, is_additional=_is_additional(subset),
937
- archive_id=i)
938
- for i in range(_GIGASPEECH_N_ARCHIVES[subset])
939
- ]
940
- )
941
- audio_archives_paths = dl_manager.download(audio_archives_urls)
942
- local_audio_archives_paths = dl_manager.extract(audio_archives_paths) if not dl_manager.is_streaming \
943
- else {}
944
-
945
- # 2. prepare sharded metadata csv files
946
- meta_urls = defaultdict(list)
947
- for split, subsets in splits_to_configs.items():
948
- for subset in subsets:
949
- meta_urls[split].extend(
950
- [
951
- _GIGASPEECH_META_URL.format(subset=subset, is_additional=_is_additional(subset), archive_id=i)
952
- for i in range(_GIGASPEECH_N_ARCHIVES[subset])
953
- ]
954
- )
955
- meta_paths = dl_manager.download_and_extract(meta_urls)
956
-
957
- return [
958
- datasets.SplitGenerator(
959
- name=datasets.Split.TRAIN,
960
- gen_kwargs={
961
- "audio_archives_iterators": [
962
- dl_manager.iter_archive(archive_path) for archive_path in audio_archives_paths["train"]
963
- ],
964
- "local_audio_archives_paths": local_audio_archives_paths.get("train"),
965
- "meta_paths": meta_paths["train"],
966
- "is_test": False,
967
- },
968
- ),
969
- datasets.SplitGenerator(
970
- name=datasets.Split.VALIDATION,
971
- gen_kwargs={
972
- "audio_archives_iterators": [
973
- dl_manager.iter_archive(archive_path) for archive_path in audio_archives_paths["dev"]
974
- ],
975
- "local_audio_archives_paths": local_audio_archives_paths.get("dev"),
976
- "meta_paths": meta_paths["dev"],
977
- "is_test": False,
978
- },
979
- ),
980
- datasets.SplitGenerator(
981
- name=datasets.Split.TEST,
982
- gen_kwargs={
983
- "audio_archives_iterators": [
984
- dl_manager.iter_archive(archive_path) for archive_path in audio_archives_paths["test"]
985
- ],
986
- "local_audio_archives_paths": local_audio_archives_paths.get("test"),
987
- "meta_paths": meta_paths["test"],
988
- "is_test": True,
989
- },
990
- ),
991
- ]
992
-
993
- def _gigaspeech_generate_examples(self, audio_archives_iterators, local_audio_archives_paths, meta_paths, is_test):
994
- assert len(audio_archives_iterators) == len(meta_paths)
995
- if local_audio_archives_paths:
996
- assert len(audio_archives_iterators) == len(local_audio_archives_paths)
997
-
998
- for i, (meta_path, audio_archive_iterator) in enumerate(zip(meta_paths, audio_archives_iterators)):
999
- meta_dict = dict()
1000
- with open(meta_path) as csvfile:
1001
- meta_csv = csv.DictReader(csvfile)
1002
- for line in meta_csv:
1003
- meta_dict[line["sid"]] = line
1004
-
1005
- for audio_path_in_archive, audio_file in audio_archive_iterator:
1006
- # `audio_path_in_archive` is like "dev_chunks_0000/YOU1000000029_S0000095.wav"
1007
- audio_filename = os.path.split(audio_path_in_archive)[1]
1008
- audio_id = audio_filename.split(".wav")[0]
1009
- audio_meta = meta_dict[audio_id]
1010
- audio_meta["id"] = audio_meta.pop("sid")
1011
- text = audio_meta.pop("text_tn")
1012
-
1013
- # Error correction
1014
- text = text.lower()
1015
- if text in ignore_segments:
1016
- continue
1017
- for junk_token in gigaspeech_junk_tokens:
1018
- text = text.replace(junk_token, "")
1019
- # convert spelled out punctuation to symbolic form
1020
- for punctuation, replacement in gigaspeech_punctuation.items():
1021
- text = text.replace(punctuation, replacement)
1022
- # JIWER compliance (for WER/CER calc.)
1023
- # remove multiple spaces
1024
- text = re.sub(r"\s\s+", " ", text)
1025
- # strip trailing spaces
1026
- text = text.strip()
1027
- if len(text) == 0:
1028
- continue
1029
-
1030
- audio_meta["text"] = text if not is_test else ""
1031
-
1032
- path = os.path.join(local_audio_archives_paths[i], audio_path_in_archive) if local_audio_archives_paths \
1033
- else audio_path_in_archive
1034
-
1035
- yield audio_id, {
1036
- "audio": {"path": path, "bytes": audio_file.read()},
1037
- "dataset": "gigaspeech",
1038
- **{feature: value for feature, value in audio_meta.items() if feature in self.info.features}
1039
- }
1040
-
1041
- def _earnings_split_generators(self, dl_manager):
1042
- meta_url = _EARNINGS_BASE_URL + "metadata.csv"
1043
- meta_path = dl_manager.download_and_extract(meta_url)
1044
-
1045
- with open(meta_path, encoding="utf-8") as f:
1046
- csvreader = csv.DictReader(f, delimiter=",")
1047
- metadata, all_ids = {}, set()
1048
- for row in csvreader:
1049
- all_ids.update([row["source_id"]])
1050
- metadata[row["file"]] = row["sentence"] # we need only text in this benchmark
1051
-
1052
- train_ids = all_ids - _EARNINGS_DEV_IDS - _EARNINGS_TEST_IDS
1053
- split_to_ids = {"train": train_ids, "dev": _EARNINGS_DEV_IDS, "test": _EARNINGS_TEST_IDS}
1054
-
1055
- dl_urls = {}
1056
- for split, split_ids in split_to_ids.items():
1057
- dl_urls[split] = [_EARNINGS_BASE_URL + f"data/{source_id}.tar.gz" for source_id in split_ids]
1058
- archive_paths = dl_manager.download(dl_urls)
1059
-
1060
- local_extracted_archive_paths = (
1061
- dl_manager.extract(archive_paths)
1062
- if not dl_manager.is_streaming
1063
- else {split: [None] * len(archive_paths[split]) for split in ["train", "dev", "test"]}
1064
- )
1065
-
1066
- return [
1067
- datasets.SplitGenerator(
1068
- name=datasets.Split.TRAIN,
1069
- gen_kwargs={
1070
- "local_extracted_archive_paths": local_extracted_archive_paths["train"],
1071
- "archives": [dl_manager.iter_archive(path) for path in archive_paths["train"]],
1072
- "metadata": metadata,
1073
- "is_test": False,
1074
- },
1075
- ),
1076
- datasets.SplitGenerator(
1077
- name=datasets.Split.VALIDATION,
1078
- gen_kwargs={
1079
- "local_extracted_archive_paths": local_extracted_archive_paths["dev"],
1080
- "archives": [dl_manager.iter_archive(path) for path in archive_paths["dev"]],
1081
- "metadata": metadata,
1082
- "is_test": False,
1083
- },
1084
- ),
1085
- datasets.SplitGenerator(
1086
- name=datasets.Split.TEST,
1087
- gen_kwargs={
1088
- "local_extracted_archive_paths": local_extracted_archive_paths["test"],
1089
- "archives": [dl_manager.iter_archive(path) for path in archive_paths["test"]],
1090
- "metadata": metadata,
1091
- "is_test": True,
1092
- },
1093
- ),
1094
- ]
1095
-
1096
- def _earnings_generate_examples(self, local_extracted_archive_paths, archives, metadata, is_test):
1097
- for local_extracted_archive_path, archive in zip(local_extracted_archive_paths, archives):
1098
- # Here we iterate over all the files within the TAR archive:
1099
- for audio_filename, audio_file in archive:
1100
- audio_filename = audio_filename.lstrip("./")
1101
- # if an audio file exists locally (i.e. in default, non-streaming mode) set the full path to it
1102
- # joining path to directory that the archive was extracted to and audio filename.
1103
- path = (
1104
- os.path.join(local_extracted_archive_path, audio_filename)
1105
- if local_extracted_archive_path
1106
- else audio_filename
1107
- )
1108
-
1109
- # Error correction
1110
- text = metadata[audio_filename]
1111
- if text.lower() in ignore_segments:
1112
- continue
1113
- # Remove junk tokens
1114
- for junk_token in earnings_junk_tokens:
1115
- text = text.replace(junk_token, "")
1116
- # JIWER compliance (for WER/CER calc.)
1117
- # remove multiple spaces
1118
- text = re.sub(r"\s\s+", " ", text)
1119
- # strip trailing spaces
1120
- text = text.strip()
1121
- if len(text) == 0:
1122
- continue
1123
-
1124
- yield audio_filename, {
1125
- "id": audio_filename,
1126
- "text": text if not is_test else "",
1127
- "dataset": "earnings22",
1128
- "audio": {"path": path, "bytes": audio_file.read()}
1129
- }
1130
-
1131
-
1132
- def _maybe_trim_suffix(transcript):
1133
- # stm files for the TEDLIUM release 1 train split contain a key (enclosed in
1134
- # parens) at the end.
1135
- splits = transcript.rsplit(" ", 1)
1136
- transcript = splits[0]
1137
- if len(splits) > 1:
1138
- suffix = splits[-1]
1139
- if not suffix.startswith("("):
1140
- transcript += " " + suffix
1141
- return transcript
1142
-
1143
-
1144
- def _extract_audio_segment(segment, sampling_rate, start_sec, end_sec):
1145
- """Extracts segment of audio samples (as an ndarray) from the given segment."""
1146
- # The dataset only contains mono audio.
1147
- start_sample = int(start_sec * sampling_rate)
1148
- end_sample = min(int(end_sec * sampling_rate), segment.shape[0])
1149
- samples = segment[start_sample:end_sample]
1150
- return samples
1151
-
1152
-
1153
- def _parse_gender(label_str):
1154
- """Parse gender string from STM "<label>" field."""
1155
- gender = re.split(",|_", label_str)[-1][:-1]
1156
- # Fix inconsistencies in the data.
1157
- if not gender:
1158
- gender = -1 # Missing label.
1159
- elif gender == "<NA": # In TEDLIUM release 3 training data.
1160
- gender = -1 # Missing label.
1161
- elif gender == "F":
1162
- gender = "female"
1163
- elif gender == "M":
1164
- gender = "male"
1165
- return gender
1166
-
1167
-
1168
- def _is_additional(name):
1169
- if name in {"s", "m", "l", "xl"}:
1170
- return "_additional"
1171
- return ""
1172
-
1173
-
1174
- _AMI_TRAIN_SAMPLE_IDS = [
1175
- "EN2001a",
1176
- "EN2001b",
1177
- "EN2001d",
1178
- "EN2001e",
1179
- "EN2003a",
1180
- "EN2004a",
1181
- "EN2005a",
1182
- "EN2006a",
1183
- "EN2006b",
1184
- "EN2009b",
1185
- "EN2009c",
1186
- "EN2009d",
1187
- "ES2002a",
1188
- "ES2002b",
1189
- "ES2002c",
1190
- "ES2002d",
1191
- "ES2003a",
1192
- "ES2003b",
1193
- "ES2003c",
1194
- "ES2003d",
1195
- "ES2005a",
1196
- "ES2005b",
1197
- "ES2005c",
1198
- "ES2005d",
1199
- "ES2006a",
1200
- "ES2006b",
1201
- "ES2006c",
1202
- "ES2006d",
1203
- "ES2007a",
1204
- "ES2007b",
1205
- "ES2007c",
1206
- "ES2007d",
1207
- "ES2008a",
1208
- "ES2008b",
1209
- "ES2008c",
1210
- "ES2008d",
1211
- "ES2009a",
1212
- "ES2009b",
1213
- "ES2009c",
1214
- "ES2009d",
1215
- "ES2010a",
1216
- "ES2010b",
1217
- "ES2010c",
1218
- "ES2010d",
1219
- "ES2012a",
1220
- "ES2012b",
1221
- "ES2012c",
1222
- "ES2012d",
1223
- "ES2013a",
1224
- "ES2013b",
1225
- "ES2013c",
1226
- "ES2013d",
1227
- "ES2014a",
1228
- "ES2014b",
1229
- "ES2014c",
1230
- "ES2014d",
1231
- "ES2015a",
1232
- "ES2015b",
1233
- "ES2015c",
1234
- "ES2015d",
1235
- "ES2016a",
1236
- "ES2016b",
1237
- "ES2016c",
1238
- "ES2016d",
1239
- "IB4005",
1240
- "IN1001",
1241
- "IN1002",
1242
- "IN1005",
1243
- "IN1007",
1244
- "IN1008",
1245
- "IN1009",
1246
- "IN1012",
1247
- "IN1013",
1248
- "IN1014",
1249
- "IN1016",
1250
- "IS1000a",
1251
- "IS1000b",
1252
- "IS1000c",
1253
- "IS1000d",
1254
- "IS1001a",
1255
- "IS1001b",
1256
- "IS1001c",
1257
- "IS1001d",
1258
- "IS1002b",
1259
- "IS1002c",
1260
- "IS1002d",
1261
- "IS1003a",
1262
- "IS1003b",
1263
- "IS1003c",
1264
- "IS1003d",
1265
- "IS1004a",
1266
- "IS1004b",
1267
- "IS1004c",
1268
- "IS1004d",
1269
- "IS1005a",
1270
- "IS1005b",
1271
- "IS1005c",
1272
- "IS1006a",
1273
- "IS1006b",
1274
- "IS1006c",
1275
- "IS1006d",
1276
- "IS1007a",
1277
- "IS1007b",
1278
- "IS1007c",
1279
- "IS1007d",
1280
- "TS3005a",
1281
- "TS3005b",
1282
- "TS3005c",
1283
- "TS3005d",
1284
- "TS3006a",
1285
- "TS3006b",
1286
- "TS3006c",
1287
- "TS3006d",
1288
- "TS3007a",
1289
- "TS3007b",
1290
- "TS3007c",
1291
- "TS3007d",
1292
- "TS3008a",
1293
- "TS3008b",
1294
- "TS3008c",
1295
- "TS3008d",
1296
- "TS3009a",
1297
- "TS3009b",
1298
- "TS3009c",
1299
- "TS3009d",
1300
- "TS3010a",
1301
- "TS3010b",
1302
- "TS3010c",
1303
- "TS3010d",
1304
- "TS3011a",
1305
- "TS3011b",
1306
- "TS3011c",
1307
- "TS3011d",
1308
- "TS3012a",
1309
- "TS3012b",
1310
- "TS3012c",
1311
- "TS3012d",
1312
- ]
1313
-
1314
- _AMI_VALIDATION_SAMPLE_IDS = [
1315
- "ES2011a",
1316
- "ES2011c",
1317
- "IB4001",
1318
- "IB4003",
1319
- "IB4010",
1320
- "IS1008a",
1321
- "IS1008c",
1322
- "TS3004a",
1323
- "TS3004c",
1324
- "ES2011b",
1325
- "ES2011d",
1326
- "IB4002",
1327
- "IB4004",
1328
- "IB4011",
1329
- "IS1008b",
1330
- "IS1008d",
1331
- "TS3004b",
1332
- "TS3004d",
1333
- ]
1334
-
1335
- _AMI_EVAL_SAMPLE_IDS = [
1336
- "EN2002a",
1337
- "EN2002b",
1338
- "EN2002c",
1339
- "EN2002d",
1340
- "ES2004a",
1341
- "ES2004b",
1342
- "ES2004c",
1343
- "ES2004d",
1344
- "IS1009a",
1345
- "IS1009b",
1346
- "IS1009c",
1347
- "IS1009d",
1348
- "TS3003a",
1349
- "TS3003b",
1350
- "TS3003c",
1351
- "TS3003d",
1352
- ]
1353
-
1354
- _AMI_SAMPLE_IDS = {
1355
- "train": _AMI_TRAIN_SAMPLE_IDS,
1356
- "dev": _AMI_VALIDATION_SAMPLE_IDS,
1357
- "eval": _AMI_EVAL_SAMPLE_IDS,
1358
- }
1359
-
1360
- _AMI_BASE_DATA_URL = "https://huggingface.co/datasets/speech-seq2seq/ami/resolve/main/"
1361
-
1362
- _AMI_AUDIO_ARCHIVE_URL = _AMI_BASE_DATA_URL + "audio/ihm/{split}/{_id}.tar.gz"
1363
-
1364
- _AMI_ANNOTATIONS_ARCHIVE_URL = _AMI_BASE_DATA_URL + "annotations/{split}/text"
1365
-
1366
- _SPGISPEECH_BASE_URL = "https://huggingface.co/datasets/kensho/spgispeech/resolve/main/data/"
1367
-
1368
- _SPGISPEECH_AUDIO_BASE_URL = _SPGISPEECH_BASE_URL + "audio"
1369
-
1370
- _SPGISPEECH_SUBSET_TO_DIR = {
1371
- "s": ["s"],
1372
- "m": ["s", "m_additional"],
1373
- "l": ["s", "m_additional", "l_additional"],
1374
- "dev": ["dev"],
1375
- "test": ["test"],
1376
- }
1377
-
1378
- # the second number in range is the number of archives (shards) in a subset
1379
- _SPGISPEECH_AUDIO_ARCHIVES_NAMES = {
1380
- "s": [f"s_part_{i}.tar.gz" for i in range(0, 6)],
1381
- "m_additional": [f"m_part_{i}.tar.gz" for i in range(0, 21)],
1382
- "l_additional": [f"l_part_{i}.tar.gz" for i in range(0, 103)],
1383
- "dev": [f"dev_part_{i}.tar.gz" for i in range(0, 3)],
1384
- "test": [f"test_part_{i}.tar.gz" for i in range(0, 3)],
1385
- }
1386
-
1387
- _SPGISPEECH_META_BASE_URL = _SPGISPEECH_BASE_URL + "meta"
1388
-
1389
- _SPGISPEECH_META_FILENAMES = {
1390
- "s": "train_small.csv",
1391
- "m": "train_medium.csv",
1392
- "l": "train.csv",
1393
- "dev": "dev.csv",
1394
- "test": "test.csv",
1395
- }
1396
-
1397
- _VOXPOPULI_BASE_DATA_DIR = "https://huggingface.co/datasets/polinaeterna/voxpopuli/resolve/main/data/"
1398
-
1399
- _VOXPOPULI_N_SHARDS_FILE = _VOXPOPULI_BASE_DATA_DIR + "n_files.json"
1400
-
1401
- _VOXPOPULI_AUDIO_ARCHIVE_PATH = _VOXPOPULI_BASE_DATA_DIR + "en/{split}/{split}_part_{n_shard}.tar.gz"
1402
-
1403
- _VOXPOPULI_METADATA_PATH = _VOXPOPULI_BASE_DATA_DIR + "en/asr_{split}.tsv"
1404
-
1405
- _LIBRISPEECH_DL_URL = "http://www.openslr.org/resources/12/"
1406
-
1407
- _LIBRISPEECH_DL_URLS = {
1408
- "dev.clean": _LIBRISPEECH_DL_URL + "dev-clean.tar.gz",
1409
- "dev.other": _LIBRISPEECH_DL_URL + "dev-other.tar.gz",
1410
- "test.clean": _LIBRISPEECH_DL_URL + "test-clean.tar.gz",
1411
- "test.other": _LIBRISPEECH_DL_URL + "test-other.tar.gz",
1412
- "train.clean.100": _LIBRISPEECH_DL_URL + "train-clean-100.tar.gz",
1413
- "train.clean.360": _LIBRISPEECH_DL_URL + "train-clean-360.tar.gz",
1414
- "train.other.500": _LIBRISPEECH_DL_URL + "train-other-500.tar.gz",
1415
- }
1416
-
1417
- _COMMON_VOICE_API_URL = "https://commonvoice.mozilla.org/api/v1"
1418
-
1419
- _COMMON_VOICE_BUNDLE_URL_TEMPLATE = 'cv-corpus-9.0-2022-04-27/cv-corpus-9.0-2022-04-27-{locale}.tar.gz'
1420
-
1421
- _TEDLIUM_BASE_URL = "https://huggingface.co/datasets/LIUM/tedlium/resolve/main/TEDLIUM_release3/legacy/"
1422
-
1423
- _TEDLIUM_URLS = {
1424
- "train": [_TEDLIUM_BASE_URL + "train_1.tar.gz", _TEDLIUM_BASE_URL + "train_2.tar.gz"],
1425
- "dev": [_TEDLIUM_BASE_URL + "dev.tar.gz"],
1426
- "test": [_TEDLIUM_BASE_URL + "test.tar.gz"],
1427
- }
1428
-
1429
- _GIGASPEECH_BASE_DATA_URL = "https://huggingface.co/datasets/speechcolab/gigaspeech/resolve/main/data/"
1430
-
1431
- _GIGASPEECH_AUDIO_ARCHIVE_URL = _GIGASPEECH_BASE_DATA_URL + "audio/{subset}_files{is_additional}/{subset}_chunks_{archive_id:04}.tar.gz"
1432
-
1433
- _GIGASPEECH_META_URL = _GIGASPEECH_BASE_DATA_URL + "metadata/{subset}_metadata{is_additional}/{subset}_chunks_{archive_id:04}_metadata.csv"
1434
-
1435
- _GIGASPEECH_CONFIGS_TO_ALL_CONFIGS = {
1436
- "xs": ["xs"],
1437
- "s": ["xs", "s"],
1438
- "m": ["xs", "s", "m"],
1439
- "l": ["xs", "s", "m", "l"],
1440
- "xl": ["xs", "s", "m", "l", "xl"],
1441
- }
1442
-
1443
- _GIGASPEECH_N_ARCHIVES = {
1444
- "xs": 1,
1445
- "s": 23,
1446
- "m": 69,
1447
- "l": 136,
1448
- "xl": 602,
1449
- "dev": 1,
1450
- "test": 3,
1451
- }
1452
-
1453
- _EARNINGS_BASE_URL = "https://huggingface.co/datasets/anton-l/earnings22_baseline_5_gram/resolve/main/"
1454
-
1455
- _EARNINGS_DEV_IDS = {
1456
- "4420696",
1457
- "4448760",
1458
- "4461799",
1459
- "4469836",
1460
- "4473238",
1461
- "4482110",
1462
- }
1463
- _EARNINGS_TEST_IDS = {
1464
- "4432298",
1465
- "4450488",
1466
- "4470290",
1467
- "4479741",
1468
- "4483338",
1469
- "4485244",
1470
- }
1471
-
1472
-
1473
- tedlium_contractions = [" 's", " 't", " 're", " 've", " 'm", " 'll", " 'd", " 'clock", " 'all"]
1474
- gigaspeech_punctuation = {" <comma>": ",", " <period>": ".", " <questionmark>": "?", " <exclamationpoint>": "!"}
1475
- gigaspeech_junk_tokens = ["<other>", "<sil>"]
1476
- swb_junk_tokens = ["[noise]", "[laughter]", "[silence]", "[vocalized-noise]", "<a_aside>", "<b_aside>", "<e_aside>",
1477
- "[laughter-", "_1", "[laugh]", "[sigh]", "[cough]", "[mn]", "[breath]", "[lipsmack]",
1478
- "[sneeze]", "[skip]", "[pause]", "(%hesitation)", "(%HESITATION)"]
1479
- swb_punctuations = ["{", "}", "[", "]-", "]", "((", "))", "(", ")", "."]
1480
- swb_fillers = r"\b(uh|uhm|um|hmm|mm|mhm|mmm)\b"
1481
- earnings_junk_tokens = ["<noise>", "<crosstalk>", "<affirmative>", "<inaudible>", "inaudible", "<laugh>", "<silence>"]
1482
- ignore_segments = ["ignore_time_segment_in_scoring", "<noise>", "<music>", "[noise]", "[laughter]", "[silence]",
1483
- "[vocalized-noise]", "<crosstalk>", "<affirmative>", "<inaudible>", "<laugh>", ""]
1484
- ignore_segments = ignore_segments + gigaspeech_junk_tokens + swb_junk_tokens + earnings_junk_tokens
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
librispeech/partial/test.clean/0000.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:785ab6a41aaaeb55d048e73413ecc47e1d1298e24f14e51fb5f0d032e6346404
3
+ size 350158103
librispeech/partial/test.other/0000.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8301dc12610184c4193b0245828734f26009397d81c56c89a6a9378b186a43e9
3
+ size 332577077
librispeech/partial/train/0000.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fffc729fdff976fd94e64ebf76e26ba97bfc0e26d93b9b3a1b91fb5d5cdcc499
3
+ size 506328032
librispeech/partial/train/0001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:864f4602ee04d388aa10158568143d2775825d8e36866ed8eef6c7d426526471
3
+ size 495897741
librispeech/partial/train/0002.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e88afc71dc20bda9e31bf8d8fde825aa4ea39de992eba07bb6f89384a43948c
3
+ size 500298280
librispeech/partial/train/0003.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a02da27218060880dc9249e91ce7994d9107cb58c813d7a1a310caaef6897db
3
+ size 502846252
librispeech/partial/train/0004.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e49b2eace2802244161cc31989503a5c3f35940b1d9b7eb3266b0fcb6a222c73
3
+ size 497976358
librispeech/partial/train/0005.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75819150e6033eae63458bdfbb5b01cafaa686a117400ae4aaef9ce8206ecdd4
3
+ size 487105154
librispeech/partial/train/0006.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc3fbc8dd195aabb9ab5abe7e88d4432854f4b1f425b305fadae8f98dc775e92
3
+ size 497462030
librispeech/partial/train/0007.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52ea582c2bee19fded4832585977ac0d176fcb5b65b0ed79d0b62da20354473d
3
+ size 486600903
librispeech/partial/train/0008.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bda2da5cd86237ca7e09006d055a83d76dd2523c2c7ee7c9618829b59b07574f
3
+ size 501139484
librispeech/partial/train/0009.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2384e748daa8643afa51336b371d36d8e0dd47f8e7eec34682555a4206372d63
3
+ size 400561237
librispeech/partial/validation.clean/0000.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7e0ec47c98a81448883b42582dfb919ee82de7ea403593cc73437df76ef646d
3
+ size 341754590
librispeech/partial/validation.other/0000.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6edd3479fdd30fd2966ae024c0a7e932804e96d2ff02119c2068ecb2e4b90f03
3
+ size 318267408