File size: 5,473 Bytes
e074eb9
 
 
 
 
 
 
 
 
39057d5
e074eb9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
39057d5
e074eb9
39057d5
e074eb9
39057d5
 
 
 
 
e074eb9
39057d5
e074eb9
 
 
 
39057d5
e074eb9
 
 
39057d5
e074eb9
 
 
 
 
 
 
 
 
39057d5
e074eb9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
# Dataset Card for Speech Wikimedia

## Table of Contents
- [Dataset Card for Speech Wikimedia](#dataset-card-for-speech-wikimedia)
  - [Table of Contents](#table-of-contents)
  - [Dataset Description](#dataset-description)
    - [Dataset Summary](#dataset-summary)
    - [Transcription languages](#transcription-languages)
    - [Hours of Audio for each language](#hours-of-audio-for-each-language)
    - [Hours of language pairs for speech translation](#hours-of-language-pairs-for-speech-translation)
  - [Dataset Structure](#dataset-structure)
    - [reformat](#reformat)
    - [transcription and transcription_2](#transcription-and-transcription_2)
    - [real_correspondence.json](#real_correspondencejson)
    - [license.json](#licensejson)
    - [Data License](#data-license)
  - [Dataset Creation](#dataset-creation)
    - [Source Data](#source-data)
      - [Initial Data Collection and Normalization](#initial-data-collection-and-normalization)
      - [Preprocessing](#preprocessing)
    - [Annotations](#annotations)
      - [Annotation process](#annotation-process)
    - [Personal and Sensitive Information](#personal-and-sensitive-information)
  - [Considerations for Using the Data](#considerations-for-using-the-data)
    - [Discussion of Biases](#discussion-of-biases)
  - [Additional Information](#additional-information)
    - [Licensing Information](#licensing-information)

## Dataset Description

- **Point of Contact:** [datasets@mlcommons.org](mailto:datasets@mlcommons.org)

### Dataset Summary

The Speech Wikimedia Dataset is a compilation of audiofiles with transcriptions extracted from wikimedia commons that is licensed for academic and commercial usage under CC and Public domain. It includes 2,000+ hours of transcribed speech in different languages with a diverse set of speakers. 
Each audiofile should have one or more transcriptions in different languages.

### Transcription languages
- English
- German
- Dutch
- Arabic
- Hindi
- Portuguese
- Spanish
- Polish
- French
- Russian
- Esperanto
- Swedish
- Korean	
- Bengali
- Hungarian
- Oriya
- Thai

### Hours of Audio for each language

We determine the amount of data available for the ASR tasks by extractign the total duration of audio we have where we also have the transcription in the same langauge. We present the duration for the top 10 languages besides from English, from which we have a total of 1488 hours of audio:

![Audios](images/ASR_Top10_Non_Eng.jpg)

### Hours of language pairs for speech translation

Our dataset contains some audios with more than one transcription, all of which correspond to a different language transcription. In total, we have 628 hours fo audio with transcripts in different languages. We present the hours of audio for the 20 most common language pairs:

![Pairs](images/Speech_Translation.jpg)


## Dataset Structure

### audios

Folder with audios in flac format and sampling_rate=16,000 Hz.

### transcription and transcription2

Folders with transcriptions in srt format.

We split this into two directories because Hugging Face does not support more than 10,000 files in a single directory.

### real_correspondence.json

File with relationship between audios and transcriptions, as one large json dictionary.

Key is the name of an audio file in the "reformat" directory, value is the list of corresponding transcript files, which sit in either the transcription or transcription2 directory.


### license.json

File with license information. The key is the name of the original audio file on Wikimedia Commons.

### Data License

Here is an excerpt from license.json:

"""
'"Berlin Wall" Speech - President Reagan\'s Address at the Brandenburg Gate - 6-12-87.webm': {'author': '<td>\n<a class="external text" href="https://www.youtube.com/user/ReaganFoundation" rel="nofollow">ReaganFoundation</a></td>',
  'source': '<td>\n<bdo dir="ltr" lang="en"><a href="/wiki/Commons:YouTube_files" [...],
  'html_license': '[\'<table class="layouttemplate mw-content-ltr" lang="en" style="width:100%; [...],
  'license': 'Public Domain'},


## Dataset Creation

### Source Data

#### Initial Data Collection and Normalization

Data was downloaded from https://commons.wikimedia.org/.

#### Preprocessing

As the original format of most of the files was video, we decided to convert them to flac format with samplerate=16_000 using ffmpeg

### Annotations

#### Annotation process

No manual annotation is done. We download only source audio with already existing transcripts. 

In particular, there is no "forced alignment" or "segmentation" done on this dataset.

### Personal and Sensitive Information

Several of our sources are legal and government proceedings, spoken stories, speeches, and so on. Given that these were intended as public documents and licensed as such, it is natural that the involved individuals are aware of this.

## Considerations for Using the Data

### Discussion of Biases

Our data is downloaded from commons.wikimedia.org. As such, the data is biased towards whatever users decide to upload there.

The data is also mostly English, though this has potential for multitask learning because several audio files have more than one transcription.

## Additional Information

### Licensing Information

The source data contains data under Public Domain and Creative Commons Licenses. 

We license this dataset under https://creativecommons.org/licenses/by-sa/4.0/

The appropriate attributions are in license.json