Datasets:
Tasks:
Automatic Speech Recognition
Formats:
parquet
Languages:
Spanish
Size:
10K - 100K
Tags:
wikipedia grabada
wikipedia spanish
ciempiess-unam
ciempiess-unam project
read speech
spanish speech
License:
File size: 10,014 Bytes
0dc6c03 ac42c19 4ab1576 0dc6c03 4ab1576 0afbf01 4ab1576 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 |
---
license: cc-by-sa-3.0
dataset_info:
config_name: wikipedia_spanish
features:
- name: audio_id
dtype: string
- name: audio
dtype:
audio:
sampling_rate: 16000
- name: speaker_id
dtype: string
- name: gender
dtype: string
- name: duration
dtype: float32
- name: normalized_text
dtype: string
splits:
- name: train
num_bytes: 1916035098.198
num_examples: 11569
download_size: 1887119130
dataset_size: 1916035098.198
configs:
- config_name: wikipedia_spanish
data_files:
- split: train
path: wikipedia_spanish/train-*
default: true
task_categories:
- automatic-speech-recognition
language:
- es
tags:
- wikipedia grabada
- wikipedia spanish
- ciempiess-unam
- ciempiess-unam project
- read speech
- spanish speech
pretty_name: WIKIPEDIA SPANISH CORPUS
size_categories:
- 10K<n<100K
---
# Dataset Card for wikipedia_spanish
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [CIEMPIESS-UNAM Project](https://ciempiess.org/)
- **Repository:** [WIKIPEDIA SPANISH CORPUS at LDC](https://catalog.ldc.upenn.edu/LDC2021S07)
- **Point of Contact:** [Carlos Mena](mailto:carlos.mena@ciempiess.org)
### Dataset Summary
According to the project page of the [WikiProject Spoken Wikipedia](https://en.wikipedia.org/wiki/Wikipedia:WikiProject_Spoken_Wikipedia):
The WikiProject Spoken Wikipedia aims to produce recordings of Wikipedia articles being read aloud. Therefore, the WIKIPEDIA SPANISH CORPUS is a dataset created from the Spanish version of the WikiProject Spoken Wikipedia, called [Wikipedia Grabada](https://es.wikipedia.org/wiki/Wikiproyecto:Wikipedia_grabada)
The WIKIPEDIA SPANISH CORPUS aims to be used in the Automatic Speech Recognition (ASR) task. It is a gender unbalanced corpus of 25 hours of duration. It contains read speech of several articles of the Wikipedia Grabada; most of such articles are recorded by male speakers. Transcriptions in this corpus were generated from the scratch by native speakers.
### Example Usage
The WIKIPEDIA SPANISH CORPUS contains only the train split:
```python
from datasets import load_dataset
wikipedia_spanish = load_dataset("ciempiess/wikipedia_spanish")
```
It is also valid to do:
```python
from datasets import load_dataset
wikipedia_spanish = load_dataset("ciempiess/wikipedia_spanish",split="train")
```
### Supported Tasks
automatic-speech-recognition: The dataset can be used to test a model for Automatic Speech Recognition (ASR). The model is presented with an audio file and asked to transcribe the audio file to written text. The most common evaluation metric is the word error rate (WER).
### Languages
The language of the corpus is Spanish.
## Dataset Structure
### Data Instances
```python
{
'audio_id': 'WKSP_F_0019_E1_0023',
'audio': {
'path': '/home/carlos/.cache/HuggingFace/datasets/downloads/extracted/d31e2de01af3d9cdc4d4b6397720048caa39f6e552a8d38b6617a50af250bdcb/train/female/F_0019/WKSP_F_0019_E1_0023.flac',
'array': array([0.08535767, 0.13946533, 0.11572266, ..., 0.13168335, 0.12426758,
0.14508057], dtype=float32), 'sampling_rate': 16000
},
'speaker_id': 'F_0019',
'gender': 'female',
'duration': 8.170000076293945,
'normalized_text': 'donde revelaba de sus placas de vidrio al colodión controversias y equivocaciones'
}
```
### Data Fields
* `audio_id` (string) - id of audio segment
* `audio` (datasets.Audio) - a dictionary containing the path to the audio, the decoded audio array, and the sampling rate. In non-streaming mode (default), the path points to the locally extracted audio. In streaming mode, the path is the relative path of an audio inside its archive (as files are not downloaded and extracted locally).
* `speaker_id` (string) - id of speaker
* `gender` (string) - gender of speaker (male or female)
* `duration` (float32) - duration of the audio file in seconds.
* `normalized_text` (string) - normalized audio segment transcription
### Data Splits
The corpus counts just with the train split which has a total of 11569 speech files from 43 female speakers and 150 male speakers with a total duration of 25 hours and 37 minutes.
## Dataset Creation
### Curation Rationale
The WIKIPEDIA SPANISH CORPUS (WSC) has the following characteristics:
* The WSC has an exact duration of 25 hours and 37 minutes. It has 11569 audio files.
* The WSC counts with 193 different speakers: 150 men and 43 women.
* Every audio file in the WSC has a duration between 3 and 10 seconds approximately.
* Data in WSC is classified by speaker. It means, all the recordings of one single speaker are stored in one single directory.
* Data is also classified according to the gender (male/female) of the speakers.
* Audio and transcriptions in the WSC are segmented and transcribed from the scratch by native speakers of the Spanish language
* Audio files in the WSC are distributed in a 16khz@16bit mono format.
* Every audio file has an ID that is compatible with ASR engines such as Kaldi and CMU-Sphinx.
### Source Data
#### Initial Data Collection and Normalization
The WIKIPEDIA SPANISH CORPUS is a speech corpus designed to train acoustic models for automatic speech recognition and it is made out of several articles of the [Wikipedia Grabada](https://es.wikipedia.org/wiki/Wikiproyecto:Wikipedia_grabada) read by volunteers.
### Annotations
#### Annotation process
The annotation process is at follows:
* 1. A whole podcast is manually segmented keeping just the portions containing good quality speech.
* 2. A second pass os segmentation is performed; this time to separate speakers and put them in different folders.
* 3. The resulting speech files between 5 and 10 seconds are transcribed by students from different departments (computing, engineering, linguistics). Most of them are native speakers but not with a particular training as transcribers.
#### Who are the annotators?
The WIKIPEDIA SPANISH CORPUS was created under the umbrella of the social service program ["Desarrollo de Tecnologías del Habla"](http://profesores.fi-b.unam.mx/carlos_mena/servicio.html) of the ["Facultad de Ingeniería"](https://www.ingenieria.unam.mx/) (FI) in the ["Universidad Nacional Autónoma de México"](https://www.unam.mx/) (UNAM) between 2018 and 2020 by Carlos Daniel Hernández Mena, head of the program.
### Personal and Sensitive Information
The dataset could contain names revealing the identity of some speakers; on the other side, the recordings come from publicly available podcasts, so, there is not a real intent of the participants to be anonymized. Anyway, you agree to not attempt to determine the identity of speakers in this dataset.
## Considerations for Using the Data
### Social Impact of Dataset
This dataset is valuable because it contains well pronounced speech with low noise.
### Discussion of Biases
The dataset is not gender balanced. It is comprised of 43 female speakers and 150 male speakers.
### Other Known Limitations
WIKIPEDIA SPANISH CORPUS by Carlos Daniel Hernández Mena is licensed under a Creative Commons Attribution-ShareAlike 3.0 Unported [CC BY-SA 3.0](https://creativecommons.org/licenses/by-sa/3.0/deed.en) and it utilizes material from [Wikipedia Grabada](https://es.wikipedia.org/wiki/Wikiproyecto:Wikipedia_grabada). This work was done with the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
### Dataset Curators
The dataset was collected by students belonging to the social service program ["Desarrollo de Tecnologías del Habla"](http://profesores.fi-b.unam.mx/carlos_mena/servicio.html). It was curated by [Carlos Daniel Hernández Mena](https://huggingface.co/carlosdanielhernandezmena) in 2020.
### Licensing Information
[CC BY-SA 3.0](https://creativecommons.org/licenses/by-sa/3.0/deed.en)
### Citation Information
```
@misc{carlosmena2021wikipediaspanish,
title={WIKIPEDIA SPANISH CORPUS: Audio and Transcriptions taken from Wikipedia Grabada},
ldc_catalog_no={LDC2021S07},
DOI={https://doi.org/10.35111/7m1j-sa17},
author={Hernandez Mena, Carlos Daniel and Meza Ruiz, Ivan Vladimir},
journal={Linguistic Data Consortium, Philadelphia},
year={2021},
url={https://catalog.ldc.upenn.edu/LDC2021S07},
}
```
### Contributions
The authors would like to thank to Alberto Templos Carbajal, Elena Vera and Angélica Gutiérrez for their support to the social service program "Desarrollo de Tecnologías del Habla" at the Facultad de Ingeniería (FI) of the Universidad Nacional Autónoma de México (UNAM). We also thank to the social service students for all the hard work.
Special thanks to the Team of "Wikipedia Grabada" for publishing all the recordings that constitute the WIKIPEDIA SPANISH CORPUS.
This dataset card was created as part of the objectives of the 16th edition of the Severo Ochoa Mobility Program (PN039300 - Severo Ochoa 2021 - E&T). |