--- license: cc0-1.0 task_categories: - automatic-speech-recognition language: - es - eu pretty_name: Basque Parliament Speech Corpus 1.0 --- # Dataset Card for Basque Parliament Speech Corpus 1.0 ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Supported Tasks](#supported-tasks) - [Languages](#languages) ## Dataset Description - **Repository:** https://huggingface.co/datasets/gttsehu/basque_parliament_1 - **Paper:** https://arxiv.org/ - **Contact:** [Luis J. Rodriguez-Fuentes](mailto:luisjavier.rodriguez@ehu.eus) ### Dataset Summary The Basque Parliament Speech Corpus 1.0 consists of 1462 hours of speech extracted from Basque Parliament plenary sessions from 2013 to 2022. Encoded as MP3 files, the dataset contains 759192 transcribed segments either spoken in Basque, Spanish or both (in Basque and Spanish). The corpus was created to help the development of speech technology for the Basque language, which is relatively low-resourced. However, the dataset is suited to the development of bilingual ASR systems, meaning to decode speech signals in Basque and/or Spanish. Given the similarity between Basque and Spanish at the phonetic/phonological level, acoustic models can be shared by both languages, which comes to circumvent the lack of training data for Basque. The dataset contains of four splits: `train`, `train_clean`, `dev` and `test`, all of them containing 3-10 second long speech segments and their corresponding transcriptions. Besides the transcription, each segment includes a speaker identifier and a language tag (Spanish, Basque or bilingual). The `train` split, aimed at estimating acoustic models, was extracted from 2013-2021 sessions, amounting to 1445 hours of speech. The `train_clean` split is a subset of the `train` split, containing only highly reliable transcriptions. The `dev` and `test` splits, amounting to 7.6 and 9.6 hours of speech respectively, were extracted from February 2022 sessions and their transcripts were manually audited. ### Languages The dataset contains segments either spoken in Basque (`eu`), Spanish (`es`) or both (`bi`). The language distribution is strongly biased towards Spanish and bilingual segments are very unfrequent. Duration (in hours) disaggregated per language: | **Split** | **es** | **eu** | **bi** | **Total** | |------------:|-------:|-------:|-------:|----------:| | train | 1018.6 | 409.5 | 17.0 | 1445.1 | | train_clean | 937.7 | 363.6 | 14.2 | 1315.5 | | dev | 4.7 | 2.6 | 0.3 | 7.6 | | test | 6.4 | 2.8 | 0.4 | 9.6 | Number of segments disaggregated per language: | **Split** | **es** | **eu** | **bi** | **Total** | |------------:|-------:|-------:|-------:|----------:| | train | 524942 | 216201 | 8802 | 749945 | | train_clean | 469937 | 184950 | 6984 | 661871 | | dev | 2567 | 1397 | 131 | 4095 | | test | 3450 | 1521 | 181 | 5152 | The dataset contains four configs that can be used to select the full set of multilingual segments or just a subset of them, constrained to a single language: * `all` : all the segments * `es` : only the Spanish segments * `eu` : only the Basque segments * `bi` : only the bilingual segments ## How to use You can use the `datasets` library to load the dataset from Python. The dataset can be downloaded in one call to your local drive by using the `load_dataset` function. For example, to download the Basque config of the `train` split, simply specify the desired language config name (i.e., "eu" for Basque) and the split: ```python from datasets import load_dataset ds = load_dataset("gttsehu/basque_parliament_1", "eu", split="train") ``` The default config is `all` and if no split is indicated all splits are prepared, so the next code prepares the full dataset: ```python from datasets import load_dataset ds = load_dataset("gttsehu/basque_parliament_1") ```