--- license: - cc-by-4.0 size_categories: ar: - n==1k task_categories: - automatic-speech-recognition task_ids: [] pretty_name: MASC dataset extra_gated_prompt: >- By clicking on “Access repository” below, you also agree to not attempt to determine the identity of speakers in the MASC dataset. language: - ar --- # Dataset Card for Common Voice Corpus 11.0 ## Table of Contents - [Dataset Description](#dataset-description) - [Dataset Summary](#dataset-summary) - [Languages](#languages) - [How to use](#how-to-use) - [Dataset Structure](#dataset-structure) - [Data Instances](#data-instances) - [Data Fields](#data-fields) - [Data Splits](#data-splits) - [Additional Information](#additional-information) - [Citation Information](#citation-information) ## Dataset Description - **Homepage:** https://ieee-dataport.org/open-access/masc-massive-arabic-speech-corpus - **Paper:** https://ieeexplore.ieee.org/document/10022652 ### Dataset Summary MASC is a dataset that contains 1,000 hours of speech sampled at 16 kHz and crawled from over 700 YouTube channels. The dataset is multi-regional, multi-genre, and multi-dialect intended to advance the research and development of Arabic speech technology with a special emphasis on Arabic speech recognition. ### Supported Tasks - Automatics Speach Recognition ### Languages ``` Arabic ``` ## How to use The `datasets` library allows you to load and pre-process your dataset in pure Python, at scale. The dataset can be downloaded and prepared in one call to your local drive by using the `load_dataset` function. ```python from datasets import load_dataset masc = load_dataset("pain/MASC", split="train") ``` Using the datasets library, you can also stream the dataset on-the-fly by adding a `streaming=True` argument to the `load_dataset` function call. Loading a dataset in streaming mode loads individual samples of the dataset at a time, rather than downloading the entire dataset to disk. ```python from datasets import load_dataset masc = load_dataset("pain/MASC", split="train", streaming=True) print(next(iter(masc))) ``` *Bonus*: create a [PyTorch dataloader](https://huggingface.co/docs/datasets/use_with_pytorch) directly with your own datasets (local/streamed). ### Local ```python from datasets import load_dataset from torch.utils.data.sampler import BatchSampler, RandomSampler masc = load_dataset("pain/MASC", split="train") batch_sampler = BatchSampler(RandomSampler(masc), batch_size=32, drop_last=False) dataloader = DataLoader(masc, batch_sampler=batch_sampler) ``` ### Streaming ```python from datasets import load_dataset from torch.utils.data import DataLoader masc = load_dataset("pain/MASC", split="train") dataloader = DataLoader(masc, batch_size=32) ``` To find out more about loading and preparing audio datasets, head over to [hf.co/blog/audio-datasets](https://huggingface.co/blog/audio-datasets). ### Example scripts Train your own CTC or Seq2Seq Automatic Speech Recognition models on MASC with `transformers` - [here](https://github.com/huggingface/transformers/tree/main/examples/pytorch/speech-recognition). ## Dataset Structure ### Data Instances A typical data point comprises the `path` to the audio file and its `sentence`. ```python {'video_id': 'OGqz9G-JO0E', 'start': 770.6, 'end': 781.835, 'duration': 11.24, 'text': 'اللهم من ارادنا وبلادنا وبلاد المسلمين بسوء اللهم فاشغله في نفسه ورد كيده في نحره واجعل تدبيره تدميره يا رب العالمين', 'type': 'c', 'file_path': '87edeceb-5349-4210-89ad-8c3e91e54062_OGqz9G-JO0E.wav', 'audio': {'path': None, 'array': array([ 0.05938721, 0.0539856, 0.03460693, ..., 0.00393677, 0.01745605, 0.03045654 ]), 'sampling_rate': 16000 } } ``` ### Data Fields `video_id` (`string`): An id for the video that the voice has been created from `start` (`float64`): The start of the audio's chunk `end` (`float64`): The end of the audio's chunk `duration` (`float64`): The duration of the chunk `text` (`string`): The text of the chunk `audio` (`dict`): A dictionary containing the path to the downloaded audio file, the decoded audio array, and the sampling rate. Note that when accessing the audio column: `dataset[0]["audio"]` the audio file is automatically decoded and resampled to `dataset.features["audio"].sampling_rate`. Decoding and resampling of a large number of audio files might take a significant amount of time. Thus it is important to first query the sample index before the `"audio"` column, *i.e.* `dataset[0]["audio"]` should **always** be preferred over `dataset["audio"][0]`. `type` (`string`): It refers to the data set type, either clean or noisy where "c: clean and n: noisy" 'file_path' (`string`): A path for the audio chunk "audio" ("audio"): Audio for the chunk ### Data Splits The speech material has been subdivided into portions for train, dev, test. The dataset splits has clean and noisy data that can be determined by type field. ### Citation Information ``` @INPROCEEDINGS{10022652, author={Al-Fetyani, Mohammad and Al-Barham, Muhammad and Abandah, Gheith and Alsharkawi, Adham and Dawas, Maha}, booktitle={2022 IEEE Spoken Language Technology Workshop (SLT)}, title={MASC: Massive Arabic Speech Corpus}, year={2023}, volume={}, number={}, pages={1006-1013}, doi={10.1109/SLT54892.2023.10022652}} } ```