--- language: - bn license: cc-by-nc-4.0 task_categories: - automatic-speech-recognition dataset_info: features: - name: audio dtype: audio - name: text dtype: string - name: duration dtype: float64 - name: category dtype: string - name: source dtype: string splits: - name: train num_bytes: 219091915.875 num_examples: 1753 download_size: 214321460 dataset_size: 219091915.875 configs: - config_name: default data_files: - split: train path: data/train-* --- # MegaBNSpeech This model is based on a study aimed at tackling one of the primary challenges in developing Automatic Speech Recognition (ASR) for low-resource languages (Bangla): the limited access to domain-specific labeled data. To address this, the study introduces a pseudo-labeling approach to develop a domain-agnostic ASR dataset. The methodology led to the creation of a robust 20k+ hours labeled Bangla speech dataset, which encompasses a wide variety of topics, speaking styles, dialects, noisy environments, and conversational scenarios. Using this data, a conformer-based ASR system was designed. The effectiveness of the model, especially when trained on pseudo-labeled data, was benchmarked against publicly available datasets and compared with other models. The research promises that experimental resources stemming from this study will be made publicly available. ## How to use: The datasets library provides the capability to load and process your dataset efficiently using just Python. You can easily download and set up the dataset on your local drive with a single call using the *load_dataset* function. ```python from datasets import load_dataset dataset = load_dataset("hishab/MegaBNSpeech", split="train") ``` With the datasets library, you have the option to stream the dataset in real-time by appending the streaming=True parameter to the load_dataset function. In streaming mode, the dataset loads one sample at a time instead of storing the whole dataset on the disk. ```python from datasets import load_dataset dataset = load_dataset("hishab/MegaBNSpeech", split="train", streaming=True) print(next(iter(dataset))) ``` ## Speech Recognition (ASR) ```python from datasets import load_dataset mega_bn_asr = load_dataset("hishab/MegaBNSpeech") # see structure print(mega_bn_asr) # load audio sample on the fly audio_input = mega_bn_asr["train"][0]["audio"] # first decoded audio sample transcription = mega_bn_asr["train"][0]["transcription"] # first transcription # use `audio_input` and `transcription` to fine-tune your model for ASR ``` ## Data Structure - The dataset was developed using a pseudo-labeling approach. - The largest collection of Bangla audio-video data was curated and cleaned from various Bangla TV channels on YouTube. This data covers varying domains, speaking styles, dialects, and communication channels. - Alignments from two ASR systems were leveraged to segment and automatically annotate the audio segments. - The created dataset was used to design an end-to-end state-of-the-art Bangla ASR system. ### Data Instances - Size of downloaded dataset files: ___ GB - Size of the generated dataset: ___ MB - Total amount of disk used: ___ GB An example of a data instance looks as follows: ``` { "id": 0, "audio_path": "data/train/wav/UCPREnbhKQP-hsVfsfKP-mCw_id_2kux6rFXMeM_85.wav", "transcription": "পরীক্ষার মূল্য তালিকা উন্মুক্ত স্থানে প্রদর্শনের আদেশ দেন এই আদেশ পাওয়ার", "duration": 5.055 } ``` ### Data Fields The data fields are written below. - **id** (int): ID of audio sample - **audio_path** (str): Path to the audio file - **transcription** (str): Transcription of the audio file - **duration** : 5.055 ### Dataset Creation The dataset was developed using a pseudo-labeling approach. An extensive, large-scale, and high-quality speech dataset of approximately 20,000 hours was developed for domain-agnostic Bangla ASR. ## Social Impact of Dataset ## Limitations ## Citation Information You can access the MegaBNSpeech paper at _________________ Please cite the paper when referencing the MegaBNSpeech corpus as: ``` @article{_______________, title = {_______________________________}, author = {___,___,___,___,___,___,___,___}, journal={_______________________________}, url = {_________________________________}, year = {2023},