KaraKaraWitch's picture
Update README.md
5e8294c verified
metadata
license: other
license_name: singapore-open-data-license
license_link: https://data.gov.sg/open-data-licence
task_categories:
  - text-generation
  - text-classification
  - automatic-speech-recognition
  - audio-classification
language:
  - en
pretty_name: The Reprocessed Singapore National Speech Corpus

Dataset Card for Reprocessed National Speech Corpus

NOTE: This is an Reprocessed version KaraKaraWitch from Recursal.
The official download can be found here.

Dataset Details

Dataset Description

Dataset Description:

The National Speech Corpus (NSC) is the first large-scale Singapore English corpus, sponsored by the Info-communications and Media Development Authority (IMDA) of Singapore. The objective is to serve as a primary resource of open speech data for automatic speech recognition (ASR) research and other applications related to speech processing.

Please note that this is an reprocessed version of the original corpus available here.

This version contains some differences compared to the original because of the unusual file formats, which were hard to process. We modified the formatting and content to make it easier for use in our dataset and for public Huggingface use. For details on the changes and a list of original files and their modified counterparts, please see the documentation in the Docs folder.

Strike through content denotes the deleted or reformatted information for enhanced usability. All original data has been preserved and made available in the updated formats.

Changes

  • Converted docx documentation into markdown text for ease of sharing. The conversation was done with verbatim in mind. Minimizing as much changes as possible.
  • Converted XLSX tables into utf-8 csv (The BOM is included due to excel's csv saving).
  • All files are in split tar files in standard json / jsonl files + FLAC Compressed audio.
  • We have not modified any text transcripts, audio files were originally saved in .wav which we have losslessly compressed to flac.

Uses

The National Speech Corpus (NSC) is a large-scale Singapore English corpus aimed to improve speech engines' accuracy of recognition and transcription for locally accented English, supporting innovative digital solutions and driving progress in Singapore's digital landscape.

Direct Use

The NSC can be used to improve Automatic Speech Recognition (ASR) research and speech-related applications, such as telco call centres to transcribe calls for auditing and sentiment analysis purposes, and chatbots that can accurately support the Singaporean accent.

Out-of-Scope Use

The NSC is not intended for speech synthesis technology, but it can contribute to producing an AI voice that is more familiar to Singaporeans, with local terms pronounced more accurately.

Dataset Structure

The entire NSC is approximately 1.2 TB (~800GB with FLAC Compression and json metadata) in size, consisting of 6 parts.

Due to HF file restrictions, we have chunked the file with split command. As such you will need to recombine the files with cat or similar methods.

Using DifferentRooms.tar from Part 3 as an Example:
cat "DifferentRooms-00.tar" "DifferentRooms-01.tar" [...] "DifferentRooms-06.tar" "DifferentRooms-07.tar" > "DifferentRooms.tar"

After concating / combining the files, refer to the Usage Example on how to use it.

Usage Example

The dataset can be loaded with webdataset.

import webdataset as wds
# After concatting, you may use the file like a regular dataset.

# The dataset is compatible with WebDataset format. Example...

# filepath to Part 1/Channel_0.tar
FILEPATH = "Channel_0.tar"

hf_dataset = wds.WebDataset(FILEPATH).shuffle(1000).to_tuple("json", "flac")

for i in hf_dataset:
    # print(i)
    # Prints something like this:
    # i = {
    #     "__key__": "SP0402-CH00-SE00-RC023",
    #     "__url__": FILEPATH,
    #     "json": b'{"SpeakerID":402,"ChannelID":0,"SessionID":0,"RecordingID":23,"original_text":"I felt happy upon hearing the good news from my parents.","read_text":"I felt happy upon hearing the good news from my parents"}',
    #     "flac": b"",
    # }
    break

If you need to use HF datasets, load it like so:

from datasets import load_dataset

# The tar files are split due to HF limits. You will need to combine them first.
# You may use the following 
# `cat "DifferentRooms-00.tar" "DifferentRooms-01.tar" ... "DifferentRooms-06.tar" "DifferentRooms-07.tar" > "DifferentRooms.tar"`
# After concatting, you may use the file like a regular dataset.

FILEPATH = "Channel_0.tar"

hf_dataset = load_dataset("webdataset", data_files={"train": FILEPATH}, split="train", streaming=True)

# NOTE: You will need to install 'librosa' and 'soundfile' to decode the flac file.

for i in hf_dataset:
    print(i)
    # Prints something like this:
    # {
    #     "__key__": "SP0402-CH00-SE00-RC001",
    #     "__url__": FILEPATH,
    #     "json": {
    #         "ChannelID": 0,
    #         "RecordingID": 1,
    #         "SessionID": 0,
    #         "SpeakerID": 402,
    #         "original_text": "Mary and her family were moving to another city.",
    #         "read_text": "Mary and her family were moving to another city",
    #     },
    #     "flac": {
    #         "path": "SP0402-CH00-SE00-RC001.flac",
    #         "array": array(
    #             [
    #                 0.00000000e00,
    #                 6.10351562e-05,
    #                 1.52587891e-04,
    #                 ...,
    #                 -2.44140625e-04,
    #                 -2.44140625e-04,
    #                 -1.83105469e-04,
    #             ]
    #         ),
    #         "sampling_rate": 16000,
    #     },
    # }
    break

Other Notes

  • Downloading from dropbox is really not optimal. I've sent a email to nsc@imda.gov.sg but they didn't respond to me. 😔
    • Managed to get a response from them. Though it is around 1 month and I've already downloaded all the files from the dropbox link.
  • The scripts has numerous text issues. Notably missing quotes and incorrect encoding which made it a absolute headache to process.
    • I'll probably release the tools at a later time.

BibTeX Citation

@ONLINE{reprocessed_nationalspeechcorpus,
  title         = {The Reprocessed National Speech Corpus},
  author        = {KaraKaraWitch},
  year          = {2024},
  howpublished  = {\url{https://huggingface.co/datasets/recursal/reprocessed_national_speech_corpus}},
}

@ONLINE{imda_nationalspeechcorpus,
  title         = {IMDA National Speech Corpus},
  author        = {Infocomm Media Development Authority},
  year          = {2024},
  howpublished  = {\url{https://www.imda.gov.sg/how-we-can-help/national-speech-corpus}},
}

Glossary

There is a compiled markdown containing documents and notes various notes.

Recursal's Vision

To make AI accessible to everyone, regardless of language, or economical status

This is the collective goal of the RWKV Open Source foundation and Recursal AI, the commercial entity who backs it.

We believe that AI should not be controlled by a select few individual organization. And that it should be made accessible regardless if you are rich or poor, or a native speaker of english.

About RWKV

RWKV is an Open Source, non profit group, under the linux foundation. Focused on developing the RWKV AI architecture, in accordence to our vision.

The RWKV architecture scales efficiently and economically. As an RNN & Transformer hybrid, it is able to provide the performance similar to leading transformer models, while having the compute and energy efficiency of an RNN based architecture.

You can find out more about the project, and latest models, at the following

About Recursal AI

Recursal AI, is the commercial entity built to provide support for RWKV model development and users, while providing commercial services via its public cloud, or private-cloud / on-premise offerings.

As part of our vision. Our commitment, is to ensure open source development and access to the best foundational AI models and datasets.

The following dataset/models provided here, is part of that commitment.

You can find out more about recursal AI here

Dataset Card Contact

For issues regarding this reprocessed dataset, you may use the community discussions thread. In regards to licensing, please refer to Singapore Open Data License listed above. We will close any issues regard licensing due to abuse as of late due to it.

For further enquiries on the original National Speech Corpus (raw data), please contact nsc@imda.gov.sg.