|
--- |
|
dataset_info: |
|
features: |
|
- name: audio |
|
dtype: |
|
audio: |
|
sampling_rate: 16000 |
|
- name: sentence |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 12264199958.656 |
|
num_examples: 49504 |
|
download_size: 11879936920 |
|
dataset_size: 12264199958.656 |
|
configs: |
|
- config_name: default |
|
data_files: |
|
- split: train |
|
path: data/train-* |
|
license: apache-2.0 |
|
task_categories: |
|
- automatic-speech-recognition |
|
- translation |
|
- text-to-speech |
|
language: |
|
- ja |
|
size_categories: |
|
- 10K<n<100K |
|
--- |
|
common voice, google fleurs, JSUTv1.1, JAS_v2 (joujiboi/japanese-anime-speech-v2) processed for whisper. Not shuffled or normalized. 50% anime speech and 50% other. Other corpus's are fully represented. |
|
|
|
if needed: |
|
|
|
import neologdn |
|
import MeCab |
|
import re |
|
from transformers.models.whisper.english_normalizer import BasicTextNormalizer |
|
|
|
wakati = MeCab.Tagger("-Owakati") |
|
special_characters = '[\,\γ\γ\οΌ\γ\γ\β¦\οΌ\γ»\!\-\;\:\"\β\%\β\β\οΏ½]' |
|
|
|
|
|
def norm_everything(batch): |
|
batch["sentence"] = neologdn.normalize(batch["sentence"]).strip() |
|
batch["sentence"] = normalizer(batch["sentence"]).strip() |
|
batch["sentence"] = wakati.parse(batch["sentence"]).strip() |
|
batch["sentence"] = re.sub(special_characters,'', batch["sentence"]).strip() |
|
return batch |
|
|
|
ds = ds.map(norm_everything) |