File size: 4,058 Bytes
18f336f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 |
# coding=utf-8
"""EmoDB paralinguistics dataset."""
import os
import textwrap
import datasets
import itertools
import typing as tp
from pathlib import Path
from ._emodb import OFFICIAL_TRAIN, OFFICIAL_TEST
SAMPLE_RATE = 16_000
_COMPRESSED_FILENAME = 'emo-db.tar.gz'
EMOTIONS_MAPPING = {
'A': 'anxiety',
'E': 'disgust',
'F': 'happiness',
'L': 'boredom',
'N': 'neutral',
'T': 'sadness',
'W': 'anger',
}
EMOTIONS = [
'anxiety', 'disgust', 'happiness', 'boredom', 'neutral', 'sadness', 'anger'
]
class EmodbConfig(datasets.BuilderConfig):
"""BuilderConfig for EmoDB."""
def __init__(self, features, **kwargs):
super(EmodbConfig, self).__init__(version=datasets.Version("0.0.1", ""), **kwargs)
self.features = features
class EmoDB(datasets.GeneratorBasedBuilder):
BUILDER_CONFIGS = [
EmodbConfig(
features=datasets.Features(
{
"file": datasets.Value("string"),
"audio": datasets.Audio(sampling_rate=SAMPLE_RATE),
"emotion": datasets.Value("string"),
"label": datasets.ClassLabel(names=EMOTIONS),
}
),
name="gtzan",
description=textwrap.dedent(
"""\
Paralinguistics classifies each audio for its emotion as a multi-class
classification, where emotions are in the same pre-defined set for both training and testing.
The evaluation metric is accuracy (ACC).
"""
),
),
]
def _info(self):
return datasets.DatasetInfo(
description="",
features=self.config.features,
supervised_keys=None,
homepage="",
citation="",
task_templates=None,
)
def _split_generators(self, dl_manager):
"""Returns SplitGenerators."""
archive_path = dl_manager.extract(_COMPRESSED_FILENAME)
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN, gen_kwargs={"archive_path": archive_path, "split": "train"}
),
datasets.SplitGenerator(
name=datasets.Split.TEST, gen_kwargs={"archive_path": archive_path, "split": "test"}
),
]
def _generate_examples(self, archive_path, split=None):
extensions = ['.wav']
_, _walker = fast_scandir(archive_path, extensions, recursive=True)
if split == 'train':
_walker = [fileid for fileid in _walker if Path(fileid).stem in OFFICIAL_TRAIN]
elif split == 'test':
_walker = [fileid for fileid in _walker if Path(fileid).stem in OFFICIAL_TEST]
for guid, audio_path in enumerate(_walker):
yield guid, {
"id": str(guid),
"file": audio_path,
"audio": audio_path,
"emotion": EMOTIONS_MAPPING.get(Path(audio_path).stem[-2]),
"label": EMOTIONS_MAPPING.get(Path(audio_path).stem[-2]),
}
def fast_scandir(path: str, exts: tp.List[str], recursive: bool = False):
# Scan files recursively faster than glob
# From github.com/drscotthawley/aeiou/blob/main/aeiou/core.py
subfolders, files = [], []
try: # hope to avoid 'permission denied' by this try
for f in os.scandir(path):
try: # 'hope to avoid too many levels of symbolic links' error
if f.is_dir():
subfolders.append(f.path)
elif f.is_file():
if os.path.splitext(f.name)[1].lower() in exts:
files.append(f.path)
except Exception:
pass
except Exception:
pass
if recursive:
for path in list(subfolders):
sf, f = fast_scandir(path, exts, recursive=recursive)
subfolders.extend(sf)
files.extend(f) # type: ignore
return subfolders, files |