clustering_segments / clustering_segments.py
Guscerra's picture
Snap
99d00a7
raw
history blame
No virus
3.51 kB
import datasets
import pandas as pd
import numpy as np
logger = datasets.logging.get_logger(__name__)
_DATA_PATH = "https://huggingface.co/datasets/conversy/clustering_segments/resolve/main/dataset.pkl"
class ClusteringSegmentsConfig(datasets.BuilderConfig):
"""BuilderConfig for Conversy Benchmark."""
def __init__(self, name, version, **kwargs):
"""BuilderConfig for Conversy Benchmark.
Args:
**kwargs: keyword arguments forwarded to super.
"""
self.name = name
self.version = version
self.features = kwargs.pop("features", None)
self.description = kwargs.pop("description", None)
self.data_url = kwargs.pop("data_url", None)
self.nb_data_shards = kwargs.pop("nb_data_shards", None)
super(ClusteringSegmentsConfig, self).__init__(
name=name,
version=version,
**kwargs
)
class ClusteringSegments(datasets.GeneratorBasedBuilder):
"""Conversy benchmark"""
VERSION = datasets.Version("1.0.0")
BUILDER_CONFIGS = [
ClusteringSegmentsConfig(
name="VPClusteringBenchmark",
version=VERSION,
description="Conversy Benchmark for ML models evaluation",
features=["segment_id", "filename", "speaker", "duration", "vp",
"start", "end", "readable_start", "readable_end",
"segment_clean"],
data_url=_DATA_PATH,
nb_data_shards=1)
]
def _info(self):
description = (
"Voice Print Clustering Benchmark"
)
features = datasets.Features(
{
"segment_id": datasets.Value("int32"),
"filename": datasets.Value("string"),
"speaker": datasets.Value("string"),
"duration": datasets.Value("float32"),
"segment_clean": datasets.Value("bool"),
"start": datasets.Value("float32"),
"end": datasets.Value("float32"),
"readable_start": datasets.Value("string"),
"readable_end": datasets.Value("string"),
"vp": datasets.Sequence(datasets.Value("float32"))
})
return datasets.DatasetInfo(
description=description,
features=features,
supervised_keys=None,
version=self.config.version
)
def _split_generators(self, dl_manager):
"""Returns SplitGenerators."""
data_url = self.config.data_url
downloaded_file = dl_manager.download_and_extract(data_url)
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN,
gen_kwargs={"file_path": downloaded_file},
),
]
def _generate_examples(self, file_path):
"""Yields examples."""
df = pd.read_pickle(file_path)
for idx, row in df.iterrows():
yield idx, {
"segment_id": row["segment_id"],
"filename": row["filename"],
"speaker": row["speaker"],
"duration": row["duration"],
"segment_clean": row["segment_clean"],
"start": row['start'],
"end": row['end'],
"readable_start": row['readable_start'],
"readable_end": row['readable_end'],
"vp": np.asarray(row["vp"], dtype=np.float32)
}