import os import datasets import pandas as pd _CITATION = """""" _DESCRIPTION = """\ This dataset is designed to be used in training models that restore punctuation marks from the output of Automatic Speech Recognition system for Polish language. """ _HOMEPAGE = "https://github.com/poleval/2021-punctuation-restoration" _URL = "https://raw.githubusercontent.com/poleval/2021-punctuation-restoration/main" _PATHS = { "train": [os.path.join(_URL, "train", "in.tsv"), os.path.join(_URL, "train", "expected.tsv")], "test-A": [os.path.join(_URL, "test-A", "in.tsv"), os.path.join(_URL, "test-A", "expected.tsv")], } _TO_DOWNLOAD = _PATHS["train"] + _PATHS["test-A"] class PunctuationDatasetConfig(datasets.BuilderConfig): """BuilderConfig for AfrikaansNerCorpus""" def __init__(self, **kwargs): """BuilderConfig for PunctuationDataset. Args: **kwargs: keyword arguments forwarded to super. """ super(PunctuationDatasetConfig, self).__init__(**kwargs) class PunctuationDataset(datasets.GeneratorBasedBuilder): """TODO: Short description of my dataset.""" VERSION = datasets.Version("1.0.0") BUILDER_CONFIGS = [ PunctuationDatasetConfig( name="punctuation_dataset", version=datasets.Version("1.0.0"), description="PunctuationDataset dataset", ), ] def _info(self): return datasets.DatasetInfo( description=_DESCRIPTION, features=datasets.Features( { "text_in": datasets.Value("string"), "text_out": datasets.Value("string"), "tokens": datasets.Sequence(datasets.Value("string")), "tags": datasets.Sequence( datasets.features.ClassLabel( names=[ 'B-.', 'B-,', 'B--', 'B-!', 'B-?', 'B-:', 'B-;', 'O', ] ) ) }), supervised_keys=None, homepage=_HOMEPAGE, citation=_CITATION, ) def _split_generators(self, dl_manager): data_paths = {key: dl_manager.download(urls) for key, urls in _PATHS.items()} print(data_paths) return [ datasets.SplitGenerator( name=datasets.Split.TRAIN, gen_kwargs={"filepaths": data_paths["train"]} ), datasets.SplitGenerator( name=datasets.Split.TEST, gen_kwargs={"filepaths": data_paths["test-A"]} ), ] def _generate_examples(self, filepaths): in_df = pd.read_csv(os.path.join(filepaths[0]), sep='\t', header=None) out_df = pd.read_csv(os.path.join(filepaths[1]), sep='\t', header=None) for key, ((_, row_in), (_, row_out)) in enumerate(zip(in_df.iterrows(), out_df.iterrows()), 1): text_in = PunctuationDataset._clean_text(row_in[1]) text_out = PunctuationDataset._clean_text(row_out[0]) tokens = [] tags = [] for token_in, token_out in zip(text_in.split(), text_out.split()): assert token_in.lower() in token_out.lower() tokens.append(token_in) if token_in.lower() == token_out.lower(): tags.append('O') else: tags.append(f'B-{token_out[-1]}') yield key, { "text_in": text_in, "text_out": text_out, "tokens": tokens, "tags": tags } @staticmethod def _clean_text(text: str, lower: bool = False) -> str: if lower: text = text.lower() text = text.replace(' -', '') text = text.replace(' .', '') text = text.replace(' ,', '') text = text.replace(' ', ' ') text = text.strip() return text