Datasets:
Tasks:
Automatic Speech Recognition
Languages:
Polish
Multilinguality:
monolingual
Size Categories:
n<1K
Language Creators:
crowdsourced
Annotations Creators:
crowdsourced
import os | |
import datasets | |
import pandas as pd | |
_CITATION = """""" | |
_DESCRIPTION = """\ | |
This dataset is designed to be used in training models | |
that restore punctuation marks from the output of | |
Automatic Speech Recognition system for Polish language. | |
""" | |
_HOMEPAGE = "https://github.com/poleval/2021-punctuation-restoration" | |
_URL = "https://raw.githubusercontent.com/poleval/2021-punctuation-restoration/main" | |
_PATHS = { | |
"train": [os.path.join(_URL, "train", "in.tsv"), os.path.join(_URL, "train", "expected.tsv")], | |
"test-A": [os.path.join(_URL, "test-A", "in.tsv"), os.path.join(_URL, "test-A", "expected.tsv")], | |
} | |
_TO_DOWNLOAD = _PATHS["train"] + _PATHS["test-A"] | |
class PunctuationDatasetConfig(datasets.BuilderConfig): | |
"""BuilderConfig for AfrikaansNerCorpus""" | |
def __init__(self, **kwargs): | |
"""BuilderConfig for PunctuationDataset. | |
Args: | |
**kwargs: keyword arguments forwarded to super. | |
""" | |
super(PunctuationDatasetConfig, self).__init__(**kwargs) | |
class PunctuationDataset(datasets.GeneratorBasedBuilder): | |
"""TODO: Short description of my dataset.""" | |
VERSION = datasets.Version("1.0.0") | |
BUILDER_CONFIGS = [ | |
PunctuationDatasetConfig( | |
name="punctuation_dataset", | |
version=datasets.Version("1.0.0"), | |
description="PunctuationDataset dataset", | |
), | |
] | |
def _info(self): | |
return datasets.DatasetInfo( | |
description=_DESCRIPTION, | |
features=datasets.Features( | |
{ | |
"text_in": datasets.Value("string"), | |
"text_out": datasets.Value("string"), | |
"tokens": datasets.Sequence(datasets.Value("string")), | |
"tags": datasets.Sequence( | |
datasets.features.ClassLabel( | |
names=[ | |
'B-.', | |
'B-,', | |
'B--', | |
'B-!', | |
'B-?', | |
'B-:', | |
'B-;', | |
'O', | |
] | |
) | |
) | |
}), | |
supervised_keys=None, | |
homepage=_HOMEPAGE, | |
citation=_CITATION, | |
) | |
def _split_generators(self, dl_manager): | |
data_paths = {key: dl_manager.download(urls) for key, urls in _PATHS.items()} | |
print(data_paths) | |
return [ | |
datasets.SplitGenerator( | |
name=datasets.Split.TRAIN, gen_kwargs={"filepaths": data_paths["train"]} | |
), | |
datasets.SplitGenerator( | |
name=datasets.Split.TEST, gen_kwargs={"filepaths": data_paths["test-A"]} | |
), | |
] | |
def _generate_examples(self, filepaths): | |
in_df = pd.read_csv(os.path.join(filepaths[0]), sep='\t', header=None) | |
out_df = pd.read_csv(os.path.join(filepaths[1]), sep='\t', header=None) | |
for key, ((_, row_in), (_, row_out)) in enumerate(zip(in_df.iterrows(), out_df.iterrows()), 1): | |
text_in = PunctuationDataset._clean_text(row_in[1]) | |
text_out = PunctuationDataset._clean_text(row_out[0]) | |
tokens = [] | |
tags = [] | |
for token_in, token_out in zip(text_in.split(), text_out.split()): | |
assert token_in.lower() in token_out.lower() | |
tokens.append(token_in) | |
if token_in.lower() == token_out.lower(): | |
tags.append('O') | |
else: | |
tags.append(f'B-{token_out[-1]}') | |
yield key, { | |
"text_in": text_in, | |
"text_out": text_out, | |
"tokens": tokens, | |
"tags": tags | |
} | |
def _clean_text(text: str, lower: bool = False) -> str: | |
if lower: | |
text = text.lower() | |
text = text.replace(' -', '') | |
text = text.replace(' .', '') | |
text = text.replace(' ,', '') | |
text = text.replace(' ', ' ') | |
text = text.strip() | |
return text | |