Datasets:
cjvt
/

Tasks:
Other
Languages:
Slovenian
License:
slo_collocations / slo_collocations.py
Matej Klemen
Add first version of script for loading Slovene collocations
9763e6d
"""A collocations dictionary of modern Slovene"""
import xml.etree.ElementTree as ET
import json
import os
import datasets
_CITATION = """\
@inproceedings{kosem2018collocations,
title={Collocations dictionary of modern Slovene},
author={Kosem, Iztok and Krek, Simon and Gantar, Polona and Arhar Holdt, {\v{S}}pela and {\v{C}}ibej, Jaka and Laskowski, Cyprian},
booktitle={Proceedings of the XVIII EURALEX International Congress: Lexicography in Global Contexts},
pages={989--997},
year={2018},
organization={Znanstvena zalo{\v{z}}ba Filozofske fakultete Univerze v Ljubljani}
}"""
_DESCRIPTION = """\
The database of the Collocations Dictionary of Modern Slovene 1.0 contains collocations that were automatically
extracted from the Gigafida 1.0 corpus and then postprocessed.
"""
_HOMEPAGE = "https://www.cjvt.si/kssj/en/"
_LICENSE = "Creative Commons - Attribution-ShareAlike 4.0 International (CC BY-SA 4.0)"
_URLS = {
"slo_collocations": "https://www.clarin.si/repository/xmlui/bitstream/handle/11356/1250/KSSS.zip"
}
class SloCollocations(datasets.GeneratorBasedBuilder):
"""A collocations dictionary of modern Slovene"""
VERSION = datasets.Version("1.0.0")
def _info(self):
features = datasets.Features({
"collocation": datasets.Value("string"),
"cluster": datasets.Value("uint32"),
"words": datasets.Sequence(datasets.Value("string")),
"sloleks_ids": datasets.Sequence(datasets.Value("string")),
"gramrel": datasets.Value("string"),
"sense": datasets.Value("uint32"),
"id_lex_unit": datasets.Value("string"),
"lex_unit": datasets.Value("string"),
"lex_unit_category": datasets.Value("string")
})
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=features,
homepage=_HOMEPAGE,
license=_LICENSE,
citation=_CITATION
)
def _split_generators(self, dl_manager):
urls = _URLS["slo_collocations"]
data_dir = dl_manager.download_and_extract(urls)
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN,
gen_kwargs={"data_dir": os.path.join(data_dir, "KSSS")}
)
]
def _generate_examples(self, data_dir):
# Map Slovene category (coarse POS tag) to its English translation
CATEGORY_MAPPING = {"samostalnik": "noun", "glagol": "verb", "pridevnik": "adjective", "prislov": "adverb"}
all_file_paths = []
for cat_slo in ["glagol", "pridevnik", "prislov", "samostalnik"]:
_curr_dir = os.path.join(data_dir, cat_slo)
for _fname in os.listdir(_curr_dir):
if os.path.isfile(os.path.join(_curr_dir, _fname)) and _fname.endswith(".xml"):
all_file_paths.append(os.path.join(_curr_dir, _fname))
idx_ex = 0
for file_path in all_file_paths:
curr_doc = ET.parse(file_path)
root = curr_doc.getroot()
header_tag = root.find("header")
lu_tag = header_tag.find("lexical_unit")
id_lu = lu_tag.attrib["id"]
lu_form = lu_tag.text
category = CATEGORY_MAPPING[header_tag.find("category").text.strip()]
for idx_sense, sense_tag in enumerate(root.findall(".//sense")):
for gramrel_tag in sense_tag.findall(".//gramrel"):
gramrel = gramrel_tag.attrib["name"]
for colloc_tag in gramrel_tag.findall(".//collocation"):
colloc_form = colloc_tag.attrib["form"]
colloc_cluster = colloc_tag.attrib["cluster"]
colloc_words, sloleks_ids = [], []
for comp_tag in colloc_tag.findall("comp"):
colloc_words.append(comp_tag.text)
sloleks_ids.append(comp_tag.attrib["sloleks"])
yield idx_ex, {
"collocation": colloc_form,
"cluster": int(colloc_cluster),
"words": colloc_words,
"sloleks_ids": sloleks_ids,
"gramrel": gramrel,
"sense": idx_sense,
"id_lex_unit": id_lu,
"lex_unit": lu_form,
"lex_unit_category": category,
}
idx_ex += 1