cngv1 / cngv1.py
jimregan's picture
fix usage hint
20292dd
# coding=utf-8
# Copyright 2021 The TensorFlow Datasets Authors and the HuggingFace Datasets Authors.
# Copyright 2021 Phonetics and Speech Laboratory, Trinity College, Dublin
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# Lint as: python3
import os
from pathlib import Path
import datasets
from bs4 import BeautifulSoup
_CITATION = """\
@article{ite2003corpas,
title={Corpas Náisiúnta na Gaeilge/National Corpus of Irish, Volume 1},
author={Institiúid Teangeolaíochta Éireann},
journal={Dublin: ITÉ},
year={2003}
}
"""
_DESCRIPTION = """\
Corpus of written Irish.
"""
_TEXTDIRS = [
"fiction", "information", "instruction", "non_fiction", "official"
]
class CNGDataset(datasets.GeneratorBasedBuilder):
"""National Corpus of Irish."""
VERSION = datasets.Version("1.1.0")
BUILDER_CONFIGS = [
datasets.BuilderConfig(name="documents", version=VERSION, description="Plain text portion of the corpus: whole documents"),
datasets.BuilderConfig(name="paragraphs", version=VERSION, description="Plain text portion of the corpus: paragraphs"),
datasets.BuilderConfig(name="pos", version=VERSION, description="Part-of-speech tagging subset"),
]
def _info(self):
if self.config.name == "documents" or self.config.name == "paragraphs":
features = datasets.Features(
{
"title": datasets.Value("string"),
"doc_id": datasets.Value("string"),
"author": datasets.Value("string"),
"date": datasets.Value("string"),
"text": datasets.Value("string"),
"classes": datasets.Sequence(datasets.Value("string"))
}
)
else:
features = datasets.Features(
{
"title": datasets.Value("string"),
"doc_id": datasets.Value("string"),
"author": datasets.Value("string"),
"date": datasets.Value("string"),
"classes": datasets.Sequence(datasets.Value("string")),
"words": datasets.Sequence(datasets.Value("string")),
"pos": datasets.Sequence(datasets.Value("string"))
}
)
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=features,
supervised_keys=None,
citation=_CITATION,
)
def _split_generators(self, dl_manager):
"""Returns SplitGenerators."""
manual_dir = os.path.abspath(os.path.expanduser(dl_manager.manual_dir))
if not os.path.exists(manual_dir):
raise FileNotFoundError(
"{} does not exist. Make sure you insert a manual dir via `datasets.load_dataset('phonlab-tcd/cngv1', data_dir=...)` with the path to the corpus directory".format(
manual_dir
)
)
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN,
gen_kwargs={
"data_dir": manual_dir,
"split": "train",
},
),
]
def _generate_examples(
self, data_dir, split
):
""" Yields examples as (key, example) tuples. """
if self.config.name == "documents" or self.config.name == "paragraphs":
dirs = _TEXTDIRS
else:
dirs = ["pos"]
cng_path = Path(data_dir)
_id = 1
for dir in dirs:
dir_path = cng_path / dir
for filepath in dir_path.glob('*.SGM'):
with open(filepath, encoding="utf-16-le") as f:
fid = filepath.stem
content = f.read()
soup = BeautifulSoup(content, 'html.parser')
title = _get_title(soup)
author = _get_author(soup)
classes = _get_categories(content)
date = _get_creation(soup)
if self.config.name == "pos":
for sent in _get_pos(soup):
words = [tok["word"] for tok in sent]
tags = [tok["msd"] for tok in sent]
yield _id, {
"title": title,
"doc_id": fid,
"author": author,
"date": date,
"classes": classes,
"words": words,
"pos": tags
}
_id += 1
else:
text = _get_paragraphs(soup)
if self.config.name == "documents":
text = ["\n".join(text)]
for para in text:
yield _id, {
"title": title,
"doc_id": fid,
"author": author,
"date": date,
"classes": classes,
"text": para
}
_id += 1
def _get_title(soup):
title = soup.find("title")
if title.text and title.text.strip() != "":
return title.text.strip()
def _get_author(soup):
author = soup.find("author")
if author.text and author.text.strip() != "":
return author.text.strip()
def _get_creation(soup):
creation = soup.find("creation")
if creation.text and creation.text.strip() != "":
return creation.text.strip()
def _get_paragraphs(soup):
import re
out = []
body = soup.find('body')
for p in body.find_all(['p', 'head']):
text = p.text.strip()
text = text.replace('\n', ' ')
text = re.sub(' +', ' ', text)
if text:
out.append(text)
return out
def _get_categories(text):
import re
out = []
for cat in re.findall('<catRef target="([^"]+)">', text):
out.append(cat)
return out
def _get_pos(soup):
out = []
for sent in soup.find_all('s'):
words = []
for word in sent.find_all('w'):
if word.text:
text = word.text.strip()
msd = word.get('msd')
if msd and text:
words.append({"msd": msd, "word": text})
out.append(words)
return out