Datasets:
Tasks:
Other
Languages:
English
Multilinguality:
monolingual
Size Categories:
100K<n<1M
Language Creators:
found
Annotations Creators:
machine-generated
Source Datasets:
original
License:
# coding=utf-8 | |
# Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor. | |
# | |
# Licensed under the Apache License, Version 2.0 (the "License"); | |
# you may not use this file except in compliance with the License. | |
# You may obtain a copy of the License at | |
# | |
# http://www.apache.org/licenses/LICENSE-2.0 | |
# | |
# Unless required by applicable law or agreed to in writing, software | |
# distributed under the License is distributed on an "AS IS" BASIS, | |
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
# See the License for the specific language governing permissions and | |
# limitations under the License. | |
"""Google Sentence Compression dataset""" | |
import gzip | |
import json | |
import datasets | |
_CITATION = """\ | |
@inproceedings{filippova-altun-2013-overcoming, | |
title = "Overcoming the Lack of Parallel Data in Sentence Compression", | |
author = "Filippova, Katja and | |
Altun, Yasemin", | |
booktitle = "Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing", | |
month = oct, | |
year = "2013", | |
address = "Seattle, Washington, USA", | |
publisher = "Association for Computational Linguistics", | |
url = "https://www.aclweb.org/anthology/D13-1155", | |
pages = "1481--1491", | |
} | |
""" | |
_DESCRIPTION = """\ | |
Large corpus of uncompressed and compressed sentences from news articles. | |
""" | |
_HOMEPAGE = "https://github.com/google-research-datasets/sentence-compression" | |
_URLs = { | |
datasets.Split.VALIDATION: [ | |
"https://github.com/google-research-datasets/sentence-compression/raw/master/data/comp-data.eval.json.gz" | |
], | |
datasets.Split.TRAIN: [ | |
f"https://github.com/google-research-datasets/sentence-compression/raw/master/data/sent-comp.train{str(i).zfill(2)}.json.gz" | |
for i in range(1, 11) | |
], | |
} | |
class SentComp(datasets.GeneratorBasedBuilder): | |
"""Google Setence Compression dataset""" | |
def _info(self): | |
node_features = { | |
"form": datasets.Value("string"), | |
"type": datasets.Value("string"), | |
"mid": datasets.Value("string"), | |
"word": datasets.features.Sequence( | |
{ | |
"id": datasets.Value("int32"), | |
"form": datasets.Value("string"), | |
"stem": datasets.Value("string"), | |
"tag": datasets.Value("string"), | |
} | |
), | |
"gender": datasets.Value("int32"), | |
"head_word_index": datasets.Value("int32"), | |
} | |
compression_edge_features = { | |
"parent_id": datasets.Value("int32"), | |
"child_id": datasets.Value("int32"), | |
} | |
edge_features = {**compression_edge_features, "label": datasets.Value("string")} | |
entity_features = { | |
"start": datasets.Value("int32"), | |
"end": datasets.Value("int32"), | |
"head": datasets.Value("int32"), | |
"name": datasets.Value("string"), | |
"type": datasets.Value("string"), | |
"mid": datasets.Value("string"), | |
"is_proper_name_entity": datasets.Value("bool"), | |
"gender": datasets.Value("int32"), | |
} | |
tree_features = { | |
"id": datasets.Value("string"), | |
"sentence": datasets.Value("string"), | |
"node": datasets.features.Sequence(node_features), | |
"edge": datasets.features.Sequence(edge_features), | |
"entity_mention": datasets.features.Sequence(entity_features), | |
} | |
compression_features = { | |
"text": datasets.Value("string"), | |
"edge": datasets.features.Sequence(compression_edge_features), | |
} | |
return datasets.DatasetInfo( | |
description=_DESCRIPTION, | |
features=datasets.Features( | |
{ | |
"graph": tree_features, | |
"compression": compression_features, | |
"headline": datasets.Value("string"), | |
"compression_ratio": datasets.Value("float"), | |
"doc_id": datasets.Value("string"), | |
"source_tree": tree_features, | |
"compression_untransformed": compression_features, | |
} | |
), | |
supervised_keys=None, | |
homepage=_HOMEPAGE, | |
citation=_CITATION, | |
) | |
def _split_generators(self, dl_manager): | |
"""Returns SplitGenerators.""" | |
return [ | |
datasets.SplitGenerator( | |
name=split, | |
# These kwargs will be passed to _generate_examples | |
gen_kwargs={"filepaths": dl_manager.download(_URLs[split])}, | |
) | |
for split in _URLs | |
] | |
def _generate_examples(self, filepaths): | |
"""Yields examples.""" | |
id_ = -1 | |
for ix, filepath in enumerate(filepaths): | |
with gzip.open(filepath, mode="rt", encoding="utf-8") as f: | |
all_text = f.read() | |
# in the data file, it's in the form of JSON objects, separated with '\n\n' characters | |
# we'll format the file to be able to read with json package | |
all_text = "[" + all_text + "]" | |
all_text = all_text.replace("}\n\n{", "},\n{") | |
samples = json.loads(all_text) | |
for sample in samples: | |
# add some default values | |
for node in sample["graph"]["node"] + sample["source_tree"]["node"]: | |
if "type" not in node: | |
node["type"] = "" | |
if "mid" not in node: | |
node["mid"] = "" | |
id_ += 1 | |
yield id_, sample | |