assin2 / assin2.py
ruanchaves's picture
ASSIN 2 dataset: replace broken Google Drive _URLS by links on github (#4004)
af0d803
raw
history blame
4.24 kB
# coding=utf-8
# Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""ASSIN 2 dataset."""
import xml.etree.ElementTree as ET
import datasets
_CITATION = """
@inproceedings{real2020assin,
title={The assin 2 shared task: a quick overview},
author={Real, Livy and Fonseca, Erick and Oliveira, Hugo Goncalo},
booktitle={International Conference on Computational Processing of the Portuguese Language},
pages={406--412},
year={2020},
organization={Springer}
}
"""
_DESCRIPTION = """
The ASSIN 2 corpus is composed of rather simple sentences. Following the procedures of SemEval 2014 Task 1.
The training and validation data are composed, respectively, of 6,500 and 500 sentence pairs in Brazilian Portuguese,
annotated for entailment and semantic similarity. Semantic similarity values range from 1 to 5, and text entailment
classes are either entailment or none. The test data are composed of approximately 3,000 sentence pairs with the same
annotation. All data were manually annotated.
"""
_HOMEPAGE = "https://sites.google.com/view/assin2"
_LICENSE = ""
_URLS = {
"train": "https://github.com/ruanchaves/assin/raw/master/sources/assin2-train-only.xml",
"dev": "https://github.com/ruanchaves/assin/raw/master/sources/assin2-dev.xml",
"test": "https://github.com/ruanchaves/assin/raw/master/sources/assin2-test.xml",
}
class Assin2(datasets.GeneratorBasedBuilder):
"""ASSIN 2 dataset."""
VERSION = datasets.Version("1.0.0")
def _info(self):
features = datasets.Features(
{
"sentence_pair_id": datasets.Value("int64"),
"premise": datasets.Value("string"),
"hypothesis": datasets.Value("string"),
"relatedness_score": datasets.Value("float32"),
"entailment_judgment": datasets.features.ClassLabel(names=["NONE", "ENTAILMENT"]),
}
)
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=features,
supervised_keys=None,
homepage=_HOMEPAGE,
license=_LICENSE,
citation=_CITATION,
)
def _split_generators(self, dl_manager):
"""Returns SplitGenerators."""
data_dir = dl_manager.download(_URLS)
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN,
gen_kwargs={
"filepath": data_dir["train"],
"split": "train",
},
),
datasets.SplitGenerator(
name=datasets.Split.TEST,
gen_kwargs={
"filepath": data_dir["test"],
"split": "test",
},
),
datasets.SplitGenerator(
name=datasets.Split.VALIDATION,
gen_kwargs={
"filepath": data_dir["dev"],
"split": "dev",
},
),
]
def _generate_examples(self, filepath, split):
"""Yields examples."""
id_ = 0
with open(filepath, "rb") as f:
tree = ET.parse(f)
root = tree.getroot()
for pair in root:
yield id_, {
"sentence_pair_id": int(pair.attrib.get("id")),
"premise": pair.find(".//t").text,
"hypothesis": pair.find(".//h").text,
"relatedness_score": float(pair.attrib.get("similarity")),
"entailment_judgment": pair.attrib.get("entailment").upper(),
}
id_ += 1