|
"""TODO(x_stance): Add a description here.""" |
|
|
|
|
|
import json |
|
import os |
|
|
|
import datasets |
|
|
|
|
|
|
|
_CITATION = """\ |
|
@inproceedings{vamvas2020xstance, |
|
author = "Vamvas, Jannis and Sennrich, Rico", |
|
title = "{X-Stance}: A Multilingual Multi-Target Dataset for Stance Detection", |
|
booktitle = "Proceedings of the 5th Swiss Text Analytics Conference (SwissText) \\& 16th Conference on Natural Language Processing (KONVENS)", |
|
address = "Zurich, Switzerland", |
|
year = "2020", |
|
month = "jun", |
|
url = "http://ceur-ws.org/Vol-2624/paper9.pdf" |
|
} |
|
""" |
|
|
|
|
|
_DESCRIPTION = """\ |
|
The x-stance dataset contains more than 150 political questions, and 67k comments written by candidates on those questions. |
|
|
|
It can be used to train and evaluate stance detection systems. |
|
|
|
""" |
|
|
|
_URL = "https://github.com/ZurichNLP/xstance/raw/v1.0.0/data/xstance-data-v1.0.zip" |
|
|
|
|
|
class XStance(datasets.GeneratorBasedBuilder): |
|
"""TODO(x_stance): Short description of my dataset.""" |
|
|
|
|
|
VERSION = datasets.Version("0.1.0") |
|
|
|
def _info(self): |
|
|
|
return datasets.DatasetInfo( |
|
|
|
description=_DESCRIPTION, |
|
|
|
features=datasets.Features( |
|
{ |
|
"question": datasets.Value("string"), |
|
"id": datasets.Value("int32"), |
|
"question_id": datasets.Value("int32"), |
|
"language": datasets.Value("string"), |
|
"comment": datasets.Value("string"), |
|
"label": datasets.Value("string"), |
|
"numerical_label": datasets.Value("int32"), |
|
"author": datasets.Value("string"), |
|
"topic": datasets.Value("string") |
|
|
|
} |
|
), |
|
|
|
|
|
|
|
supervised_keys=None, |
|
|
|
homepage="https://github.com/ZurichNLP/xstance", |
|
citation=_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
"""Returns SplitGenerators.""" |
|
|
|
|
|
|
|
dl_dir = dl_manager.download_and_extract(_URL) |
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
|
|
gen_kwargs={"filepath": os.path.join(dl_dir, "train.jsonl")}, |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TEST, |
|
|
|
gen_kwargs={"filepath": os.path.join(dl_dir, "test.jsonl")}, |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.VALIDATION, |
|
|
|
gen_kwargs={"filepath": os.path.join(dl_dir, "valid.jsonl")}, |
|
), |
|
] |
|
|
|
def _generate_examples(self, filepath): |
|
"""Yields examples.""" |
|
|
|
with open(filepath, encoding="utf-8") as f: |
|
for id_, row in enumerate(f): |
|
data = json.loads(row) |
|
|
|
yield id_, { |
|
"id": data["id"], |
|
"question_id": data["question_id"], |
|
"question": data["question"], |
|
"comment": data["comment"], |
|
"label": data["label"], |
|
"author": data["author"], |
|
"numerical_label": data["numerical_label"], |
|
"topic": data["topic"], |
|
"language": data["language"], |
|
} |
|
|