Datasets:
Tasks:
Question Answering
Sub-tasks:
open-domain-qa
Multilinguality:
multilingual
Size Categories:
10K<n<100K
Annotations Creators:
crowdsourced
ArXiv:
Tags:
License:
"""XOR QA: Cross-lingual Open-Retrieval Question Answering""" | |
import json | |
import textwrap | |
import datasets | |
_XOR_TYDI_QA_CITATION = """\ | |
@misc{asai2020xor, | |
title={XOR QA: Cross-lingual Open-Retrieval Question Answering}, | |
author={Akari Asai and Jungo Kasai and Jonathan H. Clark and Kenton Lee and Eunsol Choi and Hannaneh Hajishirzi}, | |
year={2020}, | |
eprint={2010.11856}, | |
archivePrefix={arXiv}, | |
primaryClass={cs.CL} | |
} | |
""" | |
_XOR_TYDI_QA_DESCRIPTION = """\ | |
XOR-TyDi QA brings together for the first time information-seeking questions, | |
open-retrieval QA, and multilingual QA to create a multilingual open-retrieval | |
QA dataset that enables cross-lingual answer retrieval. It consists of questions | |
written by information-seeking native speakers in 7 typologically diverse languages | |
and answer annotations that are retrieved from multilingual document collections. | |
There are three sub-tasks: XOR-Retrieve, XOR-EnglishSpan, and XOR-Full. | |
""" | |
_DESCRIPTIONS = { | |
"xor-retrieve": textwrap.dedent( | |
"""\ | |
XOR-Retrieve is a cross-lingual retrieval task where a question is written in the target | |
language (e.g., Japanese) and a system is required to retrieve English document that answers the question. | |
""" | |
), | |
"xor-full": textwrap.dedent( | |
"""\ | |
XOR-Full is a cross-lingual retrieval task where a question is written in the target | |
language (e.g., Japanese) and a system is required to output a short answer in the target language.""" | |
), | |
} | |
_DATA_URLS = { | |
"xor-retrieve": { | |
"train": "https://nlp.cs.washington.edu/xorqa/XORQA_site/data/xor_train_retrieve_eng_span.jsonl", | |
"dev": "https://nlp.cs.washington.edu/xorqa/XORQA_site/data/xor_dev_retrieve_eng_span_v1_1.jsonl", | |
"test": "https://nlp.cs.washington.edu/xorqa/XORQA_site/data/xor_test_retrieve_eng_span_q_only_v1_1.jsonl", | |
}, | |
"xor-full": { | |
"train": "https://nlp.cs.washington.edu/xorqa/XORQA_site/data/xor_train_full.jsonl", | |
"dev": "https://nlp.cs.washington.edu/xorqa/XORQA_site/data/xor_dev_full_v1_1.jsonl", | |
"test": "https://nlp.cs.washington.edu/xorqa/XORQA_site/data/xor_test_full_q_only_v1_1.jsonl", | |
}, | |
} | |
_XOR_TYDI_QA_URL = "https://nlp.cs.washington.edu/xorqa/" | |
class XORTyDiConfig(datasets.BuilderConfig): | |
"BuilderConfig for XOR-TyDi Dataset" | |
def __init__(self, data_url, citation, url, **kwargs): | |
""" | |
Args: | |
data_url: `dictionary`, dict with url for each split of data. | |
citation: `string`, citation for the dataset. | |
url: `string`, url for information about the dataset. | |
**kwargs: keyword arguments forwarded to super. | |
""" | |
super(XORTyDiConfig, self).__init__(version=datasets.Version("1.1.0", ""), **kwargs) | |
self.data_url = data_url | |
self.citation = citation | |
self.url = url | |
class XORTyDi(datasets.GeneratorBasedBuilder): | |
BUILDER_CONFIGS = [ | |
XORTyDiConfig( | |
name=name, | |
description=_DESCRIPTIONS[name], | |
data_url=_DATA_URLS[name], | |
citation=_XOR_TYDI_QA_CITATION, | |
url=_XOR_TYDI_QA_URL, | |
) | |
for name in ["xor-retrieve", "xor-full"] | |
] | |
def _info(self): | |
features = {} | |
features["question"] = datasets.Value("string") | |
features["lang"] = datasets.features.ClassLabel(names=["ar", "bn", "fi", "ja", "ko", "ru", "te"]) | |
features["answers"] = datasets.Value("string") | |
return datasets.DatasetInfo( | |
description=_XOR_TYDI_QA_DESCRIPTION + "\n" + self.config.description, | |
features=datasets.Features(features), | |
homepage=self.config.url, | |
citation=_XOR_TYDI_QA_CITATION, | |
) | |
def _split_generators(self, dl_manager): | |
train = dl_manager.download_and_extract(self.config.data_url["train"]) | |
dev = dl_manager.download_and_extract(self.config.data_url["dev"]) | |
test = dl_manager.download_and_extract(self.config.data_url["test"]) | |
return [ | |
datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": train, "split": "train"}), | |
datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={"filepath": dev, "split": "dev"}), | |
datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"filepath": test, "split": "test"}), | |
] | |
def _generate_examples(self, filepath, split): | |
with open(filepath, encoding="utf-8") as f: | |
jlines = f.read() | |
result = [json.loads(jline) for jline in jlines.splitlines()] | |
if split == "test": | |
for id_, row in enumerate(result): | |
yield id_, {"question": row["question"], "answers": "None", "lang": row["lang"].strip()} | |
else: | |
for id_, row in enumerate(result): | |
yield id_, { | |
"question": row["question"], | |
"answers": " ".join(row["answers"]), | |
"lang": row["lang"].strip(), | |
} | |