Datasets:

Languages:
English
Multilinguality:
monolingual
Size Categories:
10K<n<100K
Language Creators:
crowdsourced
Annotations Creators:
crowdsourced
Source Datasets:
original
ArXiv:
License:
hybrid_qa / hybrid_qa.py
albertvillanova's picture
Add missing JSON data files (#9)
27cd56d
raw history blame
No virus
6.66 kB
# coding=utf-8
# Copyright 2020 The HuggingFace Datasets Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""HybridQA: A Dataset of Multi-Hop Question Answering over Tabular and Textual Data"""
import json
import os
import datasets
_CITATION = """\
@article{chen2020hybridqa,
title={HybridQA: A Dataset of Multi-Hop Question Answering over Tabular and Textual Data},
author={Chen, Wenhu and Zha, Hanwen and Chen, Zhiyu and Xiong, Wenhan and Wang, Hong and Wang, William},
journal={Findings of EMNLP 2020},
year={2020}
}
"""
_DESCRIPTION = """\
Existing question answering datasets focus on dealing with homogeneous information, based either only on text or \
KB/Table information alone. However, as human knowledge is distributed over heterogeneous forms, \
using homogeneous information alone might lead to severe coverage problems. \
To fill in the gap, we present HybridQA, a new large-scale question-answering dataset that \
requires reasoning on heterogeneous information. Each question is aligned with a Wikipedia table \
and multiple free-form corpora linked with the entities in the table. The questions are designed \
to aggregate both tabular information and text information, i.e., \
lack of either form would render the question unanswerable.
"""
_HOMEPAGE = "https://hybridqa.github.io/index.html"
_WIKI_TABLES_GIT_ARCHIVE_URL = "WikiTables-WithLinks-f4ed68e54e25c495f63d309de0b89c0f97b3c508.zip"
_QA_DATA_BASE_URL = "https://raw.githubusercontent.com/wenhuchen/HybridQA/master/released_data"
_URLS = {
"train": f"{_QA_DATA_BASE_URL}/train.json",
"dev": f"{_QA_DATA_BASE_URL}/dev.json",
"test": f"{_QA_DATA_BASE_URL}/test.json",
}
class HybridQa(datasets.GeneratorBasedBuilder):
"""HybridQA: A Dataset of Multi-Hop Question Answering over Tabular and Textual Data"""
VERSION = datasets.Version("1.0.0")
BUILDER_CONFIGS = [
datasets.BuilderConfig(
name="hybrid_qa",
version=datasets.Version("1.0.0"),
),
]
def _info(self):
features = {
"question_id": datasets.Value("string"),
"question": datasets.Value("string"),
"table_id": datasets.Value("string"),
"answer_text": datasets.Value("string"),
"question_postag": datasets.Value("string"),
"table": {
"url": datasets.Value("string"),
"title": datasets.Value("string"),
"header": datasets.Sequence(datasets.Value("string")),
"data": [
{
"value": datasets.Value("string"),
"urls": [{"url": datasets.Value("string"), "summary": datasets.Value("string")}],
}
],
"section_title": datasets.Value("string"),
"section_text": datasets.Value("string"),
"uid": datasets.Value("string"),
"intro": datasets.Value("string"),
},
}
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=datasets.Features(features),
supervised_keys=None,
homepage=_HOMEPAGE,
citation=_CITATION,
)
def _split_generators(self, dl_manager):
extracted_path = dl_manager.download_and_extract(_WIKI_TABLES_GIT_ARCHIVE_URL)
downloaded_files = dl_manager.download(_URLS)
repo_path = os.path.join(extracted_path, "WikiTables-WithLinks-f4ed68e54e25c495f63d309de0b89c0f97b3c508")
tables_path = os.path.join(repo_path, "tables_tok")
requests_path = os.path.join(repo_path, "request_tok")
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN,
gen_kwargs={
"qa_filepath": downloaded_files["train"],
"tables_path": tables_path,
"requests_path": requests_path,
},
),
datasets.SplitGenerator(
name=datasets.Split.VALIDATION,
gen_kwargs={
"qa_filepath": downloaded_files["dev"],
"tables_path": tables_path,
"requests_path": requests_path,
},
),
datasets.SplitGenerator(
name=datasets.Split.TEST,
gen_kwargs={
"qa_filepath": downloaded_files["test"],
"tables_path": tables_path,
"requests_path": requests_path,
},
),
]
def _generate_examples(self, qa_filepath, tables_path, requests_path):
with open(qa_filepath, encoding="utf-8") as f:
examples = json.load(f)
for example in examples:
table_id = example["table_id"]
table_file_path = os.path.join(tables_path, f"{table_id}.json")
url_data_path = os.path.join(requests_path, f"{table_id}.json")
try:
with open(table_file_path, encoding="utf-8") as f:
table = json.load(f)
with open(url_data_path, encoding="utf-8") as f:
url_data = json.load(f)
except FileNotFoundError: # Some JSON files were not properly added to the GitHub repo: filenames with ':', '"'
continue
table["header"] = [header[0] for header in table["header"]]
# here each row is a list with two elemets, the row value and list of urls for that row
# convert it to list of dict with keys value and urls
rows = []
for row in table["data"]:
for col in row:
new_row = {"value": col[0]}
urls = col[1]
new_row["urls"] = [{"url": url, "summary": url_data[url]} for url in urls]
rows.append(new_row)
table["data"] = rows
example["answer_text"] = example.pop("answer-text") if "answer-text" in example else ""
example["table"] = table
yield example["question_id"], example