Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
Dask
License:
narrativeqa / narrativeqa.py
system's picture
system HF staff
Update files from the datasets library (from 1.6.0)
0b6eb56
raw
history blame
6.67 kB
# coding=utf-8
# Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
"""NarrativeQA Reading Comprehension Challenge"""
import csv
import os
import datasets
_CITATION = """\
@article{narrativeqa,
author = {Tom\\'a\\v s Ko\\v cisk\\'y and Jonathan Schwarz and Phil Blunsom and
Chris Dyer and Karl Moritz Hermann and G\\'abor Melis and
Edward Grefenstette},
title = {The {NarrativeQA} Reading Comprehension Challenge},
journal = {Transactions of the Association for Computational Linguistics},
url = {https://TBD},
volume = {TBD},
year = {2018},
pages = {TBD},
}
"""
_DESCRIPTION = """\
The NarrativeQA dataset for question answering on long documents (movie scripts, books). It includes the list of documents with Wikipedia summaries, links to full stories, and questions and answers.
"""
_URLS = {
"full_text": "https://storage.googleapis.com/huggingface-nlp/datasets/narrative_qa/narrativeqa_full_text.zip",
"repo": "https://github.com/deepmind/narrativeqa/archive/master.zip",
}
class NarrativeQa(datasets.GeneratorBasedBuilder):
"""NarrativeQA: Question answering on long-documents"""
def _info(self):
return datasets.DatasetInfo(
description=_DESCRIPTION,
citation=_CITATION,
features=datasets.Features(
{
"document": {
"id": datasets.Value("string"),
"kind": datasets.Value("string"),
"url": datasets.Value("string"),
"file_size": datasets.Value("int32"),
"word_count": datasets.Value("int32"),
"start": datasets.Value("string"),
"end": datasets.Value("string"),
"summary": {
"text": datasets.Value("string"),
"tokens": datasets.features.Sequence(datasets.Value("string")),
"url": datasets.Value("string"),
"title": datasets.Value("string"),
},
"text": datasets.Value("string"),
},
"question": {
"text": datasets.Value("string"),
"tokens": datasets.features.Sequence(datasets.Value("string")),
},
"answers": [
{
"text": datasets.Value("string"),
"tokens": datasets.features.Sequence(datasets.Value("string")),
}
],
}
),
homepage="https://github.com/deepmind/narrativeqa",
)
def _split_generators(self, dl_manager):
"""Returns SplitGenerators."""
dl_dir = dl_manager.download_and_extract(_URLS)
dl_dir["repo"] = os.path.join(dl_dir["repo"], "narrativeqa-master")
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN,
gen_kwargs={"repo_dir": dl_dir["repo"], "full_text_dir": dl_dir["full_text"], "split": "train"},
),
datasets.SplitGenerator(
name=datasets.Split.TEST,
gen_kwargs={"repo_dir": dl_dir["repo"], "full_text_dir": dl_dir["full_text"], "split": "test"},
),
datasets.SplitGenerator(
name=datasets.Split.VALIDATION,
gen_kwargs={"repo_dir": dl_dir["repo"], "full_text_dir": dl_dir["full_text"], "split": "valid"},
),
]
def _generate_examples(self, repo_dir, full_text_dir, split):
"""Yields examples."""
documents = {}
with open(os.path.join(repo_dir, "documents.csv"), encoding="utf-8") as f:
reader = csv.DictReader(f)
for row in reader:
if row["set"] != split:
continue
documents[row["document_id"]] = row
summaries = {}
with open(os.path.join(repo_dir, "third_party", "wikipedia", "summaries.csv"), encoding="utf-8") as f:
reader = csv.DictReader(f)
for row in reader:
if row["set"] != split:
continue
summaries[row["document_id"]] = row
with open(os.path.join(repo_dir, "qaps.csv"), encoding="utf-8") as f:
reader = csv.DictReader(f)
for id_, row in enumerate(reader):
if row["set"] != split:
continue
document_id = row["document_id"]
document = documents[document_id]
summary = summaries[document_id]
full_text = open(os.path.join(full_text_dir, document_id + ".content"), encoding="latin-1").read()
res = {
"document": {
"id": document["document_id"],
"kind": document["kind"],
"url": document["story_url"],
"file_size": document["story_file_size"],
"word_count": document["story_word_count"],
"start": document["story_start"],
"end": document["story_end"],
"summary": {
"text": summary["summary"],
"tokens": summary["summary_tokenized"].split(),
"url": document["wiki_url"],
"title": document["wiki_title"],
},
"text": full_text,
},
"question": {"text": row["question"], "tokens": row["question_tokenized"].split()},
"answers": [
{"text": row["answer1"], "tokens": row["answer1_tokenized"].split()},
{"text": row["answer2"], "tokens": row["answer2_tokenized"].split()},
],
}
yield id_, res