| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| |
| """FEVER dataset.""" |
|
|
| import json |
| import os |
| import textwrap |
|
|
| import datasets |
|
|
|
|
| class FeverConfig(datasets.BuilderConfig): |
| """BuilderConfig for FEVER.""" |
|
|
| def __init__(self, homepage: str = None, citation: str = None, base_url: str = None, urls: dict = None, **kwargs): |
| """BuilderConfig for FEVER. |
| |
| Args: |
| homepage (`str`): Homepage. |
| citation (`str`): Citation reference. |
| base_url (`str`): Data base URL that precedes all data URLs. |
| urls (`dict`): Data URLs (each URL will pe preceded by `base_url`). |
| **kwargs: keyword arguments forwarded to super. |
| """ |
| super().__init__(**kwargs) |
| self.homepage = homepage |
| self.citation = citation |
| self.base_url = base_url |
| self.urls = {key: f"{base_url}/{url}" for key, url in urls.items()} |
|
|
|
|
| class Fever(datasets.GeneratorBasedBuilder): |
| """Fact Extraction and VERification Dataset.""" |
| BUILDER_CONFIGS = [ |
| FeverConfig( |
| name="v1.0", |
| version=datasets.Version("1.0.0"), |
| description=textwrap.dedent( |
| "FEVER v1.0\n" |
| "FEVER (Fact Extraction and VERification) consists of 185,445 claims generated by altering sentences " |
| "extracted from Wikipedia and subsequently verified without knowledge of the sentence they were " |
| "derived from. The claims are classified as Supported, Refuted or NotEnoughInfo. For the first two " |
| "classes, the annotators also recorded the sentence(s) forming the necessary evidence for their " |
| "judgment." |
| ), |
| homepage="https://fever.ai/dataset/fever.html", |
| citation=textwrap.dedent( |
| """\ |
| @inproceedings{Thorne18Fever, |
| author = {Thorne, James and Vlachos, Andreas and Christodoulopoulos, Christos and Mittal, Arpit}, |
| title = {{FEVER}: a Large-scale Dataset for Fact Extraction and {VERification}}, |
| booktitle = {NAACL-HLT}, |
| year = {2018} |
| }""" |
| ), |
| base_url="https://fever.ai/download/fever", |
| urls={ |
| datasets.Split.TRAIN: "train.jsonl", |
| "dev": "shared_task_dev.jsonl", |
| "paper_dev": "paper_dev.jsonl", |
| "paper_test": "paper_test.jsonl", |
| }, |
| ), |
| FeverConfig( |
| name="v2.0", |
| version=datasets.Version("2.0.0"), |
| description=textwrap.dedent( |
| "FEVER v2.0:\n" |
| "The FEVER 2.0 Dataset consists of 1174 claims created by the submissions of participants in the " |
| "Breaker phase of the 2019 shared task. Participants (Breakers) were tasked with generating " |
| "adversarial examples that induce classification errors for the existing systems. Breakers submitted " |
| "a dataset of up to 1000 instances with equal number of instances for each of the three classes " |
| "(Supported, Refuted NotEnoughInfo). Only novel claims (i.e. not contained in the original FEVER " |
| "dataset) were considered as valid entries to the shared task. The submissions were then manually " |
| "evaluated for Correctness (grammatical, appropriately labeled and meet the FEVER annotation " |
| "guidelines requirements)." |
| ), |
| homepage="https://fever.ai/dataset/adversarial.html", |
| citation=textwrap.dedent( |
| """\ |
| @inproceedings{Thorne19FEVER2, |
| author = {Thorne, James and Vlachos, Andreas and Cocarascu, Oana and Christodoulopoulos, Christos and Mittal, Arpit}, |
| title = {The {FEVER2.0} Shared Task}, |
| booktitle = {Proceedings of the Second Workshop on {Fact Extraction and VERification (FEVER)}}, |
| year = {2018} |
| }""" |
| ), |
| base_url="https://fever.ai/download/fever2.0", |
| urls={ |
| datasets.Split.VALIDATION: "fever2-fixers-dev.jsonl", |
| }, |
| ), |
| FeverConfig( |
| name="wiki_pages", |
| version=datasets.Version("1.0.0"), |
| description=textwrap.dedent( |
| "Wikipedia pages for FEVER v1.0:\n" |
| "FEVER (Fact Extraction and VERification) consists of 185,445 claims generated by altering sentences " |
| "extracted from Wikipedia and subsequently verified without knowledge of the sentence they were " |
| "derived from. The claims are classified as Supported, Refuted or NotEnoughInfo. For the first two " |
| "classes, the annotators also recorded the sentence(s) forming the necessary evidence for their " |
| "judgment." |
| ), |
| homepage="https://fever.ai/dataset/fever.html", |
| citation=textwrap.dedent( |
| """\ |
| @inproceedings{Thorne18Fever, |
| author = {Thorne, James and Vlachos, Andreas and Christodoulopoulos, Christos and Mittal, Arpit}, |
| title = {{FEVER}: a Large-scale Dataset for Fact Extraction and {VERification}}, |
| booktitle = {NAACL-HLT}, |
| year = {2018} |
| }""" |
| ), |
| base_url="https://fever.ai/download/fever", |
| urls={ |
| "wikipedia_pages": "wiki-pages.zip", |
| }, |
| ), |
| ] |
|
|
| def __init__(self, *args, **kwargs): |
| super().__init__(*args, **kwargs) |
| self.seen_claims= set() |
|
|
| def _info(self): |
| if |
| config.name == "wiki_pages": |
| features = { |
| "id": datasets.Value("string"), |
| "text": datasets.Value("string"), |
| "lines": datasets.Value("string"), |
| } |
| elif |
| config.name == "v1.0" or |
| config.name == "v2.0": |
| features = { |
| "id": datasets.Value("int32"), |
| "label": datasets.ClassLabel(names=["REFUTES", "SUPPORTS"]), |
| "claim": datasets.Value("string"), |
| "evidence_annotation_id": datasets.Value("int32"), |
| "evidence_id": datasets.Value("int32"), |
| "evidence_wiki_url": datasets.Value("string"), |
| "evidence_sentence_id": datasets.Value("int32"), |
| } |
| return datasets.DatasetInfo( |
| description= |
| config.description, |
| features=datasets.Features(features), |
| homepage= |
| config.homepage, |
| citation= |
| config.citation, |
| ) |
|
|
| def _split_generators(self, dl_manager): |
| """Returns SplitGenerators.""" |
| dl_paths = dl_manager.download_and_extract(self.config.urls) |
| return [ |
| datasets.SplitGenerator( |
| name=split, |
| gen_kwargs={ |
| "filepath": dl_paths[split] |
| if self.config.name != "wiki_pages" |
| else dl_manager.iter_files(os.path.join(dl_paths[split], "wiki-pages")), |
| }, |
| ) |
| for split in dl_paths.keys() |
| ] |
|
|
| def _generate_examples(self, filepath): |
| """Yields examples.""" |
| if self.config.name == "v1.0" or self.config.name == "v2.0": |
| with open(filepath, encoding="utf-8") as f: |
| for row_id, row in enumerate(f): |
| data = json.loads(row) |
| id_ = data["id"] |
| label = data.get("label", "") |
|
|
| |
| if label not in ("REFUTES", "SUPPORTS"): |
| continue |
|
|
| claim = data["claim"] |
| if claim in self.seen_claims: |
| continue |
| else: |
| self.seen_claims.add(claim) |
| evidences = data.get("evidence", []) |
| if len(evidences) > 0: |
| for i in range(len(evidences)): |
| for j in range(len(evidences[i])): |
| annot_id = evidences[i][j][0] if evidences[i][j][0] else -1 |
| evidence_id = evidences[i][j][1] if evidences[i][j][1] else -1 |
| wiki_url = evidences[i][j][2] if evidences[i][j][2] else "" |
| sent_id = evidences[i][j][3] if evidences[i][j][3] else -1 |
| yield str(row_id) + "_" + str(i) + "_" + str(j), { |
| "id": id_, |
| "label": label, |
| "claim": claim, |
| "evidence_annotation_id": annot_id, |
| "evidence_id": evidence_id, |
| "evidence_wiki_url": wiki_url, |
| "evidence_sentence_id": sent_id, |
| } |
| else: |
| yield row_id, { |
| "id": id_, |
| "label": label, |
| "claim": claim, |
| "evidence_annotation_id": -1, |
| "evidence_id": -1, |
| "evidence_wiki_url": "", |
| "evidence_sentence_id": -1, |
| } |
| elif self.config.name == "wiki_pages": |
| for file_id, file in enumerate(filepath): |
| with open(file, encoding="utf-8") as f: |
| for row_id, row in enumerate(f): |
| data = json.loads(row) |
| yield f"{file_id}_{row_id}", data |
|
|