"""DocRED: A Large-Scale Document-Level Relation Extraction Dataset""" import json import os import datasets _CITATION = """\ @inproceedings{yao2019DocRED, title={{DocRED}: A Large-Scale Document-Level Relation Extraction Dataset}, author={Yao, Yuan and Ye, Deming and Li, Peng and Han, Xu and Lin, Yankai and Liu, Zhenghao and Liu, \ Zhiyuan and Huang, Lixin and Zhou, Jie and Sun, Maosong}, booktitle={Proceedings of ACL 2019}, year={2019} } """ _DESCRIPTION = """\ Multiple entities in a document generally exhibit complex inter-sentence relations, and cannot be well handled by \ existing relation extraction (RE) methods that typically focus on extracting intra-sentence relations for single \ entity pairs. In order to accelerate the research on document-level RE, we introduce DocRED, a new dataset constructed \ from Wikipedia and Wikidata with three features: - DocRED annotates both named entities and relations, and is the largest human-annotated dataset for document-level RE from plain text. - DocRED requires reading multiple sentences in a document to extract entities and infer their relations by synthesizing all information of the document. - Along with the human-annotated data, we also offer large-scale distantly supervised data, which enables DocRED to be adopted for both supervised and weakly supervised scenarios. """ _URLS = { "dev": "https://drive.google.com/uc?export=download&id=1AHUm1-_V9GCtGuDcc8XrMUCJE8B-HHoL", "train_distant": "https://drive.google.com/uc?export=download&id=1Qr4Jct2IJ9BVI86_mCk_Pz0J32ww9dYw", "train_annotated": "https://drive.google.com/uc?export=download&id=1NN33RzyETbanw4Dg2sRrhckhWpzuBQS9", "test": "https://drive.google.com/uc?export=download&id=1lAVDcD94Sigx7gR3jTfStI66o86cflum", "rel_info": "https://drive.google.com/uc?id=1y9A0zKrvETc1ddUFuFhBg3Xfr7FEL4dW&export=download", } class DocRed(datasets.GeneratorBasedBuilder): """DocRED: A Large-Scale Document-Level Relation Extraction Dataset""" def _info(self): return datasets.DatasetInfo( description=_DESCRIPTION, features=datasets.Features( { "title": datasets.Value("string"), "sents": datasets.features.Sequence(datasets.features.Sequence(datasets.Value("string"))), "vertexSet": [ [ { "name": datasets.Value("string"), "sent_id": datasets.Value("int32"), "pos": datasets.features.Sequence(datasets.Value("int32")), "type": datasets.Value("string"), } ] ], "labels": datasets.features.Sequence( { "head": datasets.Value("int32"), "tail": datasets.Value("int32"), "relation_id": datasets.Value("string"), "relation_text": datasets.Value("string"), "evidence": datasets.features.Sequence(datasets.Value("int32")), } ), } ), supervised_keys=None, homepage="https://github.com/thunlp/DocRED", citation=_CITATION, ) def _split_generators(self, dl_manager): downloads = {} for key in _URLS.keys(): downloads[key] = dl_manager.download_and_extract(_URLS[key]) # Fix for dummy data if os.path.isdir(downloads[key]): downloads[key] = os.path.join(downloads[key], key + ".json") return [ datasets.SplitGenerator( name=datasets.Split.VALIDATION, gen_kwargs={"filepath": downloads["dev"], "rel_info": downloads["rel_info"]}, ), datasets.SplitGenerator( name=datasets.Split.TEST, gen_kwargs={"filepath": downloads["test"], "rel_info": downloads["rel_info"]} ), datasets.SplitGenerator( name="train_annotated", gen_kwargs={"filepath": downloads["train_annotated"], "rel_info": downloads["rel_info"]}, ), datasets.SplitGenerator( name="train_distant", gen_kwargs={"filepath": downloads["train_distant"], "rel_info": downloads["rel_info"]}, ), ] def _generate_examples(self, filepath, rel_info): """Generate DocRED examples.""" with open(rel_info, encoding="utf-8") as f: relation_name_map = json.load(f) with open(filepath, encoding="utf-8") as f: data = json.load(f) for idx, example in enumerate(data): # Test set has no labels - Results need to be uploaded to Codalab if "labels" not in example.keys(): example["labels"] = [] for label in example["labels"]: # Rename and include full relation names label["relation_text"] = relation_name_map[label["r"]] label["relation_id"] = label["r"] label["head"] = label["h"] label["tail"] = label["t"] del label["r"] del label["h"] del label["t"] yield idx, example