# coding=utf-8 # Copyright 2022 The HuggingFace Datasets Authors and the current dataset script contributor. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ The data came from the GENIA version 3.02 corpus (Kim et al., 2003). This was formed from a controlled search on MEDLINE using the MeSH terms human, blood cells and transcription factors. From this search 2,000 abstracts were selected and hand annotated according to a small taxonomy of 48 classes based on a chemical classification. Among the classes, 36 terminal classes were used to annotate the GENIA corpus. """ from typing import Dict, List, Tuple import datasets from .bigbiohub import kb_features from .bigbiohub import BigBioConfig from .bigbiohub import Tasks _LANGUAGES = ['English'] _PUBMED = True _LOCAL = False # TODO: Add BibTeX citation _CITATION = """\ @inproceedings{collier-kim-2004-introduction, title = "Introduction to the Bio-entity Recognition Task at {JNLPBA}", author = "Collier, Nigel and Kim, Jin-Dong", booktitle = "Proceedings of the International Joint Workshop on Natural Language Processing in Biomedicine and its Applications ({NLPBA}/{B}io{NLP})", month = aug # " 28th and 29th", year = "2004", address = "Geneva, Switzerland", publisher = "COLING", url = "https://aclanthology.org/W04-1213", pages = "73--78", } """ _DATASETNAME = "jnlpba" _DISPLAYNAME = "JNLPBA" _DESCRIPTION = """\ NER For Bio-Entities """ _HOMEPAGE = "http://www.geniaproject.org/shared-tasks/bionlp-jnlpba-shared-task-2004" _LICENSE = 'Creative Commons Attribution 3.0 Unported' _URLS = { _DATASETNAME: "http://www.nactem.ac.uk/GENIA/current/Shared-tasks/JNLPBA/Train/Genia4ERtraining.tar.gz", } # TODO: add supported task by dataset. One dataset may support multiple tasks _SUPPORTED_TASKS = [ Tasks.NAMED_ENTITY_RECOGNITION ] # example: [Tasks.TRANSLATION, Tasks.NAMED_ENTITY_RECOGNITION, Tasks.RELATION_EXTRACTION] # TODO: set this to a version that is associated with the dataset. if none exists use "1.0.0" # This version doesn't have to be consistent with semantic versioning. Anything that is # provided by the original dataset as a version goes. _SOURCE_VERSION = "3.2.0" _BIGBIO_VERSION = "1.0.0" class JNLPBADataset(datasets.GeneratorBasedBuilder): """ The data came from the GENIA version 3.02 corpus (Kim et al., 2003). This was formed from a controlled search on MEDLINE using the MeSH terms human, blood cells and transcription factors. From this search 2,000 abstracts were selected and hand annotated according to a small taxonomy of 48 classes based on a chemical classification. Among the classes, 36 terminal classes were used to annotate the GENIA corpus. """ SOURCE_VERSION = datasets.Version(_SOURCE_VERSION) BIGBIO_VERSION = datasets.Version(_BIGBIO_VERSION) BUILDER_CONFIGS = [ BigBioConfig( name="jnlpba_source", version=SOURCE_VERSION, description="jnlpba source schema", schema="source", subset_id="jnlpba", ), BigBioConfig( name="jnlpba_bigbio_kb", version=BIGBIO_VERSION, description="jnlpba BigBio schema", schema="bigbio_kb", subset_id="jnlpba", ), ] DEFAULT_CONFIG_NAME = "jnlpba_source" def _info(self) -> datasets.DatasetInfo: if self.config.schema == "source": features = datasets.load_dataset("jnlpba", split="train").features elif self.config.schema == "bigbio_kb": features = kb_features return datasets.DatasetInfo( description=_DESCRIPTION, features=features, homepage=_HOMEPAGE, license=str(_LICENSE), citation=_CITATION, ) def _split_generators(self, dl_manager) -> List[datasets.SplitGenerator]: """Returns SplitGenerators.""" data = datasets.load_dataset("jnlpba") return [ datasets.SplitGenerator( name=datasets.Split.TRAIN, # Whatever you put in gen_kwargs will be passed to _generate_examples gen_kwargs={"data": data["train"]}, ), datasets.SplitGenerator( name=datasets.Split.VALIDATION, gen_kwargs={"data": data["validation"]}, ), ] def _generate_examples(self, data: datasets.Dataset) -> Tuple[int, Dict]: """Yields examples as (key, example) tuples.""" uid = 0 if self.config.schema == "source": for key, sample in enumerate(data): yield key, sample elif self.config.schema == "bigbio_kb": for i, sample in enumerate(data): feature_dict = { "id": uid, "document_id": "NULL", "passages": [], "entities": [], "relations": [], "events": [], "coreferences": [], } uid += 1 offset_start = 0 for token, tag in zip(sample["tokens"], sample["ner_tags"]): offset_start += len(token) + 1 feature_dict["entities"].append( { "id": uid, "offsets": [[offset_start, offset_start + len(token)]], "text": [token], "type": tag, "normalized": [], } ) uid += 1 # entities yield i, feature_dict