""" Dataset from https://github.com/allenai/sequential_sentence_classification Dataset maintainer: @soldni """ import json from typing import Iterable, Sequence, Tuple import datasets from datasets.builder import BuilderConfig, GeneratorBasedBuilder from datasets.info import DatasetInfo from datasets.splits import Split, SplitGenerator from datasets.utils.logging import get_logger LOGGER = get_logger(__name__) _NAME = "CSAbstruct" _CITATION = """\ @inproceedings{Cohan2019EMNLP, title={Pretrained Language Models for Sequential Sentence Classification}, author={Arman Cohan, Iz Beltagy, Daniel King, Bhavana Dalvi, Dan Weld}, year={2019}, booktitle={EMNLP}, } """ _LICENSE = "Apache License 2.0" _DESCRIPTION = """\ As a step toward better document-level understanding, we explore \ classification of a sequence of sentences into their corresponding \ categories, a task that requires understanding sentences in context \ of the document. Recent successful models for this task have used \ hierarchical models to contextualize sentence representations, and \ Conditional Random Fields (CRFs) to incorporate dependencies between \ subsequent labels. In this work, we show that pretrained language \ models, BERT (Devlin et al., 2018) in particular, can be used for \ this task to capture contextual dependencies without the need for \ hierarchical encoding nor a CRF. Specifically, we construct a joint \ sentence representation that allows BERT Transformer layers to \ directly utilize contextual information from all words in all \ sentences. Our approach achieves state-of-the-art results on four \ datasets, including a new dataset of structured scientific abstracts. """ _HOMEPAGE = "https://github.com/allenai/sequential_sentence_classification" _VERSION = "1.0.0" _URL = ( "https://raw.githubusercontent.com/allenai/" "sequential_sentence_classification/master/" ) _SPLITS = { Split.TRAIN: _URL + "data/CSAbstruct/train.jsonl", Split.VALIDATION: _URL + "data/CSAbstruct/dev.jsonl", Split.TEST: _URL + "data/CSAbstruct/test.jsonl", } class CSAbstruct(GeneratorBasedBuilder): """CSAbstruct""" BUILDER_CONFIGS = [ BuilderConfig( name=_NAME, version=datasets.Version(_VERSION), description=_DESCRIPTION, ) ] def _info(self) -> DatasetInfo: class_labels = ["background", "method", "objective", "other", "result"] features = datasets.Features( { "abstract_id": datasets.Value("string"), "sentences": [datasets.Value("string")], "labels": [datasets.ClassLabel(names=class_labels)], "confs": [datasets.Value("float")], } ) return DatasetInfo( description=_DESCRIPTION, features=features, supervised_keys=None, homepage=_HOMEPAGE, license=_LICENSE, citation=_CITATION, ) def _split_generators( self, dl_manager: datasets.DownloadManager ) -> Sequence[SplitGenerator]: archive = dl_manager.download(_SPLITS) return [ SplitGenerator( name=split_name, # type: ignore gen_kwargs={ "split_name": split_name, "filepath": archive[split_name], # type: ignore }, ) for split_name in _SPLITS ] def _generate_examples( self, split_name: str, filepath: str ) -> Iterable[Tuple[str, dict]]: """This function returns the examples in the raw (text) form.""" LOGGER.info(f"generating examples from documents in {filepath}...") with open(filepath, mode="r", encoding="utf-8") as f: data = [json.loads(ln) for ln in f] for i, row in enumerate(data): row["abstract_id"] = f"{split_name}_{i:04d}" yield row["abstract_id"], row