# coding=utf-8 # Copyright 2020 The TensorFlow Datasets Authors and the HuggingFace Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # Lint as: python3 """Reuters 21578""" from textwrap import dedent import datasets _HOMEPAGE = "https://archive.ics.uci.edu/dataset/137/reuters+21578+text+categorization+collection" _CITATION = """\ @article{APTE94, author = {Chidanand Apt{\'{e}} and Fred Damerau and Sholom M. Weiss}, title = {Automated Learning of Decision Rules for Text Categorization}, journal = {ACM Transactions on Information Systems}, year = {1994}, note = {To appear.} } @inproceedings{APTE94b, author = {Chidanand Apt{\'{e}} and Fred Damerau and Sholom M. Weiss}, title = {Toward Language Independent Automated Learning of Text Categorization Models}, booktitle = {sigir94}, year = {1994}, note = {To appear.} } @inproceedings{HAYES8}, author = {Philip J. Hayes and Peggy M. Anderson and Irene B. Nirenburg and Linda M. Schmandt}, title = {{TCS}: A Shell for Content-Based Text Categorization}, booktitle = {IEEE Conference on Artificial Intelligence Applications}, year = {1990} } @inproceedings{HAYES90b, author = {Philip J. Hayes and Steven P. Weinstein}, title = {{CONSTRUE/TIS:} A System for Content-Based Indexing of a Database of News Stories}, booktitle = {Second Annual Conference on Innovative Applications of Artificial Intelligence}, year = {1990} } @incollection{HAYES92 , author = {Philip J. Hayes}, title = {Intelligent High-Volume Text Processing using Shallow, Domain-Specific Techniques}, booktitle = {Text-Based Intelligent Systems}, publisher = {Lawrence Erlbaum}, address = {Hillsdale, NJ}, year = {1992}, editor = {Paul S. Jacobs} } @inproceedings{LEWIS91c , author = {David D. Lewis}, title = {Evaluating Text Categorization}, booktitle = {Proceedings of Speech and Natural Language Workshop}, year = {1991}, month = {feb}, organization = {Defense Advanced Research Projects Agency}, publisher = {Morgan Kaufmann}, pages = {312--318} } @phdthesis{LEWIS91d, author = {David Dolan Lewis}, title = {Representation and Learning in Information Retrieval}, school = {Computer Science Dept.; Univ. of Massachusetts; Amherst, MA 01003}, year = 1992}, note = {Technical Report 91--93.} } @inproceedings{LEWIS91e, author = {David D. Lewis}, title = {Data Extraction as Text Categorization: An Experiment with the {MUC-3} Corpus}, booktitle = {Proceedings of the Third Message Understanding Evaluation and Conference}, year = {1991}, month = {may}, organization = {Defense Advanced Research Projects Agency}, publisher = {Morgan Kaufmann}, address = {Los Altos, CA} } @inproceedings{LEWIS92b, author = {David D. Lewis}, title = {An Evaluation of Phrasal and Clustered Representations on a Text Categorization Task}, booktitle = {Fifteenth Annual International ACM SIGIR Conference on Research and Development in Information Retrieval}, year = {1992}, pages = {37--50} } @inproceedings{LEWIS92d , author = {David D. Lewis and Richard M. Tong}, title = {Text Filtering in {MUC-3} and {MUC-4}}, booktitle = {Proceedings of the Fourth Message Understanding Conference ({MUC-4})}, year = {1992}, month = {jun}, organization = {Defense Advanced Research Projects Agency}, publisher = {Morgan Kaufmann}, address = {Los Altos, CA} } @inproceedings{LEWIS92e, author = {David D. Lewis}, title = {Feature Selection and Feature Extraction for Text Categorization}, booktitle = {Proceedings of Speech and Natural Language Workshop}, year = {1992}, month = {feb} , organization = {Defense Advanced Research Projects Agency}, publisher = {Morgan Kaufmann}, pages = {212--217} } @inproceedings{LEWIS94b, author = {David D. Lewis and Marc Ringuette}, title = {A Comparison of Two Learning Algorithms for Text Categorization}, booktitle = {Symposium on Document Analysis and Information Retrieval}, year = {1994}, organization = {ISRI; Univ. of Nevada, Las Vegas}, address = {Las Vegas, NV}, month = {apr}, pages = {81--93} } @article{LEWIS94d, author = {David D. Lewis and Philip J. Hayes}, title = {Guest Editorial}, journal = {ACM Transactions on Information Systems}, year = {1994}, volume = {12}, number = {3}, pages = {231}, month = {jul} } @article{SPARCKJONES76, author = {K. {Sparck Jones} and C. J. {van Rijsbergen}}, title = {Information Retrieval Test Collections}, journal = {Journal of Documentation}, year = {1976}, volume = {32}, number = {1}, pages = {59--75} } @book{WEISS91, author = {Sholom M. Weiss and Casimir A. Kulikowski}, title = {Computer Systems That Learn}, publisher = {Morgan Kaufmann}, year = {1991}, address = {San Mateo, CA} } """ _DESCRIPTION = """\ The Reuters-21578 dataset is one of the most widely used data collections for text categorization research. It is collected from the Reuters financial newswire service in 1987. """ _DATA_URL = "data/reuters21578.tar.gz" class Reuters21578Config(datasets.BuilderConfig): """BuilderConfig for reuters-21578.""" def __init__(self, **kwargs): """BuilderConfig for Reuters21578. Args: **kwargs: keyword arguments forwarded to super. """ super(Reuters21578Config, self).__init__(version=datasets.Version("1.0.0", ""), **kwargs) class Reuters21578(datasets.GeneratorBasedBuilder): """Reuters 21578""" BUILDER_CONFIGS = [ Reuters21578Config( name="ModHayes", description=dedent( """Training Set (20856 docs): CGISPLIT="TRAINING-SET" Test Set (722 docs): CGISPLIT="PUBLISHED-TESTSET" Unused (0 docs)""" ), ), Reuters21578Config( name="ModLewis", description=dedent( """Training Set (13,625 docs): LEWISSPLIT="TRAIN"; TOPICS="YES" or "NO" Test Set (6,188 docs): LEWISSPLIT="TEST"; TOPICS="YES" or "NO" Unused (1,765): LEWISSPLIT="NOT-USED" or TOPICS="BYPASS""" ), ), Reuters21578Config( name="ModApte", description=dedent( """Training Set (9,603 docs): LEWISSPLIT="TRAIN"; TOPICS="YES" Test Set (3,299 docs): LEWISSPLIT="TEST"; TOPICS="YES" Unused (8,676 docs): LEWISSPLIT="NOT-USED"; TOPICS="YES" or TOPICS="NO" or TOPICS="BYPASS" """ ), ), ] def _info(self): return datasets.DatasetInfo( description=_DESCRIPTION, features=datasets.Features( { "text": datasets.Value("string"), "text_type": datasets.Value("string"), "topics": datasets.Sequence(datasets.Value("string")), "lewis_split": datasets.Value("string"), "cgis_split": datasets.Value("string"), "old_id": datasets.Value("string"), "new_id": datasets.Value("string"), "places": datasets.Sequence(datasets.Value("string")), "people": datasets.Sequence(datasets.Value("string")), "orgs": datasets.Sequence(datasets.Value("string")), "exchanges": datasets.Sequence(datasets.Value("string")), "date": datasets.Value("string"), "title": datasets.Value("string"), } ), # No default supervised_keys (as we have to pass both premise # and hypothesis as input). supervised_keys=None, homepage=_HOMEPAGE, citation=_CITATION, ) def _split_generators(self, dl_manager): archive = dl_manager.download(_DATA_URL) filepaths = ["reut2-" + "%03d" % i + ".sgm" for i in range(22)] if self.config.name == "ModHayes": return [ datasets.SplitGenerator( name=datasets.Split.TEST, gen_kwargs={ "filepaths": filepaths, "split": "PUBLISHED-TESTSET", "files": dl_manager.iter_archive(archive), }, ), datasets.SplitGenerator( name=datasets.Split.TRAIN, gen_kwargs={ "filepaths": filepaths, "split": "TRAINING-SET", "files": dl_manager.iter_archive(archive), }, ), ] else: return [ datasets.SplitGenerator( name=datasets.Split.TEST, gen_kwargs={"filepaths": filepaths, "split": "TEST", "files": dl_manager.iter_archive(archive)}, ), datasets.SplitGenerator( name=datasets.Split.TRAIN, gen_kwargs={"filepaths": filepaths, "split": "TRAIN", "files": dl_manager.iter_archive(archive)}, ), datasets.SplitGenerator( name="unused", gen_kwargs={ "filepaths": filepaths, "split": "NOT-USED", "files": dl_manager.iter_archive(archive), }, ), ] def _generate_examples(self, filepaths, split, files): """This function returns the examples in the raw (text) form.""" for path, f in files: if path in filepaths: # only the file reut2-017 has one line non UTF-8 encoded so we can ignore it line = f.readline().decode("utf-8", errors="ignore") while line: if line.startswith(""): if line.replace("\n", "") != "": line = line.split("") topics = [topic.replace("", "") for topic in line[1:]] topics = [topic.replace("\n", "") for topic in topics] line = f.readline().decode("utf-8", errors="ignore") elif line.startswith(""): if line.replace("\n", "") != "": line = line.split("") places = [place.replace("", "") for place in line[1:]] places = [place.replace("\n", "") for place in places] line = f.readline().decode("utf-8", errors="ignore") elif line.startswith(""): if line.replace("\n", "") != "": line = line.split("") people = [p.replace("", "") for p in line[1:]] people = [p.replace("\n", "") for p in people] line = f.readline().decode("utf-8", errors="ignore") elif line.startswith(""): if line.replace("\n", "") != "": line = line.split("") orgs = [org.replace("", "") for org in line[1:]] orgs = [org.replace("\n", "") for org in orgs] line = f.readline().decode("utf-8", errors="ignore") elif line.startswith(""): if line.replace("\n", "") != "": line = line.split("") exchanges = [ex.replace("", "") for ex in line[1:]] exchanges = [ex.replace("\n", "") for ex in exchanges] line = f.readline().decode("utf-8", errors="ignore") elif line.startswith(""): date = line.replace("\n", "") date = line[6:-8] line = f.readline().decode("utf-8", errors="ignore") elif line.startswith(""): title = line[7:-9] line = f.readline().decode("utf-8", errors="ignore") elif "*<TITLE>" in line: # These lines start with a variable number of * chars title = line.split("*<TITLE>")[1][:-1] line = f.readline().decode("utf-8", errors="ignore") while "" not in line: # Convert any \n in TYPE="BRIEF" text to spaces to match other titles title += " " + line[:-1] line = f.readline().decode("utf-8", errors="ignore") elif "" in line: text = line.split("")[1] line = f.readline().decode("utf-8", errors="ignore") while "" not in line: text += line line = f.readline().decode("utf-8", errors="ignore") elif line.startswith(''): text_type = '"UNPROC"' text = line[20:] line = f.readline().decode("utf-8", errors="ignore") while "" not in line: text += line line = f.readline().decode("utf-8", errors="ignore") elif line.startswith(''): text_type = '"BRIEF"' line = f.readline().decode("utf-8", errors="ignore") elif line.startswith(""): text_type = '"NORM"' line = f.readline().decode("utf-8", errors="ignore") elif line.startswith(""): yield new_id, { "lewis_split": lewis_split, "cgis_split": cgis_split, "old_id": old_id, "new_id": new_id, "topics": topics, "places": places, "people": people, "orgs": orgs, "exchanges": exchanges, "date": date, "title": title, "text": text, "text_type": text_type, } line = f.readline().decode("utf-8", errors="ignore") else: line = f.readline().decode("utf-8", errors="ignore")