Datasets:

Modalities:
Text
Languages:
English
Size:
< 1K
Libraries:
Datasets
License:
semeval-2010-pre / README.md
MHoubre's picture
Update README.md (#1)
c98da16
metadata
annotations_creators:
  - unknown
language_creators:
  - unknown
language:
  - en
license: cc-by-4.0
multilinguality:
  - monolingual
task_categories:
  - text-mining
  - text-generation
task_ids:
  - keyphrase-generation
  - keyphrase-extraction
size_categories:
  - n<1K
pretty_name: Preprocessed SemEval-2010 Benchmark dataset

Preprocessed SemEval-2010 Benchmark dataset for Keyphrase Generation

About

SemEval-2010 is a dataset for benchmarking keyphrase extraction and generation models. The dataset is composed of 244 full-text scientific papers collected from the ACM Digital Library. Keyphrases were annotated by readers and combined with those provided by the authors. Details about the SemEval-2010 dataset can be found in the original paper (kim et al., 2010).

This version of the dataset was produced by (Boudin et al., 2016) and provides four increasingly sophisticated levels of document preprocessing:

  • lvl-1: default text files provided by the SemEval-2010 organizers.

  • lvl-2: for each file, we manually retrieved the original PDF file from the ACM Digital Library. We then extract the enriched textual content of the PDF files using an Optical Character Recognition (OCR) system and perform document logical structure detection using ParsCit v110505. We use the detected logical structure to remove author-assigned keyphrases and select only relevant elements : title, headers, abstract, introduction, related work, body text and conclusion. We finally apply a systematic dehyphenation at line breaks.s

  • lvl-3: we further abridge the input text from level 2 preprocessed documents to the following: title, headers, abstract, introduction, related work, background and conclusion.

  • lvl-4: we abridge the input text from level 3 preprocessed documents using an unsupervised summarization technique. We keep the title and abstract and select the most content bearing sentences from the remaining contents.

Titles and abstracts, collected from the SciCorefCorpus, are also provided. Details about how they were extracted and cleaned up can be found in (Chaimongkol et al., 2014).

Reference keyphrases are provided in stemmed form (because they were provided like this for the test split in the competition). They are also categorized under the PRMU (Present-Reordered-Mixed-Unseen) scheme as proposed in (Boudin and Gallina, 2021). Text pre-processing (tokenization) is carried out using spacy (en_core_web_sm model) with a special rule to avoid splitting words with hyphens (e.g. graph-based is kept as one token). Stemming (Porter's stemmer implementation provided in nltk) is applied before reference keyphrases are matched against the source text. Details about the process can be found in prmu.py. The Present reference keyphrases are also ordered by their order of apparition in the concatenation of title and text (lvl-1).

Content and statistics

The dataset is divided into the following two splits:

Split # documents #words # keyphrases % Present % Reordered % Mixed % Unseen
Train 144 184.6 15.44 42.16 7.36 26.85 23.63
Test 100 203.1 14.66 40.11 8.34 27.12 24.43

Statistics (#words, PRMU distributions) are computed using the title/abstract and not the full text of scientific papers.

The following data fields are available :

  • id: unique identifier of the document.
  • title: title of the document.
  • abstract: abstract of the document.
  • lvl-1: content of the document with no text processing.
  • lvl-2: content of the document retrieved from original PDF files and cleaned up.
  • lvl-3: content of the document further abridged to relevant sections.
  • lvl-4: content of the document further abridged using an unsupervised summarization technique.
  • keyphrases: list of reference keyphrases.
  • prmu: list of Present-Reordered-Mixed-Unseen categories for reference keyphrases.

References