discourse_marker_qa / README.md
sileod's picture
Update README.md
ee2d5f7
|
raw
history blame
1.84 kB
metadata
annotations_creators:
  - expert-generated
language_creators:
  - crowdsourced
languages:
  - en
licenses:
  - apache-2.0
multilinguality:
  - monolingual
pretty_name: discourse_marker_qa
size_categories:
  - n<1K
source_datasets:
  - original
task_categories:
  - question-answering
  - multiple-choice
task_ids:
  - open-domain-qa
  - multiple-choice-qa

Dataset for evaluation of (zero-shot) discourse marker prediction with language models

This is the Big-Bench version of our discourse marker prediction dataset, Discovery

Design considerations: https://github.com/google/BIG-bench/tree/main/bigbench/benchmark_tasks/discourse_marker_prediction

GPT2 has to zero-shot 15% accuracy with on this multiple-choice task based on language modeling perplexity. As a comparison, a fully supervised model, trained with 10k examples per marker with ROBERTA and default hyperparameters with one epoch, leads to an accuracy of 30% with 174 possible markers. This shows that this task is hard for GPT2 and that the model didn't memorize the discourse markers, but that high accuracies are still possible.

Citation

@inproceedings{sileo-etal-2019-mining,
    title = "Mining Discourse Markers for Unsupervised Sentence Representation Learning",
    author = "Sileo, Damien  and
      Van De Cruys, Tim  and
      Pradel, Camille  and
      Muller, Philippe",
    booktitle = "Proceedings of the 2019 Conference of the North {A}merican Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)",
    month = jun,
    year = "2019",
    address = "Minneapolis, Minnesota",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/N19-1351",
    doi = "10.18653/v1/N19-1351",
    pages = "3477--3486",
}