Fix metadata in dataset card

#13
by albertvillanova HF staff - opened
Files changed (1) hide show
  1. README.md +90 -21
README.md CHANGED
@@ -6,7 +6,7 @@ language_creators:
6
  language:
7
  - en
8
  license:
9
- - unknown
10
  multilinguality:
11
  - monolingual
12
  size_categories:
@@ -378,9 +378,9 @@ dataset_info:
378
 
379
  ## Dataset Description
380
 
381
- - **Homepage:** [https://github.com/google-research-datasets/boolean-questions](https://github.com/google-research-datasets/boolean-questions)
382
  - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
383
- - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
384
  - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
385
  - **Size of downloaded dataset files:** 58.36 MB
386
  - **Size of the generated dataset:** 249.57 MB
@@ -392,11 +392,6 @@ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after
392
  GLUE with a new set of more difficult language understanding tasks, improved
393
  resources, and a new public leaderboard.
394
 
395
- BoolQ (Boolean Questions, Clark et al., 2019a) is a QA task where each example consists of a short
396
- passage and a yes/no question about the passage. The questions are provided anonymously and
397
- unsolicited by users of the Google search engine, and afterwards paired with a paragraph from a
398
- Wikipedia article containing the answer. Following the original work, we evaluate with accuracy.
399
-
400
  ### Supported Tasks and Leaderboards
401
 
402
  [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
@@ -584,30 +579,104 @@ The data fields are the same among all splits.
584
 
585
  ### Licensing Information
586
 
587
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
588
 
589
  ### Citation Information
590
 
 
591
  ```
 
 
 
 
 
 
592
  @inproceedings{clark2019boolq,
593
- title={BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions},
594
- author={Clark, Christopher and Lee, Kenton and Chang, Ming-Wei, and Kwiatkowski, Tom and Collins, Michael, and Toutanova, Kristina},
595
- booktitle={NAACL},
596
  year={2019}
597
  }
598
- @article{wang2019superglue,
599
- title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems},
600
- author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R},
601
- journal={arXiv preprint arXiv:1905.00537},
602
  year={2019}
603
  }
604
-
605
- Note that each SuperGLUE dataset has its own citation. Please see the source to
606
- get the correct citation for each contained dataset.
607
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
608
  ```
609
 
610
-
611
  ### Contributions
612
 
613
  Thanks to [@thomwolf](https://github.com/thomwolf), [@lewtun](https://github.com/lewtun), [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset.
 
6
  language:
7
  - en
8
  license:
9
+ - other
10
  multilinguality:
11
  - monolingual
12
  size_categories:
 
378
 
379
  ## Dataset Description
380
 
381
+ - **Homepage:** https://super.gluebenchmark.com/
382
  - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
383
+ - **Paper:** https://arxiv.org/abs/1905.00537
384
  - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
385
  - **Size of downloaded dataset files:** 58.36 MB
386
  - **Size of the generated dataset:** 249.57 MB
 
392
  GLUE with a new set of more difficult language understanding tasks, improved
393
  resources, and a new public leaderboard.
394
 
 
 
 
 
 
395
  ### Supported Tasks and Leaderboards
396
 
397
  [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
 
579
 
580
  ### Licensing Information
581
 
582
+ The primary SuperGLUE tasks are built on and derived from existing datasets. We refer users to the original licenses accompanying each dataset, but it is our understanding that these licenses allow for their use and redistribution in a research context.
583
 
584
  ### Citation Information
585
 
586
+ If you use SuperGLUE, please cite all the datasets you use in any papers that come out of your work. In addition, we encourage you to use the following BibTeX citation for SuperGLUE itself:
587
  ```
588
+ @article{wang2019superglue,
589
+ title={Super{GLUE}: A Stickier Benchmark for General-Purpose Language Understanding Systems},
590
+ author={Alex Wang and Yada Pruksachatkun and Nikita Nangia and Amanpreet Singh and Julian Michael and Felix Hill and Omer Levy and Samuel R. Bowman},
591
+ journal={arXiv preprint 1905.00537},
592
+ year={2019}
593
+ }
594
  @inproceedings{clark2019boolq,
595
+ title={{B}ool{Q}: Exploring the Surprising Difficulty of Natural Yes/No Questions},
596
+ author={Clark, Christopher and Lee, Kenton and Chang, Ming-Wei and Kwiatkowski, Tom and Collins, Michael and Toutanova, Kristina},
597
+ booktitle={Proceedings of NAACL-HLT 2019},
598
  year={2019}
599
  }
600
+ @inproceedings{demarneffe:cb,
601
+ title={{The CommitmentBank}: Investigating projection in naturally occurring discourse},
602
+ author={De Marneffe, Marie-Catherine and Simons, Mandy and Tonhauser, Judith},
603
+ note={To appear in proceedings of Sinn und Bedeutung 23. Data can be found at https://github.com/mcdm/CommitmentBank/},
604
  year={2019}
605
  }
606
+ @inproceedings{roemmele2011choice,
607
+ title={Choice of plausible alternatives: An evaluation of commonsense causal reasoning},
608
+ author={Roemmele, Melissa and Bejan, Cosmin Adrian and Gordon, Andrew S.},
609
+ booktitle={2011 AAAI Spring Symposium Series},
610
+ year={2011}
611
+ }
612
+ @inproceedings{khashabi2018looking,
613
+ title={Looking beyond the surface: A challenge set for reading comprehension over multiple sentences},
614
+ author={Khashabi, Daniel and Chaturvedi, Snigdha and Roth, Michael and Upadhyay, Shyam and Roth, Dan},
615
+ booktitle={Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers)},
616
+ pages={252--262},
617
+ year={2018}
618
+ }
619
+ @article{zhang2018record,
620
+ title={{ReCoRD}: Bridging the Gap between Human and Machine Commonsense Reading Comprehension},
621
+ author={Sheng Zhang and Xiaodong Liu and Jingjing Liu and Jianfeng Gao and Kevin Duh and Benjamin Van Durme},
622
+ journal={arXiv preprint 1810.12885},
623
+ year={2018}
624
+ }
625
+ @incollection{dagan2006pascal,
626
+ title={The {PASCAL} recognising textual entailment challenge},
627
+ author={Dagan, Ido and Glickman, Oren and Magnini, Bernardo},
628
+ booktitle={Machine learning challenges. evaluating predictive uncertainty, visual object classification, and recognising tectual entailment},
629
+ pages={177--190},
630
+ year={2006},
631
+ publisher={Springer}
632
+ }
633
+ @article{bar2006second,
634
+ title={The second {PASCAL} recognising textual entailment challenge},
635
+ author={Bar Haim, Roy and Dagan, Ido and Dolan, Bill and Ferro, Lisa and Giampiccolo, Danilo and Magnini, Bernardo and Szpektor, Idan},
636
+ year={2006}
637
+ }
638
+ @inproceedings{giampiccolo2007third,
639
+ title={The third {PASCAL} recognizing textual entailment challenge},
640
+ author={Giampiccolo, Danilo and Magnini, Bernardo and Dagan, Ido and Dolan, Bill},
641
+ booktitle={Proceedings of the ACL-PASCAL workshop on textual entailment and paraphrasing},
642
+ pages={1--9},
643
+ year={2007},
644
+ organization={Association for Computational Linguistics},
645
+ }
646
+ @article{bentivogli2009fifth,
647
+ title={The Fifth {PASCAL} Recognizing Textual Entailment Challenge},
648
+ author={Bentivogli, Luisa and Dagan, Ido and Dang, Hoa Trang and Giampiccolo, Danilo and Magnini, Bernardo},
649
+ booktitle={TAC},
650
+ year={2009}
651
+ }
652
+ @inproceedings{pilehvar2018wic,
653
+ title={{WiC}: The Word-in-Context Dataset for Evaluating Context-Sensitive Meaning Representations},
654
+ author={Pilehvar, Mohammad Taher and Camacho-Collados, Jose},
655
+ booktitle={Proceedings of NAACL-HLT},
656
+ year={2019}
657
+ }
658
+ @inproceedings{rudinger2018winogender,
659
+ title={Gender Bias in Coreference Resolution},
660
+ author={Rudinger, Rachel and Naradowsky, Jason and Leonard, Brian and {Van Durme}, Benjamin},
661
+ booktitle={Proceedings of NAACL-HLT},
662
+ year={2018}
663
+ }
664
+ @inproceedings{poliak2018dnc,
665
+ title={Collecting Diverse Natural Language Inference Problems for Sentence Representation Evaluation},
666
+ author={Poliak, Adam and Haldar, Aparajita and Rudinger, Rachel and Hu, J. Edward and Pavlick, Ellie and White, Aaron Steven and {Van Durme}, Benjamin},
667
+ booktitle={Proceedings of EMNLP},
668
+ year={2018}
669
+ }
670
+ @inproceedings{levesque2011winograd,
671
+ title={The {W}inograd schema challenge},
672
+ author={Levesque, Hector J and Davis, Ernest and Morgenstern, Leora},
673
+ booktitle={{AAAI} Spring Symposium: Logical Formalizations of Commonsense Reasoning},
674
+ volume={46},
675
+ pages={47},
676
+ year={2011}
677
+ }
678
  ```
679
 
 
680
  ### Contributions
681
 
682
  Thanks to [@thomwolf](https://github.com/thomwolf), [@lewtun](https://github.com/lewtun), [@patrickvonplaten](https://github.com/patrickvonplaten) for adding this dataset.