Datasets:
Tasks:
Question Answering
Formats:
parquet
Sub-tasks:
multiple-choice-qa
Languages:
English
Size:
10K - 100K
ArXiv:
License:
Add information to dataset card
#5
by
albertvillanova
HF staff
- opened
README.md
CHANGED
@@ -5,8 +5,7 @@ language_creators:
|
|
5 |
- crowdsourced
|
6 |
language:
|
7 |
- en
|
8 |
-
license:
|
9 |
-
- unknown
|
10 |
multilinguality:
|
11 |
- monolingual
|
12 |
size_categories:
|
@@ -282,14 +281,16 @@ configs:
|
|
282 |
## Dataset Description
|
283 |
|
284 |
- **Homepage:** [Add homepage URL here if available (unless it's a GitHub repository)]()
|
285 |
-
- **Repository:**
|
286 |
-
- **Paper:**
|
287 |
-
- **
|
288 |
-
- **Point of Contact:** [If known, name and email of at least one person the reader can contact for questions about the dataset.]()
|
289 |
|
290 |
### Dataset Summary
|
291 |
|
292 |
-
|
|
|
|
|
|
|
293 |
|
294 |
### Supported Tasks and Leaderboards
|
295 |
|
@@ -368,11 +369,33 @@ configs:
|
|
368 |
|
369 |
### Licensing Information
|
370 |
|
371 |
-
|
372 |
|
373 |
### Citation Information
|
374 |
|
375 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
376 |
|
377 |
### Contributions
|
378 |
|
|
|
5 |
- crowdsourced
|
6 |
language:
|
7 |
- en
|
8 |
+
license: odc-by
|
|
|
9 |
multilinguality:
|
10 |
- monolingual
|
11 |
size_categories:
|
|
|
281 |
## Dataset Description
|
282 |
|
283 |
- **Homepage:** [Add homepage URL here if available (unless it's a GitHub repository)]()
|
284 |
+
- **Repository:** https://github.com/Websail-NU/CODAH
|
285 |
+
- **Paper:** https://aclanthology.org/W19-2008/
|
286 |
+
- **Paper:** https://arxiv.org/abs/1904.04365
|
|
|
287 |
|
288 |
### Dataset Summary
|
289 |
|
290 |
+
The COmmonsense Dataset Adversarially-authored by Humans (CODAH) is an evaluation set for commonsense
|
291 |
+
question-answering in the sentence completion style of SWAG. As opposed to other automatically generated
|
292 |
+
NLI datasets, CODAH is adversarially constructed by humans who can view feedback from a pre-trained model
|
293 |
+
and use this information to design challenging commonsense questions.
|
294 |
|
295 |
### Supported Tasks and Leaderboards
|
296 |
|
|
|
369 |
|
370 |
### Licensing Information
|
371 |
|
372 |
+
The CODAH dataset is made available under the Open Data Commons Attribution License: http://opendatacommons.org/licenses/by/1.0/
|
373 |
|
374 |
### Citation Information
|
375 |
|
376 |
+
```
|
377 |
+
@inproceedings{chen-etal-2019-codah,
|
378 |
+
title = "{CODAH}: An Adversarially-Authored Question Answering Dataset for Common Sense",
|
379 |
+
author = "Chen, Michael and
|
380 |
+
D{'}Arcy, Mike and
|
381 |
+
Liu, Alisa and
|
382 |
+
Fernandez, Jared and
|
383 |
+
Downey, Doug",
|
384 |
+
editor = "Rogers, Anna and
|
385 |
+
Drozd, Aleksandr and
|
386 |
+
Rumshisky, Anna and
|
387 |
+
Goldberg, Yoav",
|
388 |
+
booktitle = "Proceedings of the 3rd Workshop on Evaluating Vector Space Representations for {NLP}",
|
389 |
+
month = jun,
|
390 |
+
year = "2019",
|
391 |
+
address = "Minneapolis, USA",
|
392 |
+
publisher = "Association for Computational Linguistics",
|
393 |
+
url = "https://aclanthology.org/W19-2008",
|
394 |
+
doi = "10.18653/v1/W19-2008",
|
395 |
+
pages = "63--69",
|
396 |
+
abstract = "Commonsense reasoning is a critical AI capability, but it is difficult to construct challenging datasets that test common sense. Recent neural question answering systems, based on large pre-trained models of language, have already achieved near-human-level performance on commonsense knowledge benchmarks. These systems do not possess human-level common sense, but are able to exploit limitations of the datasets to achieve human-level scores. We introduce the CODAH dataset, an adversarially-constructed evaluation dataset for testing common sense. CODAH forms a challenging extension to the recently-proposed SWAG dataset, which tests commonsense knowledge using sentence-completion questions that describe situations observed in video. To produce a more difficult dataset, we introduce a novel procedure for question acquisition in which workers author questions designed to target weaknesses of state-of-the-art neural question answering systems. Workers are rewarded for submissions that models fail to answer correctly both before and after fine-tuning (in cross-validation). We create 2.8k questions via this procedure and evaluate the performance of multiple state-of-the-art question answering systems on our dataset. We observe a significant gap between human performance, which is 95.3{\%}, and the performance of the best baseline accuracy of 65.3{\%} by the OpenAI GPT model.",
|
397 |
+
}
|
398 |
+
```
|
399 |
|
400 |
### Contributions
|
401 |
|