Datasets:

Languages:
English
Multilinguality:
monolingual
Size Categories:
100K<n<1M
Language Creators:
found
Annotations Creators:
machine-generated
ArXiv:
Tags:
License:
albertvillanova HF staff commited on
Commit
00b0e4a
1 Parent(s): c9f30e3

Update metadata in dataset card (#4)

Browse files

- Update metadata in dataset card (e7ba0237985f3ba7ca316f829ce4e570bf52a39a)

Files changed (1) hide show
  1. README.md +24 -12
README.md CHANGED
@@ -204,17 +204,18 @@ dataset_info:
204
 
205
  - **Homepage:** [https://compguesswhat.github.io/](https://compguesswhat.github.io/)
206
  - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
207
- - **Paper:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
208
- - **Point of Contact:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
 
209
  - **Size of downloaded dataset files:** 112.05 MB
210
  - **Size of the generated dataset:** 271.11 MB
211
  - **Total amount of disk used:** 383.16 MB
212
 
213
  ### Dataset Summary
214
 
215
- CompGuessWhat?! is an instance of a multi-task framework for evaluating the quality of learned neural representations,
216
- in particular concerning attribute grounding. Use this dataset if you want to use the set of games whose reference
217
- scene is an image in VisualGenome. Visit the website for more details: https://compguesswhat.github.io
218
 
219
  ### Supported Tasks and Leaderboards
220
 
@@ -417,13 +418,24 @@ The data fields are the same among all splits.
417
  ### Citation Information
418
 
419
  ```
420
- @inproceedings{suglia2020compguesswhat,
421
- title={CompGuessWhat?!: a Multi-task Evaluation Framework for Grounded Language Learning},
422
- author={Suglia, Alessandro, Konstas, Ioannis, Vanzo, Andrea, Bastianelli, Emanuele, Desmond Elliott, Stella Frank and Oliver Lemon},
423
- booktitle={Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics},
424
- year={2020}
425
- }
426
-
 
 
 
 
 
 
 
 
 
 
 
427
  ```
428
 
429
 
 
204
 
205
  - **Homepage:** [https://compguesswhat.github.io/](https://compguesswhat.github.io/)
206
  - **Repository:** [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
207
+ - **Paper:** https://arxiv.org/abs/2006.02174
208
+ - **Paper:** https://doi.org/10.18653/v1/2020.acl-main.682
209
+ - **Point of Contact:** [Alessandro Suglia](mailto:alessandro.suglia@gmail.com)
210
  - **Size of downloaded dataset files:** 112.05 MB
211
  - **Size of the generated dataset:** 271.11 MB
212
  - **Total amount of disk used:** 383.16 MB
213
 
214
  ### Dataset Summary
215
 
216
+ CompGuessWhat?! is an instance of a multi-task framework for evaluating the quality of learned neural representations,
217
+ in particular concerning attribute grounding. Use this dataset if you want to use the set of games whose reference
218
+ scene is an image in VisualGenome. Visit the website for more details: https://compguesswhat.github.io
219
 
220
  ### Supported Tasks and Leaderboards
221
 
 
418
  ### Citation Information
419
 
420
  ```
421
+ @inproceedings{suglia-etal-2020-compguesswhat,
422
+ title = "{C}omp{G}uess{W}hat?!: A Multi-task Evaluation Framework for Grounded Language Learning",
423
+ author = "Suglia, Alessandro and
424
+ Konstas, Ioannis and
425
+ Vanzo, Andrea and
426
+ Bastianelli, Emanuele and
427
+ Elliott, Desmond and
428
+ Frank, Stella and
429
+ Lemon, Oliver",
430
+ booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics",
431
+ month = jul,
432
+ year = "2020",
433
+ address = "Online",
434
+ publisher = "Association for Computational Linguistics",
435
+ url = "https://www.aclweb.org/anthology/2020.acl-main.682",
436
+ pages = "7625--7641",
437
+ abstract = "Approaches to Grounded Language Learning are commonly focused on a single task-based final performance measure which may not depend on desirable properties of the learned hidden representations, such as their ability to predict object attributes or generalize to unseen situations. To remedy this, we present GroLLA, an evaluation framework for Grounded Language Learning with Attributes based on three sub-tasks: 1) Goal-oriented evaluation; 2) Object attribute prediction evaluation; and 3) Zero-shot evaluation. We also propose a new dataset CompGuessWhat?! as an instance of this framework for evaluating the quality of learned neural representations, in particular with respect to attribute grounding. To this end, we extend the original GuessWhat?! dataset by including a semantic layer on top of the perceptual one. Specifically, we enrich the VisualGenome scene graphs associated with the GuessWhat?! images with several attributes from resources such as VISA and ImSitu. We then compare several hidden state representations from current state-of-the-art approaches to Grounded Language Learning. By using diagnostic classifiers, we show that current models{'} learned representations are not expressive enough to encode object attributes (average F1 of 44.27). In addition, they do not learn strategies nor representations that are robust enough to perform well when novel scenes or objects are involved in gameplay (zero-shot best accuracy 50.06{\%}).",
438
+ }
439
  ```
440
 
441