Datasets:
GEM
/

Languages:
English
Multilinguality:
unknown
Size Categories:
unknown
Language Creators:
unknown
Annotations Creators:
none
Source Datasets:
original
Tags:
data-to-text
License:
e2e_nlg / e2e_nlg.json
Sebastian Gehrmann
.
9d3698b
raw history blame
No virus
12.8 kB
{
"overview": {
"where": {
"has-leaderboard": "no",
"leaderboard-url": "N/A",
"leaderboard-description": "N/A",
"website": "[Website](http://www.macs.hw.ac.uk/InteractionLab/E2E/)",
"data-url": "[Github](https://github.com/tuetschek/e2e-cleaning)",
"paper-url": "[First data release](https://www.aclweb.org/anthology/W17-5525/), [Detailed E2E Challenge writeup](https://doi.org/10.1016/j.csl.2019.06.009), [Cleaned E2E version](https://www.aclweb.org/anthology/W19-8652/)",
"paper-bibtext": "```\n@inproceedings{e2e_cleaned,\n\taddress = {Tokyo, Japan},\n\ttitle = {Semantic {Noise} {Matters} for {Neural} {Natural} {Language} {Generation}},\n\turl = {https://www.aclweb.org/anthology/W19-8652/},\n\tbooktitle = {Proceedings of the 12th {International} {Conference} on {Natural} {Language} {Generation} ({INLG} 2019)},\n\tauthor = {Du\u0161ek, Ond\u0159ej and Howcroft, David M and Rieser, Verena},\n\tyear = {2019},\n\tpages = {421--426},\n}\n```",
"contact-name": "Ondrej Dusek",
"contact-email": "odusek@ufal.mff.cuni.cz"
},
"languages": {
"is-multilingual": "no",
"license": "cc-by-sa-4.0: Creative Commons Attribution Share Alike 4.0 International",
"task-other": "N/A",
"language-names": [
"English"
],
"language-speakers": "The original dataset was collected using the CrowdFlower (now Appen) platform using native English speakers (self-reported). No demographic information was provided, but the collection was geographically limited to English-speaking countries. ",
"language-dialects": "Dialect-specific data was not collected and the language is general British English. ",
"intended-use": "The dataset was collected to test neural model on a very well specified realization task. ",
"license-other": "N/A",
"task": "Data-to-Text",
"communicative": "Producing a text informing/recommending a restaurant, given all and only the attributes specified on the input."
},
"credit": {
"organization-type": [
"academic"
],
"organization-names": "Heriot-Watt University",
"creators": "Jekaterina Novikova, Ondrej Dusek and Verena Rieser",
"funding": "This research received funding from the EPSRC projects DILiGENt (EP/M005429/1) and MaDrIgAL (EP/N017536/1).",
"gem-added-by": "Simon Mille wrote the initial data card and Yacine Jernite the data loader. Sebastian Gehrmann migrated the data card to the v2 format and moved the data loader to the hub."
},
"structure": {
"data-fields": "The data is in a CSV format, with the following fields:\n\n* `mr` -- the meaning representation (MR, input)\n* `ref` -- reference, i.e. the corresponding natural-language description (output)\n\nThere are additional fields (`fixed`, `orig_mr`) indicating whether the data was modified in the\ncleaning process and what was the original MR before cleaning, but these aren't used for NLG.\n\nThe MR has a flat structure -- attribute-value pairs are comma separated, with values\nenclosed in brackets (see example above). There are 8 attributes:\n* `name` -- restaurant name\n* `near` -- a landmark close to the restaurant\n* `area` -- location (riverside, city centre)\n* `food` -- food type / cuisine (e.g. Japanese, Indian, English etc.)\n* `eatType` -- restaurant type (restaurant, coffee shop, pub)\n* `priceRange` -- price range (low, medium, high, <\u00a320, \u00a320-30, >\u00a330)\n* `rating` -- customer rating (low, medium, high, 1/5, 3/5, 5/5)\n* `familyFriendly` -- is the restaurant family-friendly (yes/no)\n\nThe same MR is often repeated multiple times with different synonymous references.\n",
"structure-splits": "\n| | MRs | Distinct MRs | References |\n|-------------|------|--------------|------------|\n| Training |12,568| 8,362 | 33,525 |\n| Development | 1,484| 1,132 | 4,299 |\n| Test | 1,847| 1,358 | 4,693 |\n| Total |15,899| 10,852 | 42,517 |\n\n\n\u201cDistinct MRs\u201d are MRs that remain distinct even if restaurant/place names (attributes `name`, `near`)\nare delexicalized, i.e., replaced with a placeholder.",
"structure-splits-criteria": "The data are divided so that MRs in different splits do not overlap.\n",
"structure-description": "n/a",
"structure-labels": "The source MRs were generated automatically at random from a set of valid attribute values. The labels were crowdsourced and are natural language",
"structure-example": "```\n{\n \"input\": \"name[Alimentum], area[riverside], familyFriendly[yes], near[Burger King]\",\n \"target\": \"Alimentum is a kids friendly place in the riverside area near Burger King.\" \n}\n```",
"structure-outlier": "n/a"
},
"what": {
"dataset": "The E2E NLG dataset is an English benchmark dataset for data-to-text models that verbalize a set of 2-9 key-value attribute pairs in the restaurant domain. The version used for GEM is the cleaned E2E NLG dataset, which filters examples with hallucinations and outputs that don't fully cover all input attributes. "
}
},
"gem": {
"rationale": {
"sole-task-dataset": "yes",
"sole-language-task-dataset": "no",
"distinction-description": "The dataset is much cleaner than comparable datasets, and it is also a relatively easy task, making for a straightforward evaluation. ",
"contribution": "The E2E dataset is one of the largest limited-domain NLG datasets and is frequently used as a data-to-text generation benchmark. The E2E Challenge included 20 systems of very different architectures, with system outputs available for download.\n\n",
"model-ability": "surface realization."
},
"curation": {
"has-additional-curation": "yes",
"modification-types": [],
"modification-description": "N/A",
"has-additional-splits": "yes",
"additional-splits-description": "4 special test sets for E2E were added to the GEM evaluation suite.\n\n1. We created subsets of the training and development sets of ~500 randomly selected inputs each.\n2. We applied input scrambling on a subset of 500 randomly selected test instances; the order of the input properties was randomly reassigned.\n3. For the input size, we created subpopulations based on the number of restaurant properties in the input.\n\n| Input length | Frequency English |\n|---------------|-------------------|\n| 2 | 5 |\n| 3 | 120 |\n| 4 | 389 |\n| 5 | 737 |\n| 6 | 1187 |\n| 7 | 1406 |\n| 8 | 774 |\n| 9 | 73 |\n| 10 | 2 |",
"additional-splits-capacicites": "Generalization and robustness"
},
"starting": {
"research-pointers": "n/a",
"technical-terms": "n/a"
}
},
"curation": {
"original": {
"is-aggregated": "no",
"aggregated-sources": "N/A",
"rationale": "The dataset was collected to showcase/test neural NLG models. It is larger and contains more lexical richness and syntactic variation than previous closed-domain NLG datasets.\n\n",
"communicative": "Producing a text informing/recommending a restaurant, given all and only the attributes specified on the input."
},
"language": {
"found": [],
"crowdsourced": [
"Other crowdworker platform"
],
"created": "N/A",
"machine-generated": "N/A",
"validated": "validated by data curator",
"is-filtered": "algorithmically",
"filtered-criteria": "The cleaned version of the dataset which we are using in GEM was algorithmically filtered. They used regular expressions to match all human-generated references with a more accurate input when attributes were hallucinated or dropped. Additionally, train-test overlap stemming from the transformation was removed. As a result, this data is much cleaner than the original dataset but not perfect (about 20% of instances may have misaligned slots, compared to 40% of the original data. ",
"obtained": [
"Crowdsourced"
],
"producers-description": "Human references describing the MRs were collected by crowdsourcing on the CrowdFlower (now Appen) platform,\nwith either textual or pictorial MRs as a baseline. \nThe pictorial MRs were used in 20% of cases -- these yield higher lexical variation but introduce noise.",
"topics": "The dataset is focused on descriptions of restaurants.",
"pre-processed": "There were basic checks (length, valid characters, repetition).\n"
},
"annotations": {
"origin": "none",
"rater-number": "N/A",
"rater-qualifications": "N/A",
"rater-training-num": "N/A",
"rater-test-num": "N/A",
"rater-annotation-service-bool": "no",
"rater-annotation-service": [],
"values": "N/A",
"quality-control": [],
"quality-control-details": "N/A"
},
"consent": {
"has-consent": "yes",
"consent-policy": "Since a crowdsourcing platform was used, the involved raters waived their rights to the data and are aware that the produced annotations can be publicly released. ",
"consent-other": "N/A",
"no-consent-justification": "N/A"
},
"pii": {
"has-pii": "no PII",
"no-pii-justification": "The dataset is artificial and does not contain any description of people.",
"is-pii-identified": "N/A",
"pii-identified-method": "N/A",
"is-pii-replaced": "N/A",
"pii-replaced-method": "N/A",
"pii-categories": []
},
"maintenance": {
"has-maintenance": "no",
"description": "N/A",
"contact": "N/A",
"contestation-mechanism": "N/A",
"contestation-link": "N/A",
"contestation-description": "N/A"
}
},
"results": {
"results": {
"other-metrics-definitions": "N/A",
"has-previous-results": "yes",
"current-evaluation": "Most previous results, including the shared task results, used the library provided by the dataset creators. The shared task also conducted a human evaluation using the following two criteria: \n\n- `Quality`: When collecting quality ratings, system outputs were presented to crowd workers together with the corresponding meaning representation, which implies that correctness of the NL utterance relative to the MR should also influence this ranking. The crowd workers were asked: \u201cHow do you judge the overall quality of the utterance in terms of its grammatical correctness, fluency, adequacy and other important factors?\u201d\n- `Naturalness`: When collecting naturalness ratings, system outputs were presented to crowd workers without the corresponding meaning representation. The crowd workers were asked: \u201cCould the utterance have been produced by a native speaker?\u201d",
"previous-results": "The shared task writeup has in-depth evaluations of systems (https://www.sciencedirect.com/science/article/pii/S0885230819300919) ",
"model-abilities": "Surface realization.",
"metrics": [
"BLEU",
"METEOR",
"ROUGE"
],
"original-evaluation": "The official evaluation script combines the MT-Eval and COCO Captioning libraries with the following metrics.\n\n- BLEU\n- CIDEr\n- NIST\n- METEOR\n- ROUGE-L"
}
},
"considerations": {
"pii": {
"risks-description": "n/a"
},
"licenses": {
"dataset-restrictions-other": "N/A",
"data-copyright-other": "N/A",
"dataset-restrictions": [
"open license - commercial use allowed"
],
"data-copyright": [
"open license - commercial use allowed"
]
},
"limitations": {
"data-technical-limitations": "The cleaned version still has data points with hallucinated or omitted attributes. ",
"data-unsuited-applications": "The data only pertains to the restaurant domain and the included attributes. A model cannot be expected to handle other domains or attributes. ",
"data-discouraged-use": "n/a"
}
},
"context": {
"previous": {
"is-deployed": "no",
"described-risks": "N/A",
"changes-from-observation": "N/A"
},
"underserved": {
"helps-underserved": "no",
"underserved-description": "N/A"
},
"biases": {
"has-biases": "no",
"bias-analyses": "N/A",
"speaker-distibution": "The source data is generated randomly, so it should not contain biases. The human references may be biased by the workers' demographic, but that was not investigated upon data collection."
}
}
}