v1 / README.md
albertvillanova's picture
Fix `license` metadata (#1)
a8158d1
metadata
language:
  - en
license:
  - unknown
multilinguality:
  - monolingual
size_categories:
  - 10K<n<100K

Dataset Card for CLUTRR

Table of Contents

Dataset Description

Dataset Summary

CLUTRR (Compositional Language Understanding and Text-based Relational Reasoning), a diagnostic benchmark suite, is first introduced in (https://arxiv.org/abs/1908.06177) to test the systematic generalization and inductive reasoning capabilities of NLU systems.

The CLUTRR benchmark allows us to test a model’s ability for systematic generalization by testing on stories that contain unseen combinations of logical rules, and test for the various forms of model robustness by adding different kinds of superfluous noise facts to the stories.

Dataset Task

CLUTRR contains a large set of semi-synthetic stories involving hypothetical families. The task is to infer the relationship between two family members, whose relationship is not explicitly mentioned in the given story.

Join the CLUTRR community in https://www.cs.mcgill.ca/~ksinha4/clutrr/

Dataset Structure

We show detailed information for all 14 configurations of the dataset.

configurations:

id: a unique series of characters and numbers that identify each instance
story: one semi-synthetic story involving hypothetical families
query: the target query/relation which contains two names, where the goal is to classify the relation that holds between these two entities
target: indicator for the correct relation for the query
target_text: text for the correct relation for the query
the indicator follows the rule as follows:
"aunt": 0, "son-in-law": 1, "grandfather": 2, "brother": 3, "sister": 4, "father": 5, "mother": 6, "grandmother": 7, "uncle": 8, "daughter-in-law": 9, "grandson": 10, "granddaughter": 11, "father-in-law": 12, "mother-in-law": 13, "nephew": 14, "son": 15, "daughter": 16, "niece": 17, "husband": 18, "wife": 19, "sister-in-law": 20
clean_story: the story without noise factors
proof_state: the logical rule of the kinship generation
f_comb: the kinships of the query followed by the logical rule
task_name: the task of the sub-dataset in a form of "task_[num1].[num2]"
The first number [num1] indicates the status of noise facts added in the story: 1- no noise facts; 2- Irrelevant facts*; 3- Supporting facts*; 4- Disconnected facts*.
The second number [num2] directly indicates the length of clauses for the task target.
for example:
task_1.2 -- task requiring clauses of length 2 without adding noise facts
task_2.3 -- task requiring clauses of length 3 with Irrelevant noise facts added in the story
story_edges: all the edges in the kinship graph
edge_types: similar to the f_comb, another form of the query's kinships followed by the logical rule
query_edge: the corresponding edge of the target query in the kinship graph
genders: genders of names appeared in the story
task_split: train,test

*Further explanation of Irrelevant facts, Supporting facts and Disconnected facts can be found in the 3.5 Robust Reasoning section in https://arxiv.org/abs/1908.06177

Data Instances

An example of 'train'in Task 1.2 looks as follows.

{
  "id": b2b9752f-d7fa-46a9-83ae-d474184c35b6,
  "story": "[Lillian] and her daughter [April] went to visit [Lillian]'s mother [Ashley] last Sunday.",
  "query": ('April', 'Ashley'),
  "target": 7,
  "target_text": "grandmother",
  "clean_story": [Lillian] and her daughter [April] went to visit [Lillian]'s mother [Ashley] last Sunday.,
  "proof_state": [{('April', 'grandmother', 'Ashley'): [('April', 'mother', 'Lillian'), ('Lillian', 'mother', 'Ashley')]}],
  "f_comb": "mother-mother",
  "task_name": "task_1.2",
  "story_edges": [(0, 1), (1, 2)],
  "edge_types": ['mother', 'mother'],
  "query_edge": (0, 2),
  "genders": "April:female,Lillian:female,Ashley:female",
  "task_split": trian
}

Data Splits

Data Split Name

(corresponding with the name used in the paper)

task_split split name in paper train &validation task test task
gen_train23_test2to10 data_089907f8 1.2, 1.3 1.2, 1.3, 1.4, 1.5, 1.6, 1.7, 1.8, 1.9, 1.10
gen_train234_test2to10 data_db9b8f04 1.2, 1.3, 1.4 1.2, 1.3, 1.4, 1.5, 1.6, 1.7, 1.8, 1.9, 1.10
rob_train_clean_23_test_all_23 data_7c5b0e70 1.2,1.3 1.2, 1.3, 2.3, 3.3, 4.3
rob_train_sup_23_test_all_23 data_06b8f2a1 2.2, 2.3 2.2, 2.3, 1.3, 3.3, 4.3
rob_train_irr_23_test_all_23 data_523348e6 3.2, 3.3 3.2, 3.3, 1.3, 2.3, 4.3
rob_train_disc_23_test_all_23 data_d83ecc3e 4.2, 4.3 4.2, 4.3, 1.3, 2.3, 3.3

Data Split Summary

Number of Instances in each split

task_split train validation test
gen_train23_test2to10 9074 2020 1146
gen_train234_test2to10 12064 3019 1048
rob_train_clean_23_test_all_23 8098 2026 447
rob_train_disc_23_test_all_23 8080 2020 445
rob_train_irr_23_test_all_23 8079 2020 444
rob_train_sup_23_test_all_23 8123 2031 447

Citation Information

@article{sinha2019clutrr,
  Author = {Koustuv Sinha and Shagun Sodhani and Jin Dong and Joelle Pineau and William L. Hamilton},
  Title = {CLUTRR: A Diagnostic Benchmark for Inductive Reasoning from Text},
  Year = {2019},
  journal = {Empirical Methods of Natural Language Processing (EMNLP)},
  arxiv = {1908.06177}
}