albertvillanova HF staff commited on
Commit
1f25e28
1 Parent(s): 0f91fab

Convert dataset to Parquet (#4)

Browse files

- Convert dataset to Parquet (3faa85ce535df5e2720d8aba1e466978369d3f05)
- Add presupposition_both_presupposition data files (006f6477b2674d64b28b0ea9591ee9818c2c5c7c)
- Add presupposition_change_of_state data files (c5204174310abde71ae83e17b1c347466f360c6f)
- Add presupposition_cleft_existence data files (0f4557b71fb5ed90628062e36e200434f1e12e68)
- Add presupposition_cleft_uniqueness data files (1f35490cb8cd732f8005fc606d4ed0bbfc3e851e)
- Add presupposition_only_presupposition data files (42b8ffb04ca2f8144f26ff0674f9295f0c4d392a)
- Add presupposition_possessed_definites_existence data files (62079a2e325f76ddb56086890efee775310cfd5b)
- Add presupposition_possessed_definites_uniqueness data files (53e393529f54a8d5e934bdafdf83b5db9346ee43)
- Add presupposition_question_presupposition data files (5a0266360fb9eab8b5f754c367f1ddae80636ba9)
- Add implicature_connectives data files (caa51dba1b16c05dda73427ee2674ccf10c87d61)
- Add implicature_gradable_adjective data files (0bbcb8646377d22d2e5a8fc19b1251846084847b)
- Add implicature_gradable_verb data files (f890c1ea7887e17094873aae10bd758f4841b776)
- Add implicature_modals data files (e62622d1b3be5544aa68bbafbd25d9aefe01d95c)
- Add implicature_numerals_10_100 data files (b3a53e8543c4b3af14352f5be7295f5b88e3012c)
- Add implicature_numerals_2_3 data files (4cf3b03c455e518b69ebc12c22d1d535ab1b158f)
- Add implicature_quantifiers data files (08c10033f8a292ef3b54931b207cdbedb698d7ca)
- Delete loading script (98a418448876c84df74937be2b82925b5f924145)
- Delete legacy dataset_infos.json (c99d1621660287718fbc5675deca090afc8e8e1d)

README.md CHANGED
@@ -20,106 +20,245 @@ task_ids:
20
  paperswithcode_id: imppres
21
  pretty_name: IMPPRES
22
  dataset_info:
23
- - config_name: presupposition_all_n_presupposition
24
  features:
25
  - name: premise
26
  dtype: string
27
  - name: hypothesis
28
  dtype: string
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
  - name: trigger
30
  dtype: string
31
- - name: trigger1
32
  dtype: string
33
- - name: trigger2
 
 
 
 
 
 
 
 
34
  dtype: string
35
- - name: presupposition
36
  dtype: string
37
- - name: gold_label
38
  dtype:
39
  class_label:
40
  names:
41
  '0': entailment
42
  '1': neutral
43
  '2': contradiction
44
- - name: UID
 
 
 
 
 
 
 
45
  dtype: string
46
- - name: pairID
 
 
 
 
47
  dtype: string
48
- - name: paradigmID
49
- dtype: int16
50
  splits:
51
- - name: all_n_presupposition
52
- num_bytes: 458492
53
- num_examples: 1900
54
- download_size: 335088
55
- dataset_size: 458492
56
- - config_name: presupposition_both_presupposition
57
  features:
58
  - name: premise
59
  dtype: string
60
  - name: hypothesis
61
  dtype: string
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
62
  - name: trigger
63
  dtype: string
64
- - name: trigger1
65
  dtype: string
66
- - name: trigger2
 
 
 
 
 
 
 
 
67
  dtype: string
68
- - name: presupposition
69
  dtype: string
70
- - name: gold_label
71
  dtype:
72
  class_label:
73
  names:
74
  '0': entailment
75
  '1': neutral
76
  '2': contradiction
77
- - name: UID
 
 
 
 
 
 
 
78
  dtype: string
79
- - name: pairID
 
 
 
 
80
  dtype: string
81
- - name: paradigmID
82
- dtype: int16
83
  splits:
84
- - name: both_presupposition
85
- num_bytes: 432792
86
- num_examples: 1900
87
- download_size: 335088
88
- dataset_size: 432792
89
- - config_name: presupposition_change_of_state
90
  features:
91
  - name: premise
92
  dtype: string
93
  - name: hypothesis
94
  dtype: string
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
95
  - name: trigger
96
  dtype: string
97
- - name: trigger1
98
  dtype: string
99
- - name: trigger2
 
 
 
 
 
 
 
 
100
  dtype: string
101
- - name: presupposition
102
  dtype: string
103
- - name: gold_label
104
  dtype:
105
  class_label:
106
  names:
107
  '0': entailment
108
  '1': neutral
109
  '2': contradiction
110
- - name: UID
 
 
 
 
 
 
 
111
  dtype: string
112
- - name: pairID
 
 
 
 
113
  dtype: string
114
- - name: paradigmID
115
- dtype: int16
116
  splits:
117
- - name: change_of_state
118
- num_bytes: 308627
119
- num_examples: 1900
120
- download_size: 335088
121
- dataset_size: 308627
122
- - config_name: presupposition_cleft_existence
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
123
  features:
124
  - name: premise
125
  dtype: string
@@ -147,12 +286,12 @@ dataset_info:
147
  - name: paradigmID
148
  dtype: int16
149
  splits:
150
- - name: cleft_existence
151
- num_bytes: 363238
152
  num_examples: 1900
153
- download_size: 335088
154
- dataset_size: 363238
155
- - config_name: presupposition_cleft_uniqueness
156
  features:
157
  - name: premise
158
  dtype: string
@@ -180,12 +319,12 @@ dataset_info:
180
  - name: paradigmID
181
  dtype: int16
182
  splits:
183
- - name: cleft_uniqueness
184
- num_bytes: 388779
185
  num_examples: 1900
186
- download_size: 335088
187
- dataset_size: 388779
188
- - config_name: presupposition_only_presupposition
189
  features:
190
  - name: premise
191
  dtype: string
@@ -213,12 +352,12 @@ dataset_info:
213
  - name: paradigmID
214
  dtype: int16
215
  splits:
216
- - name: only_presupposition
217
- num_bytes: 349018
218
  num_examples: 1900
219
- download_size: 335088
220
- dataset_size: 349018
221
- - config_name: presupposition_possessed_definites_existence
222
  features:
223
  - name: premise
224
  dtype: string
@@ -246,12 +385,12 @@ dataset_info:
246
  - name: paradigmID
247
  dtype: int16
248
  splits:
249
- - name: possessed_definites_existence
250
- num_bytes: 362334
251
  num_examples: 1900
252
- download_size: 335088
253
- dataset_size: 362334
254
- - config_name: presupposition_possessed_definites_uniqueness
255
  features:
256
  - name: premise
257
  dtype: string
@@ -279,12 +418,12 @@ dataset_info:
279
  - name: paradigmID
280
  dtype: int16
281
  splits:
282
- - name: possessed_definites_uniqueness
283
- num_bytes: 459403
284
  num_examples: 1900
285
- download_size: 335088
286
- dataset_size: 459403
287
- - config_name: presupposition_question_presupposition
288
  features:
289
  - name: premise
290
  dtype: string
@@ -312,249 +451,175 @@ dataset_info:
312
  - name: paradigmID
313
  dtype: int16
314
  splits:
315
- - name: question_presupposition
316
- num_bytes: 397227
317
  num_examples: 1900
318
- download_size: 335088
319
- dataset_size: 397227
320
- - config_name: implicature_connectives
321
  features:
322
  - name: premise
323
  dtype: string
324
  - name: hypothesis
325
  dtype: string
326
- - name: gold_label_log
327
- dtype:
328
- class_label:
329
- names:
330
- '0': entailment
331
- '1': neutral
332
- '2': contradiction
333
- - name: gold_label_prag
334
- dtype:
335
- class_label:
336
- names:
337
- '0': entailment
338
- '1': neutral
339
- '2': contradiction
340
- - name: spec_relation
341
- dtype: string
342
- - name: item_type
343
- dtype: string
344
  - name: trigger
345
  dtype: string
346
- - name: lexemes
347
  dtype: string
348
- splits:
349
- - name: connectives
350
- num_bytes: 221868
351
- num_examples: 1200
352
- download_size: 335088
353
- dataset_size: 221868
354
- - config_name: implicature_gradable_adjective
355
- features:
356
- - name: premise
357
  dtype: string
358
- - name: hypothesis
359
  dtype: string
360
- - name: gold_label_log
361
- dtype:
362
- class_label:
363
- names:
364
- '0': entailment
365
- '1': neutral
366
- '2': contradiction
367
- - name: gold_label_prag
368
  dtype:
369
  class_label:
370
  names:
371
  '0': entailment
372
  '1': neutral
373
  '2': contradiction
374
- - name: spec_relation
375
- dtype: string
376
- - name: item_type
377
- dtype: string
378
- - name: trigger
379
  dtype: string
380
- - name: lexemes
381
  dtype: string
 
 
382
  splits:
383
- - name: gradable_adjective
384
- num_bytes: 153672
385
- num_examples: 1200
386
- download_size: 335088
387
- dataset_size: 153672
388
- - config_name: implicature_gradable_verb
389
  features:
390
  - name: premise
391
  dtype: string
392
  - name: hypothesis
393
  dtype: string
394
- - name: gold_label_log
395
- dtype:
396
- class_label:
397
- names:
398
- '0': entailment
399
- '1': neutral
400
- '2': contradiction
401
- - name: gold_label_prag
402
- dtype:
403
- class_label:
404
- names:
405
- '0': entailment
406
- '1': neutral
407
- '2': contradiction
408
- - name: spec_relation
409
- dtype: string
410
- - name: item_type
411
- dtype: string
412
  - name: trigger
413
  dtype: string
414
- - name: lexemes
415
  dtype: string
416
- splits:
417
- - name: gradable_verb
418
- num_bytes: 180702
419
- num_examples: 1200
420
- download_size: 335088
421
- dataset_size: 180702
422
- - config_name: implicature_modals
423
- features:
424
- - name: premise
425
  dtype: string
426
- - name: hypothesis
427
  dtype: string
428
- - name: gold_label_log
429
- dtype:
430
- class_label:
431
- names:
432
- '0': entailment
433
- '1': neutral
434
- '2': contradiction
435
- - name: gold_label_prag
436
  dtype:
437
  class_label:
438
  names:
439
  '0': entailment
440
  '1': neutral
441
  '2': contradiction
442
- - name: spec_relation
443
- dtype: string
444
- - name: item_type
445
- dtype: string
446
- - name: trigger
447
  dtype: string
448
- - name: lexemes
449
  dtype: string
 
 
450
  splits:
451
- - name: modals
452
- num_bytes: 178560
453
- num_examples: 1200
454
- download_size: 335088
455
- dataset_size: 178560
456
- - config_name: implicature_numerals_10_100
457
  features:
458
  - name: premise
459
  dtype: string
460
  - name: hypothesis
461
  dtype: string
462
- - name: gold_label_log
463
- dtype:
464
- class_label:
465
- names:
466
- '0': entailment
467
- '1': neutral
468
- '2': contradiction
469
- - name: gold_label_prag
470
- dtype:
471
- class_label:
472
- names:
473
- '0': entailment
474
- '1': neutral
475
- '2': contradiction
476
- - name: spec_relation
477
- dtype: string
478
- - name: item_type
479
- dtype: string
480
  - name: trigger
481
  dtype: string
482
- - name: lexemes
483
  dtype: string
484
- splits:
485
- - name: numerals_10_100
486
- num_bytes: 208620
487
- num_examples: 1200
488
- download_size: 335088
489
- dataset_size: 208620
490
- - config_name: implicature_numerals_2_3
491
- features:
492
- - name: premise
493
  dtype: string
494
- - name: hypothesis
495
  dtype: string
496
- - name: gold_label_log
497
- dtype:
498
- class_label:
499
- names:
500
- '0': entailment
501
- '1': neutral
502
- '2': contradiction
503
- - name: gold_label_prag
504
  dtype:
505
  class_label:
506
  names:
507
  '0': entailment
508
  '1': neutral
509
  '2': contradiction
510
- - name: spec_relation
511
- dtype: string
512
- - name: item_type
513
- dtype: string
514
- - name: trigger
515
  dtype: string
516
- - name: lexemes
517
  dtype: string
 
 
518
  splits:
519
- - name: numerals_2_3
520
- num_bytes: 188784
521
- num_examples: 1200
522
- download_size: 335088
523
- dataset_size: 188784
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
524
  - config_name: implicature_quantifiers
525
- features:
526
- - name: premise
527
- dtype: string
528
- - name: hypothesis
529
- dtype: string
530
- - name: gold_label_log
531
- dtype:
532
- class_label:
533
- names:
534
- '0': entailment
535
- '1': neutral
536
- '2': contradiction
537
- - name: gold_label_prag
538
- dtype:
539
- class_label:
540
- names:
541
- '0': entailment
542
- '1': neutral
543
- '2': contradiction
544
- - name: spec_relation
545
- dtype: string
546
- - name: item_type
547
- dtype: string
548
- - name: trigger
549
- dtype: string
550
- - name: lexemes
551
- dtype: string
552
- splits:
553
- - name: quantifiers
554
- num_bytes: 176814
555
- num_examples: 1200
556
- download_size: 335088
557
- dataset_size: 176814
 
 
 
 
 
 
558
  ---
559
 
560
  # Dataset Card for IMPPRES
20
  paperswithcode_id: imppres
21
  pretty_name: IMPPRES
22
  dataset_info:
23
+ - config_name: implicature_connectives
24
  features:
25
  - name: premise
26
  dtype: string
27
  - name: hypothesis
28
  dtype: string
29
+ - name: gold_label_log
30
+ dtype:
31
+ class_label:
32
+ names:
33
+ '0': entailment
34
+ '1': neutral
35
+ '2': contradiction
36
+ - name: gold_label_prag
37
+ dtype:
38
+ class_label:
39
+ names:
40
+ '0': entailment
41
+ '1': neutral
42
+ '2': contradiction
43
+ - name: spec_relation
44
+ dtype: string
45
+ - name: item_type
46
+ dtype: string
47
  - name: trigger
48
  dtype: string
49
+ - name: lexemes
50
  dtype: string
51
+ splits:
52
+ - name: connectives
53
+ num_bytes: 221844
54
+ num_examples: 1200
55
+ download_size: 25478
56
+ dataset_size: 221844
57
+ - config_name: implicature_gradable_adjective
58
+ features:
59
+ - name: premise
60
  dtype: string
61
+ - name: hypothesis
62
  dtype: string
63
+ - name: gold_label_log
64
  dtype:
65
  class_label:
66
  names:
67
  '0': entailment
68
  '1': neutral
69
  '2': contradiction
70
+ - name: gold_label_prag
71
+ dtype:
72
+ class_label:
73
+ names:
74
+ '0': entailment
75
+ '1': neutral
76
+ '2': contradiction
77
+ - name: spec_relation
78
  dtype: string
79
+ - name: item_type
80
+ dtype: string
81
+ - name: trigger
82
+ dtype: string
83
+ - name: lexemes
84
  dtype: string
 
 
85
  splits:
86
+ - name: gradable_adjective
87
+ num_bytes: 153648
88
+ num_examples: 1200
89
+ download_size: 17337
90
+ dataset_size: 153648
91
+ - config_name: implicature_gradable_verb
92
  features:
93
  - name: premise
94
  dtype: string
95
  - name: hypothesis
96
  dtype: string
97
+ - name: gold_label_log
98
+ dtype:
99
+ class_label:
100
+ names:
101
+ '0': entailment
102
+ '1': neutral
103
+ '2': contradiction
104
+ - name: gold_label_prag
105
+ dtype:
106
+ class_label:
107
+ names:
108
+ '0': entailment
109
+ '1': neutral
110
+ '2': contradiction
111
+ - name: spec_relation
112
+ dtype: string
113
+ - name: item_type
114
+ dtype: string
115
  - name: trigger
116
  dtype: string
117
+ - name: lexemes
118
  dtype: string
119
+ splits:
120
+ - name: gradable_verb
121
+ num_bytes: 180678
122
+ num_examples: 1200
123
+ download_size: 21504
124
+ dataset_size: 180678
125
+ - config_name: implicature_modals
126
+ features:
127
+ - name: premise
128
  dtype: string
129
+ - name: hypothesis
130
  dtype: string
131
+ - name: gold_label_log
132
  dtype:
133
  class_label:
134
  names:
135
  '0': entailment
136
  '1': neutral
137
  '2': contradiction
138
+ - name: gold_label_prag
139
+ dtype:
140
+ class_label:
141
+ names:
142
+ '0': entailment
143
+ '1': neutral
144
+ '2': contradiction
145
+ - name: spec_relation
146
  dtype: string
147
+ - name: item_type
148
+ dtype: string
149
+ - name: trigger
150
+ dtype: string
151
+ - name: lexemes
152
  dtype: string
 
 
153
  splits:
154
+ - name: modals
155
+ num_bytes: 178536
156
+ num_examples: 1200
157
+ download_size: 21179
158
+ dataset_size: 178536
159
+ - config_name: implicature_numerals_10_100
160
  features:
161
  - name: premise
162
  dtype: string
163
  - name: hypothesis
164
  dtype: string
165
+ - name: gold_label_log
166
+ dtype:
167
+ class_label:
168
+ names:
169
+ '0': entailment
170
+ '1': neutral
171
+ '2': contradiction
172
+ - name: gold_label_prag
173
+ dtype:
174
+ class_label:
175
+ names:
176
+ '0': entailment
177
+ '1': neutral
178
+ '2': contradiction
179
+ - name: spec_relation
180
+ dtype: string
181
+ - name: item_type
182
+ dtype: string
183
  - name: trigger
184
  dtype: string
185
+ - name: lexemes
186
  dtype: string
187
+ splits:
188
+ - name: numerals_10_100
189
+ num_bytes: 208596
190
+ num_examples: 1200
191
+ download_size: 22640
192
+ dataset_size: 208596
193
+ - config_name: implicature_numerals_2_3
194
+ features:
195
+ - name: premise
196
  dtype: string
197
+ - name: hypothesis
198
  dtype: string
199
+ - name: gold_label_log
200
  dtype:
201
  class_label:
202
  names:
203
  '0': entailment
204
  '1': neutral
205
  '2': contradiction
206
+ - name: gold_label_prag
207
+ dtype:
208
+ class_label:
209
+ names:
210
+ '0': entailment
211
+ '1': neutral
212
+ '2': contradiction
213
+ - name: spec_relation
214
  dtype: string
215
+ - name: item_type
216
+ dtype: string
217
+ - name: trigger
218
+ dtype: string
219
+ - name: lexemes
220
  dtype: string
 
 
221
  splits:
222
+ - name: numerals_2_3
223
+ num_bytes: 188760
224
+ num_examples: 1200
225
+ download_size: 22218
226
+ dataset_size: 188760
227
+ - config_name: implicature_quantifiers
228
+ features:
229
+ - name: premise
230
+ dtype: string
231
+ - name: hypothesis
232
+ dtype: string
233
+ - name: gold_label_log
234
+ dtype:
235
+ class_label:
236
+ names:
237
+ '0': entailment
238
+ '1': neutral
239
+ '2': contradiction
240
+ - name: gold_label_prag
241
+ dtype:
242
+ class_label:
243
+ names:
244
+ '0': entailment
245
+ '1': neutral
246
+ '2': contradiction
247
+ - name: spec_relation
248
+ dtype: string
249
+ - name: item_type
250
+ dtype: string
251
+ - name: trigger
252
+ dtype: string
253
+ - name: lexemes
254
+ dtype: string
255
+ splits:
256
+ - name: quantifiers
257
+ num_bytes: 176790
258
+ num_examples: 1200
259
+ download_size: 21017
260
+ dataset_size: 176790
261
+ - config_name: presupposition_all_n_presupposition
262
  features:
263
  - name: premise
264
  dtype: string
286
  - name: paradigmID
287
  dtype: int16
288
  splits:
289
+ - name: all_n_presupposition
290
+ num_bytes: 458460
291
  num_examples: 1900
292
+ download_size: 43038
293
+ dataset_size: 458460
294
+ - config_name: presupposition_both_presupposition
295
  features:
296
  - name: premise
297
  dtype: string
319
  - name: paradigmID
320
  dtype: int16
321
  splits:
322
+ - name: both_presupposition
323
+ num_bytes: 432760
324
  num_examples: 1900
325
+ download_size: 41142
326
+ dataset_size: 432760
327
+ - config_name: presupposition_change_of_state
328
  features:
329
  - name: premise
330
  dtype: string
352
  - name: paradigmID
353
  dtype: int16
354
  splits:
355
+ - name: change_of_state
356
+ num_bytes: 308595
357
  num_examples: 1900
358
+ download_size: 35814
359
+ dataset_size: 308595
360
+ - config_name: presupposition_cleft_existence
361
  features:
362
  - name: premise
363
  dtype: string
385
  - name: paradigmID
386
  dtype: int16
387
  splits:
388
+ - name: cleft_existence
389
+ num_bytes: 363206
390
  num_examples: 1900
391
+ download_size: 37597
392
+ dataset_size: 363206
393
+ - config_name: presupposition_cleft_uniqueness
394
  features:
395
  - name: premise
396
  dtype: string
418
  - name: paradigmID
419
  dtype: int16
420
  splits:
421
+ - name: cleft_uniqueness
422
+ num_bytes: 388747
423
  num_examples: 1900
424
+ download_size: 38279
425
+ dataset_size: 388747
426
+ - config_name: presupposition_only_presupposition
427
  features:
428
  - name: premise
429
  dtype: string
451
  - name: paradigmID
452
  dtype: int16
453
  splits:
454
+ - name: only_presupposition
455
+ num_bytes: 348986
456
  num_examples: 1900
457
+ download_size: 38126
458
+ dataset_size: 348986
459
+ - config_name: presupposition_possessed_definites_existence
460
  features:
461
  - name: premise
462
  dtype: string
463
  - name: hypothesis
464
  dtype: string
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
465
  - name: trigger
466
  dtype: string
467
+ - name: trigger1
468
  dtype: string
469
+ - name: trigger2
 
 
 
 
 
 
 
 
470
  dtype: string
471
+ - name: presupposition
472
  dtype: string
473
+ - name: gold_label
 
 
 
 
 
 
 
474
  dtype:
475
  class_label:
476
  names:
477
  '0': entailment
478
  '1': neutral
479
  '2': contradiction
480
+ - name: UID
 
 
 
 
481
  dtype: string
482
+ - name: pairID
483
  dtype: string
484
+ - name: paradigmID
485
+ dtype: int16
486
  splits:
487
+ - name: possessed_definites_existence
488
+ num_bytes: 362302
489
+ num_examples: 1900
490
+ download_size: 38712
491
+ dataset_size: 362302
492
+ - config_name: presupposition_possessed_definites_uniqueness
493
  features:
494
  - name: premise
495
  dtype: string
496
  - name: hypothesis
497
  dtype: string
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
498
  - name: trigger
499
  dtype: string
500
+ - name: trigger1
501
  dtype: string
502
+ - name: trigger2
 
 
 
 
 
 
 
 
503
  dtype: string
504
+ - name: presupposition
505
  dtype: string
506
+ - name: gold_label
 
 
 
 
 
 
 
507
  dtype:
508
  class_label:
509
  names:
510
  '0': entailment
511
  '1': neutral
512
  '2': contradiction
513
+ - name: UID
 
 
 
 
514
  dtype: string
515
+ - name: pairID
516
  dtype: string
517
+ - name: paradigmID
518
+ dtype: int16
519
  splits:
520
+ - name: possessed_definites_uniqueness
521
+ num_bytes: 459371
522
+ num_examples: 1900
523
+ download_size: 42068
524
+ dataset_size: 459371
525
+ - config_name: presupposition_question_presupposition
526
  features:
527
  - name: premise
528
  dtype: string
529
  - name: hypothesis
530
  dtype: string
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
531
  - name: trigger
532
  dtype: string
533
+ - name: trigger1
534
  dtype: string
535
+ - name: trigger2
 
 
 
 
 
 
 
 
536
  dtype: string
537
+ - name: presupposition
538
  dtype: string
539
+ - name: gold_label
 
 
 
 
 
 
 
540
  dtype:
541
  class_label:
542
  names:
543
  '0': entailment
544
  '1': neutral
545
  '2': contradiction
546
+ - name: UID
 
 
 
 
547
  dtype: string
548
+ - name: pairID
549
  dtype: string
550
+ - name: paradigmID
551
+ dtype: int16
552
  splits:
553
+ - name: question_presupposition
554
+ num_bytes: 397195
555
+ num_examples: 1900
556
+ download_size: 41247
557
+ dataset_size: 397195
558
+ configs:
559
+ - config_name: implicature_connectives
560
+ data_files:
561
+ - split: connectives
562
+ path: implicature_connectives/connectives-*
563
+ - config_name: implicature_gradable_adjective
564
+ data_files:
565
+ - split: gradable_adjective
566
+ path: implicature_gradable_adjective/gradable_adjective-*
567
+ - config_name: implicature_gradable_verb
568
+ data_files:
569
+ - split: gradable_verb
570
+ path: implicature_gradable_verb/gradable_verb-*
571
+ - config_name: implicature_modals
572
+ data_files:
573
+ - split: modals
574
+ path: implicature_modals/modals-*
575
+ - config_name: implicature_numerals_10_100
576
+ data_files:
577
+ - split: numerals_10_100
578
+ path: implicature_numerals_10_100/numerals_10_100-*
579
+ - config_name: implicature_numerals_2_3
580
+ data_files:
581
+ - split: numerals_2_3
582
+ path: implicature_numerals_2_3/numerals_2_3-*
583
  - config_name: implicature_quantifiers
584
+ data_files:
585
+ - split: quantifiers
586
+ path: implicature_quantifiers/quantifiers-*
587
+ - config_name: presupposition_all_n_presupposition
588
+ data_files:
589
+ - split: all_n_presupposition
590
+ path: presupposition_all_n_presupposition/all_n_presupposition-*
591
+ - config_name: presupposition_both_presupposition
592
+ data_files:
593
+ - split: both_presupposition
594
+ path: presupposition_both_presupposition/both_presupposition-*
595
+ - config_name: presupposition_change_of_state
596
+ data_files:
597
+ - split: change_of_state
598
+ path: presupposition_change_of_state/change_of_state-*
599
+ - config_name: presupposition_cleft_existence
600
+ data_files:
601
+ - split: cleft_existence
602
+ path: presupposition_cleft_existence/cleft_existence-*
603
+ - config_name: presupposition_cleft_uniqueness
604
+ data_files:
605
+ - split: cleft_uniqueness
606
+ path: presupposition_cleft_uniqueness/cleft_uniqueness-*
607
+ - config_name: presupposition_only_presupposition
608
+ data_files:
609
+ - split: only_presupposition
610
+ path: presupposition_only_presupposition/only_presupposition-*
611
+ - config_name: presupposition_possessed_definites_existence
612
+ data_files:
613
+ - split: possessed_definites_existence
614
+ path: presupposition_possessed_definites_existence/possessed_definites_existence-*
615
+ - config_name: presupposition_possessed_definites_uniqueness
616
+ data_files:
617
+ - split: possessed_definites_uniqueness
618
+ path: presupposition_possessed_definites_uniqueness/possessed_definites_uniqueness-*
619
+ - config_name: presupposition_question_presupposition
620
+ data_files:
621
+ - split: question_presupposition
622
+ path: presupposition_question_presupposition/question_presupposition-*
623
  ---
624
 
625
  # Dataset Card for IMPPRES
dataset_infos.json DELETED
@@ -1 +0,0 @@
1
- {"presupposition_all_n_presupposition": {"description": "Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures.", "citation": "@inproceedings{jeretic-etal-2020-natural,\n title = \"Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}\",\n author = \"Jereti\u000b{c}, Paloma and\n Warstadt, Alex and\n Bhooshan, Suvrat and\n Williams, Adina\",\n booktitle = \"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics\",\n month = jul,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.acl-main.768\",\n doi = \"10.18653/v1/2020.acl-main.768\",\n pages = \"8690--8705\",\n abstract = \"Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.\",\n}\n", "homepage": "https://github.com/facebookresearch/Imppres", "license": "Creative Commons Attribution-NonCommercial 4.0 International Public License", "features": {"premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "trigger": {"dtype": "string", "id": null, "_type": "Value"}, "trigger1": {"dtype": "string", "id": null, "_type": "Value"}, "trigger2": {"dtype": "string", "id": null, "_type": "Value"}, "presupposition": {"dtype": "string", "id": null, "_type": "Value"}, "gold_label": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "UID": {"dtype": "string", "id": null, "_type": "Value"}, "pairID": {"dtype": "string", "id": null, "_type": "Value"}, "paradigmID": {"dtype": "int16", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "imppres", "config_name": "presupposition_all_n_presupposition", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"all_n_presupposition": {"name": "all_n_presupposition", "num_bytes": 458492, "num_examples": 1900, "dataset_name": "imppres"}}, "download_checksums": {"https://github.com/facebookresearch/Imppres/blob/master/dataset/IMPPRES.zip?raw=true": {"num_bytes": 335088, "checksum": "f1f4ab03aec2248dcfbcb00b80e6099c592751cca2b542b208c6cf46f2926937"}}, "download_size": 335088, "post_processing_size": null, "dataset_size": 458492, "size_in_bytes": 793580}, "presupposition_both_presupposition": {"description": "Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures.", "citation": "@inproceedings{jeretic-etal-2020-natural,\n title = \"Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}\",\n author = \"Jereti\u000b{c}, Paloma and\n Warstadt, Alex and\n Bhooshan, Suvrat and\n Williams, Adina\",\n booktitle = \"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics\",\n month = jul,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.acl-main.768\",\n doi = \"10.18653/v1/2020.acl-main.768\",\n pages = \"8690--8705\",\n abstract = \"Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.\",\n}\n", "homepage": "https://github.com/facebookresearch/Imppres", "license": "Creative Commons Attribution-NonCommercial 4.0 International Public License", "features": {"premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "trigger": {"dtype": "string", "id": null, "_type": "Value"}, "trigger1": {"dtype": "string", "id": null, "_type": "Value"}, "trigger2": {"dtype": "string", "id": null, "_type": "Value"}, "presupposition": {"dtype": "string", "id": null, "_type": "Value"}, "gold_label": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "UID": {"dtype": "string", "id": null, "_type": "Value"}, "pairID": {"dtype": "string", "id": null, "_type": "Value"}, "paradigmID": {"dtype": "int16", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "imppres", "config_name": "presupposition_both_presupposition", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"both_presupposition": {"name": "both_presupposition", "num_bytes": 432792, "num_examples": 1900, "dataset_name": "imppres"}}, "download_checksums": {"https://github.com/facebookresearch/Imppres/blob/master/dataset/IMPPRES.zip?raw=true": {"num_bytes": 335088, "checksum": "f1f4ab03aec2248dcfbcb00b80e6099c592751cca2b542b208c6cf46f2926937"}}, "download_size": 335088, "post_processing_size": null, "dataset_size": 432792, "size_in_bytes": 767880}, "presupposition_change_of_state": {"description": "Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures.", "citation": "@inproceedings{jeretic-etal-2020-natural,\n title = \"Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}\",\n author = \"Jereti\u000b{c}, Paloma and\n Warstadt, Alex and\n Bhooshan, Suvrat and\n Williams, Adina\",\n booktitle = \"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics\",\n month = jul,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.acl-main.768\",\n doi = \"10.18653/v1/2020.acl-main.768\",\n pages = \"8690--8705\",\n abstract = \"Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.\",\n}\n", "homepage": "https://github.com/facebookresearch/Imppres", "license": "Creative Commons Attribution-NonCommercial 4.0 International Public License", "features": {"premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "trigger": {"dtype": "string", "id": null, "_type": "Value"}, "trigger1": {"dtype": "string", "id": null, "_type": "Value"}, "trigger2": {"dtype": "string", "id": null, "_type": "Value"}, "presupposition": {"dtype": "string", "id": null, "_type": "Value"}, "gold_label": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "UID": {"dtype": "string", "id": null, "_type": "Value"}, "pairID": {"dtype": "string", "id": null, "_type": "Value"}, "paradigmID": {"dtype": "int16", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "imppres", "config_name": "presupposition_change_of_state", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"change_of_state": {"name": "change_of_state", "num_bytes": 308627, "num_examples": 1900, "dataset_name": "imppres"}}, "download_checksums": {"https://github.com/facebookresearch/Imppres/blob/master/dataset/IMPPRES.zip?raw=true": {"num_bytes": 335088, "checksum": "f1f4ab03aec2248dcfbcb00b80e6099c592751cca2b542b208c6cf46f2926937"}}, "download_size": 335088, "post_processing_size": null, "dataset_size": 308627, "size_in_bytes": 643715}, "presupposition_cleft_existence": {"description": "Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures.", "citation": "@inproceedings{jeretic-etal-2020-natural,\n title = \"Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}\",\n author = \"Jereti\u000b{c}, Paloma and\n Warstadt, Alex and\n Bhooshan, Suvrat and\n Williams, Adina\",\n booktitle = \"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics\",\n month = jul,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.acl-main.768\",\n doi = \"10.18653/v1/2020.acl-main.768\",\n pages = \"8690--8705\",\n abstract = \"Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.\",\n}\n", "homepage": "https://github.com/facebookresearch/Imppres", "license": "Creative Commons Attribution-NonCommercial 4.0 International Public License", "features": {"premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "trigger": {"dtype": "string", "id": null, "_type": "Value"}, "trigger1": {"dtype": "string", "id": null, "_type": "Value"}, "trigger2": {"dtype": "string", "id": null, "_type": "Value"}, "presupposition": {"dtype": "string", "id": null, "_type": "Value"}, "gold_label": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "UID": {"dtype": "string", "id": null, "_type": "Value"}, "pairID": {"dtype": "string", "id": null, "_type": "Value"}, "paradigmID": {"dtype": "int16", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "imppres", "config_name": "presupposition_cleft_existence", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"cleft_existence": {"name": "cleft_existence", "num_bytes": 363238, "num_examples": 1900, "dataset_name": "imppres"}}, "download_checksums": {"https://github.com/facebookresearch/Imppres/blob/master/dataset/IMPPRES.zip?raw=true": {"num_bytes": 335088, "checksum": "f1f4ab03aec2248dcfbcb00b80e6099c592751cca2b542b208c6cf46f2926937"}}, "download_size": 335088, "post_processing_size": null, "dataset_size": 363238, "size_in_bytes": 698326}, "presupposition_cleft_uniqueness": {"description": "Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures.", "citation": "@inproceedings{jeretic-etal-2020-natural,\n title = \"Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}\",\n author = \"Jereti\u000b{c}, Paloma and\n Warstadt, Alex and\n Bhooshan, Suvrat and\n Williams, Adina\",\n booktitle = \"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics\",\n month = jul,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.acl-main.768\",\n doi = \"10.18653/v1/2020.acl-main.768\",\n pages = \"8690--8705\",\n abstract = \"Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.\",\n}\n", "homepage": "https://github.com/facebookresearch/Imppres", "license": "Creative Commons Attribution-NonCommercial 4.0 International Public License", "features": {"premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "trigger": {"dtype": "string", "id": null, "_type": "Value"}, "trigger1": {"dtype": "string", "id": null, "_type": "Value"}, "trigger2": {"dtype": "string", "id": null, "_type": "Value"}, "presupposition": {"dtype": "string", "id": null, "_type": "Value"}, "gold_label": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "UID": {"dtype": "string", "id": null, "_type": "Value"}, "pairID": {"dtype": "string", "id": null, "_type": "Value"}, "paradigmID": {"dtype": "int16", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "imppres", "config_name": "presupposition_cleft_uniqueness", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"cleft_uniqueness": {"name": "cleft_uniqueness", "num_bytes": 388779, "num_examples": 1900, "dataset_name": "imppres"}}, "download_checksums": {"https://github.com/facebookresearch/Imppres/blob/master/dataset/IMPPRES.zip?raw=true": {"num_bytes": 335088, "checksum": "f1f4ab03aec2248dcfbcb00b80e6099c592751cca2b542b208c6cf46f2926937"}}, "download_size": 335088, "post_processing_size": null, "dataset_size": 388779, "size_in_bytes": 723867}, "presupposition_only_presupposition": {"description": "Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures.", "citation": "@inproceedings{jeretic-etal-2020-natural,\n title = \"Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}\",\n author = \"Jereti\u000b{c}, Paloma and\n Warstadt, Alex and\n Bhooshan, Suvrat and\n Williams, Adina\",\n booktitle = \"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics\",\n month = jul,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.acl-main.768\",\n doi = \"10.18653/v1/2020.acl-main.768\",\n pages = \"8690--8705\",\n abstract = \"Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.\",\n}\n", "homepage": "https://github.com/facebookresearch/Imppres", "license": "Creative Commons Attribution-NonCommercial 4.0 International Public License", "features": {"premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "trigger": {"dtype": "string", "id": null, "_type": "Value"}, "trigger1": {"dtype": "string", "id": null, "_type": "Value"}, "trigger2": {"dtype": "string", "id": null, "_type": "Value"}, "presupposition": {"dtype": "string", "id": null, "_type": "Value"}, "gold_label": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "UID": {"dtype": "string", "id": null, "_type": "Value"}, "pairID": {"dtype": "string", "id": null, "_type": "Value"}, "paradigmID": {"dtype": "int16", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "imppres", "config_name": "presupposition_only_presupposition", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"only_presupposition": {"name": "only_presupposition", "num_bytes": 349018, "num_examples": 1900, "dataset_name": "imppres"}}, "download_checksums": {"https://github.com/facebookresearch/Imppres/blob/master/dataset/IMPPRES.zip?raw=true": {"num_bytes": 335088, "checksum": "f1f4ab03aec2248dcfbcb00b80e6099c592751cca2b542b208c6cf46f2926937"}}, "download_size": 335088, "post_processing_size": null, "dataset_size": 349018, "size_in_bytes": 684106}, "presupposition_possessed_definites_existence": {"description": "Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures.", "citation": "@inproceedings{jeretic-etal-2020-natural,\n title = \"Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}\",\n author = \"Jereti\u000b{c}, Paloma and\n Warstadt, Alex and\n Bhooshan, Suvrat and\n Williams, Adina\",\n booktitle = \"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics\",\n month = jul,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.acl-main.768\",\n doi = \"10.18653/v1/2020.acl-main.768\",\n pages = \"8690--8705\",\n abstract = \"Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.\",\n}\n", "homepage": "https://github.com/facebookresearch/Imppres", "license": "Creative Commons Attribution-NonCommercial 4.0 International Public License", "features": {"premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "trigger": {"dtype": "string", "id": null, "_type": "Value"}, "trigger1": {"dtype": "string", "id": null, "_type": "Value"}, "trigger2": {"dtype": "string", "id": null, "_type": "Value"}, "presupposition": {"dtype": "string", "id": null, "_type": "Value"}, "gold_label": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "UID": {"dtype": "string", "id": null, "_type": "Value"}, "pairID": {"dtype": "string", "id": null, "_type": "Value"}, "paradigmID": {"dtype": "int16", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "imppres", "config_name": "presupposition_possessed_definites_existence", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"possessed_definites_existence": {"name": "possessed_definites_existence", "num_bytes": 362334, "num_examples": 1900, "dataset_name": "imppres"}}, "download_checksums": {"https://github.com/facebookresearch/Imppres/blob/master/dataset/IMPPRES.zip?raw=true": {"num_bytes": 335088, "checksum": "f1f4ab03aec2248dcfbcb00b80e6099c592751cca2b542b208c6cf46f2926937"}}, "download_size": 335088, "post_processing_size": null, "dataset_size": 362334, "size_in_bytes": 697422}, "presupposition_possessed_definites_uniqueness": {"description": "Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures.", "citation": "@inproceedings{jeretic-etal-2020-natural,\n title = \"Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}\",\n author = \"Jereti\u000b{c}, Paloma and\n Warstadt, Alex and\n Bhooshan, Suvrat and\n Williams, Adina\",\n booktitle = \"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics\",\n month = jul,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.acl-main.768\",\n doi = \"10.18653/v1/2020.acl-main.768\",\n pages = \"8690--8705\",\n abstract = \"Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.\",\n}\n", "homepage": "https://github.com/facebookresearch/Imppres", "license": "Creative Commons Attribution-NonCommercial 4.0 International Public License", "features": {"premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "trigger": {"dtype": "string", "id": null, "_type": "Value"}, "trigger1": {"dtype": "string", "id": null, "_type": "Value"}, "trigger2": {"dtype": "string", "id": null, "_type": "Value"}, "presupposition": {"dtype": "string", "id": null, "_type": "Value"}, "gold_label": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "UID": {"dtype": "string", "id": null, "_type": "Value"}, "pairID": {"dtype": "string", "id": null, "_type": "Value"}, "paradigmID": {"dtype": "int16", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "imppres", "config_name": "presupposition_possessed_definites_uniqueness", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"possessed_definites_uniqueness": {"name": "possessed_definites_uniqueness", "num_bytes": 459403, "num_examples": 1900, "dataset_name": "imppres"}}, "download_checksums": {"https://github.com/facebookresearch/Imppres/blob/master/dataset/IMPPRES.zip?raw=true": {"num_bytes": 335088, "checksum": "f1f4ab03aec2248dcfbcb00b80e6099c592751cca2b542b208c6cf46f2926937"}}, "download_size": 335088, "post_processing_size": null, "dataset_size": 459403, "size_in_bytes": 794491}, "presupposition_question_presupposition": {"description": "Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures.", "citation": "@inproceedings{jeretic-etal-2020-natural,\n title = \"Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}\",\n author = \"Jereti\u000b{c}, Paloma and\n Warstadt, Alex and\n Bhooshan, Suvrat and\n Williams, Adina\",\n booktitle = \"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics\",\n month = jul,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.acl-main.768\",\n doi = \"10.18653/v1/2020.acl-main.768\",\n pages = \"8690--8705\",\n abstract = \"Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.\",\n}\n", "homepage": "https://github.com/facebookresearch/Imppres", "license": "Creative Commons Attribution-NonCommercial 4.0 International Public License", "features": {"premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "trigger": {"dtype": "string", "id": null, "_type": "Value"}, "trigger1": {"dtype": "string", "id": null, "_type": "Value"}, "trigger2": {"dtype": "string", "id": null, "_type": "Value"}, "presupposition": {"dtype": "string", "id": null, "_type": "Value"}, "gold_label": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "UID": {"dtype": "string", "id": null, "_type": "Value"}, "pairID": {"dtype": "string", "id": null, "_type": "Value"}, "paradigmID": {"dtype": "int16", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "imppres", "config_name": "presupposition_question_presupposition", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"question_presupposition": {"name": "question_presupposition", "num_bytes": 397227, "num_examples": 1900, "dataset_name": "imppres"}}, "download_checksums": {"https://github.com/facebookresearch/Imppres/blob/master/dataset/IMPPRES.zip?raw=true": {"num_bytes": 335088, "checksum": "f1f4ab03aec2248dcfbcb00b80e6099c592751cca2b542b208c6cf46f2926937"}}, "download_size": 335088, "post_processing_size": null, "dataset_size": 397227, "size_in_bytes": 732315}, "implicature_connectives": {"description": "Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures.", "citation": "@inproceedings{jeretic-etal-2020-natural,\n title = \"Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}\",\n author = \"Jereti\u000b{c}, Paloma and\n Warstadt, Alex and\n Bhooshan, Suvrat and\n Williams, Adina\",\n booktitle = \"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics\",\n month = jul,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.acl-main.768\",\n doi = \"10.18653/v1/2020.acl-main.768\",\n pages = \"8690--8705\",\n abstract = \"Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.\",\n}\n", "homepage": "https://github.com/facebookresearch/Imppres", "license": "Creative Commons Attribution-NonCommercial 4.0 International Public License", "features": {"premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "gold_label_log": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "gold_label_prag": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "spec_relation": {"dtype": "string", "id": null, "_type": "Value"}, "item_type": {"dtype": "string", "id": null, "_type": "Value"}, "trigger": {"dtype": "string", "id": null, "_type": "Value"}, "lexemes": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "imppres", "config_name": "implicature_connectives", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"connectives": {"name": "connectives", "num_bytes": 221868, "num_examples": 1200, "dataset_name": "imppres"}}, "download_checksums": {"https://github.com/facebookresearch/Imppres/blob/master/dataset/IMPPRES.zip?raw=true": {"num_bytes": 335088, "checksum": "f1f4ab03aec2248dcfbcb00b80e6099c592751cca2b542b208c6cf46f2926937"}}, "download_size": 335088, "post_processing_size": null, "dataset_size": 221868, "size_in_bytes": 556956}, "implicature_gradable_adjective": {"description": "Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures.", "citation": "@inproceedings{jeretic-etal-2020-natural,\n title = \"Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}\",\n author = \"Jereti\u000b{c}, Paloma and\n Warstadt, Alex and\n Bhooshan, Suvrat and\n Williams, Adina\",\n booktitle = \"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics\",\n month = jul,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.acl-main.768\",\n doi = \"10.18653/v1/2020.acl-main.768\",\n pages = \"8690--8705\",\n abstract = \"Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.\",\n}\n", "homepage": "https://github.com/facebookresearch/Imppres", "license": "Creative Commons Attribution-NonCommercial 4.0 International Public License", "features": {"premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "gold_label_log": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "gold_label_prag": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "spec_relation": {"dtype": "string", "id": null, "_type": "Value"}, "item_type": {"dtype": "string", "id": null, "_type": "Value"}, "trigger": {"dtype": "string", "id": null, "_type": "Value"}, "lexemes": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "imppres", "config_name": "implicature_gradable_adjective", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"gradable_adjective": {"name": "gradable_adjective", "num_bytes": 153672, "num_examples": 1200, "dataset_name": "imppres"}}, "download_checksums": {"https://github.com/facebookresearch/Imppres/blob/master/dataset/IMPPRES.zip?raw=true": {"num_bytes": 335088, "checksum": "f1f4ab03aec2248dcfbcb00b80e6099c592751cca2b542b208c6cf46f2926937"}}, "download_size": 335088, "post_processing_size": null, "dataset_size": 153672, "size_in_bytes": 488760}, "implicature_gradable_verb": {"description": "Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures.", "citation": "@inproceedings{jeretic-etal-2020-natural,\n title = \"Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}\",\n author = \"Jereti\u000b{c}, Paloma and\n Warstadt, Alex and\n Bhooshan, Suvrat and\n Williams, Adina\",\n booktitle = \"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics\",\n month = jul,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.acl-main.768\",\n doi = \"10.18653/v1/2020.acl-main.768\",\n pages = \"8690--8705\",\n abstract = \"Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.\",\n}\n", "homepage": "https://github.com/facebookresearch/Imppres", "license": "Creative Commons Attribution-NonCommercial 4.0 International Public License", "features": {"premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "gold_label_log": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "gold_label_prag": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "spec_relation": {"dtype": "string", "id": null, "_type": "Value"}, "item_type": {"dtype": "string", "id": null, "_type": "Value"}, "trigger": {"dtype": "string", "id": null, "_type": "Value"}, "lexemes": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "imppres", "config_name": "implicature_gradable_verb", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"gradable_verb": {"name": "gradable_verb", "num_bytes": 180702, "num_examples": 1200, "dataset_name": "imppres"}}, "download_checksums": {"https://github.com/facebookresearch/Imppres/blob/master/dataset/IMPPRES.zip?raw=true": {"num_bytes": 335088, "checksum": "f1f4ab03aec2248dcfbcb00b80e6099c592751cca2b542b208c6cf46f2926937"}}, "download_size": 335088, "post_processing_size": null, "dataset_size": 180702, "size_in_bytes": 515790}, "implicature_modals": {"description": "Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures.", "citation": "@inproceedings{jeretic-etal-2020-natural,\n title = \"Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}\",\n author = \"Jereti\u000b{c}, Paloma and\n Warstadt, Alex and\n Bhooshan, Suvrat and\n Williams, Adina\",\n booktitle = \"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics\",\n month = jul,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.acl-main.768\",\n doi = \"10.18653/v1/2020.acl-main.768\",\n pages = \"8690--8705\",\n abstract = \"Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.\",\n}\n", "homepage": "https://github.com/facebookresearch/Imppres", "license": "Creative Commons Attribution-NonCommercial 4.0 International Public License", "features": {"premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "gold_label_log": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "gold_label_prag": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "spec_relation": {"dtype": "string", "id": null, "_type": "Value"}, "item_type": {"dtype": "string", "id": null, "_type": "Value"}, "trigger": {"dtype": "string", "id": null, "_type": "Value"}, "lexemes": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "imppres", "config_name": "implicature_modals", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"modals": {"name": "modals", "num_bytes": 178560, "num_examples": 1200, "dataset_name": "imppres"}}, "download_checksums": {"https://github.com/facebookresearch/Imppres/blob/master/dataset/IMPPRES.zip?raw=true": {"num_bytes": 335088, "checksum": "f1f4ab03aec2248dcfbcb00b80e6099c592751cca2b542b208c6cf46f2926937"}}, "download_size": 335088, "post_processing_size": null, "dataset_size": 178560, "size_in_bytes": 513648}, "implicature_numerals_10_100": {"description": "Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures.", "citation": "@inproceedings{jeretic-etal-2020-natural,\n title = \"Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}\",\n author = \"Jereti\u000b{c}, Paloma and\n Warstadt, Alex and\n Bhooshan, Suvrat and\n Williams, Adina\",\n booktitle = \"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics\",\n month = jul,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.acl-main.768\",\n doi = \"10.18653/v1/2020.acl-main.768\",\n pages = \"8690--8705\",\n abstract = \"Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.\",\n}\n", "homepage": "https://github.com/facebookresearch/Imppres", "license": "Creative Commons Attribution-NonCommercial 4.0 International Public License", "features": {"premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "gold_label_log": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "gold_label_prag": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "spec_relation": {"dtype": "string", "id": null, "_type": "Value"}, "item_type": {"dtype": "string", "id": null, "_type": "Value"}, "trigger": {"dtype": "string", "id": null, "_type": "Value"}, "lexemes": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "imppres", "config_name": "implicature_numerals_10_100", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"numerals_10_100": {"name": "numerals_10_100", "num_bytes": 208620, "num_examples": 1200, "dataset_name": "imppres"}}, "download_checksums": {"https://github.com/facebookresearch/Imppres/blob/master/dataset/IMPPRES.zip?raw=true": {"num_bytes": 335088, "checksum": "f1f4ab03aec2248dcfbcb00b80e6099c592751cca2b542b208c6cf46f2926937"}}, "download_size": 335088, "post_processing_size": null, "dataset_size": 208620, "size_in_bytes": 543708}, "implicature_numerals_2_3": {"description": "Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures.", "citation": "@inproceedings{jeretic-etal-2020-natural,\n title = \"Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}\",\n author = \"Jereti\u000b{c}, Paloma and\n Warstadt, Alex and\n Bhooshan, Suvrat and\n Williams, Adina\",\n booktitle = \"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics\",\n month = jul,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.acl-main.768\",\n doi = \"10.18653/v1/2020.acl-main.768\",\n pages = \"8690--8705\",\n abstract = \"Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.\",\n}\n", "homepage": "https://github.com/facebookresearch/Imppres", "license": "Creative Commons Attribution-NonCommercial 4.0 International Public License", "features": {"premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "gold_label_log": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "gold_label_prag": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "spec_relation": {"dtype": "string", "id": null, "_type": "Value"}, "item_type": {"dtype": "string", "id": null, "_type": "Value"}, "trigger": {"dtype": "string", "id": null, "_type": "Value"}, "lexemes": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "imppres", "config_name": "implicature_numerals_2_3", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"numerals_2_3": {"name": "numerals_2_3", "num_bytes": 188784, "num_examples": 1200, "dataset_name": "imppres"}}, "download_checksums": {"https://github.com/facebookresearch/Imppres/blob/master/dataset/IMPPRES.zip?raw=true": {"num_bytes": 335088, "checksum": "f1f4ab03aec2248dcfbcb00b80e6099c592751cca2b542b208c6cf46f2926937"}}, "download_size": 335088, "post_processing_size": null, "dataset_size": 188784, "size_in_bytes": 523872}, "implicature_quantifiers": {"description": "Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures.", "citation": "@inproceedings{jeretic-etal-2020-natural,\n title = \"Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}\",\n author = \"Jereti\u000b{c}, Paloma and\n Warstadt, Alex and\n Bhooshan, Suvrat and\n Williams, Adina\",\n booktitle = \"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics\",\n month = jul,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.acl-main.768\",\n doi = \"10.18653/v1/2020.acl-main.768\",\n pages = \"8690--8705\",\n abstract = \"Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.\",\n}\n", "homepage": "https://github.com/facebookresearch/Imppres", "license": "Creative Commons Attribution-NonCommercial 4.0 International Public License", "features": {"premise": {"dtype": "string", "id": null, "_type": "Value"}, "hypothesis": {"dtype": "string", "id": null, "_type": "Value"}, "gold_label_log": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "gold_label_prag": {"num_classes": 3, "names": ["entailment", "neutral", "contradiction"], "names_file": null, "id": null, "_type": "ClassLabel"}, "spec_relation": {"dtype": "string", "id": null, "_type": "Value"}, "item_type": {"dtype": "string", "id": null, "_type": "Value"}, "trigger": {"dtype": "string", "id": null, "_type": "Value"}, "lexemes": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "imppres", "config_name": "implicature_quantifiers", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"quantifiers": {"name": "quantifiers", "num_bytes": 176814, "num_examples": 1200, "dataset_name": "imppres"}}, "download_checksums": {"https://github.com/facebookresearch/Imppres/blob/master/dataset/IMPPRES.zip?raw=true": {"num_bytes": 335088, "checksum": "f1f4ab03aec2248dcfbcb00b80e6099c592751cca2b542b208c6cf46f2926937"}}, "download_size": 335088, "post_processing_size": null, "dataset_size": 176814, "size_in_bytes": 511902}}
 
implicature_connectives/connectives-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80926df2d07638140cd32b8f7e1a4fbc9664c8626e9b4b913b27ab6953e1876b
3
+ size 25478
implicature_gradable_adjective/gradable_adjective-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:665847d1cbc1f66cbca03034de631fed35d8a58aa61637b4b56d4e52cfde9a8f
3
+ size 17337
implicature_gradable_verb/gradable_verb-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd7e8d9a15e7813913b9121e6d5d877673a14b9ee0cc48ae4bcc88f583044596
3
+ size 21504
implicature_modals/modals-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93cb8c075318e9d10278c9c7b04009b4ad0b30779d615bdeee9dfaffa8f86873
3
+ size 21179
implicature_numerals_10_100/numerals_10_100-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3e402e712d586dfd5ee7daf9728df89cc4c3ff016ad9cfcbe7459cb28765a72
3
+ size 22640
implicature_numerals_2_3/numerals_2_3-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d70c965cce5ce06bd62a3573808d0ef0a4084e9828d429a1f33824cd4cfd9a80
3
+ size 22218
implicature_quantifiers/quantifiers-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7530f406d6ee49cf2c171803ccdfc69a97ad12fe15c7f4eab6e381161482e16d
3
+ size 21017
imppres.py DELETED
@@ -1,273 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
- """Over 25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures."""
16
-
17
-
18
- import json
19
- import os
20
-
21
- import datasets
22
-
23
-
24
- # Find for instance the citation on arxiv or on the dataset repo/website
25
- _CITATION = """\
26
- @inproceedings{jeretic-etal-2020-natural,
27
- title = "Are Natural Language Inference Models {IMPPRESsive}? {L}earning {IMPlicature} and {PRESupposition}",
28
- author = "Jereti\v{c}, Paloma and
29
- Warstadt, Alex and
30
- Bhooshan, Suvrat and
31
- Williams, Adina",
32
- booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics",
33
- month = jul,
34
- year = "2020",
35
- address = "Online",
36
- publisher = "Association for Computational Linguistics",
37
- url = "https://www.aclweb.org/anthology/2020.acl-main.768",
38
- doi = "10.18653/v1/2020.acl-main.768",
39
- pages = "8690--8705",
40
- abstract = "Natural language inference (NLI) is an increasingly important task for natural language understanding, which requires one to infer whether a sentence entails another. However, the ability of NLI models to make pragmatic inferences remains understudied. We create an IMPlicature and PRESupposition diagnostic dataset (IMPPRES), consisting of 32K semi-automatically generated sentence pairs illustrating well-studied pragmatic inference types. We use IMPPRES to evaluate whether BERT, InferSent, and BOW NLI models trained on MultiNLI (Williams et al., 2018) learn to make pragmatic inferences. Although MultiNLI appears to contain very few pairs illustrating these inference types, we find that BERT learns to draw pragmatic inferences. It reliably treats scalar implicatures triggered by {``}some{''} as entailments. For some presupposition triggers like {``}only{''}, BERT reliably recognizes the presupposition as an entailment, even when the trigger is embedded under an entailment canceling operator like negation. BOW and InferSent show weaker evidence of pragmatic reasoning. We conclude that NLI training encourages models to learn some, but not all, pragmatic inferences.",
41
- }
42
- """
43
-
44
- # You can copy an official description
45
- _DESCRIPTION = """Over >25k semiautomatically generated sentence pairs illustrating well-studied pragmatic inference types. IMPPRES is an NLI dataset following the format of SNLI (Bowman et al., 2015), MultiNLI (Williams et al., 2018) and XNLI (Conneau et al., 2018), which was created to evaluate how well trained NLI models recognize several classes of presuppositions and scalar implicatures."""
46
- _HOMEPAGE = "https://github.com/facebookresearch/Imppres"
47
- _LICENSE = "Creative Commons Attribution-NonCommercial 4.0 International Public License"
48
-
49
- # The HuggingFace dataset library don't host the datasets but only point to the original files
50
- # This can be an arbitrary nested dict/list of URLs (see below in `_split_generators` method)
51
- _URLs = {"default": "https://github.com/facebookresearch/Imppres/raw/main/dataset/IMPPRES.zip"}
52
-
53
-
54
- class Imppres(datasets.GeneratorBasedBuilder):
55
- """Each sentence type in IMPPRES is generated according to a template that specifies the linear order of the constituents in the sentence. The constituents are sampled from a vocabulary of over 3000 lexical items annotated with grammatical features needed to ensure wellformedness. We semiautomatically generate IMPPRES using a codebase developed by Warstadt et al. (2019a) and significantly expanded for the BLiMP dataset (Warstadt et al., 2019b)."""
56
-
57
- VERSION = datasets.Version("1.1.0")
58
-
59
- # This is an example of a dataset with multiple configurations.
60
- # If you don't want/need to define several sub-sets in your dataset,
61
- # just remove the BUILDER_CONFIG_CLASS and the BUILDER_CONFIGS attributes.
62
-
63
- # If you need to make complex sub-parts in the datasets with configurable options
64
- # You can create your own builder configuration class to store attribute, inheriting from datasets.BuilderConfig
65
- # BUILDER_CONFIG_CLASS = MyBuilderConfig
66
-
67
- # You will be able to load one or the other configurations in the following list with
68
- # data = datasets.load_dataset('my_dataset', 'first_domain')
69
- # data = datasets.load_dataset('my_dataset', 'second_domain')
70
- BUILDER_CONFIGS = [
71
- datasets.BuilderConfig(
72
- name="presupposition_all_n_presupposition",
73
- version=VERSION,
74
- description="Presuppositions are facts that the speaker takes for granted when uttering a sentence.",
75
- ),
76
- datasets.BuilderConfig(
77
- name="presupposition_both_presupposition",
78
- version=VERSION,
79
- description="Presuppositions are facts that the speaker takes for granted when uttering a sentence.",
80
- ),
81
- datasets.BuilderConfig(
82
- name="presupposition_change_of_state",
83
- version=VERSION,
84
- description="Presuppositions are facts that the speaker takes for granted when uttering a sentence.",
85
- ),
86
- datasets.BuilderConfig(
87
- name="presupposition_cleft_existence",
88
- version=VERSION,
89
- description="Presuppositions are facts that the speaker takes for granted when uttering a sentence.",
90
- ),
91
- datasets.BuilderConfig(
92
- name="presupposition_cleft_uniqueness",
93
- version=VERSION,
94
- description="Presuppositions are facts that the speaker takes for granted when uttering a sentence.",
95
- ),
96
- datasets.BuilderConfig(
97
- name="presupposition_only_presupposition",
98
- version=VERSION,
99
- description="Presuppositions are facts that the speaker takes for granted when uttering a sentence.",
100
- ),
101
- datasets.BuilderConfig(
102
- name="presupposition_possessed_definites_existence",
103
- version=VERSION,
104
- description="Presuppositions are facts that the speaker takes for granted when uttering a sentence.",
105
- ),
106
- datasets.BuilderConfig(
107
- name="presupposition_possessed_definites_uniqueness",
108
- version=VERSION,
109
- description="Presuppositions are facts that the speaker takes for granted when uttering a sentence.",
110
- ),
111
- datasets.BuilderConfig(
112
- name="presupposition_question_presupposition",
113
- version=VERSION,
114
- description="Presuppositions are facts that the speaker takes for granted when uttering a sentence.",
115
- ),
116
- datasets.BuilderConfig(
117
- name="implicature_connectives",
118
- version=VERSION,
119
- description="Scalar implicatures are inferences which can be drawn when one member of a memorized lexical scale is uttered.",
120
- ),
121
- datasets.BuilderConfig(
122
- name="implicature_gradable_adjective",
123
- version=VERSION,
124
- description="Scalar implicatures are inferences which can be drawn when one member of a memorized lexical scale is uttered.",
125
- ),
126
- datasets.BuilderConfig(
127
- name="implicature_gradable_verb",
128
- version=VERSION,
129
- description="Scalar implicatures are inferences which can be drawn when one member of a memorized lexical scale is uttered.",
130
- ),
131
- datasets.BuilderConfig(
132
- name="implicature_modals",
133
- version=VERSION,
134
- description="Scalar implicatures are inferences which can be drawn when one member of a memorized lexical scale is uttered.",
135
- ),
136
- datasets.BuilderConfig(
137
- name="implicature_numerals_10_100",
138
- version=VERSION,
139
- description="Scalar implicatures are inferences which can be drawn when one member of a memorized lexical scale is uttered.",
140
- ),
141
- datasets.BuilderConfig(
142
- name="implicature_numerals_2_3",
143
- version=VERSION,
144
- description="Scalar implicatures are inferences which can be drawn when one member of a memorized lexical scale is uttered.",
145
- ),
146
- datasets.BuilderConfig(
147
- name="implicature_quantifiers",
148
- version=VERSION,
149
- description="Scalar implicatures are inferences which can be drawn when one member of a memorized lexical scale is uttered.",
150
- ),
151
- ]
152
-
153
- def _info(self):
154
- if (
155
- "presupposition" in self.config.name
156
- ): # This is the name of the configuration selected in BUILDER_CONFIGS above
157
- features = datasets.Features(
158
- {
159
- "premise": datasets.Value("string"),
160
- "hypothesis": datasets.Value("string"),
161
- "trigger": datasets.Value("string"),
162
- "trigger1": datasets.Value("string"),
163
- "trigger2": datasets.Value("string"),
164
- "presupposition": datasets.Value("string"),
165
- "gold_label": datasets.ClassLabel(names=["entailment", "neutral", "contradiction"]),
166
- "UID": datasets.Value("string"),
167
- "pairID": datasets.Value("string"),
168
- "paradigmID": datasets.Value("int16")
169
- # These are the features of your dataset like images, labels ...
170
- }
171
- )
172
- else: # This is an example to show how to have different features for "first_domain" and "second_domain"
173
- features = datasets.Features(
174
- {
175
- "premise": datasets.Value("string"),
176
- "hypothesis": datasets.Value("string"),
177
- "gold_label_log": datasets.ClassLabel(names=["entailment", "neutral", "contradiction"]),
178
- "gold_label_prag": datasets.ClassLabel(names=["entailment", "neutral", "contradiction"]),
179
- "spec_relation": datasets.Value("string"),
180
- "item_type": datasets.Value("string"),
181
- "trigger": datasets.Value("string"),
182
- "lexemes": datasets.Value("string"),
183
- # These are the features of your dataset like images, labels ...
184
- }
185
- )
186
- return datasets.DatasetInfo(
187
- # This is the description that will appear on the datasets page.
188
- description=_DESCRIPTION,
189
- # This defines the different columns of the dataset and their types
190
- features=features, # Here we define them above because they are different between the two configurations
191
- # If there's a common (input, target) tuple from the features,
192
- # specify them here. They'll be used if as_supervised=True in
193
- # builder.as_dataset.
194
- supervised_keys=None,
195
- # Homepage of the dataset for documentation
196
- homepage=_HOMEPAGE,
197
- # License for the dataset if available
198
- license=_LICENSE,
199
- # Citation for the dataset
200
- citation=_CITATION,
201
- )
202
-
203
- def _split_generators(self, dl_manager):
204
- """Returns SplitGenerators."""
205
- # TODO: This method is tasked with downloading/extracting the data and defining the splits depending on the configuration
206
- # If several configurations are possible (listed in BUILDER_CONFIGS), the configuration selected by the user is in self.config.name
207
-
208
- # dl_manager is a datasets.download.DownloadManager that can be used to download and extract URLs
209
- # It can accept any type or nested list/dict and will give back the same structure with the url replaced with path to local files.
210
- # By default the archives will be extracted and a path to a cached folder where they are extracted is returned instead of the archive
211
- my_urls = _URLs["default"]
212
- base_config = self.config.name.split("_")[0]
213
- secondary_config = self.config.name.split(base_config + "_")[1]
214
- data_dir = os.path.join(dl_manager.download_and_extract(my_urls), "IMPPRES", base_config)
215
- return [
216
- datasets.SplitGenerator(
217
- name=secondary_config,
218
- # These kwargs will be passed to _generate_examples
219
- gen_kwargs={
220
- "filepath": os.path.join(data_dir, secondary_config + ".jsonl"),
221
- "split": "test",
222
- },
223
- )
224
- ]
225
-
226
- def _generate_examples(self, filepath, split):
227
- """Yields examples."""
228
- # TODO: This method will receive as arguments the `gen_kwargs` defined in the previous `_split_generators` method.
229
- # It is in charge of opening the given file and yielding (key, example) tuples from the dataset
230
- # The key is not important, it's more here for legacy reason (legacy from tfds)
231
-
232
- with open(filepath, encoding="utf-8") as f:
233
- for id_, row in enumerate(f):
234
- data = json.loads(row)
235
- if "presupposition" in self.config.name:
236
- if "trigger1" not in list(data.keys()):
237
- yield id_, {
238
- "premise": data["sentence1"],
239
- "hypothesis": data["sentence2"],
240
- "trigger": data["trigger"],
241
- "trigger1": "Not_In_Example",
242
- "trigger2": "Not_In_Example",
243
- "presupposition": data["presupposition"],
244
- "gold_label": data["gold_label"],
245
- "UID": data["UID"],
246
- "pairID": data["pairID"],
247
- "paradigmID": data["paradigmID"],
248
- }
249
- else:
250
- yield id_, {
251
- "premise": data["sentence1"],
252
- "hypothesis": data["sentence2"],
253
- "trigger": "Not_In_Example",
254
- "trigger1": data["trigger1"],
255
- "trigger2": data["trigger2"],
256
- "presupposition": "Not_In_Example",
257
- "gold_label": data["gold_label"],
258
- "UID": data["UID"],
259
- "pairID": data["pairID"],
260
- "paradigmID": data["paradigmID"],
261
- }
262
-
263
- else:
264
- yield id_, {
265
- "premise": data["sentence1"],
266
- "hypothesis": data["sentence2"],
267
- "gold_label_log": data["gold_label_log"],
268
- "gold_label_prag": data["gold_label_prag"],
269
- "spec_relation": data["spec_relation"],
270
- "item_type": data["item_type"],
271
- "trigger": data["trigger"],
272
- "lexemes": data["lexemes"],
273
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
presupposition_all_n_presupposition/all_n_presupposition-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b23643269b3e2e5b07dd419aa97ea5d48d7ab9aa1df90d802caa8eff2f8e025
3
+ size 43038
presupposition_both_presupposition/both_presupposition-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2556f0b3980aa715d292f7d2513e8473d7d63e8631064526a706259f067d2b86
3
+ size 41142
presupposition_change_of_state/change_of_state-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68e3a61af82f43b12004c20721fd77b41f6d6aef0493675d1f28f0cf84d2f04d
3
+ size 35814
presupposition_cleft_existence/cleft_existence-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cfc0beca3d51dd94e975e6c97863e2c563f8cbe8ba65f90fc03f9a7326ffbcf
3
+ size 37597
presupposition_cleft_uniqueness/cleft_uniqueness-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cfffe6d01c4de4532a2d9f6bd7a279266d1039aafc2072eb0ed6e5ae4eee86b
3
+ size 38279
presupposition_only_presupposition/only_presupposition-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ec7e3112aba19e2285878aa582e275374276d305a27df6b220a1d1f455047ea
3
+ size 38126
presupposition_possessed_definites_existence/possessed_definites_existence-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:567496b59e5a13a2492d5edf9967c00a8e6c8ad51c12b3ab3929123aeaa5875e
3
+ size 38712
presupposition_possessed_definites_uniqueness/possessed_definites_uniqueness-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6b2ee13a4e2605d1aaf33dd282a06c84f99b7a9261ce297871d48e4b9f48a1e
3
+ size 42068
presupposition_question_presupposition/question_presupposition-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c25b1f61067defba8f886fbefc87d615f0e6b3d7b3ccaa7cf7fa68fc6700e663
3
+ size 41247