parquet-converter commited on
Commit
e5c4703
1 Parent(s): a7d837b

Update parquet files

Browse files
README.md DELETED
@@ -1,2105 +0,0 @@
1
- ---
2
- annotations_creators:
3
- - crowdsourced
4
- - found
5
- language_creators:
6
- - crowdsourced
7
- - found
8
- - machine-generated
9
- language:
10
- - cs
11
- - de
12
- - en
13
- - es
14
- - ru
15
- - tr
16
- - vi
17
- license:
18
- - other
19
- multilinguality:
20
- - monolingual
21
- - multilingual
22
- size_categories:
23
- - 100K<n<1M
24
- - 10K<n<100K
25
- - 1K<n<10K
26
- source_datasets:
27
- - extended|other-vision-datasets
28
- - original
29
- task_categories:
30
- - fill-mask
31
- - summarization
32
- - table-to-text
33
- - tabular-to-text
34
- - text-generation
35
- - text2text-generation
36
- task_ids:
37
- - dialogue-modeling
38
- - rdf-to-text
39
- - news-articles-summarization
40
- - text-simplification
41
- paperswithcode_id: gem
42
- pretty_name: GEM
43
- configs:
44
- - common_gen
45
- - cs_restaurants
46
- - dart
47
- - e2e_nlg
48
- - mlsum_de
49
- - mlsum_es
50
- - schema_guided_dialog
51
- - totto
52
- - web_nlg_en
53
- - web_nlg_ru
54
- - wiki_auto_asset_turk
55
- - wiki_lingua_es_en
56
- - wiki_lingua_ru_en
57
- - wiki_lingua_tr_en
58
- - wiki_lingua_vi_en
59
- - xsum
60
- tags:
61
- - intent-to-text
62
- - meaning-representation-to-text
63
- - concepts-to-text
64
- dataset_info:
65
- - config_name: mlsum_de
66
- features:
67
- - name: gem_id
68
- dtype: string
69
- - name: gem_parent_id
70
- dtype: string
71
- - name: text
72
- dtype: string
73
- - name: topic
74
- dtype: string
75
- - name: url
76
- dtype: string
77
- - name: title
78
- dtype: string
79
- - name: date
80
- dtype: string
81
- - name: target
82
- dtype: string
83
- - name: references
84
- list: string
85
- splits:
86
- - name: train
87
- num_bytes: 858060337
88
- num_examples: 220748
89
- - name: validation
90
- num_bytes: 49712791
91
- num_examples: 11392
92
- - name: test
93
- num_bytes: 49146354
94
- num_examples: 10695
95
- - name: challenge_train_sample
96
- num_bytes: 1894220
97
- num_examples: 500
98
- - name: challenge_validation_sample
99
- num_bytes: 2202723
100
- num_examples: 500
101
- - name: challenge_test_covid
102
- num_bytes: 19771285
103
- num_examples: 5058
104
- download_size: 362783528
105
- dataset_size: 980787710
106
- - config_name: mlsum_es
107
- features:
108
- - name: gem_id
109
- dtype: string
110
- - name: gem_parent_id
111
- dtype: string
112
- - name: text
113
- dtype: string
114
- - name: topic
115
- dtype: string
116
- - name: url
117
- dtype: string
118
- - name: title
119
- dtype: string
120
- - name: date
121
- dtype: string
122
- - name: target
123
- dtype: string
124
- - name: references
125
- list: string
126
- splits:
127
- - name: train
128
- num_bytes: 1211240956
129
- num_examples: 259888
130
- - name: validation
131
- num_bytes: 51611723
132
- num_examples: 9977
133
- - name: test
134
- num_bytes: 72117564
135
- num_examples: 13366
136
- - name: challenge_train_sample
137
- num_bytes: 2366443
138
- num_examples: 500
139
- - name: challenge_validation_sample
140
- num_bytes: 2658596
141
- num_examples: 500
142
- - name: challenge_test_covid
143
- num_bytes: 13576624
144
- num_examples: 1938
145
- download_size: 525621426
146
- dataset_size: 1353571906
147
- - config_name: wiki_lingua_es_en_v0
148
- features:
149
- - name: gem_id
150
- dtype: string
151
- - name: gem_parent_id
152
- dtype: string
153
- - name: source
154
- dtype: string
155
- - name: target
156
- dtype: string
157
- - name: references
158
- list: string
159
- splits:
160
- - name: train
161
- num_bytes: 215665468
162
- num_examples: 79515
163
- - name: validation
164
- num_bytes: 25891008
165
- num_examples: 8835
166
- - name: test
167
- num_bytes: 50195305
168
- num_examples: 19797
169
- download_size: 169406387
170
- dataset_size: 291751781
171
- - config_name: wiki_lingua_ru_en_v0
172
- features:
173
- - name: gem_id
174
- dtype: string
175
- - name: gem_parent_id
176
- dtype: string
177
- - name: source
178
- dtype: string
179
- - name: target
180
- dtype: string
181
- - name: references
182
- list: string
183
- splits:
184
- - name: train
185
- num_bytes: 159631205
186
- num_examples: 36898
187
- - name: validation
188
- num_bytes: 18626973
189
- num_examples: 4100
190
- - name: test
191
- num_bytes: 34865311
192
- num_examples: 9094
193
- download_size: 169406387
194
- dataset_size: 213123489
195
- - config_name: wiki_lingua_tr_en_v0
196
- features:
197
- - name: gem_id
198
- dtype: string
199
- - name: gem_parent_id
200
- dtype: string
201
- - name: source
202
- dtype: string
203
- - name: target
204
- dtype: string
205
- - name: references
206
- list: string
207
- splits:
208
- - name: train
209
- num_bytes: 7689845
210
- num_examples: 3193
211
- - name: validation
212
- num_bytes: 942122
213
- num_examples: 355
214
- - name: test
215
- num_bytes: 1875110
216
- num_examples: 808
217
- download_size: 169406387
218
- dataset_size: 10507077
219
- - config_name: wiki_lingua_vi_en_v0
220
- features:
221
- - name: gem_id
222
- dtype: string
223
- - name: gem_parent_id
224
- dtype: string
225
- - name: source
226
- dtype: string
227
- - name: target
228
- dtype: string
229
- - name: references
230
- list: string
231
- splits:
232
- - name: train
233
- num_bytes: 31599580
234
- num_examples: 9206
235
- - name: validation
236
- num_bytes: 3618660
237
- num_examples: 1023
238
- - name: test
239
- num_bytes: 6267359
240
- num_examples: 2167
241
- download_size: 169406387
242
- dataset_size: 41485599
243
- - config_name: wiki_lingua_arabic_ar
244
- features:
245
- - name: gem_id
246
- dtype: string
247
- - name: gem_parent_id
248
- dtype: string
249
- - name: source_aligned
250
- dtype:
251
- translation:
252
- languages:
253
- - ar
254
- - en
255
- - name: target_aligned
256
- dtype:
257
- translation:
258
- languages:
259
- - ar
260
- - en
261
- - name: source
262
- dtype: string
263
- - name: target
264
- dtype: string
265
- - name: references
266
- list: string
267
- splits:
268
- - name: train
269
- num_bytes: 208106335
270
- num_examples: 20441
271
- - name: validation
272
- num_bytes: 31126187
273
- num_examples: 2919
274
- - name: test
275
- num_bytes: 60915220
276
- num_examples: 5841
277
- download_size: 58984103
278
- dataset_size: 300147742
279
- - config_name: wiki_lingua_chinese_zh
280
- features:
281
- - name: gem_id
282
- dtype: string
283
- - name: gem_parent_id
284
- dtype: string
285
- - name: source_aligned
286
- dtype:
287
- translation:
288
- languages:
289
- - zh
290
- - en
291
- - name: target_aligned
292
- dtype:
293
- translation:
294
- languages:
295
- - zh
296
- - en
297
- - name: source
298
- dtype: string
299
- - name: target
300
- dtype: string
301
- - name: references
302
- list: string
303
- splits:
304
- - name: train
305
- num_bytes: 86130302
306
- num_examples: 13211
307
- - name: validation
308
- num_bytes: 13060918
309
- num_examples: 1886
310
- - name: test
311
- num_bytes: 25310021
312
- num_examples: 3775
313
- download_size: 32899156
314
- dataset_size: 124501241
315
- - config_name: wiki_lingua_czech_cs
316
- features:
317
- - name: gem_id
318
- dtype: string
319
- - name: gem_parent_id
320
- dtype: string
321
- - name: source_aligned
322
- dtype:
323
- translation:
324
- languages:
325
- - cs
326
- - en
327
- - name: target_aligned
328
- dtype:
329
- translation:
330
- languages:
331
- - cs
332
- - en
333
- - name: source
334
- dtype: string
335
- - name: target
336
- dtype: string
337
- - name: references
338
- list: string
339
- splits:
340
- - name: train
341
- num_bytes: 41107318
342
- num_examples: 5033
343
- - name: validation
344
- num_bytes: 6305328
345
- num_examples: 718
346
- - name: test
347
- num_bytes: 12124770
348
- num_examples: 1438
349
- download_size: 14515534
350
- dataset_size: 59537416
351
- - config_name: wiki_lingua_dutch_nl
352
- features:
353
- - name: gem_id
354
- dtype: string
355
- - name: gem_parent_id
356
- dtype: string
357
- - name: source_aligned
358
- dtype:
359
- translation:
360
- languages:
361
- - nl
362
- - en
363
- - name: target_aligned
364
- dtype:
365
- translation:
366
- languages:
367
- - nl
368
- - en
369
- - name: source
370
- dtype: string
371
- - name: target
372
- dtype: string
373
- - name: references
374
- list: string
375
- splits:
376
- - name: train
377
- num_bytes: 169067454
378
- num_examples: 21866
379
- - name: validation
380
- num_bytes: 25521003
381
- num_examples: 3123
382
- - name: test
383
- num_bytes: 49165151
384
- num_examples: 6248
385
- download_size: 56492150
386
- dataset_size: 243753608
387
- - config_name: wiki_lingua_english_en
388
- features:
389
- - name: gem_id
390
- dtype: string
391
- - name: gem_parent_id
392
- dtype: string
393
- - name: source_aligned
394
- dtype:
395
- translation:
396
- languages:
397
- - en
398
- - en
399
- - name: target_aligned
400
- dtype:
401
- translation:
402
- languages:
403
- - en
404
- - en
405
- - name: source
406
- dtype: string
407
- - name: target
408
- dtype: string
409
- - name: references
410
- list: string
411
- splits:
412
- - name: train
413
- num_bytes: 464171624
414
- num_examples: 99020
415
- - name: validation
416
- num_bytes: 67652281
417
- num_examples: 13823
418
- - name: test
419
- num_bytes: 138944243
420
- num_examples: 28614
421
- download_size: 118031903
422
- dataset_size: 670768148
423
- - config_name: wiki_lingua_french_fr
424
- features:
425
- - name: gem_id
426
- dtype: string
427
- - name: gem_parent_id
428
- dtype: string
429
- - name: source_aligned
430
- dtype:
431
- translation:
432
- languages:
433
- - fr
434
- - en
435
- - name: target_aligned
436
- dtype:
437
- translation:
438
- languages:
439
- - fr
440
- - en
441
- - name: source
442
- dtype: string
443
- - name: target
444
- dtype: string
445
- - name: references
446
- list: string
447
- splits:
448
- - name: train
449
- num_bytes: 372039357
450
- num_examples: 44556
451
- - name: validation
452
- num_bytes: 54992250
453
- num_examples: 6364
454
- - name: test
455
- num_bytes: 108831855
456
- num_examples: 12731
457
- download_size: 118758047
458
- dataset_size: 535863462
459
- - config_name: wiki_lingua_german_de
460
- features:
461
- - name: gem_id
462
- dtype: string
463
- - name: gem_parent_id
464
- dtype: string
465
- - name: source_aligned
466
- dtype:
467
- translation:
468
- languages:
469
- - de
470
- - en
471
- - name: target_aligned
472
- dtype:
473
- translation:
474
- languages:
475
- - de
476
- - en
477
- - name: source
478
- dtype: string
479
- - name: target
480
- dtype: string
481
- - name: references
482
- list: string
483
- splits:
484
- - name: train
485
- num_bytes: 322276536
486
- num_examples: 40839
487
- - name: validation
488
- num_bytes: 47631883
489
- num_examples: 5833
490
- - name: test
491
- num_bytes: 93715331
492
- num_examples: 11669
493
- download_size: 107638803
494
- dataset_size: 463623750
495
- - config_name: wiki_lingua_hindi_hi
496
- features:
497
- - name: gem_id
498
- dtype: string
499
- - name: gem_parent_id
500
- dtype: string
501
- - name: source_aligned
502
- dtype:
503
- translation:
504
- languages:
505
- - hi
506
- - en
507
- - name: target_aligned
508
- dtype:
509
- translation:
510
- languages:
511
- - hi
512
- - en
513
- - name: source
514
- dtype: string
515
- - name: target
516
- dtype: string
517
- - name: references
518
- list: string
519
- splits:
520
- - name: train
521
- num_bytes: 99672133
522
- num_examples: 6942
523
- - name: validation
524
- num_bytes: 14706378
525
- num_examples: 991
526
- - name: test
527
- num_bytes: 28543048
528
- num_examples: 1984
529
- download_size: 21042040
530
- dataset_size: 142921559
531
- - config_name: wiki_lingua_indonesian_id
532
- features:
533
- - name: gem_id
534
- dtype: string
535
- - name: gem_parent_id
536
- dtype: string
537
- - name: source_aligned
538
- dtype:
539
- translation:
540
- languages:
541
- - id
542
- - en
543
- - name: target_aligned
544
- dtype:
545
- translation:
546
- languages:
547
- - id
548
- - en
549
- - name: source
550
- dtype: string
551
- - name: target
552
- dtype: string
553
- - name: references
554
- list: string
555
- splits:
556
- - name: train
557
- num_bytes: 263974954
558
- num_examples: 33237
559
- - name: validation
560
- num_bytes: 39297987
561
- num_examples: 4747
562
- - name: test
563
- num_bytes: 76567819
564
- num_examples: 9497
565
- download_size: 83968162
566
- dataset_size: 379840760
567
- - config_name: wiki_lingua_italian_it
568
- features:
569
- - name: gem_id
570
- dtype: string
571
- - name: gem_parent_id
572
- dtype: string
573
- - name: source_aligned
574
- dtype:
575
- translation:
576
- languages:
577
- - it
578
- - en
579
- - name: target_aligned
580
- dtype:
581
- translation:
582
- languages:
583
- - it
584
- - en
585
- - name: source
586
- dtype: string
587
- - name: target
588
- dtype: string
589
- - name: references
590
- list: string
591
- splits:
592
- - name: train
593
- num_bytes: 267090482
594
- num_examples: 35661
595
- - name: validation
596
- num_bytes: 39227425
597
- num_examples: 5093
598
- - name: test
599
- num_bytes: 76840429
600
- num_examples: 10189
601
- download_size: 88921209
602
- dataset_size: 383158336
603
- - config_name: wiki_lingua_japanese_ja
604
- features:
605
- - name: gem_id
606
- dtype: string
607
- - name: gem_parent_id
608
- dtype: string
609
- - name: source_aligned
610
- dtype:
611
- translation:
612
- languages:
613
- - ja
614
- - en
615
- - name: target_aligned
616
- dtype:
617
- translation:
618
- languages:
619
- - ja
620
- - en
621
- - name: source
622
- dtype: string
623
- - name: target
624
- dtype: string
625
- - name: references
626
- list: string
627
- splits:
628
- - name: train
629
- num_bytes: 73871019
630
- num_examples: 8853
631
- - name: validation
632
- num_bytes: 10807006
633
- num_examples: 1264
634
- - name: test
635
- num_bytes: 21175951
636
- num_examples: 2530
637
- download_size: 22803299
638
- dataset_size: 105853976
639
- - config_name: wiki_lingua_korean_ko
640
- features:
641
- - name: gem_id
642
- dtype: string
643
- - name: gem_parent_id
644
- dtype: string
645
- - name: source_aligned
646
- dtype:
647
- translation:
648
- languages:
649
- - ko
650
- - en
651
- - name: target_aligned
652
- dtype:
653
- translation:
654
- languages:
655
- - ko
656
- - en
657
- - name: source
658
- dtype: string
659
- - name: target
660
- dtype: string
661
- - name: references
662
- list: string
663
- splits:
664
- - name: train
665
- num_bytes: 73106687
666
- num_examples: 8524
667
- - name: validation
668
- num_bytes: 10788276
669
- num_examples: 1216
670
- - name: test
671
- num_bytes: 21172641
672
- num_examples: 2436
673
- download_size: 23336917
674
- dataset_size: 105067604
675
- - config_name: wiki_lingua_portuguese_pt
676
- features:
677
- - name: gem_id
678
- dtype: string
679
- - name: gem_parent_id
680
- dtype: string
681
- - name: source_aligned
682
- dtype:
683
- translation:
684
- languages:
685
- - pt
686
- - en
687
- - name: target_aligned
688
- dtype:
689
- translation:
690
- languages:
691
- - pt
692
- - en
693
- - name: source
694
- dtype: string
695
- - name: target
696
- dtype: string
697
- - name: references
698
- list: string
699
- splits:
700
- - name: train
701
- num_bytes: 405546332
702
- num_examples: 57159
703
- - name: validation
704
- num_bytes: 59729210
705
- num_examples: 8165
706
- - name: test
707
- num_bytes: 117775356
708
- num_examples: 16331
709
- download_size: 137542940
710
- dataset_size: 583050898
711
- - config_name: wiki_lingua_russian_ru
712
- features:
713
- - name: gem_id
714
- dtype: string
715
- - name: gem_parent_id
716
- dtype: string
717
- - name: source_aligned
718
- dtype:
719
- translation:
720
- languages:
721
- - ru
722
- - en
723
- - name: target_aligned
724
- dtype:
725
- translation:
726
- languages:
727
- - ru
728
- - en
729
- - name: source
730
- dtype: string
731
- - name: target
732
- dtype: string
733
- - name: references
734
- list: string
735
- splits:
736
- - name: train
737
- num_bytes: 406299624
738
- num_examples: 37028
739
- - name: validation
740
- num_bytes: 59651340
741
- num_examples: 5288
742
- - name: test
743
- num_bytes: 116330937
744
- num_examples: 10580
745
- download_size: 106281321
746
- dataset_size: 582281901
747
- - config_name: wiki_lingua_spanish_es
748
- features:
749
- - name: gem_id
750
- dtype: string
751
- - name: gem_parent_id
752
- dtype: string
753
- - name: source_aligned
754
- dtype:
755
- translation:
756
- languages:
757
- - es
758
- - en
759
- - name: target_aligned
760
- dtype:
761
- translation:
762
- languages:
763
- - es
764
- - en
765
- - name: source
766
- dtype: string
767
- - name: target
768
- dtype: string
769
- - name: references
770
- list: string
771
- splits:
772
- - name: train
773
- num_bytes: 604276564
774
- num_examples: 79212
775
- - name: validation
776
- num_bytes: 88677656
777
- num_examples: 11316
778
- - name: test
779
- num_bytes: 177096288
780
- num_examples: 22632
781
- download_size: 198247534
782
- dataset_size: 870050508
783
- - config_name: wiki_lingua_thai_th
784
- features:
785
- - name: gem_id
786
- dtype: string
787
- - name: gem_parent_id
788
- dtype: string
789
- - name: source_aligned
790
- dtype:
791
- translation:
792
- languages:
793
- - th
794
- - en
795
- - name: target_aligned
796
- dtype:
797
- translation:
798
- languages:
799
- - th
800
- - en
801
- - name: source
802
- dtype: string
803
- - name: target
804
- dtype: string
805
- - name: references
806
- list: string
807
- splits:
808
- - name: train
809
- num_bytes: 139287649
810
- num_examples: 10325
811
- - name: validation
812
- num_bytes: 21097845
813
- num_examples: 1475
814
- - name: test
815
- num_bytes: 40049968
816
- num_examples: 2950
817
- download_size: 29988180
818
- dataset_size: 200435462
819
- - config_name: wiki_lingua_turkish_tr
820
- features:
821
- - name: gem_id
822
- dtype: string
823
- - name: gem_parent_id
824
- dtype: string
825
- - name: source_aligned
826
- dtype:
827
- translation:
828
- languages:
829
- - tr
830
- - en
831
- - name: target_aligned
832
- dtype:
833
- translation:
834
- languages:
835
- - tr
836
- - en
837
- - name: source
838
- dtype: string
839
- - name: target
840
- dtype: string
841
- - name: references
842
- list: string
843
- splits:
844
- - name: train
845
- num_bytes: 21987247
846
- num_examples: 3148
847
- - name: validation
848
- num_bytes: 3229714
849
- num_examples: 449
850
- - name: test
851
- num_bytes: 6197850
852
- num_examples: 900
853
- download_size: 7055820
854
- dataset_size: 31414811
855
- - config_name: wiki_lingua_vietnamese_vi
856
- features:
857
- - name: gem_id
858
- dtype: string
859
- - name: gem_parent_id
860
- dtype: string
861
- - name: source_aligned
862
- dtype:
863
- translation:
864
- languages:
865
- - vi
866
- - en
867
- - name: target_aligned
868
- dtype:
869
- translation:
870
- languages:
871
- - vi
872
- - en
873
- - name: source
874
- dtype: string
875
- - name: target
876
- dtype: string
877
- - name: references
878
- list: string
879
- splits:
880
- - name: train
881
- num_bytes: 128025008
882
- num_examples: 13707
883
- - name: validation
884
- num_bytes: 19414734
885
- num_examples: 1957
886
- - name: test
887
- num_bytes: 37430208
888
- num_examples: 3917
889
- download_size: 38035490
890
- dataset_size: 184869950
891
- - config_name: xsum
892
- features:
893
- - name: gem_id
894
- dtype: string
895
- - name: gem_parent_id
896
- dtype: string
897
- - name: xsum_id
898
- dtype: string
899
- - name: document
900
- dtype: string
901
- - name: target
902
- dtype: string
903
- - name: references
904
- list: string
905
- splits:
906
- - name: train
907
- num_bytes: 66299136
908
- num_examples: 23206
909
- - name: validation
910
- num_bytes: 2270306
911
- num_examples: 1117
912
- - name: test
913
- num_bytes: 2598509
914
- num_examples: 1166
915
- - name: challenge_train_sample
916
- num_bytes: 1429145
917
- num_examples: 500
918
- - name: challenge_validation_sample
919
- num_bytes: 1012689
920
- num_examples: 500
921
- - name: challenge_test_backtranslation
922
- num_bytes: 1262047
923
- num_examples: 500
924
- - name: challenge_test_bfp_02
925
- num_bytes: 1090364
926
- num_examples: 500
927
- - name: challenge_test_bfp_05
928
- num_bytes: 1078076
929
- num_examples: 500
930
- - name: challenge_test_nopunc
931
- num_bytes: 1127796
932
- num_examples: 500
933
- - name: challenge_test_covid
934
- num_bytes: 1867180
935
- num_examples: 401
936
- download_size: 258277147
937
- dataset_size: 80035248
938
- - config_name: common_gen
939
- features:
940
- - name: gem_id
941
- dtype: string
942
- - name: gem_parent_id
943
- dtype: string
944
- - name: concept_set_id
945
- dtype: int32
946
- - name: concepts
947
- list: string
948
- - name: target
949
- dtype: string
950
- - name: references
951
- list: string
952
- splits:
953
- - name: train
954
- num_bytes: 10475926
955
- num_examples: 67389
956
- - name: validation
957
- num_bytes: 405872
958
- num_examples: 993
959
- - name: test
960
- num_bytes: 153170
961
- num_examples: 1497
962
- - name: challenge_train_sample
963
- num_bytes: 85413
964
- num_examples: 500
965
- - name: challenge_validation_sample
966
- num_bytes: 215192
967
- num_examples: 500
968
- - name: challenge_test_scramble
969
- num_bytes: 60411
970
- num_examples: 500
971
- download_size: 1933517
972
- dataset_size: 11395984
973
- - config_name: cs_restaurants
974
- features:
975
- - name: gem_id
976
- dtype: string
977
- - name: gem_parent_id
978
- dtype: string
979
- - name: dialog_act
980
- dtype: string
981
- - name: dialog_act_delexicalized
982
- dtype: string
983
- - name: target_delexicalized
984
- dtype: string
985
- - name: target
986
- dtype: string
987
- - name: references
988
- list: string
989
- splits:
990
- - name: train
991
- num_bytes: 873145
992
- num_examples: 3569
993
- - name: validation
994
- num_bytes: 288222
995
- num_examples: 781
996
- - name: test
997
- num_bytes: 295696
998
- num_examples: 842
999
- - name: challenge_train_sample
1000
- num_bytes: 127869
1001
- num_examples: 500
1002
- - name: challenge_validation_sample
1003
- num_bytes: 193239
1004
- num_examples: 500
1005
- - name: challenge_test_scramble
1006
- num_bytes: 185574
1007
- num_examples: 500
1008
- download_size: 1531111
1009
- dataset_size: 1963745
1010
- - config_name: dart
1011
- features:
1012
- - name: gem_id
1013
- dtype: string
1014
- - name: gem_parent_id
1015
- dtype: string
1016
- - name: dart_id
1017
- dtype: int32
1018
- - name: tripleset
1019
- list:
1020
- list: string
1021
- - name: subtree_was_extended
1022
- dtype: bool
1023
- - name: target_sources
1024
- list: string
1025
- - name: target
1026
- dtype: string
1027
- - name: references
1028
- list: string
1029
- splits:
1030
- - name: train
1031
- num_bytes: 23047610
1032
- num_examples: 62659
1033
- - name: validation
1034
- num_bytes: 1934054
1035
- num_examples: 2768
1036
- - name: test
1037
- num_bytes: 3476953
1038
- num_examples: 5097
1039
- download_size: 29939366
1040
- dataset_size: 28458617
1041
- - config_name: e2e_nlg
1042
- features:
1043
- - name: gem_id
1044
- dtype: string
1045
- - name: gem_parent_id
1046
- dtype: string
1047
- - name: meaning_representation
1048
- dtype: string
1049
- - name: target
1050
- dtype: string
1051
- - name: references
1052
- list: string
1053
- splits:
1054
- - name: train
1055
- num_bytes: 9129030
1056
- num_examples: 33525
1057
- - name: validation
1058
- num_bytes: 1856097
1059
- num_examples: 4299
1060
- - name: test
1061
- num_bytes: 2133695
1062
- num_examples: 4693
1063
- - name: challenge_train_sample
1064
- num_bytes: 145319
1065
- num_examples: 500
1066
- - name: challenge_validation_sample
1067
- num_bytes: 226525
1068
- num_examples: 500
1069
- - name: challenge_test_scramble
1070
- num_bytes: 236199
1071
- num_examples: 500
1072
- download_size: 14668048
1073
- dataset_size: 13726865
1074
- - config_name: totto
1075
- features:
1076
- - name: gem_id
1077
- dtype: string
1078
- - name: gem_parent_id
1079
- dtype: string
1080
- - name: totto_id
1081
- dtype: int32
1082
- - name: table_page_title
1083
- dtype: string
1084
- - name: table_webpage_url
1085
- dtype: string
1086
- - name: table_section_title
1087
- dtype: string
1088
- - name: table_section_text
1089
- dtype: string
1090
- - name: table
1091
- list:
1092
- list:
1093
- - name: column_span
1094
- dtype: int32
1095
- - name: is_header
1096
- dtype: bool
1097
- - name: row_span
1098
- dtype: int32
1099
- - name: value
1100
- dtype: string
1101
- - name: highlighted_cells
1102
- list:
1103
- list: int32
1104
- - name: example_id
1105
- dtype: string
1106
- - name: sentence_annotations
1107
- list:
1108
- - name: original_sentence
1109
- dtype: string
1110
- - name: sentence_after_deletion
1111
- dtype: string
1112
- - name: sentence_after_ambiguity
1113
- dtype: string
1114
- - name: final_sentence
1115
- dtype: string
1116
- - name: overlap_subset
1117
- dtype: string
1118
- - name: target
1119
- dtype: string
1120
- - name: references
1121
- list: string
1122
- splits:
1123
- - name: train
1124
- num_bytes: 676032144
1125
- num_examples: 121153
1126
- - name: validation
1127
- num_bytes: 50736204
1128
- num_examples: 7700
1129
- - name: test
1130
- num_bytes: 41330062
1131
- num_examples: 7700
1132
- - name: challenge_train_sample
1133
- num_bytes: 2283076
1134
- num_examples: 500
1135
- - name: challenge_validation_sample
1136
- num_bytes: 3398639
1137
- num_examples: 500
1138
- - name: challenge_test_scramble
1139
- num_bytes: 2638966
1140
- num_examples: 500
1141
- download_size: 189534609
1142
- dataset_size: 776419091
1143
- - config_name: web_nlg_en
1144
- features:
1145
- - name: gem_id
1146
- dtype: string
1147
- - name: gem_parent_id
1148
- dtype: string
1149
- - name: input
1150
- list: string
1151
- - name: target
1152
- dtype: string
1153
- - name: references
1154
- list: string
1155
- - name: category
1156
- dtype: string
1157
- - name: webnlg_id
1158
- dtype: string
1159
- splits:
1160
- - name: train
1161
- num_bytes: 13067615
1162
- num_examples: 35426
1163
- - name: validation
1164
- num_bytes: 1153995
1165
- num_examples: 1667
1166
- - name: test
1167
- num_bytes: 1403601
1168
- num_examples: 1779
1169
- - name: challenge_train_sample
1170
- num_bytes: 193198
1171
- num_examples: 502
1172
- - name: challenge_validation_sample
1173
- num_bytes: 359868
1174
- num_examples: 499
1175
- - name: challenge_test_scramble
1176
- num_bytes: 402407
1177
- num_examples: 500
1178
- - name: challenge_test_numbers
1179
- num_bytes: 409213
1180
- num_examples: 500
1181
- download_size: 13181969
1182
- dataset_size: 16989897
1183
- - config_name: web_nlg_ru
1184
- features:
1185
- - name: gem_id
1186
- dtype: string
1187
- - name: gem_parent_id
1188
- dtype: string
1189
- - name: input
1190
- list: string
1191
- - name: target
1192
- dtype: string
1193
- - name: references
1194
- list: string
1195
- - name: category
1196
- dtype: string
1197
- - name: webnlg_id
1198
- dtype: string
1199
- splits:
1200
- - name: train
1201
- num_bytes: 6888009
1202
- num_examples: 14630
1203
- - name: validation
1204
- num_bytes: 795998
1205
- num_examples: 790
1206
- - name: test
1207
- num_bytes: 1145282
1208
- num_examples: 1102
1209
- - name: challenge_train_sample
1210
- num_bytes: 247089
1211
- num_examples: 501
1212
- - name: challenge_validation_sample
1213
- num_bytes: 514117
1214
- num_examples: 500
1215
- - name: challenge_test_scramble
1216
- num_bytes: 521625
1217
- num_examples: 500
1218
- download_size: 7854845
1219
- dataset_size: 10112120
1220
- - config_name: wiki_auto_asset_turk
1221
- features:
1222
- - name: gem_id
1223
- dtype: string
1224
- - name: gem_parent_id
1225
- dtype: string
1226
- - name: source
1227
- dtype: string
1228
- - name: target
1229
- dtype: string
1230
- - name: references
1231
- list: string
1232
- splits:
1233
- - name: train
1234
- num_bytes: 161095379
1235
- num_examples: 483801
1236
- - name: validation
1237
- num_bytes: 8211308
1238
- num_examples: 20000
1239
- - name: test_asset
1240
- num_bytes: 475336
1241
- num_examples: 359
1242
- - name: test_turk
1243
- num_bytes: 406842
1244
- num_examples: 359
1245
- - name: challenge_train_sample
1246
- num_bytes: 219542
1247
- num_examples: 500
1248
- - name: challenge_validation_sample
1249
- num_bytes: 213048
1250
- num_examples: 500
1251
- - name: challenge_test_asset_backtranslation
1252
- num_bytes: 436820
1253
- num_examples: 359
1254
- - name: challenge_test_asset_bfp02
1255
- num_bytes: 432742
1256
- num_examples: 359
1257
- - name: challenge_test_asset_bfp05
1258
- num_bytes: 432742
1259
- num_examples: 359
1260
- - name: challenge_test_asset_nopunc
1261
- num_bytes: 432735
1262
- num_examples: 359
1263
- - name: challenge_test_turk_backtranslation
1264
- num_bytes: 417204
1265
- num_examples: 359
1266
- - name: challenge_test_turk_bfp02
1267
- num_bytes: 414381
1268
- num_examples: 359
1269
- - name: challenge_test_turk_bfp05
1270
- num_bytes: 414383
1271
- num_examples: 359
1272
- - name: challenge_test_turk_nopunc
1273
- num_bytes: 414388
1274
- num_examples: 359
1275
- download_size: 126927527
1276
- dataset_size: 174016850
1277
- - config_name: schema_guided_dialog
1278
- features:
1279
- - name: gem_id
1280
- dtype: string
1281
- - name: gem_parent_id
1282
- dtype: string
1283
- - name: dialog_acts
1284
- list:
1285
- - name: act
1286
- dtype:
1287
- class_label:
1288
- names:
1289
- 0: AFFIRM
1290
- 1: AFFIRM_INTENT
1291
- 2: CONFIRM
1292
- 3: GOODBYE
1293
- 4: INFORM
1294
- 5: INFORM_COUNT
1295
- 6: INFORM_INTENT
1296
- 7: NEGATE
1297
- 8: NEGATE_INTENT
1298
- 9: NOTIFY_FAILURE
1299
- 10: NOTIFY_SUCCESS
1300
- 11: OFFER
1301
- 12: OFFER_INTENT
1302
- 13: REQUEST
1303
- 14: REQUEST_ALTS
1304
- 15: REQ_MORE
1305
- 16: SELECT
1306
- 17: THANK_YOU
1307
- - name: slot
1308
- dtype: string
1309
- - name: values
1310
- list: string
1311
- - name: context
1312
- list: string
1313
- - name: dialog_id
1314
- dtype: string
1315
- - name: service
1316
- dtype: string
1317
- - name: turn_id
1318
- dtype: int32
1319
- - name: prompt
1320
- dtype: string
1321
- - name: target
1322
- dtype: string
1323
- - name: references
1324
- list: string
1325
- splits:
1326
- - name: train
1327
- num_bytes: 146648117
1328
- num_examples: 164982
1329
- - name: validation
1330
- num_bytes: 9376504
1331
- num_examples: 10000
1332
- - name: test
1333
- num_bytes: 10160596
1334
- num_examples: 10000
1335
- - name: challenge_train_sample
1336
- num_bytes: 441326
1337
- num_examples: 500
1338
- - name: challenge_validation_sample
1339
- num_bytes: 491492
1340
- num_examples: 500
1341
- - name: challenge_test_backtranslation
1342
- num_bytes: 512834
1343
- num_examples: 500
1344
- - name: challenge_test_bfp02
1345
- num_bytes: 529404
1346
- num_examples: 500
1347
- - name: challenge_test_bfp05
1348
- num_bytes: 515151
1349
- num_examples: 500
1350
- - name: challenge_test_nopunc
1351
- num_bytes: 509332
1352
- num_examples: 500
1353
- - name: challenge_test_scramble
1354
- num_bytes: 514644
1355
- num_examples: 500
1356
- download_size: 17826468
1357
- dataset_size: 169699400
1358
- ---
1359
-
1360
- # Dataset Card for GEM
1361
-
1362
- ## Table of Contents
1363
- - [Dataset Description](#dataset-description)
1364
- - [Dataset Summary](#dataset-summary)
1365
- - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
1366
- - [Languages](#languages)
1367
- - [Dataset Structure](#dataset-structure)
1368
- - [Data Instances](#data-instances)
1369
- - [Data Fields](#data-fields)
1370
- - [Data Splits](#data-splits)
1371
- - [Dataset Creation](#dataset-creation)
1372
- - [Curation Rationale](#curation-rationale)
1373
- - [Source Data](#source-data)
1374
- - [Annotations](#annotations)
1375
- - [Personal and Sensitive Information](#personal-and-sensitive-information)
1376
- - [Considerations for Using the Data](#considerations-for-using-the-data)
1377
- - [Social Impact of Dataset](#social-impact-of-dataset)
1378
- - [Discussion of Biases](#discussion-of-biases)
1379
- - [Other Known Limitations](#other-known-limitations)
1380
- - [Additional Information](#additional-information)
1381
- - [Dataset Curators](#dataset-curators)
1382
- - [Licensing Information](#licensing-information)
1383
- - [Citation Information](#citation-information)
1384
- - [Contributions](#contributions)
1385
-
1386
- ## Dataset Description
1387
-
1388
- - **Homepage:** [https://gem-benchmark.github.io/](https://gem-benchmark.github.io/)
1389
- - **Repository:**
1390
- - **Paper:** [The GEM Benchmark: Natural Language Generation, its Evaluation and Metrics](https://arxiv.org/abs/2102.01672)
1391
- - **Point of Contact:** [Sebastian Gehrman](gehrmann@google.com)
1392
- - **Size of downloaded dataset files:** 2084.23 MB
1393
- - **Size of the generated dataset:** 3734.73 MB
1394
- - **Total amount of disk used:** 5818.96 MB
1395
-
1396
- ### Dataset Summary
1397
-
1398
- GEM is a benchmark environment for Natural Language Generation with a focus on its Evaluation,
1399
- both through human annotations and automated Metrics.
1400
-
1401
- GEM aims to:
1402
- - measure NLG progress across 13 datasets spanning many NLG tasks and languages.
1403
- - provide an in-depth analysis of data and models presented via data statements and challenge sets.
1404
- - develop standards for evaluation of generated text using both automated and human metrics.
1405
-
1406
- It is our goal to regularly update GEM and to encourage toward more inclusive practices in dataset development
1407
- by extending existing data or developing datasets for additional languages.
1408
-
1409
- You can find more complete information in the dataset cards for each of the subsets:
1410
- - [CommonGen](https://gem-benchmark.com/data_cards/common_gen)
1411
- - [Czech Restaurant](https://gem-benchmark.com/data_cards/cs_restaurants)
1412
- - [DART](https://gem-benchmark.com/data_cards/dart)
1413
- - [E2E](https://gem-benchmark.com/data_cards/e2e_nlg)
1414
- - [MLSum](https://gem-benchmark.com/data_cards/mlsum)
1415
- - [Schema-Guided Dialog](https://gem-benchmark.com/data_cards/schema_guided_dialog)
1416
- - [WebNLG](https://gem-benchmark.com/data_cards/web_nlg)
1417
- - [Wiki-Auto/ASSET/TURK](https://gem-benchmark.com/data_cards/wiki_auto_asset_turk)
1418
- - [WikiLingua](https://gem-benchmark.com/data_cards/wiki_lingua)
1419
- - [XSum](https://gem-benchmark.com/data_cards/xsum)
1420
-
1421
- The subsets are organized by task:
1422
- ```
1423
- {
1424
- "summarization": {
1425
- "mlsum": ["mlsum_de", "mlsum_es"],
1426
- "wiki_lingua": ["wiki_lingua_es_en", "wiki_lingua_ru_en", "wiki_lingua_tr_en", "wiki_lingua_vi_en"],
1427
- "xsum": ["xsum"],
1428
- },
1429
- "struct2text": {
1430
- "common_gen": ["common_gen"],
1431
- "cs_restaurants": ["cs_restaurants"],
1432
- "dart": ["dart"],
1433
- "e2e": ["e2e_nlg"],
1434
- "totto": ["totto"],
1435
- "web_nlg": ["web_nlg_en", "web_nlg_ru"],
1436
- },
1437
- "simplification": {
1438
- "wiki_auto_asset_turk": ["wiki_auto_asset_turk"],
1439
- },
1440
- "dialog": {
1441
- "schema_guided_dialog": ["schema_guided_dialog"],
1442
- },
1443
- }
1444
- ```
1445
-
1446
- Each example has one `target` per example in its training set, and a set of `references` (with one or more items) in its validation and test set.
1447
-
1448
- ### Supported Tasks and Leaderboards
1449
-
1450
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
1451
-
1452
- ### Languages
1453
-
1454
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
1455
-
1456
- ## Dataset Structure
1457
-
1458
- ### Data Instances
1459
-
1460
- #### common_gen
1461
-
1462
- - **Size of downloaded dataset files:** 1.76 MB
1463
- - **Size of the generated dataset:** 8.80 MB
1464
- - **Total amount of disk used:** 10.56 MB
1465
-
1466
- An example of `validation` looks as follows.
1467
- ```
1468
- {'concept_set_id': 0,
1469
- 'concepts': ['field', 'look', 'stand'],
1470
- 'gem_id': 'common_gen-validation-0',
1471
- 'references': ['The player stood in the field looking at the batter.',
1472
- 'The coach stands along the field, looking at the goalkeeper.',
1473
- 'I stood and looked across the field, peacefully.',
1474
- 'Someone stands, looking around the empty field.'],
1475
- 'target': 'The player stood in the field looking at the batter.'}
1476
- ```
1477
-
1478
- #### cs_restaurants
1479
-
1480
- - **Size of downloaded dataset files:** 1.40 MB
1481
- - **Size of the generated dataset:** 1.25 MB
1482
- - **Total amount of disk used:** 2.64 MB
1483
-
1484
- An example of `validation` looks as follows.
1485
- ```
1486
- {'dialog_act': '?request(area)',
1487
- 'dialog_act_delexicalized': '?request(area)',
1488
- 'gem_id': 'cs_restaurants-validation-0',
1489
- 'references': ['Jakou lokalitu hledáte ?'],
1490
- 'target': 'Jakou lokalitu hledáte ?',
1491
- 'target_delexicalized': 'Jakou lokalitu hledáte ?'}
1492
- ```
1493
-
1494
- #### dart
1495
-
1496
- - **Size of downloaded dataset files:** 28.01 MB
1497
- - **Size of the generated dataset:** 26.17 MB
1498
- - **Total amount of disk used:** 54.18 MB
1499
-
1500
- An example of `validation` looks as follows.
1501
- ```
1502
- {'dart_id': 0,
1503
- 'gem_id': 'dart-validation-0',
1504
- 'references': ['A school from Mars Hill, North Carolina, joined in 1973.'],
1505
- 'subtree_was_extended': True,
1506
- 'target': 'A school from Mars Hill, North Carolina, joined in 1973.',
1507
- 'target_sources': ['WikiSQL_decl_sents'],
1508
- 'tripleset': [['Mars Hill College', 'JOINED', '1973'], ['Mars Hill College', 'LOCATION', 'Mars Hill, North Carolina']]}
1509
- ```
1510
-
1511
- #### e2e_nlg
1512
-
1513
- - **Size of downloaded dataset files:** 13.92 MB
1514
- - **Size of the generated dataset:** 11.58 MB
1515
- - **Total amount of disk used:** 25.50 MB
1516
-
1517
- An example of `validation` looks as follows.
1518
- ```
1519
- {'gem_id': 'e2e_nlg-validation-0',
1520
- 'meaning_representation': 'name[Alimentum], area[city centre], familyFriendly[no]',
1521
- 'references': ['There is a place in the city centre, Alimentum, that is not family-friendly.'],
1522
- 'target': 'There is a place in the city centre, Alimentum, that is not family-friendly.'}
1523
- ```
1524
-
1525
- #### mlsum_de
1526
-
1527
- - **Size of downloaded dataset files:** 331.27 MB
1528
- - **Size of the generated dataset:** 907.00 MB
1529
- - **Total amount of disk used:** 1238.27 MB
1530
-
1531
- An example of `validation` looks as follows.
1532
- ```
1533
- {'date': '00/04/2019',
1534
- 'gem_id': 'mlsum_de-validation-0',
1535
- 'references': ['In einer Kleinstadt auf der Insel Usedom war eine junge Frau tot in ihrer Wohnung gefunden worden. Nun stehen zwei Bekannte unter Verdacht.'],
1536
- 'target': 'In einer Kleinstadt auf der Insel Usedom war eine junge Frau tot in ihrer Wohnung gefunden worden. Nun stehen zwei Bekannte unter Verdacht.',
1537
- 'text': 'Kerzen und Blumen stehen vor dem Eingang eines Hauses, in dem eine 18-jährige Frau tot aufgefunden wurde. In einer Kleinstadt auf der Insel Usedom war eine junge Frau tot in ...',
1538
- 'title': 'Tod von 18-Jähriger auf Usedom: Zwei Festnahmen',
1539
- 'topic': 'panorama',
1540
- 'url': 'https://www.sueddeutsche.de/panorama/usedom-frau-tot-festnahme-verdaechtige-1.4412256'}
1541
- ```
1542
-
1543
- #### mlsum_es
1544
-
1545
- - **Size of downloaded dataset files:** 490.29 MB
1546
- - **Size of the generated dataset:** 1253.63 MB
1547
- - **Total amount of disk used:** 1743.92 MB
1548
-
1549
- An example of `validation` looks as follows.
1550
- ```
1551
- {'date': '05/01/2019',
1552
- 'gem_id': 'mlsum_es-validation-0',
1553
- 'references': ['El diseñador que dio carta de naturaleza al estilo genuinamente americano celebra el medio siglo de su marca entre grandes fastos y problemas financieros. Conectar con las nuevas generaciones es el regalo que precisa más que nunca'],
1554
- 'target': 'El diseñador que dio carta de naturaleza al estilo genuinamente americano celebra el medio siglo de su marca entre grandes fastos y problemas financieros. Conectar con las nuevas generaciones es el regalo que precisa más que nunca',
1555
- 'text': 'Un oso de peluche marcándose un heelflip de monopatín es todo lo que Ralph Lauren necesitaba esta Navidad. Estampado en un jersey de lana azul marino, supone la guinda que corona ...',
1556
- 'title': 'Ralph Lauren busca el secreto de la eterna juventud',
1557
- 'topic': 'elpais estilo',
1558
- 'url': 'http://elpais.com/elpais/2019/01/04/estilo/1546617396_933318.html'}
1559
- ```
1560
-
1561
- #### schema_guided_dialog
1562
-
1563
- - **Size of downloaded dataset files:** 8.24 MB
1564
- - **Size of the generated dataset:** 43.66 MB
1565
- - **Total amount of disk used:** 51.91 MB
1566
-
1567
- An example of `validation` looks as follows.
1568
- ```
1569
- {'dialog_acts': [{'act': 2, 'slot': 'song_name', 'values': ['Carnivore']}, {'act': 2, 'slot': 'playback_device', 'values': ['TV']}],
1570
- 'dialog_id': '10_00054',
1571
- 'gem_id': 'schema_guided_dialog-validation-0',
1572
- 'prompt': 'Yes, I would.',
1573
- 'references': ['Please confirm the song Carnivore on tv.'],
1574
- 'target': 'Please confirm the song Carnivore on tv.',
1575
- 'turn_id': 15}
1576
- ```
1577
-
1578
- #### totto
1579
-
1580
- - **Size of downloaded dataset files:** 179.03 MB
1581
- - **Size of the generated dataset:** 722.88 MB
1582
- - **Total amount of disk used:** 901.91 MB
1583
-
1584
- An example of `validation` looks as follows.
1585
- ```
1586
- {'example_id': '7391450717765563190',
1587
- 'gem_id': 'totto-validation-0',
1588
- 'highlighted_cells': [[3, 0], [3, 2], [3, 3]],
1589
- 'overlap_subset': 'True',
1590
- 'references': ['Daniel Henry Chamberlain was the 76th Governor of South Carolina from 1874.',
1591
- 'Daniel Henry Chamberlain was the 76th Governor of South Carolina, beginning in 1874.',
1592
- 'Daniel Henry Chamberlain was the 76th Governor of South Carolina who took office in 1874.'],
1593
- 'sentence_annotations': [{'final_sentence': 'Daniel Henry Chamberlain was the 76th Governor of South Carolina from 1874.',
1594
- 'original_sentence': 'Daniel Henry Chamberlain (June 23, 1835 – April 13, 1907) was an American planter, lawyer, author and the 76th Governor of South Carolina '
1595
- 'from 1874 until 1877.',
1596
- 'sentence_after_ambiguity': 'Daniel Henry Chamberlain was the 76th Governor of South Carolina from 1874.',
1597
- 'sentence_after_deletion': 'Daniel Henry Chamberlain was the 76th Governor of South Carolina from 1874.'},
1598
- ...
1599
- ],
1600
- 'table': [[{'column_span': 1, 'is_header': True, 'row_span': 1, 'value': '#'},
1601
- {'column_span': 2, 'is_header': True, 'row_span': 1, 'value': 'Governor'},
1602
- {'column_span': 1, 'is_header': True, 'row_span': 1, 'value': 'Took Office'},
1603
- {'column_span': 1, 'is_header': True, 'row_span': 1, 'value': 'Left Office'}],
1604
- [{'column_span': 1, 'is_header': True, 'row_span': 1, 'value': '74'},
1605
- {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': '-'},
1606
- {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'Robert Kingston Scott'},
1607
- {'column_span': 1, 'is_header': False, 'row_span': 1, 'value': 'July 6, 1868'}],
1608
- ...
1609
- ],
1610
- 'table_page_title': 'List of Governors of South Carolina',
1611
- 'table_section_text': 'Parties Democratic Republican',
1612
- 'table_section_title': 'Governors under the Constitution of 1868',
1613
- 'table_webpage_url': 'http://en.wikipedia.org/wiki/List_of_Governors_of_South_Carolina',
1614
- 'target': 'Daniel Henry Chamberlain was the 76th Governor of South Carolina from 1874.',
1615
- 'totto_id': 0}
1616
- ```
1617
-
1618
- #### web_nlg_en
1619
-
1620
- - **Size of downloaded dataset files:** 12.35 MB
1621
- - **Size of the generated dataset:** 13.95 MB
1622
- - **Total amount of disk used:** 26.29 MB
1623
-
1624
- An example of `validation` looks as follows.
1625
- ```
1626
- {'category': 'Airport',
1627
- 'gem_id': 'web_nlg_en-validation-0',
1628
- 'input': ['Aarhus | leader | Jacob_Bundsgaard'],
1629
- 'references': ['The leader of Aarhus is Jacob Bundsgaard.'],
1630
- 'target': 'The leader of Aarhus is Jacob Bundsgaard.',
1631
- 'webnlg_id': 'dev/Airport/1/Id1'}
1632
- ```
1633
-
1634
- #### web_nlg_ru
1635
-
1636
- - **Size of downloaded dataset files:** 7.28 MB
1637
- - **Size of the generated dataset:** 8.02 MB
1638
- - **Total amount of disk used:** 15.30 MB
1639
-
1640
- An example of `validation` looks as follows.
1641
- ```
1642
- {'category': 'Airport',
1643
- 'gem_id': 'web_nlg_ru-validation-0',
1644
- 'input': ['Punjab,_Pakistan | leaderTitle | Provincial_Assembly_of_the_Punjab'],
1645
- 'references': ['Пенджаб, Пакистан, возглавляется Провинциальной ассамблеей Пенджаба.', 'Пенджаб, Пакистан возглавляется Провинциальной ассамблеей Пенджаба.'],
1646
- 'target': 'Пенджаб, Пакистан, возглавляется Провинциальной ассамблеей Пенджаба.',
1647
- 'webnlg_id': 'dev/Airport/1/Id1'}
1648
- ```
1649
-
1650
- #### wiki_auto_asset_turk
1651
-
1652
- - **Size of downloaded dataset files:** 121.37 MB
1653
- - **Size of the generated dataset:** 145.69 MB
1654
- - **Total amount of disk used:** 267.07 MB
1655
-
1656
- An example of `validation` looks as follows.
1657
- ```
1658
- {'gem_id': 'wiki_auto_asset_turk-validation-0',
1659
- 'references': ['The Gandalf Awards honor excellent writing in in fantasy literature.'],
1660
- 'source': 'The Gandalf Awards, honoring achievement in fantasy literature, were conferred by the World Science Fiction Society annually from 1974 to 1981.',
1661
- 'source_id': '350_691837-1-0-0',
1662
- 'target': 'The Gandalf Awards honor excellent writing in in fantasy literature.',
1663
- 'target_id': '350_691837-0-0-0'}
1664
- ```
1665
-
1666
- #### wiki_lingua_es_en
1667
-
1668
- - **Size of downloaded dataset files:** 161.56 MB
1669
- - **Size of the generated dataset:** 274.28 MB
1670
- - **Total amount of disk used:** 435.84 MB
1671
-
1672
- An example of `validation` looks as follows.
1673
- ```
1674
- 'references': ["Practice matted hair prevention from early in your cat's life. Make sure that your cat is grooming itself effectively. Keep a close eye on cats with long hair."],
1675
- 'source': 'Muchas personas presentan problemas porque no cepillaron el pelaje de sus gatos en una etapa temprana de su vida, ya que no lo consideraban necesario. Sin embargo, a medida que...',
1676
- 'target': "Practice matted hair prevention from early in your cat's life. Make sure that your cat is grooming itself effectively. Keep a close eye on cats with long hair."}
1677
- ```
1678
-
1679
- #### wiki_lingua_ru_en
1680
-
1681
- - **Size of downloaded dataset files:** 161.56 MB
1682
- - **Size of the generated dataset:** 201.43 MB
1683
- - **Total amount of disk used:** 362.99 MB
1684
-
1685
- An example of `validation` looks as follows.
1686
- ```
1687
- {'gem_id': 'wiki_lingua_ru_en-val-0',
1688
- 'references': ['Get immediate medical care if you notice signs of a complication. Undergo diagnostic tests to check for gallstones and complications. Ask your doctor about your treatment '
1689
- 'options.'],
1690
- 'source': 'И хотя, скорее всего, вам не о чем волноваться, следует незамедлительно обратиться к врачу, если вы подозреваете, что у вас возникло осложнение желчекаменной болезни. Это ...',
1691
- 'target': 'Get immediate medical care if you notice signs of a complication. Undergo diagnostic tests to check for gallstones and complications. Ask your doctor about your treatment '
1692
- 'options.'}
1693
- ```
1694
-
1695
- #### wiki_lingua_tr_en
1696
-
1697
- - **Size of downloaded dataset files:** 161.56 MB
1698
- - **Size of the generated dataset:** 9.87 MB
1699
- - **Total amount of disk used:** 171.42 MB
1700
-
1701
- An example of `validation` looks as follows.
1702
- ```
1703
- {'gem_id': 'wiki_lingua_tr_en-val-0',
1704
- 'references': ['Open Instagram. Go to the video you want to download. Tap ⋮. Tap Copy Link. Open Google Chrome. Tap the address bar. Go to the SaveFromWeb site. Tap the "Paste Instagram Video" text box. Tap and hold the text box. Tap PASTE. Tap Download. Download the video. Find the video on your Android.'],
1705
- 'source': 'Instagram uygulamasının çok renkli kamera şeklindeki simgesine dokun. Daha önce giriş yaptıysan Instagram haber kaynağı açılır. Giriş yapmadıysan istendiğinde e-posta adresini ...',
1706
- 'target': 'Open Instagram. Go to the video you want to download. Tap ⋮. Tap Copy Link. Open Google Chrome. Tap the address bar. Go to the SaveFromWeb site. Tap the "Paste Instagram Video" text box. Tap and hold the text box. Tap PASTE. Tap Download. Download the video. Find the video on your Android.'}
1707
- ```
1708
-
1709
- #### wiki_lingua_vi_en
1710
-
1711
- - **Size of downloaded dataset files:** 161.56 MB
1712
- - **Size of the generated dataset:** 39.12 MB
1713
- - **Total amount of disk used:** 200.68 MB
1714
-
1715
- An example of `validation` looks as follows.
1716
- ```
1717
- {'gem_id': 'wiki_lingua_vi_en-val-0',
1718
- 'references': ['Select the right time of year for planting the tree. You will usually want to plant your tree when it is dormant, or not flowering, during cooler or colder times of year.'],
1719
- 'source': 'Bạn muốn cung cấp cho cây cơ hội tốt nhất để phát triển và sinh tồn. Trồng cây đúng thời điểm trong năm chính là yếu tố then chốt. Thời điểm sẽ thay đổi phụ thuộc vào loài cây ...',
1720
- 'target': 'Select the right time of year for planting the tree. You will usually want to plant your tree when it is dormant, or not flowering, during cooler or colder times of year.'}
1721
- ```
1722
-
1723
- #### xsum
1724
-
1725
- - **Size of downloaded dataset files:** 243.08 MB
1726
- - **Size of the generated dataset:** 67.40 MB
1727
- - **Total amount of disk used:** 310.48 MB
1728
-
1729
- An example of `validation` looks as follows.
1730
- ```
1731
- {'document': 'Burberry reported pre-tax profits of £166m for the year to March. A year ago it made a loss of £16.1m, hit by charges at its Spanish operations.\n'
1732
- 'In the past year it has opened 21 new stores and closed nine. It plans to open 20-30 stores this year worldwide.\n'
1733
- 'The group has also focused on promoting the Burberry brand online...',
1734
- 'gem_id': 'xsum-validation-0',
1735
- 'references': ['Luxury fashion designer Burberry has returned to profit after opening new stores and spending more on online marketing'],
1736
- 'target': 'Luxury fashion designer Burberry has returned to profit after opening new stores and spending more on online marketing',
1737
- 'xsum_id': '10162122'}
1738
- ```
1739
-
1740
- ### Data Fields
1741
-
1742
- The data fields are the same among all splits.
1743
-
1744
- #### common_gen
1745
- - `gem_id`: a `string` feature.
1746
- - `concept_set_id`: a `int32` feature.
1747
- - `concepts`: a `list` of `string` features.
1748
- - `target`: a `string` feature.
1749
- - `references`: a `list` of `string` features.
1750
-
1751
- #### cs_restaurants
1752
- - `gem_id`: a `string` feature.
1753
- - `dialog_act`: a `string` feature.
1754
- - `dialog_act_delexicalized`: a `string` feature.
1755
- - `target_delexicalized`: a `string` feature.
1756
- - `target`: a `string` feature.
1757
- - `references`: a `list` of `string` features.
1758
-
1759
- #### dart
1760
- - `gem_id`: a `string` feature.
1761
- - `dart_id`: a `int32` feature.
1762
- - `tripleset`: a `list` of `string` features.
1763
- - `subtree_was_extended`: a `bool` feature.
1764
- - `target_sources`: a `list` of `string` features.
1765
- - `target`: a `string` feature.
1766
- - `references`: a `list` of `string` features.
1767
-
1768
- #### e2e_nlg
1769
- - `gem_id`: a `string` feature.
1770
- - `meaning_representation`: a `string` feature.
1771
- - `target`: a `string` feature.
1772
- - `references`: a `list` of `string` features.
1773
-
1774
- #### mlsum_de
1775
- - `gem_id`: a `string` feature.
1776
- - `text`: a `string` feature.
1777
- - `topic`: a `string` feature.
1778
- - `url`: a `string` feature.
1779
- - `title`: a `string` feature.
1780
- - `date`: a `string` feature.
1781
- - `target`: a `string` feature.
1782
- - `references`: a `list` of `string` features.
1783
-
1784
- #### mlsum_es
1785
- - `gem_id`: a `string` feature.
1786
- - `text`: a `string` feature.
1787
- - `topic`: a `string` feature.
1788
- - `url`: a `string` feature.
1789
- - `title`: a `string` feature.
1790
- - `date`: a `string` feature.
1791
- - `target`: a `string` feature.
1792
- - `references`: a `list` of `string` features.
1793
-
1794
- #### schema_guided_dialog
1795
- - `gem_id`: a `string` feature.
1796
- - `act`: a classification label, with possible values including `AFFIRM` (0), `AFFIRM_INTENT` (1), `CONFIRM` (2), `GOODBYE` (3), `INFORM` (4).
1797
- - `slot`: a `string` feature.
1798
- - `values`: a `list` of `string` features.
1799
- - `dialog_id`: a `string` feature.
1800
- - `turn_id`: a `int32` feature.
1801
- - `prompt`: a `string` feature.
1802
- - `target`: a `string` feature.
1803
- - `references`: a `list` of `string` features.
1804
-
1805
- #### totto
1806
- - `gem_id`: a `string` feature.
1807
- - `totto_id`: a `int32` feature.
1808
- - `table_page_title`: a `string` feature.
1809
- - `table_webpage_url`: a `string` feature.
1810
- - `table_section_title`: a `string` feature.
1811
- - `table_section_text`: a `string` feature.
1812
- - `column_span`: a `int32` feature.
1813
- - `is_header`: a `bool` feature.
1814
- - `row_span`: a `int32` feature.
1815
- - `value`: a `string` feature.
1816
- - `highlighted_cells`: a `list` of `int32` features.
1817
- - `example_id`: a `string` feature.
1818
- - `original_sentence`: a `string` feature.
1819
- - `sentence_after_deletion`: a `string` feature.
1820
- - `sentence_after_ambiguity`: a `string` feature.
1821
- - `final_sentence`: a `string` feature.
1822
- - `overlap_subset`: a `string` feature.
1823
- - `target`: a `string` feature.
1824
- - `references`: a `list` of `string` features.
1825
-
1826
- #### web_nlg_en
1827
- - `gem_id`: a `string` feature.
1828
- - `input`: a `list` of `string` features.
1829
- - `target`: a `string` feature.
1830
- - `references`: a `list` of `string` features.
1831
- - `category`: a `string` feature.
1832
- - `webnlg_id`: a `string` feature.
1833
-
1834
- #### web_nlg_ru
1835
- - `gem_id`: a `string` feature.
1836
- - `input`: a `list` of `string` features.
1837
- - `target`: a `string` feature.
1838
- - `references`: a `list` of `string` features.
1839
- - `category`: a `string` feature.
1840
- - `webnlg_id`: a `string` feature.
1841
-
1842
- #### wiki_auto_asset_turk
1843
- - `gem_id`: a `string` feature.
1844
- - `source_id`: a `string` feature.
1845
- - `target_id`: a `string` feature.
1846
- - `source`: a `string` feature.
1847
- - `target`: a `string` feature.
1848
- - `references`: a `list` of `string` features.
1849
-
1850
- #### wiki_lingua_es_en
1851
- - `gem_id`: a `string` feature.
1852
- - `source`: a `string` feature.
1853
- - `target`: a `string` feature.
1854
- - `references`: a `list` of `string` features.
1855
-
1856
- #### wiki_lingua_ru_en
1857
- - `gem_id`: a `string` feature.
1858
- - `source`: a `string` feature.
1859
- - `target`: a `string` feature.
1860
- - `references`: a `list` of `string` features.
1861
-
1862
- #### wiki_lingua_tr_en
1863
- - `gem_id`: a `string` feature.
1864
- - `source`: a `string` feature.
1865
- - `target`: a `string` feature.
1866
- - `references`: a `list` of `string` features.
1867
-
1868
- #### wiki_lingua_vi_en
1869
- - `gem_id`: a `string` feature.
1870
- - `source`: a `string` feature.
1871
- - `target`: a `string` feature.
1872
- - `references`: a `list` of `string` features.
1873
-
1874
- #### xsum
1875
- - `gem_id`: a `string` feature.
1876
- - `xsum_id`: a `string` feature.
1877
- - `document`: a `string` feature.
1878
- - `target`: a `string` feature.
1879
- - `references`: a `list` of `string` features.
1880
-
1881
- ### Data Splits
1882
-
1883
- #### common_gen
1884
-
1885
- | |train|validation|test|
1886
- |----------|----:|---------:|---:|
1887
- |common_gen|67389| 993|1497|
1888
-
1889
- #### cs_restaurants
1890
-
1891
- | |train|validation|test|
1892
- |--------------|----:|---------:|---:|
1893
- |cs_restaurants| 3569| 781| 842|
1894
-
1895
- #### dart
1896
-
1897
- | |train|validation|test|
1898
- |----|----:|---------:|---:|
1899
- |dart|62659| 2768|6959|
1900
-
1901
- #### e2e_nlg
1902
-
1903
- | |train|validation|test|
1904
- |-------|----:|---------:|---:|
1905
- |e2e_nlg|33525| 4299|4693|
1906
-
1907
- #### mlsum_de
1908
-
1909
- | |train |validation|test |
1910
- |--------|-----:|---------:|----:|
1911
- |mlsum_de|220748| 11392|10695|
1912
-
1913
- #### mlsum_es
1914
-
1915
- | |train |validation|test |
1916
- |--------|-----:|---------:|----:|
1917
- |mlsum_es|259886| 9977|13365|
1918
-
1919
- #### schema_guided_dialog
1920
-
1921
- | |train |validation|test |
1922
- |--------------------|-----:|---------:|----:|
1923
- |schema_guided_dialog|164982| 10000|10000|
1924
-
1925
- #### totto
1926
-
1927
- | |train |validation|test|
1928
- |-----|-----:|---------:|---:|
1929
- |totto|121153| 7700|7700|
1930
-
1931
- #### web_nlg_en
1932
-
1933
- | |train|validation|test|
1934
- |----------|----:|---------:|---:|
1935
- |web_nlg_en|35426| 1667|1779|
1936
-
1937
- #### web_nlg_ru
1938
-
1939
- | |train|validation|test|
1940
- |----------|----:|---------:|---:|
1941
- |web_nlg_ru|14630| 790|1102|
1942
-
1943
- #### wiki_auto_asset_turk
1944
-
1945
- | |train |validation|test_asset|test_turk|
1946
- |--------------------|-----:|---------:|---------:|--------:|
1947
- |wiki_auto_asset_turk|373801| 73249| 359| 359|
1948
-
1949
- #### wiki_lingua_es_en
1950
-
1951
- | |train|validation|test |
1952
- |-----------------|----:|---------:|----:|
1953
- |wiki_lingua_es_en|79515| 8835|19797|
1954
-
1955
- #### wiki_lingua_ru_en
1956
-
1957
- | |train|validation|test|
1958
- |-----------------|----:|---------:|---:|
1959
- |wiki_lingua_ru_en|36898| 4100|9094|
1960
-
1961
- #### wiki_lingua_tr_en
1962
-
1963
- | |train|validation|test|
1964
- |-----------------|----:|---------:|---:|
1965
- |wiki_lingua_tr_en| 3193| 355| 808|
1966
-
1967
- #### wiki_lingua_vi_en
1968
-
1969
- | |train|validation|test|
1970
- |-----------------|----:|---------:|---:|
1971
- |wiki_lingua_vi_en| 9206| 1023|2167|
1972
-
1973
- #### xsum
1974
-
1975
- | |train|validation|test|
1976
- |----|----:|---------:|---:|
1977
- |xsum|23206| 1117|1166|
1978
-
1979
- ## Dataset Creation
1980
-
1981
- ### Curation Rationale
1982
-
1983
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
1984
-
1985
- ### Source Data
1986
-
1987
- #### Initial Data Collection and Normalization
1988
-
1989
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
1990
-
1991
- #### Who are the source language producers?
1992
-
1993
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
1994
-
1995
- ### Annotations
1996
-
1997
- #### Annotation process
1998
-
1999
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
2000
-
2001
- #### Who are the annotators?
2002
-
2003
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
2004
-
2005
- ### Personal and Sensitive Information
2006
-
2007
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
2008
-
2009
- ## Considerations for Using the Data
2010
-
2011
- ### Social Impact of Dataset
2012
-
2013
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
2014
-
2015
- ### Discussion of Biases
2016
-
2017
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
2018
-
2019
- ### Other Known Limitations
2020
-
2021
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
2022
-
2023
- ## Additional Information
2024
-
2025
- ### Dataset Curators
2026
-
2027
- [More Information Needed](https://github.com/huggingface/datasets/blob/master/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
2028
-
2029
- ### Licensing Information
2030
-
2031
- CC-BY-SA-4.0
2032
-
2033
- ### Citation Information
2034
-
2035
- ```
2036
- @article{gem_benchmark,
2037
- author = {Sebastian Gehrmann and
2038
- Tosin P. Adewumi and
2039
- Karmanya Aggarwal and
2040
- Pawan Sasanka Ammanamanchi and
2041
- Aremu Anuoluwapo and
2042
- Antoine Bosselut and
2043
- Khyathi Raghavi Chandu and
2044
- Miruna{-}Adriana Clinciu and
2045
- Dipanjan Das and
2046
- Kaustubh D. Dhole and
2047
- Wanyu Du and
2048
- Esin Durmus and
2049
- Ondrej Dusek and
2050
- Chris Emezue and
2051
- Varun Gangal and
2052
- Cristina Garbacea and
2053
- Tatsunori Hashimoto and
2054
- Yufang Hou and
2055
- Yacine Jernite and
2056
- Harsh Jhamtani and
2057
- Yangfeng Ji and
2058
- Shailza Jolly and
2059
- Dhruv Kumar and
2060
- Faisal Ladhak and
2061
- Aman Madaan and
2062
- Mounica Maddela and
2063
- Khyati Mahajan and
2064
- Saad Mahamood and
2065
- Bodhisattwa Prasad Majumder and
2066
- Pedro Henrique Martins and
2067
- Angelina McMillan{-}Major and
2068
- Simon Mille and
2069
- Emiel van Miltenburg and
2070
- Moin Nadeem and
2071
- Shashi Narayan and
2072
- Vitaly Nikolaev and
2073
- Rubungo Andre Niyongabo and
2074
- Salomey Osei and
2075
- Ankur P. Parikh and
2076
- Laura Perez{-}Beltrachini and
2077
- Niranjan Ramesh Rao and
2078
- Vikas Raunak and
2079
- Juan Diego Rodriguez and
2080
- Sashank Santhanam and
2081
- Jo{\~{a}}o Sedoc and
2082
- Thibault Sellam and
2083
- Samira Shaikh and
2084
- Anastasia Shimorina and
2085
- Marco Antonio Sobrevilla Cabezudo and
2086
- Hendrik Strobelt and
2087
- Nishant Subramani and
2088
- Wei Xu and
2089
- Diyi Yang and
2090
- Akhila Yerukola and
2091
- Jiawei Zhou},
2092
- title = {The {GEM} Benchmark: Natural Language Generation, its Evaluation and
2093
- Metrics},
2094
- journal = {CoRR},
2095
- volume = {abs/2102.01672},
2096
- year = {2021},
2097
- url = {https://arxiv.org/abs/2102.01672},
2098
- archivePrefix = {arXiv},
2099
- eprint = {2102.01672}
2100
- }
2101
- ```
2102
-
2103
- ### Contributions
2104
-
2105
- Thanks to [@yjernite](https://github.com/yjernite) for adding this dataset.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
dataset_infos.json DELETED
The diff for this file is too large to render. See raw diff
 
gem.py DELETED
@@ -1,1335 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
- """GEM: Generation Evaluation Metrics supporting datasets"""
16
-
17
-
18
- import csv
19
- import json
20
- import os
21
-
22
- import datasets
23
-
24
-
25
- _CITATION = """\
26
- @article{gem_benchmark,
27
- author = {Sebastian Gehrmann and
28
- Tosin P. Adewumi and
29
- Karmanya Aggarwal and
30
- Pawan Sasanka Ammanamanchi and
31
- Aremu Anuoluwapo and
32
- Antoine Bosselut and
33
- Khyathi Raghavi Chandu and
34
- Miruna{-}Adriana Clinciu and
35
- Dipanjan Das and
36
- Kaustubh D. Dhole and
37
- Wanyu Du and
38
- Esin Durmus and
39
- Ondrej Dusek and
40
- Chris Emezue and
41
- Varun Gangal and
42
- Cristina Garbacea and
43
- Tatsunori Hashimoto and
44
- Yufang Hou and
45
- Yacine Jernite and
46
- Harsh Jhamtani and
47
- Yangfeng Ji and
48
- Shailza Jolly and
49
- Dhruv Kumar and
50
- Faisal Ladhak and
51
- Aman Madaan and
52
- Mounica Maddela and
53
- Khyati Mahajan and
54
- Saad Mahamood and
55
- Bodhisattwa Prasad Majumder and
56
- Pedro Henrique Martins and
57
- Angelina McMillan{-}Major and
58
- Simon Mille and
59
- Emiel van Miltenburg and
60
- Moin Nadeem and
61
- Shashi Narayan and
62
- Vitaly Nikolaev and
63
- Rubungo Andre Niyongabo and
64
- Salomey Osei and
65
- Ankur P. Parikh and
66
- Laura Perez{-}Beltrachini and
67
- Niranjan Ramesh Rao and
68
- Vikas Raunak and
69
- Juan Diego Rodriguez and
70
- Sashank Santhanam and
71
- Joao Sedoc and
72
- Thibault Sellam and
73
- Samira Shaikh and
74
- Anastasia Shimorina and
75
- Marco Antonio Sobrevilla Cabezudo and
76
- Hendrik Strobelt and
77
- Nishant Subramani and
78
- Wei Xu and
79
- Diyi Yang and
80
- Akhila Yerukola and
81
- Jiawei Zhou},
82
- title = {The {GEM} Benchmark: Natural Language Generation, its Evaluation and
83
- Metrics},
84
- journal = {CoRR},
85
- volume = {abs/2102.01672},
86
- year = {2021},
87
- url = {https://arxiv.org/abs/2102.01672},
88
- archivePrefix = {arXiv},
89
- eprint = {2102.01672}
90
- }
91
- """
92
-
93
- _DESCRIPTION = """\
94
- GEM is a benchmark environment for Natural Language Generation with a focus on its Evaluation,
95
- both through human annotations and automated Metrics.
96
-
97
- GEM aims to:
98
- - measure NLG progress across 13 datasets spanning many NLG tasks and languages.
99
- - provide an in-depth analysis of data and models presented via data statements and challenge sets.
100
- - develop standards for evaluation of generated text using both automated and human metrics.
101
-
102
- It is our goal to regularly update GEM and to encourage toward more inclusive practices in dataset development
103
- by extending existing data or developing datasets for additional languages.
104
- """
105
-
106
- _HOMEPAGE = "https://gem-benchmark.github.io/"
107
-
108
- _LICENSE = "CC-BY-SA-4.0"
109
-
110
- _TASKS = {
111
- "summarization": {
112
- "mlsum": ["mlsum_de", "mlsum_es"],
113
- "wiki_lingua": [
114
- "wiki_lingua_es_en_v0",
115
- "wiki_lingua_ru_en_v0",
116
- "wiki_lingua_tr_en_v0",
117
- "wiki_lingua_vi_en_v0",
118
- "wiki_lingua_arabic_ar",
119
- "wiki_lingua_chinese_zh",
120
- "wiki_lingua_czech_cs",
121
- "wiki_lingua_dutch_nl",
122
- "wiki_lingua_english_en",
123
- "wiki_lingua_french_fr",
124
- "wiki_lingua_german_de",
125
- "wiki_lingua_hindi_hi",
126
- "wiki_lingua_indonesian_id",
127
- "wiki_lingua_italian_it",
128
- "wiki_lingua_japanese_ja",
129
- "wiki_lingua_korean_ko",
130
- "wiki_lingua_portuguese_pt",
131
- "wiki_lingua_russian_ru",
132
- "wiki_lingua_spanish_es",
133
- "wiki_lingua_thai_th",
134
- "wiki_lingua_turkish_tr",
135
- "wiki_lingua_vietnamese_vi",
136
- ],
137
- "xsum": ["xsum"],
138
- },
139
- "struct2text": {
140
- "common_gen": ["common_gen"],
141
- "cs_restaurants": ["cs_restaurants"],
142
- "dart": ["dart"],
143
- "e2e": ["e2e_nlg"],
144
- "totto": ["totto"],
145
- "web_nlg": ["web_nlg_en", "web_nlg_ru"],
146
- },
147
- "simplification": {
148
- "wiki_auto_asset_turk": ["wiki_auto_asset_turk"],
149
- },
150
- "dialog": {
151
- "schema_guided_dialog": ["schema_guided_dialog"],
152
- },
153
- }
154
-
155
- _URLs = {
156
- "common_gen": {
157
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/common_gen/commongen_data.zip",
158
- "challenge_set": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_challenge_sets/common_gen.zip",
159
- },
160
- "cs_restaurants": {
161
- "train": "https://raw.githubusercontent.com/UFAL-DSG/cs_restaurant_dataset/master/train.json",
162
- "validation": "https://raw.githubusercontent.com/UFAL-DSG/cs_restaurant_dataset/master/devel.json",
163
- "test": "https://raw.githubusercontent.com/UFAL-DSG/cs_restaurant_dataset/master/test.json",
164
- "challenge_set": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_challenge_sets/cs_restaurants.zip",
165
- },
166
- "dart": {
167
- "train": "https://raw.githubusercontent.com/Yale-LILY/dart/master/data/v1.1.1/dart-v1.1.1-full-train.json",
168
- "validation": "https://raw.githubusercontent.com/Yale-LILY/dart/master/data/v1.1.1/dart-v1.1.1-full-dev.json",
169
- "test": "https://raw.githubusercontent.com/Yale-LILY/dart/master/data/v1.1.1/dart-v1.1.1-full-test.json",
170
- },
171
- "e2e_nlg": {
172
- "train": "https://github.com/tuetschek/e2e-cleaning/raw/master/cleaned-data/train-fixed.no-ol.csv",
173
- "validation": "https://github.com/tuetschek/e2e-cleaning/raw/master/cleaned-data/devel-fixed.no-ol.csv",
174
- "test": "https://github.com/tuetschek/e2e-cleaning/raw/master/cleaned-data/test-fixed.csv",
175
- "challenge_set": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_challenge_sets/e2e_nlg.zip",
176
- },
177
- "mlsum_de": {
178
- "train": "https://gitlab.lip6.fr/scialom/mlsum_data/-/raw/master/MLSUM/de_train.zip",
179
- "validation": "https://gitlab.lip6.fr/scialom/mlsum_data/-/raw/master/MLSUM/de_val.zip",
180
- "test": "https://gitlab.lip6.fr/scialom/mlsum_data/-/raw/master/MLSUM/de_test.zip",
181
- "bad_ids": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_mlsum_bad_ids_fixed.json",
182
- "challenge_set": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_challenge_sets/mlsum_de.zip",
183
- },
184
- "mlsum_es": {
185
- "train": "https://gitlab.lip6.fr/scialom/mlsum_data/-/raw/master/MLSUM/es_train.zip",
186
- "validation": "https://gitlab.lip6.fr/scialom/mlsum_data/-/raw/master/MLSUM/es_val.zip",
187
- "test": "https://gitlab.lip6.fr/scialom/mlsum_data/-/raw/master/MLSUM/es_test.zip",
188
- "bad_ids": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_mlsum_bad_ids_fixed.json",
189
- "challenge_set": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_challenge_sets/mlsum_es.zip",
190
- },
191
- "schema_guided_dialog": {
192
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_sgd_context.zip",
193
- "challenge_set": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_challenge_sets/schema_guided_dialog.zip",
194
- },
195
- "totto": {
196
- "data": "https://storage.googleapis.com/totto-public/totto_data.zip",
197
- "challenge_set": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_challenge_sets/totto.zip",
198
- },
199
- "web_nlg_en": {
200
- "train": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_web_nlg/webnlg_en_train.json",
201
- "validation": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_web_nlg/webnlg_en_val.json",
202
- "test": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_web_nlg/webnlg_en_test.json",
203
- "challenge_set": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_challenge_sets/web_nlg_en.zip",
204
- },
205
- "web_nlg_ru": {
206
- "train": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_web_nlg/webnlg_ru_train.json",
207
- "validation": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_web_nlg/webnlg_ru_val.json",
208
- "test": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_web_nlg/webnlg_ru_test.json",
209
- "challenge_set": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_challenge_sets/web_nlg_ru.zip",
210
- },
211
- "wiki_auto_asset_turk": {
212
- "train": "https://github.com/chaojiang06/wiki-auto/raw/master/wiki-auto/GEM2021/full_with_split/train.tsv",
213
- "validation": "https://github.com/chaojiang06/wiki-auto/raw/master/wiki-auto/GEM2021/full_with_split/valid.tsv",
214
- "test_turk": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_turk_detokenized.json",
215
- "challenge_set": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_challenge_sets/wiki_auto_asset_turk_train_valid.zip",
216
- },
217
- "wiki_lingua_es_en_v0": {
218
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua.zip",
219
- },
220
- "wiki_lingua_ru_en_v0": {
221
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua.zip",
222
- },
223
- "wiki_lingua_tr_en_v0": {
224
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua.zip",
225
- },
226
- "wiki_lingua_vi_en_v0": {
227
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua.zip",
228
- },
229
- "wiki_lingua_arabic_ar": {
230
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/arabic.zip",
231
- },
232
- "wiki_lingua_chinese_zh": {
233
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/chinese.zip",
234
- },
235
- "wiki_lingua_czech_cs": {
236
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/czech.zip",
237
- },
238
- "wiki_lingua_dutch_nl": {
239
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/dutch.zip",
240
- },
241
- "wiki_lingua_english_en": {
242
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/english.zip",
243
- },
244
- "wiki_lingua_french_fr": {
245
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/french.zip",
246
- },
247
- "wiki_lingua_german_de": {
248
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/german.zip",
249
- },
250
- "wiki_lingua_hindi_hi": {
251
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/hindi.zip",
252
- },
253
- "wiki_lingua_indonesian_id": {
254
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/indonesian.zip",
255
- },
256
- "wiki_lingua_italian_it": {
257
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/italian.zip",
258
- },
259
- "wiki_lingua_japanese_ja": {
260
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/japanese.zip",
261
- },
262
- "wiki_lingua_korean_ko": {
263
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/korean.zip",
264
- },
265
- "wiki_lingua_portuguese_pt": {
266
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/portuguese.zip",
267
- },
268
- "wiki_lingua_russian_ru": {
269
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/russian.zip",
270
- },
271
- "wiki_lingua_spanish_es": {
272
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/spanish.zip",
273
- },
274
- "wiki_lingua_thai_th": {
275
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/thai.zip",
276
- },
277
- "wiki_lingua_turkish_tr": {
278
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/turkish.zip",
279
- },
280
- "wiki_lingua_vietnamese_vi": {
281
- "data": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_wikilingua_full/vietnamese.zip",
282
- },
283
- "xsum": {
284
- "data": "http://bollin.inf.ed.ac.uk/public/direct/XSUM-EMNLP18-Summary-Data-Original.tar.gz",
285
- "splits": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_xsum_confidence_0.8.json",
286
- "challenge_set": "https://storage.googleapis.com/huggingface-nlp/datasets/gem/gem_challenge_sets/xsum.zip",
287
- },
288
- }
289
-
290
- # Add Asset files
291
- _URLs["wiki_auto_asset_turk"][
292
- "test_asset_orig"
293
- ] = "https://github.com/facebookresearch/asset/raw/main/dataset/asset.test.orig"
294
- for i in range(10):
295
- _URLs["wiki_auto_asset_turk"][
296
- f"test_asset_{i}"
297
- ] = f"https://github.com/facebookresearch/asset/raw/main/dataset/asset.test.simp.{i}"
298
-
299
- _SGD_ACTS = [
300
- "AFFIRM",
301
- "AFFIRM_INTENT",
302
- "CONFIRM",
303
- "GOODBYE",
304
- "INFORM",
305
- "INFORM_COUNT",
306
- "INFORM_INTENT",
307
- "NEGATE",
308
- "NEGATE_INTENT",
309
- "NOTIFY_FAILURE",
310
- "NOTIFY_SUCCESS",
311
- "OFFER",
312
- "OFFER_INTENT",
313
- "REQUEST",
314
- "REQUEST_ALTS",
315
- "REQ_MORE",
316
- "SELECT",
317
- "THANK_YOU",
318
- ]
319
-
320
- _XSUM_REMOVE_LINES = set(
321
- [
322
- "Share this with\n",
323
- "Email\n",
324
- "Facebook\n",
325
- "Messenger\n",
326
- "Twitter\n",
327
- "Pinterest\n",
328
- "WhatsApp\n",
329
- "Linkedin\n",
330
- "LinkedIn\n",
331
- "Copy this link\n",
332
- "These are external links and will open in a new window\n",
333
- ]
334
- )
335
-
336
-
337
- class Gem(datasets.GeneratorBasedBuilder):
338
- """GEM: datasets supporting the Generation Evaluation Metrics 2021 shared task."""
339
-
340
- BUILDER_CONFIGS = [
341
- datasets.BuilderConfig(
342
- name=conf,
343
- version=datasets.Version("1.1.0"),
344
- description=f"GEM benchmark: {task} task, {conf} subset",
345
- )
346
- for task, dset_confs in _TASKS.items()
347
- for conf_list in dset_confs.values()
348
- for conf in conf_list
349
- ]
350
-
351
- DEFAULT_CONFIG_NAME = "common_gen" # First alphabetical
352
-
353
- def _info(self):
354
- if self.config.name == "common_gen":
355
- features = datasets.Features(
356
- {
357
- "gem_id": datasets.Value("string"),
358
- "gem_parent_id": datasets.Value("string"),
359
- "concept_set_id": datasets.Value("int32"),
360
- "concepts": [datasets.Value("string")],
361
- "target": datasets.Value("string"), # single target for train
362
- "references": [datasets.Value("string")], # multiple references for validation
363
- }
364
- )
365
- elif self.config.name == "cs_restaurants":
366
- features = datasets.Features(
367
- {
368
- "gem_id": datasets.Value("string"),
369
- "gem_parent_id": datasets.Value("string"),
370
- "dialog_act": datasets.Value("string"),
371
- "dialog_act_delexicalized": datasets.Value("string"),
372
- "target_delexicalized": datasets.Value("string"),
373
- "target": datasets.Value("string"),
374
- "references": [datasets.Value("string")],
375
- }
376
- )
377
- elif self.config.name == "dart":
378
- features = datasets.Features(
379
- {
380
- "gem_id": datasets.Value("string"),
381
- "gem_parent_id": datasets.Value("string"),
382
- "dart_id": datasets.Value("int32"),
383
- "tripleset": [[datasets.Value("string")]], # list of triples
384
- "subtree_was_extended": datasets.Value("bool"),
385
- "target_sources": [datasets.Value("string")],
386
- "target": datasets.Value("string"), # single target for train
387
- "references": [datasets.Value("string")],
388
- }
389
- )
390
- elif self.config.name == "e2e_nlg":
391
- features = datasets.Features(
392
- {
393
- "gem_id": datasets.Value("string"),
394
- "gem_parent_id": datasets.Value("string"),
395
- "meaning_representation": datasets.Value("string"),
396
- "target": datasets.Value("string"),
397
- "references": [datasets.Value("string")],
398
- }
399
- )
400
- elif self.config.name.startswith("mlsum"):
401
- features = datasets.Features(
402
- {
403
- "gem_id": datasets.Value("string"),
404
- "gem_parent_id": datasets.Value("string"),
405
- "text": datasets.Value("string"),
406
- "topic": datasets.Value("string"),
407
- "url": datasets.Value("string"),
408
- "title": datasets.Value("string"),
409
- "date": datasets.Value("string"),
410
- "target": datasets.Value("string"),
411
- "references": [datasets.Value("string")],
412
- }
413
- )
414
- elif self.config.name == "schema_guided_dialog":
415
- features = datasets.Features(
416
- {
417
- "gem_id": datasets.Value("string"),
418
- "gem_parent_id": datasets.Value("string"),
419
- "dialog_acts": [
420
- {
421
- "act": datasets.ClassLabel(names=_SGD_ACTS),
422
- "slot": datasets.Value("string"),
423
- "values": [datasets.Value("string")],
424
- }
425
- ],
426
- "context": [datasets.Value("string")],
427
- "dialog_id": datasets.Value("string"),
428
- "service": datasets.Value("string"),
429
- "turn_id": datasets.Value("int32"),
430
- "prompt": datasets.Value("string"),
431
- "target": datasets.Value("string"),
432
- "references": [datasets.Value("string")],
433
- }
434
- )
435
- elif self.config.name == "totto":
436
- features = datasets.Features(
437
- {
438
- "gem_id": datasets.Value("string"),
439
- "gem_parent_id": datasets.Value("string"),
440
- "totto_id": datasets.Value("int32"),
441
- "table_page_title": datasets.Value("string"),
442
- "table_webpage_url": datasets.Value("string"),
443
- "table_section_title": datasets.Value("string"),
444
- "table_section_text": datasets.Value("string"),
445
- "table": [
446
- [
447
- {
448
- "column_span": datasets.Value("int32"),
449
- "is_header": datasets.Value("bool"),
450
- "row_span": datasets.Value("int32"),
451
- "value": datasets.Value("string"),
452
- }
453
- ]
454
- ],
455
- "highlighted_cells": [[datasets.Value("int32")]],
456
- "example_id": datasets.Value("string"),
457
- "sentence_annotations": [
458
- {
459
- "original_sentence": datasets.Value("string"),
460
- "sentence_after_deletion": datasets.Value("string"),
461
- "sentence_after_ambiguity": datasets.Value("string"),
462
- "final_sentence": datasets.Value("string"),
463
- }
464
- ],
465
- "overlap_subset": datasets.Value("string"),
466
- "target": datasets.Value("string"), # single target for train
467
- "references": [datasets.Value("string")],
468
- },
469
- )
470
- elif self.config.name.startswith("web_nlg"):
471
- features = datasets.Features(
472
- {
473
- "gem_id": datasets.Value("string"),
474
- "gem_parent_id": datasets.Value("string"),
475
- "input": [datasets.Value("string")],
476
- "target": datasets.Value("string"), # single target for train
477
- "references": [datasets.Value("string")],
478
- "category": datasets.Value("string"),
479
- "webnlg_id": datasets.Value("string"),
480
- }
481
- )
482
- elif self.config.name == "wiki_auto_asset_turk":
483
- features = datasets.Features(
484
- {
485
- "gem_id": datasets.Value("string"),
486
- "gem_parent_id": datasets.Value("string"),
487
- "source": datasets.Value("string"),
488
- "target": datasets.Value("string"),
489
- "references": [datasets.Value("string")],
490
- }
491
- )
492
- elif self.config.name.startswith("wiki_lingua"):
493
- if "v0" in self.config.name:
494
- features = datasets.Features(
495
- {
496
- "gem_id": datasets.Value("string"),
497
- "gem_parent_id": datasets.Value("string"),
498
- "source": datasets.Value("string"),
499
- "target": datasets.Value("string"),
500
- "references": [datasets.Value("string")],
501
- }
502
- )
503
- else:
504
- ln = self.config.name.split("_")[-1]
505
- features = datasets.Features(
506
- {
507
- "gem_id": datasets.Value("string"),
508
- "gem_parent_id": datasets.Value("string"),
509
- "source_aligned": datasets.Translation(languages=[ln, "en"]),
510
- "target_aligned": datasets.Translation(languages=[ln, "en"]),
511
- "source": datasets.Value("string"),
512
- "target": datasets.Value("string"),
513
- "references": [datasets.Value("string")],
514
- }
515
- )
516
- elif self.config.name == "xsum":
517
- features = datasets.Features(
518
- {
519
- "gem_id": datasets.Value("string"),
520
- "gem_parent_id": datasets.Value("string"),
521
- "xsum_id": datasets.Value("string"),
522
- "document": datasets.Value("string"),
523
- "target": datasets.Value("string"),
524
- "references": [datasets.Value("string")],
525
- }
526
- )
527
- return datasets.DatasetInfo(
528
- description=_DESCRIPTION,
529
- features=features,
530
- supervised_keys=None,
531
- homepage=_HOMEPAGE,
532
- license=_LICENSE,
533
- citation=_CITATION,
534
- )
535
-
536
- def _split_generators(self, dl_manager):
537
- """Returns SplitGenerators."""
538
- dl_dir = dl_manager.download_and_extract(_URLs[self.config.name])
539
- if self.config.name == "common_gen":
540
- challenge_sets = [
541
- ("challenge_train_sample", "train_common_gen_RandomSample500.json"),
542
- ("challenge_validation_sample", "validation_common_gen_RandomSample500.json"),
543
- ("challenge_test_scramble", "test_common_gen_ScrambleInputStructure500.json"),
544
- ]
545
- return [
546
- datasets.SplitGenerator(
547
- name=datasets.Split.TRAIN,
548
- gen_kwargs={
549
- "filepath": os.path.join(dl_dir["data"], "commongen.train.jsonl"),
550
- "split": "train",
551
- },
552
- ),
553
- datasets.SplitGenerator(
554
- name=datasets.Split.VALIDATION,
555
- gen_kwargs={
556
- "filepath": os.path.join(dl_dir["data"], "commongen.dev.jsonl"),
557
- "split": "validation",
558
- },
559
- ),
560
- datasets.SplitGenerator(
561
- name=datasets.Split.TEST,
562
- gen_kwargs={
563
- "filepath": os.path.join(dl_dir["data"], "commongen.test_noref.jsonl"),
564
- "split": "test",
565
- },
566
- ),
567
- ] + [
568
- datasets.SplitGenerator(
569
- name=challenge_split,
570
- gen_kwargs={
571
- "filepath": os.path.join(dl_dir["challenge_set"], self.config.name, filename),
572
- "split": challenge_split,
573
- },
574
- )
575
- for challenge_split, filename in challenge_sets
576
- ]
577
- elif self.config.name == "cs_restaurants":
578
- challenge_sets = [
579
- ("challenge_train_sample", "train_cs_restaurants_RandomSample500.json"),
580
- ("challenge_validation_sample", "validation_cs_restaurants_RandomSample500.json"),
581
- ("challenge_test_scramble", "test_cs_restaurants_ScrambleInputStructure500.json"),
582
- ]
583
- return [
584
- datasets.SplitGenerator(name=spl, gen_kwargs={"filepath": dl_dir[spl], "split": spl})
585
- for spl in ["train", "validation", "test"]
586
- ] + [
587
- datasets.SplitGenerator(
588
- name=challenge_split,
589
- gen_kwargs={
590
- "filepath": os.path.join(dl_dir["challenge_set"], self.config.name, filename),
591
- "split": challenge_split,
592
- },
593
- )
594
- for challenge_split, filename in challenge_sets
595
- ]
596
- elif self.config.name == "dart":
597
- return [
598
- datasets.SplitGenerator(name=spl, gen_kwargs={"filepath": dl_dir[spl], "split": spl})
599
- for spl in ["train", "validation", "test"]
600
- ]
601
- elif self.config.name == "e2e_nlg":
602
- challenge_sets = [
603
- ("challenge_train_sample", "train_e2e_nlg_RandomSample500.json"),
604
- ("challenge_validation_sample", "validation_e2e_nlg_RandomSample500.json"),
605
- ("challenge_test_scramble", "test_e2e_nlg_ScrambleInputStructure500.json"),
606
- ]
607
- return [
608
- datasets.SplitGenerator(name=spl, gen_kwargs={"filepath": dl_dir[spl], "split": spl})
609
- for spl in ["train", "validation", "test"]
610
- ] + [
611
- datasets.SplitGenerator(
612
- name=challenge_split,
613
- gen_kwargs={
614
- "filepath": os.path.join(dl_dir["challenge_set"], self.config.name, filename),
615
- "split": challenge_split,
616
- },
617
- )
618
- for challenge_split, filename in challenge_sets
619
- ]
620
- elif self.config.name.startswith("mlsum"):
621
- lang = self.config.name.split("_")[1]
622
- challenge_sets = [
623
- ("challenge_train_sample", f"train_mlsum_{lang}_RandomSample500.json"),
624
- ("challenge_validation_sample", f"validation_mlsum_{lang}_RandomSample500.json"),
625
- ("challenge_test_covid", f"{lang}_test_covid19_cleaned.jsonl"),
626
- ]
627
- return [
628
- datasets.SplitGenerator(
629
- name=datasets.Split.TRAIN,
630
- gen_kwargs={
631
- "filepath": os.path.join(dl_dir["train"], lang + "_train.jsonl"),
632
- "split": "train",
633
- "lang": lang,
634
- "filepaths": dl_dir["bad_ids"],
635
- },
636
- ),
637
- datasets.SplitGenerator(
638
- name=datasets.Split.VALIDATION,
639
- gen_kwargs={
640
- "filepath": os.path.join(dl_dir["validation"], lang + "_val.jsonl"),
641
- "split": "validation",
642
- "lang": lang,
643
- "filepaths": dl_dir["bad_ids"],
644
- },
645
- ),
646
- datasets.SplitGenerator(
647
- name=datasets.Split.TEST,
648
- gen_kwargs={
649
- "filepath": os.path.join(dl_dir["test"], lang + "_test.jsonl"),
650
- "split": "test",
651
- "lang": lang,
652
- "filepaths": dl_dir["bad_ids"],
653
- },
654
- ),
655
- ] + [
656
- datasets.SplitGenerator(
657
- name=challenge_split,
658
- gen_kwargs={
659
- "filepath": os.path.join(dl_dir["challenge_set"], self.config.name, filename),
660
- "split": challenge_split,
661
- },
662
- )
663
- for challenge_split, filename in challenge_sets
664
- ]
665
- elif self.config.name == "schema_guided_dialog":
666
- challenge_sets = [
667
- ("challenge_train_sample", "train_schema_guided_dialog_RandomSample500_reformatted.json"),
668
- ("challenge_validation_sample", "validation_schema_guided_dialog_RandomSample500_reformatted.json"),
669
- ("challenge_test_backtranslation", "test_schema_guided_dialog_BackTranslation500_reformatted.json"),
670
- (
671
- "challenge_test_bfp02",
672
- "test_schema_guided_dialog_ButterFingersPerturbation_p=0.02_500_reformatted.json",
673
- ),
674
- (
675
- "challenge_test_bfp05",
676
- "test_schema_guided_dialog_ButterFingersPerturbation_p=0.05_500_reformatted.json",
677
- ),
678
- ("challenge_test_nopunc", "test_schema_guided_dialog_WithoutPunctuation500_reformatted.json"),
679
- ("challenge_test_scramble", "test_schema_guided_dialog_ScrambleInputStructure500_reformatted.json"),
680
- ]
681
- return [
682
- datasets.SplitGenerator(
683
- name=spl, gen_kwargs={"filepath": os.path.join(dl_dir["data"], "gem_sgd.json"), "split": spl}
684
- )
685
- for spl in ["train", "validation", "test"]
686
- ] + [
687
- datasets.SplitGenerator(
688
- name=challenge_split,
689
- gen_kwargs={
690
- "filepath": os.path.join(dl_dir["challenge_set"], self.config.name, filename),
691
- "split": challenge_split,
692
- },
693
- )
694
- for challenge_split, filename in challenge_sets
695
- ]
696
- elif self.config.name == "totto":
697
- challenge_sets = [
698
- ("challenge_train_sample", "train_totto_RandomSample500.json"),
699
- ("challenge_validation_sample", "validation_totto_RandomSample500.json"),
700
- ("challenge_test_scramble", "test_totto_ScrambleInputStructure500.json"),
701
- ]
702
- return [
703
- datasets.SplitGenerator(
704
- name=datasets.Split.TRAIN,
705
- gen_kwargs={
706
- "filepath": os.path.join(dl_dir["data"], "totto_data/totto_train_data.jsonl"),
707
- "split": "train",
708
- },
709
- ),
710
- datasets.SplitGenerator(
711
- name=datasets.Split.VALIDATION,
712
- gen_kwargs={
713
- "filepath": os.path.join(dl_dir["data"], "totto_data/totto_dev_data.jsonl"),
714
- "split": "validation",
715
- },
716
- ),
717
- datasets.SplitGenerator(
718
- name=datasets.Split.TEST,
719
- gen_kwargs={
720
- "filepath": os.path.join(dl_dir["data"], "totto_data/unlabeled_totto_test_data.jsonl"),
721
- "split": "test",
722
- },
723
- ),
724
- ] + [
725
- datasets.SplitGenerator(
726
- name=challenge_split,
727
- gen_kwargs={
728
- "filepath": os.path.join(dl_dir["challenge_set"], self.config.name, filename),
729
- "split": challenge_split,
730
- },
731
- )
732
- for challenge_split, filename in challenge_sets
733
- ]
734
- elif self.config.name.startswith("web_nlg"):
735
- ln = self.config.name.split("_")[-1]
736
- challenge_sets = [
737
- ("challenge_train_sample", f"train_web_nlg_{ln}_RandomSample500.json"),
738
- ("challenge_validation_sample", f"validation_web_nlg_{ln}_RandomSample500.json"),
739
- ("challenge_test_scramble", f"test_web_nlg_{ln}_ScrambleInputStructure500.json"),
740
- ]
741
- if ln == "en":
742
- challenge_sets += [("challenge_test_numbers", f"test_web_nlg_{ln}_replace_numbers_500.json")]
743
- return [
744
- datasets.SplitGenerator(name=spl, gen_kwargs={"filepath": dl_dir[spl], "split": spl})
745
- for spl in ["train", "validation", "test"]
746
- ] + [
747
- datasets.SplitGenerator(
748
- name=challenge_split,
749
- gen_kwargs={
750
- "filepath": os.path.join(dl_dir["challenge_set"], self.config.name, filename),
751
- "split": challenge_split,
752
- },
753
- )
754
- for challenge_split, filename in challenge_sets
755
- ]
756
- elif self.config.name == "wiki_auto_asset_turk":
757
- challenge_sets = [
758
- ("challenge_train_sample", "train_wiki_auto_asset_turk_RandomSample500.json"),
759
- ("challenge_validation_sample", "validation_wiki_auto_asset_turk_RandomSample500.json"),
760
- ("challenge_test_asset_backtranslation", "test_asset_wiki_auto_asset_turk_BackTranslation.json"),
761
- (
762
- "challenge_test_asset_bfp02",
763
- "test_asset_wiki_auto_asset_turk_ButterFingersPerturbation_p=0.02.json",
764
- ),
765
- (
766
- "challenge_test_asset_bfp05",
767
- "test_asset_wiki_auto_asset_turk_ButterFingersPerturbation_p=0.05.json",
768
- ),
769
- ("challenge_test_asset_nopunc", "test_asset_wiki_auto_asset_turk_WithoutPunctuation.json"),
770
- ("challenge_test_turk_backtranslation", "detok_test_turk_wiki_auto_asset_turk_BackTranslation.json"),
771
- (
772
- "challenge_test_turk_bfp02",
773
- "detok_test_turk_wiki_auto_asset_turk_ButterFingersPerturbation_p=0.02.json",
774
- ),
775
- (
776
- "challenge_test_turk_bfp05",
777
- "detok_test_turk_wiki_auto_asset_turk_ButterFingersPerturbation_p=0.05.json",
778
- ),
779
- ("challenge_test_turk_nopunc", "detok_test_turk_wiki_auto_asset_turk_WithoutPunctuation.json"),
780
- ]
781
- return [
782
- datasets.SplitGenerator(
783
- name=datasets.Split.TRAIN,
784
- gen_kwargs={
785
- "filepath": dl_dir["train"],
786
- "split": "train",
787
- },
788
- ),
789
- datasets.SplitGenerator(
790
- name=datasets.Split.VALIDATION,
791
- gen_kwargs={
792
- "filepath": dl_dir["validation"],
793
- "split": "validation",
794
- },
795
- ),
796
- datasets.SplitGenerator(
797
- name="test_asset",
798
- gen_kwargs={
799
- "filepath": "",
800
- "split": "test_asset",
801
- "filepaths": [dl_dir["test_asset_orig"]] + [dl_dir[f"test_asset_{i}"] for i in range(10)],
802
- },
803
- ),
804
- datasets.SplitGenerator(
805
- name="test_turk",
806
- gen_kwargs={
807
- "filepath": dl_dir["test_turk"],
808
- "split": "test_turk",
809
- },
810
- ),
811
- ] + [
812
- datasets.SplitGenerator(
813
- name=challenge_split,
814
- gen_kwargs={
815
- "filepath": os.path.join(dl_dir["challenge_set"], "wiki_auto_asset_turk", filename),
816
- "split": challenge_split,
817
- },
818
- )
819
- for challenge_split, filename in challenge_sets
820
- ]
821
- elif self.config.name.startswith("wiki_lingua"):
822
- if "v0" in self.config.name:
823
- lang = self.config.name.split("_")[-3]
824
- base_dir = os.path.join(dl_dir["data"], "GEM_data_crosslingual", f"{lang}_en")
825
- return [
826
- datasets.SplitGenerator(
827
- name=datasets.Split.TRAIN,
828
- gen_kwargs={
829
- "filepath": base_dir,
830
- "split": "train",
831
- },
832
- ),
833
- datasets.SplitGenerator(
834
- name=datasets.Split.VALIDATION,
835
- gen_kwargs={
836
- "filepath": base_dir,
837
- "split": "val",
838
- },
839
- ),
840
- datasets.SplitGenerator(
841
- name=datasets.Split.TEST,
842
- gen_kwargs={
843
- "filepath": base_dir,
844
- "split": "test",
845
- },
846
- ),
847
- ]
848
- else:
849
- lang_name = self.config.name.split("_")[-2]
850
- lang = self.config.name.split("_")[-1]
851
- base_dir = os.path.join(dl_dir["data"], lang_name)
852
- return [
853
- datasets.SplitGenerator(
854
- name=datasets.Split.TRAIN,
855
- gen_kwargs={
856
- "filepath": base_dir,
857
- "split": "train",
858
- "lang": lang,
859
- },
860
- ),
861
- datasets.SplitGenerator(
862
- name=datasets.Split.VALIDATION,
863
- gen_kwargs={
864
- "filepath": base_dir,
865
- "split": "val",
866
- "lang": lang,
867
- },
868
- ),
869
- datasets.SplitGenerator(
870
- name=datasets.Split.TEST,
871
- gen_kwargs={
872
- "filepath": base_dir,
873
- "split": "test",
874
- "lang": lang,
875
- },
876
- ),
877
- ]
878
- elif self.config.name == "xsum":
879
- challenge_sets = [
880
- ("challenge_train_sample", "train_xsum_RandomSample500.json"),
881
- ("challenge_validation_sample", "validation_xsum_RandomSample500.json"),
882
- ("challenge_test_backtranslation", "test_xsum_BackTranslation500.json"),
883
- ("challenge_test_bfp_02", "test_xsum_ButterFingersPerturbation_p=0.02_500.json"),
884
- ("challenge_test_bfp_05", "test_xsum_ButterFingersPerturbation_p=0.05_500.json"),
885
- ("challenge_test_nopunc", "test_xsum_WithoutPunctuation500.json"),
886
- ("challenge_test_covid", "en_test_covid19.jsonl"),
887
- ]
888
- return [
889
- datasets.SplitGenerator(
890
- name=datasets.Split.TRAIN,
891
- gen_kwargs={
892
- "filepath": dl_dir["splits"],
893
- "split": "train",
894
- "filepaths": os.path.join(dl_dir["data"], "bbc-summary-data"),
895
- },
896
- ),
897
- datasets.SplitGenerator(
898
- name=datasets.Split.VALIDATION,
899
- gen_kwargs={
900
- "filepath": dl_dir["splits"],
901
- "split": "validation",
902
- "filepaths": os.path.join(dl_dir["data"], "bbc-summary-data"),
903
- },
904
- ),
905
- datasets.SplitGenerator(
906
- name=datasets.Split.TEST,
907
- gen_kwargs={
908
- "filepath": dl_dir["splits"],
909
- "split": "test",
910
- "filepaths": os.path.join(dl_dir["data"], "bbc-summary-data"),
911
- },
912
- ),
913
- ] + [
914
- datasets.SplitGenerator(
915
- name=challenge_split,
916
- gen_kwargs={
917
- "filepath": os.path.join(dl_dir["challenge_set"], "xsum", filename),
918
- "split": challenge_split,
919
- },
920
- )
921
- for challenge_split, filename in challenge_sets
922
- ]
923
-
924
- def _generate_examples(self, filepath, split, filepaths=None, lang=None):
925
- """Yields examples."""
926
- if self.config.name == "common_gen":
927
- if split.startswith("challenge"):
928
- exples = json.load(open(filepath, encoding="utf-8"))
929
- if isinstance(exples, dict):
930
- assert len(exples) == 1, "multiple entries found"
931
- exples = list(exples.values())[0]
932
- for id_, exple in enumerate(exples):
933
- if len(exple) == 0:
934
- continue
935
- exple["gem_parent_id"] = exple["gem_id"]
936
- exple["gem_id"] = f"{self.config.name}-{split}-{id_}"
937
- yield id_, exple
938
- else:
939
- with open(filepath, encoding="utf-8") as f:
940
- id_ = -1
941
- i = -1
942
- for row in f:
943
- row = row.replace(", }", "}") # Fix possible JSON format error
944
- data = json.loads(row)
945
- concepts = [word for word in data["concept_set"].split("#")]
946
- if split == "train":
947
- i += 1
948
- for scene in data["scene"]:
949
- id_ += 1
950
- yield id_, {
951
- "gem_id": f"{self.config.name}-{split}-{id_}",
952
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
953
- "concept_set_id": i,
954
- "concepts": concepts,
955
- "target": scene,
956
- "references": [],
957
- }
958
- else:
959
- id_ += 1
960
- yield id_, {
961
- "gem_id": f"{self.config.name}-{split}-{id_}",
962
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
963
- "concept_set_id": id_,
964
- "concepts": concepts,
965
- "target": "" if split == "test" else data["scene"][0],
966
- "references": [] if split == "test" else data["scene"],
967
- }
968
- elif self.config.name == "cs_restaurants":
969
- if split.startswith("challenge"):
970
- exples = json.load(open(filepath, encoding="utf-8"))
971
- if isinstance(exples, dict):
972
- assert len(exples) == 1, "multiple entries found"
973
- exples = list(exples.values())[0]
974
- for id_, exple in enumerate(exples):
975
- if len(exple) == 0:
976
- continue
977
- exple["gem_parent_id"] = exple["gem_id"]
978
- exple["gem_id"] = f"{self.config.name}-{split}-{id_}"
979
- yield id_, exple
980
- else:
981
- with open(filepath, encoding="utf8") as f:
982
- data = json.load(f)
983
- for id_, instance in enumerate(data):
984
- yield id_, {
985
- "gem_id": f"{self.config.name}-{split}-{id_}",
986
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
987
- "dialog_act": instance["da"],
988
- "dialog_act_delexicalized": instance["delex_da"],
989
- "target": instance["text"],
990
- "target_delexicalized": instance["delex_text"],
991
- "references": [] if split == "train" else [instance["text"]],
992
- }
993
- elif self.config.name == "dart":
994
- with open(filepath, encoding="utf-8") as f:
995
- data = json.loads(f.read())
996
- id_ = -1
997
- i = -1
998
- for example in data:
999
- if split == "train":
1000
- i += 1
1001
- for annotation in example["annotations"]:
1002
- id_ += 1
1003
- yield id_, {
1004
- "gem_id": f"{self.config.name}-{split}-{id_}",
1005
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
1006
- "dart_id": i,
1007
- "tripleset": example["tripleset"],
1008
- "subtree_was_extended": example.get("subtree_was_extended", None), # some are missing
1009
- "target_sources": [annotation["source"] for annotation in example["annotations"]],
1010
- "target": annotation["text"],
1011
- "references": [],
1012
- }
1013
- else:
1014
- id_ += 1
1015
- yield id_, {
1016
- "gem_id": f"{self.config.name}-{split}-{id_}",
1017
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
1018
- "dart_id": id_,
1019
- "tripleset": example["tripleset"],
1020
- "subtree_was_extended": example.get("subtree_was_extended", None), # some are missing
1021
- "target_sources": [annotation["source"] for annotation in example["annotations"]],
1022
- "target": example["annotations"][0]["text"] if len(example["annotations"]) > 0 else "",
1023
- "references": [annotation["text"] for annotation in example["annotations"]],
1024
- }
1025
- elif self.config.name == "e2e_nlg":
1026
- if split.startswith("challenge"):
1027
- exples = json.load(open(filepath, encoding="utf-8"))
1028
- if isinstance(exples, dict):
1029
- assert len(exples) == 1, "multiple entries found"
1030
- exples = list(exples.values())[0]
1031
- for id_, exple in enumerate(exples):
1032
- if len(exple) == 0:
1033
- continue
1034
- exple["gem_parent_id"] = exple["gem_id"]
1035
- exple["gem_id"] = f"{self.config.name}-{split}-{id_}"
1036
- yield id_, exple
1037
- else:
1038
- with open(filepath, encoding="utf-8") as f:
1039
- reader = csv.DictReader(f)
1040
- for id_, example in enumerate(reader):
1041
- yield id_, {
1042
- "gem_id": f"{self.config.name}-{split}-{id_}",
1043
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
1044
- "meaning_representation": example["mr"],
1045
- "target": example["ref"],
1046
- "references": [] if split == "train" else [example["ref"]],
1047
- }
1048
- elif self.config.name.startswith("mlsum"):
1049
- if split in ["train", "validation", "test", "challenge_test_covid"]:
1050
- if split == "challenge_test_covid":
1051
- bad_ids = {}
1052
- else:
1053
- bad_ids_dct = json.load(open(filepaths, encoding="utf-8"))
1054
- bad_ids = dict((bad_url, True) for _, bad_url in bad_ids_dct[f"{lang}-{split}"])
1055
- with open(filepath, encoding="utf-8") as f:
1056
- id_ = -1
1057
- for line in f:
1058
- data = json.loads(line)
1059
- if data["url"] in bad_ids:
1060
- continue
1061
- else:
1062
- id_ += 1
1063
- yield id_, {
1064
- "gem_id": f"{self.config.name}-{split}-{id_}",
1065
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
1066
- "text": data["text"],
1067
- "target": data["summary"],
1068
- "references": [] if split == "train" else [data["summary"]],
1069
- "topic": data["topic"],
1070
- "url": data["url"],
1071
- "title": data["title"],
1072
- "date": data["date"],
1073
- }
1074
- else:
1075
- exples = json.load(open(filepath, encoding="utf-8"))
1076
- if isinstance(exples, dict):
1077
- assert len(exples) == 1, "multiple entries found"
1078
- exples = list(exples.values())[0]
1079
- for id_, exple in enumerate(exples):
1080
- if len(exple) == 0:
1081
- continue
1082
- exple["gem_parent_id"] = exple["gem_id"]
1083
- exple["gem_id"] = f"{self.config.name}-{split}-{id_}"
1084
- yield id_, exple
1085
- elif self.config.name == "schema_guided_dialog":
1086
- if "challenge" in split:
1087
- exples = json.load(open(filepath, encoding="utf-8"))
1088
- if isinstance(exples, dict):
1089
- assert len(exples) == 1, "multiple entries found"
1090
- exples = list(exples.values())[0]
1091
- for id_, exple in enumerate(exples):
1092
- if len(exple) == 0:
1093
- continue
1094
- exple["gem_parent_id"] = exple["gem_id"]
1095
- exple["gem_id"] = f"{self.config.name}-{split}-{id_}"
1096
- yield id_, exple
1097
- else:
1098
- examples = json.load(open(filepath, encoding="utf-8"))[split]
1099
- for id_, example in enumerate(examples):
1100
- yield id_, {
1101
- "gem_id": f"{self.config.name}-{split}-{id_}",
1102
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
1103
- "dialog_acts": [
1104
- {
1105
- "act": act_id,
1106
- "slot": slot,
1107
- "values": values,
1108
- }
1109
- for act_id, slot, values in example["da"]
1110
- ],
1111
- "context": example["context"],
1112
- "dialog_id": example["dialog_id"],
1113
- "service": example["service"],
1114
- "turn_id": example["turn_ix"],
1115
- "prompt": example["prompt"],
1116
- "target": example["target"],
1117
- "references": [] if split == "train" else [example["target"]],
1118
- }
1119
- elif self.config.name == "totto":
1120
- if "challenge" in split:
1121
- exples = json.load(open(filepath, encoding="utf-8"))
1122
- if isinstance(exples, dict):
1123
- assert len(exples) == 1, "multiple entries found"
1124
- exples = list(exples.values())[0]
1125
- for id_, exple in enumerate(exples):
1126
- if len(exple) == 0:
1127
- continue
1128
- exple["gem_parent_id"] = exple["gem_id"]
1129
- exple["gem_id"] = f"{self.config.name}-{split}-{id_}"
1130
- yield id_, exple
1131
- else:
1132
- with open(filepath, "r", encoding="utf-8") as json_file:
1133
- json_list = list(json_file)
1134
- id_ = -1
1135
- i = -1
1136
- for json_str in json_list:
1137
- result = json.loads(json_str)
1138
- if split == "train":
1139
- i += 1
1140
- for sentence in result["sentence_annotations"]:
1141
- id_ += 1
1142
- response = {
1143
- "gem_id": f"{self.config.name}-{split}-{id_}",
1144
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
1145
- "totto_id": i,
1146
- "table_page_title": result["table_page_title"],
1147
- "table_webpage_url": result["table_webpage_url"],
1148
- "table_section_title": result["table_section_title"],
1149
- "table_section_text": result["table_section_text"],
1150
- "table": result["table"],
1151
- "highlighted_cells": result["highlighted_cells"],
1152
- "example_id": str(result["example_id"]),
1153
- "overlap_subset": "none",
1154
- "sentence_annotations": [sentence],
1155
- "references": [],
1156
- "target": sentence["final_sentence"],
1157
- }
1158
- yield id_, response
1159
- else:
1160
- id_ += 1
1161
- response = {
1162
- "gem_id": f"{self.config.name}-{split}-{id_}",
1163
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
1164
- "totto_id": id_,
1165
- "table_page_title": result["table_page_title"],
1166
- "table_webpage_url": result["table_webpage_url"],
1167
- "table_section_title": result["table_section_title"],
1168
- "table_section_text": result["table_section_text"],
1169
- "table": result["table"],
1170
- "highlighted_cells": result["highlighted_cells"],
1171
- "example_id": str(result["example_id"]),
1172
- "overlap_subset": str(result["overlap_subset"]),
1173
- "sentence_annotations": [] if split == "test" else result["sentence_annotations"],
1174
- }
1175
- response["references"] = [
1176
- sentence["final_sentence"] for sentence in response["sentence_annotations"]
1177
- ]
1178
- response["target"] = response["references"][0] if len(response["references"]) > 0 else ""
1179
- yield id_, response
1180
- elif self.config.name.startswith("web_nlg"):
1181
- if "challenge" in split:
1182
- exples = json.load(open(filepath, encoding="utf-8"))
1183
- if isinstance(exples, dict):
1184
- assert len(exples) == 1, "multiple entries found"
1185
- exples = list(exples.values())[0]
1186
- for id_, exple in enumerate(exples):
1187
- if len(exple) == 0:
1188
- continue
1189
- exple["gem_parent_id"] = exple["gem_id"]
1190
- exple["gem_id"] = f"{self.config.name}-{split}-{id_}"
1191
- yield id_, exple
1192
- else:
1193
- with open(filepath, encoding="utf-8") as f:
1194
- examples = json.load(f)
1195
- id_ = -1
1196
- for example in examples["values"]:
1197
- if split == "train":
1198
- for target in example["target"]:
1199
- id_ += 1
1200
- yield id_, {
1201
- "gem_id": f"{self.config.name}-{split}-{id_}",
1202
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
1203
- "input": example["input"],
1204
- "target": target,
1205
- "references": [] if split == "train" else example["target"],
1206
- "category": example["category"],
1207
- "webnlg_id": example["webnlg-id"],
1208
- }
1209
- else:
1210
- id_ += 1
1211
- yield id_, {
1212
- "gem_id": f"{self.config.name}-{split}-{id_}",
1213
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
1214
- "input": example["input"],
1215
- "target": example["target"][0] if len(example["target"]) > 0 else "",
1216
- "references": example["target"],
1217
- "category": example["category"],
1218
- "webnlg_id": example["webnlg-id"],
1219
- }
1220
- elif self.config.name == "wiki_auto_asset_turk":
1221
- if split in ["train", "validation"]:
1222
- keys = [
1223
- "source",
1224
- "target",
1225
- ]
1226
- with open(filepath, encoding="utf-8") as f:
1227
- for id_, line in enumerate(f):
1228
- values = line.strip().split("\t")
1229
- assert len(values) == 2, f"Not enough fields in ---- {line} --- {values}"
1230
- example = dict([(k, val) for k, val in zip(keys, values)])
1231
- example["gem_id"] = f"{self.config.name}-{split}-{id_}"
1232
- example["gem_parent_id"] = example["gem_id"]
1233
- example["references"] = [] if split == "train" else [example["target"]]
1234
- yield id_, example
1235
- elif split == "test_turk":
1236
- examples = json.load(open(filepath, encoding="utf-8"))
1237
- for id_, example in enumerate(examples):
1238
- example["gem_parent_id"] = example["gem_id"]
1239
- for k in ["source_id", "target_id"]:
1240
- if k in example:
1241
- del example[k]
1242
- yield id_, example
1243
- elif split == "test_asset":
1244
- files = [open(f_name, encoding="utf-8") for f_name in filepaths]
1245
- for id_, lines in enumerate(zip(*files)):
1246
- yield id_, {
1247
- "gem_id": f"{self.config.name}-{split}-{id_}",
1248
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
1249
- "target": lines[1].strip(),
1250
- "source": lines[0].strip(),
1251
- "references": [line.strip() for line in lines[1:]],
1252
- }
1253
- else:
1254
- exples = json.load(open(filepath, encoding="utf-8"))
1255
- if isinstance(exples, dict):
1256
- assert len(exples) == 1, "multiple entries found"
1257
- exples = list(exples.values())[0]
1258
- for id_, exple in enumerate(exples):
1259
- exple["gem_parent_id"] = exple["gem_id"]
1260
- exple["gem_id"] = f"{self.config.name}-{split}-{id_}"
1261
- for k in ["source_id", "target_id"]:
1262
- if k in exple:
1263
- del exple[k]
1264
- yield id_, exple
1265
- elif self.config.name.startswith("wiki_lingua"):
1266
- if "v0" in self.config.name:
1267
- with open(os.path.join(filepath, f"{split}.src"), encoding="utf-8") as f_in:
1268
- with open(os.path.join(filepath, f"{split}.tgt"), encoding="utf-8") as f_out:
1269
- for id_, (src, tgt) in enumerate(zip(f_in, f_out)):
1270
- yield id_, {
1271
- "gem_id": f"{self.config.name}-{split}-{id_}",
1272
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
1273
- "source": src.strip(),
1274
- "target": tgt.strip(),
1275
- "references": [] if split == "train" else [tgt.strip()],
1276
- }
1277
- else:
1278
- with open(os.path.join(filepath, f"{split}.src.{lang}"), encoding="utf-8") as f_in_ln:
1279
- with open(os.path.join(filepath, f"{split}.src.en"), encoding="utf-8") as f_in_en:
1280
- with open(os.path.join(filepath, f"{split}.tgt.{lang}"), encoding="utf-8") as f_out_ln:
1281
- with open(os.path.join(filepath, f"{split}.tgt.en"), encoding="utf-8") as f_out_en:
1282
- for id_, (src_ln, src_en, tgt_ln, tgt_en) in enumerate(
1283
- zip(f_in_ln, f_in_en, f_out_ln, f_out_en)
1284
- ):
1285
- yield id_, {
1286
- "gem_id": f"{self.config.name}-{split}-{id_}",
1287
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
1288
- "source_aligned": {lang: src_ln.strip(), "en": src_en.strip()},
1289
- "target_aligned": {lang: tgt_ln.strip(), "en": tgt_en.strip()},
1290
- "source": src_ln.strip(),
1291
- "target": tgt_en.strip(),
1292
- "references": [] if split == "train" else [tgt_en.strip()],
1293
- }
1294
- elif self.config.name == "xsum":
1295
- if "challenge" in split:
1296
- if "covid" in split:
1297
- with open(filepath, encoding="utf-8") as f:
1298
- id_ = -1
1299
- for line in f:
1300
- data = json.loads(line)
1301
- id_ += 1
1302
- yield id_, {
1303
- "gem_id": f"{self.config.name}-{split}-{id_}",
1304
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
1305
- "xsum_id": data["url"],
1306
- "document": data["text"],
1307
- "target": data["summary"],
1308
- "references": [] if split == "train" else [data["summary"]],
1309
- }
1310
- else:
1311
- exples = json.load(open(filepath, encoding="utf-8"))
1312
- if isinstance(exples, dict):
1313
- assert len(exples) == 1, "multiple entries found"
1314
- exples = list(exples.values())[0]
1315
- for id_, exple in enumerate(exples):
1316
- exple["gem_parent_id"] = exple["gem_id"]
1317
- exple["gem_id"] = f"{self.config.name}-{split}-{id_}"
1318
- yield id_, exple
1319
- else:
1320
- with open(filepath, "r", encoding="utf-8") as f:
1321
- split_ids = json.load(f)
1322
- for id_, i in enumerate(split_ids[split]):
1323
- with open(os.path.join(filepaths, i + ".summary"), "r", encoding="utf-8") as f:
1324
- text = "".join(
1325
- [line for line in f.readlines() if line not in _XSUM_REMOVE_LINES and line.strip()]
1326
- )
1327
- segs = text.split("[SN]")
1328
- yield id_, {
1329
- "gem_id": f"{self.config.name}-{split}-{id_}",
1330
- "gem_parent_id": f"{self.config.name}-{split}-{id_}",
1331
- "xsum_id": i,
1332
- "document": segs[8].strip(),
1333
- "target": segs[6].strip(),
1334
- "references": [] if split == "train" else [segs[6].strip()],
1335
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
wiki_lingua_portuguese_pt/gem-test.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a00c5f658cd7f48ae8ae20b4ef3c769a6566710fcd23a5bbecd7f20b5bffc61
3
+ size 69554025
wiki_lingua_portuguese_pt/gem-train.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e437bc350d159790ed897077d37b421026e3832e424b0edb3e5d7bd91684e53e
3
+ size 238868335
wiki_lingua_portuguese_pt/gem-validation.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c2f9ef60fd3b409580df8d16b4e0ec14cabcc6f0ac180f27a65923c4395060b
3
+ size 35266771