albertvillanova HF staff commited on
Commit
5423c47
1 Parent(s): 246cdd1

Convert dataset to Parquet (#3)

Browse files

- Convert dataset to Parquet (cec9478cc7c05d52698f1604093557df4d6171f7)
- Add alignments data files (b878bfae7771fe6bb41485b63a32d81754566f4a)
- Add multilingual data files (2cfaf4eff8155be45690bdf3bca0652e487822e9)
- Add crosslingual_test data files (6d9e778a205a1897dca512c76e8ecca1e259e36e)
- Add crosslingual_with_para_test data files (b97f2de1292ae913dce46d5267429508a0f06743)
- Add crosslingual_bg data files (4d9ca37028471b6128a13dec0c741f215e65694e)
- Add crosslingual_with_para_bg data files (7e09f78756b88508c64b7928ae001feb64656c03)
- Add crosslingual_hr data files (8c0039c33cdad7058278726329360b161d2a4604)
- Add crosslingual_with_para_hr data files (95a9b70e7325bd7cddb1f87923128a83781e3d18)
- Add crosslingual_hu data files (d168bbe928b23acf6588975f472b90198a4ac79c)
- Add crosslingual_with_para_hu data files (23c3a4a36186feaa3bce66b62ed372b2fc99fb95)
- Add crosslingual_it data files (6b1f4321bc5951692c8d9786da37daf79ab5998d)
- Add crosslingual_with_para_it data files (71da5a021089ac6ec76d913bd2270625eb9e572f)
- Add crosslingual_mk data files (4026bb96a56bb3fd529bd5e15b7a512a57bacf81)
- Add crosslingual_with_para_mk data files (893fa7f8f9d215e6cad95144281680fb5c041ed1)
- Add crosslingual_pl data files (efd4f02499b28e24dac9b437f870638db0913bc3)
- Add crosslingual_with_para_pl data files (dede1311f6a38aa9b81bc9502044f1c2e1a4a9d1)
- Add crosslingual_pt data files (6d47a82192c560aa455475c3f2c32fb24dbc92e7)
- Add crosslingual_with_para_pt data files (db23686ea2fc8e9be8dbcaf1191c9ce35c73f247)
- Add crosslingual_sq data files (ac9f5fc03d4a4563a8e2e4e0b3d418d7c81309da)
- Add crosslingual_with_para_sq data files (d14dcee3528b18d10a22d79d701bc5d0d2efb37e)
- Add crosslingual_sr data files (7b54a9839cc9d275095499cfe6a05a07117df42d)
- Add crosslingual_with_para_sr data files (9018cb3a52630cf0778eee025204a58db1338aa6)
- Add crosslingual_tr data files (33786108438f8662df5de918a347001d144fe393)
- Add crosslingual_with_para_tr data files (25f8b66a9c4599bb80dfa839f2cc67c700fa249f)
- Add crosslingual_vi data files (902ea15c17727c366dc467720ead35afe5a984f5)
- Add crosslingual_with_para_vi data files (b074c819cd69caf659f1f7530f6f147171b0ab3b)
- Delete loading script (d5e1adb6337466d7abb0588ccffe996074b53ac3)
- Delete legacy dataset_infos.json (ae496195d6100627278b2f721eff48d7fd91b244)

This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. README.md +377 -215
  2. alignments/full-00000-of-00001.parquet +3 -0
  3. crosslingual_bg/train-00000-of-00001.parquet +3 -0
  4. crosslingual_bg/validation-00000-of-00001.parquet +3 -0
  5. crosslingual_hr/train-00000-of-00001.parquet +3 -0
  6. crosslingual_hr/validation-00000-of-00001.parquet +3 -0
  7. crosslingual_hu/train-00000-of-00001.parquet +3 -0
  8. crosslingual_hu/validation-00000-of-00001.parquet +3 -0
  9. crosslingual_it/train-00000-of-00001.parquet +3 -0
  10. crosslingual_it/validation-00000-of-00001.parquet +3 -0
  11. crosslingual_mk/train-00000-of-00001.parquet +3 -0
  12. crosslingual_mk/validation-00000-of-00001.parquet +3 -0
  13. crosslingual_pl/train-00000-of-00001.parquet +3 -0
  14. crosslingual_pl/validation-00000-of-00001.parquet +3 -0
  15. crosslingual_pt/train-00000-of-00001.parquet +3 -0
  16. crosslingual_pt/validation-00000-of-00001.parquet +3 -0
  17. crosslingual_sq/train-00000-of-00001.parquet +3 -0
  18. crosslingual_sq/validation-00000-of-00001.parquet +3 -0
  19. crosslingual_sr/train-00000-of-00001.parquet +3 -0
  20. crosslingual_sr/validation-00000-of-00001.parquet +3 -0
  21. crosslingual_test/test-00000-of-00001.parquet +3 -0
  22. crosslingual_tr/train-00000-of-00001.parquet +3 -0
  23. crosslingual_tr/validation-00000-of-00001.parquet +3 -0
  24. crosslingual_vi/train-00000-of-00001.parquet +3 -0
  25. crosslingual_vi/validation-00000-of-00001.parquet +3 -0
  26. crosslingual_with_para_bg/train-00000-of-00001.parquet +3 -0
  27. crosslingual_with_para_bg/validation-00000-of-00001.parquet +3 -0
  28. crosslingual_with_para_hr/train-00000-of-00001.parquet +3 -0
  29. crosslingual_with_para_hr/validation-00000-of-00001.parquet +3 -0
  30. crosslingual_with_para_hu/train-00000-of-00001.parquet +3 -0
  31. crosslingual_with_para_hu/validation-00000-of-00001.parquet +3 -0
  32. crosslingual_with_para_it/train-00000-of-00001.parquet +3 -0
  33. crosslingual_with_para_it/validation-00000-of-00001.parquet +3 -0
  34. crosslingual_with_para_mk/train-00000-of-00001.parquet +3 -0
  35. crosslingual_with_para_mk/validation-00000-of-00001.parquet +3 -0
  36. crosslingual_with_para_pl/train-00000-of-00001.parquet +3 -0
  37. crosslingual_with_para_pl/validation-00000-of-00001.parquet +3 -0
  38. crosslingual_with_para_pt/train-00000-of-00001.parquet +3 -0
  39. crosslingual_with_para_pt/validation-00000-of-00001.parquet +3 -0
  40. crosslingual_with_para_sq/train-00000-of-00001.parquet +3 -0
  41. crosslingual_with_para_sq/validation-00000-of-00001.parquet +3 -0
  42. crosslingual_with_para_sr/train-00000-of-00001.parquet +3 -0
  43. crosslingual_with_para_sr/validation-00000-of-00001.parquet +3 -0
  44. crosslingual_with_para_test/test-00000-of-00001.parquet +3 -0
  45. crosslingual_with_para_tr/train-00000-of-00001.parquet +3 -0
  46. crosslingual_with_para_tr/validation-00000-of-00001.parquet +3 -0
  47. crosslingual_with_para_vi/train-00000-of-00001.parquet +3 -0
  48. crosslingual_with_para_vi/validation-00000-of-00001.parquet +3 -0
  49. dataset_infos.json +0 -0
  50. exams.py +0 -266
README.md CHANGED
@@ -1,5 +1,4 @@
1
  ---
2
- pretty_name: EXAMS
3
  annotations_creators:
4
  - found
5
  language_creators:
@@ -37,6 +36,35 @@ task_categories:
37
  task_ids:
38
  - multiple-choice-qa
39
  paperswithcode_id: exams
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
  dataset_info:
41
  - config_name: alignments
42
  features:
@@ -46,11 +74,11 @@ dataset_info:
46
  sequence: string
47
  splits:
48
  - name: full
49
- num_bytes: 1265280
50
  num_examples: 10834
51
- download_size: 169745177
52
- dataset_size: 1265280
53
- - config_name: multilingual
54
  features:
55
  - name: id
56
  dtype: string
@@ -78,17 +106,14 @@ dataset_info:
78
  dtype: string
79
  splits:
80
  - name: train
81
- num_bytes: 3385865
82
- num_examples: 7961
83
  - name: validation
84
- num_bytes: 1143067
85
- num_examples: 2672
86
- - name: test
87
- num_bytes: 5753625
88
- num_examples: 13510
89
- download_size: 169745177
90
- dataset_size: 10282557
91
- - config_name: multilingual_with_para
92
  features:
93
  - name: id
94
  dtype: string
@@ -116,17 +141,14 @@ dataset_info:
116
  dtype: string
117
  splits:
118
  - name: train
119
- num_bytes: 127298595
120
- num_examples: 7961
121
  - name: validation
122
- num_bytes: 42713069
123
- num_examples: 2672
124
- - name: test
125
- num_bytes: 207981218
126
- num_examples: 13510
127
- download_size: 169745177
128
- dataset_size: 377992882
129
- - config_name: crosslingual_test
130
  features:
131
  - name: id
132
  dtype: string
@@ -153,12 +175,15 @@ dataset_info:
153
  - name: language
154
  dtype: string
155
  splits:
156
- - name: test
157
- num_bytes: 8412531
158
- num_examples: 19736
159
- download_size: 169745177
160
- dataset_size: 8412531
161
- - config_name: crosslingual_with_para_test
 
 
 
162
  features:
163
  - name: id
164
  dtype: string
@@ -185,12 +210,15 @@ dataset_info:
185
  - name: language
186
  dtype: string
187
  splits:
188
- - name: test
189
- num_bytes: 207981218
190
- num_examples: 13510
191
- download_size: 169745177
192
- dataset_size: 207981218
193
- - config_name: crosslingual_bg
 
 
 
194
  features:
195
  - name: id
196
  dtype: string
@@ -218,14 +246,14 @@ dataset_info:
218
  dtype: string
219
  splits:
220
  - name: train
221
- num_bytes: 1078545
222
- num_examples: 2344
223
  - name: validation
224
- num_bytes: 282115
225
- num_examples: 593
226
- download_size: 169745177
227
- dataset_size: 1360660
228
- - config_name: crosslingual_with_para_bg
229
  features:
230
  - name: id
231
  dtype: string
@@ -253,14 +281,14 @@ dataset_info:
253
  dtype: string
254
  splits:
255
  - name: train
256
- num_bytes: 47068024
257
- num_examples: 2344
258
  - name: validation
259
- num_bytes: 11916370
260
- num_examples: 593
261
- download_size: 169745177
262
- dataset_size: 58984394
263
- - config_name: crosslingual_hr
264
  features:
265
  - name: id
266
  dtype: string
@@ -288,14 +316,14 @@ dataset_info:
288
  dtype: string
289
  splits:
290
  - name: train
291
- num_bytes: 808320
292
- num_examples: 2341
293
  - name: validation
294
- num_bytes: 176910
295
- num_examples: 538
296
- download_size: 169745177
297
- dataset_size: 985230
298
- - config_name: crosslingual_with_para_hr
299
  features:
300
  - name: id
301
  dtype: string
@@ -323,14 +351,14 @@ dataset_info:
323
  dtype: string
324
  splits:
325
  - name: train
326
- num_bytes: 24890820
327
- num_examples: 2341
328
  - name: validation
329
- num_bytes: 5695382
330
- num_examples: 538
331
- download_size: 169745177
332
- dataset_size: 30586202
333
- - config_name: crosslingual_hu
334
  features:
335
  - name: id
336
  dtype: string
@@ -358,14 +386,14 @@ dataset_info:
358
  dtype: string
359
  splits:
360
  - name: train
361
- num_bytes: 678447
362
- num_examples: 1731
363
  - name: validation
364
- num_bytes: 202324
365
- num_examples: 536
366
- download_size: 169745177
367
- dataset_size: 880771
368
- - config_name: crosslingual_with_para_hu
369
  features:
370
  - name: id
371
  dtype: string
@@ -392,15 +420,12 @@ dataset_info:
392
  - name: language
393
  dtype: string
394
  splits:
395
- - name: train
396
- num_bytes: 19036575
397
- num_examples: 1731
398
- - name: validation
399
- num_bytes: 6043577
400
- num_examples: 536
401
- download_size: 169745177
402
- dataset_size: 25080152
403
- - config_name: crosslingual_it
404
  features:
405
  - name: id
406
  dtype: string
@@ -428,14 +453,14 @@ dataset_info:
428
  dtype: string
429
  splits:
430
  - name: train
431
- num_bytes: 399864
432
- num_examples: 1010
433
  - name: validation
434
- num_bytes: 93343
435
- num_examples: 246
436
- download_size: 169745177
437
- dataset_size: 493207
438
- - config_name: crosslingual_with_para_it
439
  features:
440
  - name: id
441
  dtype: string
@@ -463,14 +488,14 @@ dataset_info:
463
  dtype: string
464
  splits:
465
  - name: train
466
- num_bytes: 16409787
467
- num_examples: 1010
468
  - name: validation
469
- num_bytes: 4018497
470
- num_examples: 246
471
- download_size: 169745177
472
- dataset_size: 20428284
473
- - config_name: crosslingual_mk
474
  features:
475
  - name: id
476
  dtype: string
@@ -498,14 +523,14 @@ dataset_info:
498
  dtype: string
499
  splits:
500
  - name: train
501
- num_bytes: 826582
502
- num_examples: 1665
503
  - name: validation
504
- num_bytes: 204570
505
- num_examples: 410
506
- download_size: 169745177
507
- dataset_size: 1031152
508
- - config_name: crosslingual_with_para_mk
509
  features:
510
  - name: id
511
  dtype: string
@@ -533,14 +558,14 @@ dataset_info:
533
  dtype: string
534
  splits:
535
  - name: train
536
- num_bytes: 38446774
537
- num_examples: 1665
538
  - name: validation
539
- num_bytes: 9673826
540
- num_examples: 410
541
- download_size: 169745177
542
- dataset_size: 48120600
543
- - config_name: crosslingual_pl
544
  features:
545
  - name: id
546
  dtype: string
@@ -568,14 +593,14 @@ dataset_info:
568
  dtype: string
569
  splits:
570
  - name: train
571
- num_bytes: 574246
572
- num_examples: 1577
573
  - name: validation
574
- num_bytes: 141877
575
- num_examples: 394
576
- download_size: 169745177
577
- dataset_size: 716123
578
- - config_name: crosslingual_with_para_pl
579
  features:
580
  - name: id
581
  dtype: string
@@ -603,14 +628,14 @@ dataset_info:
603
  dtype: string
604
  splits:
605
  - name: train
606
- num_bytes: 16374617
607
- num_examples: 1577
608
  - name: validation
609
- num_bytes: 4159076
610
- num_examples: 394
611
- download_size: 169745177
612
- dataset_size: 20533693
613
- - config_name: crosslingual_pt
614
  features:
615
  - name: id
616
  dtype: string
@@ -638,14 +663,14 @@ dataset_info:
638
  dtype: string
639
  splits:
640
  - name: train
641
- num_bytes: 375214
642
- num_examples: 740
643
  - name: validation
644
- num_bytes: 87850
645
- num_examples: 184
646
- download_size: 169745177
647
- dataset_size: 463064
648
- - config_name: crosslingual_with_para_pt
649
  features:
650
  - name: id
651
  dtype: string
@@ -673,14 +698,14 @@ dataset_info:
673
  dtype: string
674
  splits:
675
  - name: train
676
- num_bytes: 12185799
677
- num_examples: 740
678
  - name: validation
679
- num_bytes: 3093848
680
- num_examples: 184
681
- download_size: 169745177
682
- dataset_size: 15279647
683
- - config_name: crosslingual_sq
684
  features:
685
  - name: id
686
  dtype: string
@@ -708,13 +733,13 @@ dataset_info:
708
  dtype: string
709
  splits:
710
  - name: train
711
- num_bytes: 424388
712
- num_examples: 1194
713
  - name: validation
714
- num_bytes: 110293
715
- num_examples: 311
716
- download_size: 169745177
717
- dataset_size: 534681
718
  - config_name: crosslingual_with_para_sq
719
  features:
720
  - name: id
@@ -743,14 +768,14 @@ dataset_info:
743
  dtype: string
744
  splits:
745
  - name: train
746
- num_bytes: 17341921
747
  num_examples: 1194
748
  - name: validation
749
- num_bytes: 4450152
750
  num_examples: 311
751
- download_size: 169745177
752
- dataset_size: 21792073
753
- - config_name: crosslingual_sr
754
  features:
755
  - name: id
756
  dtype: string
@@ -778,14 +803,14 @@ dataset_info:
778
  dtype: string
779
  splits:
780
  - name: train
781
- num_bytes: 650268
782
  num_examples: 1323
783
  - name: validation
784
- num_bytes: 145928
785
  num_examples: 314
786
- download_size: 169745177
787
- dataset_size: 796196
788
- - config_name: crosslingual_with_para_sr
789
  features:
790
  - name: id
791
  dtype: string
@@ -812,15 +837,12 @@ dataset_info:
812
  - name: language
813
  dtype: string
814
  splits:
815
- - name: train
816
- num_bytes: 24576553
817
- num_examples: 1323
818
- - name: validation
819
- num_bytes: 5772713
820
- num_examples: 314
821
- download_size: 169745177
822
- dataset_size: 30349266
823
- - config_name: crosslingual_tr
824
  features:
825
  - name: id
826
  dtype: string
@@ -848,14 +870,14 @@ dataset_info:
848
  dtype: string
849
  splits:
850
  - name: train
851
- num_bytes: 718431
852
  num_examples: 1571
853
  - name: validation
854
- num_bytes: 182974
855
  num_examples: 393
856
- download_size: 169745177
857
- dataset_size: 901405
858
- - config_name: crosslingual_with_para_tr
859
  features:
860
  - name: id
861
  dtype: string
@@ -883,14 +905,14 @@ dataset_info:
883
  dtype: string
884
  splits:
885
  - name: train
886
- num_bytes: 18597963
887
- num_examples: 1571
888
  - name: validation
889
- num_bytes: 4763341
890
- num_examples: 393
891
- download_size: 169745177
892
- dataset_size: 23361304
893
- - config_name: crosslingual_vi
894
  features:
895
  - name: id
896
  dtype: string
@@ -918,14 +940,17 @@ dataset_info:
918
  dtype: string
919
  splits:
920
  - name: train
921
- num_bytes: 954191
922
- num_examples: 1955
923
  - name: validation
924
- num_bytes: 232264
925
- num_examples: 488
926
- download_size: 169745177
927
- dataset_size: 1186455
928
- - config_name: crosslingual_with_para_vi
 
 
 
929
  features:
930
  - name: id
931
  dtype: string
@@ -953,41 +978,178 @@ dataset_info:
953
  dtype: string
954
  splits:
955
  - name: train
956
- num_bytes: 40884023
957
- num_examples: 1955
958
  - name: validation
959
- num_bytes: 10260662
960
- num_examples: 488
961
- download_size: 169745177
962
- dataset_size: 51144685
963
- config_names:
964
- - alignments
965
- - crosslingual_bg
966
- - crosslingual_hr
967
- - crosslingual_hu
968
- - crosslingual_it
969
- - crosslingual_mk
970
- - crosslingual_pl
971
- - crosslingual_pt
972
- - crosslingual_sq
973
- - crosslingual_sr
974
- - crosslingual_test
975
- - crosslingual_tr
976
- - crosslingual_vi
977
- - crosslingual_with_para_bg
978
- - crosslingual_with_para_hr
979
- - crosslingual_with_para_hu
980
- - crosslingual_with_para_it
981
- - crosslingual_with_para_mk
982
- - crosslingual_with_para_pl
983
- - crosslingual_with_para_pt
984
- - crosslingual_with_para_sq
985
- - crosslingual_with_para_sr
986
- - crosslingual_with_para_test
987
- - crosslingual_with_para_tr
988
- - crosslingual_with_para_vi
989
- - multilingual
990
- - multilingual_with_para
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
991
  ---
992
 
993
  # Dataset Card for [Dataset Name]
 
1
  ---
 
2
  annotations_creators:
3
  - found
4
  language_creators:
 
36
  task_ids:
37
  - multiple-choice-qa
38
  paperswithcode_id: exams
39
+ pretty_name: EXAMS
40
+ config_names:
41
+ - alignments
42
+ - crosslingual_bg
43
+ - crosslingual_hr
44
+ - crosslingual_hu
45
+ - crosslingual_it
46
+ - crosslingual_mk
47
+ - crosslingual_pl
48
+ - crosslingual_pt
49
+ - crosslingual_sq
50
+ - crosslingual_sr
51
+ - crosslingual_test
52
+ - crosslingual_tr
53
+ - crosslingual_vi
54
+ - crosslingual_with_para_bg
55
+ - crosslingual_with_para_hr
56
+ - crosslingual_with_para_hu
57
+ - crosslingual_with_para_it
58
+ - crosslingual_with_para_mk
59
+ - crosslingual_with_para_pl
60
+ - crosslingual_with_para_pt
61
+ - crosslingual_with_para_sq
62
+ - crosslingual_with_para_sr
63
+ - crosslingual_with_para_test
64
+ - crosslingual_with_para_tr
65
+ - crosslingual_with_para_vi
66
+ - multilingual
67
+ - multilingual_with_para
68
  dataset_info:
69
  - config_name: alignments
70
  features:
 
74
  sequence: string
75
  splits:
76
  - name: full
77
+ num_bytes: 1265256
78
  num_examples: 10834
79
+ download_size: 184096
80
+ dataset_size: 1265256
81
+ - config_name: crosslingual_bg
82
  features:
83
  - name: id
84
  dtype: string
 
106
  dtype: string
107
  splits:
108
  - name: train
109
+ num_bytes: 1077329
110
+ num_examples: 2344
111
  - name: validation
112
+ num_bytes: 281771
113
+ num_examples: 593
114
+ download_size: 514922
115
+ dataset_size: 1359100
116
+ - config_name: crosslingual_hr
 
 
 
117
  features:
118
  - name: id
119
  dtype: string
 
141
  dtype: string
142
  splits:
143
  - name: train
144
+ num_bytes: 807104
145
+ num_examples: 2341
146
  - name: validation
147
+ num_bytes: 176594
148
+ num_examples: 538
149
+ download_size: 450090
150
+ dataset_size: 983698
151
+ - config_name: crosslingual_hu
 
 
 
152
  features:
153
  - name: id
154
  dtype: string
 
175
  - name: language
176
  dtype: string
177
  splits:
178
+ - name: train
179
+ num_bytes: 677535
180
+ num_examples: 1731
181
+ - name: validation
182
+ num_bytes: 202012
183
+ num_examples: 536
184
+ download_size: 401455
185
+ dataset_size: 879547
186
+ - config_name: crosslingual_it
187
  features:
188
  - name: id
189
  dtype: string
 
210
  - name: language
211
  dtype: string
212
  splits:
213
+ - name: train
214
+ num_bytes: 399312
215
+ num_examples: 1010
216
+ - name: validation
217
+ num_bytes: 93175
218
+ num_examples: 246
219
+ download_size: 226376
220
+ dataset_size: 492487
221
+ - config_name: crosslingual_mk
222
  features:
223
  - name: id
224
  dtype: string
 
246
  dtype: string
247
  splits:
248
  - name: train
249
+ num_bytes: 825702
250
+ num_examples: 1665
251
  - name: validation
252
+ num_bytes: 204318
253
+ num_examples: 410
254
+ download_size: 394548
255
+ dataset_size: 1030020
256
+ - config_name: crosslingual_pl
257
  features:
258
  - name: id
259
  dtype: string
 
281
  dtype: string
282
  splits:
283
  - name: train
284
+ num_bytes: 573410
285
+ num_examples: 1577
286
  - name: validation
287
+ num_bytes: 141633
288
+ num_examples: 394
289
+ download_size: 341925
290
+ dataset_size: 715043
291
+ - config_name: crosslingual_pt
292
  features:
293
  - name: id
294
  dtype: string
 
316
  dtype: string
317
  splits:
318
  - name: train
319
+ num_bytes: 374798
320
+ num_examples: 740
321
  - name: validation
322
+ num_bytes: 87714
323
+ num_examples: 184
324
+ download_size: 208021
325
+ dataset_size: 462512
326
+ - config_name: crosslingual_sq
327
  features:
328
  - name: id
329
  dtype: string
 
351
  dtype: string
352
  splits:
353
  - name: train
354
+ num_bytes: 423744
355
+ num_examples: 1194
356
  - name: validation
357
+ num_bytes: 110093
358
+ num_examples: 311
359
+ download_size: 247052
360
+ dataset_size: 533837
361
+ - config_name: crosslingual_sr
362
  features:
363
  - name: id
364
  dtype: string
 
386
  dtype: string
387
  splits:
388
  - name: train
389
+ num_bytes: 649560
390
+ num_examples: 1323
391
  - name: validation
392
+ num_bytes: 145724
393
+ num_examples: 314
394
+ download_size: 327466
395
+ dataset_size: 795284
396
+ - config_name: crosslingual_test
397
  features:
398
  - name: id
399
  dtype: string
 
420
  - name: language
421
  dtype: string
422
  splits:
423
+ - name: test
424
+ num_bytes: 8402575
425
+ num_examples: 19736
426
+ download_size: 3438526
427
+ dataset_size: 8402575
428
+ - config_name: crosslingual_tr
 
 
 
429
  features:
430
  - name: id
431
  dtype: string
 
453
  dtype: string
454
  splits:
455
  - name: train
456
+ num_bytes: 717599
457
+ num_examples: 1571
458
  - name: validation
459
+ num_bytes: 182730
460
+ num_examples: 393
461
+ download_size: 440914
462
+ dataset_size: 900329
463
+ - config_name: crosslingual_vi
464
  features:
465
  - name: id
466
  dtype: string
 
488
  dtype: string
489
  splits:
490
  - name: train
491
+ num_bytes: 953167
492
+ num_examples: 1955
493
  - name: validation
494
+ num_bytes: 231976
495
+ num_examples: 488
496
+ download_size: 462940
497
+ dataset_size: 1185143
498
+ - config_name: crosslingual_with_para_bg
499
  features:
500
  - name: id
501
  dtype: string
 
523
  dtype: string
524
  splits:
525
  - name: train
526
+ num_bytes: 47066808
527
+ num_examples: 2344
528
  - name: validation
529
+ num_bytes: 11916026
530
+ num_examples: 593
531
+ download_size: 15794611
532
+ dataset_size: 58982834
533
+ - config_name: crosslingual_with_para_hr
534
  features:
535
  - name: id
536
  dtype: string
 
558
  dtype: string
559
  splits:
560
  - name: train
561
+ num_bytes: 24889604
562
+ num_examples: 2341
563
  - name: validation
564
+ num_bytes: 5695066
565
+ num_examples: 538
566
+ download_size: 9839452
567
+ dataset_size: 30584670
568
+ - config_name: crosslingual_with_para_hu
569
  features:
570
  - name: id
571
  dtype: string
 
593
  dtype: string
594
  splits:
595
  - name: train
596
+ num_bytes: 19035663
597
+ num_examples: 1731
598
  - name: validation
599
+ num_bytes: 6043265
600
+ num_examples: 536
601
+ download_size: 9263625
602
+ dataset_size: 25078928
603
+ - config_name: crosslingual_with_para_it
604
  features:
605
  - name: id
606
  dtype: string
 
628
  dtype: string
629
  splits:
630
  - name: train
631
+ num_bytes: 16409235
632
+ num_examples: 1010
633
  - name: validation
634
+ num_bytes: 4018329
635
+ num_examples: 246
636
+ download_size: 6907617
637
+ dataset_size: 20427564
638
+ - config_name: crosslingual_with_para_mk
639
  features:
640
  - name: id
641
  dtype: string
 
663
  dtype: string
664
  splits:
665
  - name: train
666
+ num_bytes: 38445894
667
+ num_examples: 1665
668
  - name: validation
669
+ num_bytes: 9673574
670
+ num_examples: 410
671
+ download_size: 12878474
672
+ dataset_size: 48119468
673
+ - config_name: crosslingual_with_para_pl
674
  features:
675
  - name: id
676
  dtype: string
 
698
  dtype: string
699
  splits:
700
  - name: train
701
+ num_bytes: 16373781
702
+ num_examples: 1577
703
  - name: validation
704
+ num_bytes: 4158832
705
+ num_examples: 394
706
+ download_size: 6539172
707
+ dataset_size: 20532613
708
+ - config_name: crosslingual_with_para_pt
709
  features:
710
  - name: id
711
  dtype: string
 
733
  dtype: string
734
  splits:
735
  - name: train
736
+ num_bytes: 12185383
737
+ num_examples: 740
738
  - name: validation
739
+ num_bytes: 3093712
740
+ num_examples: 184
741
+ download_size: 4956969
742
+ dataset_size: 15279095
743
  - config_name: crosslingual_with_para_sq
744
  features:
745
  - name: id
 
768
  dtype: string
769
  splits:
770
  - name: train
771
+ num_bytes: 17341277
772
  num_examples: 1194
773
  - name: validation
774
+ num_bytes: 4449952
775
  num_examples: 311
776
+ download_size: 7112236
777
+ dataset_size: 21791229
778
+ - config_name: crosslingual_with_para_sr
779
  features:
780
  - name: id
781
  dtype: string
 
803
  dtype: string
804
  splits:
805
  - name: train
806
+ num_bytes: 24575845
807
  num_examples: 1323
808
  - name: validation
809
+ num_bytes: 5772509
810
  num_examples: 314
811
+ download_size: 8035415
812
+ dataset_size: 30348354
813
+ - config_name: crosslingual_with_para_test
814
  features:
815
  - name: id
816
  dtype: string
 
837
  - name: language
838
  dtype: string
839
  splits:
840
+ - name: test
841
+ num_bytes: 207974374
842
+ num_examples: 13510
843
+ download_size: 62878029
844
+ dataset_size: 207974374
845
+ - config_name: crosslingual_with_para_tr
 
 
 
846
  features:
847
  - name: id
848
  dtype: string
 
870
  dtype: string
871
  splits:
872
  - name: train
873
+ num_bytes: 18597131
874
  num_examples: 1571
875
  - name: validation
876
+ num_bytes: 4763097
877
  num_examples: 393
878
+ download_size: 7346658
879
+ dataset_size: 23360228
880
+ - config_name: crosslingual_with_para_vi
881
  features:
882
  - name: id
883
  dtype: string
 
905
  dtype: string
906
  splits:
907
  - name: train
908
+ num_bytes: 40882999
909
+ num_examples: 1955
910
  - name: validation
911
+ num_bytes: 10260374
912
+ num_examples: 488
913
+ download_size: 13028078
914
+ dataset_size: 51143373
915
+ - config_name: multilingual
916
  features:
917
  - name: id
918
  dtype: string
 
940
  dtype: string
941
  splits:
942
  - name: train
943
+ num_bytes: 3381837
944
+ num_examples: 7961
945
  - name: validation
946
+ num_bytes: 1141687
947
+ num_examples: 2672
948
+ - name: test
949
+ num_bytes: 5746781
950
+ num_examples: 13510
951
+ download_size: 4323915
952
+ dataset_size: 10270305
953
+ - config_name: multilingual_with_para
954
  features:
955
  - name: id
956
  dtype: string
 
978
  dtype: string
979
  splits:
980
  - name: train
981
+ num_bytes: 127294567
982
+ num_examples: 7961
983
  - name: validation
984
+ num_bytes: 42711689
985
+ num_examples: 2672
986
+ - name: test
987
+ num_bytes: 207974374
988
+ num_examples: 13510
989
+ download_size: 112597818
990
+ dataset_size: 377980630
991
+ configs:
992
+ - config_name: alignments
993
+ data_files:
994
+ - split: full
995
+ path: alignments/full-*
996
+ - config_name: crosslingual_bg
997
+ data_files:
998
+ - split: train
999
+ path: crosslingual_bg/train-*
1000
+ - split: validation
1001
+ path: crosslingual_bg/validation-*
1002
+ - config_name: crosslingual_hr
1003
+ data_files:
1004
+ - split: train
1005
+ path: crosslingual_hr/train-*
1006
+ - split: validation
1007
+ path: crosslingual_hr/validation-*
1008
+ - config_name: crosslingual_hu
1009
+ data_files:
1010
+ - split: train
1011
+ path: crosslingual_hu/train-*
1012
+ - split: validation
1013
+ path: crosslingual_hu/validation-*
1014
+ - config_name: crosslingual_it
1015
+ data_files:
1016
+ - split: train
1017
+ path: crosslingual_it/train-*
1018
+ - split: validation
1019
+ path: crosslingual_it/validation-*
1020
+ - config_name: crosslingual_mk
1021
+ data_files:
1022
+ - split: train
1023
+ path: crosslingual_mk/train-*
1024
+ - split: validation
1025
+ path: crosslingual_mk/validation-*
1026
+ - config_name: crosslingual_pl
1027
+ data_files:
1028
+ - split: train
1029
+ path: crosslingual_pl/train-*
1030
+ - split: validation
1031
+ path: crosslingual_pl/validation-*
1032
+ - config_name: crosslingual_pt
1033
+ data_files:
1034
+ - split: train
1035
+ path: crosslingual_pt/train-*
1036
+ - split: validation
1037
+ path: crosslingual_pt/validation-*
1038
+ - config_name: crosslingual_sq
1039
+ data_files:
1040
+ - split: train
1041
+ path: crosslingual_sq/train-*
1042
+ - split: validation
1043
+ path: crosslingual_sq/validation-*
1044
+ - config_name: crosslingual_sr
1045
+ data_files:
1046
+ - split: train
1047
+ path: crosslingual_sr/train-*
1048
+ - split: validation
1049
+ path: crosslingual_sr/validation-*
1050
+ - config_name: crosslingual_test
1051
+ data_files:
1052
+ - split: test
1053
+ path: crosslingual_test/test-*
1054
+ - config_name: crosslingual_tr
1055
+ data_files:
1056
+ - split: train
1057
+ path: crosslingual_tr/train-*
1058
+ - split: validation
1059
+ path: crosslingual_tr/validation-*
1060
+ - config_name: crosslingual_vi
1061
+ data_files:
1062
+ - split: train
1063
+ path: crosslingual_vi/train-*
1064
+ - split: validation
1065
+ path: crosslingual_vi/validation-*
1066
+ - config_name: crosslingual_with_para_bg
1067
+ data_files:
1068
+ - split: train
1069
+ path: crosslingual_with_para_bg/train-*
1070
+ - split: validation
1071
+ path: crosslingual_with_para_bg/validation-*
1072
+ - config_name: crosslingual_with_para_hr
1073
+ data_files:
1074
+ - split: train
1075
+ path: crosslingual_with_para_hr/train-*
1076
+ - split: validation
1077
+ path: crosslingual_with_para_hr/validation-*
1078
+ - config_name: crosslingual_with_para_hu
1079
+ data_files:
1080
+ - split: train
1081
+ path: crosslingual_with_para_hu/train-*
1082
+ - split: validation
1083
+ path: crosslingual_with_para_hu/validation-*
1084
+ - config_name: crosslingual_with_para_it
1085
+ data_files:
1086
+ - split: train
1087
+ path: crosslingual_with_para_it/train-*
1088
+ - split: validation
1089
+ path: crosslingual_with_para_it/validation-*
1090
+ - config_name: crosslingual_with_para_mk
1091
+ data_files:
1092
+ - split: train
1093
+ path: crosslingual_with_para_mk/train-*
1094
+ - split: validation
1095
+ path: crosslingual_with_para_mk/validation-*
1096
+ - config_name: crosslingual_with_para_pl
1097
+ data_files:
1098
+ - split: train
1099
+ path: crosslingual_with_para_pl/train-*
1100
+ - split: validation
1101
+ path: crosslingual_with_para_pl/validation-*
1102
+ - config_name: crosslingual_with_para_pt
1103
+ data_files:
1104
+ - split: train
1105
+ path: crosslingual_with_para_pt/train-*
1106
+ - split: validation
1107
+ path: crosslingual_with_para_pt/validation-*
1108
+ - config_name: crosslingual_with_para_sq
1109
+ data_files:
1110
+ - split: train
1111
+ path: crosslingual_with_para_sq/train-*
1112
+ - split: validation
1113
+ path: crosslingual_with_para_sq/validation-*
1114
+ - config_name: crosslingual_with_para_sr
1115
+ data_files:
1116
+ - split: train
1117
+ path: crosslingual_with_para_sr/train-*
1118
+ - split: validation
1119
+ path: crosslingual_with_para_sr/validation-*
1120
+ - config_name: crosslingual_with_para_test
1121
+ data_files:
1122
+ - split: test
1123
+ path: crosslingual_with_para_test/test-*
1124
+ - config_name: crosslingual_with_para_tr
1125
+ data_files:
1126
+ - split: train
1127
+ path: crosslingual_with_para_tr/train-*
1128
+ - split: validation
1129
+ path: crosslingual_with_para_tr/validation-*
1130
+ - config_name: crosslingual_with_para_vi
1131
+ data_files:
1132
+ - split: train
1133
+ path: crosslingual_with_para_vi/train-*
1134
+ - split: validation
1135
+ path: crosslingual_with_para_vi/validation-*
1136
+ - config_name: multilingual
1137
+ data_files:
1138
+ - split: train
1139
+ path: multilingual/train-*
1140
+ - split: validation
1141
+ path: multilingual/validation-*
1142
+ - split: test
1143
+ path: multilingual/test-*
1144
+ - config_name: multilingual_with_para
1145
+ data_files:
1146
+ - split: train
1147
+ path: multilingual_with_para/train-*
1148
+ - split: validation
1149
+ path: multilingual_with_para/validation-*
1150
+ - split: test
1151
+ path: multilingual_with_para/test-*
1152
+ default: true
1153
  ---
1154
 
1155
  # Dataset Card for [Dataset Name]
alignments/full-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b29e157eefe6fd0518a8956564b888d96e83662170720bf9d31b3c1a2de9ce6
3
+ size 184096
crosslingual_bg/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2661015ad0eff676acab87fed6a99768b6782a616fbe0cd09d49eb3dadf94bd
3
+ size 403502
crosslingual_bg/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:041ce81fa04d83356ea13dea98835d3ac1ef566a9af5b419bcd78f21727311bd
3
+ size 111420
crosslingual_hr/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9704d89c6aa31e0ffbd9abcd5ba534748b1b6bd54dddbd4fc6c9eed2f0090ed8
3
+ size 365783
crosslingual_hr/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d62fcdbc31d96829092b0d9c4e442896a01d097f94017d51b85076a23b80981a
3
+ size 84307
crosslingual_hu/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3c3154ff716d40c7ed15debbe9e82631f944f3a2f4ae4535af59d09f0055d81
3
+ size 305804
crosslingual_hu/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1461df538b82497e92175f35610682fd5108e03b9f1007a5c6cc3cd9f3cd859d
3
+ size 95651
crosslingual_it/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ceefc63ff4bf30000850b0d5c7713be5b646ab21786edda2ce15b91e1851dfc4
3
+ size 178661
crosslingual_it/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4744304f4803a7d8451a9833ec4ec31909440269d96dbeed797c1a1f375a58bc
3
+ size 47715
crosslingual_mk/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21514491c215094848d5c9c0717ccd69aae0d2d1503ec182d78c20517b2744ef
3
+ size 310559
crosslingual_mk/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7974cd49f0ca0307696a1ff3dfe94968ce1587b3a2be39f4efb31019f1f5047
3
+ size 83989
crosslingual_pl/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0208006a6436b3b1929fff649978e4b313dce7661dbae63d79fb85028a027c5
3
+ size 271310
crosslingual_pl/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fedcdf86a5a43f5007c54beda4c308e43d0ea1d97b2f3e235cd2304024a8493
3
+ size 70615
crosslingual_pt/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d92737677095b00de7d194c99da09b82016e9b9f251d777c632a308a406561a5
3
+ size 163195
crosslingual_pt/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af3b73b848a52c4584a81cf2b232380a9b94da465e7086011004e63b030b8c68
3
+ size 44826
crosslingual_sq/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee35f50add1857c5148de8e7f5e31c51e4b7f6fd41544929f753fe43a1ad4d36
3
+ size 191271
crosslingual_sq/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff7f7756181d1685b714a7e54e7ec400f5657bc1c6464a85239414f73007d209
3
+ size 55781
crosslingual_sr/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33a516d6c68800aa00dcde49ed3d86ced9b4bb80dbe000f42ba2cdc136c0fdcc
3
+ size 262335
crosslingual_sr/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d67144cc23fda9b38c78d447f5b1253dbbca7468ebc35f78b6c454162ace2ae
3
+ size 65131
crosslingual_test/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:807134617aea0ae6958bc05798d8f71d0f9d5dbcf690b2ef0ec189e8c545249e
3
+ size 3438526
crosslingual_tr/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ee94d6b1dd71e2362d79de40da4604bb0d0188f45ce2ee974f20a2b8df645e8
3
+ size 346493
crosslingual_tr/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fb8caf5b7b937c53356d816031bbb4cabd045d94782840a6f91effc363e9c67
3
+ size 94421
crosslingual_vi/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:864a2bb0d96bebda26dbbd8645db2d96135dd7cf8e690fe43481213d929d6165
3
+ size 366031
crosslingual_vi/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c0d8e76970c491cc3027af0347675e2070ef6dc18aeb89e87d77c1a510e9f94
3
+ size 96909
crosslingual_with_para_bg/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c095a226e38375c00741df05298f7d34b52791242d928759434a759a887639f2
3
+ size 12606675
crosslingual_with_para_bg/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:093e37c7a797b8dee06afb6757064ce09ef696a5ca2df4301cfbd2f86d72d4c3
3
+ size 3187936
crosslingual_with_para_hr/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6369b88b7b23c40f2f6a5eed98b968a781da8a2ece5d65ab474064cf76fc9328
3
+ size 8021778
crosslingual_with_para_hr/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fa559aa8723b0be712931660be3e91d4fab67bccd0660faeebcf1fd943e5111
3
+ size 1817674
crosslingual_with_para_hu/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2545f2f64f24989eaae48f6de2c6af71cc88339d1ede0143006fc7993097582f
3
+ size 7064584
crosslingual_with_para_hu/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4730990290df58fda67654dbb4125f2515af5292fbcf21cf6827611f850f3952
3
+ size 2199041
crosslingual_with_para_it/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b95e3e83479806119995a03a96436a22564348265fe1bc42e263a65c7b5bbf00
3
+ size 5572126
crosslingual_with_para_it/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ceb521f2416c45d53c2e50754eea02db374a4675ba4ff0618f9c5c801059ff5
3
+ size 1335491
crosslingual_with_para_mk/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a65ee5f5d0730ffc458746043632038680dcaefed18a863393d1b16b4ba2e400
3
+ size 10283648
crosslingual_with_para_mk/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bd0a3a3f422fc58bee977f61b28a4d79214372e91dfb29fb75a8e3d983ddfa4
3
+ size 2594826
crosslingual_with_para_pl/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfe8bffe5187745e2efd8d8354f29b835eab3326e593aa5da02d6a4131b04706
3
+ size 5225951
crosslingual_with_para_pl/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:670310fc6baf8b89fb283b260aa2f1b13e73a82890f56957350a3c086368e0e2
3
+ size 1313221
crosslingual_with_para_pt/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:935795674ccffe9470a1cff5f0be87b573ab2fafdec38b458f3350aa4cc5701f
3
+ size 3961124
crosslingual_with_para_pt/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ee30998f4d77188eea641cfd0b5e2616db0c1031d495029b43567ab11edb99e
3
+ size 995845
crosslingual_with_para_sq/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b193ccb1664d06189e863deb1c8f3d35cb4ebaf3721c7b7ef5025005d852444
3
+ size 5623573
crosslingual_with_para_sq/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:076693b308981ef8468c753215c52238a74c8187c9d0e53928e7036dd8fdfa3a
3
+ size 1488663
crosslingual_with_para_sr/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d288f0fc3112a86e9297943cdecd070e867c2346396065d4dc50163b83a01cb
3
+ size 6526994
crosslingual_with_para_sr/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edcf0d6210bcb2de0f1ffc69a02c961187fa0967a1609c2bae96ffa70945f9d3
3
+ size 1508421
crosslingual_with_para_test/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f795460d626ac0611394cb06e8ab0ff0944d0aab79822d74e2caa2dda4ae4f61
3
+ size 62878029
crosslingual_with_para_tr/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:368e985cf012bc248d2d235f0c40c0314e1934e222b5d5c4f1d770a2ee17f9ca
3
+ size 5832098
crosslingual_with_para_tr/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d5938d0b2b3289144cccac3ab7d961ed991ae141b24241c43c0f5facca7fd92
3
+ size 1514560
crosslingual_with_para_vi/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0368b09798743a1c41150463a455fc5472c3c6c43abbcd8cffd41395c68385f
3
+ size 10473199
crosslingual_with_para_vi/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17e4c1cf0deaf19ee61a0db7f007e1dfb9668f765392441923f5f5b4332457ad
3
+ size 2554879
dataset_infos.json DELETED
The diff for this file is too large to render. See raw diff
 
exams.py DELETED
@@ -1,266 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
- """EXAMS: a benchmark dataset for multilingual and cross-lingual question answering"""
16
-
17
-
18
- import json
19
-
20
- import datasets
21
-
22
-
23
- _CITATION = """\
24
- @article{hardalov2020exams,
25
- title={EXAMS: A Multi-subject High School Examinations Dataset for Cross-lingual and Multilingual Question Answering},
26
- author={Hardalov, Momchil and Mihaylov, Todor and Dimitrina Zlatkova and Yoan Dinkov and Ivan Koychev and Preslav Nvakov},
27
- journal={arXiv preprint arXiv:2011.03080},
28
- year={2020}
29
- }
30
- """
31
-
32
- _DESCRIPTION = """\
33
- EXAMS is a benchmark dataset for multilingual and cross-lingual question answering from high school examinations.
34
- It consists of more than 24,000 high-quality high school exam questions in 16 languages,
35
- covering 8 language families and 24 school subjects from Natural Sciences and Social Sciences, among others.
36
- """
37
-
38
- _HOMEPAGE = "https://github.com/mhardalov/exams-qa"
39
-
40
- _LICENSE = "CC-BY-SA-4.0"
41
-
42
- _URLS_LIST = [
43
- ("alignments", "https://github.com/mhardalov/exams-qa/raw/main/data/exams/parallel_questions.jsonl"),
44
- ]
45
- _URLS_LIST += [
46
- (
47
- "multilingual_train",
48
- "https://github.com/mhardalov/exams-qa/raw/main/data/exams/multilingual/train.jsonl.tar.gz",
49
- ),
50
- ("multilingual_dev", "https://github.com/mhardalov/exams-qa/raw/main/data/exams/multilingual/dev.jsonl.tar.gz"),
51
- ("multilingual_test", "https://github.com/mhardalov/exams-qa/raw/main/data/exams/multilingual/test.jsonl.tar.gz"),
52
- (
53
- "multilingual_with_para_train",
54
- "https://github.com/mhardalov/exams-qa/raw/main/data/exams/multilingual/with_paragraphs/train_with_para.jsonl.tar.gz",
55
- ),
56
- (
57
- "multilingual_with_para_dev",
58
- "https://github.com/mhardalov/exams-qa/raw/main/data/exams/multilingual/with_paragraphs/dev_with_para.jsonl.tar.gz",
59
- ),
60
- (
61
- "multilingual_with_para_test",
62
- "https://github.com/mhardalov/exams-qa/raw/main/data/exams/multilingual/with_paragraphs/test_with_para.jsonl.tar.gz",
63
- ),
64
- ]
65
-
66
- _CROSS_LANGUAGES = ["bg", "hr", "hu", "it", "mk", "pl", "pt", "sq", "sr", "tr", "vi"]
67
- _URLS_LIST += [
68
- ("crosslingual_test", "https://github.com/mhardalov/exams-qa/raw/main/data/exams/cross-lingual/test.jsonl.tar.gz"),
69
- (
70
- "crosslingual_with_para_test",
71
- "https://github.com/mhardalov/exams-qa/raw/main/data/exams/cross-lingual/with_paragraphs/test_with_para.jsonl.tar.gz",
72
- ),
73
- ]
74
- for ln in _CROSS_LANGUAGES:
75
- _URLS_LIST += [
76
- (
77
- f"crosslingual_{ln}_train",
78
- f"https://github.com/mhardalov/exams-qa/raw/main/data/exams/cross-lingual/train_{ln}.jsonl.tar.gz",
79
- ),
80
- (
81
- f"crosslingual_with_para_{ln}_train",
82
- f"https://github.com/mhardalov/exams-qa/raw/main/data/exams/cross-lingual/with_paragraphs/train_{ln}_with_para.jsonl.tar.gz",
83
- ),
84
- (
85
- f"crosslingual_{ln}_dev",
86
- f"https://github.com/mhardalov/exams-qa/raw/main/data/exams/cross-lingual/dev_{ln}.jsonl.tar.gz",
87
- ),
88
- (
89
- f"crosslingual_with_para_{ln}_dev",
90
- f"https://github.com/mhardalov/exams-qa/raw/main/data/exams/cross-lingual/with_paragraphs/dev_{ln}_with_para.jsonl.tar.gz",
91
- ),
92
- ]
93
- _URLs = dict(_URLS_LIST)
94
-
95
-
96
- class ExamsConfig(datasets.BuilderConfig):
97
- def __init__(self, lang, with_para, **kwargs):
98
- super(ExamsConfig, self).__init__(version=datasets.Version("1.0.0", ""), **kwargs)
99
- self.lang = lang
100
- self.with_para = "_with_para" if with_para else ""
101
-
102
-
103
- class Exams(datasets.GeneratorBasedBuilder):
104
- """Exams dataset"""
105
-
106
- VERSION = datasets.Version("1.0.0")
107
- BUILDER_CONFIG_CLASS = ExamsConfig
108
- BUILDER_CONFIGS = [
109
- ExamsConfig(
110
- lang="",
111
- with_para=False,
112
- name="alignments",
113
- description="loads the alignment between question IDs across languages",
114
- ),
115
- ExamsConfig(
116
- lang="all",
117
- with_para=False,
118
- name="multilingual",
119
- description="Loads the unified multilingual train/dev/test split",
120
- ),
121
- ExamsConfig(
122
- lang="all",
123
- with_para=True,
124
- name="multilingual_with_para",
125
- description="Loads the unified multilingual train/dev/test split with Wikipedia support paragraphs",
126
- ),
127
- ExamsConfig(
128
- lang="all", with_para=False, name="crosslingual_test", description="Loads crosslingual test set only"
129
- ),
130
- ExamsConfig(
131
- lang="all",
132
- with_para=True,
133
- name="crosslingual_with_para_test",
134
- description="Loads crosslingual test set only with Wikipedia support paragraphs",
135
- ),
136
- ]
137
- for ln in _CROSS_LANGUAGES:
138
- BUILDER_CONFIGS += [
139
- ExamsConfig(
140
- lang=ln,
141
- with_para=False,
142
- name=f"crosslingual_{ln}",
143
- description=f"Loads crosslingual train and dev set for {ln}",
144
- ),
145
- ExamsConfig(
146
- lang=ln,
147
- with_para=True,
148
- name=f"crosslingual_with_para_{ln}",
149
- description=f"Loads crosslingual train and dev set for {ln} with Wikipedia support paragraphs",
150
- ),
151
- ]
152
-
153
- DEFAULT_CONFIG_NAME = (
154
- "multilingual_with_para" # It's not mandatory to have a default configuration. Just use one if it make sense.
155
- )
156
-
157
- def _info(self):
158
- if self.config.name == "alignments": # This is the name of the configuration selected in BUILDER_CONFIGS above
159
- features = datasets.Features(
160
- {
161
- "source_id": datasets.Value("string"),
162
- "target_id_list": datasets.Sequence(datasets.Value("string")),
163
- }
164
- )
165
- else: # This is an example to show how to have different features for "first_domain" and "second_domain"
166
- features = datasets.Features(
167
- {
168
- "id": datasets.Value("string"),
169
- "question": {
170
- "stem": datasets.Value("string"),
171
- "choices": datasets.Sequence(
172
- {
173
- "text": datasets.Value("string"),
174
- "label": datasets.Value("string"),
175
- "para": datasets.Value("string"),
176
- }
177
- ),
178
- },
179
- "answerKey": datasets.Value("string"),
180
- "info": {
181
- "grade": datasets.Value("int32"),
182
- "subject": datasets.Value("string"),
183
- "language": datasets.Value("string"),
184
- },
185
- }
186
- )
187
- return datasets.DatasetInfo(
188
- description=_DESCRIPTION,
189
- features=features, # Here we define them above because they are different between the two configurations
190
- supervised_keys=None,
191
- homepage=_HOMEPAGE,
192
- license=_LICENSE,
193
- citation=_CITATION,
194
- )
195
-
196
- def _split_generators(self, dl_manager):
197
- """Returns SplitGenerators."""
198
- archives = dl_manager.download(_URLs)
199
- if self.config.name == "alignments":
200
- return [
201
- datasets.SplitGenerator(
202
- name="full",
203
- gen_kwargs={"filepath": archives["alignments"]},
204
- ),
205
- ]
206
- elif self.config.name in ["multilingual", "multilingual_with_para"]:
207
- return [
208
- datasets.SplitGenerator(
209
- name=spl_enum,
210
- gen_kwargs={
211
- "filepath": f"{spl}{self.config.with_para}.jsonl",
212
- "files": dl_manager.iter_archive(archives[f"{self.config.name}_{spl}"]),
213
- },
214
- )
215
- for spl, spl_enum in [
216
- ("train", datasets.Split.TRAIN),
217
- ("dev", datasets.Split.VALIDATION),
218
- ("test", datasets.Split.TEST),
219
- ]
220
- ]
221
- elif self.config.name in ["crosslingual_test", "crosslingual_with_para_test"]:
222
- return [
223
- datasets.SplitGenerator(
224
- name=datasets.Split.TEST,
225
- gen_kwargs={
226
- "filepath": f"test{self.config.with_para}.jsonl",
227
- "files": dl_manager.iter_archive(archives[self.config.name]),
228
- },
229
- ),
230
- ]
231
- else:
232
- return [
233
- datasets.SplitGenerator(
234
- name=spl_enum,
235
- gen_kwargs={
236
- "filepath": f"{spl}_{self.config.lang}{self.config.with_para}.jsonl",
237
- "files": dl_manager.iter_archive(archives[f"{self.config.name}_{spl}"]),
238
- },
239
- )
240
- for spl, spl_enum in [
241
- ("train", datasets.Split.TRAIN),
242
- ("dev", datasets.Split.VALIDATION),
243
- ]
244
- ]
245
-
246
- def _generate_examples(self, filepath, files=None):
247
- if self.config.name == "alignments":
248
- with open(filepath, encoding="utf-8") as f:
249
- for id_, line in enumerate(f):
250
- line_dict = json.loads(line.strip())
251
- in_id, out_list = list(line_dict.items())[0]
252
- yield id_, {"source_id": in_id, "target_id_list": out_list}
253
- else:
254
- for path, f in files:
255
- if path == filepath:
256
- for id_, line in enumerate(f):
257
- line_dict = json.loads(line.strip())
258
- for choice in line_dict["question"]["choices"]:
259
- choice["para"] = choice.get("para", "")
260
- yield id_, {
261
- "id": line_dict["id"],
262
- "question": line_dict["question"],
263
- "answerKey": line_dict["answerKey"],
264
- "info": line_dict["info"],
265
- }
266
- break