albertvillanova HF staff commited on
Commit
26b4712
1 Parent(s): ea7dceb

Add rte data files

Browse files
README.md CHANGED
@@ -260,17 +260,17 @@ dataset_info:
260
  - name: idx
261
  dtype: int32
262
  splits:
263
- - name: test
264
- num_bytes: 975936
265
- num_examples: 3000
266
  - name: train
267
- num_bytes: 848888
268
  num_examples: 2490
269
  - name: validation
270
- num_bytes: 90911
271
  num_examples: 277
272
- download_size: 697150
273
- dataset_size: 1915735
 
 
 
274
  - config_name: sst2
275
  features:
276
  - name: sentence
@@ -400,6 +400,14 @@ configs:
400
  path: qqp/validation-*
401
  - split: test
402
  path: qqp/test-*
 
 
 
 
 
 
 
 
403
  - config_name: sst2
404
  data_files:
405
  - split: train
 
260
  - name: idx
261
  dtype: int32
262
  splits:
 
 
 
263
  - name: train
264
+ num_bytes: 847320
265
  num_examples: 2490
266
  - name: validation
267
+ num_bytes: 90728
268
  num_examples: 277
269
+ - name: test
270
+ num_bytes: 974053
271
+ num_examples: 3000
272
+ download_size: 1274409
273
+ dataset_size: 1912101
274
  - config_name: sst2
275
  features:
276
  - name: sentence
 
400
  path: qqp/validation-*
401
  - split: test
402
  path: qqp/test-*
403
+ - config_name: rte
404
+ data_files:
405
+ - split: train
406
+ path: rte/train-*
407
+ - split: validation
408
+ path: rte/validation-*
409
+ - split: test
410
+ path: rte/test-*
411
  - config_name: sst2
412
  data_files:
413
  - split: train
dataset_infos.json CHANGED
@@ -533,39 +533,32 @@
533
  },
534
  "rte": {
535
  "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
536
- "citation": "@inproceedings{dagan2005pascal,\n title={The PASCAL recognising textual entailment challenge},\n author={Dagan, Ido and Glickman, Oren and Magnini, Bernardo},\n booktitle={Machine Learning Challenges Workshop},\n pages={177--190},\n year={2005},\n organization={Springer}\n}\n@inproceedings{bar2006second,\n title={The second pascal recognising textual entailment challenge},\n author={Bar-Haim, Roy and Dagan, Ido and Dolan, Bill and Ferro, Lisa and Giampiccolo, Danilo and Magnini, Bernardo and Szpektor, Idan},\n booktitle={Proceedings of the second PASCAL challenges workshop on recognising textual entailment},\n volume={6},\n number={1},\n pages={6--4},\n year={2006},\n organization={Venice}\n}\n@inproceedings{giampiccolo2007third,\n title={The third pascal recognizing textual entailment challenge},\n author={Giampiccolo, Danilo and Magnini, Bernardo and Dagan, Ido and Dolan, Bill},\n booktitle={Proceedings of the ACL-PASCAL workshop on textual entailment and paraphrasing},\n pages={1--9},\n year={2007},\n organization={Association for Computational Linguistics}\n}\n@inproceedings{bentivogli2009fifth,\n title={The Fifth PASCAL Recognizing Textual Entailment Challenge.},\n author={Bentivogli, Luisa and Clark, Peter and Dagan, Ido and Giampiccolo, Danilo},\n booktitle={TAC},\n year={2009}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n\nNote that each GLUE dataset has its own citation. Please see the source to see\nthe correct citation for each contained dataset.",
537
  "homepage": "https://aclweb.org/aclwiki/Recognizing_Textual_Entailment",
538
  "license": "",
539
  "features": {
540
  "sentence1": {
541
  "dtype": "string",
542
- "id": null,
543
  "_type": "Value"
544
  },
545
  "sentence2": {
546
  "dtype": "string",
547
- "id": null,
548
  "_type": "Value"
549
  },
550
  "label": {
551
- "num_classes": 2,
552
  "names": [
553
  "entailment",
554
  "not_entailment"
555
  ],
556
- "names_file": null,
557
- "id": null,
558
  "_type": "ClassLabel"
559
  },
560
  "idx": {
561
  "dtype": "int32",
562
- "id": null,
563
  "_type": "Value"
564
  }
565
  },
566
- "post_processed": null,
567
- "supervised_keys": null,
568
  "builder_name": "glue",
 
569
  "config_name": "rte",
570
  "version": {
571
  "version_str": "1.0.0",
@@ -575,35 +568,28 @@
575
  "patch": 0
576
  },
577
  "splits": {
578
- "test": {
579
- "name": "test",
580
- "num_bytes": 975936,
581
- "num_examples": 3000,
582
- "dataset_name": "glue"
583
- },
584
  "train": {
585
  "name": "train",
586
- "num_bytes": 848888,
587
  "num_examples": 2490,
588
- "dataset_name": "glue"
589
  },
590
  "validation": {
591
  "name": "validation",
592
- "num_bytes": 90911,
593
  "num_examples": 277,
594
- "dataset_name": "glue"
595
- }
596
- },
597
- "download_checksums": {
598
- "https://dl.fbaipublicfiles.com/glue/data/RTE.zip": {
599
- "num_bytes": 697150,
600
- "checksum": "6bf86de103ecd335f3441bd43574d23fef87ecc695977a63b82d5efb206556ee"
601
  }
602
  },
603
- "download_size": 697150,
604
- "post_processing_size": null,
605
- "dataset_size": 1915735,
606
- "size_in_bytes": 2612885
607
  },
608
  "wnli": {
609
  "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
 
533
  },
534
  "rte": {
535
  "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
536
+ "citation": "@inproceedings{dagan2005pascal,\n title={The PASCAL recognising textual entailment challenge},\n author={Dagan, Ido and Glickman, Oren and Magnini, Bernardo},\n booktitle={Machine Learning Challenges Workshop},\n pages={177--190},\n year={2005},\n organization={Springer}\n}\n@inproceedings{bar2006second,\n title={The second pascal recognising textual entailment challenge},\n author={Bar-Haim, Roy and Dagan, Ido and Dolan, Bill and Ferro, Lisa and Giampiccolo, Danilo and Magnini, Bernardo and Szpektor, Idan},\n booktitle={Proceedings of the second PASCAL challenges workshop on recognising textual entailment},\n volume={6},\n number={1},\n pages={6--4},\n year={2006},\n organization={Venice}\n}\n@inproceedings{giampiccolo2007third,\n title={The third pascal recognizing textual entailment challenge},\n author={Giampiccolo, Danilo and Magnini, Bernardo and Dagan, Ido and Dolan, Bill},\n booktitle={Proceedings of the ACL-PASCAL workshop on textual entailment and paraphrasing},\n pages={1--9},\n year={2007},\n organization={Association for Computational Linguistics}\n}\n@inproceedings{bentivogli2009fifth,\n title={The Fifth PASCAL Recognizing Textual Entailment Challenge.},\n author={Bentivogli, Luisa and Clark, Peter and Dagan, Ido and Giampiccolo, Danilo},\n booktitle={TAC},\n year={2009}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n",
537
  "homepage": "https://aclweb.org/aclwiki/Recognizing_Textual_Entailment",
538
  "license": "",
539
  "features": {
540
  "sentence1": {
541
  "dtype": "string",
 
542
  "_type": "Value"
543
  },
544
  "sentence2": {
545
  "dtype": "string",
 
546
  "_type": "Value"
547
  },
548
  "label": {
 
549
  "names": [
550
  "entailment",
551
  "not_entailment"
552
  ],
 
 
553
  "_type": "ClassLabel"
554
  },
555
  "idx": {
556
  "dtype": "int32",
 
557
  "_type": "Value"
558
  }
559
  },
 
 
560
  "builder_name": "glue",
561
+ "dataset_name": "glue",
562
  "config_name": "rte",
563
  "version": {
564
  "version_str": "1.0.0",
 
568
  "patch": 0
569
  },
570
  "splits": {
 
 
 
 
 
 
571
  "train": {
572
  "name": "train",
573
+ "num_bytes": 847320,
574
  "num_examples": 2490,
575
+ "dataset_name": null
576
  },
577
  "validation": {
578
  "name": "validation",
579
+ "num_bytes": 90728,
580
  "num_examples": 277,
581
+ "dataset_name": null
582
+ },
583
+ "test": {
584
+ "name": "test",
585
+ "num_bytes": 974053,
586
+ "num_examples": 3000,
587
+ "dataset_name": null
588
  }
589
  },
590
+ "download_size": 1274409,
591
+ "dataset_size": 1912101,
592
+ "size_in_bytes": 3186510
 
593
  },
594
  "wnli": {
595
  "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
rte/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f44aadbfb8bbb7a64ba0674bd26ff77b66e88fdf7a6d64255a5ba6ae9057383
3
+ size 621413
rte/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6252ab17015d718f6de1effe0980f7b158df63e3d16207cd8bd396b608e5147
3
+ size 583976
rte/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb2aa2e04f551133ba663617a15ae133dc22b0f6a969bc0629b5ea6003ee9cf8
3
+ size 69020