--- configs: - config_name: default data_files: - split: train path: data/train-* - split: validation path: data/validation-* - split: test path: data/test-* dataset_info: features: - name: label dtype: float64 - name: dataset_name dtype: string - name: input_ids sequence: int32 - name: token_type_ids sequence: int8 - name: attention_mask sequence: int8 splits: - name: train num_bytes: 232895922 num_examples: 949728 - name: validation num_bytes: 17255970 num_examples: 69711 - name: test num_bytes: 96102951 num_examples: 425205 download_size: 123150665 dataset_size: 346254843 --- Original Dataset from: https://huggingface.co/datasets/glue This dataset is adapted from https://huggingface.co/datasets/gmongaras/BERT_Base_Cased_512_GLUE Every split besides the ax split is in this dataset. Lines above 512 tokens from the BERT-cased (bert-base-cased) tokenizer are removed in the original dataset If in any case the sentences are longer than 512 tokens, they are subsetted. Original labels and dataset categories are retained.