gmongaras's picture
Upload README.md with huggingface_hub
c6074d4
|
raw
history blame
No virus
1.12 kB
---
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
- split: test
path: data/test-*
dataset_info:
features:
- name: label
dtype: float64
- name: dataset_name
dtype: string
- name: input_ids
sequence: int32
- name: token_type_ids
sequence: int8
- name: attention_mask
sequence: int8
splits:
- name: train
num_bytes: 232895922
num_examples: 949728
- name: validation
num_bytes: 17255970
num_examples: 69711
- name: test
num_bytes: 96102951
num_examples: 425205
download_size: 123150665
dataset_size: 346254843
---
Original Dataset from: https://huggingface.co/datasets/glue
This dataset is adapted from https://huggingface.co/datasets/gmongaras/BERT_Base_Cased_512_GLUE
Every split besides the ax split is in this dataset.
Lines above 512 tokens from the BERT-cased (bert-base-cased) tokenizer are removed in the original dataset
If in any case the sentences are longer than 512 tokens, they are subsetted.
Original labels and dataset categories are retained.