fnet-large-finetuned-cola-copy4 / tokenizer_config.json
gchhablani's picture
Training in progress, epoch 1
e98b06d
raw
history blame
618 Bytes
{"do_lower_case": false, "remove_space": true, "keep_accents": true, "unk_token": "<unk>", "sep_token": "[SEP]", "pad_token": "<pad>", "cls_token": "[CLS]", "mask_token": {"content": "[MASK]", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "sp_model_kwargs": {}, "model_max_length": 512, "special_tokens_map_file": "/home/gunjan/.cache/huggingface/transformers/9a50cd9b3771023230b1128ab5e112461bf36f5826d5f7eb654348e956979a54.a2cfb41d43ad2ac50fa89998bfa88393398e3bb3439ee99e1b00933bce2eb1ba", "name_or_path": "google/fnet-large", "tokenizer_class": "FNetTokenizer"}