{ "added_tokens_decoder": { "0": { "content": "<|unused0|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "1": { "content": "<|unused1|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "2": { "content": "<|endoftext|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "3": { "content": "<|sep|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "30000": { "content": "<|acc|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "30001": { "content": "<|tel|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "30002": { "content": "<|rrn|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true } }, "additional_special_tokens": [ "<|endoftext|>", "<|sep|>", "<|acc|>", "<|tel|>", "<|rrn|>" ], "clean_up_tokenization_spaces": true, "eos_token": "<|endoftext|>", "model_max_length": 1000000000000000019884624838656, "pad_token": "<|endoftext|>", "tokenizer_class": "PreTrainedTokenizerFast" }