tiny-random-chatglm2 / tokenizer_config.json
katuni4ka's picture
End of training
0be9a13 verified
raw
history blame contribute delete
458 Bytes
{
"added_tokens_decoder": {},
"auto_map": {
"AutoTokenizer": [
"katuni4ka/tiny-random-chatglm2--tokenization_chatglm.ChatGLMTokenizer",
null
]
},
"clean_up_tokenization_spaces": false,
"do_lower_case": false,
"eos_token": "</s>",
"model_max_length": 1000000000000000019884624838656,
"pad_token": "<unk>",
"padding_side": "left",
"remove_space": false,
"tokenizer_class": "ChatGLMTokenizer",
"unk_token": "<unk>"
}