{ "add_bos_token": true, "add_eos_token": false, "added_tokens_decoder": { "50256": { "content": "<|endoftext|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "58940": { "content": "", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "58941": { "content": "", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "58943": { "content": "<|im_end|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true } }, "additional_special_tokens": [], "bos_token": "", "chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{'A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the users questions.\nhuman: ' + message['content'] + '\n'}}{% else %}{{'Assistant:\n' + message['content']}}{% endif %}{% endfor %}", "clean_up_tokenization_spaces": false, "eos_token": "<|im_end|>", "legacy": true, "model_max_length": 2048, "pad_token": "", "sp_model_kwargs": {}, "spaces_between_special_tokens": false, "tokenizer_class": "LlamaTokenizer", "trust_remote_code": false, "unk_token": "<|endoftext|>", "use_default_system_prompt": false, "use_fast": true }