{ "add_prefix_space": false, "added_tokens_decoder": { "0": { "content": "", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "1": { "content": "", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "2": { "content": "<|im_end|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true }, "10": { "content": "", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true } }, "bos_token": "", "chat_template": "{%- set ns = namespace(found=false) -%}\n{%- for message in messages -%}\n {%- if message['role'] == 'system' -%}\n {%- set ns.found = true -%}\n {%- endif -%}\n{%- endfor -%}\n{%- for message in messages -%}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n' + message['content'].rstrip() + '<|eot_id|>' -}}\n{%- endfor -%}\n{%- if add_generation_prompt -%}\n {{-'<|start_header_id|>assistant<|end_header_id|>\n\n'-}}\n{%- endif -%}", "clean_up_tokenization_spaces": true, "eos_token": "<|im_end|>", "model_max_length": 1000000000000000019884624838656, "pad_token": "", "tokenizer_class": "GPT2Tokenizer", "unk_token": "" }