Upload tokenizer_config.json with huggingface_hub
Browse files- tokenizer_config.json +2 -2
 
    	
        tokenizer_config.json
    CHANGED
    
    | 
         @@ -277,7 +277,7 @@ 
     | 
|
| 277 | 
         
             
                "</box>"
         
     | 
| 278 | 
         
             
              ],
         
     | 
| 279 | 
         
             
              "bos_token": null,
         
     | 
| 280 | 
         
            -
              "chat_template": "{ 
     | 
| 281 | 
         
             
              "clean_up_tokenization_spaces": false,
         
     | 
| 282 | 
         
             
              "eos_token": "<|im_end|>",
         
     | 
| 283 | 
         
             
              "errors": "replace",
         
     | 
| 
         @@ -286,4 +286,4 @@ 
     | 
|
| 286 | 
         
             
              "split_special_tokens": false,
         
     | 
| 287 | 
         
             
              "tokenizer_class": "Qwen2Tokenizer",
         
     | 
| 288 | 
         
             
              "unk_token": null
         
     | 
| 289 | 
         
            -
            }
         
     | 
| 
         | 
|
| 277 | 
         
             
                "</box>"
         
     | 
| 278 | 
         
             
              ],
         
     | 
| 279 | 
         
             
              "bos_token": null,
         
     | 
| 280 | 
         
            +
              "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n'}}{% if message['content'] is string %}{{ message['content'] }}{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' %}{{ '<image>\n' }}{% elif content['type'] == 'video' %}{{ '<video>\n' }}{% elif content['type'] == 'text' %}{{ content['text'] }}{% endif %}{% endfor %}{% endif %}{{'<|im_end|>\n'}}{% endfor %}{% if add_generation_prompt %}{{'<|im_start|>assistant\n' }}{% endif %}",
         
     | 
| 281 | 
         
             
              "clean_up_tokenization_spaces": false,
         
     | 
| 282 | 
         
             
              "eos_token": "<|im_end|>",
         
     | 
| 283 | 
         
             
              "errors": "replace",
         
     | 
| 
         | 
|
| 286 | 
         
             
              "split_special_tokens": false,
         
     | 
| 287 | 
         
             
              "tokenizer_class": "Qwen2Tokenizer",
         
     | 
| 288 | 
         
             
              "unk_token": null
         
     | 
| 289 | 
         
            +
            }
         
     |