{ "add_prefix_space": false, "additional_special_tokens": [ ">>TITLE<<", ">>ABSTRACT<<", ">>INTRODUCTION<<", ">>SUMMARY<<", ">>COMMENT<<", ">>ANSWER<<", ">>QUESTION<<", ">>DOMAIN<<", ">>PREFIX<<", ">>SUFFIX<<", ">>MIDDLE<<" ], "clean_up_tokenization_spaces": true, "eos_token": "<|endoftext|>", "model_max_length": 2048, "tokenizer_class": "PreTrainedTokenizerFast" }