{"max_len": 42, "padding": "max_length", "add_special_tokens": true, "return_tensors": "pt", "truncation": true, "bos_token": "", "eos_token": "", "unk_token": "", "pad_token": "", "mask_token": "", "special_tokens_map_file": "/opt/ml/.cache/huggingface/transformers/c2ab65b9d700d0871fd407d489869d7b93f69fb5f1a58fb1fac796fd43b9ea27.1f5b09bb43973b9fbd2ba75c9fe44ffab036b980c4e6a9d779aa7707913416fe", "name_or_path": "skt/ko-gpt-trinity-1.2B-v0.5", "tokenizer_class": "PreTrainedTokenizerFast"}