anikethjr's picture
Training in progress, step 250
c017b5d
raw
history blame
443 Bytes
{
"model_max_length": 512,
"name_or_path": "/global/scratch/users/aniketh/PromoGen/HepG2_SentencePieceUnigramTokenizer_4096_fast",
"special_tokens": [
"<BOS>",
"<EOS>",
"<PAD>",
"<UNK>",
"<CLS>",
"<SEP>",
"<MASK>"
],
"special_tokens_map_file": "/global/scratch/users/aniketh/PromoGen/HepG2_SentencePieceUnigramTokenizer_4096_fast/special_tokens_map.json",
"tokenizer_class": "PreTrainedTokenizerFast"
}