import sentencepiece as spm spm.SentencePieceTrainer.train(input="/researchdisk/training_dataset_sentences/train.txt", model_prefix='spiece', vocab_size=32000, character_coverage=1.0, pad_id=0, unk_id=2, eos_id=1, bos_id=-1, user_defined_symbols=['[NLU]', '[NLG]', '[S2S]'], train_extremely_large_corpus=True, num_threads=96, input_sentence_size=50000000, shuffle_input_sentence=True)