# from tensorflow.keras.preprocessing.text import Tokenizer
from tensorflow.keras.preprocessing.text import Tokenizer

import joblib


def one_hot_vector():
    # todo:1 - 初始化词表
    vocabs = ['周杰伦','成绩辅导费','王力宏','反间谍法就','吴亦凡','鹿晗']
    # todo:2 - 实例化tokenizer对象,分词器对象
    tokenizer = Tokenizer(num_woros = len(vocabs))
    print('tokenizer---->',tokenizer)
