eson's picture
update
d10ecd7
raw history blame
No virus
183 Bytes
from transformers import AutoTokenizer, AutoModelForMaskedLM
tokenizer = AutoTokenizer.from_pretrained("bert-base-chinese")
tokens = tokenizer.tokenize("明天搜索")
print(tokens)