import os | |
from transformers import AutoTokenizer | |
tokenizer = AutoTokenizer.from_pretrained("bert-base-chinese") | |
# vocab_size = len(tokenizer.get_vocab()) | |
# vocab_size = tokenizer.vocab_size | |
import os | |
from transformers import AutoTokenizer | |
tokenizer = AutoTokenizer.from_pretrained("bert-base-chinese") | |
# vocab_size = len(tokenizer.get_vocab()) | |
# vocab_size = tokenizer.vocab_size | |