## vocab_file | |
- ice_text.model | |
- 二进制文件 | |
- num_image_tokens = 20000 | |
词典大小 150528 | |
``` | |
tokens: ['▁good', '▁morning'] ; id: [20315, 21774] ; text: good morning | |
tokens: ['▁good', '<|blank_2|>', 'morning'] ; id: [20315, 150009, 60813] ; text: good morning | |
tokens: ['▁', 'goog', '▁morning', 'abc'] ; id: [20005, 46456, 21774, 27415] ; text: goog morningabc | |
tokens: ['▁', '你是谁'] ; id: [20005, 128293] ; text: 你是谁 | |
``` | |
`▁` 是啥,空格吗?注意区分 `_` | |
## | |
``` | |
tokenizer = TextTokenizer(self.vocab_file) | |
``` | |