Upload tokenizer.json
Browse files- tokenizer.json +1 -0
tokenizer.json
CHANGED
@@ -257,6 +257,7 @@
|
|
257 |
"continuing_subword_prefix": null,
|
258 |
"end_of_word_suffix": null,
|
259 |
"fuse_unk": false,
|
|
|
260 |
"vocab": {
|
261 |
"<|endoftext|>": 0,
|
262 |
"<|padding|>": 1,
|
|
|
257 |
"continuing_subword_prefix": null,
|
258 |
"end_of_word_suffix": null,
|
259 |
"fuse_unk": false,
|
260 |
+
"byte_fallback": false,
|
261 |
"vocab": {
|
262 |
"<|endoftext|>": 0,
|
263 |
"<|padding|>": 1,
|