Upload tokenizer
Browse files- merges.txt +1 -1
- tokenizer.json +0 -1
merges.txt
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
-
#version: 0.2
|
2 |
Ġ t
|
3 |
Ġ a
|
4 |
h e
|
|
|
1 |
+
#version: 0.2 - Trained by `huggingface/tokenizers`
|
2 |
Ġ t
|
3 |
Ġ a
|
4 |
h e
|
tokenizer.json
CHANGED
@@ -87,7 +87,6 @@
|
|
87 |
"continuing_subword_prefix": "",
|
88 |
"end_of_word_suffix": "",
|
89 |
"fuse_unk": false,
|
90 |
-
"byte_fallback": false,
|
91 |
"vocab": {
|
92 |
"<s>": 0,
|
93 |
"<pad>": 1,
|
|
|
87 |
"continuing_subword_prefix": "",
|
88 |
"end_of_word_suffix": "",
|
89 |
"fuse_unk": false,
|
|
|
90 |
"vocab": {
|
91 |
"<s>": 0,
|
92 |
"<pad>": 1,
|