Upload tokenizer
Browse files- merges.txt +1 -1
- tokenizer.json +0 -1
merges.txt
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
-
#version: 0.2
|
2 |
Ġ Ġ
|
3 |
ĠĠ ĠĠ
|
4 |
ĠĠĠĠ ĠĠĠĠ
|
|
|
1 |
+
#version: 0.2 - Trained by `huggingface/tokenizers`
|
2 |
Ġ Ġ
|
3 |
ĠĠ ĠĠ
|
4 |
ĠĠĠĠ ĠĠĠĠ
|
tokenizer.json
CHANGED
@@ -205,7 +205,6 @@
|
|
205 |
"continuing_subword_prefix": null,
|
206 |
"end_of_word_suffix": null,
|
207 |
"fuse_unk": false,
|
208 |
-
"byte_fallback": false,
|
209 |
"vocab": {
|
210 |
"<|endoftext|>": 0,
|
211 |
"<fim_prefix>": 1,
|
|
|
205 |
"continuing_subword_prefix": null,
|
206 |
"end_of_word_suffix": null,
|
207 |
"fuse_unk": false,
|
|
|
208 |
"vocab": {
|
209 |
"<|endoftext|>": 0,
|
210 |
"<fim_prefix>": 1,
|