Upload tokenizer
Browse files- merges.txt +1 -1
- tokenizer.json +0 -1
merges.txt
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
-
#version: 0.2
|
2 |
Ġ Ġ
|
3 |
ĠĠ ĠĠ
|
4 |
ĠĠ Ġ
|
|
|
1 |
+
#version: 0.2 - Trained by `huggingface/tokenizers`
|
2 |
Ġ Ġ
|
3 |
ĠĠ ĠĠ
|
4 |
ĠĠ Ġ
|
tokenizer.json
CHANGED
@@ -39,7 +39,6 @@
|
|
39 |
"continuing_subword_prefix": "",
|
40 |
"end_of_word_suffix": "",
|
41 |
"fuse_unk": false,
|
42 |
-
"byte_fallback": false,
|
43 |
"vocab": {
|
44 |
"<|endoftext|>": 0,
|
45 |
"!": 1,
|
|
|
39 |
"continuing_subword_prefix": "",
|
40 |
"end_of_word_suffix": "",
|
41 |
"fuse_unk": false,
|
|
|
42 |
"vocab": {
|
43 |
"<|endoftext|>": 0,
|
44 |
"!": 1,
|