Update tokenizer.json
Browse files- tokenizer.json +4 -5
tokenizer.json
CHANGED
@@ -4,7 +4,7 @@
|
|
4 |
"padding": null,
|
5 |
"added_tokens": [
|
6 |
{
|
7 |
-
"id":
|
8 |
"content": "<|endoftext|>",
|
9 |
"single_word": false,
|
10 |
"lstrip": false,
|
@@ -13,7 +13,7 @@
|
|
13 |
"special": true
|
14 |
},
|
15 |
{
|
16 |
-
"id":
|
17 |
"content": "<|beginoftext|>",
|
18 |
"single_word": false,
|
19 |
"lstrip": false,
|
@@ -22,7 +22,7 @@
|
|
22 |
"special": true
|
23 |
},
|
24 |
{
|
25 |
-
"id":
|
26 |
"content": "<|unknown|>",
|
27 |
"single_word": false,
|
28 |
"lstrip": false,
|
@@ -31,7 +31,7 @@
|
|
31 |
"special": true
|
32 |
},
|
33 |
{
|
34 |
-
"id":
|
35 |
"content": "<|padding|>",
|
36 |
"single_word": false,
|
37 |
"lstrip": false,
|
@@ -73,7 +73,6 @@
|
|
73 |
"vocab": {
|
74 |
"<|endoftext|>": 0,
|
75 |
"<|padding|>": 1,
|
76 |
-
|
77 |
"a": 28,
|
78 |
"c": 29,
|
79 |
"g": 30,
|
|
|
4 |
"padding": null,
|
5 |
"added_tokens": [
|
6 |
{
|
7 |
+
"id": 1,
|
8 |
"content": "<|endoftext|>",
|
9 |
"single_word": false,
|
10 |
"lstrip": false,
|
|
|
13 |
"special": true
|
14 |
},
|
15 |
{
|
16 |
+
"id": 0,
|
17 |
"content": "<|beginoftext|>",
|
18 |
"single_word": false,
|
19 |
"lstrip": false,
|
|
|
22 |
"special": true
|
23 |
},
|
24 |
{
|
25 |
+
"id": 2,
|
26 |
"content": "<|unknown|>",
|
27 |
"single_word": false,
|
28 |
"lstrip": false,
|
|
|
31 |
"special": true
|
32 |
},
|
33 |
{
|
34 |
+
"id": 1,
|
35 |
"content": "<|padding|>",
|
36 |
"single_word": false,
|
37 |
"lstrip": false,
|
|
|
73 |
"vocab": {
|
74 |
"<|endoftext|>": 0,
|
75 |
"<|padding|>": 1,
|
|
|
76 |
"a": 28,
|
77 |
"c": 29,
|
78 |
"g": 30,
|