Upload tokenizer
Browse files- README.md +1 -1
- tokenizer.json +0 -0
- tokenizer_config.json +3 -3
README.md
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
---
|
|
|
2 |
library_name: transformers
|
3 |
license: apache-2.0
|
4 |
-
base_model: AndrewOficial/metrology
|
5 |
tags:
|
6 |
- generated_from_trainer
|
7 |
model-index:
|
|
|
1 |
---
|
2 |
+
base_model: AndrewOficial/metrology
|
3 |
library_name: transformers
|
4 |
license: apache-2.0
|
|
|
5 |
tags:
|
6 |
- generated_from_trainer
|
7 |
model-index:
|
tokenizer.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
CHANGED
@@ -16,7 +16,7 @@
|
|
16 |
"single_word": false,
|
17 |
"special": true
|
18 |
},
|
19 |
-
"
|
20 |
"content": "кПа",
|
21 |
"lstrip": false,
|
22 |
"normalized": true,
|
@@ -24,7 +24,7 @@
|
|
24 |
"single_word": false,
|
25 |
"special": false
|
26 |
},
|
27 |
-
"
|
28 |
"content": " ",
|
29 |
"lstrip": false,
|
30 |
"normalized": true,
|
@@ -32,7 +32,7 @@
|
|
32 |
"single_word": false,
|
33 |
"special": false
|
34 |
},
|
35 |
-
"
|
36 |
"content": "/t",
|
37 |
"lstrip": false,
|
38 |
"normalized": true,
|
|
|
16 |
"single_word": false,
|
17 |
"special": true
|
18 |
},
|
19 |
+
"1629": {
|
20 |
"content": "кПа",
|
21 |
"lstrip": false,
|
22 |
"normalized": true,
|
|
|
24 |
"single_word": false,
|
25 |
"special": false
|
26 |
},
|
27 |
+
"3572": {
|
28 |
"content": " ",
|
29 |
"lstrip": false,
|
30 |
"normalized": true,
|
|
|
32 |
"single_word": false,
|
33 |
"special": false
|
34 |
},
|
35 |
+
"3573": {
|
36 |
"content": "/t",
|
37 |
"lstrip": false,
|
38 |
"normalized": true,
|