Xenova HF staff commited on
Commit
e26400f
1 Parent(s): ac44bfe

Upload tokenizer.json

Browse files
Files changed (1) hide show
  1. tokenizer.json +99 -0
tokenizer.json ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
5
+ "added_tokens": [
6
+ {
7
+ "id": 34,
8
+ "content": "<unk>",
9
+ "single_word": false,
10
+ "lstrip": false,
11
+ "rstrip": false,
12
+ "normalized": false,
13
+ "special": true
14
+ }
15
+ ],
16
+ "normalizer": {
17
+ "type": "Sequence",
18
+ "normalizers": [
19
+ {
20
+ "type": "Lowercase"
21
+ },
22
+ {
23
+ "type": "Replace",
24
+ "pattern": {
25
+ "String": "\u021b"
26
+ },
27
+ "content": "\u0163"
28
+ },
29
+ {
30
+ "type": "Replace",
31
+ "pattern": {
32
+ "Regex": "[^cqubg_n\u2013mzitl\u0103f\u0219a1\u00eep\\-hs'xej\u00e2 rvo\u0163d]"
33
+ },
34
+ "content": ""
35
+ },
36
+ {
37
+ "type": "Strip",
38
+ "strip_left": true,
39
+ "strip_right": true
40
+ },
41
+ {
42
+ "type": "Replace",
43
+ "pattern": {
44
+ "Regex": "(?=.)|(?<!^)$"
45
+ },
46
+ "content": "c"
47
+ }
48
+ ]
49
+ },
50
+ "pre_tokenizer": {
51
+ "type": "Split",
52
+ "pattern": {
53
+ "Regex": ""
54
+ },
55
+ "behavior": "Isolated",
56
+ "invert": false
57
+ },
58
+ "post_processor": null,
59
+ "decoder": null,
60
+ "model": {
61
+ "vocab": {
62
+ "c": 0,
63
+ "q": 1,
64
+ "u": 2,
65
+ "b": 3,
66
+ "g": 4,
67
+ "_": 5,
68
+ "n": 6,
69
+ "\u2013": 7,
70
+ "m": 8,
71
+ "z": 9,
72
+ "i": 10,
73
+ "t": 11,
74
+ "l": 12,
75
+ "\u0103": 13,
76
+ "f": 14,
77
+ "\u0219": 15,
78
+ "a": 16,
79
+ "1": 17,
80
+ "\u00ee": 18,
81
+ "p": 19,
82
+ "-": 20,
83
+ "h": 21,
84
+ "s": 22,
85
+ "'": 23,
86
+ "x": 24,
87
+ "e": 25,
88
+ "j": 26,
89
+ "\u00e2": 27,
90
+ " ": 28,
91
+ "r": 29,
92
+ "v": 30,
93
+ "o": 31,
94
+ "\u0163": 32,
95
+ "d": 33,
96
+ "<unk>": 34
97
+ }
98
+ }
99
+ }