Xenova HF staff commited on
Commit
763623f
1 Parent(s): 3e7ebf5

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -40,7 +40,7 @@
40
  "pooler_size_per_head": 128,
41
  "pooler_type": "first_token_transform",
42
  "position_embedding_type": "absolute",
43
- "transformers_version": "4.26.1",
44
  "type_vocab_size": 2,
45
  "use_cache": true,
46
  "vocab_size": 105879
 
40
  "pooler_size_per_head": 128,
41
  "pooler_type": "first_token_transform",
42
  "position_embedding_type": "absolute",
43
+ "transformers_version": "4.29.2",
44
  "type_vocab_size": 2,
45
  "use_cache": true,
46
  "vocab_size": 105879
onnx/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f096a03bd1a36a82e5783f8a4a1bf62c740540e24c4af32aa58d19df4a96d50
3
+ size 669707799
onnx/model_quantized.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9b78da8a15fb254408c5295815df4a512dfa3e3d29a900b5187d870f4c502141
3
- size 170409471
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70a87c0bfae76da0440b5f48f146d82270206db68625310cc01989fdb04f3504
3
+ size 168174746
tokenizer.json CHANGED
@@ -5,48 +5,48 @@
5
  "added_tokens": [
6
  {
7
  "id": 0,
8
- "special": true,
9
  "content": "[PAD]",
10
  "single_word": false,
11
  "lstrip": false,
12
  "rstrip": false,
13
- "normalized": false
 
14
  },
15
  {
16
  "id": 100,
17
- "special": true,
18
  "content": "[UNK]",
19
  "single_word": false,
20
  "lstrip": false,
21
  "rstrip": false,
22
- "normalized": false
 
23
  },
24
  {
25
  "id": 101,
26
- "special": true,
27
  "content": "[CLS]",
28
  "single_word": false,
29
  "lstrip": false,
30
  "rstrip": false,
31
- "normalized": false
 
32
  },
33
  {
34
  "id": 102,
35
- "special": true,
36
  "content": "[SEP]",
37
  "single_word": false,
38
  "lstrip": false,
39
  "rstrip": false,
40
- "normalized": false
 
41
  },
42
  {
43
  "id": 103,
44
- "special": true,
45
  "content": "[MASK]",
46
  "single_word": false,
47
  "lstrip": false,
48
  "rstrip": false,
49
- "normalized": false
 
50
  }
51
  ],
52
  "normalizer": {
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
8
  "content": "[PAD]",
9
  "single_word": false,
10
  "lstrip": false,
11
  "rstrip": false,
12
+ "normalized": false,
13
+ "special": true
14
  },
15
  {
16
  "id": 100,
 
17
  "content": "[UNK]",
18
  "single_word": false,
19
  "lstrip": false,
20
  "rstrip": false,
21
+ "normalized": false,
22
+ "special": true
23
  },
24
  {
25
  "id": 101,
 
26
  "content": "[CLS]",
27
  "single_word": false,
28
  "lstrip": false,
29
  "rstrip": false,
30
+ "normalized": false,
31
+ "special": true
32
  },
33
  {
34
  "id": 102,
 
35
  "content": "[SEP]",
36
  "single_word": false,
37
  "lstrip": false,
38
  "rstrip": false,
39
+ "normalized": false,
40
+ "special": true
41
  },
42
  {
43
  "id": 103,
 
44
  "content": "[MASK]",
45
  "single_word": false,
46
  "lstrip": false,
47
  "rstrip": false,
48
+ "normalized": false,
49
+ "special": true
50
  }
51
  ],
52
  "normalizer": {
tokenizer_config.json CHANGED
@@ -1,15 +1,14 @@
1
  {
 
2
  "cls_token": "[CLS]",
3
  "do_basic_tokenize": true,
4
  "do_lower_case": true,
5
  "mask_token": "[MASK]",
6
  "max_len": 512,
7
  "model_max_length": 512,
8
- "name_or_path": "nlptown/bert-base-multilingual-uncased-sentiment",
9
  "never_split": null,
10
  "pad_token": "[PAD]",
11
  "sep_token": "[SEP]",
12
- "special_tokens_map_file": null,
13
  "strip_accents": null,
14
  "tokenize_chinese_chars": true,
15
  "tokenizer_class": "BertTokenizer",
 
1
  {
2
+ "clean_up_tokenization_spaces": true,
3
  "cls_token": "[CLS]",
4
  "do_basic_tokenize": true,
5
  "do_lower_case": true,
6
  "mask_token": "[MASK]",
7
  "max_len": 512,
8
  "model_max_length": 512,
 
9
  "never_split": null,
10
  "pad_token": "[PAD]",
11
  "sep_token": "[SEP]",
 
12
  "strip_accents": null,
13
  "tokenize_chinese_chars": true,
14
  "tokenizer_class": "BertTokenizer",