Joshua Lochner commited on
Commit
9f7ab1e
1 Parent(s): b85a77d

Add v1 models

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. quantized/bert-base-cased/default/config.json +25 -0
  2. quantized/bert-base-cased/default/model.onnx +3 -0
  3. quantized/bert-base-cased/default/special_tokens_map.json +7 -0
  4. quantized/bert-base-cased/default/tokenizer.json +0 -0
  5. quantized/bert-base-cased/default/tokenizer_config.json +14 -0
  6. quantized/bert-base-cased/default/vocab.txt +0 -0
  7. quantized/bert-base-cased/masked-lm/config.json +25 -0
  8. quantized/bert-base-cased/masked-lm/model.onnx +3 -0
  9. quantized/bert-base-cased/masked-lm/special_tokens_map.json +7 -0
  10. quantized/bert-base-cased/masked-lm/tokenizer.json +0 -0
  11. quantized/bert-base-cased/masked-lm/tokenizer_config.json +14 -0
  12. quantized/bert-base-cased/masked-lm/vocab.txt +0 -0
  13. quantized/bert-base-multilingual-cased/default/config.json +30 -0
  14. quantized/bert-base-multilingual-cased/default/model.onnx +3 -0
  15. quantized/bert-base-multilingual-cased/default/special_tokens_map.json +7 -0
  16. quantized/bert-base-multilingual-cased/default/tokenizer.json +0 -0
  17. quantized/bert-base-multilingual-cased/default/tokenizer_config.json +14 -0
  18. quantized/bert-base-multilingual-cased/default/vocab.txt +0 -0
  19. quantized/bert-base-multilingual-cased/masked-lm/config.json +30 -0
  20. quantized/bert-base-multilingual-cased/masked-lm/model.onnx +3 -0
  21. quantized/bert-base-multilingual-cased/masked-lm/special_tokens_map.json +7 -0
  22. quantized/bert-base-multilingual-cased/masked-lm/tokenizer.json +0 -0
  23. quantized/bert-base-multilingual-cased/masked-lm/tokenizer_config.json +14 -0
  24. quantized/bert-base-multilingual-cased/masked-lm/vocab.txt +0 -0
  25. quantized/bert-base-multilingual-cased/sequence-classification/config.json +30 -0
  26. quantized/bert-base-multilingual-cased/sequence-classification/model.onnx +3 -0
  27. quantized/bert-base-multilingual-cased/sequence-classification/special_tokens_map.json +7 -0
  28. quantized/bert-base-multilingual-cased/sequence-classification/tokenizer.json +0 -0
  29. quantized/bert-base-multilingual-cased/sequence-classification/tokenizer_config.json +14 -0
  30. quantized/bert-base-multilingual-cased/sequence-classification/vocab.txt +0 -0
  31. quantized/bert-base-multilingual-uncased/default/config.json +30 -0
  32. quantized/bert-base-multilingual-uncased/default/model.onnx +3 -0
  33. quantized/bert-base-multilingual-uncased/default/special_tokens_map.json +7 -0
  34. quantized/bert-base-multilingual-uncased/default/tokenizer.json +0 -0
  35. quantized/bert-base-multilingual-uncased/default/tokenizer_config.json +14 -0
  36. quantized/bert-base-multilingual-uncased/default/vocab.txt +0 -0
  37. quantized/bert-base-multilingual-uncased/masked-lm/config.json +30 -0
  38. quantized/bert-base-multilingual-uncased/masked-lm/model.onnx +3 -0
  39. quantized/bert-base-multilingual-uncased/masked-lm/special_tokens_map.json +7 -0
  40. quantized/bert-base-multilingual-uncased/masked-lm/tokenizer.json +0 -0
  41. quantized/bert-base-multilingual-uncased/masked-lm/tokenizer_config.json +14 -0
  42. quantized/bert-base-multilingual-uncased/masked-lm/vocab.txt +0 -0
  43. quantized/bert-base-multilingual-uncased/sequence-classification/config.json +30 -0
  44. quantized/bert-base-multilingual-uncased/sequence-classification/model.onnx +3 -0
  45. quantized/bert-base-multilingual-uncased/sequence-classification/special_tokens_map.json +7 -0
  46. quantized/bert-base-multilingual-uncased/sequence-classification/tokenizer.json +0 -0
  47. quantized/bert-base-multilingual-uncased/sequence-classification/tokenizer_config.json +14 -0
  48. quantized/bert-base-multilingual-uncased/sequence-classification/vocab.txt +0 -0
  49. quantized/bert-base-uncased/default/config.json +25 -0
  50. quantized/bert-base-uncased/default/model.onnx +3 -0
quantized/bert-base-cased/default/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-cased",
3
+ "architectures": [
4
+ "BertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "transformers_version": "4.26.1",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 28996
25
+ }
quantized/bert-base-cased/default/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a488ed90daa60e1f8a320220d1cc2abeeed02c1f4d596b526fea1db410da2a8
3
+ size 111073262
quantized/bert-base-cased/default/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
quantized/bert-base-cased/default/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
quantized/bert-base-cased/default/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": false,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 512,
6
+ "name_or_path": "bert-base-cased",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
quantized/bert-base-cased/default/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
quantized/bert-base-cased/masked-lm/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-cased",
3
+ "architectures": [
4
+ "BertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "transformers_version": "4.26.1",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 28996
25
+ }
quantized/bert-base-cased/masked-lm/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d92a9cd900e9cdd793a40344596c53357e45287bf76ec4e513bce8f4d2e81e16
3
+ size 131848370
quantized/bert-base-cased/masked-lm/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
quantized/bert-base-cased/masked-lm/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
quantized/bert-base-cased/masked-lm/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": false,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 512,
6
+ "name_or_path": "bert-base-cased",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
quantized/bert-base-cased/masked-lm/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
quantized/bert-base-multilingual-cased/default/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-multilingual-cased",
3
+ "architectures": [
4
+ "BertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "directionality": "bidi",
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "pooler_fc_size": 768,
21
+ "pooler_num_attention_heads": 12,
22
+ "pooler_num_fc_layers": 3,
23
+ "pooler_size_per_head": 128,
24
+ "pooler_type": "first_token_transform",
25
+ "position_embedding_type": "absolute",
26
+ "transformers_version": "4.26.1",
27
+ "type_vocab_size": 2,
28
+ "use_cache": true,
29
+ "vocab_size": 119547
30
+ }
quantized/bert-base-multilingual-cased/default/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efdbda1fd0e2a668e9ccb5cebdce1674b31edcfeecaf33d8fb9aa3cc67c06bc3
3
+ size 180616430
quantized/bert-base-multilingual-cased/default/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
quantized/bert-base-multilingual-cased/default/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
quantized/bert-base-multilingual-cased/default/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": false,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 512,
6
+ "name_or_path": "bert-base-multilingual-cased",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
quantized/bert-base-multilingual-cased/default/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
quantized/bert-base-multilingual-cased/masked-lm/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-multilingual-cased",
3
+ "architectures": [
4
+ "BertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "directionality": "bidi",
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "pooler_fc_size": 768,
21
+ "pooler_num_attention_heads": 12,
22
+ "pooler_num_fc_layers": 3,
23
+ "pooler_size_per_head": 128,
24
+ "pooler_type": "first_token_transform",
25
+ "position_embedding_type": "absolute",
26
+ "transformers_version": "4.26.1",
27
+ "type_vocab_size": 2,
28
+ "use_cache": true,
29
+ "vocab_size": 119547
30
+ }
quantized/bert-base-multilingual-cased/masked-lm/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8563a3f5cab6cfc6e3969c03ab5a2e9dd718b95479c120f1995d905efa3d56e9
3
+ size 271749666
quantized/bert-base-multilingual-cased/masked-lm/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
quantized/bert-base-multilingual-cased/masked-lm/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
quantized/bert-base-multilingual-cased/masked-lm/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": false,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 512,
6
+ "name_or_path": "bert-base-multilingual-cased",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
quantized/bert-base-multilingual-cased/masked-lm/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
quantized/bert-base-multilingual-cased/sequence-classification/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-multilingual-cased",
3
+ "architectures": [
4
+ "BertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "directionality": "bidi",
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "pooler_fc_size": 768,
21
+ "pooler_num_attention_heads": 12,
22
+ "pooler_num_fc_layers": 3,
23
+ "pooler_size_per_head": 128,
24
+ "pooler_type": "first_token_transform",
25
+ "position_embedding_type": "absolute",
26
+ "transformers_version": "4.26.1",
27
+ "type_vocab_size": 2,
28
+ "use_cache": true,
29
+ "vocab_size": 119547
30
+ }
quantized/bert-base-multilingual-cased/sequence-classification/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30b54e696722bc9cd526016cd7d215dfed320e15f63fa14aba2013e1b6c04b02
3
+ size 711528862
quantized/bert-base-multilingual-cased/sequence-classification/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
quantized/bert-base-multilingual-cased/sequence-classification/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
quantized/bert-base-multilingual-cased/sequence-classification/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": false,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 512,
6
+ "name_or_path": "bert-base-multilingual-cased",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
quantized/bert-base-multilingual-cased/sequence-classification/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
quantized/bert-base-multilingual-uncased/default/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-multilingual-uncased",
3
+ "architectures": [
4
+ "BertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "directionality": "bidi",
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "pooler_fc_size": 768,
21
+ "pooler_num_attention_heads": 12,
22
+ "pooler_num_fc_layers": 3,
23
+ "pooler_size_per_head": 128,
24
+ "pooler_type": "first_token_transform",
25
+ "position_embedding_type": "absolute",
26
+ "transformers_version": "4.26.1",
27
+ "type_vocab_size": 2,
28
+ "use_cache": true,
29
+ "vocab_size": 105879
30
+ }
quantized/bert-base-multilingual-uncased/default/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b82ea957068e3acc6a14b2ea35bb8963a2fcd3128fe51defda4294af36b86b0
3
+ size 170119406
quantized/bert-base-multilingual-uncased/default/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
quantized/bert-base-multilingual-uncased/default/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
quantized/bert-base-multilingual-uncased/default/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": true,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 512,
6
+ "name_or_path": "bert-base-multilingual-uncased",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
quantized/bert-base-multilingual-uncased/default/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
quantized/bert-base-multilingual-uncased/masked-lm/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-multilingual-uncased",
3
+ "architectures": [
4
+ "BertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "directionality": "bidi",
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "pooler_fc_size": 768,
21
+ "pooler_num_attention_heads": 12,
22
+ "pooler_num_fc_layers": 3,
23
+ "pooler_size_per_head": 128,
24
+ "pooler_type": "first_token_transform",
25
+ "position_embedding_type": "absolute",
26
+ "transformers_version": "4.26.1",
27
+ "type_vocab_size": 2,
28
+ "use_cache": true,
29
+ "vocab_size": 105879
30
+ }
quantized/bert-base-multilingual-uncased/masked-lm/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbf2d7991514ea760db7432fc5efd84b01d5da01ba8e924fc21ca340338f23c7
3
+ size 250632605
quantized/bert-base-multilingual-uncased/masked-lm/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
quantized/bert-base-multilingual-uncased/masked-lm/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
quantized/bert-base-multilingual-uncased/masked-lm/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": true,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 512,
6
+ "name_or_path": "bert-base-multilingual-uncased",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
quantized/bert-base-multilingual-uncased/masked-lm/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
quantized/bert-base-multilingual-uncased/sequence-classification/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-multilingual-uncased",
3
+ "architectures": [
4
+ "BertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "directionality": "bidi",
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "pooler_fc_size": 768,
21
+ "pooler_num_attention_heads": 12,
22
+ "pooler_num_fc_layers": 3,
23
+ "pooler_size_per_head": 128,
24
+ "pooler_type": "first_token_transform",
25
+ "position_embedding_type": "absolute",
26
+ "transformers_version": "4.26.1",
27
+ "type_vocab_size": 2,
28
+ "use_cache": true,
29
+ "vocab_size": 105879
30
+ }
quantized/bert-base-multilingual-uncased/sequence-classification/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bee715fe94ecc95e65161aaf938af33e8a4323f8f50d66465f5b06f96d286c7e
3
+ size 170127027
quantized/bert-base-multilingual-uncased/sequence-classification/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
quantized/bert-base-multilingual-uncased/sequence-classification/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
quantized/bert-base-multilingual-uncased/sequence-classification/tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": true,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 512,
6
+ "name_or_path": "bert-base-multilingual-uncased",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
quantized/bert-base-multilingual-uncased/sequence-classification/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
quantized/bert-base-uncased/default/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-uncased",
3
+ "architectures": [
4
+ "BertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "transformers_version": "4.26.1",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 30522
25
+ }
quantized/bert-base-uncased/default/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:728f16136ce5f0457949c572d6153a2ccbd22795a457bcce748f4057c54e2100
3
+ size 112245230