Text Classification
Transformers
PyTorch
ONNX
Safetensors
English
deberta
Trained with AutoTrain
Inference Endpoints
DarwinAnim8or commited on
Commit
838ade0
·
verified ·
1 Parent(s): f3015b1

Quantized Models (#9)

Browse files

- updated quantized models (a0604e28df2d1ed8f79810872b85c027f02945ce)

config.json CHANGED
@@ -1,5 +1,6 @@
1
  {
2
- "_name_or_path": "AutoTrain",
 
3
  "_num_labels": 9,
4
  "architectures": [
5
  "DebertaForSequenceClassification"
@@ -50,8 +51,7 @@
50
  ],
51
  "position_biased_input": false,
52
  "relative_attention": true,
53
- "torch_dtype": "float32",
54
- "transformers_version": "4.29.2",
55
  "type_vocab_size": 0,
56
  "vocab_size": 50265
57
  }
 
1
  {
2
+ "_attn_implementation_autoset": true,
3
+ "_name_or_path": "KoalaAI/Text-Moderation",
4
  "_num_labels": 9,
5
  "architectures": [
6
  "DebertaForSequenceClassification"
 
51
  ],
52
  "position_biased_input": false,
53
  "relative_attention": true,
54
+ "transformers_version": "4.46.1",
 
55
  "type_vocab_size": 0,
56
  "vocab_size": 50265
57
  }
onnx/model.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:697d2cce4b6468b385f45b5b6504767dd7ed663f6d2abf4522d3350ee88ea9c7
3
- size 557347903
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fad115a8cd91eb803fb278249024fd9be85ec3566e996a19ba0c3d55f49379c5
3
+ size 557224900
onnx/model_bnb4.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b5245778f5c4f80f3ffa9bc756a6e07b2749b6bfe24b7cdfb513d756c687085
3
+ size 216612754
onnx/model_fp16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d5bb69d6f7ea1b3fe17aa70fea1498b0752e6380c5502de588dee781ca55c2c
3
+ size 278826418
onnx/model_int8.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f453d21e8e4dd42acf90bd8cc67e04ca1490b5987279778df65208d05e45f5c1
3
+ size 142967535
onnx/model_q4.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9c8d652d6cc8c599dcc7131421fea58bf6af0cf5c6544a3a474bce2f2a1a270
3
+ size 222805366
onnx/model_q4f16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6afc17ea9577a47c0bcb14f0e23a6101ea43b5af5c08c89b8099de5fc35b50f
3
+ size 136394668
onnx/model_quantized.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:28dcab041de741692239119e5967efcb197f1d4a1d1da3713bbbeee3a3e139a6
3
- size 143281992
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f453d21e8e4dd42acf90bd8cc67e04ca1490b5987279778df65208d05e45f5c1
3
+ size 142967535
onnx/model_uint8.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6584dc5c064021c74314e463b11c1af853df4c662f28fce3db2af423ad31f977
3
+ size 142967535
quantize_config.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "modes": [
3
+ "fp16",
4
+ "q8",
5
+ "int8",
6
+ "uint8",
7
+ "q4",
8
+ "q4f16",
9
+ "bnb4"
10
+ ],
11
+ "per_channel": true,
12
+ "reduce_range": true,
13
+ "block_size": null,
14
+ "is_symmetric": true,
15
+ "accuracy_level": null,
16
+ "quant_type": 1
17
+ }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc39e4a9ef588149e8513a2d7e0d1cda450b1884a5f6c62a945d9ec1e5bdbfe4
3
- size 2109876
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02eccb4743ef6eeb5bf1fa77bca58cf5a42094ec91852a767e3825b1617f550b
3
+ size 3559904
tokenizer_config.json CHANGED
@@ -1,66 +1,66 @@
1
  {
2
  "add_bos_token": false,
3
  "add_prefix_space": false,
4
- "bos_token": {
5
- "__type": "AddedToken",
6
- "content": "[CLS]",
7
- "lstrip": false,
8
- "normalized": true,
9
- "rstrip": false,
10
- "single_word": false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  },
 
12
  "clean_up_tokenization_spaces": true,
13
- "cls_token": {
14
- "__type": "AddedToken",
15
- "content": "[CLS]",
16
- "lstrip": false,
17
- "normalized": true,
18
- "rstrip": false,
19
- "single_word": false
20
- },
21
  "do_lower_case": false,
22
- "eos_token": {
23
- "__type": "AddedToken",
24
- "content": "[SEP]",
25
- "lstrip": false,
26
- "normalized": true,
27
- "rstrip": false,
28
- "single_word": false
29
- },
30
  "errors": "replace",
31
- "mask_token": {
32
- "__type": "AddedToken",
33
- "content": "[MASK]",
34
- "lstrip": true,
35
- "normalized": true,
36
- "rstrip": false,
37
- "single_word": false
38
- },
39
  "model_max_length": 512,
40
- "pad_token": {
41
- "__type": "AddedToken",
42
- "content": "[PAD]",
43
- "lstrip": false,
44
- "normalized": true,
45
- "rstrip": false,
46
- "single_word": false
47
- },
48
- "sep_token": {
49
- "__type": "AddedToken",
50
- "content": "[SEP]",
51
- "lstrip": false,
52
- "normalized": true,
53
- "rstrip": false,
54
- "single_word": false
55
- },
56
  "tokenizer_class": "DebertaTokenizer",
57
- "unk_token": {
58
- "__type": "AddedToken",
59
- "content": "[UNK]",
60
- "lstrip": false,
61
- "normalized": true,
62
- "rstrip": false,
63
- "single_word": false
64
- },
65
  "vocab_type": "gpt2"
66
  }
 
1
  {
2
  "add_bos_token": false,
3
  "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "[PAD]",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "[CLS]",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "[SEP]",
23
+ "lstrip": false,
24
+ "normalized": true,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "3": {
30
+ "content": "[UNK]",
31
+ "lstrip": false,
32
+ "normalized": true,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "50264": {
38
+ "content": "[MASK]",
39
+ "lstrip": true,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ }
45
  },
46
+ "bos_token": "[CLS]",
47
  "clean_up_tokenization_spaces": true,
48
+ "cls_token": "[CLS]",
 
 
 
 
 
 
 
49
  "do_lower_case": false,
50
+ "eos_token": "[SEP]",
 
 
 
 
 
 
 
51
  "errors": "replace",
52
+ "mask_token": "[MASK]",
53
+ "max_length": 384,
 
 
 
 
 
 
54
  "model_max_length": 512,
55
+ "pad_to_multiple_of": null,
56
+ "pad_token": "[PAD]",
57
+ "pad_token_type_id": 0,
58
+ "padding_side": "right",
59
+ "sep_token": "[SEP]",
60
+ "stride": 0,
 
 
 
 
 
 
 
 
 
 
61
  "tokenizer_class": "DebertaTokenizer",
62
+ "truncation_side": "right",
63
+ "truncation_strategy": "longest_first",
64
+ "unk_token": "[UNK]",
 
 
 
 
 
65
  "vocab_type": "gpt2"
66
  }