Text Classification
Transformers
PyTorch
ONNX
Safetensors
English
deberta
Trained with AutoTrain
Inference Endpoints
config.json CHANGED
@@ -1,6 +1,5 @@
1
  {
2
- "_attn_implementation_autoset": true,
3
- "_name_or_path": "KoalaAI/Text-Moderation",
4
  "_num_labels": 9,
5
  "architectures": [
6
  "DebertaForSequenceClassification"
@@ -51,7 +50,8 @@
51
  ],
52
  "position_biased_input": false,
53
  "relative_attention": true,
54
- "transformers_version": "4.46.1",
 
55
  "type_vocab_size": 0,
56
  "vocab_size": 50265
57
  }
 
1
  {
2
+ "_name_or_path": "AutoTrain",
 
3
  "_num_labels": 9,
4
  "architectures": [
5
  "DebertaForSequenceClassification"
 
50
  ],
51
  "position_biased_input": false,
52
  "relative_attention": true,
53
+ "torch_dtype": "float32",
54
+ "transformers_version": "4.29.2",
55
  "type_vocab_size": 0,
56
  "vocab_size": 50265
57
  }
onnx/model.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fad115a8cd91eb803fb278249024fd9be85ec3566e996a19ba0c3d55f49379c5
3
- size 557224900
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:697d2cce4b6468b385f45b5b6504767dd7ed663f6d2abf4522d3350ee88ea9c7
3
+ size 557347903
onnx/model_bnb4.onnx DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b5245778f5c4f80f3ffa9bc756a6e07b2749b6bfe24b7cdfb513d756c687085
3
- size 216612754
 
 
 
 
onnx/model_fp16.onnx DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3d5bb69d6f7ea1b3fe17aa70fea1498b0752e6380c5502de588dee781ca55c2c
3
- size 278826418
 
 
 
 
onnx/model_int8.onnx DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f453d21e8e4dd42acf90bd8cc67e04ca1490b5987279778df65208d05e45f5c1
3
- size 142967535
 
 
 
 
onnx/model_q4.onnx DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d9c8d652d6cc8c599dcc7131421fea58bf6af0cf5c6544a3a474bce2f2a1a270
3
- size 222805366
 
 
 
 
onnx/model_q4f16.onnx DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f6afc17ea9577a47c0bcb14f0e23a6101ea43b5af5c08c89b8099de5fc35b50f
3
- size 136394668
 
 
 
 
onnx/model_quantized.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f453d21e8e4dd42acf90bd8cc67e04ca1490b5987279778df65208d05e45f5c1
3
- size 142967535
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28dcab041de741692239119e5967efcb197f1d4a1d1da3713bbbeee3a3e139a6
3
+ size 143281992
onnx/model_uint8.onnx DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6584dc5c064021c74314e463b11c1af853df4c662f28fce3db2af423ad31f977
3
- size 142967535
 
 
 
 
quantize_config.json DELETED
@@ -1,17 +0,0 @@
1
- {
2
- "modes": [
3
- "fp16",
4
- "q8",
5
- "int8",
6
- "uint8",
7
- "q4",
8
- "q4f16",
9
- "bnb4"
10
- ],
11
- "per_channel": true,
12
- "reduce_range": true,
13
- "block_size": null,
14
- "is_symmetric": true,
15
- "accuracy_level": null,
16
- "quant_type": 1
17
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:02eccb4743ef6eeb5bf1fa77bca58cf5a42094ec91852a767e3825b1617f550b
3
- size 3559904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc39e4a9ef588149e8513a2d7e0d1cda450b1884a5f6c62a945d9ec1e5bdbfe4
3
+ size 2109876
tokenizer_config.json CHANGED
@@ -1,66 +1,66 @@
1
  {
2
  "add_bos_token": false,
3
  "add_prefix_space": false,
4
- "added_tokens_decoder": {
5
- "0": {
6
- "content": "[PAD]",
7
- "lstrip": false,
8
- "normalized": true,
9
- "rstrip": false,
10
- "single_word": false,
11
- "special": true
12
- },
13
- "1": {
14
- "content": "[CLS]",
15
- "lstrip": false,
16
- "normalized": true,
17
- "rstrip": false,
18
- "single_word": false,
19
- "special": true
20
- },
21
- "2": {
22
- "content": "[SEP]",
23
- "lstrip": false,
24
- "normalized": true,
25
- "rstrip": false,
26
- "single_word": false,
27
- "special": true
28
- },
29
- "3": {
30
- "content": "[UNK]",
31
- "lstrip": false,
32
- "normalized": true,
33
- "rstrip": false,
34
- "single_word": false,
35
- "special": true
36
- },
37
- "50264": {
38
- "content": "[MASK]",
39
- "lstrip": true,
40
- "normalized": true,
41
- "rstrip": false,
42
- "single_word": false,
43
- "special": true
44
- }
45
  },
46
- "bos_token": "[CLS]",
47
  "clean_up_tokenization_spaces": true,
48
- "cls_token": "[CLS]",
 
 
 
 
 
 
 
49
  "do_lower_case": false,
50
- "eos_token": "[SEP]",
 
 
 
 
 
 
 
51
  "errors": "replace",
52
- "mask_token": "[MASK]",
53
- "max_length": 384,
 
 
 
 
 
 
54
  "model_max_length": 512,
55
- "pad_to_multiple_of": null,
56
- "pad_token": "[PAD]",
57
- "pad_token_type_id": 0,
58
- "padding_side": "right",
59
- "sep_token": "[SEP]",
60
- "stride": 0,
 
 
 
 
 
 
 
 
 
 
61
  "tokenizer_class": "DebertaTokenizer",
62
- "truncation_side": "right",
63
- "truncation_strategy": "longest_first",
64
- "unk_token": "[UNK]",
 
 
 
 
 
65
  "vocab_type": "gpt2"
66
  }
 
1
  {
2
  "add_bos_token": false,
3
  "add_prefix_space": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "[CLS]",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  },
 
12
  "clean_up_tokenization_spaces": true,
13
+ "cls_token": {
14
+ "__type": "AddedToken",
15
+ "content": "[CLS]",
16
+ "lstrip": false,
17
+ "normalized": true,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
  "do_lower_case": false,
22
+ "eos_token": {
23
+ "__type": "AddedToken",
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
  "errors": "replace",
31
+ "mask_token": {
32
+ "__type": "AddedToken",
33
+ "content": "[MASK]",
34
+ "lstrip": true,
35
+ "normalized": true,
36
+ "rstrip": false,
37
+ "single_word": false
38
+ },
39
  "model_max_length": 512,
40
+ "pad_token": {
41
+ "__type": "AddedToken",
42
+ "content": "[PAD]",
43
+ "lstrip": false,
44
+ "normalized": true,
45
+ "rstrip": false,
46
+ "single_word": false
47
+ },
48
+ "sep_token": {
49
+ "__type": "AddedToken",
50
+ "content": "[SEP]",
51
+ "lstrip": false,
52
+ "normalized": true,
53
+ "rstrip": false,
54
+ "single_word": false
55
+ },
56
  "tokenizer_class": "DebertaTokenizer",
57
+ "unk_token": {
58
+ "__type": "AddedToken",
59
+ "content": "[UNK]",
60
+ "lstrip": false,
61
+ "normalized": true,
62
+ "rstrip": false,
63
+ "single_word": false
64
+ },
65
  "vocab_type": "gpt2"
66
  }