harshildarji commited on
Commit
0f2b4ca
1 Parent(s): 1fc4e85
Files changed (8) hide show
  1. .DS_Store +0 -0
  2. README.md +6 -0
  3. config.json +73 -0
  4. merges.txt +0 -0
  5. model.safetensors +3 -0
  6. special_tokens_map.json +51 -0
  7. tokenizer_config.json +59 -0
  8. vocab.json +0 -0
.DS_Store ADDED
Binary file (6.15 kB). View file
 
README.md CHANGED
@@ -1,3 +1,9 @@
1
  ---
2
  license: mit
 
 
3
  ---
 
 
 
 
 
1
  ---
2
  license: mit
3
+ widget:
4
+ - text: "We may collect information about your activity on third-party services that use cookies and other technologies provided by us. We use this information to improve your user experience with our services and to improve your advertising services, including measuring the performance of ads and showing you relevant ads."
5
  ---
6
+
7
+ #### Performance:
8
+
9
+ ![model performance](https://i.imgur.com/xoqfljE.png)
config.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/deberta-base",
3
+ "architectures": [
4
+ "DebertaForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 768,
10
+ "id2label": {
11
+ "0": "receives_from",
12
+ "1": "utilises",
13
+ "2": "manages",
14
+ "3": "processes",
15
+ "4": "has_legal_basis",
16
+ "5": "ensures",
17
+ "6": "complies_with",
18
+ "7": "processes_for",
19
+ "8": "delegates",
20
+ "9": "retains",
21
+ "10": "oversees",
22
+ "11": "implements",
23
+ "12": "subjected_to",
24
+ "13": "appoints",
25
+ "14": "provides",
26
+ "15": "can_lodge",
27
+ "16": "transfers_to",
28
+ "17": "has_right"
29
+ },
30
+ "initializer_range": 0.02,
31
+ "intermediate_size": 3072,
32
+ "label2id": {
33
+ "appoints": 13,
34
+ "can_lodge": 15,
35
+ "complies_with": 6,
36
+ "delegates": 8,
37
+ "ensures": 5,
38
+ "has_legal_basis": 4,
39
+ "has_right": 17,
40
+ "implements": 11,
41
+ "manages": 2,
42
+ "oversees": 10,
43
+ "processes": 3,
44
+ "processes_for": 7,
45
+ "provides": 14,
46
+ "receives_from": 0,
47
+ "retains": 9,
48
+ "subjected_to": 12,
49
+ "transfers_to": 16,
50
+ "utilises": 1
51
+ },
52
+ "layer_norm_eps": 1e-07,
53
+ "max_position_embeddings": 512,
54
+ "max_relative_positions": -1,
55
+ "model_type": "deberta",
56
+ "num_attention_heads": 12,
57
+ "num_hidden_layers": 12,
58
+ "pad_token_id": 0,
59
+ "pooler_dropout": 0,
60
+ "pooler_hidden_act": "gelu",
61
+ "pooler_hidden_size": 768,
62
+ "pos_att_type": [
63
+ "c2p",
64
+ "p2c"
65
+ ],
66
+ "position_biased_input": false,
67
+ "problem_type": "multi_label_classification",
68
+ "relative_attention": true,
69
+ "torch_dtype": "float32",
70
+ "transformers_version": "4.41.2",
71
+ "type_vocab_size": 0,
72
+ "vocab_size": 50265
73
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:599f5301c71b34639b88818adbcf0c425da154df27599abbeac3b3bda2cf8c0e
3
+ size 556848784
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "[CLS]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "[SEP]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "[MASK]",
25
+ "lstrip": true,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "[PAD]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "[SEP]",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "[UNK]",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "[PAD]",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "[CLS]",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "[SEP]",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "3": {
30
+ "content": "[UNK]",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "50264": {
38
+ "content": "[MASK]",
39
+ "lstrip": true,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ }
45
+ },
46
+ "bos_token": "[CLS]",
47
+ "clean_up_tokenization_spaces": true,
48
+ "cls_token": "[CLS]",
49
+ "do_lower_case": false,
50
+ "eos_token": "[SEP]",
51
+ "errors": "replace",
52
+ "mask_token": "[MASK]",
53
+ "model_max_length": 1000000000000000019884624838656,
54
+ "pad_token": "[PAD]",
55
+ "sep_token": "[SEP]",
56
+ "tokenizer_class": "DebertaTokenizer",
57
+ "unk_token": "[UNK]",
58
+ "vocab_type": "gpt2"
59
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff