KoichiYasuoka commited on
Commit
5821aad
1 Parent(s): 2496be9

model improved

Browse files
config.json CHANGED
@@ -20,12 +20,15 @@
20
  "pooler_dropout": 0,
21
  "pooler_hidden_act": "gelu",
22
  "pooler_hidden_size": 1024,
23
- "pos_att_type": null,
24
- "position_biased_input": true,
25
- "relative_attention": false,
 
 
 
26
  "tokenizer_class": "DebertaV2TokenizerFast",
27
  "torch_dtype": "float32",
28
- "transformers_version": "4.19.4",
29
  "type_vocab_size": 0,
30
  "vocab_size": 32000
31
  }
20
  "pooler_dropout": 0,
21
  "pooler_hidden_act": "gelu",
22
  "pooler_hidden_size": 1024,
23
+ "pos_att_type": [
24
+ "p2c",
25
+ "c2p"
26
+ ],
27
+ "position_biased_input": false,
28
+ "relative_attention": true,
29
  "tokenizer_class": "DebertaV2TokenizerFast",
30
  "torch_dtype": "float32",
31
+ "transformers_version": "4.22.1",
32
  "type_vocab_size": 0,
33
  "vocab_size": 32000
34
  }
deprel/config.json CHANGED
@@ -5,7 +5,6 @@
5
  "attention_probs_dropout_prob": 0.1,
6
  "bos_token_id": 0,
7
  "eos_token_id": 2,
8
- "finetuning_task": "pos",
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 1024,
@@ -21,43 +20,45 @@
21
  "8": "B-compound",
22
  "9": "B-cop",
23
  "10": "B-csubj",
24
- "11": "B-dep",
25
- "12": "B-det",
26
- "13": "B-discourse",
27
- "14": "B-dislocated",
28
  "15": "B-fixed",
29
  "16": "B-mark",
30
  "17": "B-nmod",
31
  "18": "B-nsubj",
32
- "19": "B-nummod",
33
- "20": "B-obj",
34
- "21": "B-obl",
35
- "22": "B-punct",
36
- "23": "B-root",
37
- "24": "I-acl",
38
- "25": "I-advcl",
39
- "26": "I-advmod",
40
- "27": "I-amod",
41
- "28": "I-aux",
42
- "29": "I-case",
43
- "30": "I-cc",
44
- "31": "I-ccomp",
45
- "32": "I-compound",
46
- "33": "I-cop",
47
- "34": "I-csubj",
48
- "35": "I-dep",
49
- "36": "I-det",
50
- "37": "I-discourse",
51
- "38": "I-dislocated",
52
- "39": "I-fixed",
53
- "40": "I-mark",
54
- "41": "I-nmod",
55
- "42": "I-nsubj",
56
- "43": "I-nummod",
57
- "44": "I-obj",
58
- "45": "I-obl",
59
- "46": "I-punct",
60
- "47": "I-root"
 
 
61
  },
62
  "initializer_range": 0.02,
63
  "intermediate_size": 4096,
@@ -73,43 +74,45 @@
73
  "B-compound": 8,
74
  "B-cop": 9,
75
  "B-csubj": 10,
76
- "B-dep": 11,
77
- "B-det": 12,
78
- "B-discourse": 13,
79
- "B-dislocated": 14,
80
  "B-fixed": 15,
81
  "B-mark": 16,
82
  "B-nmod": 17,
83
  "B-nsubj": 18,
84
- "B-nummod": 19,
85
- "B-obj": 20,
86
- "B-obl": 21,
87
- "B-punct": 22,
88
- "B-root": 23,
89
- "I-acl": 24,
90
- "I-advcl": 25,
91
- "I-advmod": 26,
92
- "I-amod": 27,
93
- "I-aux": 28,
94
- "I-case": 29,
95
- "I-cc": 30,
96
- "I-ccomp": 31,
97
- "I-compound": 32,
98
- "I-cop": 33,
99
- "I-csubj": 34,
100
- "I-dep": 35,
101
- "I-det": 36,
102
- "I-discourse": 37,
103
- "I-dislocated": 38,
104
- "I-fixed": 39,
105
- "I-mark": 40,
106
- "I-nmod": 41,
107
- "I-nsubj": 42,
108
- "I-nummod": 43,
109
- "I-obj": 44,
110
- "I-obl": 45,
111
- "I-punct": 46,
112
- "I-root": 47
 
 
113
  },
114
  "layer_norm_eps": 1e-07,
115
  "max_position_embeddings": 512,
@@ -121,12 +124,15 @@
121
  "pooler_dropout": 0,
122
  "pooler_hidden_act": "gelu",
123
  "pooler_hidden_size": 1024,
124
- "pos_att_type": null,
125
- "position_biased_input": true,
126
- "relative_attention": false,
 
 
 
127
  "tokenizer_class": "DebertaV2TokenizerFast",
128
  "torch_dtype": "float32",
129
- "transformers_version": "4.19.4",
130
  "type_vocab_size": 0,
131
  "vocab_size": 32000
132
  }
5
  "attention_probs_dropout_prob": 0.1,
6
  "bos_token_id": 0,
7
  "eos_token_id": 2,
 
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.1,
10
  "hidden_size": 1024,
20
  "8": "B-compound",
21
  "9": "B-cop",
22
  "10": "B-csubj",
23
+ "11": "B-csubj:outer",
24
+ "12": "B-dep",
25
+ "13": "B-det",
26
+ "14": "B-discourse",
27
  "15": "B-fixed",
28
  "16": "B-mark",
29
  "17": "B-nmod",
30
  "18": "B-nsubj",
31
+ "19": "B-nsubj:outer",
32
+ "20": "B-nummod",
33
+ "21": "B-obj",
34
+ "22": "B-obl",
35
+ "23": "B-punct",
36
+ "24": "B-root",
37
+ "25": "I-acl",
38
+ "26": "I-advcl",
39
+ "27": "I-advmod",
40
+ "28": "I-amod",
41
+ "29": "I-aux",
42
+ "30": "I-case",
43
+ "31": "I-cc",
44
+ "32": "I-ccomp",
45
+ "33": "I-compound",
46
+ "34": "I-cop",
47
+ "35": "I-csubj",
48
+ "36": "I-csubj:outer",
49
+ "37": "I-dep",
50
+ "38": "I-det",
51
+ "39": "I-discourse",
52
+ "40": "I-fixed",
53
+ "41": "I-mark",
54
+ "42": "I-nmod",
55
+ "43": "I-nsubj",
56
+ "44": "I-nsubj:outer",
57
+ "45": "I-nummod",
58
+ "46": "I-obj",
59
+ "47": "I-obl",
60
+ "48": "I-punct",
61
+ "49": "I-root"
62
  },
63
  "initializer_range": 0.02,
64
  "intermediate_size": 4096,
74
  "B-compound": 8,
75
  "B-cop": 9,
76
  "B-csubj": 10,
77
+ "B-csubj:outer": 11,
78
+ "B-dep": 12,
79
+ "B-det": 13,
80
+ "B-discourse": 14,
81
  "B-fixed": 15,
82
  "B-mark": 16,
83
  "B-nmod": 17,
84
  "B-nsubj": 18,
85
+ "B-nsubj:outer": 19,
86
+ "B-nummod": 20,
87
+ "B-obj": 21,
88
+ "B-obl": 22,
89
+ "B-punct": 23,
90
+ "B-root": 24,
91
+ "I-acl": 25,
92
+ "I-advcl": 26,
93
+ "I-advmod": 27,
94
+ "I-amod": 28,
95
+ "I-aux": 29,
96
+ "I-case": 30,
97
+ "I-cc": 31,
98
+ "I-ccomp": 32,
99
+ "I-compound": 33,
100
+ "I-cop": 34,
101
+ "I-csubj": 35,
102
+ "I-csubj:outer": 36,
103
+ "I-dep": 37,
104
+ "I-det": 38,
105
+ "I-discourse": 39,
106
+ "I-fixed": 40,
107
+ "I-mark": 41,
108
+ "I-nmod": 42,
109
+ "I-nsubj": 43,
110
+ "I-nsubj:outer": 44,
111
+ "I-nummod": 45,
112
+ "I-obj": 46,
113
+ "I-obl": 47,
114
+ "I-punct": 48,
115
+ "I-root": 49
116
  },
117
  "layer_norm_eps": 1e-07,
118
  "max_position_embeddings": 512,
124
  "pooler_dropout": 0,
125
  "pooler_hidden_act": "gelu",
126
  "pooler_hidden_size": 1024,
127
+ "pos_att_type": [
128
+ "p2c",
129
+ "c2p"
130
+ ],
131
+ "position_biased_input": false,
132
+ "relative_attention": true,
133
  "tokenizer_class": "DebertaV2TokenizerFast",
134
  "torch_dtype": "float32",
135
+ "transformers_version": "4.22.1",
136
  "type_vocab_size": 0,
137
  "vocab_size": 32000
138
  }
deprel/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef2d5e996deaf7254efc9261b9abb01ffc068f0a12d217da7a2737364971b3b4
3
- size 1342748467
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:559a6b3b5670fa1151f701d9b47155793669c10d487ac8c63bfa29a6d8baee87
3
+ size 1546410675
deprel/special_tokens_map.json CHANGED
@@ -1 +1,9 @@
1
- {"bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": "[UNK]"
9
+ }
deprel/tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
deprel/tokenizer_config.json CHANGED
@@ -1 +1,14 @@
1
- {"do_lower_case": false, "bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "split_by_punct": true, "keep_accents": true, "model_max_length": 512, "tokenizer_class": "DebertaV2TokenizerFast"}
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": false,
5
+ "eos_token": "[SEP]",
6
+ "keep_accents": true,
7
+ "mask_token": "[MASK]",
8
+ "model_max_length": 512,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "split_by_punct": true,
12
+ "tokenizer_class": "DebertaV2TokenizerFast",
13
+ "unk_token": "[UNK]"
14
+ }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5953ef557ad39cdc7a2a7d83f820b72104e70d314d83f2d371f42d6003ec40d3
3
- size 1342559923
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0eb71213259b99844953acf8fb4a2ce8bcb0f88b172a5a78c5557cee487b973c
3
+ size 1546213875
special_tokens_map.json CHANGED
@@ -1 +1,9 @@
1
- {"bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": "[UNK]"
9
+ }
tagger/config.json CHANGED
@@ -5,7 +5,6 @@
5
  "attention_probs_dropout_prob": 0.1,
6
  "bos_token_id": 0,
7
  "eos_token_id": 2,
8
- "finetuning_task": "pos",
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 1024,
@@ -101,12 +100,15 @@
101
  "pooler_dropout": 0,
102
  "pooler_hidden_act": "gelu",
103
  "pooler_hidden_size": 1024,
104
- "pos_att_type": null,
105
- "position_biased_input": true,
106
- "relative_attention": false,
 
 
 
107
  "tokenizer_class": "DebertaV2TokenizerFast",
108
  "torch_dtype": "float32",
109
- "transformers_version": "4.19.4",
110
  "type_vocab_size": 0,
111
  "vocab_size": 32000
112
  }
5
  "attention_probs_dropout_prob": 0.1,
6
  "bos_token_id": 0,
7
  "eos_token_id": 2,
 
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.1,
10
  "hidden_size": 1024,
100
  "pooler_dropout": 0,
101
  "pooler_hidden_act": "gelu",
102
  "pooler_hidden_size": 1024,
103
+ "pos_att_type": [
104
+ "p2c",
105
+ "c2p"
106
+ ],
107
+ "position_biased_input": false,
108
+ "relative_attention": true,
109
  "tokenizer_class": "DebertaV2TokenizerFast",
110
  "torch_dtype": "float32",
111
+ "transformers_version": "4.22.1",
112
  "type_vocab_size": 0,
113
  "vocab_size": 32000
114
  }
tagger/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6ce99dd88b1df984c8969c5238679936bd9733ceb4111379e06d623c4e58175d
3
- size 1342707507
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb18628c956c64104d2e34a55be4d94084b0305728ead8117bf31233ded6cd12
3
+ size 1546361459
tagger/special_tokens_map.json CHANGED
@@ -1 +1,9 @@
1
- {"bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": "[UNK]"
9
+ }
tagger/tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
tagger/tokenizer_config.json CHANGED
@@ -1 +1,14 @@
1
- {"do_lower_case": false, "bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "split_by_punct": true, "keep_accents": true, "model_max_length": 512, "tokenizer_class": "DebertaV2TokenizerFast"}
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": false,
5
+ "eos_token": "[SEP]",
6
+ "keep_accents": true,
7
+ "mask_token": "[MASK]",
8
+ "model_max_length": 512,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "split_by_punct": true,
12
+ "tokenizer_class": "DebertaV2TokenizerFast",
13
+ "unk_token": "[UNK]"
14
+ }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
tokenizer_config.json CHANGED
@@ -1 +1,14 @@
1
- {"do_lower_case": false, "bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "split_by_punct": true, "keep_accents": true, "model_max_length": 512, "tokenizer_class": "DebertaV2TokenizerFast"}
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "do_lower_case": false,
5
+ "eos_token": "[SEP]",
6
+ "keep_accents": true,
7
+ "mask_token": "[MASK]",
8
+ "model_max_length": 512,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "split_by_punct": true,
12
+ "tokenizer_class": "DebertaV2TokenizerFast",
13
+ "unk_token": "[UNK]"
14
+ }