asahi417 commited on
Commit
15a54fb
1 Parent(s): 81f7f5a
Files changed (2) hide show
  1. config.json +167 -0
  2. pytorch_model.bin +3 -0
config.json ADDED
@@ -0,0 +1,167 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "tner_ckpt/bc5cdr_deberta_v3_large/best_model",
3
+ "architectures": [
4
+ "DebertaV2ForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "crf_state_dict": {
8
+ "_constraint_mask": [
9
+ [
10
+ 1.0,
11
+ 1.0,
12
+ 1.0,
13
+ 0.0,
14
+ 0.0,
15
+ 0.0,
16
+ 1.0
17
+ ],
18
+ [
19
+ 1.0,
20
+ 1.0,
21
+ 1.0,
22
+ 0.0,
23
+ 1.0,
24
+ 0.0,
25
+ 1.0
26
+ ],
27
+ [
28
+ 1.0,
29
+ 1.0,
30
+ 1.0,
31
+ 1.0,
32
+ 0.0,
33
+ 0.0,
34
+ 1.0
35
+ ],
36
+ [
37
+ 1.0,
38
+ 1.0,
39
+ 1.0,
40
+ 1.0,
41
+ 0.0,
42
+ 0.0,
43
+ 1.0
44
+ ],
45
+ [
46
+ 1.0,
47
+ 1.0,
48
+ 1.0,
49
+ 0.0,
50
+ 1.0,
51
+ 0.0,
52
+ 1.0
53
+ ],
54
+ [
55
+ 1.0,
56
+ 1.0,
57
+ 1.0,
58
+ 0.0,
59
+ 0.0,
60
+ 0.0,
61
+ 0.0
62
+ ],
63
+ [
64
+ 0.0,
65
+ 0.0,
66
+ 0.0,
67
+ 0.0,
68
+ 0.0,
69
+ 0.0,
70
+ 0.0
71
+ ]
72
+ ],
73
+ "end_transitions": [
74
+ 1.0797829627990723,
75
+ 2.337775468826294,
76
+ -0.5702832341194153,
77
+ -0.09888350963592529,
78
+ 0.44394630193710327
79
+ ],
80
+ "start_transitions": [
81
+ 1.050474762916565,
82
+ -0.9998553395271301,
83
+ 0.6489607691764832,
84
+ -1.3081002235412598,
85
+ 0.8147163391113281
86
+ ],
87
+ "transitions": [
88
+ [
89
+ 0.38620230555534363,
90
+ 0.46338361501693726,
91
+ -0.2693702280521393,
92
+ -0.027202147990465164,
93
+ 0.2154194712638855
94
+ ],
95
+ [
96
+ -0.24631035327911377,
97
+ 0.7250676155090332,
98
+ -0.05480072647333145,
99
+ 0.05186671018600464,
100
+ 0.564211368560791
101
+ ],
102
+ [
103
+ 0.12546269595623016,
104
+ -0.22872965037822723,
105
+ -0.1702175885438919,
106
+ -0.40474018454551697,
107
+ 0.6381494402885437
108
+ ],
109
+ [
110
+ 0.5764445066452026,
111
+ 0.6380641460418701,
112
+ -0.6308369636535645,
113
+ 0.9163250923156738,
114
+ 0.7476255893707275
115
+ ],
116
+ [
117
+ -0.14198975265026093,
118
+ 0.41760584712028503,
119
+ 0.3615993857383728,
120
+ 0.5377429723739624,
121
+ 0.029010772705078125
122
+ ]
123
+ ]
124
+ },
125
+ "hidden_act": "gelu",
126
+ "hidden_dropout_prob": 0.1,
127
+ "hidden_size": 1024,
128
+ "id2label": {
129
+ "0": "O",
130
+ "1": "B-Chemical",
131
+ "2": "B-Disease",
132
+ "3": "I-Disease",
133
+ "4": "I-Chemical"
134
+ },
135
+ "initializer_range": 0.02,
136
+ "intermediate_size": 4096,
137
+ "label2id": {
138
+ "B-Chemical": 1,
139
+ "B-Disease": 2,
140
+ "I-Chemical": 4,
141
+ "I-Disease": 3,
142
+ "O": 0
143
+ },
144
+ "layer_norm_eps": 1e-07,
145
+ "max_position_embeddings": 512,
146
+ "max_relative_positions": -1,
147
+ "model_type": "deberta-v2",
148
+ "norm_rel_ebd": "layer_norm",
149
+ "num_attention_heads": 16,
150
+ "num_hidden_layers": 24,
151
+ "pad_token_id": 0,
152
+ "pooler_dropout": 0,
153
+ "pooler_hidden_act": "gelu",
154
+ "pooler_hidden_size": 1024,
155
+ "pos_att_type": [
156
+ "p2c",
157
+ "c2p"
158
+ ],
159
+ "position_biased_input": false,
160
+ "position_buckets": 256,
161
+ "relative_attention": true,
162
+ "share_att_key": true,
163
+ "torch_dtype": "float32",
164
+ "transformers_version": "4.20.1",
165
+ "type_vocab_size": 0,
166
+ "vocab_size": 128100
167
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48b16a3c69f179b4728b87bf14ccedfda1cbd4d741f56b524b6f4cc7e38b00a4
3
+ size 1736201135