asahi417 commited on
Commit
6bd665d
1 Parent(s): 1698d97
Files changed (2) hide show
  1. config.json +319 -0
  2. pytorch_model.bin +3 -0
config.json ADDED
@@ -0,0 +1,319 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "tner_ckpt/conll2003_roberta_large/best_model",
3
+ "architectures": [
4
+ "RobertaForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "crf_state_dict": {
10
+ "_constraint_mask": [
11
+ [
12
+ 1.0,
13
+ 1.0,
14
+ 1.0,
15
+ 1.0,
16
+ 0.0,
17
+ 1.0,
18
+ 0.0,
19
+ 0.0,
20
+ 0.0,
21
+ 0.0,
22
+ 1.0
23
+ ],
24
+ [
25
+ 1.0,
26
+ 1.0,
27
+ 1.0,
28
+ 1.0,
29
+ 0.0,
30
+ 1.0,
31
+ 1.0,
32
+ 0.0,
33
+ 0.0,
34
+ 0.0,
35
+ 1.0
36
+ ],
37
+ [
38
+ 1.0,
39
+ 1.0,
40
+ 1.0,
41
+ 1.0,
42
+ 0.0,
43
+ 1.0,
44
+ 0.0,
45
+ 1.0,
46
+ 0.0,
47
+ 0.0,
48
+ 1.0
49
+ ],
50
+ [
51
+ 1.0,
52
+ 1.0,
53
+ 1.0,
54
+ 1.0,
55
+ 1.0,
56
+ 1.0,
57
+ 0.0,
58
+ 0.0,
59
+ 0.0,
60
+ 0.0,
61
+ 1.0
62
+ ],
63
+ [
64
+ 1.0,
65
+ 1.0,
66
+ 1.0,
67
+ 1.0,
68
+ 1.0,
69
+ 1.0,
70
+ 0.0,
71
+ 0.0,
72
+ 0.0,
73
+ 0.0,
74
+ 1.0
75
+ ],
76
+ [
77
+ 1.0,
78
+ 1.0,
79
+ 1.0,
80
+ 1.0,
81
+ 0.0,
82
+ 1.0,
83
+ 0.0,
84
+ 0.0,
85
+ 1.0,
86
+ 0.0,
87
+ 1.0
88
+ ],
89
+ [
90
+ 1.0,
91
+ 1.0,
92
+ 1.0,
93
+ 1.0,
94
+ 0.0,
95
+ 1.0,
96
+ 1.0,
97
+ 0.0,
98
+ 0.0,
99
+ 0.0,
100
+ 1.0
101
+ ],
102
+ [
103
+ 1.0,
104
+ 1.0,
105
+ 1.0,
106
+ 1.0,
107
+ 0.0,
108
+ 1.0,
109
+ 0.0,
110
+ 1.0,
111
+ 0.0,
112
+ 0.0,
113
+ 1.0
114
+ ],
115
+ [
116
+ 1.0,
117
+ 1.0,
118
+ 1.0,
119
+ 1.0,
120
+ 0.0,
121
+ 1.0,
122
+ 0.0,
123
+ 0.0,
124
+ 1.0,
125
+ 0.0,
126
+ 1.0
127
+ ],
128
+ [
129
+ 1.0,
130
+ 1.0,
131
+ 1.0,
132
+ 1.0,
133
+ 0.0,
134
+ 1.0,
135
+ 0.0,
136
+ 0.0,
137
+ 0.0,
138
+ 0.0,
139
+ 0.0
140
+ ],
141
+ [
142
+ 0.0,
143
+ 0.0,
144
+ 0.0,
145
+ 0.0,
146
+ 0.0,
147
+ 0.0,
148
+ 0.0,
149
+ 0.0,
150
+ 0.0,
151
+ 0.0,
152
+ 0.0
153
+ ]
154
+ ],
155
+ "end_transitions": [
156
+ 1.6543885469436646,
157
+ 1.4006357192993164,
158
+ 0.4444950819015503,
159
+ -0.8417924046516418,
160
+ -1.5104246139526367,
161
+ 3.0435879230499268,
162
+ 2.1962571144104004,
163
+ -0.8038378357887268,
164
+ -1.0722748041152954
165
+ ],
166
+ "start_transitions": [
167
+ 1.2061455249786377,
168
+ 0.6586089134216309,
169
+ 0.4692096412181854,
170
+ 0.5858622789382935,
171
+ -0.630755603313446,
172
+ -0.32020196318626404,
173
+ -0.5469271540641785,
174
+ -0.3013189435005188,
175
+ -0.14686475694179535
176
+ ],
177
+ "transitions": [
178
+ [
179
+ 0.13537588715553284,
180
+ -0.13092884421348572,
181
+ 0.4625251293182373,
182
+ 0.23847511410713196,
183
+ 0.34067708253860474,
184
+ 0.12975099682807922,
185
+ 0.5387805700302124,
186
+ 0.09070442616939545,
187
+ 0.3408451974391937
188
+ ],
189
+ [
190
+ -0.14204159379005432,
191
+ 0.37123164534568787,
192
+ -0.21415242552757263,
193
+ 0.011411471292376518,
194
+ -0.4560433328151703,
195
+ -0.34037908911705017,
196
+ -0.2033482789993286,
197
+ -0.906518816947937,
198
+ -0.47004014253616333
199
+ ],
200
+ [
201
+ -0.17460417747497559,
202
+ -0.1531568169593811,
203
+ -0.006479573901742697,
204
+ -0.4537687301635742,
205
+ -0.050648175179958344,
206
+ 0.38808903098106384,
207
+ -0.16089072823524475,
208
+ -0.04831279069185257,
209
+ 0.05039094015955925
210
+ ],
211
+ [
212
+ -0.11159854382276535,
213
+ -0.41543281078338623,
214
+ 0.555843710899353,
215
+ 0.014951681718230247,
216
+ 0.08906713128089905,
217
+ -0.13042980432510376,
218
+ -0.09327378123998642,
219
+ 0.5290833711624146,
220
+ 0.10889127105474472
221
+ ],
222
+ [
223
+ 0.33919236063957214,
224
+ -0.4331152141094208,
225
+ -0.38547754287719727,
226
+ 0.32877588272094727,
227
+ -0.5369617342948914,
228
+ 0.07536065578460693,
229
+ -0.31710729002952576,
230
+ 0.037764519453048706,
231
+ -0.3209592401981354
232
+ ],
233
+ [
234
+ 0.027896448969841003,
235
+ 0.2133970558643341,
236
+ 0.5522003173828125,
237
+ -0.2629242539405823,
238
+ -0.01897781528532505,
239
+ -0.07080112397670746,
240
+ -0.8012456297874451,
241
+ -0.007121096830815077,
242
+ 0.5570559501647949
243
+ ],
244
+ [
245
+ -0.130000039935112,
246
+ -0.26081007719039917,
247
+ -0.159550741314888,
248
+ -0.4839116334915161,
249
+ 0.49926015734672546,
250
+ 0.2511492371559143,
251
+ 0.014935985207557678,
252
+ -0.04589146375656128,
253
+ 0.4296587407588959
254
+ ],
255
+ [
256
+ 0.17752425372600555,
257
+ -0.0608215294778347,
258
+ 0.03037858009338379,
259
+ 0.09913276135921478,
260
+ 0.2518308758735657,
261
+ -0.5100915431976318,
262
+ 0.1673133820295334,
263
+ -0.06604345887899399,
264
+ 0.18997260928153992
265
+ ],
266
+ [
267
+ -0.2999376654624939,
268
+ 0.03706861287355423,
269
+ -0.3692086338996887,
270
+ 0.14070719480514526,
271
+ -0.05091197043657303,
272
+ 0.14978167414665222,
273
+ 0.547923743724823,
274
+ 0.015628628432750702,
275
+ 0.17064540088176727
276
+ ]
277
+ ]
278
+ },
279
+ "eos_token_id": 2,
280
+ "hidden_act": "gelu",
281
+ "hidden_dropout_prob": 0.1,
282
+ "hidden_size": 1024,
283
+ "id2label": {
284
+ "0": "O",
285
+ "1": "B-ORG",
286
+ "2": "B-MISC",
287
+ "3": "B-PER",
288
+ "4": "I-PER",
289
+ "5": "B-LOC",
290
+ "6": "I-ORG",
291
+ "7": "I-MISC",
292
+ "8": "I-LOC"
293
+ },
294
+ "initializer_range": 0.02,
295
+ "intermediate_size": 4096,
296
+ "label2id": {
297
+ "B-LOC": 5,
298
+ "B-MISC": 2,
299
+ "B-ORG": 1,
300
+ "B-PER": 3,
301
+ "I-LOC": 8,
302
+ "I-MISC": 7,
303
+ "I-ORG": 6,
304
+ "I-PER": 4,
305
+ "O": 0
306
+ },
307
+ "layer_norm_eps": 1e-05,
308
+ "max_position_embeddings": 514,
309
+ "model_type": "roberta",
310
+ "num_attention_heads": 16,
311
+ "num_hidden_layers": 24,
312
+ "pad_token_id": 1,
313
+ "position_embedding_type": "absolute",
314
+ "torch_dtype": "float32",
315
+ "transformers_version": "4.20.1",
316
+ "type_vocab_size": 1,
317
+ "use_cache": true,
318
+ "vocab_size": 50265
319
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9209b8f0b9bee7269c6c02648dd604825e787ac03776ce8335018adcb309d9f
3
+ size 1417408561