madlag commited on
Commit
98f6bb5
1 Parent(s): 7e06c87

Initial commit.

Browse files
config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-large-uncased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "finetuning_task": "mnli",
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 1024,
12
+ "id2label": {
13
+ "0": "LABEL_0",
14
+ "1": "LABEL_1",
15
+ "2": "LABEL_2"
16
+ },
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 4096,
19
+ "label2id": {
20
+ "LABEL_0": 0,
21
+ "LABEL_1": 1,
22
+ "LABEL_2": 2
23
+ },
24
+ "layer_norm_eps": 1e-12,
25
+ "max_position_embeddings": 512,
26
+ "model_type": "bert",
27
+ "num_attention_heads": 16,
28
+ "num_hidden_layers": 24,
29
+ "pad_token_id": 0,
30
+ "position_embedding_type": "absolute",
31
+ "transformers_version": "4.5.0.dev0",
32
+ "type_vocab_size": 2,
33
+ "use_cache": true,
34
+ "vocab_size": 30522
35
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:257a98e2c089768f0c4a7b6eaa5b138baa7e9777e301c09b08e965102a07c540
3
+ size 1340750921
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "bert-large-uncased"}
trainer_state.json ADDED
@@ -0,0 +1,358 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.8334419817470664,
5
+ "global_step": 22500,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.04,
12
+ "learning_rate": 1.9728378965667103e-05,
13
+ "loss": 0.755,
14
+ "step": 500
15
+ },
16
+ {
17
+ "epoch": 0.08,
18
+ "learning_rate": 1.9456757931334205e-05,
19
+ "loss": 0.5771,
20
+ "step": 1000
21
+ },
22
+ {
23
+ "epoch": 0.12,
24
+ "learning_rate": 1.9185136897001307e-05,
25
+ "loss": 0.5303,
26
+ "step": 1500
27
+ },
28
+ {
29
+ "epoch": 0.16,
30
+ "learning_rate": 1.8913515862668405e-05,
31
+ "loss": 0.5153,
32
+ "step": 2000
33
+ },
34
+ {
35
+ "epoch": 0.2,
36
+ "learning_rate": 1.8641894828335507e-05,
37
+ "loss": 0.5023,
38
+ "step": 2500
39
+ },
40
+ {
41
+ "epoch": 0.2,
42
+ "eval_accuracy": 0.8247580234335201,
43
+ "eval_loss": 0.4500664174556732,
44
+ "eval_runtime": 38.363,
45
+ "eval_samples_per_second": 255.846,
46
+ "step": 2500
47
+ },
48
+ {
49
+ "epoch": 0.24,
50
+ "learning_rate": 1.837027379400261e-05,
51
+ "loss": 0.4797,
52
+ "step": 3000
53
+ },
54
+ {
55
+ "epoch": 0.29,
56
+ "learning_rate": 1.809865275966971e-05,
57
+ "loss": 0.4613,
58
+ "step": 3500
59
+ },
60
+ {
61
+ "epoch": 0.33,
62
+ "learning_rate": 1.7827031725336812e-05,
63
+ "loss": 0.4665,
64
+ "step": 4000
65
+ },
66
+ {
67
+ "epoch": 0.37,
68
+ "learning_rate": 1.7555410691003914e-05,
69
+ "loss": 0.4516,
70
+ "step": 4500
71
+ },
72
+ {
73
+ "epoch": 0.41,
74
+ "learning_rate": 1.7283789656671015e-05,
75
+ "loss": 0.4485,
76
+ "step": 5000
77
+ },
78
+ {
79
+ "epoch": 0.41,
80
+ "eval_accuracy": 0.8420784513499745,
81
+ "eval_loss": 0.40373361110687256,
82
+ "eval_runtime": 38.331,
83
+ "eval_samples_per_second": 256.059,
84
+ "step": 5000
85
+ },
86
+ {
87
+ "epoch": 0.45,
88
+ "learning_rate": 1.7012168622338114e-05,
89
+ "loss": 0.4394,
90
+ "step": 5500
91
+ },
92
+ {
93
+ "epoch": 0.49,
94
+ "learning_rate": 1.6740547588005215e-05,
95
+ "loss": 0.4262,
96
+ "step": 6000
97
+ },
98
+ {
99
+ "epoch": 0.53,
100
+ "learning_rate": 1.6468926553672317e-05,
101
+ "loss": 0.4374,
102
+ "step": 6500
103
+ },
104
+ {
105
+ "epoch": 0.57,
106
+ "learning_rate": 1.619730551933942e-05,
107
+ "loss": 0.4239,
108
+ "step": 7000
109
+ },
110
+ {
111
+ "epoch": 0.61,
112
+ "learning_rate": 1.592568448500652e-05,
113
+ "loss": 0.4311,
114
+ "step": 7500
115
+ },
116
+ {
117
+ "epoch": 0.61,
118
+ "eval_accuracy": 0.845746306673459,
119
+ "eval_loss": 0.43105846643447876,
120
+ "eval_runtime": 38.3496,
121
+ "eval_samples_per_second": 255.935,
122
+ "step": 7500
123
+ },
124
+ {
125
+ "epoch": 0.65,
126
+ "learning_rate": 1.5654063450673622e-05,
127
+ "loss": 0.41,
128
+ "step": 8000
129
+ },
130
+ {
131
+ "epoch": 0.69,
132
+ "learning_rate": 1.5382442416340724e-05,
133
+ "loss": 0.4248,
134
+ "step": 8500
135
+ },
136
+ {
137
+ "epoch": 0.73,
138
+ "learning_rate": 1.5110821382007822e-05,
139
+ "loss": 0.4041,
140
+ "step": 9000
141
+ },
142
+ {
143
+ "epoch": 0.77,
144
+ "learning_rate": 1.4839200347674924e-05,
145
+ "loss": 0.4047,
146
+ "step": 9500
147
+ },
148
+ {
149
+ "epoch": 0.81,
150
+ "learning_rate": 1.4567579313342026e-05,
151
+ "loss": 0.4062,
152
+ "step": 10000
153
+ },
154
+ {
155
+ "epoch": 0.81,
156
+ "eval_accuracy": 0.8529801324503311,
157
+ "eval_loss": 0.38434121012687683,
158
+ "eval_runtime": 38.3942,
159
+ "eval_samples_per_second": 255.638,
160
+ "step": 10000
161
+ },
162
+ {
163
+ "epoch": 0.86,
164
+ "learning_rate": 1.4295958279009128e-05,
165
+ "loss": 0.4095,
166
+ "step": 10500
167
+ },
168
+ {
169
+ "epoch": 0.9,
170
+ "learning_rate": 1.402433724467623e-05,
171
+ "loss": 0.3963,
172
+ "step": 11000
173
+ },
174
+ {
175
+ "epoch": 0.94,
176
+ "learning_rate": 1.3752716210343331e-05,
177
+ "loss": 0.3997,
178
+ "step": 11500
179
+ },
180
+ {
181
+ "epoch": 0.98,
182
+ "learning_rate": 1.3481095176010431e-05,
183
+ "loss": 0.4054,
184
+ "step": 12000
185
+ },
186
+ {
187
+ "epoch": 1.02,
188
+ "learning_rate": 1.3209474141677531e-05,
189
+ "loss": 0.3341,
190
+ "step": 12500
191
+ },
192
+ {
193
+ "epoch": 1.02,
194
+ "eval_accuracy": 0.8573611818644932,
195
+ "eval_loss": 0.4230915606021881,
196
+ "eval_runtime": 38.3424,
197
+ "eval_samples_per_second": 255.983,
198
+ "step": 12500
199
+ },
200
+ {
201
+ "epoch": 1.06,
202
+ "learning_rate": 1.2937853107344633e-05,
203
+ "loss": 0.2561,
204
+ "step": 13000
205
+ },
206
+ {
207
+ "epoch": 1.1,
208
+ "learning_rate": 1.2666232073011735e-05,
209
+ "loss": 0.2567,
210
+ "step": 13500
211
+ },
212
+ {
213
+ "epoch": 1.14,
214
+ "learning_rate": 1.2394611038678836e-05,
215
+ "loss": 0.2583,
216
+ "step": 14000
217
+ },
218
+ {
219
+ "epoch": 1.18,
220
+ "learning_rate": 1.2122990004345938e-05,
221
+ "loss": 0.2511,
222
+ "step": 14500
223
+ },
224
+ {
225
+ "epoch": 1.22,
226
+ "learning_rate": 1.185136897001304e-05,
227
+ "loss": 0.2584,
228
+ "step": 15000
229
+ },
230
+ {
231
+ "epoch": 1.22,
232
+ "eval_accuracy": 0.8566479877738156,
233
+ "eval_loss": 0.4263673424720764,
234
+ "eval_runtime": 38.3835,
235
+ "eval_samples_per_second": 255.709,
236
+ "step": 15000
237
+ },
238
+ {
239
+ "epoch": 1.26,
240
+ "learning_rate": 1.157974793568014e-05,
241
+ "loss": 0.2588,
242
+ "step": 15500
243
+ },
244
+ {
245
+ "epoch": 1.3,
246
+ "learning_rate": 1.130812690134724e-05,
247
+ "loss": 0.2712,
248
+ "step": 16000
249
+ },
250
+ {
251
+ "epoch": 1.34,
252
+ "learning_rate": 1.1036505867014341e-05,
253
+ "loss": 0.2543,
254
+ "step": 16500
255
+ },
256
+ {
257
+ "epoch": 1.39,
258
+ "learning_rate": 1.0764884832681443e-05,
259
+ "loss": 0.2602,
260
+ "step": 17000
261
+ },
262
+ {
263
+ "epoch": 1.43,
264
+ "learning_rate": 1.0493263798348545e-05,
265
+ "loss": 0.2615,
266
+ "step": 17500
267
+ },
268
+ {
269
+ "epoch": 1.43,
270
+ "eval_accuracy": 0.85206316861946,
271
+ "eval_loss": 0.43541985750198364,
272
+ "eval_runtime": 38.3513,
273
+ "eval_samples_per_second": 255.924,
274
+ "step": 17500
275
+ },
276
+ {
277
+ "epoch": 1.47,
278
+ "learning_rate": 1.0221642764015647e-05,
279
+ "loss": 0.2601,
280
+ "step": 18000
281
+ },
282
+ {
283
+ "epoch": 1.51,
284
+ "learning_rate": 9.950021729682747e-06,
285
+ "loss": 0.2624,
286
+ "step": 18500
287
+ },
288
+ {
289
+ "epoch": 1.55,
290
+ "learning_rate": 9.678400695349848e-06,
291
+ "loss": 0.2611,
292
+ "step": 19000
293
+ },
294
+ {
295
+ "epoch": 1.59,
296
+ "learning_rate": 9.40677966101695e-06,
297
+ "loss": 0.264,
298
+ "step": 19500
299
+ },
300
+ {
301
+ "epoch": 1.63,
302
+ "learning_rate": 9.13515862668405e-06,
303
+ "loss": 0.2589,
304
+ "step": 20000
305
+ },
306
+ {
307
+ "epoch": 1.63,
308
+ "eval_accuracy": 0.8614365766683647,
309
+ "eval_loss": 0.40027427673339844,
310
+ "eval_runtime": 38.3743,
311
+ "eval_samples_per_second": 255.77,
312
+ "step": 20000
313
+ },
314
+ {
315
+ "epoch": 1.67,
316
+ "learning_rate": 8.863537592351152e-06,
317
+ "loss": 0.2596,
318
+ "step": 20500
319
+ },
320
+ {
321
+ "epoch": 1.71,
322
+ "learning_rate": 8.591916558018254e-06,
323
+ "loss": 0.2558,
324
+ "step": 21000
325
+ },
326
+ {
327
+ "epoch": 1.75,
328
+ "learning_rate": 8.320295523685355e-06,
329
+ "loss": 0.2599,
330
+ "step": 21500
331
+ },
332
+ {
333
+ "epoch": 1.79,
334
+ "learning_rate": 8.048674489352455e-06,
335
+ "loss": 0.2562,
336
+ "step": 22000
337
+ },
338
+ {
339
+ "epoch": 1.83,
340
+ "learning_rate": 7.777053455019557e-06,
341
+ "loss": 0.2641,
342
+ "step": 22500
343
+ },
344
+ {
345
+ "epoch": 1.83,
346
+ "eval_accuracy": 0.8667345899133979,
347
+ "eval_loss": 0.3984006643295288,
348
+ "eval_runtime": 38.3342,
349
+ "eval_samples_per_second": 256.038,
350
+ "step": 22500
351
+ }
352
+ ],
353
+ "max_steps": 36816,
354
+ "num_train_epochs": 3,
355
+ "total_flos": 1.8532124395781683e+17,
356
+ "trial_name": null,
357
+ "trial_params": null
358
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5c706b3c2054f5c39cd48f3e866aa84e83602c7e6b71d65bd3548823e8f9b2f
3
+ size 2351
vocab.txt ADDED
The diff for this file is too large to render. See raw diff