Nikita Moghe Moghe commited on
Commit
572813d
1 Parent(s): 1cb59e7

First model version

Browse files
config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/fs/startiger0/nmoghe/experiments/pbert/base/small/conv/en-it/mbert-tlm-chat-pe/checkpoint-225000",
3
+ "architectures": [
4
+ "BertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "directionality": "bidi",
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "pooler_fc_size": 768,
21
+ "pooler_num_attention_heads": 12,
22
+ "pooler_num_fc_layers": 3,
23
+ "pooler_size_per_head": 128,
24
+ "pooler_type": "first_token_transform",
25
+ "position_embedding_type": "absolute",
26
+ "transformers_version": "4.2.2",
27
+ "type_vocab_size": 2,
28
+ "use_cache": true,
29
+ "vocab_size": 105879
30
+ }
eval_results_mlm.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ perplexity = 2.974910322494858
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d7d49afd0a18d280c3d985eba90894d7a1ba1c6ed3fd3d5b04fb77773a5359d
3
+ size 669910510
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "<S>", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "name_or_path": "bert-base-multilingual-uncased"}
train_results.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ epoch = 20.0
2
+ train_runtime = 70464.5842
3
+ train_samples_per_second = 3.548
trainer_state.json ADDED
@@ -0,0 +1,762 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 1.0504688024520874,
3
+ "best_model_checkpoint": "/fs/startiger0/nmoghe/experiments/pbert/base/small/conv/en-it/mbert-tlm-chat-pe/checkpoint-225000",
4
+ "epoch": 20.0,
5
+ "global_step": 250000,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.2,
12
+ "learning_rate": 4.9500000000000004e-05,
13
+ "loss": 1.8222,
14
+ "step": 2500
15
+ },
16
+ {
17
+ "epoch": 0.4,
18
+ "learning_rate": 4.9e-05,
19
+ "loss": 1.6035,
20
+ "step": 5000
21
+ },
22
+ {
23
+ "epoch": 0.6,
24
+ "learning_rate": 4.85e-05,
25
+ "loss": 1.4988,
26
+ "step": 7500
27
+ },
28
+ {
29
+ "epoch": 0.8,
30
+ "learning_rate": 4.8e-05,
31
+ "loss": 1.4465,
32
+ "step": 10000
33
+ },
34
+ {
35
+ "epoch": 1.0,
36
+ "learning_rate": 4.75e-05,
37
+ "loss": 1.3893,
38
+ "step": 12500
39
+ },
40
+ {
41
+ "epoch": 1.0,
42
+ "eval_loss": 1.4548144340515137,
43
+ "eval_runtime": 9.5125,
44
+ "eval_samples_per_second": 105.125,
45
+ "step": 12500
46
+ },
47
+ {
48
+ "epoch": 1.2,
49
+ "learning_rate": 4.7e-05,
50
+ "loss": 1.3425,
51
+ "step": 15000
52
+ },
53
+ {
54
+ "epoch": 1.4,
55
+ "learning_rate": 4.6500000000000005e-05,
56
+ "loss": 1.3139,
57
+ "step": 17500
58
+ },
59
+ {
60
+ "epoch": 1.6,
61
+ "learning_rate": 4.600000000000001e-05,
62
+ "loss": 1.2891,
63
+ "step": 20000
64
+ },
65
+ {
66
+ "epoch": 1.8,
67
+ "learning_rate": 4.55e-05,
68
+ "loss": 1.2596,
69
+ "step": 22500
70
+ },
71
+ {
72
+ "epoch": 2.0,
73
+ "learning_rate": 4.5e-05,
74
+ "loss": 1.2512,
75
+ "step": 25000
76
+ },
77
+ {
78
+ "epoch": 2.0,
79
+ "eval_loss": 1.3658138513565063,
80
+ "eval_runtime": 9.3984,
81
+ "eval_samples_per_second": 106.401,
82
+ "step": 25000
83
+ },
84
+ {
85
+ "epoch": 2.2,
86
+ "learning_rate": 4.4500000000000004e-05,
87
+ "loss": 1.2094,
88
+ "step": 27500
89
+ },
90
+ {
91
+ "epoch": 2.4,
92
+ "learning_rate": 4.4000000000000006e-05,
93
+ "loss": 1.187,
94
+ "step": 30000
95
+ },
96
+ {
97
+ "epoch": 2.6,
98
+ "learning_rate": 4.35e-05,
99
+ "loss": 1.1791,
100
+ "step": 32500
101
+ },
102
+ {
103
+ "epoch": 2.8,
104
+ "learning_rate": 4.3e-05,
105
+ "loss": 1.1635,
106
+ "step": 35000
107
+ },
108
+ {
109
+ "epoch": 3.0,
110
+ "learning_rate": 4.25e-05,
111
+ "loss": 1.1478,
112
+ "step": 37500
113
+ },
114
+ {
115
+ "epoch": 3.0,
116
+ "eval_loss": 1.2877916097640991,
117
+ "eval_runtime": 9.4132,
118
+ "eval_samples_per_second": 106.233,
119
+ "step": 37500
120
+ },
121
+ {
122
+ "epoch": 3.2,
123
+ "learning_rate": 4.2e-05,
124
+ "loss": 1.1201,
125
+ "step": 40000
126
+ },
127
+ {
128
+ "epoch": 3.4,
129
+ "learning_rate": 4.15e-05,
130
+ "loss": 1.1054,
131
+ "step": 42500
132
+ },
133
+ {
134
+ "epoch": 3.6,
135
+ "learning_rate": 4.1e-05,
136
+ "loss": 1.1036,
137
+ "step": 45000
138
+ },
139
+ {
140
+ "epoch": 3.8,
141
+ "learning_rate": 4.05e-05,
142
+ "loss": 1.0881,
143
+ "step": 47500
144
+ },
145
+ {
146
+ "epoch": 4.0,
147
+ "learning_rate": 4e-05,
148
+ "loss": 1.0839,
149
+ "step": 50000
150
+ },
151
+ {
152
+ "epoch": 4.0,
153
+ "eval_loss": 1.2922987937927246,
154
+ "eval_runtime": 9.404,
155
+ "eval_samples_per_second": 106.338,
156
+ "step": 50000
157
+ },
158
+ {
159
+ "epoch": 4.2,
160
+ "learning_rate": 3.9500000000000005e-05,
161
+ "loss": 1.0512,
162
+ "step": 52500
163
+ },
164
+ {
165
+ "epoch": 4.4,
166
+ "learning_rate": 3.9000000000000006e-05,
167
+ "loss": 1.0497,
168
+ "step": 55000
169
+ },
170
+ {
171
+ "epoch": 4.6,
172
+ "learning_rate": 3.85e-05,
173
+ "loss": 1.045,
174
+ "step": 57500
175
+ },
176
+ {
177
+ "epoch": 4.8,
178
+ "learning_rate": 3.8e-05,
179
+ "loss": 1.0245,
180
+ "step": 60000
181
+ },
182
+ {
183
+ "epoch": 5.0,
184
+ "learning_rate": 3.7500000000000003e-05,
185
+ "loss": 1.0315,
186
+ "step": 62500
187
+ },
188
+ {
189
+ "epoch": 5.0,
190
+ "eval_loss": 1.2622771263122559,
191
+ "eval_runtime": 9.3953,
192
+ "eval_samples_per_second": 106.436,
193
+ "step": 62500
194
+ },
195
+ {
196
+ "epoch": 5.2,
197
+ "learning_rate": 3.7e-05,
198
+ "loss": 0.994,
199
+ "step": 65000
200
+ },
201
+ {
202
+ "epoch": 5.4,
203
+ "learning_rate": 3.65e-05,
204
+ "loss": 0.9962,
205
+ "step": 67500
206
+ },
207
+ {
208
+ "epoch": 5.6,
209
+ "learning_rate": 3.6e-05,
210
+ "loss": 0.9982,
211
+ "step": 70000
212
+ },
213
+ {
214
+ "epoch": 5.8,
215
+ "learning_rate": 3.55e-05,
216
+ "loss": 0.9833,
217
+ "step": 72500
218
+ },
219
+ {
220
+ "epoch": 6.0,
221
+ "learning_rate": 3.5e-05,
222
+ "loss": 0.9806,
223
+ "step": 75000
224
+ },
225
+ {
226
+ "epoch": 6.0,
227
+ "eval_loss": 1.2437635660171509,
228
+ "eval_runtime": 9.3976,
229
+ "eval_samples_per_second": 106.41,
230
+ "step": 75000
231
+ },
232
+ {
233
+ "epoch": 6.2,
234
+ "learning_rate": 3.45e-05,
235
+ "loss": 0.9458,
236
+ "step": 77500
237
+ },
238
+ {
239
+ "epoch": 6.4,
240
+ "learning_rate": 3.4000000000000007e-05,
241
+ "loss": 0.9548,
242
+ "step": 80000
243
+ },
244
+ {
245
+ "epoch": 6.6,
246
+ "learning_rate": 3.35e-05,
247
+ "loss": 0.9523,
248
+ "step": 82500
249
+ },
250
+ {
251
+ "epoch": 6.8,
252
+ "learning_rate": 3.3e-05,
253
+ "loss": 0.9439,
254
+ "step": 85000
255
+ },
256
+ {
257
+ "epoch": 7.0,
258
+ "learning_rate": 3.2500000000000004e-05,
259
+ "loss": 0.9384,
260
+ "step": 87500
261
+ },
262
+ {
263
+ "epoch": 7.0,
264
+ "eval_loss": 1.1912026405334473,
265
+ "eval_runtime": 9.4045,
266
+ "eval_samples_per_second": 106.332,
267
+ "step": 87500
268
+ },
269
+ {
270
+ "epoch": 7.2,
271
+ "learning_rate": 3.2000000000000005e-05,
272
+ "loss": 0.9196,
273
+ "step": 90000
274
+ },
275
+ {
276
+ "epoch": 7.4,
277
+ "learning_rate": 3.15e-05,
278
+ "loss": 0.9098,
279
+ "step": 92500
280
+ },
281
+ {
282
+ "epoch": 7.6,
283
+ "learning_rate": 3.1e-05,
284
+ "loss": 0.9152,
285
+ "step": 95000
286
+ },
287
+ {
288
+ "epoch": 7.8,
289
+ "learning_rate": 3.05e-05,
290
+ "loss": 0.9073,
291
+ "step": 97500
292
+ },
293
+ {
294
+ "epoch": 8.0,
295
+ "learning_rate": 3e-05,
296
+ "loss": 0.8991,
297
+ "step": 100000
298
+ },
299
+ {
300
+ "epoch": 8.0,
301
+ "eval_loss": 1.1714017391204834,
302
+ "eval_runtime": 9.4059,
303
+ "eval_samples_per_second": 106.316,
304
+ "step": 100000
305
+ },
306
+ {
307
+ "epoch": 8.2,
308
+ "learning_rate": 2.95e-05,
309
+ "loss": 0.8676,
310
+ "step": 102500
311
+ },
312
+ {
313
+ "epoch": 8.4,
314
+ "learning_rate": 2.9e-05,
315
+ "loss": 0.8792,
316
+ "step": 105000
317
+ },
318
+ {
319
+ "epoch": 8.6,
320
+ "learning_rate": 2.8499999999999998e-05,
321
+ "loss": 0.872,
322
+ "step": 107500
323
+ },
324
+ {
325
+ "epoch": 8.8,
326
+ "learning_rate": 2.8000000000000003e-05,
327
+ "loss": 0.8728,
328
+ "step": 110000
329
+ },
330
+ {
331
+ "epoch": 9.0,
332
+ "learning_rate": 2.7500000000000004e-05,
333
+ "loss": 0.8617,
334
+ "step": 112500
335
+ },
336
+ {
337
+ "epoch": 9.0,
338
+ "eval_loss": 1.1749364137649536,
339
+ "eval_runtime": 9.4187,
340
+ "eval_samples_per_second": 106.171,
341
+ "step": 112500
342
+ },
343
+ {
344
+ "epoch": 9.2,
345
+ "learning_rate": 2.7000000000000002e-05,
346
+ "loss": 0.8444,
347
+ "step": 115000
348
+ },
349
+ {
350
+ "epoch": 9.4,
351
+ "learning_rate": 2.6500000000000004e-05,
352
+ "loss": 0.8463,
353
+ "step": 117500
354
+ },
355
+ {
356
+ "epoch": 9.6,
357
+ "learning_rate": 2.6000000000000002e-05,
358
+ "loss": 0.8334,
359
+ "step": 120000
360
+ },
361
+ {
362
+ "epoch": 9.8,
363
+ "learning_rate": 2.5500000000000003e-05,
364
+ "loss": 0.8304,
365
+ "step": 122500
366
+ },
367
+ {
368
+ "epoch": 10.0,
369
+ "learning_rate": 2.5e-05,
370
+ "loss": 0.8291,
371
+ "step": 125000
372
+ },
373
+ {
374
+ "epoch": 10.0,
375
+ "eval_loss": 1.1757138967514038,
376
+ "eval_runtime": 9.3955,
377
+ "eval_samples_per_second": 106.433,
378
+ "step": 125000
379
+ },
380
+ {
381
+ "epoch": 10.2,
382
+ "learning_rate": 2.45e-05,
383
+ "loss": 0.8166,
384
+ "step": 127500
385
+ },
386
+ {
387
+ "epoch": 10.4,
388
+ "learning_rate": 2.4e-05,
389
+ "loss": 0.8126,
390
+ "step": 130000
391
+ },
392
+ {
393
+ "epoch": 10.6,
394
+ "learning_rate": 2.35e-05,
395
+ "loss": 0.8102,
396
+ "step": 132500
397
+ },
398
+ {
399
+ "epoch": 10.8,
400
+ "learning_rate": 2.3000000000000003e-05,
401
+ "loss": 0.8029,
402
+ "step": 135000
403
+ },
404
+ {
405
+ "epoch": 11.0,
406
+ "learning_rate": 2.25e-05,
407
+ "loss": 0.8031,
408
+ "step": 137500
409
+ },
410
+ {
411
+ "epoch": 11.0,
412
+ "eval_loss": 1.1473666429519653,
413
+ "eval_runtime": 9.3894,
414
+ "eval_samples_per_second": 106.503,
415
+ "step": 137500
416
+ },
417
+ {
418
+ "epoch": 11.2,
419
+ "learning_rate": 2.2000000000000003e-05,
420
+ "loss": 0.7793,
421
+ "step": 140000
422
+ },
423
+ {
424
+ "epoch": 11.4,
425
+ "learning_rate": 2.15e-05,
426
+ "loss": 0.7842,
427
+ "step": 142500
428
+ },
429
+ {
430
+ "epoch": 11.6,
431
+ "learning_rate": 2.1e-05,
432
+ "loss": 0.7734,
433
+ "step": 145000
434
+ },
435
+ {
436
+ "epoch": 11.8,
437
+ "learning_rate": 2.05e-05,
438
+ "loss": 0.7817,
439
+ "step": 147500
440
+ },
441
+ {
442
+ "epoch": 12.0,
443
+ "learning_rate": 2e-05,
444
+ "loss": 0.7701,
445
+ "step": 150000
446
+ },
447
+ {
448
+ "epoch": 12.0,
449
+ "eval_loss": 1.1467667818069458,
450
+ "eval_runtime": 9.4066,
451
+ "eval_samples_per_second": 106.308,
452
+ "step": 150000
453
+ },
454
+ {
455
+ "epoch": 12.2,
456
+ "learning_rate": 1.9500000000000003e-05,
457
+ "loss": 0.7573,
458
+ "step": 152500
459
+ },
460
+ {
461
+ "epoch": 12.4,
462
+ "learning_rate": 1.9e-05,
463
+ "loss": 0.7483,
464
+ "step": 155000
465
+ },
466
+ {
467
+ "epoch": 12.6,
468
+ "learning_rate": 1.85e-05,
469
+ "loss": 0.753,
470
+ "step": 157500
471
+ },
472
+ {
473
+ "epoch": 12.8,
474
+ "learning_rate": 1.8e-05,
475
+ "loss": 0.7475,
476
+ "step": 160000
477
+ },
478
+ {
479
+ "epoch": 13.0,
480
+ "learning_rate": 1.75e-05,
481
+ "loss": 0.7395,
482
+ "step": 162500
483
+ },
484
+ {
485
+ "epoch": 13.0,
486
+ "eval_loss": 1.1321425437927246,
487
+ "eval_runtime": 9.3951,
488
+ "eval_samples_per_second": 106.438,
489
+ "step": 162500
490
+ },
491
+ {
492
+ "epoch": 13.2,
493
+ "learning_rate": 1.7000000000000003e-05,
494
+ "loss": 0.7292,
495
+ "step": 165000
496
+ },
497
+ {
498
+ "epoch": 13.4,
499
+ "learning_rate": 1.65e-05,
500
+ "loss": 0.7296,
501
+ "step": 167500
502
+ },
503
+ {
504
+ "epoch": 13.6,
505
+ "learning_rate": 1.6000000000000003e-05,
506
+ "loss": 0.719,
507
+ "step": 170000
508
+ },
509
+ {
510
+ "epoch": 13.8,
511
+ "learning_rate": 1.55e-05,
512
+ "loss": 0.7197,
513
+ "step": 172500
514
+ },
515
+ {
516
+ "epoch": 14.0,
517
+ "learning_rate": 1.5e-05,
518
+ "loss": 0.7122,
519
+ "step": 175000
520
+ },
521
+ {
522
+ "epoch": 14.0,
523
+ "eval_loss": 1.1006277799606323,
524
+ "eval_runtime": 9.3983,
525
+ "eval_samples_per_second": 106.402,
526
+ "step": 175000
527
+ },
528
+ {
529
+ "epoch": 14.2,
530
+ "learning_rate": 1.45e-05,
531
+ "loss": 0.7011,
532
+ "step": 177500
533
+ },
534
+ {
535
+ "epoch": 14.4,
536
+ "learning_rate": 1.4000000000000001e-05,
537
+ "loss": 0.7032,
538
+ "step": 180000
539
+ },
540
+ {
541
+ "epoch": 14.6,
542
+ "learning_rate": 1.3500000000000001e-05,
543
+ "loss": 0.6977,
544
+ "step": 182500
545
+ },
546
+ {
547
+ "epoch": 14.8,
548
+ "learning_rate": 1.3000000000000001e-05,
549
+ "loss": 0.7004,
550
+ "step": 185000
551
+ },
552
+ {
553
+ "epoch": 15.0,
554
+ "learning_rate": 1.25e-05,
555
+ "loss": 0.6871,
556
+ "step": 187500
557
+ },
558
+ {
559
+ "epoch": 15.0,
560
+ "eval_loss": 1.1152338981628418,
561
+ "eval_runtime": 9.3979,
562
+ "eval_samples_per_second": 106.407,
563
+ "step": 187500
564
+ },
565
+ {
566
+ "epoch": 15.2,
567
+ "learning_rate": 1.2e-05,
568
+ "loss": 0.6793,
569
+ "step": 190000
570
+ },
571
+ {
572
+ "epoch": 15.4,
573
+ "learning_rate": 1.1500000000000002e-05,
574
+ "loss": 0.6775,
575
+ "step": 192500
576
+ },
577
+ {
578
+ "epoch": 15.6,
579
+ "learning_rate": 1.1000000000000001e-05,
580
+ "loss": 0.6764,
581
+ "step": 195000
582
+ },
583
+ {
584
+ "epoch": 15.8,
585
+ "learning_rate": 1.05e-05,
586
+ "loss": 0.6731,
587
+ "step": 197500
588
+ },
589
+ {
590
+ "epoch": 16.0,
591
+ "learning_rate": 1e-05,
592
+ "loss": 0.6773,
593
+ "step": 200000
594
+ },
595
+ {
596
+ "epoch": 16.0,
597
+ "eval_loss": 1.076183795928955,
598
+ "eval_runtime": 9.4003,
599
+ "eval_samples_per_second": 106.38,
600
+ "step": 200000
601
+ },
602
+ {
603
+ "epoch": 16.2,
604
+ "learning_rate": 9.5e-06,
605
+ "loss": 0.6625,
606
+ "step": 202500
607
+ },
608
+ {
609
+ "epoch": 16.4,
610
+ "learning_rate": 9e-06,
611
+ "loss": 0.6525,
612
+ "step": 205000
613
+ },
614
+ {
615
+ "epoch": 16.6,
616
+ "learning_rate": 8.500000000000002e-06,
617
+ "loss": 0.6566,
618
+ "step": 207500
619
+ },
620
+ {
621
+ "epoch": 16.8,
622
+ "learning_rate": 8.000000000000001e-06,
623
+ "loss": 0.6557,
624
+ "step": 210000
625
+ },
626
+ {
627
+ "epoch": 17.0,
628
+ "learning_rate": 7.5e-06,
629
+ "loss": 0.653,
630
+ "step": 212500
631
+ },
632
+ {
633
+ "epoch": 17.0,
634
+ "eval_loss": 1.0961084365844727,
635
+ "eval_runtime": 9.3914,
636
+ "eval_samples_per_second": 106.48,
637
+ "step": 212500
638
+ },
639
+ {
640
+ "epoch": 17.2,
641
+ "learning_rate": 7.000000000000001e-06,
642
+ "loss": 0.6431,
643
+ "step": 215000
644
+ },
645
+ {
646
+ "epoch": 17.4,
647
+ "learning_rate": 6.5000000000000004e-06,
648
+ "loss": 0.6416,
649
+ "step": 217500
650
+ },
651
+ {
652
+ "epoch": 17.6,
653
+ "learning_rate": 6e-06,
654
+ "loss": 0.6391,
655
+ "step": 220000
656
+ },
657
+ {
658
+ "epoch": 17.8,
659
+ "learning_rate": 5.500000000000001e-06,
660
+ "loss": 0.6345,
661
+ "step": 222500
662
+ },
663
+ {
664
+ "epoch": 18.0,
665
+ "learning_rate": 5e-06,
666
+ "loss": 0.6365,
667
+ "step": 225000
668
+ },
669
+ {
670
+ "epoch": 18.0,
671
+ "eval_loss": 1.0504688024520874,
672
+ "eval_runtime": 9.4106,
673
+ "eval_samples_per_second": 106.263,
674
+ "step": 225000
675
+ },
676
+ {
677
+ "epoch": 18.2,
678
+ "learning_rate": 4.5e-06,
679
+ "loss": 0.623,
680
+ "step": 227500
681
+ },
682
+ {
683
+ "epoch": 18.4,
684
+ "learning_rate": 4.000000000000001e-06,
685
+ "loss": 0.6205,
686
+ "step": 230000
687
+ },
688
+ {
689
+ "epoch": 18.6,
690
+ "learning_rate": 3.5000000000000004e-06,
691
+ "loss": 0.6231,
692
+ "step": 232500
693
+ },
694
+ {
695
+ "epoch": 18.8,
696
+ "learning_rate": 3e-06,
697
+ "loss": 0.6235,
698
+ "step": 235000
699
+ },
700
+ {
701
+ "epoch": 19.0,
702
+ "learning_rate": 2.5e-06,
703
+ "loss": 0.6104,
704
+ "step": 237500
705
+ },
706
+ {
707
+ "epoch": 19.0,
708
+ "eval_loss": 1.0633997917175293,
709
+ "eval_runtime": 9.3934,
710
+ "eval_samples_per_second": 106.458,
711
+ "step": 237500
712
+ },
713
+ {
714
+ "epoch": 19.2,
715
+ "learning_rate": 2.0000000000000003e-06,
716
+ "loss": 0.6154,
717
+ "step": 240000
718
+ },
719
+ {
720
+ "epoch": 19.4,
721
+ "learning_rate": 1.5e-06,
722
+ "loss": 0.6103,
723
+ "step": 242500
724
+ },
725
+ {
726
+ "epoch": 19.6,
727
+ "learning_rate": 1.0000000000000002e-06,
728
+ "loss": 0.5998,
729
+ "step": 245000
730
+ },
731
+ {
732
+ "epoch": 19.8,
733
+ "learning_rate": 5.000000000000001e-07,
734
+ "loss": 0.6125,
735
+ "step": 247500
736
+ },
737
+ {
738
+ "epoch": 20.0,
739
+ "learning_rate": 0.0,
740
+ "loss": 0.6093,
741
+ "step": 250000
742
+ },
743
+ {
744
+ "epoch": 20.0,
745
+ "eval_loss": 1.0744267702102661,
746
+ "eval_runtime": 9.4035,
747
+ "eval_samples_per_second": 106.344,
748
+ "step": 250000
749
+ },
750
+ {
751
+ "epoch": 20.0,
752
+ "step": 250000,
753
+ "train_runtime": 70464.5842,
754
+ "train_samples_per_second": 3.548
755
+ }
756
+ ],
757
+ "max_steps": 250000,
758
+ "num_train_epochs": 20,
759
+ "total_flos": 305681719606891488,
760
+ "trial_name": null,
761
+ "trial_params": null
762
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a71e183cfb2a0607b1d5c20c2795f97aa7bce72cdd31360831726c8f5910f4a8
3
+ size 1835
vocab.txt ADDED
The diff for this file is too large to render. See raw diff