Check commited on
Commit
3418e9a
Β·
1 Parent(s): 6fa0334

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/rng_state.pth +2 -2
  6. model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/trainer_state.json +793 -4
  9. model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629778472.5431197/events.out.tfevents.1629778472.c435e1c5ee04.920.41 +3 -0
  11. model-bin/finetune/base/log/1629779133.0815535/events.out.tfevents.1629779133.c435e1c5ee04.920.43 +3 -0
  12. model-bin/finetune/base/log/1629779785.3045485/events.out.tfevents.1629779785.c435e1c5ee04.920.45 +3 -0
  13. model-bin/finetune/base/log/1629780436.4732363/events.out.tfevents.1629780436.c435e1c5ee04.920.47 +3 -0
  14. model-bin/finetune/base/log/1629781090.5327764/events.out.tfevents.1629781090.c435e1c5ee04.920.49 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629778472.c435e1c5ee04.920.40 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629779133.c435e1c5ee04.920.42 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629779785.c435e1c5ee04.920.44 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629780436.c435e1c5ee04.920.46 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629781090.c435e1c5ee04.920.48 +3 -0
model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9588d1cdc17839a81d3b9833fa479a88e833d5e7e4445b67e88528d92457cd28
3
  size 722165009
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30cf6d09d0b50a9f67e11e8c67c49356a46eab3a11783da12312b2ad6c07427d
3
  size 722165009
model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0bdd0211a727cf8c3094eb7dd0798468d83876617b84e685844e1e82c9ef74b5
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25ab05e0b9cc5b5c4f58d78d8b2e9f01de46ec6c559978d368a049852d0476fd
3
  size 377909911
model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:886ff01e12101694ba90eee9a9b7d639459bf4fb95e67d4c35cce050d60d612e
3
- size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc67ac4d86710e96e4aff956d7725e59bf4ef61f39d585f959501871aaef7a35
3
+ size 14567
model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:93631ed3f39e4cb03d22fbc0af7e360abac36a33350849fe88351d54a6afd41c
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46c4f16ce8829f5786060a04c5b5a8cbb455321b5dfbceaa0c763f4777791489
3
  size 559
model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:90699fcee6c06088cbb78a3b5245a2e25cdf72d6b45818c2c1fb1b391e326908
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7db297422c4c558e216638cad9666c70a88ca6d2485e276ec8d633302b42fd19
3
  size 623
model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.18631571186315712,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-55501",
4
- "epoch": 453.99598393574297,
5
- "global_step": 56372,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -171087,11 +171087,800 @@
171087
  "eval_steps_per_second": 0.67,
171088
  "eval_wer": 0.20751998847511344,
171089
  "step": 56372
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
171090
  }
171091
  ],
171092
- "max_steps": 620000,
171093
  "num_train_epochs": 5000,
171094
- "total_flos": 1.5864105616332263e+20,
171095
  "trial_name": null,
171096
  "trial_params": null
171097
  }
 
1
  {
2
  "best_metric": 0.18631571186315712,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-55501",
4
+ "epoch": 462.99595141700405,
5
+ "global_step": 56994,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
171087
  "eval_steps_per_second": 0.67,
171088
  "eval_wer": 0.20751998847511344,
171089
  "step": 56372
171090
+ },
171091
+ {
171092
+ "epoch": 454.02,
171093
+ "learning_rate": 9.112916666666667e-06,
171094
+ "loss": 0.4064,
171095
+ "step": 56375
171096
+ },
171097
+ {
171098
+ "epoch": 454.06,
171099
+ "learning_rate": 9.112836538461539e-06,
171100
+ "loss": 0.319,
171101
+ "step": 56380
171102
+ },
171103
+ {
171104
+ "epoch": 454.1,
171105
+ "learning_rate": 9.112756410256412e-06,
171106
+ "loss": 0.3691,
171107
+ "step": 56385
171108
+ },
171109
+ {
171110
+ "epoch": 454.14,
171111
+ "learning_rate": 9.112676282051283e-06,
171112
+ "loss": 0.5006,
171113
+ "step": 56390
171114
+ },
171115
+ {
171116
+ "epoch": 454.18,
171117
+ "learning_rate": 9.112596153846154e-06,
171118
+ "loss": 0.8926,
171119
+ "step": 56395
171120
+ },
171121
+ {
171122
+ "epoch": 454.22,
171123
+ "learning_rate": 9.112516025641026e-06,
171124
+ "loss": 0.9786,
171125
+ "step": 56400
171126
+ },
171127
+ {
171128
+ "epoch": 454.27,
171129
+ "learning_rate": 9.112435897435899e-06,
171130
+ "loss": 0.3246,
171131
+ "step": 56405
171132
+ },
171133
+ {
171134
+ "epoch": 454.31,
171135
+ "learning_rate": 9.11235576923077e-06,
171136
+ "loss": 0.3865,
171137
+ "step": 56410
171138
+ },
171139
+ {
171140
+ "epoch": 454.35,
171141
+ "learning_rate": 9.112275641025641e-06,
171142
+ "loss": 0.3918,
171143
+ "step": 56415
171144
+ },
171145
+ {
171146
+ "epoch": 454.39,
171147
+ "learning_rate": 9.112195512820513e-06,
171148
+ "loss": 0.8994,
171149
+ "step": 56420
171150
+ },
171151
+ {
171152
+ "epoch": 454.43,
171153
+ "learning_rate": 9.112115384615386e-06,
171154
+ "loss": 1.0022,
171155
+ "step": 56425
171156
+ },
171157
+ {
171158
+ "epoch": 454.47,
171159
+ "learning_rate": 9.112035256410257e-06,
171160
+ "loss": 0.3168,
171161
+ "step": 56430
171162
+ },
171163
+ {
171164
+ "epoch": 454.51,
171165
+ "learning_rate": 9.111955128205129e-06,
171166
+ "loss": 0.5756,
171167
+ "step": 56435
171168
+ },
171169
+ {
171170
+ "epoch": 454.55,
171171
+ "learning_rate": 9.111875000000002e-06,
171172
+ "loss": 0.4405,
171173
+ "step": 56440
171174
+ },
171175
+ {
171176
+ "epoch": 454.59,
171177
+ "learning_rate": 9.111794871794873e-06,
171178
+ "loss": 0.8018,
171179
+ "step": 56445
171180
+ },
171181
+ {
171182
+ "epoch": 454.63,
171183
+ "learning_rate": 9.111714743589744e-06,
171184
+ "loss": 0.8747,
171185
+ "step": 56450
171186
+ },
171187
+ {
171188
+ "epoch": 454.67,
171189
+ "learning_rate": 9.111634615384616e-06,
171190
+ "loss": 0.2681,
171191
+ "step": 56455
171192
+ },
171193
+ {
171194
+ "epoch": 454.71,
171195
+ "learning_rate": 9.111554487179489e-06,
171196
+ "loss": 0.3752,
171197
+ "step": 56460
171198
+ },
171199
+ {
171200
+ "epoch": 454.75,
171201
+ "learning_rate": 9.11147435897436e-06,
171202
+ "loss": 0.4424,
171203
+ "step": 56465
171204
+ },
171205
+ {
171206
+ "epoch": 454.79,
171207
+ "learning_rate": 9.111394230769231e-06,
171208
+ "loss": 0.8706,
171209
+ "step": 56470
171210
+ },
171211
+ {
171212
+ "epoch": 454.83,
171213
+ "learning_rate": 9.111314102564103e-06,
171214
+ "loss": 1.0377,
171215
+ "step": 56475
171216
+ },
171217
+ {
171218
+ "epoch": 454.87,
171219
+ "learning_rate": 9.111233974358976e-06,
171220
+ "loss": 0.3031,
171221
+ "step": 56480
171222
+ },
171223
+ {
171224
+ "epoch": 454.91,
171225
+ "learning_rate": 9.111153846153847e-06,
171226
+ "loss": 0.3767,
171227
+ "step": 56485
171228
+ },
171229
+ {
171230
+ "epoch": 454.95,
171231
+ "learning_rate": 9.111073717948719e-06,
171232
+ "loss": 0.5039,
171233
+ "step": 56490
171234
+ },
171235
+ {
171236
+ "epoch": 454.99,
171237
+ "learning_rate": 9.110993589743592e-06,
171238
+ "loss": 0.9137,
171239
+ "step": 56495
171240
+ },
171241
+ {
171242
+ "epoch": 455.0,
171243
+ "eval_loss": 0.38671064376831055,
171244
+ "eval_runtime": 39.4284,
171245
+ "eval_samples_per_second": 21.304,
171246
+ "eval_steps_per_second": 0.685,
171247
+ "eval_wer": 0.1874216618742166,
171248
+ "step": 56496
171249
+ },
171250
+ {
171251
+ "epoch": 451.03,
171252
+ "learning_rate": 9.110913461538461e-06,
171253
+ "loss": 0.3197,
171254
+ "step": 56500
171255
+ },
171256
+ {
171257
+ "epoch": 451.07,
171258
+ "learning_rate": 9.110833333333334e-06,
171259
+ "loss": 0.3031,
171260
+ "step": 56505
171261
+ },
171262
+ {
171263
+ "epoch": 451.11,
171264
+ "learning_rate": 9.110753205128206e-06,
171265
+ "loss": 0.365,
171266
+ "step": 56510
171267
+ },
171268
+ {
171269
+ "epoch": 451.15,
171270
+ "learning_rate": 9.110673076923077e-06,
171271
+ "loss": 0.5421,
171272
+ "step": 56515
171273
+ },
171274
+ {
171275
+ "epoch": 451.19,
171276
+ "learning_rate": 9.110592948717948e-06,
171277
+ "loss": 1.0469,
171278
+ "step": 56520
171279
+ },
171280
+ {
171281
+ "epoch": 451.23,
171282
+ "learning_rate": 9.110512820512821e-06,
171283
+ "loss": 0.8389,
171284
+ "step": 56525
171285
+ },
171286
+ {
171287
+ "epoch": 451.27,
171288
+ "learning_rate": 9.110432692307693e-06,
171289
+ "loss": 0.3759,
171290
+ "step": 56530
171291
+ },
171292
+ {
171293
+ "epoch": 451.31,
171294
+ "learning_rate": 9.110352564102564e-06,
171295
+ "loss": 0.3245,
171296
+ "step": 56535
171297
+ },
171298
+ {
171299
+ "epoch": 451.35,
171300
+ "learning_rate": 9.110272435897437e-06,
171301
+ "loss": 0.4346,
171302
+ "step": 56540
171303
+ },
171304
+ {
171305
+ "epoch": 451.39,
171306
+ "learning_rate": 9.110192307692309e-06,
171307
+ "loss": 0.9923,
171308
+ "step": 56545
171309
+ },
171310
+ {
171311
+ "epoch": 451.43,
171312
+ "learning_rate": 9.11011217948718e-06,
171313
+ "loss": 0.6995,
171314
+ "step": 56550
171315
+ },
171316
+ {
171317
+ "epoch": 451.47,
171318
+ "learning_rate": 9.110032051282051e-06,
171319
+ "loss": 0.2944,
171320
+ "step": 56555
171321
+ },
171322
+ {
171323
+ "epoch": 451.51,
171324
+ "learning_rate": 9.109951923076924e-06,
171325
+ "loss": 0.3371,
171326
+ "step": 56560
171327
+ },
171328
+ {
171329
+ "epoch": 451.55,
171330
+ "learning_rate": 9.109871794871796e-06,
171331
+ "loss": 0.4263,
171332
+ "step": 56565
171333
+ },
171334
+ {
171335
+ "epoch": 451.59,
171336
+ "learning_rate": 9.109791666666667e-06,
171337
+ "loss": 1.1097,
171338
+ "step": 56570
171339
+ },
171340
+ {
171341
+ "epoch": 451.63,
171342
+ "learning_rate": 9.109711538461538e-06,
171343
+ "loss": 0.6258,
171344
+ "step": 56575
171345
+ },
171346
+ {
171347
+ "epoch": 451.67,
171348
+ "learning_rate": 9.109631410256411e-06,
171349
+ "loss": 0.3704,
171350
+ "step": 56580
171351
+ },
171352
+ {
171353
+ "epoch": 451.71,
171354
+ "learning_rate": 9.109551282051283e-06,
171355
+ "loss": 0.2776,
171356
+ "step": 56585
171357
+ },
171358
+ {
171359
+ "epoch": 451.75,
171360
+ "learning_rate": 9.109471153846154e-06,
171361
+ "loss": 0.4297,
171362
+ "step": 56590
171363
+ },
171364
+ {
171365
+ "epoch": 451.79,
171366
+ "learning_rate": 9.109391025641027e-06,
171367
+ "loss": 1.1215,
171368
+ "step": 56595
171369
+ },
171370
+ {
171371
+ "epoch": 451.83,
171372
+ "learning_rate": 9.109310897435899e-06,
171373
+ "loss": 0.707,
171374
+ "step": 56600
171375
+ },
171376
+ {
171377
+ "epoch": 451.87,
171378
+ "learning_rate": 9.10923076923077e-06,
171379
+ "loss": 0.3058,
171380
+ "step": 56605
171381
+ },
171382
+ {
171383
+ "epoch": 451.91,
171384
+ "learning_rate": 9.109150641025641e-06,
171385
+ "loss": 0.3418,
171386
+ "step": 56610
171387
+ },
171388
+ {
171389
+ "epoch": 451.95,
171390
+ "learning_rate": 9.109070512820514e-06,
171391
+ "loss": 0.4603,
171392
+ "step": 56615
171393
+ },
171394
+ {
171395
+ "epoch": 451.99,
171396
+ "learning_rate": 9.108990384615384e-06,
171397
+ "loss": 1.0587,
171398
+ "step": 56620
171399
+ },
171400
+ {
171401
+ "epoch": 452.0,
171402
+ "eval_loss": 0.4021945893764496,
171403
+ "eval_runtime": 39.9783,
171404
+ "eval_samples_per_second": 21.011,
171405
+ "eval_steps_per_second": 0.675,
171406
+ "eval_wer": 0.19840631625118796,
171407
+ "step": 56621
171408
+ },
171409
+ {
171410
+ "epoch": 452.03,
171411
+ "learning_rate": 9.108910256410257e-06,
171412
+ "loss": 0.3565,
171413
+ "step": 56625
171414
+ },
171415
+ {
171416
+ "epoch": 452.07,
171417
+ "learning_rate": 9.10883012820513e-06,
171418
+ "loss": 0.2871,
171419
+ "step": 56630
171420
+ },
171421
+ {
171422
+ "epoch": 452.11,
171423
+ "learning_rate": 9.10875e-06,
171424
+ "loss": 0.3495,
171425
+ "step": 56635
171426
+ },
171427
+ {
171428
+ "epoch": 452.15,
171429
+ "learning_rate": 9.108669871794873e-06,
171430
+ "loss": 0.4982,
171431
+ "step": 56640
171432
+ },
171433
+ {
171434
+ "epoch": 452.19,
171435
+ "learning_rate": 9.108589743589744e-06,
171436
+ "loss": 0.9314,
171437
+ "step": 56645
171438
+ },
171439
+ {
171440
+ "epoch": 452.23,
171441
+ "learning_rate": 9.108509615384616e-06,
171442
+ "loss": 0.8041,
171443
+ "step": 56650
171444
+ },
171445
+ {
171446
+ "epoch": 452.27,
171447
+ "learning_rate": 9.108429487179487e-06,
171448
+ "loss": 0.3507,
171449
+ "step": 56655
171450
+ },
171451
+ {
171452
+ "epoch": 452.31,
171453
+ "learning_rate": 9.10834935897436e-06,
171454
+ "loss": 0.3273,
171455
+ "step": 56660
171456
+ },
171457
+ {
171458
+ "epoch": 452.35,
171459
+ "learning_rate": 9.108269230769231e-06,
171460
+ "loss": 0.4386,
171461
+ "step": 56665
171462
+ },
171463
+ {
171464
+ "epoch": 452.39,
171465
+ "learning_rate": 9.108189102564103e-06,
171466
+ "loss": 0.9502,
171467
+ "step": 56670
171468
+ },
171469
+ {
171470
+ "epoch": 452.43,
171471
+ "learning_rate": 9.108108974358974e-06,
171472
+ "loss": 0.7826,
171473
+ "step": 56675
171474
+ },
171475
+ {
171476
+ "epoch": 452.47,
171477
+ "learning_rate": 9.108028846153847e-06,
171478
+ "loss": 0.2844,
171479
+ "step": 56680
171480
+ },
171481
+ {
171482
+ "epoch": 452.51,
171483
+ "learning_rate": 9.107948717948718e-06,
171484
+ "loss": 0.3316,
171485
+ "step": 56685
171486
+ },
171487
+ {
171488
+ "epoch": 452.55,
171489
+ "learning_rate": 9.10786858974359e-06,
171490
+ "loss": 0.423,
171491
+ "step": 56690
171492
+ },
171493
+ {
171494
+ "epoch": 452.59,
171495
+ "learning_rate": 9.107788461538463e-06,
171496
+ "loss": 0.9898,
171497
+ "step": 56695
171498
+ },
171499
+ {
171500
+ "epoch": 452.63,
171501
+ "learning_rate": 9.107708333333334e-06,
171502
+ "loss": 0.696,
171503
+ "step": 56700
171504
+ },
171505
+ {
171506
+ "epoch": 452.67,
171507
+ "learning_rate": 9.107628205128206e-06,
171508
+ "loss": 0.2807,
171509
+ "step": 56705
171510
+ },
171511
+ {
171512
+ "epoch": 452.71,
171513
+ "learning_rate": 9.107548076923077e-06,
171514
+ "loss": 0.3262,
171515
+ "step": 56710
171516
+ },
171517
+ {
171518
+ "epoch": 452.75,
171519
+ "learning_rate": 9.10746794871795e-06,
171520
+ "loss": 0.4433,
171521
+ "step": 56715
171522
+ },
171523
+ {
171524
+ "epoch": 452.79,
171525
+ "learning_rate": 9.107387820512821e-06,
171526
+ "loss": 0.9766,
171527
+ "step": 56720
171528
+ },
171529
+ {
171530
+ "epoch": 452.83,
171531
+ "learning_rate": 9.107307692307693e-06,
171532
+ "loss": 0.8275,
171533
+ "step": 56725
171534
+ },
171535
+ {
171536
+ "epoch": 452.87,
171537
+ "learning_rate": 9.107227564102566e-06,
171538
+ "loss": 0.3268,
171539
+ "step": 56730
171540
+ },
171541
+ {
171542
+ "epoch": 452.91,
171543
+ "learning_rate": 9.107147435897437e-06,
171544
+ "loss": 0.3838,
171545
+ "step": 56735
171546
+ },
171547
+ {
171548
+ "epoch": 452.95,
171549
+ "learning_rate": 9.107067307692309e-06,
171550
+ "loss": 0.4729,
171551
+ "step": 56740
171552
+ },
171553
+ {
171554
+ "epoch": 452.99,
171555
+ "learning_rate": 9.10698717948718e-06,
171556
+ "loss": 0.9755,
171557
+ "step": 56745
171558
+ },
171559
+ {
171560
+ "epoch": 453.0,
171561
+ "eval_loss": 0.4409070611000061,
171562
+ "eval_runtime": 39.436,
171563
+ "eval_samples_per_second": 21.3,
171564
+ "eval_steps_per_second": 0.685,
171565
+ "eval_wer": 0.20334678021813835,
171566
+ "step": 56746
171567
+ },
171568
+ {
171569
+ "epoch": 453.03,
171570
+ "learning_rate": 9.106907051282053e-06,
171571
+ "loss": 0.3777,
171572
+ "step": 56750
171573
+ },
171574
+ {
171575
+ "epoch": 453.07,
171576
+ "learning_rate": 9.106826923076924e-06,
171577
+ "loss": 0.3611,
171578
+ "step": 56755
171579
+ },
171580
+ {
171581
+ "epoch": 453.11,
171582
+ "learning_rate": 9.106746794871796e-06,
171583
+ "loss": 0.36,
171584
+ "step": 56760
171585
+ },
171586
+ {
171587
+ "epoch": 453.15,
171588
+ "learning_rate": 9.106666666666667e-06,
171589
+ "loss": 0.4627,
171590
+ "step": 56765
171591
+ },
171592
+ {
171593
+ "epoch": 453.19,
171594
+ "learning_rate": 9.10658653846154e-06,
171595
+ "loss": 0.8728,
171596
+ "step": 56770
171597
+ },
171598
+ {
171599
+ "epoch": 453.23,
171600
+ "learning_rate": 9.10650641025641e-06,
171601
+ "loss": 0.6227,
171602
+ "step": 56775
171603
+ },
171604
+ {
171605
+ "epoch": 453.27,
171606
+ "learning_rate": 9.106426282051283e-06,
171607
+ "loss": 0.3291,
171608
+ "step": 56780
171609
+ },
171610
+ {
171611
+ "epoch": 453.31,
171612
+ "learning_rate": 9.106346153846156e-06,
171613
+ "loss": 0.2861,
171614
+ "step": 56785
171615
+ },
171616
+ {
171617
+ "epoch": 453.35,
171618
+ "learning_rate": 9.106266025641025e-06,
171619
+ "loss": 0.4064,
171620
+ "step": 56790
171621
+ },
171622
+ {
171623
+ "epoch": 453.39,
171624
+ "learning_rate": 9.106185897435899e-06,
171625
+ "loss": 0.9348,
171626
+ "step": 56795
171627
+ },
171628
+ {
171629
+ "epoch": 453.43,
171630
+ "learning_rate": 9.10610576923077e-06,
171631
+ "loss": 0.6904,
171632
+ "step": 56800
171633
+ },
171634
+ {
171635
+ "epoch": 453.47,
171636
+ "learning_rate": 9.106025641025641e-06,
171637
+ "loss": 0.3354,
171638
+ "step": 56805
171639
+ },
171640
+ {
171641
+ "epoch": 453.51,
171642
+ "learning_rate": 9.105945512820513e-06,
171643
+ "loss": 0.3073,
171644
+ "step": 56810
171645
+ },
171646
+ {
171647
+ "epoch": 453.55,
171648
+ "learning_rate": 9.105865384615386e-06,
171649
+ "loss": 0.3934,
171650
+ "step": 56815
171651
+ },
171652
+ {
171653
+ "epoch": 453.59,
171654
+ "learning_rate": 9.105785256410257e-06,
171655
+ "loss": 1.0902,
171656
+ "step": 56820
171657
+ },
171658
+ {
171659
+ "epoch": 453.63,
171660
+ "learning_rate": 9.105705128205128e-06,
171661
+ "loss": 0.7669,
171662
+ "step": 56825
171663
+ },
171664
+ {
171665
+ "epoch": 453.67,
171666
+ "learning_rate": 9.105625000000001e-06,
171667
+ "loss": 0.3124,
171668
+ "step": 56830
171669
+ },
171670
+ {
171671
+ "epoch": 453.71,
171672
+ "learning_rate": 9.105544871794873e-06,
171673
+ "loss": 0.3776,
171674
+ "step": 56835
171675
+ },
171676
+ {
171677
+ "epoch": 453.75,
171678
+ "learning_rate": 9.105464743589744e-06,
171679
+ "loss": 0.4458,
171680
+ "step": 56840
171681
+ },
171682
+ {
171683
+ "epoch": 453.79,
171684
+ "learning_rate": 9.105384615384616e-06,
171685
+ "loss": 0.8865,
171686
+ "step": 56845
171687
+ },
171688
+ {
171689
+ "epoch": 453.83,
171690
+ "learning_rate": 9.105304487179489e-06,
171691
+ "loss": 0.8748,
171692
+ "step": 56850
171693
+ },
171694
+ {
171695
+ "epoch": 453.87,
171696
+ "learning_rate": 9.10522435897436e-06,
171697
+ "loss": 0.5539,
171698
+ "step": 56855
171699
+ },
171700
+ {
171701
+ "epoch": 453.91,
171702
+ "learning_rate": 9.105144230769231e-06,
171703
+ "loss": 0.3233,
171704
+ "step": 56860
171705
+ },
171706
+ {
171707
+ "epoch": 453.95,
171708
+ "learning_rate": 9.105064102564103e-06,
171709
+ "loss": 0.4554,
171710
+ "step": 56865
171711
+ },
171712
+ {
171713
+ "epoch": 453.99,
171714
+ "learning_rate": 9.104983974358976e-06,
171715
+ "loss": 0.9328,
171716
+ "step": 56870
171717
+ },
171718
+ {
171719
+ "epoch": 454.0,
171720
+ "eval_loss": 0.44104063510894775,
171721
+ "eval_runtime": 40.1551,
171722
+ "eval_samples_per_second": 20.894,
171723
+ "eval_steps_per_second": 0.672,
171724
+ "eval_wer": 0.2034719710669078,
171725
+ "step": 56871
171726
+ },
171727
+ {
171728
+ "epoch": 462.03,
171729
+ "learning_rate": 9.104903846153847e-06,
171730
+ "loss": 0.4305,
171731
+ "step": 56875
171732
+ },
171733
+ {
171734
+ "epoch": 462.07,
171735
+ "learning_rate": 9.104823717948718e-06,
171736
+ "loss": 0.3532,
171737
+ "step": 56880
171738
+ },
171739
+ {
171740
+ "epoch": 462.11,
171741
+ "learning_rate": 9.104743589743591e-06,
171742
+ "loss": 0.3565,
171743
+ "step": 56885
171744
+ },
171745
+ {
171746
+ "epoch": 462.15,
171747
+ "learning_rate": 9.104663461538463e-06,
171748
+ "loss": 0.5439,
171749
+ "step": 56890
171750
+ },
171751
+ {
171752
+ "epoch": 462.19,
171753
+ "learning_rate": 9.104583333333334e-06,
171754
+ "loss": 0.975,
171755
+ "step": 56895
171756
+ },
171757
+ {
171758
+ "epoch": 462.23,
171759
+ "learning_rate": 9.104503205128206e-06,
171760
+ "loss": 0.6971,
171761
+ "step": 56900
171762
+ },
171763
+ {
171764
+ "epoch": 462.28,
171765
+ "learning_rate": 9.104423076923079e-06,
171766
+ "loss": 0.287,
171767
+ "step": 56905
171768
+ },
171769
+ {
171770
+ "epoch": 462.32,
171771
+ "learning_rate": 9.104342948717948e-06,
171772
+ "loss": 0.3621,
171773
+ "step": 56910
171774
+ },
171775
+ {
171776
+ "epoch": 462.36,
171777
+ "learning_rate": 9.104262820512821e-06,
171778
+ "loss": 0.4668,
171779
+ "step": 56915
171780
+ },
171781
+ {
171782
+ "epoch": 462.4,
171783
+ "learning_rate": 9.104182692307693e-06,
171784
+ "loss": 0.9692,
171785
+ "step": 56920
171786
+ },
171787
+ {
171788
+ "epoch": 462.44,
171789
+ "learning_rate": 9.104102564102566e-06,
171790
+ "loss": 0.6832,
171791
+ "step": 56925
171792
+ },
171793
+ {
171794
+ "epoch": 462.48,
171795
+ "learning_rate": 9.104022435897437e-06,
171796
+ "loss": 0.3668,
171797
+ "step": 56930
171798
+ },
171799
+ {
171800
+ "epoch": 462.52,
171801
+ "learning_rate": 9.103942307692308e-06,
171802
+ "loss": 0.3313,
171803
+ "step": 56935
171804
+ },
171805
+ {
171806
+ "epoch": 462.56,
171807
+ "learning_rate": 9.103862179487181e-06,
171808
+ "loss": 0.4229,
171809
+ "step": 56940
171810
+ },
171811
+ {
171812
+ "epoch": 462.6,
171813
+ "learning_rate": 9.103782051282051e-06,
171814
+ "loss": 0.9475,
171815
+ "step": 56945
171816
+ },
171817
+ {
171818
+ "epoch": 462.64,
171819
+ "learning_rate": 9.103701923076924e-06,
171820
+ "loss": 0.7394,
171821
+ "step": 56950
171822
+ },
171823
+ {
171824
+ "epoch": 462.68,
171825
+ "learning_rate": 9.103621794871796e-06,
171826
+ "loss": 0.306,
171827
+ "step": 56955
171828
+ },
171829
+ {
171830
+ "epoch": 462.72,
171831
+ "learning_rate": 9.103541666666667e-06,
171832
+ "loss": 0.2872,
171833
+ "step": 56960
171834
+ },
171835
+ {
171836
+ "epoch": 462.76,
171837
+ "learning_rate": 9.103461538461538e-06,
171838
+ "loss": 0.4024,
171839
+ "step": 56965
171840
+ },
171841
+ {
171842
+ "epoch": 462.8,
171843
+ "learning_rate": 9.103381410256411e-06,
171844
+ "loss": 0.9516,
171845
+ "step": 56970
171846
+ },
171847
+ {
171848
+ "epoch": 462.84,
171849
+ "learning_rate": 9.103301282051283e-06,
171850
+ "loss": 0.7913,
171851
+ "step": 56975
171852
+ },
171853
+ {
171854
+ "epoch": 462.88,
171855
+ "learning_rate": 9.103221153846154e-06,
171856
+ "loss": 0.3061,
171857
+ "step": 56980
171858
+ },
171859
+ {
171860
+ "epoch": 462.92,
171861
+ "learning_rate": 9.103141025641027e-06,
171862
+ "loss": 0.4049,
171863
+ "step": 56985
171864
+ },
171865
+ {
171866
+ "epoch": 462.96,
171867
+ "learning_rate": 9.103060897435898e-06,
171868
+ "loss": 0.4995,
171869
+ "step": 56990
171870
+ },
171871
+ {
171872
+ "epoch": 463.0,
171873
+ "eval_loss": 0.38458994030952454,
171874
+ "eval_runtime": 40.5268,
171875
+ "eval_samples_per_second": 20.702,
171876
+ "eval_steps_per_second": 0.666,
171877
+ "eval_wer": 0.1988261188554659,
171878
+ "step": 56994
171879
  }
171880
  ],
171881
+ "max_steps": 615000,
171882
  "num_train_epochs": 5000,
171883
+ "total_flos": 1.6039691475592757e+20,
171884
  "trial_name": null,
171885
  "trial_params": null
171886
  }
model-bin/finetune/base/{checkpoint-56372 β†’ checkpoint-56994}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629778472.5431197/events.out.tfevents.1629778472.c435e1c5ee04.920.41 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd4455d0a891e6f5a00d31a4462d63094040cdb10d0a46c1db5a0d46ff972af5
3
+ size 4194
model-bin/finetune/base/log/1629779133.0815535/events.out.tfevents.1629779133.c435e1c5ee04.920.43 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9680d8d37e919800b1254384b37840a50341d557714bd38ba236c6e0fe996ff5
3
+ size 4194
model-bin/finetune/base/log/1629779785.3045485/events.out.tfevents.1629779785.c435e1c5ee04.920.45 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee94580bf491928c6738e5c944d6013c9cec35d37b3a9625920f7229985d6965
3
+ size 4194
model-bin/finetune/base/log/1629780436.4732363/events.out.tfevents.1629780436.c435e1c5ee04.920.47 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37c6d96d0c2be205fe016a35592a3ef051169be88525e373fd581cd556266145
3
+ size 4194
model-bin/finetune/base/log/1629781090.5327764/events.out.tfevents.1629781090.c435e1c5ee04.920.49 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49f019551c5dd52a891e2dc4a6fc21d09861d20be5eb27f09e533c9a9c82802e
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629778472.c435e1c5ee04.920.40 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1751399f588b9effc54dd7e7367165c47a52e73cefa571f4b33aecfd699afb1b
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629779133.c435e1c5ee04.920.42 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c240a440bb6c8ff1993bc3eb5c21851b0c0c31e07baf833620b89c9822e43809
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629779785.c435e1c5ee04.920.44 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd2cd03fefff7ac84ffb1d9abe4c882573c8539fcb787644c74f6cf152de04c3
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629780436.c435e1c5ee04.920.46 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4b19575163b50308202cef4a37d406146eb02794863b51c5696db008e67af57
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629781090.c435e1c5ee04.920.48 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8533a5c3626250d1ea47e29be6f2c15af0ee72559fa9d25af58f76c29fbc2c60
3
+ size 8462