Check commited on
Commit
d905eab
Β·
1 Parent(s): 612c9e4

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/trainer_state.json +799 -4
  9. model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630214833.9430623/events.out.tfevents.1630214833.cc93b136ebf5.1086.49 +3 -0
  11. model-bin/finetune/base/log/1630215268.1180196/events.out.tfevents.1630215268.cc93b136ebf5.1086.51 +3 -0
  12. model-bin/finetune/base/log/1630215709.9436204/events.out.tfevents.1630215709.cc93b136ebf5.1086.53 +3 -0
  13. model-bin/finetune/base/log/1630216147.602746/events.out.tfevents.1630216147.cc93b136ebf5.1086.55 +3 -0
  14. model-bin/finetune/base/log/1630216585.8185782/events.out.tfevents.1630216585.cc93b136ebf5.1086.57 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630214833.cc93b136ebf5.1086.48 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630215268.cc93b136ebf5.1086.50 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630215709.cc93b136ebf5.1086.52 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630216147.cc93b136ebf5.1086.54 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630216585.cc93b136ebf5.1086.56 +3 -0
model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fabeaa58e79bc630b4651cb5de62d3becf8f3d011d2a72a8769bd12ad8ec2b2d
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13943fb05087ba1925fc85cf6d8ccdcd436e67bce7b73387491c31d71a0d9a4d
3
  size 722165393
model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c8ca5c08add489f83f8fbbe54f7967a3267d0c0a25ffcaf253ac3312f999a598
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60daade2e90914f426e13eda15e69959b6dfbcf964cfa92071ea99fff71da02f
3
  size 377909911
model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3ead055a10b6a20dc53b007fa47951821da9955300fa2f176322d948a4b8ef3b
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e42e08f28ef8b1308862c7a353c1fb5d4b225eeb6cba470f033516a5bfc09a8a
3
  size 14503
model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:435c7949f76b27d1892fd86718a853997f446c51f752754e7e9bab37c34db055
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08811403bda30859be9404fa2f69eb552c512d466566eb146d6719d79f4b49f8
3
  size 559
model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e654bbb2cd6ccd57472ee266ae110834ec44311eb0593d1db337b55a5107ef7
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ef285b8f55ac571b7d259714b492a3bca4226307e36590c40ca749e5758e1e7
3
  size 623
model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1689111747851003,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
4
- "epoch": 1146.0,
5
- "global_step": 142119,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -280182,11 +280182,806 @@
280182
  "eval_steps_per_second": 0.652,
280183
  "eval_wer": 0.18064516129032257,
280184
  "step": 142119
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
280185
  }
280186
  ],
280187
- "max_steps": 620000,
280188
  "num_train_epochs": 5000,
280189
- "total_flos": 3.999266636171066e+20,
280190
  "trial_name": null,
280191
  "trial_params": null
280192
  }
 
1
  {
2
  "best_metric": 0.1689111747851003,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
4
+ "epoch": 1141.0,
5
+ "global_step": 142742,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
280182
  "eval_steps_per_second": 0.652,
280183
  "eval_wer": 0.18064516129032257,
280184
  "step": 142119
280185
+ },
280186
+ {
280187
+ "epoch": 1136.01,
280188
+ "learning_rate": 7.721082390953151e-06,
280189
+ "loss": 0.3707,
280190
+ "step": 142120
280191
+ },
280192
+ {
280193
+ "epoch": 1136.05,
280194
+ "learning_rate": 7.721001615508887e-06,
280195
+ "loss": 0.2748,
280196
+ "step": 142125
280197
+ },
280198
+ {
280199
+ "epoch": 1136.09,
280200
+ "learning_rate": 7.720920840064621e-06,
280201
+ "loss": 0.2521,
280202
+ "step": 142130
280203
+ },
280204
+ {
280205
+ "epoch": 1136.13,
280206
+ "learning_rate": 7.720840064620357e-06,
280207
+ "loss": 0.2975,
280208
+ "step": 142135
280209
+ },
280210
+ {
280211
+ "epoch": 1136.17,
280212
+ "learning_rate": 7.720759289176091e-06,
280213
+ "loss": 0.5674,
280214
+ "step": 142140
280215
+ },
280216
+ {
280217
+ "epoch": 1136.21,
280218
+ "learning_rate": 7.720678513731827e-06,
280219
+ "loss": 1.1433,
280220
+ "step": 142145
280221
+ },
280222
+ {
280223
+ "epoch": 1136.25,
280224
+ "learning_rate": 7.720597738287561e-06,
280225
+ "loss": 0.3099,
280226
+ "step": 142150
280227
+ },
280228
+ {
280229
+ "epoch": 1136.29,
280230
+ "learning_rate": 7.720516962843297e-06,
280231
+ "loss": 0.2866,
280232
+ "step": 142155
280233
+ },
280234
+ {
280235
+ "epoch": 1136.33,
280236
+ "learning_rate": 7.720436187399031e-06,
280237
+ "loss": 0.3084,
280238
+ "step": 142160
280239
+ },
280240
+ {
280241
+ "epoch": 1136.37,
280242
+ "learning_rate": 7.720355411954767e-06,
280243
+ "loss": 0.5635,
280244
+ "step": 142165
280245
+ },
280246
+ {
280247
+ "epoch": 1136.41,
280248
+ "learning_rate": 7.720274636510501e-06,
280249
+ "loss": 1.0665,
280250
+ "step": 142170
280251
+ },
280252
+ {
280253
+ "epoch": 1136.45,
280254
+ "learning_rate": 7.720193861066237e-06,
280255
+ "loss": 0.2342,
280256
+ "step": 142175
280257
+ },
280258
+ {
280259
+ "epoch": 1136.49,
280260
+ "learning_rate": 7.720113085621971e-06,
280261
+ "loss": 0.3193,
280262
+ "step": 142180
280263
+ },
280264
+ {
280265
+ "epoch": 1136.53,
280266
+ "learning_rate": 7.720032310177707e-06,
280267
+ "loss": 0.333,
280268
+ "step": 142185
280269
+ },
280270
+ {
280271
+ "epoch": 1136.57,
280272
+ "learning_rate": 7.719951534733441e-06,
280273
+ "loss": 0.4402,
280274
+ "step": 142190
280275
+ },
280276
+ {
280277
+ "epoch": 1136.61,
280278
+ "learning_rate": 7.719870759289177e-06,
280279
+ "loss": 1.1355,
280280
+ "step": 142195
280281
+ },
280282
+ {
280283
+ "epoch": 1136.65,
280284
+ "learning_rate": 7.719789983844913e-06,
280285
+ "loss": 0.2931,
280286
+ "step": 142200
280287
+ },
280288
+ {
280289
+ "epoch": 1136.69,
280290
+ "learning_rate": 7.719709208400647e-06,
280291
+ "loss": 0.2583,
280292
+ "step": 142205
280293
+ },
280294
+ {
280295
+ "epoch": 1136.73,
280296
+ "learning_rate": 7.719628432956383e-06,
280297
+ "loss": 0.3204,
280298
+ "step": 142210
280299
+ },
280300
+ {
280301
+ "epoch": 1136.77,
280302
+ "learning_rate": 7.719547657512117e-06,
280303
+ "loss": 0.4336,
280304
+ "step": 142215
280305
+ },
280306
+ {
280307
+ "epoch": 1136.81,
280308
+ "learning_rate": 7.719466882067853e-06,
280309
+ "loss": 1.0034,
280310
+ "step": 142220
280311
+ },
280312
+ {
280313
+ "epoch": 1136.85,
280314
+ "learning_rate": 7.719386106623587e-06,
280315
+ "loss": 0.2522,
280316
+ "step": 142225
280317
+ },
280318
+ {
280319
+ "epoch": 1136.89,
280320
+ "learning_rate": 7.719305331179323e-06,
280321
+ "loss": 0.2879,
280322
+ "step": 142230
280323
+ },
280324
+ {
280325
+ "epoch": 1136.93,
280326
+ "learning_rate": 7.719224555735057e-06,
280327
+ "loss": 0.317,
280328
+ "step": 142235
280329
+ },
280330
+ {
280331
+ "epoch": 1136.97,
280332
+ "learning_rate": 7.719143780290793e-06,
280333
+ "loss": 0.5506,
280334
+ "step": 142240
280335
+ },
280336
+ {
280337
+ "epoch": 1137.0,
280338
+ "eval_loss": 0.3600503206253052,
280339
+ "eval_runtime": 41.6622,
280340
+ "eval_samples_per_second": 20.114,
280341
+ "eval_steps_per_second": 0.648,
280342
+ "eval_wer": 0.1694229797067957,
280343
+ "step": 142244
280344
+ },
280345
+ {
280346
+ "epoch": 1147.01,
280347
+ "learning_rate": 7.71907915993538e-06,
280348
+ "loss": 0.325,
280349
+ "step": 142245
280350
+ },
280351
+ {
280352
+ "epoch": 1147.05,
280353
+ "learning_rate": 7.718998384491116e-06,
280354
+ "loss": 0.3236,
280355
+ "step": 142250
280356
+ },
280357
+ {
280358
+ "epoch": 1147.09,
280359
+ "learning_rate": 7.71891760904685e-06,
280360
+ "loss": 0.2434,
280361
+ "step": 142255
280362
+ },
280363
+ {
280364
+ "epoch": 1147.13,
280365
+ "learning_rate": 7.718836833602586e-06,
280366
+ "loss": 0.3158,
280367
+ "step": 142260
280368
+ },
280369
+ {
280370
+ "epoch": 1147.17,
280371
+ "learning_rate": 7.718756058158322e-06,
280372
+ "loss": 0.5659,
280373
+ "step": 142265
280374
+ },
280375
+ {
280376
+ "epoch": 1147.21,
280377
+ "learning_rate": 7.718675282714056e-06,
280378
+ "loss": 0.9828,
280379
+ "step": 142270
280380
+ },
280381
+ {
280382
+ "epoch": 1147.25,
280383
+ "learning_rate": 7.718594507269792e-06,
280384
+ "loss": 0.3256,
280385
+ "step": 142275
280386
+ },
280387
+ {
280388
+ "epoch": 1147.29,
280389
+ "learning_rate": 7.718513731825526e-06,
280390
+ "loss": 0.2851,
280391
+ "step": 142280
280392
+ },
280393
+ {
280394
+ "epoch": 1147.33,
280395
+ "learning_rate": 7.718432956381262e-06,
280396
+ "loss": 0.3086,
280397
+ "step": 142285
280398
+ },
280399
+ {
280400
+ "epoch": 1147.37,
280401
+ "learning_rate": 7.718352180936996e-06,
280402
+ "loss": 0.4722,
280403
+ "step": 142290
280404
+ },
280405
+ {
280406
+ "epoch": 1147.41,
280407
+ "learning_rate": 7.718271405492732e-06,
280408
+ "loss": 1.1545,
280409
+ "step": 142295
280410
+ },
280411
+ {
280412
+ "epoch": 1147.45,
280413
+ "learning_rate": 7.718190630048466e-06,
280414
+ "loss": 0.2882,
280415
+ "step": 142300
280416
+ },
280417
+ {
280418
+ "epoch": 1147.49,
280419
+ "learning_rate": 7.718109854604202e-06,
280420
+ "loss": 0.4264,
280421
+ "step": 142305
280422
+ },
280423
+ {
280424
+ "epoch": 1147.53,
280425
+ "learning_rate": 7.718029079159936e-06,
280426
+ "loss": 0.4184,
280427
+ "step": 142310
280428
+ },
280429
+ {
280430
+ "epoch": 1147.57,
280431
+ "learning_rate": 7.717948303715672e-06,
280432
+ "loss": 0.5507,
280433
+ "step": 142315
280434
+ },
280435
+ {
280436
+ "epoch": 1147.61,
280437
+ "learning_rate": 7.717867528271407e-06,
280438
+ "loss": 0.9842,
280439
+ "step": 142320
280440
+ },
280441
+ {
280442
+ "epoch": 1147.65,
280443
+ "learning_rate": 7.717786752827141e-06,
280444
+ "loss": 0.2279,
280445
+ "step": 142325
280446
+ },
280447
+ {
280448
+ "epoch": 1147.69,
280449
+ "learning_rate": 7.717705977382877e-06,
280450
+ "loss": 0.2926,
280451
+ "step": 142330
280452
+ },
280453
+ {
280454
+ "epoch": 1147.73,
280455
+ "learning_rate": 7.717625201938611e-06,
280456
+ "loss": 0.381,
280457
+ "step": 142335
280458
+ },
280459
+ {
280460
+ "epoch": 1147.77,
280461
+ "learning_rate": 7.717544426494347e-06,
280462
+ "loss": 0.4725,
280463
+ "step": 142340
280464
+ },
280465
+ {
280466
+ "epoch": 1147.81,
280467
+ "learning_rate": 7.717463651050081e-06,
280468
+ "loss": 1.0339,
280469
+ "step": 142345
280470
+ },
280471
+ {
280472
+ "epoch": 1147.85,
280473
+ "learning_rate": 7.717382875605817e-06,
280474
+ "loss": 0.3305,
280475
+ "step": 142350
280476
+ },
280477
+ {
280478
+ "epoch": 1147.89,
280479
+ "learning_rate": 7.717302100161551e-06,
280480
+ "loss": 0.2422,
280481
+ "step": 142355
280482
+ },
280483
+ {
280484
+ "epoch": 1147.93,
280485
+ "learning_rate": 7.717221324717287e-06,
280486
+ "loss": 0.3134,
280487
+ "step": 142360
280488
+ },
280489
+ {
280490
+ "epoch": 1147.97,
280491
+ "learning_rate": 7.717140549273021e-06,
280492
+ "loss": 0.5877,
280493
+ "step": 142365
280494
+ },
280495
+ {
280496
+ "epoch": 1148.0,
280497
+ "eval_loss": 0.3255111873149872,
280498
+ "eval_runtime": 41.3411,
280499
+ "eval_samples_per_second": 20.27,
280500
+ "eval_steps_per_second": 0.653,
280501
+ "eval_wer": 0.18185745140388768,
280502
+ "step": 142368
280503
+ },
280504
+ {
280505
+ "epoch": 1148.02,
280506
+ "learning_rate": 7.717059773828757e-06,
280507
+ "loss": 0.3388,
280508
+ "step": 142370
280509
+ },
280510
+ {
280511
+ "epoch": 1148.06,
280512
+ "learning_rate": 7.716978998384493e-06,
280513
+ "loss": 0.2702,
280514
+ "step": 142375
280515
+ },
280516
+ {
280517
+ "epoch": 1148.1,
280518
+ "learning_rate": 7.716898222940227e-06,
280519
+ "loss": 0.2971,
280520
+ "step": 142380
280521
+ },
280522
+ {
280523
+ "epoch": 1148.14,
280524
+ "learning_rate": 7.716817447495963e-06,
280525
+ "loss": 0.328,
280526
+ "step": 142385
280527
+ },
280528
+ {
280529
+ "epoch": 1148.18,
280530
+ "learning_rate": 7.716736672051697e-06,
280531
+ "loss": 0.6229,
280532
+ "step": 142390
280533
+ },
280534
+ {
280535
+ "epoch": 1148.22,
280536
+ "learning_rate": 7.716655896607433e-06,
280537
+ "loss": 1.0692,
280538
+ "step": 142395
280539
+ },
280540
+ {
280541
+ "epoch": 1148.26,
280542
+ "learning_rate": 7.716575121163167e-06,
280543
+ "loss": 0.3869,
280544
+ "step": 142400
280545
+ },
280546
+ {
280547
+ "epoch": 1148.3,
280548
+ "learning_rate": 7.716494345718903e-06,
280549
+ "loss": 0.2883,
280550
+ "step": 142405
280551
+ },
280552
+ {
280553
+ "epoch": 1148.34,
280554
+ "learning_rate": 7.716413570274637e-06,
280555
+ "loss": 0.3579,
280556
+ "step": 142410
280557
+ },
280558
+ {
280559
+ "epoch": 1148.38,
280560
+ "learning_rate": 7.716332794830373e-06,
280561
+ "loss": 0.643,
280562
+ "step": 142415
280563
+ },
280564
+ {
280565
+ "epoch": 1148.42,
280566
+ "learning_rate": 7.716252019386107e-06,
280567
+ "loss": 1.0174,
280568
+ "step": 142420
280569
+ },
280570
+ {
280571
+ "epoch": 1148.46,
280572
+ "learning_rate": 7.716171243941843e-06,
280573
+ "loss": 0.3438,
280574
+ "step": 142425
280575
+ },
280576
+ {
280577
+ "epoch": 1148.5,
280578
+ "learning_rate": 7.716090468497577e-06,
280579
+ "loss": 0.2701,
280580
+ "step": 142430
280581
+ },
280582
+ {
280583
+ "epoch": 1148.54,
280584
+ "learning_rate": 7.716009693053313e-06,
280585
+ "loss": 0.3322,
280586
+ "step": 142435
280587
+ },
280588
+ {
280589
+ "epoch": 1148.58,
280590
+ "learning_rate": 7.715928917609049e-06,
280591
+ "loss": 0.5148,
280592
+ "step": 142440
280593
+ },
280594
+ {
280595
+ "epoch": 1148.62,
280596
+ "learning_rate": 7.715848142164783e-06,
280597
+ "loss": 0.9305,
280598
+ "step": 142445
280599
+ },
280600
+ {
280601
+ "epoch": 1148.66,
280602
+ "learning_rate": 7.715767366720519e-06,
280603
+ "loss": 0.2918,
280604
+ "step": 142450
280605
+ },
280606
+ {
280607
+ "epoch": 1148.7,
280608
+ "learning_rate": 7.715686591276253e-06,
280609
+ "loss": 0.2845,
280610
+ "step": 142455
280611
+ },
280612
+ {
280613
+ "epoch": 1148.74,
280614
+ "learning_rate": 7.715605815831989e-06,
280615
+ "loss": 0.368,
280616
+ "step": 142460
280617
+ },
280618
+ {
280619
+ "epoch": 1148.78,
280620
+ "learning_rate": 7.715525040387723e-06,
280621
+ "loss": 0.6089,
280622
+ "step": 142465
280623
+ },
280624
+ {
280625
+ "epoch": 1148.82,
280626
+ "learning_rate": 7.715444264943459e-06,
280627
+ "loss": 0.993,
280628
+ "step": 142470
280629
+ },
280630
+ {
280631
+ "epoch": 1148.86,
280632
+ "learning_rate": 7.715363489499193e-06,
280633
+ "loss": 0.2292,
280634
+ "step": 142475
280635
+ },
280636
+ {
280637
+ "epoch": 1148.9,
280638
+ "learning_rate": 7.715282714054929e-06,
280639
+ "loss": 0.315,
280640
+ "step": 142480
280641
+ },
280642
+ {
280643
+ "epoch": 1148.94,
280644
+ "learning_rate": 7.715201938610663e-06,
280645
+ "loss": 0.3183,
280646
+ "step": 142485
280647
+ },
280648
+ {
280649
+ "epoch": 1148.98,
280650
+ "learning_rate": 7.715121163166399e-06,
280651
+ "loss": 0.6854,
280652
+ "step": 142490
280653
+ },
280654
+ {
280655
+ "epoch": 1149.0,
280656
+ "eval_loss": 0.35613495111465454,
280657
+ "eval_runtime": 42.22,
280658
+ "eval_samples_per_second": 19.825,
280659
+ "eval_steps_per_second": 0.64,
280660
+ "eval_wer": 0.17750628817822495,
280661
+ "step": 142492
280662
+ },
280663
+ {
280664
+ "epoch": 1139.02,
280665
+ "learning_rate": 7.715040387722133e-06,
280666
+ "loss": 0.3242,
280667
+ "step": 142495
280668
+ },
280669
+ {
280670
+ "epoch": 1139.06,
280671
+ "learning_rate": 7.714959612277869e-06,
280672
+ "loss": 0.3263,
280673
+ "step": 142500
280674
+ },
280675
+ {
280676
+ "epoch": 1139.1,
280677
+ "learning_rate": 7.714878836833604e-06,
280678
+ "loss": 0.2828,
280679
+ "step": 142505
280680
+ },
280681
+ {
280682
+ "epoch": 1139.14,
280683
+ "learning_rate": 7.714798061389339e-06,
280684
+ "loss": 0.343,
280685
+ "step": 142510
280686
+ },
280687
+ {
280688
+ "epoch": 1139.18,
280689
+ "learning_rate": 7.714717285945074e-06,
280690
+ "loss": 0.859,
280691
+ "step": 142515
280692
+ },
280693
+ {
280694
+ "epoch": 1139.22,
280695
+ "learning_rate": 7.714636510500809e-06,
280696
+ "loss": 0.8421,
280697
+ "step": 142520
280698
+ },
280699
+ {
280700
+ "epoch": 1139.26,
280701
+ "learning_rate": 7.714555735056544e-06,
280702
+ "loss": 0.3022,
280703
+ "step": 142525
280704
+ },
280705
+ {
280706
+ "epoch": 1139.3,
280707
+ "learning_rate": 7.714474959612279e-06,
280708
+ "loss": 0.2578,
280709
+ "step": 142530
280710
+ },
280711
+ {
280712
+ "epoch": 1139.34,
280713
+ "learning_rate": 7.714394184168014e-06,
280714
+ "loss": 0.3935,
280715
+ "step": 142535
280716
+ },
280717
+ {
280718
+ "epoch": 1139.38,
280719
+ "learning_rate": 7.714313408723748e-06,
280720
+ "loss": 0.6255,
280721
+ "step": 142540
280722
+ },
280723
+ {
280724
+ "epoch": 1139.42,
280725
+ "learning_rate": 7.714232633279484e-06,
280726
+ "loss": 0.7272,
280727
+ "step": 142545
280728
+ },
280729
+ {
280730
+ "epoch": 1139.46,
280731
+ "learning_rate": 7.714151857835218e-06,
280732
+ "loss": 0.2663,
280733
+ "step": 142550
280734
+ },
280735
+ {
280736
+ "epoch": 1139.5,
280737
+ "learning_rate": 7.714071082390954e-06,
280738
+ "loss": 0.2698,
280739
+ "step": 142555
280740
+ },
280741
+ {
280742
+ "epoch": 1139.54,
280743
+ "learning_rate": 7.713990306946688e-06,
280744
+ "loss": 0.3076,
280745
+ "step": 142560
280746
+ },
280747
+ {
280748
+ "epoch": 1139.58,
280749
+ "learning_rate": 7.713909531502424e-06,
280750
+ "loss": 0.735,
280751
+ "step": 142565
280752
+ },
280753
+ {
280754
+ "epoch": 1139.62,
280755
+ "learning_rate": 7.71382875605816e-06,
280756
+ "loss": 0.8629,
280757
+ "step": 142570
280758
+ },
280759
+ {
280760
+ "epoch": 1139.66,
280761
+ "learning_rate": 7.713747980613894e-06,
280762
+ "loss": 0.3157,
280763
+ "step": 142575
280764
+ },
280765
+ {
280766
+ "epoch": 1139.7,
280767
+ "learning_rate": 7.71366720516963e-06,
280768
+ "loss": 0.2521,
280769
+ "step": 142580
280770
+ },
280771
+ {
280772
+ "epoch": 1139.74,
280773
+ "learning_rate": 7.713586429725364e-06,
280774
+ "loss": 0.3982,
280775
+ "step": 142585
280776
+ },
280777
+ {
280778
+ "epoch": 1139.78,
280779
+ "learning_rate": 7.7135056542811e-06,
280780
+ "loss": 0.7916,
280781
+ "step": 142590
280782
+ },
280783
+ {
280784
+ "epoch": 1139.82,
280785
+ "learning_rate": 7.713424878836834e-06,
280786
+ "loss": 0.9487,
280787
+ "step": 142595
280788
+ },
280789
+ {
280790
+ "epoch": 1139.86,
280791
+ "learning_rate": 7.71334410339257e-06,
280792
+ "loss": 0.2883,
280793
+ "step": 142600
280794
+ },
280795
+ {
280796
+ "epoch": 1139.9,
280797
+ "learning_rate": 7.713263327948304e-06,
280798
+ "loss": 0.4633,
280799
+ "step": 142605
280800
+ },
280801
+ {
280802
+ "epoch": 1139.94,
280803
+ "learning_rate": 7.71318255250404e-06,
280804
+ "loss": 0.3633,
280805
+ "step": 142610
280806
+ },
280807
+ {
280808
+ "epoch": 1139.98,
280809
+ "learning_rate": 7.713101777059774e-06,
280810
+ "loss": 0.7164,
280811
+ "step": 142615
280812
+ },
280813
+ {
280814
+ "epoch": 1140.0,
280815
+ "eval_loss": 0.40156617760658264,
280816
+ "eval_runtime": 41.5385,
280817
+ "eval_samples_per_second": 20.15,
280818
+ "eval_steps_per_second": 0.65,
280819
+ "eval_wer": 0.1809362827956199,
280820
+ "step": 142617
280821
+ },
280822
+ {
280823
+ "epoch": 1140.02,
280824
+ "learning_rate": 7.71302100161551e-06,
280825
+ "loss": 0.3018,
280826
+ "step": 142620
280827
+ },
280828
+ {
280829
+ "epoch": 1140.06,
280830
+ "learning_rate": 7.712940226171244e-06,
280831
+ "loss": 0.2666,
280832
+ "step": 142625
280833
+ },
280834
+ {
280835
+ "epoch": 1140.1,
280836
+ "learning_rate": 7.71285945072698e-06,
280837
+ "loss": 0.2767,
280838
+ "step": 142630
280839
+ },
280840
+ {
280841
+ "epoch": 1140.14,
280842
+ "learning_rate": 7.712778675282716e-06,
280843
+ "loss": 0.3826,
280844
+ "step": 142635
280845
+ },
280846
+ {
280847
+ "epoch": 1140.18,
280848
+ "learning_rate": 7.71269789983845e-06,
280849
+ "loss": 0.6727,
280850
+ "step": 142640
280851
+ },
280852
+ {
280853
+ "epoch": 1140.22,
280854
+ "learning_rate": 7.712633279483039e-06,
280855
+ "loss": 0.8758,
280856
+ "step": 142645
280857
+ },
280858
+ {
280859
+ "epoch": 1140.26,
280860
+ "learning_rate": 7.712552504038773e-06,
280861
+ "loss": 0.3001,
280862
+ "step": 142650
280863
+ },
280864
+ {
280865
+ "epoch": 1140.3,
280866
+ "learning_rate": 7.712471728594509e-06,
280867
+ "loss": 0.31,
280868
+ "step": 142655
280869
+ },
280870
+ {
280871
+ "epoch": 1140.34,
280872
+ "learning_rate": 7.712390953150243e-06,
280873
+ "loss": 0.3155,
280874
+ "step": 142660
280875
+ },
280876
+ {
280877
+ "epoch": 1140.38,
280878
+ "learning_rate": 7.712310177705979e-06,
280879
+ "loss": 0.746,
280880
+ "step": 142665
280881
+ },
280882
+ {
280883
+ "epoch": 1140.42,
280884
+ "learning_rate": 7.712229402261713e-06,
280885
+ "loss": 0.867,
280886
+ "step": 142670
280887
+ },
280888
+ {
280889
+ "epoch": 1140.46,
280890
+ "learning_rate": 7.712148626817449e-06,
280891
+ "loss": 0.261,
280892
+ "step": 142675
280893
+ },
280894
+ {
280895
+ "epoch": 1140.5,
280896
+ "learning_rate": 7.712067851373183e-06,
280897
+ "loss": 0.309,
280898
+ "step": 142680
280899
+ },
280900
+ {
280901
+ "epoch": 1140.54,
280902
+ "learning_rate": 7.711987075928919e-06,
280903
+ "loss": 0.3376,
280904
+ "step": 142685
280905
+ },
280906
+ {
280907
+ "epoch": 1140.58,
280908
+ "learning_rate": 7.711906300484653e-06,
280909
+ "loss": 0.836,
280910
+ "step": 142690
280911
+ },
280912
+ {
280913
+ "epoch": 1140.62,
280914
+ "learning_rate": 7.711825525040389e-06,
280915
+ "loss": 0.8401,
280916
+ "step": 142695
280917
+ },
280918
+ {
280919
+ "epoch": 1140.66,
280920
+ "learning_rate": 7.711744749596125e-06,
280921
+ "loss": 0.2793,
280922
+ "step": 142700
280923
+ },
280924
+ {
280925
+ "epoch": 1140.7,
280926
+ "learning_rate": 7.711663974151859e-06,
280927
+ "loss": 0.3063,
280928
+ "step": 142705
280929
+ },
280930
+ {
280931
+ "epoch": 1140.74,
280932
+ "learning_rate": 7.711583198707595e-06,
280933
+ "loss": 0.3074,
280934
+ "step": 142710
280935
+ },
280936
+ {
280937
+ "epoch": 1140.78,
280938
+ "learning_rate": 7.711502423263329e-06,
280939
+ "loss": 0.7983,
280940
+ "step": 142715
280941
+ },
280942
+ {
280943
+ "epoch": 1140.82,
280944
+ "learning_rate": 7.711421647819065e-06,
280945
+ "loss": 0.8764,
280946
+ "step": 142720
280947
+ },
280948
+ {
280949
+ "epoch": 1140.86,
280950
+ "learning_rate": 7.711340872374799e-06,
280951
+ "loss": 0.2742,
280952
+ "step": 142725
280953
+ },
280954
+ {
280955
+ "epoch": 1140.9,
280956
+ "learning_rate": 7.711260096930535e-06,
280957
+ "loss": 0.3481,
280958
+ "step": 142730
280959
+ },
280960
+ {
280961
+ "epoch": 1140.94,
280962
+ "learning_rate": 7.711179321486269e-06,
280963
+ "loss": 0.3124,
280964
+ "step": 142735
280965
+ },
280966
+ {
280967
+ "epoch": 1140.98,
280968
+ "learning_rate": 7.711098546042005e-06,
280969
+ "loss": 0.8163,
280970
+ "step": 142740
280971
+ },
280972
+ {
280973
+ "epoch": 1141.0,
280974
+ "eval_loss": 0.37549889087677,
280975
+ "eval_runtime": 41.1914,
280976
+ "eval_samples_per_second": 20.32,
280977
+ "eval_steps_per_second": 0.655,
280978
+ "eval_wer": 0.18999632217727105,
280979
+ "step": 142742
280980
  }
280981
  ],
280982
+ "max_steps": 625000,
280983
  "num_train_epochs": 5000,
280984
+ "total_flos": 4.0167170527047575e+20,
280985
  "trial_name": null,
280986
  "trial_params": null
280987
  }
model-bin/finetune/base/{checkpoint-142119 β†’ checkpoint-142742}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630214833.9430623/events.out.tfevents.1630214833.cc93b136ebf5.1086.49 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f143c1996764d184ea6e0142485429f172d965a5beee99c8d12635385968f4ba
3
+ size 4194
model-bin/finetune/base/log/1630215268.1180196/events.out.tfevents.1630215268.cc93b136ebf5.1086.51 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfbe1ada41cc7d3cc8cba21ccfef201ccd24706f6294b62d6f1db3b299da27e1
3
+ size 4194
model-bin/finetune/base/log/1630215709.9436204/events.out.tfevents.1630215709.cc93b136ebf5.1086.53 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8991206a1452b6cf9c05c982a5f3a8d9c2138da33486b32f4c94a31bc711bcd7
3
+ size 4194
model-bin/finetune/base/log/1630216147.602746/events.out.tfevents.1630216147.cc93b136ebf5.1086.55 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98e3c0e7d6e201e8688302d20444d2a50524d7fa3bcea16ecaf8f14bf5776fdc
3
+ size 4194
model-bin/finetune/base/log/1630216585.8185782/events.out.tfevents.1630216585.cc93b136ebf5.1086.57 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c378405f199baeb6e346553baa7f54b6cb213ebc7c86aaa0482bfa189053278
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630214833.cc93b136ebf5.1086.48 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d86e40343bd585bcbd5d65c3e7326f3a5e490b9bf51dbe05a59f5cc1e3a10ff0
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630215268.cc93b136ebf5.1086.50 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c6af5f14ffe58fc7865a60458f9244806462798491b9b175869cadbbe303b31
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630215709.cc93b136ebf5.1086.52 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:363d8446ccdaf00a932c78a1481e7c27f004b1c679fdf0ea960c2a0a6816d940
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630216147.cc93b136ebf5.1086.54 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f0084b3c04544b2030f6f07fd8c6ef86f4137babca34e875305cb7761985d0b
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630216585.cc93b136ebf5.1086.56 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec5a183ccbf22dba5982b3551c42c7c10552ac78bbc45c8030bd3152945d15f6
3
+ size 8622