Check commited on
Commit
d7c7bb6
Β·
1 Parent(s): c3f613d

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/trainer_state.json +799 -4
  9. model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629946430.8098779/events.out.tfevents.1629946430.8e89bd551565.924.31 +3 -0
  11. model-bin/finetune/base/log/1629946866.7205796/events.out.tfevents.1629946866.8e89bd551565.924.33 +3 -0
  12. model-bin/finetune/base/log/1629947293.5282943/events.out.tfevents.1629947293.8e89bd551565.924.35 +3 -0
  13. model-bin/finetune/base/log/1629947731.4501543/events.out.tfevents.1629947731.8e89bd551565.924.37 +3 -0
  14. model-bin/finetune/base/log/1629948172.226034/events.out.tfevents.1629948172.8e89bd551565.924.39 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629946430.8e89bd551565.924.30 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629946865.8e89bd551565.924.32 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629947293.8e89bd551565.924.34 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629947731.8e89bd551565.924.36 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629948172.8e89bd551565.924.38 +3 -0
model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8572d131ce8d59b8860c47164157d951991c9c0ef7b171cad275b160c84bd30b
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:059f1becf8b39dcee76e6892a2b081bc70757f47a310129d0b5ea65b31bf37ad
3
  size 722165393
model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a46bfb9ac87cdbf1e5989ab2bf749ab9a164a555e55d2c2eec7b5e09680b9616
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fb13c34fb11dd331464d813a241c63c38b52eb802c711b52f064013c38883db
3
  size 377909911
model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b0b05680653568416b05354e34ad62d4f0ac27a86e0a07c5052a36b53751ea2d
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04abce0d0284c6da65004c6da2c9ba91b9646db5efd6f5375a9b927541f4a0fd
3
  size 14503
model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:43cf4a0bc3cccd678f9d1d7aac0abd075111f7eb824da38827242e3d76f31e4d
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a28146e60e7482fd8bf1168040f8ce03267a19d616aee5d2b7fadfd2ef5a1a7
3
  size 559
model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4919a489fdfa86cdcf94ef6754b2bfe5e67807bc39be182ad4fe93c03c95ae2a
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:597f477ce136a0d19546bbeaf1d90265ad546212e55347ca9da84764d96083ef
3
  size 623
model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.17731766728757703,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-88857",
4
- "epoch": 721.0,
5
- "global_step": 90224,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -214155,11 +214155,806 @@
214155
  "eval_steps_per_second": 0.685,
214156
  "eval_wer": 0.19095368633674795,
214157
  "step": 90224
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
214158
  }
214159
  ],
214160
- "max_steps": 625000,
214161
  "num_train_epochs": 5000,
214162
- "total_flos": 2.5390227896927135e+20,
214163
  "trial_name": null,
214164
  "trial_params": null
214165
  }
 
1
  {
2
  "best_metric": 0.17731766728757703,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-88857",
4
+ "epoch": 731.995983935743,
5
+ "global_step": 90847,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
214155
  "eval_steps_per_second": 0.685,
214156
  "eval_wer": 0.19095368633674795,
214157
  "step": 90224
214158
+ },
214159
+ {
214160
+ "epoch": 721.01,
214161
+ "learning_rate": 8.570689102564103e-06,
214162
+ "loss": 0.3959,
214163
+ "step": 90225
214164
+ },
214165
+ {
214166
+ "epoch": 721.05,
214167
+ "learning_rate": 8.570608974358975e-06,
214168
+ "loss": 0.3308,
214169
+ "step": 90230
214170
+ },
214171
+ {
214172
+ "epoch": 721.09,
214173
+ "learning_rate": 8.570528846153848e-06,
214174
+ "loss": 0.314,
214175
+ "step": 90235
214176
+ },
214177
+ {
214178
+ "epoch": 721.13,
214179
+ "learning_rate": 8.570448717948719e-06,
214180
+ "loss": 0.318,
214181
+ "step": 90240
214182
+ },
214183
+ {
214184
+ "epoch": 721.17,
214185
+ "learning_rate": 8.57036858974359e-06,
214186
+ "loss": 0.5316,
214187
+ "step": 90245
214188
+ },
214189
+ {
214190
+ "epoch": 721.21,
214191
+ "learning_rate": 8.570288461538462e-06,
214192
+ "loss": 1.0654,
214193
+ "step": 90250
214194
+ },
214195
+ {
214196
+ "epoch": 721.25,
214197
+ "learning_rate": 8.570208333333335e-06,
214198
+ "loss": 0.3506,
214199
+ "step": 90255
214200
+ },
214201
+ {
214202
+ "epoch": 721.29,
214203
+ "learning_rate": 8.570128205128206e-06,
214204
+ "loss": 0.3143,
214205
+ "step": 90260
214206
+ },
214207
+ {
214208
+ "epoch": 721.33,
214209
+ "learning_rate": 8.570048076923077e-06,
214210
+ "loss": 0.2882,
214211
+ "step": 90265
214212
+ },
214213
+ {
214214
+ "epoch": 721.37,
214215
+ "learning_rate": 8.56996794871795e-06,
214216
+ "loss": 0.5043,
214217
+ "step": 90270
214218
+ },
214219
+ {
214220
+ "epoch": 721.41,
214221
+ "learning_rate": 8.569887820512822e-06,
214222
+ "loss": 1.2249,
214223
+ "step": 90275
214224
+ },
214225
+ {
214226
+ "epoch": 721.45,
214227
+ "learning_rate": 8.569807692307693e-06,
214228
+ "loss": 0.2814,
214229
+ "step": 90280
214230
+ },
214231
+ {
214232
+ "epoch": 721.49,
214233
+ "learning_rate": 8.569727564102565e-06,
214234
+ "loss": 0.3371,
214235
+ "step": 90285
214236
+ },
214237
+ {
214238
+ "epoch": 721.53,
214239
+ "learning_rate": 8.569647435897438e-06,
214240
+ "loss": 0.323,
214241
+ "step": 90290
214242
+ },
214243
+ {
214244
+ "epoch": 721.57,
214245
+ "learning_rate": 8.569567307692307e-06,
214246
+ "loss": 0.5027,
214247
+ "step": 90295
214248
+ },
214249
+ {
214250
+ "epoch": 721.61,
214251
+ "learning_rate": 8.56948717948718e-06,
214252
+ "loss": 0.9879,
214253
+ "step": 90300
214254
+ },
214255
+ {
214256
+ "epoch": 721.65,
214257
+ "learning_rate": 8.569407051282052e-06,
214258
+ "loss": 0.2975,
214259
+ "step": 90305
214260
+ },
214261
+ {
214262
+ "epoch": 721.69,
214263
+ "learning_rate": 8.569326923076923e-06,
214264
+ "loss": 0.2783,
214265
+ "step": 90310
214266
+ },
214267
+ {
214268
+ "epoch": 721.73,
214269
+ "learning_rate": 8.569246794871796e-06,
214270
+ "loss": 0.3157,
214271
+ "step": 90315
214272
+ },
214273
+ {
214274
+ "epoch": 721.77,
214275
+ "learning_rate": 8.569166666666667e-06,
214276
+ "loss": 0.5318,
214277
+ "step": 90320
214278
+ },
214279
+ {
214280
+ "epoch": 721.81,
214281
+ "learning_rate": 8.569086538461539e-06,
214282
+ "loss": 1.0942,
214283
+ "step": 90325
214284
+ },
214285
+ {
214286
+ "epoch": 721.85,
214287
+ "learning_rate": 8.56900641025641e-06,
214288
+ "loss": 0.3259,
214289
+ "step": 90330
214290
+ },
214291
+ {
214292
+ "epoch": 721.89,
214293
+ "learning_rate": 8.568926282051283e-06,
214294
+ "loss": 0.2652,
214295
+ "step": 90335
214296
+ },
214297
+ {
214298
+ "epoch": 721.93,
214299
+ "learning_rate": 8.568846153846155e-06,
214300
+ "loss": 0.3069,
214301
+ "step": 90340
214302
+ },
214303
+ {
214304
+ "epoch": 721.97,
214305
+ "learning_rate": 8.568766025641026e-06,
214306
+ "loss": 0.5629,
214307
+ "step": 90345
214308
+ },
214309
+ {
214310
+ "epoch": 722.0,
214311
+ "eval_loss": 0.39299121499061584,
214312
+ "eval_runtime": 39.1237,
214313
+ "eval_samples_per_second": 21.496,
214314
+ "eval_steps_per_second": 0.69,
214315
+ "eval_wer": 0.18836970851032664,
214316
+ "step": 90349
214317
+ },
214318
+ {
214319
+ "epoch": 722.01,
214320
+ "learning_rate": 8.568685897435897e-06,
214321
+ "loss": 0.5475,
214322
+ "step": 90350
214323
+ },
214324
+ {
214325
+ "epoch": 722.05,
214326
+ "learning_rate": 8.56860576923077e-06,
214327
+ "loss": 0.2935,
214328
+ "step": 90355
214329
+ },
214330
+ {
214331
+ "epoch": 722.09,
214332
+ "learning_rate": 8.568525641025642e-06,
214333
+ "loss": 0.306,
214334
+ "step": 90360
214335
+ },
214336
+ {
214337
+ "epoch": 722.13,
214338
+ "learning_rate": 8.568445512820513e-06,
214339
+ "loss": 0.3145,
214340
+ "step": 90365
214341
+ },
214342
+ {
214343
+ "epoch": 722.17,
214344
+ "learning_rate": 8.568365384615386e-06,
214345
+ "loss": 0.5436,
214346
+ "step": 90370
214347
+ },
214348
+ {
214349
+ "epoch": 722.21,
214350
+ "learning_rate": 8.568285256410258e-06,
214351
+ "loss": 1.26,
214352
+ "step": 90375
214353
+ },
214354
+ {
214355
+ "epoch": 722.25,
214356
+ "learning_rate": 8.568205128205129e-06,
214357
+ "loss": 0.295,
214358
+ "step": 90380
214359
+ },
214360
+ {
214361
+ "epoch": 722.29,
214362
+ "learning_rate": 8.568125e-06,
214363
+ "loss": 0.2853,
214364
+ "step": 90385
214365
+ },
214366
+ {
214367
+ "epoch": 722.33,
214368
+ "learning_rate": 8.568044871794873e-06,
214369
+ "loss": 0.3319,
214370
+ "step": 90390
214371
+ },
214372
+ {
214373
+ "epoch": 722.37,
214374
+ "learning_rate": 8.567964743589745e-06,
214375
+ "loss": 0.5341,
214376
+ "step": 90395
214377
+ },
214378
+ {
214379
+ "epoch": 722.41,
214380
+ "learning_rate": 8.567884615384616e-06,
214381
+ "loss": 1.1765,
214382
+ "step": 90400
214383
+ },
214384
+ {
214385
+ "epoch": 722.45,
214386
+ "learning_rate": 8.567804487179487e-06,
214387
+ "loss": 0.336,
214388
+ "step": 90405
214389
+ },
214390
+ {
214391
+ "epoch": 722.49,
214392
+ "learning_rate": 8.56772435897436e-06,
214393
+ "loss": 0.2876,
214394
+ "step": 90410
214395
+ },
214396
+ {
214397
+ "epoch": 722.53,
214398
+ "learning_rate": 8.567644230769232e-06,
214399
+ "loss": 0.3503,
214400
+ "step": 90415
214401
+ },
214402
+ {
214403
+ "epoch": 722.57,
214404
+ "learning_rate": 8.567564102564103e-06,
214405
+ "loss": 0.5086,
214406
+ "step": 90420
214407
+ },
214408
+ {
214409
+ "epoch": 722.61,
214410
+ "learning_rate": 8.567483974358976e-06,
214411
+ "loss": 1.1796,
214412
+ "step": 90425
214413
+ },
214414
+ {
214415
+ "epoch": 722.65,
214416
+ "learning_rate": 8.567403846153846e-06,
214417
+ "loss": 0.3368,
214418
+ "step": 90430
214419
+ },
214420
+ {
214421
+ "epoch": 722.69,
214422
+ "learning_rate": 8.567323717948719e-06,
214423
+ "loss": 0.2575,
214424
+ "step": 90435
214425
+ },
214426
+ {
214427
+ "epoch": 722.73,
214428
+ "learning_rate": 8.56724358974359e-06,
214429
+ "loss": 0.2965,
214430
+ "step": 90440
214431
+ },
214432
+ {
214433
+ "epoch": 722.77,
214434
+ "learning_rate": 8.567163461538462e-06,
214435
+ "loss": 0.52,
214436
+ "step": 90445
214437
+ },
214438
+ {
214439
+ "epoch": 722.81,
214440
+ "learning_rate": 8.567083333333333e-06,
214441
+ "loss": 1.0366,
214442
+ "step": 90450
214443
+ },
214444
+ {
214445
+ "epoch": 722.85,
214446
+ "learning_rate": 8.567003205128206e-06,
214447
+ "loss": 0.3497,
214448
+ "step": 90455
214449
+ },
214450
+ {
214451
+ "epoch": 722.89,
214452
+ "learning_rate": 8.566923076923077e-06,
214453
+ "loss": 0.3024,
214454
+ "step": 90460
214455
+ },
214456
+ {
214457
+ "epoch": 722.93,
214458
+ "learning_rate": 8.566842948717949e-06,
214459
+ "loss": 0.3712,
214460
+ "step": 90465
214461
+ },
214462
+ {
214463
+ "epoch": 722.97,
214464
+ "learning_rate": 8.566762820512822e-06,
214465
+ "loss": 0.5973,
214466
+ "step": 90470
214467
+ },
214468
+ {
214469
+ "epoch": 723.0,
214470
+ "eval_loss": 0.40732017159461975,
214471
+ "eval_runtime": 37.8665,
214472
+ "eval_samples_per_second": 22.21,
214473
+ "eval_steps_per_second": 0.713,
214474
+ "eval_wer": 0.18586561967553153,
214475
+ "step": 90474
214476
+ },
214477
+ {
214478
+ "epoch": 729.01,
214479
+ "learning_rate": 8.566682692307693e-06,
214480
+ "loss": 0.3401,
214481
+ "step": 90475
214482
+ },
214483
+ {
214484
+ "epoch": 729.05,
214485
+ "learning_rate": 8.566602564102565e-06,
214486
+ "loss": 0.2985,
214487
+ "step": 90480
214488
+ },
214489
+ {
214490
+ "epoch": 729.09,
214491
+ "learning_rate": 8.566522435897436e-06,
214492
+ "loss": 0.3321,
214493
+ "step": 90485
214494
+ },
214495
+ {
214496
+ "epoch": 729.13,
214497
+ "learning_rate": 8.566442307692309e-06,
214498
+ "loss": 0.3825,
214499
+ "step": 90490
214500
+ },
214501
+ {
214502
+ "epoch": 729.17,
214503
+ "learning_rate": 8.56636217948718e-06,
214504
+ "loss": 0.5136,
214505
+ "step": 90495
214506
+ },
214507
+ {
214508
+ "epoch": 729.21,
214509
+ "learning_rate": 8.566282051282052e-06,
214510
+ "loss": 1.1282,
214511
+ "step": 90500
214512
+ },
214513
+ {
214514
+ "epoch": 729.25,
214515
+ "learning_rate": 8.566201923076923e-06,
214516
+ "loss": 0.3183,
214517
+ "step": 90505
214518
+ },
214519
+ {
214520
+ "epoch": 729.29,
214521
+ "learning_rate": 8.566121794871796e-06,
214522
+ "loss": 0.2711,
214523
+ "step": 90510
214524
+ },
214525
+ {
214526
+ "epoch": 729.33,
214527
+ "learning_rate": 8.566041666666667e-06,
214528
+ "loss": 0.3284,
214529
+ "step": 90515
214530
+ },
214531
+ {
214532
+ "epoch": 729.37,
214533
+ "learning_rate": 8.565961538461539e-06,
214534
+ "loss": 0.5451,
214535
+ "step": 90520
214536
+ },
214537
+ {
214538
+ "epoch": 729.41,
214539
+ "learning_rate": 8.565881410256412e-06,
214540
+ "loss": 1.1487,
214541
+ "step": 90525
214542
+ },
214543
+ {
214544
+ "epoch": 729.45,
214545
+ "learning_rate": 8.565801282051283e-06,
214546
+ "loss": 0.3511,
214547
+ "step": 90530
214548
+ },
214549
+ {
214550
+ "epoch": 729.49,
214551
+ "learning_rate": 8.565721153846155e-06,
214552
+ "loss": 0.3186,
214553
+ "step": 90535
214554
+ },
214555
+ {
214556
+ "epoch": 729.53,
214557
+ "learning_rate": 8.565641025641026e-06,
214558
+ "loss": 0.3456,
214559
+ "step": 90540
214560
+ },
214561
+ {
214562
+ "epoch": 729.57,
214563
+ "learning_rate": 8.565560897435899e-06,
214564
+ "loss": 0.5414,
214565
+ "step": 90545
214566
+ },
214567
+ {
214568
+ "epoch": 729.61,
214569
+ "learning_rate": 8.56548076923077e-06,
214570
+ "loss": 1.1278,
214571
+ "step": 90550
214572
+ },
214573
+ {
214574
+ "epoch": 729.65,
214575
+ "learning_rate": 8.565400641025642e-06,
214576
+ "loss": 0.3432,
214577
+ "step": 90555
214578
+ },
214579
+ {
214580
+ "epoch": 729.69,
214581
+ "learning_rate": 8.565320512820513e-06,
214582
+ "loss": 0.2526,
214583
+ "step": 90560
214584
+ },
214585
+ {
214586
+ "epoch": 729.73,
214587
+ "learning_rate": 8.565240384615386e-06,
214588
+ "loss": 0.4858,
214589
+ "step": 90565
214590
+ },
214591
+ {
214592
+ "epoch": 729.77,
214593
+ "learning_rate": 8.565160256410257e-06,
214594
+ "loss": 0.5345,
214595
+ "step": 90570
214596
+ },
214597
+ {
214598
+ "epoch": 729.81,
214599
+ "learning_rate": 8.565080128205129e-06,
214600
+ "loss": 1.1488,
214601
+ "step": 90575
214602
+ },
214603
+ {
214604
+ "epoch": 729.85,
214605
+ "learning_rate": 8.565000000000002e-06,
214606
+ "loss": 0.249,
214607
+ "step": 90580
214608
+ },
214609
+ {
214610
+ "epoch": 729.89,
214611
+ "learning_rate": 8.564919871794872e-06,
214612
+ "loss": 0.3071,
214613
+ "step": 90585
214614
+ },
214615
+ {
214616
+ "epoch": 729.93,
214617
+ "learning_rate": 8.564839743589745e-06,
214618
+ "loss": 0.3299,
214619
+ "step": 90590
214620
+ },
214621
+ {
214622
+ "epoch": 729.97,
214623
+ "learning_rate": 8.564759615384616e-06,
214624
+ "loss": 0.5956,
214625
+ "step": 90595
214626
+ },
214627
+ {
214628
+ "epoch": 730.0,
214629
+ "eval_loss": 0.4340682923793793,
214630
+ "eval_runtime": 39.1955,
214631
+ "eval_samples_per_second": 21.482,
214632
+ "eval_steps_per_second": 0.689,
214633
+ "eval_wer": 0.18922451350566366,
214634
+ "step": 90598
214635
+ },
214636
+ {
214637
+ "epoch": 724.02,
214638
+ "learning_rate": 8.564679487179487e-06,
214639
+ "loss": 0.3322,
214640
+ "step": 90600
214641
+ },
214642
+ {
214643
+ "epoch": 724.06,
214644
+ "learning_rate": 8.564599358974359e-06,
214645
+ "loss": 0.305,
214646
+ "step": 90605
214647
+ },
214648
+ {
214649
+ "epoch": 724.1,
214650
+ "learning_rate": 8.564519230769232e-06,
214651
+ "loss": 0.3531,
214652
+ "step": 90610
214653
+ },
214654
+ {
214655
+ "epoch": 724.14,
214656
+ "learning_rate": 8.564439102564103e-06,
214657
+ "loss": 0.3253,
214658
+ "step": 90615
214659
+ },
214660
+ {
214661
+ "epoch": 724.18,
214662
+ "learning_rate": 8.564358974358974e-06,
214663
+ "loss": 0.7224,
214664
+ "step": 90620
214665
+ },
214666
+ {
214667
+ "epoch": 724.22,
214668
+ "learning_rate": 8.564278846153847e-06,
214669
+ "loss": 1.1113,
214670
+ "step": 90625
214671
+ },
214672
+ {
214673
+ "epoch": 724.26,
214674
+ "learning_rate": 8.564198717948719e-06,
214675
+ "loss": 0.2906,
214676
+ "step": 90630
214677
+ },
214678
+ {
214679
+ "epoch": 724.3,
214680
+ "learning_rate": 8.56411858974359e-06,
214681
+ "loss": 0.3082,
214682
+ "step": 90635
214683
+ },
214684
+ {
214685
+ "epoch": 724.34,
214686
+ "learning_rate": 8.564038461538462e-06,
214687
+ "loss": 0.3138,
214688
+ "step": 90640
214689
+ },
214690
+ {
214691
+ "epoch": 724.38,
214692
+ "learning_rate": 8.563958333333335e-06,
214693
+ "loss": 0.6324,
214694
+ "step": 90645
214695
+ },
214696
+ {
214697
+ "epoch": 724.42,
214698
+ "learning_rate": 8.563878205128206e-06,
214699
+ "loss": 0.8806,
214700
+ "step": 90650
214701
+ },
214702
+ {
214703
+ "epoch": 724.46,
214704
+ "learning_rate": 8.563798076923077e-06,
214705
+ "loss": 0.2908,
214706
+ "step": 90655
214707
+ },
214708
+ {
214709
+ "epoch": 724.5,
214710
+ "learning_rate": 8.563717948717949e-06,
214711
+ "loss": 0.2909,
214712
+ "step": 90660
214713
+ },
214714
+ {
214715
+ "epoch": 724.54,
214716
+ "learning_rate": 8.563637820512822e-06,
214717
+ "loss": 0.3203,
214718
+ "step": 90665
214719
+ },
214720
+ {
214721
+ "epoch": 724.58,
214722
+ "learning_rate": 8.563557692307693e-06,
214723
+ "loss": 0.6635,
214724
+ "step": 90670
214725
+ },
214726
+ {
214727
+ "epoch": 724.62,
214728
+ "learning_rate": 8.563477564102564e-06,
214729
+ "loss": 1.0963,
214730
+ "step": 90675
214731
+ },
214732
+ {
214733
+ "epoch": 724.66,
214734
+ "learning_rate": 8.563397435897437e-06,
214735
+ "loss": 0.3117,
214736
+ "step": 90680
214737
+ },
214738
+ {
214739
+ "epoch": 724.7,
214740
+ "learning_rate": 8.563317307692309e-06,
214741
+ "loss": 0.2845,
214742
+ "step": 90685
214743
+ },
214744
+ {
214745
+ "epoch": 724.74,
214746
+ "learning_rate": 8.56323717948718e-06,
214747
+ "loss": 0.4016,
214748
+ "step": 90690
214749
+ },
214750
+ {
214751
+ "epoch": 724.78,
214752
+ "learning_rate": 8.563157051282052e-06,
214753
+ "loss": 0.6272,
214754
+ "step": 90695
214755
+ },
214756
+ {
214757
+ "epoch": 724.82,
214758
+ "learning_rate": 8.563076923076925e-06,
214759
+ "loss": 1.0456,
214760
+ "step": 90700
214761
+ },
214762
+ {
214763
+ "epoch": 724.86,
214764
+ "learning_rate": 8.562996794871796e-06,
214765
+ "loss": 0.3503,
214766
+ "step": 90705
214767
+ },
214768
+ {
214769
+ "epoch": 724.9,
214770
+ "learning_rate": 8.562916666666667e-06,
214771
+ "loss": 0.4135,
214772
+ "step": 90710
214773
+ },
214774
+ {
214775
+ "epoch": 724.94,
214776
+ "learning_rate": 8.562836538461539e-06,
214777
+ "loss": 0.4302,
214778
+ "step": 90715
214779
+ },
214780
+ {
214781
+ "epoch": 724.98,
214782
+ "learning_rate": 8.562756410256412e-06,
214783
+ "loss": 0.6428,
214784
+ "step": 90720
214785
+ },
214786
+ {
214787
+ "epoch": 725.0,
214788
+ "eval_loss": 0.39161810278892517,
214789
+ "eval_runtime": 39.4302,
214790
+ "eval_samples_per_second": 21.354,
214791
+ "eval_steps_per_second": 0.685,
214792
+ "eval_wer": 0.18582597364992037,
214793
+ "step": 90723
214794
+ },
214795
+ {
214796
+ "epoch": 731.02,
214797
+ "learning_rate": 8.562676282051283e-06,
214798
+ "loss": 0.4568,
214799
+ "step": 90725
214800
+ },
214801
+ {
214802
+ "epoch": 731.06,
214803
+ "learning_rate": 8.562596153846154e-06,
214804
+ "loss": 0.2456,
214805
+ "step": 90730
214806
+ },
214807
+ {
214808
+ "epoch": 731.1,
214809
+ "learning_rate": 8.562516025641028e-06,
214810
+ "loss": 0.3276,
214811
+ "step": 90735
214812
+ },
214813
+ {
214814
+ "epoch": 731.14,
214815
+ "learning_rate": 8.562435897435897e-06,
214816
+ "loss": 0.3748,
214817
+ "step": 90740
214818
+ },
214819
+ {
214820
+ "epoch": 731.18,
214821
+ "learning_rate": 8.56235576923077e-06,
214822
+ "loss": 0.675,
214823
+ "step": 90745
214824
+ },
214825
+ {
214826
+ "epoch": 731.22,
214827
+ "learning_rate": 8.562275641025642e-06,
214828
+ "loss": 1.0248,
214829
+ "step": 90750
214830
+ },
214831
+ {
214832
+ "epoch": 731.26,
214833
+ "learning_rate": 8.562195512820513e-06,
214834
+ "loss": 0.3,
214835
+ "step": 90755
214836
+ },
214837
+ {
214838
+ "epoch": 731.3,
214839
+ "learning_rate": 8.562115384615384e-06,
214840
+ "loss": 0.285,
214841
+ "step": 90760
214842
+ },
214843
+ {
214844
+ "epoch": 731.34,
214845
+ "learning_rate": 8.562035256410257e-06,
214846
+ "loss": 0.3437,
214847
+ "step": 90765
214848
+ },
214849
+ {
214850
+ "epoch": 731.38,
214851
+ "learning_rate": 8.561955128205129e-06,
214852
+ "loss": 0.6141,
214853
+ "step": 90770
214854
+ },
214855
+ {
214856
+ "epoch": 731.42,
214857
+ "learning_rate": 8.561875e-06,
214858
+ "loss": 1.1171,
214859
+ "step": 90775
214860
+ },
214861
+ {
214862
+ "epoch": 731.46,
214863
+ "learning_rate": 8.561794871794873e-06,
214864
+ "loss": 0.2953,
214865
+ "step": 90780
214866
+ },
214867
+ {
214868
+ "epoch": 731.5,
214869
+ "learning_rate": 8.561714743589744e-06,
214870
+ "loss": 0.3753,
214871
+ "step": 90785
214872
+ },
214873
+ {
214874
+ "epoch": 731.54,
214875
+ "learning_rate": 8.561634615384616e-06,
214876
+ "loss": 0.3691,
214877
+ "step": 90790
214878
+ },
214879
+ {
214880
+ "epoch": 731.58,
214881
+ "learning_rate": 8.561554487179487e-06,
214882
+ "loss": 0.6694,
214883
+ "step": 90795
214884
+ },
214885
+ {
214886
+ "epoch": 731.62,
214887
+ "learning_rate": 8.56147435897436e-06,
214888
+ "loss": 1.198,
214889
+ "step": 90800
214890
+ },
214891
+ {
214892
+ "epoch": 731.66,
214893
+ "learning_rate": 8.561394230769232e-06,
214894
+ "loss": 0.2969,
214895
+ "step": 90805
214896
+ },
214897
+ {
214898
+ "epoch": 731.7,
214899
+ "learning_rate": 8.561314102564103e-06,
214900
+ "loss": 0.2967,
214901
+ "step": 90810
214902
+ },
214903
+ {
214904
+ "epoch": 731.74,
214905
+ "learning_rate": 8.561233974358974e-06,
214906
+ "loss": 0.4696,
214907
+ "step": 90815
214908
+ },
214909
+ {
214910
+ "epoch": 731.78,
214911
+ "learning_rate": 8.561153846153847e-06,
214912
+ "loss": 0.6796,
214913
+ "step": 90820
214914
+ },
214915
+ {
214916
+ "epoch": 731.82,
214917
+ "learning_rate": 8.561073717948719e-06,
214918
+ "loss": 1.075,
214919
+ "step": 90825
214920
+ },
214921
+ {
214922
+ "epoch": 731.86,
214923
+ "learning_rate": 8.56099358974359e-06,
214924
+ "loss": 0.3209,
214925
+ "step": 90830
214926
+ },
214927
+ {
214928
+ "epoch": 731.9,
214929
+ "learning_rate": 8.560913461538463e-06,
214930
+ "loss": 0.3053,
214931
+ "step": 90835
214932
+ },
214933
+ {
214934
+ "epoch": 731.94,
214935
+ "learning_rate": 8.560833333333335e-06,
214936
+ "loss": 0.3938,
214937
+ "step": 90840
214938
+ },
214939
+ {
214940
+ "epoch": 731.98,
214941
+ "learning_rate": 8.560753205128206e-06,
214942
+ "loss": 0.7524,
214943
+ "step": 90845
214944
+ },
214945
+ {
214946
+ "epoch": 732.0,
214947
+ "eval_loss": 0.3970924913883209,
214948
+ "eval_runtime": 39.9515,
214949
+ "eval_samples_per_second": 21.076,
214950
+ "eval_steps_per_second": 0.676,
214951
+ "eval_wer": 0.17860725032239577,
214952
+ "step": 90847
214953
  }
214954
  ],
214955
+ "max_steps": 620000,
214956
  "num_train_epochs": 5000,
214957
+ "total_flos": 2.556513208848127e+20,
214958
  "trial_name": null,
214959
  "trial_params": null
214960
  }
model-bin/finetune/base/{checkpoint-90224 β†’ checkpoint-90847}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629946430.8098779/events.out.tfevents.1629946430.8e89bd551565.924.31 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0618a1807167f1dcba33af2fffd998886f3965b98f6dbcfde0bb73d3aa883d4c
3
+ size 4194
model-bin/finetune/base/log/1629946866.7205796/events.out.tfevents.1629946866.8e89bd551565.924.33 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8461f1cda7357540abedb29d6df5a00a94f312a1d0cf63d78c1bb313dcf6640
3
+ size 4194
model-bin/finetune/base/log/1629947293.5282943/events.out.tfevents.1629947293.8e89bd551565.924.35 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9bb91f64684415110a78bcf66515040a66195052e35a55962e99a1800e6e2e3
3
+ size 4194
model-bin/finetune/base/log/1629947731.4501543/events.out.tfevents.1629947731.8e89bd551565.924.37 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f85ba0707a05a4eac1d3b6c23404c5140f2a7ac188ddbf1d36288650efcbb4fb
3
+ size 4194
model-bin/finetune/base/log/1629948172.226034/events.out.tfevents.1629948172.8e89bd551565.924.39 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4befea1585251bc7c99b383b5bfd7684f3a4860350acf26f3a68bff29eab240
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629946430.8e89bd551565.924.30 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e2718ce0307e47ebd148df27a8de268e07787c55f85ff7f88559b55142085cc
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629946865.8e89bd551565.924.32 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8926a6ebf4ccc025156d8e7d5f8295105dabc670214fae5e7bb09fedc7e5cc51
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629947293.8e89bd551565.924.34 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4c2821a24b944ef3287bd7b53f04a6866aad0d533603765ff98281bc9705a55
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629947731.8e89bd551565.924.36 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89a8ec8a55692aadd934e9cf43ff933855378f67141e4dff560486c1043bbdab
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629948172.8e89bd551565.924.38 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ab273cf29802d043bb6104ea525d62f2b868b25aa762204b1073c9a93428b56
3
+ size 8622