Check commited on
Commit
261f8c2
Β·
1 Parent(s): 533c8ff

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/trainer_state.json +792 -3
  9. model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1629980056.6178455/events.out.tfevents.1629980056.8e89bd551565.924.181 +3 -0
  11. model-bin/finetune/base/log/1629980506.0580995/events.out.tfevents.1629980506.8e89bd551565.924.183 +3 -0
  12. model-bin/finetune/base/log/1629980941.2291443/events.out.tfevents.1629980941.8e89bd551565.924.185 +3 -0
  13. model-bin/finetune/base/log/1629981376.296084/events.out.tfevents.1629981376.8e89bd551565.924.187 +3 -0
  14. model-bin/finetune/base/log/1629981805.4906385/events.out.tfevents.1629981805.8e89bd551565.924.189 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1629980056.8e89bd551565.924.180 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1629980506.8e89bd551565.924.182 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1629980940.8e89bd551565.924.184 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1629981375.8e89bd551565.924.186 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1629981805.8e89bd551565.924.188 +3 -0
model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2f5597ab892dc7bbd063f5a173c1dd124665bf7d5e017c2f7fb8b03707277384
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5ca2134f4b3ba0da8edcb8e6cd1388b203548ff4a96352c981561cd13fb011f
3
  size 722165393
model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ebc5f970b96c488c19c31e1cb0e4f66aaa8ff3d2a45e1e91dc593de1a8d70888
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0466fb5a9f526bdd62d1df1d0510a451ed10596dd03df9cda2a39d5cc3c2d193
3
  size 377909911
model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:193ad6dc9913f8bd9eee397c4a62b1b7b066bdf26b0dc1694e324b4569f39718
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93aa4f1f1b23c8a0be7433fdd4a15ed2e3afdb6362c025c63304d955b95cc4e1
3
  size 14503
model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4166c0b981defd24252fdcfa267326c23bb9a67917e23b9f16f5ee97f24c853d
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0704821450ad2e472fe6ec90dc8448a1296ca59e4ae62277041ac58c687cc511
3
  size 559
model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bdfc4589d4b1984faf3a1eeeac0648257c835994e7c2e69d5928ff2a41e4c405
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d929aa920368a347fd351edf6478d569d94686bcf1b97c344d99188ee6c8acb
3
  size 623
model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.17565318086415285,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-94333",
4
- "epoch": 801.995983935743,
5
- "global_step": 99560,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -226038,11 +226038,800 @@
226038
  "eval_steps_per_second": 0.688,
226039
  "eval_wer": 0.17951169289601412,
226040
  "step": 99560
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
226041
  }
226042
  ],
226043
  "max_steps": 620000,
226044
  "num_train_epochs": 5000,
226045
- "total_flos": 2.801814193016803e+20,
226046
  "trial_name": null,
226047
  "trial_params": null
226048
  }
 
1
  {
2
  "best_metric": 0.17565318086415285,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-94333",
4
+ "epoch": 807.0,
5
+ "global_step": 100181,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
226038
  "eval_steps_per_second": 0.688,
226039
  "eval_wer": 0.17951169289601412,
226040
  "step": 99560
226041
+ },
226042
+ {
226043
+ "epoch": 802.04,
226044
+ "learning_rate": 8.421073717948718e-06,
226045
+ "loss": 0.3382,
226046
+ "step": 99565
226047
+ },
226048
+ {
226049
+ "epoch": 802.08,
226050
+ "learning_rate": 8.42099358974359e-06,
226051
+ "loss": 0.286,
226052
+ "step": 99570
226053
+ },
226054
+ {
226055
+ "epoch": 802.12,
226056
+ "learning_rate": 8.420913461538463e-06,
226057
+ "loss": 0.2989,
226058
+ "step": 99575
226059
+ },
226060
+ {
226061
+ "epoch": 802.16,
226062
+ "learning_rate": 8.420833333333334e-06,
226063
+ "loss": 0.4523,
226064
+ "step": 99580
226065
+ },
226066
+ {
226067
+ "epoch": 802.2,
226068
+ "learning_rate": 8.420753205128206e-06,
226069
+ "loss": 1.4162,
226070
+ "step": 99585
226071
+ },
226072
+ {
226073
+ "epoch": 802.24,
226074
+ "learning_rate": 8.420673076923077e-06,
226075
+ "loss": 0.33,
226076
+ "step": 99590
226077
+ },
226078
+ {
226079
+ "epoch": 802.28,
226080
+ "learning_rate": 8.42059294871795e-06,
226081
+ "loss": 0.3307,
226082
+ "step": 99595
226083
+ },
226084
+ {
226085
+ "epoch": 802.32,
226086
+ "learning_rate": 8.420512820512821e-06,
226087
+ "loss": 0.3105,
226088
+ "step": 99600
226089
+ },
226090
+ {
226091
+ "epoch": 802.36,
226092
+ "learning_rate": 8.420432692307693e-06,
226093
+ "loss": 0.4323,
226094
+ "step": 99605
226095
+ },
226096
+ {
226097
+ "epoch": 802.4,
226098
+ "learning_rate": 8.420352564102566e-06,
226099
+ "loss": 1.2248,
226100
+ "step": 99610
226101
+ },
226102
+ {
226103
+ "epoch": 802.44,
226104
+ "learning_rate": 8.420272435897437e-06,
226105
+ "loss": 0.3468,
226106
+ "step": 99615
226107
+ },
226108
+ {
226109
+ "epoch": 802.48,
226110
+ "learning_rate": 8.420192307692308e-06,
226111
+ "loss": 0.3016,
226112
+ "step": 99620
226113
+ },
226114
+ {
226115
+ "epoch": 802.52,
226116
+ "learning_rate": 8.42011217948718e-06,
226117
+ "loss": 0.3349,
226118
+ "step": 99625
226119
+ },
226120
+ {
226121
+ "epoch": 802.56,
226122
+ "learning_rate": 8.420032051282053e-06,
226123
+ "loss": 0.4675,
226124
+ "step": 99630
226125
+ },
226126
+ {
226127
+ "epoch": 802.6,
226128
+ "learning_rate": 8.419951923076922e-06,
226129
+ "loss": 1.047,
226130
+ "step": 99635
226131
+ },
226132
+ {
226133
+ "epoch": 802.64,
226134
+ "learning_rate": 8.419871794871796e-06,
226135
+ "loss": 0.3183,
226136
+ "step": 99640
226137
+ },
226138
+ {
226139
+ "epoch": 802.68,
226140
+ "learning_rate": 8.419791666666667e-06,
226141
+ "loss": 0.2828,
226142
+ "step": 99645
226143
+ },
226144
+ {
226145
+ "epoch": 802.72,
226146
+ "learning_rate": 8.419711538461538e-06,
226147
+ "loss": 0.327,
226148
+ "step": 99650
226149
+ },
226150
+ {
226151
+ "epoch": 802.76,
226152
+ "learning_rate": 8.419631410256411e-06,
226153
+ "loss": 0.473,
226154
+ "step": 99655
226155
+ },
226156
+ {
226157
+ "epoch": 802.8,
226158
+ "learning_rate": 8.419551282051283e-06,
226159
+ "loss": 1.1045,
226160
+ "step": 99660
226161
+ },
226162
+ {
226163
+ "epoch": 802.84,
226164
+ "learning_rate": 8.419471153846154e-06,
226165
+ "loss": 0.3454,
226166
+ "step": 99665
226167
+ },
226168
+ {
226169
+ "epoch": 802.88,
226170
+ "learning_rate": 8.419391025641025e-06,
226171
+ "loss": 0.2748,
226172
+ "step": 99670
226173
+ },
226174
+ {
226175
+ "epoch": 802.92,
226176
+ "learning_rate": 8.419310897435898e-06,
226177
+ "loss": 0.5234,
226178
+ "step": 99675
226179
+ },
226180
+ {
226181
+ "epoch": 802.96,
226182
+ "learning_rate": 8.41923076923077e-06,
226183
+ "loss": 0.477,
226184
+ "step": 99680
226185
+ },
226186
+ {
226187
+ "epoch": 803.0,
226188
+ "eval_loss": 0.47172078490257263,
226189
+ "eval_runtime": 40.19,
226190
+ "eval_samples_per_second": 20.727,
226191
+ "eval_steps_per_second": 0.672,
226192
+ "eval_wer": 0.18185773330432892,
226193
+ "step": 99684
226194
+ },
226195
+ {
226196
+ "epoch": 797.01,
226197
+ "learning_rate": 8.419150641025641e-06,
226198
+ "loss": 0.4009,
226199
+ "step": 99685
226200
+ },
226201
+ {
226202
+ "epoch": 797.05,
226203
+ "learning_rate": 8.419070512820513e-06,
226204
+ "loss": 0.2643,
226205
+ "step": 99690
226206
+ },
226207
+ {
226208
+ "epoch": 797.09,
226209
+ "learning_rate": 8.418990384615386e-06,
226210
+ "loss": 0.4993,
226211
+ "step": 99695
226212
+ },
226213
+ {
226214
+ "epoch": 797.13,
226215
+ "learning_rate": 8.418910256410257e-06,
226216
+ "loss": 0.4462,
226217
+ "step": 99700
226218
+ },
226219
+ {
226220
+ "epoch": 797.17,
226221
+ "learning_rate": 8.418830128205128e-06,
226222
+ "loss": 0.5344,
226223
+ "step": 99705
226224
+ },
226225
+ {
226226
+ "epoch": 797.21,
226227
+ "learning_rate": 8.418750000000001e-06,
226228
+ "loss": 1.177,
226229
+ "step": 99710
226230
+ },
226231
+ {
226232
+ "epoch": 797.25,
226233
+ "learning_rate": 8.418669871794873e-06,
226234
+ "loss": 0.3522,
226235
+ "step": 99715
226236
+ },
226237
+ {
226238
+ "epoch": 797.29,
226239
+ "learning_rate": 8.418589743589744e-06,
226240
+ "loss": 0.2821,
226241
+ "step": 99720
226242
+ },
226243
+ {
226244
+ "epoch": 797.33,
226245
+ "learning_rate": 8.418509615384615e-06,
226246
+ "loss": 0.2643,
226247
+ "step": 99725
226248
+ },
226249
+ {
226250
+ "epoch": 797.37,
226251
+ "learning_rate": 8.418429487179488e-06,
226252
+ "loss": 0.5044,
226253
+ "step": 99730
226254
+ },
226255
+ {
226256
+ "epoch": 797.41,
226257
+ "learning_rate": 8.41834935897436e-06,
226258
+ "loss": 1.3069,
226259
+ "step": 99735
226260
+ },
226261
+ {
226262
+ "epoch": 797.45,
226263
+ "learning_rate": 8.418269230769231e-06,
226264
+ "loss": 0.3163,
226265
+ "step": 99740
226266
+ },
226267
+ {
226268
+ "epoch": 797.49,
226269
+ "learning_rate": 8.418189102564103e-06,
226270
+ "loss": 0.2883,
226271
+ "step": 99745
226272
+ },
226273
+ {
226274
+ "epoch": 797.53,
226275
+ "learning_rate": 8.418108974358976e-06,
226276
+ "loss": 0.3393,
226277
+ "step": 99750
226278
+ },
226279
+ {
226280
+ "epoch": 797.57,
226281
+ "learning_rate": 8.418028846153847e-06,
226282
+ "loss": 0.5629,
226283
+ "step": 99755
226284
+ },
226285
+ {
226286
+ "epoch": 797.61,
226287
+ "learning_rate": 8.417948717948718e-06,
226288
+ "loss": 1.0497,
226289
+ "step": 99760
226290
+ },
226291
+ {
226292
+ "epoch": 797.65,
226293
+ "learning_rate": 8.417868589743591e-06,
226294
+ "loss": 0.3956,
226295
+ "step": 99765
226296
+ },
226297
+ {
226298
+ "epoch": 797.69,
226299
+ "learning_rate": 8.417788461538463e-06,
226300
+ "loss": 0.2644,
226301
+ "step": 99770
226302
+ },
226303
+ {
226304
+ "epoch": 797.73,
226305
+ "learning_rate": 8.417708333333334e-06,
226306
+ "loss": 0.3281,
226307
+ "step": 99775
226308
+ },
226309
+ {
226310
+ "epoch": 797.76,
226311
+ "learning_rate": 8.417628205128205e-06,
226312
+ "loss": 0.5832,
226313
+ "step": 99780
226314
+ },
226315
+ {
226316
+ "epoch": 797.8,
226317
+ "learning_rate": 8.417548076923079e-06,
226318
+ "loss": 1.1425,
226319
+ "step": 99785
226320
+ },
226321
+ {
226322
+ "epoch": 797.84,
226323
+ "learning_rate": 8.417467948717948e-06,
226324
+ "loss": 0.2996,
226325
+ "step": 99790
226326
+ },
226327
+ {
226328
+ "epoch": 797.88,
226329
+ "learning_rate": 8.417387820512821e-06,
226330
+ "loss": 0.3637,
226331
+ "step": 99795
226332
+ },
226333
+ {
226334
+ "epoch": 797.92,
226335
+ "learning_rate": 8.417307692307694e-06,
226336
+ "loss": 0.4234,
226337
+ "step": 99800
226338
+ },
226339
+ {
226340
+ "epoch": 797.96,
226341
+ "learning_rate": 8.417227564102564e-06,
226342
+ "loss": 0.5232,
226343
+ "step": 99805
226344
+ },
226345
+ {
226346
+ "epoch": 798.0,
226347
+ "eval_loss": 0.3478337228298187,
226348
+ "eval_runtime": 37.8125,
226349
+ "eval_samples_per_second": 22.03,
226350
+ "eval_steps_per_second": 0.714,
226351
+ "eval_wer": 0.17775073376762832,
226352
+ "step": 99809
226353
+ },
226354
+ {
226355
+ "epoch": 804.01,
226356
+ "learning_rate": 8.417147435897437e-06,
226357
+ "loss": 4.0085,
226358
+ "step": 99810
226359
+ },
226360
+ {
226361
+ "epoch": 804.05,
226362
+ "learning_rate": 8.417067307692308e-06,
226363
+ "loss": 0.3344,
226364
+ "step": 99815
226365
+ },
226366
+ {
226367
+ "epoch": 804.09,
226368
+ "learning_rate": 8.41698717948718e-06,
226369
+ "loss": 0.413,
226370
+ "step": 99820
226371
+ },
226372
+ {
226373
+ "epoch": 804.13,
226374
+ "learning_rate": 8.416907051282051e-06,
226375
+ "loss": 0.3468,
226376
+ "step": 99825
226377
+ },
226378
+ {
226379
+ "epoch": 804.17,
226380
+ "learning_rate": 8.416826923076924e-06,
226381
+ "loss": 0.4792,
226382
+ "step": 99830
226383
+ },
226384
+ {
226385
+ "epoch": 804.21,
226386
+ "learning_rate": 8.416746794871795e-06,
226387
+ "loss": 1.1734,
226388
+ "step": 99835
226389
+ },
226390
+ {
226391
+ "epoch": 804.25,
226392
+ "learning_rate": 8.416666666666667e-06,
226393
+ "loss": 0.319,
226394
+ "step": 99840
226395
+ },
226396
+ {
226397
+ "epoch": 804.29,
226398
+ "learning_rate": 8.416586538461538e-06,
226399
+ "loss": 0.3238,
226400
+ "step": 99845
226401
+ },
226402
+ {
226403
+ "epoch": 804.33,
226404
+ "learning_rate": 8.416506410256411e-06,
226405
+ "loss": 0.3412,
226406
+ "step": 99850
226407
+ },
226408
+ {
226409
+ "epoch": 804.37,
226410
+ "learning_rate": 8.416426282051283e-06,
226411
+ "loss": 0.5476,
226412
+ "step": 99855
226413
+ },
226414
+ {
226415
+ "epoch": 804.41,
226416
+ "learning_rate": 8.416346153846154e-06,
226417
+ "loss": 1.0643,
226418
+ "step": 99860
226419
+ },
226420
+ {
226421
+ "epoch": 804.45,
226422
+ "learning_rate": 8.416266025641027e-06,
226423
+ "loss": 1.1687,
226424
+ "step": 99865
226425
+ },
226426
+ {
226427
+ "epoch": 804.49,
226428
+ "learning_rate": 8.416185897435898e-06,
226429
+ "loss": 0.3379,
226430
+ "step": 99870
226431
+ },
226432
+ {
226433
+ "epoch": 804.53,
226434
+ "learning_rate": 8.41610576923077e-06,
226435
+ "loss": 0.3514,
226436
+ "step": 99875
226437
+ },
226438
+ {
226439
+ "epoch": 804.57,
226440
+ "learning_rate": 8.416025641025641e-06,
226441
+ "loss": 0.6156,
226442
+ "step": 99880
226443
+ },
226444
+ {
226445
+ "epoch": 804.61,
226446
+ "learning_rate": 8.415945512820514e-06,
226447
+ "loss": 1.1068,
226448
+ "step": 99885
226449
+ },
226450
+ {
226451
+ "epoch": 804.65,
226452
+ "learning_rate": 8.415865384615386e-06,
226453
+ "loss": 0.3151,
226454
+ "step": 99890
226455
+ },
226456
+ {
226457
+ "epoch": 804.69,
226458
+ "learning_rate": 8.415785256410257e-06,
226459
+ "loss": 0.2707,
226460
+ "step": 99895
226461
+ },
226462
+ {
226463
+ "epoch": 804.73,
226464
+ "learning_rate": 8.41570512820513e-06,
226465
+ "loss": 0.3389,
226466
+ "step": 99900
226467
+ },
226468
+ {
226469
+ "epoch": 804.77,
226470
+ "learning_rate": 8.415625000000001e-06,
226471
+ "loss": 0.5223,
226472
+ "step": 99905
226473
+ },
226474
+ {
226475
+ "epoch": 804.81,
226476
+ "learning_rate": 8.415544871794873e-06,
226477
+ "loss": 1.1362,
226478
+ "step": 99910
226479
+ },
226480
+ {
226481
+ "epoch": 804.85,
226482
+ "learning_rate": 8.415464743589744e-06,
226483
+ "loss": 0.2897,
226484
+ "step": 99915
226485
+ },
226486
+ {
226487
+ "epoch": 804.89,
226488
+ "learning_rate": 8.415384615384617e-06,
226489
+ "loss": 0.2626,
226490
+ "step": 99920
226491
+ },
226492
+ {
226493
+ "epoch": 804.93,
226494
+ "learning_rate": 8.415304487179488e-06,
226495
+ "loss": 0.3401,
226496
+ "step": 99925
226497
+ },
226498
+ {
226499
+ "epoch": 804.97,
226500
+ "learning_rate": 8.41522435897436e-06,
226501
+ "loss": 0.6187,
226502
+ "step": 99930
226503
+ },
226504
+ {
226505
+ "epoch": 805.0,
226506
+ "eval_loss": 0.37858957052230835,
226507
+ "eval_runtime": 38.0372,
226508
+ "eval_samples_per_second": 21.9,
226509
+ "eval_steps_per_second": 0.71,
226510
+ "eval_wer": 0.18095796963250935,
226511
+ "step": 99933
226512
+ },
226513
+ {
226514
+ "epoch": 805.02,
226515
+ "learning_rate": 8.415144230769231e-06,
226516
+ "loss": 0.3085,
226517
+ "step": 99935
226518
+ },
226519
+ {
226520
+ "epoch": 805.06,
226521
+ "learning_rate": 8.415064102564104e-06,
226522
+ "loss": 0.2646,
226523
+ "step": 99940
226524
+ },
226525
+ {
226526
+ "epoch": 805.1,
226527
+ "learning_rate": 8.414983974358974e-06,
226528
+ "loss": 0.3165,
226529
+ "step": 99945
226530
+ },
226531
+ {
226532
+ "epoch": 805.14,
226533
+ "learning_rate": 8.414903846153847e-06,
226534
+ "loss": 0.3132,
226535
+ "step": 99950
226536
+ },
226537
+ {
226538
+ "epoch": 805.18,
226539
+ "learning_rate": 8.41482371794872e-06,
226540
+ "loss": 0.6313,
226541
+ "step": 99955
226542
+ },
226543
+ {
226544
+ "epoch": 805.22,
226545
+ "learning_rate": 8.41474358974359e-06,
226546
+ "loss": 1.0675,
226547
+ "step": 99960
226548
+ },
226549
+ {
226550
+ "epoch": 805.26,
226551
+ "learning_rate": 8.414663461538463e-06,
226552
+ "loss": 0.286,
226553
+ "step": 99965
226554
+ },
226555
+ {
226556
+ "epoch": 805.3,
226557
+ "learning_rate": 8.414583333333334e-06,
226558
+ "loss": 0.2669,
226559
+ "step": 99970
226560
+ },
226561
+ {
226562
+ "epoch": 805.34,
226563
+ "learning_rate": 8.414503205128205e-06,
226564
+ "loss": 0.3894,
226565
+ "step": 99975
226566
+ },
226567
+ {
226568
+ "epoch": 805.38,
226569
+ "learning_rate": 8.414423076923077e-06,
226570
+ "loss": 0.628,
226571
+ "step": 99980
226572
+ },
226573
+ {
226574
+ "epoch": 805.42,
226575
+ "learning_rate": 8.41434294871795e-06,
226576
+ "loss": 0.9502,
226577
+ "step": 99985
226578
+ },
226579
+ {
226580
+ "epoch": 805.46,
226581
+ "learning_rate": 8.414262820512821e-06,
226582
+ "loss": 0.3369,
226583
+ "step": 99990
226584
+ },
226585
+ {
226586
+ "epoch": 805.5,
226587
+ "learning_rate": 8.414182692307693e-06,
226588
+ "loss": 0.267,
226589
+ "step": 99995
226590
+ },
226591
+ {
226592
+ "epoch": 805.54,
226593
+ "learning_rate": 8.414102564102566e-06,
226594
+ "loss": 0.3449,
226595
+ "step": 100000
226596
+ },
226597
+ {
226598
+ "epoch": 805.58,
226599
+ "learning_rate": 8.414022435897437e-06,
226600
+ "loss": 0.6615,
226601
+ "step": 100005
226602
+ },
226603
+ {
226604
+ "epoch": 805.62,
226605
+ "learning_rate": 8.413942307692308e-06,
226606
+ "loss": 1.0243,
226607
+ "step": 100010
226608
+ },
226609
+ {
226610
+ "epoch": 805.66,
226611
+ "learning_rate": 8.41386217948718e-06,
226612
+ "loss": 0.2643,
226613
+ "step": 100015
226614
+ },
226615
+ {
226616
+ "epoch": 805.7,
226617
+ "learning_rate": 8.413782051282053e-06,
226618
+ "loss": 0.3039,
226619
+ "step": 100020
226620
+ },
226621
+ {
226622
+ "epoch": 805.74,
226623
+ "learning_rate": 8.413701923076924e-06,
226624
+ "loss": 0.3028,
226625
+ "step": 100025
226626
+ },
226627
+ {
226628
+ "epoch": 805.78,
226629
+ "learning_rate": 8.413621794871795e-06,
226630
+ "loss": 0.5747,
226631
+ "step": 100030
226632
+ },
226633
+ {
226634
+ "epoch": 805.82,
226635
+ "learning_rate": 8.413541666666667e-06,
226636
+ "loss": 1.0076,
226637
+ "step": 100035
226638
+ },
226639
+ {
226640
+ "epoch": 805.86,
226641
+ "learning_rate": 8.41346153846154e-06,
226642
+ "loss": 0.3063,
226643
+ "step": 100040
226644
+ },
226645
+ {
226646
+ "epoch": 805.9,
226647
+ "learning_rate": 8.413381410256411e-06,
226648
+ "loss": 0.311,
226649
+ "step": 100045
226650
+ },
226651
+ {
226652
+ "epoch": 805.94,
226653
+ "learning_rate": 8.413301282051283e-06,
226654
+ "loss": 0.3488,
226655
+ "step": 100050
226656
+ },
226657
+ {
226658
+ "epoch": 805.98,
226659
+ "learning_rate": 8.413221153846156e-06,
226660
+ "loss": 0.6481,
226661
+ "step": 100055
226662
+ },
226663
+ {
226664
+ "epoch": 806.0,
226665
+ "eval_loss": 0.35248780250549316,
226666
+ "eval_runtime": 40.524,
226667
+ "eval_samples_per_second": 20.556,
226668
+ "eval_steps_per_second": 0.666,
226669
+ "eval_wer": 0.18215773595747736,
226670
+ "step": 100057
226671
+ },
226672
+ {
226673
+ "epoch": 806.02,
226674
+ "learning_rate": 8.413141025641027e-06,
226675
+ "loss": 0.2858,
226676
+ "step": 100060
226677
+ },
226678
+ {
226679
+ "epoch": 806.06,
226680
+ "learning_rate": 8.413060897435898e-06,
226681
+ "loss": 0.275,
226682
+ "step": 100065
226683
+ },
226684
+ {
226685
+ "epoch": 806.1,
226686
+ "learning_rate": 8.41298076923077e-06,
226687
+ "loss": 0.2892,
226688
+ "step": 100070
226689
+ },
226690
+ {
226691
+ "epoch": 806.15,
226692
+ "learning_rate": 8.412900641025643e-06,
226693
+ "loss": 0.4414,
226694
+ "step": 100075
226695
+ },
226696
+ {
226697
+ "epoch": 806.19,
226698
+ "learning_rate": 8.412820512820512e-06,
226699
+ "loss": 0.7002,
226700
+ "step": 100080
226701
+ },
226702
+ {
226703
+ "epoch": 806.23,
226704
+ "learning_rate": 8.412740384615385e-06,
226705
+ "loss": 0.7833,
226706
+ "step": 100085
226707
+ },
226708
+ {
226709
+ "epoch": 806.27,
226710
+ "learning_rate": 8.412660256410257e-06,
226711
+ "loss": 0.3272,
226712
+ "step": 100090
226713
+ },
226714
+ {
226715
+ "epoch": 806.31,
226716
+ "learning_rate": 8.412580128205128e-06,
226717
+ "loss": 0.3191,
226718
+ "step": 100095
226719
+ },
226720
+ {
226721
+ "epoch": 806.35,
226722
+ "learning_rate": 8.412500000000001e-06,
226723
+ "loss": 0.3924,
226724
+ "step": 100100
226725
+ },
226726
+ {
226727
+ "epoch": 806.39,
226728
+ "learning_rate": 8.412419871794873e-06,
226729
+ "loss": 0.8189,
226730
+ "step": 100105
226731
+ },
226732
+ {
226733
+ "epoch": 806.43,
226734
+ "learning_rate": 8.412339743589744e-06,
226735
+ "loss": 0.9466,
226736
+ "step": 100110
226737
+ },
226738
+ {
226739
+ "epoch": 806.47,
226740
+ "learning_rate": 8.412259615384615e-06,
226741
+ "loss": 0.2812,
226742
+ "step": 100115
226743
+ },
226744
+ {
226745
+ "epoch": 806.51,
226746
+ "learning_rate": 8.412179487179488e-06,
226747
+ "loss": 0.3069,
226748
+ "step": 100120
226749
+ },
226750
+ {
226751
+ "epoch": 806.55,
226752
+ "learning_rate": 8.41209935897436e-06,
226753
+ "loss": 0.3893,
226754
+ "step": 100125
226755
+ },
226756
+ {
226757
+ "epoch": 806.59,
226758
+ "learning_rate": 8.412019230769231e-06,
226759
+ "loss": 0.6543,
226760
+ "step": 100130
226761
+ },
226762
+ {
226763
+ "epoch": 806.63,
226764
+ "learning_rate": 8.411939102564102e-06,
226765
+ "loss": 0.8736,
226766
+ "step": 100135
226767
+ },
226768
+ {
226769
+ "epoch": 806.67,
226770
+ "learning_rate": 8.411858974358975e-06,
226771
+ "loss": 0.2459,
226772
+ "step": 100140
226773
+ },
226774
+ {
226775
+ "epoch": 806.71,
226776
+ "learning_rate": 8.411778846153847e-06,
226777
+ "loss": 0.3367,
226778
+ "step": 100145
226779
+ },
226780
+ {
226781
+ "epoch": 806.75,
226782
+ "learning_rate": 8.411698717948718e-06,
226783
+ "loss": 0.3835,
226784
+ "step": 100150
226785
+ },
226786
+ {
226787
+ "epoch": 806.79,
226788
+ "learning_rate": 8.411618589743591e-06,
226789
+ "loss": 0.7286,
226790
+ "step": 100155
226791
+ },
226792
+ {
226793
+ "epoch": 806.83,
226794
+ "learning_rate": 8.411538461538463e-06,
226795
+ "loss": 0.8448,
226796
+ "step": 100160
226797
+ },
226798
+ {
226799
+ "epoch": 806.87,
226800
+ "learning_rate": 8.411458333333334e-06,
226801
+ "loss": 0.2768,
226802
+ "step": 100165
226803
+ },
226804
+ {
226805
+ "epoch": 806.91,
226806
+ "learning_rate": 8.411378205128205e-06,
226807
+ "loss": 0.3267,
226808
+ "step": 100170
226809
+ },
226810
+ {
226811
+ "epoch": 806.95,
226812
+ "learning_rate": 8.411298076923078e-06,
226813
+ "loss": 0.3316,
226814
+ "step": 100175
226815
+ },
226816
+ {
226817
+ "epoch": 806.99,
226818
+ "learning_rate": 8.41121794871795e-06,
226819
+ "loss": 0.9675,
226820
+ "step": 100180
226821
+ },
226822
+ {
226823
+ "epoch": 807.0,
226824
+ "eval_loss": 0.4434155821800232,
226825
+ "eval_runtime": 38.2932,
226826
+ "eval_samples_per_second": 21.753,
226827
+ "eval_steps_per_second": 0.705,
226828
+ "eval_wer": 0.18565555718690355,
226829
+ "step": 100181
226830
  }
226831
  ],
226832
  "max_steps": 620000,
226833
  "num_train_epochs": 5000,
226834
+ "total_flos": 2.819371019425974e+20,
226835
  "trial_name": null,
226836
  "trial_params": null
226837
  }
model-bin/finetune/base/{checkpoint-99560 β†’ checkpoint-100181}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1629980056.6178455/events.out.tfevents.1629980056.8e89bd551565.924.181 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c22752d5e9d46587ee5483c6167738121534b5d2bdde66f150ffdafb66afdcd8
3
+ size 4194
model-bin/finetune/base/log/1629980506.0580995/events.out.tfevents.1629980506.8e89bd551565.924.183 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7be4ae668a669990af1b64f28be2b644cbe322156e930ceef12167821d59bc82
3
+ size 4194
model-bin/finetune/base/log/1629980941.2291443/events.out.tfevents.1629980941.8e89bd551565.924.185 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17a0b7a2e7e9a0b6b13f908ec88e35e6d4e5f3a07a9dbf19167403655e20033c
3
+ size 4194
model-bin/finetune/base/log/1629981376.296084/events.out.tfevents.1629981376.8e89bd551565.924.187 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df5a95646d67b5d8b0098832d77b54d7c5d74b8238a6fc9c93852a8217151ec9
3
+ size 4194
model-bin/finetune/base/log/1629981805.4906385/events.out.tfevents.1629981805.8e89bd551565.924.189 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b80bea72f6d24ec071e069b3a61125b31c158a75ee1b97395653a1ac1703572
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1629980056.8e89bd551565.924.180 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45a4bef00a6173531f71d2a4df9a89fed341e8c5edcd49f5caa61f3374908a5f
3
+ size 8462
model-bin/finetune/base/log/events.out.tfevents.1629980506.8e89bd551565.924.182 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:414d773ac83a6e2e0049f36fbb3ec46960411818101a9e9c6f0440e9c0e12879
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629980940.8e89bd551565.924.184 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb2220d7b6f4dd7eb4a0d1d3790872248f76682fa83ef73148fb7f2f2606c1fe
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629981375.8e89bd551565.924.186 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d126c57a4b4ffed8e1a472d38e83fdde836b589a2cf158bbd0b6134cf899201
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1629981805.8e89bd551565.924.188 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5772cba43f717e6e25dcf5b339279ce7260183939d437229aaf3ddfb23886a2
3
+ size 8622