Check commited on
Commit
bc2c169
β€’
1 Parent(s): a9d369f

"auto-commit"

Browse files
Files changed (19) hide show
  1. model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/config.json +0 -0
  2. model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/optimizer.pt +1 -1
  3. model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/preprocessor_config.json +0 -0
  4. model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/pytorch_model.bin +1 -1
  5. model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/rng_state.pth +1 -1
  6. model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/scaler.pt +1 -1
  7. model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/scheduler.pt +1 -1
  8. model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/trainer_state.json +798 -3
  9. model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/training_args.bin +0 -0
  10. model-bin/finetune/base/log/1630259719.897293/events.out.tfevents.1630259719.cc93b136ebf5.1086.249 +3 -0
  11. model-bin/finetune/base/log/1630260157.7758684/events.out.tfevents.1630260157.cc93b136ebf5.1086.251 +3 -0
  12. model-bin/finetune/base/log/1630260591.4709191/events.out.tfevents.1630260591.cc93b136ebf5.1086.253 +3 -0
  13. model-bin/finetune/base/log/1630261026.7725165/events.out.tfevents.1630261026.cc93b136ebf5.1086.255 +3 -0
  14. model-bin/finetune/base/log/1630261460.190783/events.out.tfevents.1630261460.cc93b136ebf5.1086.257 +3 -0
  15. model-bin/finetune/base/log/events.out.tfevents.1630259719.cc93b136ebf5.1086.248 +3 -0
  16. model-bin/finetune/base/log/events.out.tfevents.1630260157.cc93b136ebf5.1086.250 +3 -0
  17. model-bin/finetune/base/log/events.out.tfevents.1630260591.cc93b136ebf5.1086.252 +3 -0
  18. model-bin/finetune/base/log/events.out.tfevents.1630261026.cc93b136ebf5.1086.254 +3 -0
  19. model-bin/finetune/base/log/events.out.tfevents.1630261460.cc93b136ebf5.1086.256 +3 -0
model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/optimizer.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:74ba9b2dfdf6dee25d004cb565b958e98dcf9c4fbd59c57c5467560d1f4d0fa0
3
  size 722165393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:470c3ea26795136bd282cb4eb7e01b82e7f74de658b159421d45cd8357c5df2e
3
  size 722165393
model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/preprocessor_config.json RENAMED
File without changes
model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/pytorch_model.bin RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf7e26fb4cb3b91846741a8ca38d407f10f0f1e4bc6d6d1c322f960abd8cf940
3
  size 377909911
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b99febcccd6a1f50c15709024b6aadd57bf558412f09f39446be17d92b116544
3
  size 377909911
model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/rng_state.pth RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:71bad233cfa0a1d396b5ebc324d1525576abd564039e669d5d2f749724b873a3
3
  size 14503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:847621cdc9516237f2d0544ce00f3369310442cbe0ff46e33f4438f8246622c7
3
  size 14503
model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/scaler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:53047867f5dbec0185f704d1730a6309e0150cb39a35badc8cf7348be089b5bb
3
  size 559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1adce13f705e5edbbf832291b3d4b862eb38f0d353dd7a919d7e7015308b5f53
3
  size 559
model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/scheduler.pt RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e6d2ae467a50189b89589e279f888de8f1fcdfcd87203d6852aac1cd1477f181
3
  size 623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccc55077eebec856cae793ae16f745e98931efbbdb640f5646e5c41e6164ecd5
3
  size 623
model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/trainer_state.json RENAMED
@@ -1,8 +1,8 @@
1
  {
2
  "best_metric": 0.1668011442822563,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
4
- "epoch": 1246.0,
5
- "global_step": 154557,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -296010,11 +296010,806 @@
296010
  "eval_steps_per_second": 0.654,
296011
  "eval_wer": 0.17398648648648649,
296012
  "step": 154557
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
296013
  }
296014
  ],
296015
  "max_steps": 620000,
296016
  "num_train_epochs": 5000,
296017
- "total_flos": 4.349352656879201e+20,
296018
  "trial_name": null,
296019
  "trial_params": null
296020
  }
 
1
  {
2
  "best_metric": 0.1668011442822563,
3
  "best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
4
+ "epoch": 1250.995983935743,
5
+ "global_step": 155180,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
296010
  "eval_steps_per_second": 0.654,
296011
  "eval_wer": 0.17398648648648649,
296012
  "step": 154557
296013
+ },
296014
+ {
296015
+ "epoch": 1246.02,
296016
+ "learning_rate": 7.5202100161550895e-06,
296017
+ "loss": 0.3359,
296018
+ "step": 154560
296019
+ },
296020
+ {
296021
+ "epoch": 1246.06,
296022
+ "learning_rate": 7.520129240710825e-06,
296023
+ "loss": 0.2438,
296024
+ "step": 154565
296025
+ },
296026
+ {
296027
+ "epoch": 1246.1,
296028
+ "learning_rate": 7.5200484652665595e-06,
296029
+ "loss": 0.2681,
296030
+ "step": 154570
296031
+ },
296032
+ {
296033
+ "epoch": 1246.14,
296034
+ "learning_rate": 7.519967689822295e-06,
296035
+ "loss": 0.3789,
296036
+ "step": 154575
296037
+ },
296038
+ {
296039
+ "epoch": 1246.18,
296040
+ "learning_rate": 7.5198869143780295e-06,
296041
+ "loss": 0.6661,
296042
+ "step": 154580
296043
+ },
296044
+ {
296045
+ "epoch": 1246.22,
296046
+ "learning_rate": 7.519806138933765e-06,
296047
+ "loss": 0.8786,
296048
+ "step": 154585
296049
+ },
296050
+ {
296051
+ "epoch": 1246.27,
296052
+ "learning_rate": 7.5197253634894995e-06,
296053
+ "loss": 0.2525,
296054
+ "step": 154590
296055
+ },
296056
+ {
296057
+ "epoch": 1246.31,
296058
+ "learning_rate": 7.519644588045235e-06,
296059
+ "loss": 0.2866,
296060
+ "step": 154595
296061
+ },
296062
+ {
296063
+ "epoch": 1246.35,
296064
+ "learning_rate": 7.5195638126009694e-06,
296065
+ "loss": 0.3159,
296066
+ "step": 154600
296067
+ },
296068
+ {
296069
+ "epoch": 1246.39,
296070
+ "learning_rate": 7.519483037156705e-06,
296071
+ "loss": 0.6285,
296072
+ "step": 154605
296073
+ },
296074
+ {
296075
+ "epoch": 1246.43,
296076
+ "learning_rate": 7.519402261712439e-06,
296077
+ "loss": 0.8862,
296078
+ "step": 154610
296079
+ },
296080
+ {
296081
+ "epoch": 1246.47,
296082
+ "learning_rate": 7.519321486268175e-06,
296083
+ "loss": 0.2578,
296084
+ "step": 154615
296085
+ },
296086
+ {
296087
+ "epoch": 1246.51,
296088
+ "learning_rate": 7.519240710823909e-06,
296089
+ "loss": 0.2415,
296090
+ "step": 154620
296091
+ },
296092
+ {
296093
+ "epoch": 1246.55,
296094
+ "learning_rate": 7.519159935379645e-06,
296095
+ "loss": 0.3066,
296096
+ "step": 154625
296097
+ },
296098
+ {
296099
+ "epoch": 1246.59,
296100
+ "learning_rate": 7.519079159935381e-06,
296101
+ "loss": 0.5628,
296102
+ "step": 154630
296103
+ },
296104
+ {
296105
+ "epoch": 1246.63,
296106
+ "learning_rate": 7.518998384491115e-06,
296107
+ "loss": 0.865,
296108
+ "step": 154635
296109
+ },
296110
+ {
296111
+ "epoch": 1246.67,
296112
+ "learning_rate": 7.518917609046851e-06,
296113
+ "loss": 0.2865,
296114
+ "step": 154640
296115
+ },
296116
+ {
296117
+ "epoch": 1246.71,
296118
+ "learning_rate": 7.518836833602585e-06,
296119
+ "loss": 0.2618,
296120
+ "step": 154645
296121
+ },
296122
+ {
296123
+ "epoch": 1246.75,
296124
+ "learning_rate": 7.518756058158321e-06,
296125
+ "loss": 0.3727,
296126
+ "step": 154650
296127
+ },
296128
+ {
296129
+ "epoch": 1246.79,
296130
+ "learning_rate": 7.518675282714055e-06,
296131
+ "loss": 0.8115,
296132
+ "step": 154655
296133
+ },
296134
+ {
296135
+ "epoch": 1246.83,
296136
+ "learning_rate": 7.518594507269791e-06,
296137
+ "loss": 0.7482,
296138
+ "step": 154660
296139
+ },
296140
+ {
296141
+ "epoch": 1246.87,
296142
+ "learning_rate": 7.518513731825525e-06,
296143
+ "loss": 0.3001,
296144
+ "step": 154665
296145
+ },
296146
+ {
296147
+ "epoch": 1246.91,
296148
+ "learning_rate": 7.518432956381261e-06,
296149
+ "loss": 0.3185,
296150
+ "step": 154670
296151
+ },
296152
+ {
296153
+ "epoch": 1246.95,
296154
+ "learning_rate": 7.518352180936995e-06,
296155
+ "loss": 0.3938,
296156
+ "step": 154675
296157
+ },
296158
+ {
296159
+ "epoch": 1246.99,
296160
+ "learning_rate": 7.518271405492731e-06,
296161
+ "loss": 0.8753,
296162
+ "step": 154680
296163
+ },
296164
+ {
296165
+ "epoch": 1247.0,
296166
+ "eval_loss": 0.3975435793399811,
296167
+ "eval_runtime": 41.5728,
296168
+ "eval_samples_per_second": 20.326,
296169
+ "eval_steps_per_second": 0.649,
296170
+ "eval_wer": 0.1845796090400407,
296171
+ "step": 154681
296172
+ },
296173
+ {
296174
+ "epoch": 1237.03,
296175
+ "learning_rate": 7.518190630048467e-06,
296176
+ "loss": 0.3444,
296177
+ "step": 154685
296178
+ },
296179
+ {
296180
+ "epoch": 1237.07,
296181
+ "learning_rate": 7.518109854604201e-06,
296182
+ "loss": 0.2737,
296183
+ "step": 154690
296184
+ },
296185
+ {
296186
+ "epoch": 1237.11,
296187
+ "learning_rate": 7.518029079159937e-06,
296188
+ "loss": 0.2676,
296189
+ "step": 154695
296190
+ },
296191
+ {
296192
+ "epoch": 1237.15,
296193
+ "learning_rate": 7.517948303715671e-06,
296194
+ "loss": 0.3919,
296195
+ "step": 154700
296196
+ },
296197
+ {
296198
+ "epoch": 1237.19,
296199
+ "learning_rate": 7.517867528271407e-06,
296200
+ "loss": 1.0205,
296201
+ "step": 154705
296202
+ },
296203
+ {
296204
+ "epoch": 1237.23,
296205
+ "learning_rate": 7.517786752827141e-06,
296206
+ "loss": 0.6594,
296207
+ "step": 154710
296208
+ },
296209
+ {
296210
+ "epoch": 1237.27,
296211
+ "learning_rate": 7.517705977382877e-06,
296212
+ "loss": 0.2955,
296213
+ "step": 154715
296214
+ },
296215
+ {
296216
+ "epoch": 1237.31,
296217
+ "learning_rate": 7.517625201938611e-06,
296218
+ "loss": 0.2682,
296219
+ "step": 154720
296220
+ },
296221
+ {
296222
+ "epoch": 1237.35,
296223
+ "learning_rate": 7.517544426494347e-06,
296224
+ "loss": 0.3546,
296225
+ "step": 154725
296226
+ },
296227
+ {
296228
+ "epoch": 1237.39,
296229
+ "learning_rate": 7.517463651050081e-06,
296230
+ "loss": 0.8458,
296231
+ "step": 154730
296232
+ },
296233
+ {
296234
+ "epoch": 1237.43,
296235
+ "learning_rate": 7.517382875605817e-06,
296236
+ "loss": 0.6329,
296237
+ "step": 154735
296238
+ },
296239
+ {
296240
+ "epoch": 1237.47,
296241
+ "learning_rate": 7.517302100161552e-06,
296242
+ "loss": 0.2381,
296243
+ "step": 154740
296244
+ },
296245
+ {
296246
+ "epoch": 1237.51,
296247
+ "learning_rate": 7.517221324717287e-06,
296248
+ "loss": 0.3116,
296249
+ "step": 154745
296250
+ },
296251
+ {
296252
+ "epoch": 1237.55,
296253
+ "learning_rate": 7.5171405492730225e-06,
296254
+ "loss": 0.3383,
296255
+ "step": 154750
296256
+ },
296257
+ {
296258
+ "epoch": 1237.59,
296259
+ "learning_rate": 7.517059773828757e-06,
296260
+ "loss": 0.8708,
296261
+ "step": 154755
296262
+ },
296263
+ {
296264
+ "epoch": 1237.63,
296265
+ "learning_rate": 7.5169789983844924e-06,
296266
+ "loss": 0.5733,
296267
+ "step": 154760
296268
+ },
296269
+ {
296270
+ "epoch": 1237.67,
296271
+ "learning_rate": 7.5168982229402266e-06,
296272
+ "loss": 0.247,
296273
+ "step": 154765
296274
+ },
296275
+ {
296276
+ "epoch": 1237.71,
296277
+ "learning_rate": 7.516817447495962e-06,
296278
+ "loss": 0.3016,
296279
+ "step": 154770
296280
+ },
296281
+ {
296282
+ "epoch": 1237.75,
296283
+ "learning_rate": 7.5167366720516965e-06,
296284
+ "loss": 0.4686,
296285
+ "step": 154775
296286
+ },
296287
+ {
296288
+ "epoch": 1237.79,
296289
+ "learning_rate": 7.516655896607432e-06,
296290
+ "loss": 0.8122,
296291
+ "step": 154780
296292
+ },
296293
+ {
296294
+ "epoch": 1237.83,
296295
+ "learning_rate": 7.5165751211631665e-06,
296296
+ "loss": 0.5604,
296297
+ "step": 154785
296298
+ },
296299
+ {
296300
+ "epoch": 1237.87,
296301
+ "learning_rate": 7.516494345718902e-06,
296302
+ "loss": 0.3368,
296303
+ "step": 154790
296304
+ },
296305
+ {
296306
+ "epoch": 1237.91,
296307
+ "learning_rate": 7.5164135702746365e-06,
296308
+ "loss": 0.2458,
296309
+ "step": 154795
296310
+ },
296311
+ {
296312
+ "epoch": 1237.95,
296313
+ "learning_rate": 7.516332794830372e-06,
296314
+ "loss": 0.3618,
296315
+ "step": 154800
296316
+ },
296317
+ {
296318
+ "epoch": 1237.99,
296319
+ "learning_rate": 7.516252019386107e-06,
296320
+ "loss": 0.931,
296321
+ "step": 154805
296322
+ },
296323
+ {
296324
+ "epoch": 1238.0,
296325
+ "eval_loss": 0.42138466238975525,
296326
+ "eval_runtime": 40.226,
296327
+ "eval_samples_per_second": 20.981,
296328
+ "eval_steps_per_second": 0.671,
296329
+ "eval_wer": 0.17503371796793046,
296330
+ "step": 154806
296331
+ },
296332
+ {
296333
+ "epoch": 1238.03,
296334
+ "learning_rate": 7.516171243941842e-06,
296335
+ "loss": 0.2782,
296336
+ "step": 154810
296337
+ },
296338
+ {
296339
+ "epoch": 1238.07,
296340
+ "learning_rate": 7.516090468497578e-06,
296341
+ "loss": 0.2432,
296342
+ "step": 154815
296343
+ },
296344
+ {
296345
+ "epoch": 1238.11,
296346
+ "learning_rate": 7.516009693053312e-06,
296347
+ "loss": 0.2524,
296348
+ "step": 154820
296349
+ },
296350
+ {
296351
+ "epoch": 1238.15,
296352
+ "learning_rate": 7.515928917609048e-06,
296353
+ "loss": 0.3604,
296354
+ "step": 154825
296355
+ },
296356
+ {
296357
+ "epoch": 1238.19,
296358
+ "learning_rate": 7.515848142164782e-06,
296359
+ "loss": 0.934,
296360
+ "step": 154830
296361
+ },
296362
+ {
296363
+ "epoch": 1238.23,
296364
+ "learning_rate": 7.515767366720518e-06,
296365
+ "loss": 0.5878,
296366
+ "step": 154835
296367
+ },
296368
+ {
296369
+ "epoch": 1238.27,
296370
+ "learning_rate": 7.515686591276252e-06,
296371
+ "loss": 0.2449,
296372
+ "step": 154840
296373
+ },
296374
+ {
296375
+ "epoch": 1238.31,
296376
+ "learning_rate": 7.515605815831988e-06,
296377
+ "loss": 0.2886,
296378
+ "step": 154845
296379
+ },
296380
+ {
296381
+ "epoch": 1238.35,
296382
+ "learning_rate": 7.515525040387722e-06,
296383
+ "loss": 0.3381,
296384
+ "step": 154850
296385
+ },
296386
+ {
296387
+ "epoch": 1238.39,
296388
+ "learning_rate": 7.515444264943458e-06,
296389
+ "loss": 0.7834,
296390
+ "step": 154855
296391
+ },
296392
+ {
296393
+ "epoch": 1238.43,
296394
+ "learning_rate": 7.515363489499193e-06,
296395
+ "loss": 0.7959,
296396
+ "step": 154860
296397
+ },
296398
+ {
296399
+ "epoch": 1238.47,
296400
+ "learning_rate": 7.515282714054928e-06,
296401
+ "loss": 0.2812,
296402
+ "step": 154865
296403
+ },
296404
+ {
296405
+ "epoch": 1238.51,
296406
+ "learning_rate": 7.515201938610663e-06,
296407
+ "loss": 0.2775,
296408
+ "step": 154870
296409
+ },
296410
+ {
296411
+ "epoch": 1238.55,
296412
+ "learning_rate": 7.515121163166398e-06,
296413
+ "loss": 0.4103,
296414
+ "step": 154875
296415
+ },
296416
+ {
296417
+ "epoch": 1238.59,
296418
+ "learning_rate": 7.515040387722134e-06,
296419
+ "loss": 0.8922,
296420
+ "step": 154880
296421
+ },
296422
+ {
296423
+ "epoch": 1238.63,
296424
+ "learning_rate": 7.514959612277868e-06,
296425
+ "loss": 0.696,
296426
+ "step": 154885
296427
+ },
296428
+ {
296429
+ "epoch": 1238.67,
296430
+ "learning_rate": 7.514878836833604e-06,
296431
+ "loss": 0.2623,
296432
+ "step": 154890
296433
+ },
296434
+ {
296435
+ "epoch": 1238.71,
296436
+ "learning_rate": 7.514798061389338e-06,
296437
+ "loss": 0.2932,
296438
+ "step": 154895
296439
+ },
296440
+ {
296441
+ "epoch": 1238.75,
296442
+ "learning_rate": 7.514717285945074e-06,
296443
+ "loss": 0.3314,
296444
+ "step": 154900
296445
+ },
296446
+ {
296447
+ "epoch": 1238.79,
296448
+ "learning_rate": 7.514636510500808e-06,
296449
+ "loss": 0.8592,
296450
+ "step": 154905
296451
+ },
296452
+ {
296453
+ "epoch": 1238.83,
296454
+ "learning_rate": 7.514555735056544e-06,
296455
+ "loss": 0.5791,
296456
+ "step": 154910
296457
+ },
296458
+ {
296459
+ "epoch": 1238.87,
296460
+ "learning_rate": 7.514474959612279e-06,
296461
+ "loss": 0.2748,
296462
+ "step": 154915
296463
+ },
296464
+ {
296465
+ "epoch": 1238.91,
296466
+ "learning_rate": 7.514394184168014e-06,
296467
+ "loss": 0.3728,
296468
+ "step": 154920
296469
+ },
296470
+ {
296471
+ "epoch": 1238.95,
296472
+ "learning_rate": 7.514313408723749e-06,
296473
+ "loss": 0.4352,
296474
+ "step": 154925
296475
+ },
296476
+ {
296477
+ "epoch": 1238.99,
296478
+ "learning_rate": 7.514232633279484e-06,
296479
+ "loss": 0.8688,
296480
+ "step": 154930
296481
+ },
296482
+ {
296483
+ "epoch": 1239.0,
296484
+ "eval_loss": 0.36569637060165405,
296485
+ "eval_runtime": 40.9116,
296486
+ "eval_samples_per_second": 20.654,
296487
+ "eval_steps_per_second": 0.66,
296488
+ "eval_wer": 0.18007352941176472,
296489
+ "step": 154931
296490
+ },
296491
+ {
296492
+ "epoch": 1239.03,
296493
+ "learning_rate": 7.514151857835219e-06,
296494
+ "loss": 0.3529,
296495
+ "step": 154935
296496
+ },
296497
+ {
296498
+ "epoch": 1239.07,
296499
+ "learning_rate": 7.514071082390954e-06,
296500
+ "loss": 0.2858,
296501
+ "step": 154940
296502
+ },
296503
+ {
296504
+ "epoch": 1239.11,
296505
+ "learning_rate": 7.513990306946689e-06,
296506
+ "loss": 0.3375,
296507
+ "step": 154945
296508
+ },
296509
+ {
296510
+ "epoch": 1239.15,
296511
+ "learning_rate": 7.513909531502424e-06,
296512
+ "loss": 0.3088,
296513
+ "step": 154950
296514
+ },
296515
+ {
296516
+ "epoch": 1239.19,
296517
+ "learning_rate": 7.5138287560581595e-06,
296518
+ "loss": 0.7962,
296519
+ "step": 154955
296520
+ },
296521
+ {
296522
+ "epoch": 1239.23,
296523
+ "learning_rate": 7.513747980613894e-06,
296524
+ "loss": 0.6496,
296525
+ "step": 154960
296526
+ },
296527
+ {
296528
+ "epoch": 1239.27,
296529
+ "learning_rate": 7.5136672051696295e-06,
296530
+ "loss": 0.3379,
296531
+ "step": 154965
296532
+ },
296533
+ {
296534
+ "epoch": 1239.31,
296535
+ "learning_rate": 7.513586429725364e-06,
296536
+ "loss": 0.2975,
296537
+ "step": 154970
296538
+ },
296539
+ {
296540
+ "epoch": 1239.35,
296541
+ "learning_rate": 7.5135056542810994e-06,
296542
+ "loss": 0.3822,
296543
+ "step": 154975
296544
+ },
296545
+ {
296546
+ "epoch": 1239.39,
296547
+ "learning_rate": 7.513424878836834e-06,
296548
+ "loss": 0.9315,
296549
+ "step": 154980
296550
+ },
296551
+ {
296552
+ "epoch": 1239.43,
296553
+ "learning_rate": 7.513344103392569e-06,
296554
+ "loss": 0.6303,
296555
+ "step": 154985
296556
+ },
296557
+ {
296558
+ "epoch": 1239.47,
296559
+ "learning_rate": 7.513263327948304e-06,
296560
+ "loss": 0.3489,
296561
+ "step": 154990
296562
+ },
296563
+ {
296564
+ "epoch": 1239.51,
296565
+ "learning_rate": 7.513182552504039e-06,
296566
+ "loss": 0.3152,
296567
+ "step": 154995
296568
+ },
296569
+ {
296570
+ "epoch": 1239.55,
296571
+ "learning_rate": 7.513101777059774e-06,
296572
+ "loss": 0.3954,
296573
+ "step": 155000
296574
+ },
296575
+ {
296576
+ "epoch": 1239.59,
296577
+ "learning_rate": 7.513021001615509e-06,
296578
+ "loss": 0.878,
296579
+ "step": 155005
296580
+ },
296581
+ {
296582
+ "epoch": 1239.63,
296583
+ "learning_rate": 7.512940226171244e-06,
296584
+ "loss": 0.627,
296585
+ "step": 155010
296586
+ },
296587
+ {
296588
+ "epoch": 1239.67,
296589
+ "learning_rate": 7.512859450726979e-06,
296590
+ "loss": 0.2331,
296591
+ "step": 155015
296592
+ },
296593
+ {
296594
+ "epoch": 1239.71,
296595
+ "learning_rate": 7.512778675282715e-06,
296596
+ "loss": 0.2644,
296597
+ "step": 155020
296598
+ },
296599
+ {
296600
+ "epoch": 1239.75,
296601
+ "learning_rate": 7.512697899838449e-06,
296602
+ "loss": 0.3715,
296603
+ "step": 155025
296604
+ },
296605
+ {
296606
+ "epoch": 1239.79,
296607
+ "learning_rate": 7.512617124394185e-06,
296608
+ "loss": 0.9216,
296609
+ "step": 155030
296610
+ },
296611
+ {
296612
+ "epoch": 1239.83,
296613
+ "learning_rate": 7.51253634894992e-06,
296614
+ "loss": 0.6866,
296615
+ "step": 155035
296616
+ },
296617
+ {
296618
+ "epoch": 1239.87,
296619
+ "learning_rate": 7.512455573505655e-06,
296620
+ "loss": 0.2765,
296621
+ "step": 155040
296622
+ },
296623
+ {
296624
+ "epoch": 1239.91,
296625
+ "learning_rate": 7.51237479806139e-06,
296626
+ "loss": 0.2844,
296627
+ "step": 155045
296628
+ },
296629
+ {
296630
+ "epoch": 1239.95,
296631
+ "learning_rate": 7.512294022617125e-06,
296632
+ "loss": 0.3856,
296633
+ "step": 155050
296634
+ },
296635
+ {
296636
+ "epoch": 1239.99,
296637
+ "learning_rate": 7.51221324717286e-06,
296638
+ "loss": 1.0202,
296639
+ "step": 155055
296640
+ },
296641
+ {
296642
+ "epoch": 1240.0,
296643
+ "eval_loss": 0.30646491050720215,
296644
+ "eval_runtime": 40.3619,
296645
+ "eval_samples_per_second": 20.936,
296646
+ "eval_steps_per_second": 0.669,
296647
+ "eval_wer": 0.17656983568075119,
296648
+ "step": 155056
296649
+ },
296650
+ {
296651
+ "epoch": 1250.03,
296652
+ "learning_rate": 7.512132471728595e-06,
296653
+ "loss": 0.2716,
296654
+ "step": 155060
296655
+ },
296656
+ {
296657
+ "epoch": 1250.07,
296658
+ "learning_rate": 7.51205169628433e-06,
296659
+ "loss": 0.2815,
296660
+ "step": 155065
296661
+ },
296662
+ {
296663
+ "epoch": 1250.11,
296664
+ "learning_rate": 7.511970920840065e-06,
296665
+ "loss": 0.3168,
296666
+ "step": 155070
296667
+ },
296668
+ {
296669
+ "epoch": 1250.15,
296670
+ "learning_rate": 7.5118901453958e-06,
296671
+ "loss": 0.3727,
296672
+ "step": 155075
296673
+ },
296674
+ {
296675
+ "epoch": 1250.19,
296676
+ "learning_rate": 7.511809369951535e-06,
296677
+ "loss": 0.9729,
296678
+ "step": 155080
296679
+ },
296680
+ {
296681
+ "epoch": 1250.23,
296682
+ "learning_rate": 7.511728594507271e-06,
296683
+ "loss": 0.7132,
296684
+ "step": 155085
296685
+ },
296686
+ {
296687
+ "epoch": 1250.27,
296688
+ "learning_rate": 7.511647819063006e-06,
296689
+ "loss": 0.3155,
296690
+ "step": 155090
296691
+ },
296692
+ {
296693
+ "epoch": 1250.31,
296694
+ "learning_rate": 7.511567043618741e-06,
296695
+ "loss": 0.3005,
296696
+ "step": 155095
296697
+ },
296698
+ {
296699
+ "epoch": 1250.35,
296700
+ "learning_rate": 7.511486268174476e-06,
296701
+ "loss": 0.3456,
296702
+ "step": 155100
296703
+ },
296704
+ {
296705
+ "epoch": 1250.39,
296706
+ "learning_rate": 7.511405492730211e-06,
296707
+ "loss": 0.7762,
296708
+ "step": 155105
296709
+ },
296710
+ {
296711
+ "epoch": 1250.43,
296712
+ "learning_rate": 7.511324717285946e-06,
296713
+ "loss": 0.6201,
296714
+ "step": 155110
296715
+ },
296716
+ {
296717
+ "epoch": 1250.47,
296718
+ "learning_rate": 7.511243941841681e-06,
296719
+ "loss": 0.2818,
296720
+ "step": 155115
296721
+ },
296722
+ {
296723
+ "epoch": 1250.51,
296724
+ "learning_rate": 7.511163166397416e-06,
296725
+ "loss": 0.2924,
296726
+ "step": 155120
296727
+ },
296728
+ {
296729
+ "epoch": 1250.55,
296730
+ "learning_rate": 7.511082390953151e-06,
296731
+ "loss": 0.3261,
296732
+ "step": 155125
296733
+ },
296734
+ {
296735
+ "epoch": 1250.59,
296736
+ "learning_rate": 7.511001615508886e-06,
296737
+ "loss": 0.8687,
296738
+ "step": 155130
296739
+ },
296740
+ {
296741
+ "epoch": 1250.63,
296742
+ "learning_rate": 7.510920840064621e-06,
296743
+ "loss": 0.5005,
296744
+ "step": 155135
296745
+ },
296746
+ {
296747
+ "epoch": 1250.67,
296748
+ "learning_rate": 7.510840064620356e-06,
296749
+ "loss": 0.2698,
296750
+ "step": 155140
296751
+ },
296752
+ {
296753
+ "epoch": 1250.71,
296754
+ "learning_rate": 7.510759289176091e-06,
296755
+ "loss": 0.2844,
296756
+ "step": 155145
296757
+ },
296758
+ {
296759
+ "epoch": 1250.76,
296760
+ "learning_rate": 7.510678513731826e-06,
296761
+ "loss": 0.3468,
296762
+ "step": 155150
296763
+ },
296764
+ {
296765
+ "epoch": 1250.8,
296766
+ "learning_rate": 7.5105977382875615e-06,
296767
+ "loss": 0.9883,
296768
+ "step": 155155
296769
+ },
296770
+ {
296771
+ "epoch": 1250.84,
296772
+ "learning_rate": 7.5105169628432965e-06,
296773
+ "loss": 0.6705,
296774
+ "step": 155160
296775
+ },
296776
+ {
296777
+ "epoch": 1250.88,
296778
+ "learning_rate": 7.5104361873990315e-06,
296779
+ "loss": 0.2834,
296780
+ "step": 155165
296781
+ },
296782
+ {
296783
+ "epoch": 1250.92,
296784
+ "learning_rate": 7.5103554119547665e-06,
296785
+ "loss": 0.3121,
296786
+ "step": 155170
296787
+ },
296788
+ {
296789
+ "epoch": 1250.96,
296790
+ "learning_rate": 7.5102746365105015e-06,
296791
+ "loss": 0.4157,
296792
+ "step": 155175
296793
+ },
296794
+ {
296795
+ "epoch": 1251.0,
296796
+ "learning_rate": 7.5101938610662365e-06,
296797
+ "loss": 0.9646,
296798
+ "step": 155180
296799
+ },
296800
+ {
296801
+ "epoch": 1251.0,
296802
+ "eval_loss": 0.518424391746521,
296803
+ "eval_runtime": 41.1845,
296804
+ "eval_samples_per_second": 20.517,
296805
+ "eval_steps_per_second": 0.656,
296806
+ "eval_wer": 0.18026912388474478,
296807
+ "step": 155180
296808
  }
296809
  ],
296810
  "max_steps": 620000,
296811
  "num_train_epochs": 5000,
296812
+ "total_flos": 4.366899653515725e+20,
296813
  "trial_name": null,
296814
  "trial_params": null
296815
  }
model-bin/finetune/base/{checkpoint-154557 β†’ checkpoint-155180}/training_args.bin RENAMED
File without changes
model-bin/finetune/base/log/1630259719.897293/events.out.tfevents.1630259719.cc93b136ebf5.1086.249 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b026cb9d789aece99a3e446b51eef6d6eafe7463ff3affb3acdff973c9cecafb
3
+ size 4194
model-bin/finetune/base/log/1630260157.7758684/events.out.tfevents.1630260157.cc93b136ebf5.1086.251 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb971f3865430bebfa1d58eda62f42d75ec70030bc30ab5b02fcf85757e2658f
3
+ size 4194
model-bin/finetune/base/log/1630260591.4709191/events.out.tfevents.1630260591.cc93b136ebf5.1086.253 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a1abfdef8f5e60f550201561885525bc0f08b046f43b14dd16bd319ab43f679
3
+ size 4194
model-bin/finetune/base/log/1630261026.7725165/events.out.tfevents.1630261026.cc93b136ebf5.1086.255 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e4d315fde2bca0802d4b45f0f764ad61cbd51cadef039b9a244dc44c1e01520
3
+ size 4194
model-bin/finetune/base/log/1630261460.190783/events.out.tfevents.1630261460.cc93b136ebf5.1086.257 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:620ac2c75253b61fa00bf008d66e61a212b0eeb51edec85242ba1b84ca784dd7
3
+ size 4194
model-bin/finetune/base/log/events.out.tfevents.1630259719.cc93b136ebf5.1086.248 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8e93d8cc698e64dca2ac9e57763f5c78feb18455eabfac9c1e2580edc192eb2
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630260157.cc93b136ebf5.1086.250 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb210509e463781ed894247a0beaf3c1bbd3f02168ab0b5313909da570cfed24
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630260591.cc93b136ebf5.1086.252 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1cac3e5992df5b67b83acf2fc649f479413bf5f72958cbdefca5ba2ed0761bc1
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630261026.cc93b136ebf5.1086.254 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42bd2ba8b85647d66baebed9e3e1cff725f407db4785030a2cf15ecbc6bdb6ca
3
+ size 8622
model-bin/finetune/base/log/events.out.tfevents.1630261460.cc93b136ebf5.1086.256 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:541f0e49276ec7aa96edb6c944e0410f0b0b1cd347b64a03ba4fdbd5c0b0d642
3
+ size 8622