|
{ |
|
"best_metric": 2.074415922164917, |
|
"best_model_checkpoint": "ai-light-dance_drums_ft_pretrain_wav2vec2-base/checkpoint-744", |
|
"epoch": 199.90140845070422, |
|
"global_step": 1600, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.9, |
|
"eval_loss": 2.379915952682495, |
|
"eval_runtime": 7.7959, |
|
"eval_samples_per_second": 8.081, |
|
"eval_steps_per_second": 1.026, |
|
"eval_wer": 0.7797066746483089, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.0002, |
|
"loss": 0.9677, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"eval_loss": 2.4062252044677734, |
|
"eval_runtime": 7.9931, |
|
"eval_samples_per_second": 7.882, |
|
"eval_steps_per_second": 1.001, |
|
"eval_wer": 0.7913798263992816, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 0.0004, |
|
"loss": 0.9433, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"eval_loss": 2.6954824924468994, |
|
"eval_runtime": 8.0202, |
|
"eval_samples_per_second": 7.855, |
|
"eval_steps_per_second": 0.997, |
|
"eval_wer": 0.7683328344806944, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 0.00039746835443037974, |
|
"loss": 1.0276, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"eval_loss": 2.5257680416107178, |
|
"eval_runtime": 8.0107, |
|
"eval_samples_per_second": 7.864, |
|
"eval_steps_per_second": 0.999, |
|
"eval_wer": 0.7892846453157737, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"learning_rate": 0.0003949367088607595, |
|
"loss": 0.9923, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"eval_loss": 2.417393445968628, |
|
"eval_runtime": 8.1407, |
|
"eval_samples_per_second": 7.739, |
|
"eval_steps_per_second": 0.983, |
|
"eval_wer": 0.7824004788985334, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 5.9, |
|
"eval_loss": 2.7146248817443848, |
|
"eval_runtime": 8.0327, |
|
"eval_samples_per_second": 7.843, |
|
"eval_steps_per_second": 0.996, |
|
"eval_wer": 0.7788087398982341, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 6.23, |
|
"learning_rate": 0.0003924050632911393, |
|
"loss": 1.0927, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 6.9, |
|
"eval_loss": 2.9690921306610107, |
|
"eval_runtime": 7.8332, |
|
"eval_samples_per_second": 8.043, |
|
"eval_steps_per_second": 1.021, |
|
"eval_wer": 0.7883867105656989, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 7.45, |
|
"learning_rate": 0.000389873417721519, |
|
"loss": 1.0133, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 7.9, |
|
"eval_loss": 3.11049485206604, |
|
"eval_runtime": 8.0174, |
|
"eval_samples_per_second": 7.858, |
|
"eval_steps_per_second": 0.998, |
|
"eval_wer": 0.7644417838970368, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 8.68, |
|
"learning_rate": 0.00038734177215189877, |
|
"loss": 0.98, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 8.9, |
|
"eval_loss": 2.8638393878936768, |
|
"eval_runtime": 8.0331, |
|
"eval_samples_per_second": 7.843, |
|
"eval_steps_per_second": 0.996, |
|
"eval_wer": 0.7644417838970368, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 9.9, |
|
"learning_rate": 0.0003848101265822785, |
|
"loss": 0.9688, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 9.9, |
|
"eval_loss": 2.7538065910339355, |
|
"eval_runtime": 7.9304, |
|
"eval_samples_per_second": 7.944, |
|
"eval_steps_per_second": 1.009, |
|
"eval_wer": 0.758156240646513, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 10.9, |
|
"eval_loss": 2.6010308265686035, |
|
"eval_runtime": 8.0107, |
|
"eval_samples_per_second": 7.865, |
|
"eval_steps_per_second": 0.999, |
|
"eval_wer": 0.7713259503142772, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 11.23, |
|
"learning_rate": 0.00038227848101265825, |
|
"loss": 1.0453, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 11.9, |
|
"eval_loss": 2.6365110874176025, |
|
"eval_runtime": 7.9124, |
|
"eval_samples_per_second": 7.962, |
|
"eval_steps_per_second": 1.011, |
|
"eval_wer": 0.7482789583956899, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 12.45, |
|
"learning_rate": 0.00037974683544303797, |
|
"loss": 0.9397, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 12.9, |
|
"eval_loss": 2.474029302597046, |
|
"eval_runtime": 7.927, |
|
"eval_samples_per_second": 7.948, |
|
"eval_steps_per_second": 1.009, |
|
"eval_wer": 0.7629452259802454, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 13.68, |
|
"learning_rate": 0.00037721518987341774, |
|
"loss": 0.9466, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 13.9, |
|
"eval_loss": 2.6466243267059326, |
|
"eval_runtime": 7.9192, |
|
"eval_samples_per_second": 7.955, |
|
"eval_steps_per_second": 1.01, |
|
"eval_wer": 0.7617479796468123, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 14.9, |
|
"learning_rate": 0.00037468354430379746, |
|
"loss": 0.9665, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 14.9, |
|
"eval_loss": 2.6643269062042236, |
|
"eval_runtime": 7.9716, |
|
"eval_samples_per_second": 7.903, |
|
"eval_steps_per_second": 1.004, |
|
"eval_wer": 0.7644417838970368, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 15.9, |
|
"eval_loss": 2.4003283977508545, |
|
"eval_runtime": 7.9753, |
|
"eval_samples_per_second": 7.899, |
|
"eval_steps_per_second": 1.003, |
|
"eval_wer": 0.740796168811733, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 16.23, |
|
"learning_rate": 0.00037215189873417723, |
|
"loss": 1.0577, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 16.9, |
|
"eval_loss": 2.6306591033935547, |
|
"eval_runtime": 7.9695, |
|
"eval_samples_per_second": 7.905, |
|
"eval_steps_per_second": 1.004, |
|
"eval_wer": 0.7707273271475606, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 17.45, |
|
"learning_rate": 0.000369620253164557, |
|
"loss": 0.9154, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 17.9, |
|
"eval_loss": 2.6734743118286133, |
|
"eval_runtime": 7.957, |
|
"eval_samples_per_second": 7.918, |
|
"eval_steps_per_second": 1.005, |
|
"eval_wer": 0.7491768931457647, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 18.68, |
|
"learning_rate": 0.0003670886075949367, |
|
"loss": 0.9352, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 18.9, |
|
"eval_loss": 2.4403460025787354, |
|
"eval_runtime": 7.9548, |
|
"eval_samples_per_second": 7.92, |
|
"eval_steps_per_second": 1.006, |
|
"eval_wer": 0.7491768931457647, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 19.9, |
|
"learning_rate": 0.0003645569620253165, |
|
"loss": 0.8999, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 19.9, |
|
"eval_loss": 2.71948504447937, |
|
"eval_runtime": 7.9309, |
|
"eval_samples_per_second": 7.944, |
|
"eval_steps_per_second": 1.009, |
|
"eval_wer": 0.7650404070637533, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 20.9, |
|
"eval_loss": 2.2872610092163086, |
|
"eval_runtime": 7.9835, |
|
"eval_samples_per_second": 7.891, |
|
"eval_steps_per_second": 1.002, |
|
"eval_wer": 0.7602514217300209, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 21.23, |
|
"learning_rate": 0.00036202531645569626, |
|
"loss": 0.9935, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 21.9, |
|
"eval_loss": 2.844008207321167, |
|
"eval_runtime": 7.9178, |
|
"eval_samples_per_second": 7.957, |
|
"eval_steps_per_second": 1.01, |
|
"eval_wer": 0.7512720742292727, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 22.45, |
|
"learning_rate": 0.000359493670886076, |
|
"loss": 0.885, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 22.9, |
|
"eval_loss": 2.7574303150177, |
|
"eval_runtime": 7.9538, |
|
"eval_samples_per_second": 7.921, |
|
"eval_steps_per_second": 1.006, |
|
"eval_wer": 0.7485782699790482, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 23.68, |
|
"learning_rate": 0.0003569620253164557, |
|
"loss": 0.8979, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 23.9, |
|
"eval_loss": 3.1470837593078613, |
|
"eval_runtime": 7.9957, |
|
"eval_samples_per_second": 7.879, |
|
"eval_steps_per_second": 1.001, |
|
"eval_wer": 0.7536665668961389, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 24.9, |
|
"learning_rate": 0.00035443037974683546, |
|
"loss": 0.9074, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 24.9, |
|
"eval_loss": 3.089177131652832, |
|
"eval_runtime": 7.9191, |
|
"eval_samples_per_second": 7.955, |
|
"eval_steps_per_second": 1.01, |
|
"eval_wer": 0.7288237054774019, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 25.9, |
|
"eval_loss": 2.8640966415405273, |
|
"eval_runtime": 7.907, |
|
"eval_samples_per_second": 7.968, |
|
"eval_steps_per_second": 1.012, |
|
"eval_wer": 0.7518706973959892, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 26.23, |
|
"learning_rate": 0.00035189873417721524, |
|
"loss": 0.9498, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 26.9, |
|
"eval_loss": 2.879786729812622, |
|
"eval_runtime": 7.8826, |
|
"eval_samples_per_second": 7.992, |
|
"eval_steps_per_second": 1.015, |
|
"eval_wer": 0.7398982340616582, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 27.45, |
|
"learning_rate": 0.00034936708860759495, |
|
"loss": 0.8821, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 27.9, |
|
"eval_loss": 2.8114750385284424, |
|
"eval_runtime": 7.9479, |
|
"eval_samples_per_second": 7.927, |
|
"eval_steps_per_second": 1.007, |
|
"eval_wer": 0.749476204729123, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 28.68, |
|
"learning_rate": 0.00034683544303797467, |
|
"loss": 0.8968, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 28.9, |
|
"eval_loss": 2.922487735748291, |
|
"eval_runtime": 7.8659, |
|
"eval_samples_per_second": 8.009, |
|
"eval_steps_per_second": 1.017, |
|
"eval_wer": 0.7446872193953906, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 29.9, |
|
"learning_rate": 0.00034430379746835444, |
|
"loss": 0.8783, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 29.9, |
|
"eval_loss": 2.541382074356079, |
|
"eval_runtime": 8.0089, |
|
"eval_samples_per_second": 7.866, |
|
"eval_steps_per_second": 0.999, |
|
"eval_wer": 0.7503741394791978, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 30.9, |
|
"eval_loss": 2.352771759033203, |
|
"eval_runtime": 7.9156, |
|
"eval_samples_per_second": 7.959, |
|
"eval_steps_per_second": 1.011, |
|
"eval_wer": 0.7378030529781503, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 31.23, |
|
"learning_rate": 0.0003417721518987342, |
|
"loss": 0.9428, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 31.9, |
|
"eval_loss": 2.807304859161377, |
|
"eval_runtime": 7.8598, |
|
"eval_samples_per_second": 8.015, |
|
"eval_steps_per_second": 1.018, |
|
"eval_wer": 0.7141574378928465, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 32.45, |
|
"learning_rate": 0.00033924050632911393, |
|
"loss": 0.8184, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 32.9, |
|
"eval_loss": 2.575679063796997, |
|
"eval_runtime": 8.0158, |
|
"eval_samples_per_second": 7.859, |
|
"eval_steps_per_second": 0.998, |
|
"eval_wer": 0.7192457348099371, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 33.68, |
|
"learning_rate": 0.0003367088607594937, |
|
"loss": 0.9092, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 33.9, |
|
"eval_loss": 2.4403328895568848, |
|
"eval_runtime": 7.9646, |
|
"eval_samples_per_second": 7.91, |
|
"eval_steps_per_second": 1.004, |
|
"eval_wer": 0.7093684525591141, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 34.9, |
|
"learning_rate": 0.0003341772151898734, |
|
"loss": 0.8749, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 34.9, |
|
"eval_loss": 2.691164493560791, |
|
"eval_runtime": 7.9176, |
|
"eval_samples_per_second": 7.957, |
|
"eval_steps_per_second": 1.01, |
|
"eval_wer": 0.7219395390601616, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 35.9, |
|
"eval_loss": 2.4073445796966553, |
|
"eval_runtime": 8.0386, |
|
"eval_samples_per_second": 7.837, |
|
"eval_steps_per_second": 0.995, |
|
"eval_wer": 0.7327147560610595, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 36.23, |
|
"learning_rate": 0.0003316455696202532, |
|
"loss": 0.9235, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 36.9, |
|
"eval_loss": 2.444603681564331, |
|
"eval_runtime": 7.9142, |
|
"eval_samples_per_second": 7.96, |
|
"eval_steps_per_second": 1.011, |
|
"eval_wer": 0.726728524393894, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 37.45, |
|
"learning_rate": 0.00032911392405063296, |
|
"loss": 0.8654, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 37.9, |
|
"eval_loss": 2.821065902709961, |
|
"eval_runtime": 7.9214, |
|
"eval_samples_per_second": 7.953, |
|
"eval_steps_per_second": 1.01, |
|
"eval_wer": 0.7360071834780006, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 38.68, |
|
"learning_rate": 0.0003265822784810127, |
|
"loss": 0.8428, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 38.9, |
|
"eval_loss": 2.481106996536255, |
|
"eval_runtime": 7.9459, |
|
"eval_samples_per_second": 7.929, |
|
"eval_steps_per_second": 1.007, |
|
"eval_wer": 0.7243340317270278, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 39.9, |
|
"learning_rate": 0.0003240506329113924, |
|
"loss": 0.8355, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 39.9, |
|
"eval_loss": 2.357530355453491, |
|
"eval_runtime": 8.0775, |
|
"eval_samples_per_second": 7.799, |
|
"eval_steps_per_second": 0.99, |
|
"eval_wer": 0.7192457348099371, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 40.9, |
|
"eval_loss": 2.3957395553588867, |
|
"eval_runtime": 7.951, |
|
"eval_samples_per_second": 7.924, |
|
"eval_steps_per_second": 1.006, |
|
"eval_wer": 0.7138581263094882, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 41.23, |
|
"learning_rate": 0.00032151898734177216, |
|
"loss": 0.8992, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 41.9, |
|
"eval_loss": 2.4372870922088623, |
|
"eval_runtime": 7.962, |
|
"eval_samples_per_second": 7.913, |
|
"eval_steps_per_second": 1.005, |
|
"eval_wer": 0.7138581263094882, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 42.45, |
|
"learning_rate": 0.00031898734177215193, |
|
"loss": 0.8221, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 42.9, |
|
"eval_loss": 2.4234611988067627, |
|
"eval_runtime": 7.963, |
|
"eval_samples_per_second": 7.912, |
|
"eval_steps_per_second": 1.005, |
|
"eval_wer": 0.7126608799760551, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 43.68, |
|
"learning_rate": 0.00031645569620253165, |
|
"loss": 0.8305, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 43.9, |
|
"eval_loss": 2.3404624462127686, |
|
"eval_runtime": 7.9486, |
|
"eval_samples_per_second": 7.926, |
|
"eval_steps_per_second": 1.006, |
|
"eval_wer": 0.7111643220592637, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 44.9, |
|
"learning_rate": 0.00031392405063291137, |
|
"loss": 0.8328, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 44.9, |
|
"eval_loss": 2.340630292892456, |
|
"eval_runtime": 8.0091, |
|
"eval_samples_per_second": 7.866, |
|
"eval_steps_per_second": 0.999, |
|
"eval_wer": 0.7216402274768033, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 45.9, |
|
"eval_loss": 2.4469380378723145, |
|
"eval_runtime": 7.9319, |
|
"eval_samples_per_second": 7.943, |
|
"eval_steps_per_second": 1.009, |
|
"eval_wer": 0.7165519305597127, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 46.23, |
|
"learning_rate": 0.0003113924050632912, |
|
"loss": 0.8611, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 46.9, |
|
"eval_loss": 2.429659366607666, |
|
"eval_runtime": 7.8903, |
|
"eval_samples_per_second": 7.984, |
|
"eval_steps_per_second": 1.014, |
|
"eval_wer": 0.7156539958096378, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 47.45, |
|
"learning_rate": 0.0003088607594936709, |
|
"loss": 0.8092, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 47.9, |
|
"eval_loss": 2.5867514610290527, |
|
"eval_runtime": 8.0153, |
|
"eval_samples_per_second": 7.86, |
|
"eval_steps_per_second": 0.998, |
|
"eval_wer": 0.7093684525591141, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 48.68, |
|
"learning_rate": 0.00030632911392405063, |
|
"loss": 0.8173, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 48.9, |
|
"eval_loss": 2.2557618618011475, |
|
"eval_runtime": 8.0103, |
|
"eval_samples_per_second": 7.865, |
|
"eval_steps_per_second": 0.999, |
|
"eval_wer": 0.7003891050583657, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 49.9, |
|
"learning_rate": 0.0003037974683544304, |
|
"loss": 0.7772, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 49.9, |
|
"eval_loss": 2.359837770462036, |
|
"eval_runtime": 7.7486, |
|
"eval_samples_per_second": 8.13, |
|
"eval_steps_per_second": 1.032, |
|
"eval_wer": 0.7003891050583657, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 50.9, |
|
"eval_loss": 2.308309316635132, |
|
"eval_runtime": 7.7464, |
|
"eval_samples_per_second": 8.133, |
|
"eval_steps_per_second": 1.033, |
|
"eval_wer": 0.6881173301406764, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 51.23, |
|
"learning_rate": 0.00030126582278481017, |
|
"loss": 0.8494, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 51.9, |
|
"eval_loss": 2.4430971145629883, |
|
"eval_runtime": 7.7316, |
|
"eval_samples_per_second": 8.148, |
|
"eval_steps_per_second": 1.035, |
|
"eval_wer": 0.7012870398084405, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 52.45, |
|
"learning_rate": 0.0002987341772151899, |
|
"loss": 0.7997, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 52.9, |
|
"eval_loss": 2.3005003929138184, |
|
"eval_runtime": 7.7273, |
|
"eval_samples_per_second": 8.153, |
|
"eval_steps_per_second": 1.035, |
|
"eval_wer": 0.7111643220592637, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 53.68, |
|
"learning_rate": 0.0002962025316455696, |
|
"loss": 0.7879, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 53.9, |
|
"eval_loss": 2.1985087394714355, |
|
"eval_runtime": 7.8272, |
|
"eval_samples_per_second": 8.049, |
|
"eval_steps_per_second": 1.022, |
|
"eval_wer": 0.7297216402274768, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 54.9, |
|
"learning_rate": 0.0002936708860759494, |
|
"loss": 0.7694, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 54.9, |
|
"eval_loss": 2.337602376937866, |
|
"eval_runtime": 7.7096, |
|
"eval_samples_per_second": 8.172, |
|
"eval_steps_per_second": 1.038, |
|
"eval_wer": 0.708171206225681, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 55.9, |
|
"eval_loss": 2.371609926223755, |
|
"eval_runtime": 7.7243, |
|
"eval_samples_per_second": 8.156, |
|
"eval_steps_per_second": 1.036, |
|
"eval_wer": 0.7012870398084405, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 56.23, |
|
"learning_rate": 0.00029113924050632915, |
|
"loss": 0.8397, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 56.9, |
|
"eval_loss": 2.381723403930664, |
|
"eval_runtime": 7.7601, |
|
"eval_samples_per_second": 8.118, |
|
"eval_steps_per_second": 1.031, |
|
"eval_wer": 0.711463633642622, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 57.45, |
|
"learning_rate": 0.00028860759493670886, |
|
"loss": 0.7868, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 57.9, |
|
"eval_loss": 2.257676601409912, |
|
"eval_runtime": 7.7658, |
|
"eval_samples_per_second": 8.112, |
|
"eval_steps_per_second": 1.03, |
|
"eval_wer": 0.7090691409757558, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 58.68, |
|
"learning_rate": 0.00028607594936708863, |
|
"loss": 0.7311, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 58.9, |
|
"eval_loss": 2.3895084857940674, |
|
"eval_runtime": 7.7486, |
|
"eval_samples_per_second": 8.13, |
|
"eval_steps_per_second": 1.032, |
|
"eval_wer": 0.7126608799760551, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 59.9, |
|
"learning_rate": 0.00028354430379746835, |
|
"loss": 0.7796, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 59.9, |
|
"eval_loss": 2.276035785675049, |
|
"eval_runtime": 7.7565, |
|
"eval_samples_per_second": 8.122, |
|
"eval_steps_per_second": 1.031, |
|
"eval_wer": 0.7099670757258306, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 60.9, |
|
"eval_loss": 2.5685431957244873, |
|
"eval_runtime": 7.7725, |
|
"eval_samples_per_second": 8.105, |
|
"eval_steps_per_second": 1.029, |
|
"eval_wer": 0.7072732714756061, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 61.23, |
|
"learning_rate": 0.0002810126582278481, |
|
"loss": 0.8272, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 61.9, |
|
"eval_loss": 2.3881263732910156, |
|
"eval_runtime": 7.7155, |
|
"eval_samples_per_second": 8.165, |
|
"eval_steps_per_second": 1.037, |
|
"eval_wer": 0.7027835977252319, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 62.45, |
|
"learning_rate": 0.0002784810126582279, |
|
"loss": 0.7639, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 62.9, |
|
"eval_loss": 2.3456814289093018, |
|
"eval_runtime": 7.7642, |
|
"eval_samples_per_second": 8.114, |
|
"eval_steps_per_second": 1.03, |
|
"eval_wer": 0.7084705178090392, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 63.68, |
|
"learning_rate": 0.0002759493670886076, |
|
"loss": 0.789, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 63.9, |
|
"eval_loss": 2.3291287422180176, |
|
"eval_runtime": 7.7437, |
|
"eval_samples_per_second": 8.136, |
|
"eval_steps_per_second": 1.033, |
|
"eval_wer": 0.700688416641724, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 64.9, |
|
"learning_rate": 0.0002734177215189873, |
|
"loss": 0.7472, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 64.9, |
|
"eval_loss": 2.5174083709716797, |
|
"eval_runtime": 7.7856, |
|
"eval_samples_per_second": 8.092, |
|
"eval_steps_per_second": 1.028, |
|
"eval_wer": 0.70487877880874, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 65.9, |
|
"eval_loss": 2.399650812149048, |
|
"eval_runtime": 7.7605, |
|
"eval_samples_per_second": 8.118, |
|
"eval_steps_per_second": 1.031, |
|
"eval_wer": 0.7153546842262796, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 66.23, |
|
"learning_rate": 0.0002708860759493671, |
|
"loss": 0.8056, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 66.9, |
|
"eval_loss": 2.4573962688446045, |
|
"eval_runtime": 7.7606, |
|
"eval_samples_per_second": 8.118, |
|
"eval_steps_per_second": 1.031, |
|
"eval_wer": 0.7237354085603113, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 67.45, |
|
"learning_rate": 0.00026835443037974687, |
|
"loss": 0.7752, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 67.9, |
|
"eval_loss": 2.4979681968688965, |
|
"eval_runtime": 7.7515, |
|
"eval_samples_per_second": 8.127, |
|
"eval_steps_per_second": 1.032, |
|
"eval_wer": 0.7003891050583657, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 68.68, |
|
"learning_rate": 0.0002658227848101266, |
|
"loss": 0.7084, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 68.9, |
|
"eval_loss": 2.2370431423187256, |
|
"eval_runtime": 7.792, |
|
"eval_samples_per_second": 8.085, |
|
"eval_steps_per_second": 1.027, |
|
"eval_wer": 0.7084705178090392, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 69.9, |
|
"learning_rate": 0.0002632911392405063, |
|
"loss": 0.7824, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 69.9, |
|
"eval_loss": 2.359494924545288, |
|
"eval_runtime": 7.722, |
|
"eval_samples_per_second": 8.159, |
|
"eval_steps_per_second": 1.036, |
|
"eval_wer": 0.6970966776414247, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 70.9, |
|
"eval_loss": 2.19962477684021, |
|
"eval_runtime": 7.7269, |
|
"eval_samples_per_second": 8.153, |
|
"eval_steps_per_second": 1.035, |
|
"eval_wer": 0.7003891050583657, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 71.23, |
|
"learning_rate": 0.00026075949367088613, |
|
"loss": 0.7776, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 71.9, |
|
"eval_loss": 2.2957143783569336, |
|
"eval_runtime": 7.7186, |
|
"eval_samples_per_second": 8.162, |
|
"eval_steps_per_second": 1.036, |
|
"eval_wer": 0.6902125112241844, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 72.45, |
|
"learning_rate": 0.00025822784810126584, |
|
"loss": 0.7205, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 72.9, |
|
"eval_loss": 2.2435786724090576, |
|
"eval_runtime": 7.7517, |
|
"eval_samples_per_second": 8.127, |
|
"eval_steps_per_second": 1.032, |
|
"eval_wer": 0.6908111343909009, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 73.68, |
|
"learning_rate": 0.00025569620253164556, |
|
"loss": 0.7074, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 73.9, |
|
"eval_loss": 2.23608136177063, |
|
"eval_runtime": 7.7615, |
|
"eval_samples_per_second": 8.117, |
|
"eval_steps_per_second": 1.031, |
|
"eval_wer": 0.6932056270577671, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 74.9, |
|
"learning_rate": 0.00025316455696202533, |
|
"loss": 0.7237, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 74.9, |
|
"eval_loss": 2.207817792892456, |
|
"eval_runtime": 7.6966, |
|
"eval_samples_per_second": 8.185, |
|
"eval_steps_per_second": 1.039, |
|
"eval_wer": 0.6857228374738102, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 75.9, |
|
"eval_loss": 2.233380079269409, |
|
"eval_runtime": 7.7786, |
|
"eval_samples_per_second": 8.099, |
|
"eval_steps_per_second": 1.028, |
|
"eval_wer": 0.6905118228075426, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 76.23, |
|
"learning_rate": 0.0002506329113924051, |
|
"loss": 0.7862, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 76.9, |
|
"eval_loss": 2.3564865589141846, |
|
"eval_runtime": 7.7597, |
|
"eval_samples_per_second": 8.119, |
|
"eval_steps_per_second": 1.031, |
|
"eval_wer": 0.6976953008081412, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 77.45, |
|
"learning_rate": 0.0002481012658227848, |
|
"loss": 0.7299, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 77.9, |
|
"eval_loss": 2.1293139457702637, |
|
"eval_runtime": 7.6982, |
|
"eval_samples_per_second": 8.184, |
|
"eval_steps_per_second": 1.039, |
|
"eval_wer": 0.6779407363064951, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 78.68, |
|
"learning_rate": 0.0002455696202531646, |
|
"loss": 0.6755, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 78.9, |
|
"eval_loss": 2.2523531913757324, |
|
"eval_runtime": 7.7601, |
|
"eval_samples_per_second": 8.118, |
|
"eval_steps_per_second": 1.031, |
|
"eval_wer": 0.6860221490571685, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 79.9, |
|
"learning_rate": 0.00024303797468354434, |
|
"loss": 0.724, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 79.9, |
|
"eval_loss": 2.206925630569458, |
|
"eval_runtime": 7.7271, |
|
"eval_samples_per_second": 8.153, |
|
"eval_steps_per_second": 1.035, |
|
"eval_wer": 0.688715953307393, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 80.9, |
|
"eval_loss": 2.5267446041107178, |
|
"eval_runtime": 7.7853, |
|
"eval_samples_per_second": 8.092, |
|
"eval_steps_per_second": 1.028, |
|
"eval_wer": 0.6785393594732116, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 81.23, |
|
"learning_rate": 0.00024050632911392405, |
|
"loss": 0.7878, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 81.9, |
|
"eval_loss": 2.6393752098083496, |
|
"eval_runtime": 7.709, |
|
"eval_samples_per_second": 8.172, |
|
"eval_steps_per_second": 1.038, |
|
"eval_wer": 0.6824304100568692, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 82.45, |
|
"learning_rate": 0.0002379746835443038, |
|
"loss": 0.6882, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 82.9, |
|
"eval_loss": 2.4647722244262695, |
|
"eval_runtime": 7.8339, |
|
"eval_samples_per_second": 8.042, |
|
"eval_steps_per_second": 1.021, |
|
"eval_wer": 0.6764441783897037, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 83.68, |
|
"learning_rate": 0.00023544303797468357, |
|
"loss": 0.6996, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 83.9, |
|
"eval_loss": 2.4115612506866455, |
|
"eval_runtime": 7.7276, |
|
"eval_samples_per_second": 8.153, |
|
"eval_steps_per_second": 1.035, |
|
"eval_wer": 0.6890152648907513, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 84.9, |
|
"learning_rate": 0.0002329113924050633, |
|
"loss": 0.7149, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 84.9, |
|
"eval_loss": 2.1044137477874756, |
|
"eval_runtime": 7.7519, |
|
"eval_samples_per_second": 8.127, |
|
"eval_steps_per_second": 1.032, |
|
"eval_wer": 0.6893145764741095, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 85.9, |
|
"eval_loss": 2.1447622776031494, |
|
"eval_runtime": 7.7429, |
|
"eval_samples_per_second": 8.136, |
|
"eval_steps_per_second": 1.033, |
|
"eval_wer": 0.6917090691409757, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 86.23, |
|
"learning_rate": 0.00023037974683544303, |
|
"loss": 0.7499, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 86.9, |
|
"eval_loss": 2.2603471279144287, |
|
"eval_runtime": 7.7748, |
|
"eval_samples_per_second": 8.103, |
|
"eval_steps_per_second": 1.029, |
|
"eval_wer": 0.6875187069739599, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 87.45, |
|
"learning_rate": 0.0002278481012658228, |
|
"loss": 0.6881, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 87.9, |
|
"eval_loss": 2.1306064128875732, |
|
"eval_runtime": 7.7761, |
|
"eval_samples_per_second": 8.102, |
|
"eval_steps_per_second": 1.029, |
|
"eval_wer": 0.6815324753067944, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 88.68, |
|
"learning_rate": 0.00022531645569620254, |
|
"loss": 0.6652, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 88.9, |
|
"eval_loss": 2.195241689682007, |
|
"eval_runtime": 7.7296, |
|
"eval_samples_per_second": 8.151, |
|
"eval_steps_per_second": 1.035, |
|
"eval_wer": 0.6905118228075426, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 89.9, |
|
"learning_rate": 0.0002227848101265823, |
|
"loss": 0.7093, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 89.9, |
|
"eval_loss": 2.3550162315368652, |
|
"eval_runtime": 7.7395, |
|
"eval_samples_per_second": 8.14, |
|
"eval_steps_per_second": 1.034, |
|
"eval_wer": 0.676743489973062, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 90.9, |
|
"eval_loss": 2.261043071746826, |
|
"eval_runtime": 7.7396, |
|
"eval_samples_per_second": 8.14, |
|
"eval_steps_per_second": 1.034, |
|
"eval_wer": 0.6749476204729123, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 91.23, |
|
"learning_rate": 0.00022025316455696206, |
|
"loss": 0.7439, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 91.9, |
|
"eval_loss": 2.147209882736206, |
|
"eval_runtime": 7.7312, |
|
"eval_samples_per_second": 8.149, |
|
"eval_steps_per_second": 1.035, |
|
"eval_wer": 0.6857228374738102, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 92.45, |
|
"learning_rate": 0.00021772151898734177, |
|
"loss": 0.6898, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 92.9, |
|
"eval_loss": 2.074415922164917, |
|
"eval_runtime": 7.7637, |
|
"eval_samples_per_second": 8.115, |
|
"eval_steps_per_second": 1.03, |
|
"eval_wer": 0.6881173301406764, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 93.68, |
|
"learning_rate": 0.00021518987341772152, |
|
"loss": 0.6734, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 93.9, |
|
"eval_loss": 2.0897610187530518, |
|
"eval_runtime": 7.7107, |
|
"eval_samples_per_second": 8.17, |
|
"eval_steps_per_second": 1.038, |
|
"eval_wer": 0.6929063154744088, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 94.9, |
|
"learning_rate": 0.00021265822784810126, |
|
"loss": 0.6926, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 94.9, |
|
"eval_loss": 2.089553117752075, |
|
"eval_runtime": 7.721, |
|
"eval_samples_per_second": 8.16, |
|
"eval_steps_per_second": 1.036, |
|
"eval_wer": 0.6683627656390302, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 95.9, |
|
"eval_loss": 2.1928741931915283, |
|
"eval_runtime": 7.7808, |
|
"eval_samples_per_second": 8.097, |
|
"eval_steps_per_second": 1.028, |
|
"eval_wer": 0.6812331637234361, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 96.23, |
|
"learning_rate": 0.00021012658227848103, |
|
"loss": 0.7154, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 96.9, |
|
"eval_loss": 2.153787612915039, |
|
"eval_runtime": 7.7478, |
|
"eval_samples_per_second": 8.131, |
|
"eval_steps_per_second": 1.033, |
|
"eval_wer": 0.6860221490571685, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 97.45, |
|
"learning_rate": 0.00020759493670886078, |
|
"loss": 0.6493, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 97.9, |
|
"eval_loss": 2.143815279006958, |
|
"eval_runtime": 7.7358, |
|
"eval_samples_per_second": 8.144, |
|
"eval_steps_per_second": 1.034, |
|
"eval_wer": 0.6815324753067944, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 98.68, |
|
"learning_rate": 0.0002050632911392405, |
|
"loss": 0.6755, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 98.9, |
|
"eval_loss": 2.1560962200164795, |
|
"eval_runtime": 7.7208, |
|
"eval_samples_per_second": 8.16, |
|
"eval_steps_per_second": 1.036, |
|
"eval_wer": 0.6902125112241844, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 99.9, |
|
"learning_rate": 0.00020253164556962027, |
|
"loss": 0.6667, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 99.9, |
|
"eval_loss": 2.076709032058716, |
|
"eval_runtime": 7.7317, |
|
"eval_samples_per_second": 8.148, |
|
"eval_steps_per_second": 1.035, |
|
"eval_wer": 0.6908111343909009, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 100.9, |
|
"eval_loss": 2.106370449066162, |
|
"eval_runtime": 7.7472, |
|
"eval_samples_per_second": 8.132, |
|
"eval_steps_per_second": 1.033, |
|
"eval_wer": 0.6785393594732116, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 101.23, |
|
"learning_rate": 0.0002, |
|
"loss": 0.7016, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 101.9, |
|
"eval_loss": 2.227837085723877, |
|
"eval_runtime": 8.0181, |
|
"eval_samples_per_second": 7.857, |
|
"eval_steps_per_second": 0.998, |
|
"eval_wer": 0.676743489973062, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 102.45, |
|
"learning_rate": 0.00019746835443037975, |
|
"loss": 0.6726, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 102.9, |
|
"eval_loss": 2.261566162109375, |
|
"eval_runtime": 7.7943, |
|
"eval_samples_per_second": 8.083, |
|
"eval_steps_per_second": 1.026, |
|
"eval_wer": 0.6689613888057467, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 103.68, |
|
"learning_rate": 0.0001949367088607595, |
|
"loss": 0.6725, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 103.9, |
|
"eval_loss": 2.133070230484009, |
|
"eval_runtime": 7.7265, |
|
"eval_samples_per_second": 8.154, |
|
"eval_steps_per_second": 1.035, |
|
"eval_wer": 0.6878180185573182, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 104.9, |
|
"learning_rate": 0.00019240506329113924, |
|
"loss": 0.6657, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 104.9, |
|
"eval_loss": 2.1497416496276855, |
|
"eval_runtime": 7.7253, |
|
"eval_samples_per_second": 8.155, |
|
"eval_steps_per_second": 1.036, |
|
"eval_wer": 0.6731517509727627, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 105.9, |
|
"eval_loss": 2.160142421722412, |
|
"eval_runtime": 7.7477, |
|
"eval_samples_per_second": 8.131, |
|
"eval_steps_per_second": 1.033, |
|
"eval_wer": 0.6737503741394792, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 106.23, |
|
"learning_rate": 0.00018987341772151899, |
|
"loss": 0.6989, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 106.9, |
|
"eval_loss": 2.319141149520874, |
|
"eval_runtime": 7.732, |
|
"eval_samples_per_second": 8.148, |
|
"eval_steps_per_second": 1.035, |
|
"eval_wer": 0.6674648308889554, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 107.45, |
|
"learning_rate": 0.00018734177215189873, |
|
"loss": 0.6658, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 107.9, |
|
"eval_loss": 2.354734420776367, |
|
"eval_runtime": 7.7273, |
|
"eval_samples_per_second": 8.153, |
|
"eval_steps_per_second": 1.035, |
|
"eval_wer": 0.6788386710565699, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 108.68, |
|
"learning_rate": 0.0001848101265822785, |
|
"loss": 0.6398, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 108.9, |
|
"eval_loss": 2.3368043899536133, |
|
"eval_runtime": 7.7551, |
|
"eval_samples_per_second": 8.124, |
|
"eval_steps_per_second": 1.032, |
|
"eval_wer": 0.6740496857228375, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 109.9, |
|
"learning_rate": 0.00018227848101265824, |
|
"loss": 0.6465, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 109.9, |
|
"eval_loss": 2.1896259784698486, |
|
"eval_runtime": 7.7364, |
|
"eval_samples_per_second": 8.143, |
|
"eval_steps_per_second": 1.034, |
|
"eval_wer": 0.6806345405567196, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 110.9, |
|
"eval_loss": 2.1210110187530518, |
|
"eval_runtime": 7.7217, |
|
"eval_samples_per_second": 8.159, |
|
"eval_steps_per_second": 1.036, |
|
"eval_wer": 0.6797366058066447, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 111.23, |
|
"learning_rate": 0.000179746835443038, |
|
"loss": 0.727, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 111.9, |
|
"eval_loss": 2.3508195877075195, |
|
"eval_runtime": 7.7774, |
|
"eval_samples_per_second": 8.1, |
|
"eval_steps_per_second": 1.029, |
|
"eval_wer": 0.6686620772223885, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 112.45, |
|
"learning_rate": 0.00017721518987341773, |
|
"loss": 0.6409, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 112.9, |
|
"eval_loss": 2.3439958095550537, |
|
"eval_runtime": 7.7324, |
|
"eval_samples_per_second": 8.148, |
|
"eval_steps_per_second": 1.035, |
|
"eval_wer": 0.6752469320562706, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 113.68, |
|
"learning_rate": 0.00017468354430379748, |
|
"loss": 0.6573, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 113.9, |
|
"eval_loss": 2.269505500793457, |
|
"eval_runtime": 7.7658, |
|
"eval_samples_per_second": 8.112, |
|
"eval_steps_per_second": 1.03, |
|
"eval_wer": 0.6623765339718647, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 114.9, |
|
"learning_rate": 0.00017215189873417722, |
|
"loss": 0.645, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 114.9, |
|
"eval_loss": 2.1471199989318848, |
|
"eval_runtime": 7.7501, |
|
"eval_samples_per_second": 8.129, |
|
"eval_steps_per_second": 1.032, |
|
"eval_wer": 0.6770428015564203, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 115.9, |
|
"eval_loss": 2.186685562133789, |
|
"eval_runtime": 7.7952, |
|
"eval_samples_per_second": 8.082, |
|
"eval_steps_per_second": 1.026, |
|
"eval_wer": 0.6743489973061958, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 116.23, |
|
"learning_rate": 0.00016962025316455696, |
|
"loss": 0.7103, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 116.9, |
|
"eval_loss": 2.232990026473999, |
|
"eval_runtime": 7.7463, |
|
"eval_samples_per_second": 8.133, |
|
"eval_steps_per_second": 1.033, |
|
"eval_wer": 0.6701586351391798, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 117.45, |
|
"learning_rate": 0.0001670886075949367, |
|
"loss": 0.6214, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 117.9, |
|
"eval_loss": 2.217428207397461, |
|
"eval_runtime": 7.7992, |
|
"eval_samples_per_second": 8.078, |
|
"eval_steps_per_second": 1.026, |
|
"eval_wer": 0.6686620772223885, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 118.68, |
|
"learning_rate": 0.00016455696202531648, |
|
"loss": 0.6134, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 118.9, |
|
"eval_loss": 2.198005437850952, |
|
"eval_runtime": 7.8436, |
|
"eval_samples_per_second": 8.032, |
|
"eval_steps_per_second": 1.02, |
|
"eval_wer": 0.6620772223885064, |
|
"step": 952 |
|
}, |
|
{ |
|
"epoch": 119.9, |
|
"learning_rate": 0.0001620253164556962, |
|
"loss": 0.6612, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 119.9, |
|
"eval_loss": 2.2890524864196777, |
|
"eval_runtime": 7.8296, |
|
"eval_samples_per_second": 8.046, |
|
"eval_steps_per_second": 1.022, |
|
"eval_wer": 0.6749476204729123, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 120.9, |
|
"eval_loss": 2.2862656116485596, |
|
"eval_runtime": 7.7538, |
|
"eval_samples_per_second": 8.125, |
|
"eval_steps_per_second": 1.032, |
|
"eval_wer": 0.6641724034720143, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 121.23, |
|
"learning_rate": 0.00015949367088607597, |
|
"loss": 0.688, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 121.9, |
|
"eval_loss": 2.319775342941284, |
|
"eval_runtime": 7.7964, |
|
"eval_samples_per_second": 8.081, |
|
"eval_steps_per_second": 1.026, |
|
"eval_wer": 0.6668662077222388, |
|
"step": 976 |
|
}, |
|
{ |
|
"epoch": 122.45, |
|
"learning_rate": 0.00015696202531645568, |
|
"loss": 0.6451, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 122.9, |
|
"eval_loss": 2.169647216796875, |
|
"eval_runtime": 7.8494, |
|
"eval_samples_per_second": 8.026, |
|
"eval_steps_per_second": 1.019, |
|
"eval_wer": 0.6668662077222388, |
|
"step": 984 |
|
}, |
|
{ |
|
"epoch": 123.68, |
|
"learning_rate": 0.00015443037974683546, |
|
"loss": 0.6308, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 123.9, |
|
"eval_loss": 2.125494956970215, |
|
"eval_runtime": 7.8606, |
|
"eval_samples_per_second": 8.015, |
|
"eval_steps_per_second": 1.018, |
|
"eval_wer": 0.6596827297216402, |
|
"step": 992 |
|
}, |
|
{ |
|
"epoch": 124.9, |
|
"learning_rate": 0.0001518987341772152, |
|
"loss": 0.6359, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 124.9, |
|
"eval_loss": 2.2053353786468506, |
|
"eval_runtime": 7.7903, |
|
"eval_samples_per_second": 8.087, |
|
"eval_steps_per_second": 1.027, |
|
"eval_wer": 0.6569889254714157, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 125.9, |
|
"eval_loss": 2.1914987564086914, |
|
"eval_runtime": 7.7966, |
|
"eval_samples_per_second": 8.08, |
|
"eval_steps_per_second": 1.026, |
|
"eval_wer": 0.6581861718048488, |
|
"step": 1008 |
|
}, |
|
{ |
|
"epoch": 126.23, |
|
"learning_rate": 0.00014936708860759494, |
|
"loss": 0.6845, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 126.9, |
|
"eval_loss": 2.1405885219573975, |
|
"eval_runtime": 7.7604, |
|
"eval_samples_per_second": 8.118, |
|
"eval_steps_per_second": 1.031, |
|
"eval_wer": 0.6656689613888057, |
|
"step": 1016 |
|
}, |
|
{ |
|
"epoch": 127.45, |
|
"learning_rate": 0.0001468354430379747, |
|
"loss": 0.6609, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 127.9, |
|
"eval_loss": 2.1851718425750732, |
|
"eval_runtime": 7.745, |
|
"eval_samples_per_second": 8.134, |
|
"eval_steps_per_second": 1.033, |
|
"eval_wer": 0.6752469320562706, |
|
"step": 1024 |
|
}, |
|
{ |
|
"epoch": 128.68, |
|
"learning_rate": 0.00014430379746835443, |
|
"loss": 0.6345, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 128.9, |
|
"eval_loss": 2.1838414669036865, |
|
"eval_runtime": 7.7504, |
|
"eval_samples_per_second": 8.129, |
|
"eval_steps_per_second": 1.032, |
|
"eval_wer": 0.6620772223885064, |
|
"step": 1032 |
|
}, |
|
{ |
|
"epoch": 129.9, |
|
"learning_rate": 0.00014177215189873418, |
|
"loss": 0.6055, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 129.9, |
|
"eval_loss": 2.158566474914551, |
|
"eval_runtime": 7.7809, |
|
"eval_samples_per_second": 8.097, |
|
"eval_steps_per_second": 1.028, |
|
"eval_wer": 0.6701586351391798, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 130.9, |
|
"eval_loss": 2.1627261638641357, |
|
"eval_runtime": 7.7707, |
|
"eval_samples_per_second": 8.107, |
|
"eval_steps_per_second": 1.03, |
|
"eval_wer": 0.6680634540556719, |
|
"step": 1048 |
|
}, |
|
{ |
|
"epoch": 131.23, |
|
"learning_rate": 0.00013924050632911395, |
|
"loss": 0.6737, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 131.9, |
|
"eval_loss": 2.263141632080078, |
|
"eval_runtime": 7.8874, |
|
"eval_samples_per_second": 7.987, |
|
"eval_steps_per_second": 1.014, |
|
"eval_wer": 0.6761448668063454, |
|
"step": 1056 |
|
}, |
|
{ |
|
"epoch": 132.45, |
|
"learning_rate": 0.00013670886075949366, |
|
"loss": 0.6237, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 132.9, |
|
"eval_loss": 2.2553627490997314, |
|
"eval_runtime": 7.755, |
|
"eval_samples_per_second": 8.124, |
|
"eval_steps_per_second": 1.032, |
|
"eval_wer": 0.6620772223885064, |
|
"step": 1064 |
|
}, |
|
{ |
|
"epoch": 133.68, |
|
"learning_rate": 0.00013417721518987343, |
|
"loss": 0.6468, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 133.9, |
|
"eval_loss": 2.2538866996765137, |
|
"eval_runtime": 7.7947, |
|
"eval_samples_per_second": 8.082, |
|
"eval_steps_per_second": 1.026, |
|
"eval_wer": 0.6668662077222388, |
|
"step": 1072 |
|
}, |
|
{ |
|
"epoch": 134.9, |
|
"learning_rate": 0.00013164556962025315, |
|
"loss": 0.5948, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 134.9, |
|
"eval_loss": 2.2463722229003906, |
|
"eval_runtime": 7.7617, |
|
"eval_samples_per_second": 8.117, |
|
"eval_steps_per_second": 1.031, |
|
"eval_wer": 0.6516013169709668, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 135.9, |
|
"eval_loss": 2.349128484725952, |
|
"eval_runtime": 7.7593, |
|
"eval_samples_per_second": 8.119, |
|
"eval_steps_per_second": 1.031, |
|
"eval_wer": 0.6620772223885064, |
|
"step": 1088 |
|
}, |
|
{ |
|
"epoch": 136.23, |
|
"learning_rate": 0.00012911392405063292, |
|
"loss": 0.6645, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 136.9, |
|
"eval_loss": 2.2536532878875732, |
|
"eval_runtime": 7.8036, |
|
"eval_samples_per_second": 8.073, |
|
"eval_steps_per_second": 1.025, |
|
"eval_wer": 0.6620772223885064, |
|
"step": 1096 |
|
}, |
|
{ |
|
"epoch": 137.45, |
|
"learning_rate": 0.00012658227848101267, |
|
"loss": 0.6195, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 137.9, |
|
"eval_loss": 2.371696710586548, |
|
"eval_runtime": 7.7793, |
|
"eval_samples_per_second": 8.098, |
|
"eval_steps_per_second": 1.028, |
|
"eval_wer": 0.6665668961388805, |
|
"step": 1104 |
|
}, |
|
{ |
|
"epoch": 138.68, |
|
"learning_rate": 0.0001240506329113924, |
|
"loss": 0.6317, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 138.9, |
|
"eval_loss": 2.2024903297424316, |
|
"eval_runtime": 7.8765, |
|
"eval_samples_per_second": 7.998, |
|
"eval_steps_per_second": 1.016, |
|
"eval_wer": 0.6551930559712661, |
|
"step": 1112 |
|
}, |
|
{ |
|
"epoch": 139.9, |
|
"learning_rate": 0.00012151898734177217, |
|
"loss": 0.6336, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 139.9, |
|
"eval_loss": 2.142206907272339, |
|
"eval_runtime": 7.8378, |
|
"eval_samples_per_second": 8.038, |
|
"eval_steps_per_second": 1.021, |
|
"eval_wer": 0.6623765339718647, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 140.9, |
|
"eval_loss": 2.106228828430176, |
|
"eval_runtime": 7.7968, |
|
"eval_samples_per_second": 8.08, |
|
"eval_steps_per_second": 1.026, |
|
"eval_wer": 0.660580664471715, |
|
"step": 1128 |
|
}, |
|
{ |
|
"epoch": 141.23, |
|
"learning_rate": 0.0001189873417721519, |
|
"loss": 0.664, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 141.9, |
|
"eval_loss": 2.2253739833831787, |
|
"eval_runtime": 7.7492, |
|
"eval_samples_per_second": 8.13, |
|
"eval_steps_per_second": 1.032, |
|
"eval_wer": 0.6596827297216402, |
|
"step": 1136 |
|
}, |
|
{ |
|
"epoch": 142.45, |
|
"learning_rate": 0.00011645569620253166, |
|
"loss": 0.6047, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 142.9, |
|
"eval_loss": 2.322563886642456, |
|
"eval_runtime": 7.7874, |
|
"eval_samples_per_second": 8.09, |
|
"eval_steps_per_second": 1.027, |
|
"eval_wer": 0.653995809637833, |
|
"step": 1144 |
|
}, |
|
{ |
|
"epoch": 143.68, |
|
"learning_rate": 0.0001139240506329114, |
|
"loss": 0.6173, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 143.9, |
|
"eval_loss": 2.227854013442993, |
|
"eval_runtime": 7.8279, |
|
"eval_samples_per_second": 8.048, |
|
"eval_steps_per_second": 1.022, |
|
"eval_wer": 0.6683627656390302, |
|
"step": 1152 |
|
}, |
|
{ |
|
"epoch": 144.9, |
|
"learning_rate": 0.00011139240506329114, |
|
"loss": 0.6466, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 144.9, |
|
"eval_loss": 2.1866044998168945, |
|
"eval_runtime": 7.7744, |
|
"eval_samples_per_second": 8.104, |
|
"eval_steps_per_second": 1.029, |
|
"eval_wer": 0.657288237054774, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 145.9, |
|
"eval_loss": 2.2489023208618164, |
|
"eval_runtime": 7.7715, |
|
"eval_samples_per_second": 8.107, |
|
"eval_steps_per_second": 1.029, |
|
"eval_wer": 0.6590841065549237, |
|
"step": 1168 |
|
}, |
|
{ |
|
"epoch": 146.23, |
|
"learning_rate": 0.00010886075949367089, |
|
"loss": 0.6585, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 146.9, |
|
"eval_loss": 2.2273900508880615, |
|
"eval_runtime": 7.821, |
|
"eval_samples_per_second": 8.055, |
|
"eval_steps_per_second": 1.023, |
|
"eval_wer": 0.6480095779706675, |
|
"step": 1176 |
|
}, |
|
{ |
|
"epoch": 147.45, |
|
"learning_rate": 0.00010632911392405063, |
|
"loss": 0.6244, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 147.9, |
|
"eval_loss": 2.1959400177001953, |
|
"eval_runtime": 7.8697, |
|
"eval_samples_per_second": 8.005, |
|
"eval_steps_per_second": 1.017, |
|
"eval_wer": 0.662675845555223, |
|
"step": 1184 |
|
}, |
|
{ |
|
"epoch": 148.68, |
|
"learning_rate": 0.00010379746835443039, |
|
"loss": 0.6527, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 148.9, |
|
"eval_loss": 2.2115025520324707, |
|
"eval_runtime": 7.8828, |
|
"eval_samples_per_second": 7.992, |
|
"eval_steps_per_second": 1.015, |
|
"eval_wer": 0.6593834181382819, |
|
"step": 1192 |
|
}, |
|
{ |
|
"epoch": 149.9, |
|
"learning_rate": 0.00010126582278481013, |
|
"loss": 0.6247, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 149.9, |
|
"eval_loss": 2.2804923057556152, |
|
"eval_runtime": 7.7687, |
|
"eval_samples_per_second": 8.109, |
|
"eval_steps_per_second": 1.03, |
|
"eval_wer": 0.6620772223885064, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 150.9, |
|
"eval_loss": 2.2128920555114746, |
|
"eval_runtime": 7.7584, |
|
"eval_samples_per_second": 8.12, |
|
"eval_steps_per_second": 1.031, |
|
"eval_wer": 0.6578868602214906, |
|
"step": 1208 |
|
}, |
|
{ |
|
"epoch": 151.23, |
|
"learning_rate": 9.873417721518988e-05, |
|
"loss": 0.6614, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 151.9, |
|
"eval_loss": 2.23846697807312, |
|
"eval_runtime": 7.7588, |
|
"eval_samples_per_second": 8.12, |
|
"eval_steps_per_second": 1.031, |
|
"eval_wer": 0.6635737803052978, |
|
"step": 1216 |
|
}, |
|
{ |
|
"epoch": 152.45, |
|
"learning_rate": 9.620253164556962e-05, |
|
"loss": 0.6309, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 152.9, |
|
"eval_loss": 2.275683641433716, |
|
"eval_runtime": 7.8149, |
|
"eval_samples_per_second": 8.061, |
|
"eval_steps_per_second": 1.024, |
|
"eval_wer": 0.6614785992217899, |
|
"step": 1224 |
|
}, |
|
{ |
|
"epoch": 153.68, |
|
"learning_rate": 9.367088607594936e-05, |
|
"loss": 0.6501, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 153.9, |
|
"eval_loss": 2.3265960216522217, |
|
"eval_runtime": 7.8195, |
|
"eval_samples_per_second": 8.057, |
|
"eval_steps_per_second": 1.023, |
|
"eval_wer": 0.6647710266387309, |
|
"step": 1232 |
|
}, |
|
{ |
|
"epoch": 154.9, |
|
"learning_rate": 9.113924050632912e-05, |
|
"loss": 0.5869, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 154.9, |
|
"eval_loss": 2.336057662963867, |
|
"eval_runtime": 7.8329, |
|
"eval_samples_per_second": 8.043, |
|
"eval_steps_per_second": 1.021, |
|
"eval_wer": 0.6632744687219395, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 155.9, |
|
"eval_loss": 2.345245838165283, |
|
"eval_runtime": 7.7323, |
|
"eval_samples_per_second": 8.148, |
|
"eval_steps_per_second": 1.035, |
|
"eval_wer": 0.653995809637833, |
|
"step": 1248 |
|
}, |
|
{ |
|
"epoch": 156.23, |
|
"learning_rate": 8.860759493670887e-05, |
|
"loss": 0.6676, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 156.9, |
|
"eval_loss": 2.280003070831299, |
|
"eval_runtime": 7.8173, |
|
"eval_samples_per_second": 8.059, |
|
"eval_steps_per_second": 1.023, |
|
"eval_wer": 0.6614785992217899, |
|
"step": 1256 |
|
}, |
|
{ |
|
"epoch": 157.45, |
|
"learning_rate": 8.607594936708861e-05, |
|
"loss": 0.6494, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 157.9, |
|
"eval_loss": 2.305755376815796, |
|
"eval_runtime": 7.8527, |
|
"eval_samples_per_second": 8.023, |
|
"eval_steps_per_second": 1.019, |
|
"eval_wer": 0.6662675845555223, |
|
"step": 1264 |
|
}, |
|
{ |
|
"epoch": 158.68, |
|
"learning_rate": 8.354430379746835e-05, |
|
"loss": 0.6017, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 158.9, |
|
"eval_loss": 2.2905781269073486, |
|
"eval_runtime": 7.8309, |
|
"eval_samples_per_second": 8.045, |
|
"eval_steps_per_second": 1.022, |
|
"eval_wer": 0.6662675845555223, |
|
"step": 1272 |
|
}, |
|
{ |
|
"epoch": 159.9, |
|
"learning_rate": 8.10126582278481e-05, |
|
"loss": 0.6266, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 159.9, |
|
"eval_loss": 2.2315995693206787, |
|
"eval_runtime": 7.7807, |
|
"eval_samples_per_second": 8.097, |
|
"eval_steps_per_second": 1.028, |
|
"eval_wer": 0.6596827297216402, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 160.9, |
|
"eval_loss": 2.188622236251831, |
|
"eval_runtime": 7.8175, |
|
"eval_samples_per_second": 8.059, |
|
"eval_steps_per_second": 1.023, |
|
"eval_wer": 0.6710565698892547, |
|
"step": 1288 |
|
}, |
|
{ |
|
"epoch": 161.23, |
|
"learning_rate": 7.848101265822784e-05, |
|
"loss": 0.6704, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 161.9, |
|
"eval_loss": 2.3184380531311035, |
|
"eval_runtime": 7.8037, |
|
"eval_samples_per_second": 8.073, |
|
"eval_steps_per_second": 1.025, |
|
"eval_wer": 0.6590841065549237, |
|
"step": 1296 |
|
}, |
|
{ |
|
"epoch": 162.45, |
|
"learning_rate": 7.59493670886076e-05, |
|
"loss": 0.6239, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 162.9, |
|
"eval_loss": 2.3544297218322754, |
|
"eval_runtime": 7.7214, |
|
"eval_samples_per_second": 8.159, |
|
"eval_steps_per_second": 1.036, |
|
"eval_wer": 0.6617779108051481, |
|
"step": 1304 |
|
}, |
|
{ |
|
"epoch": 163.68, |
|
"learning_rate": 7.341772151898734e-05, |
|
"loss": 0.5997, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 163.9, |
|
"eval_loss": 2.298438310623169, |
|
"eval_runtime": 7.7258, |
|
"eval_samples_per_second": 8.155, |
|
"eval_steps_per_second": 1.035, |
|
"eval_wer": 0.6677641424723136, |
|
"step": 1312 |
|
}, |
|
{ |
|
"epoch": 164.9, |
|
"learning_rate": 7.088607594936709e-05, |
|
"loss": 0.6228, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 164.9, |
|
"eval_loss": 2.293006658554077, |
|
"eval_runtime": 7.7223, |
|
"eval_samples_per_second": 8.158, |
|
"eval_steps_per_second": 1.036, |
|
"eval_wer": 0.669260700389105, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 165.9, |
|
"eval_loss": 2.3272392749786377, |
|
"eval_runtime": 7.7917, |
|
"eval_samples_per_second": 8.086, |
|
"eval_steps_per_second": 1.027, |
|
"eval_wer": 0.6584854833882071, |
|
"step": 1328 |
|
}, |
|
{ |
|
"epoch": 166.23, |
|
"learning_rate": 6.835443037974683e-05, |
|
"loss": 0.6683, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 166.9, |
|
"eval_loss": 2.3456509113311768, |
|
"eval_runtime": 7.7579, |
|
"eval_samples_per_second": 8.121, |
|
"eval_steps_per_second": 1.031, |
|
"eval_wer": 0.657288237054774, |
|
"step": 1336 |
|
}, |
|
{ |
|
"epoch": 167.45, |
|
"learning_rate": 6.582278481012658e-05, |
|
"loss": 0.598, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 167.9, |
|
"eval_loss": 2.2177622318267822, |
|
"eval_runtime": 7.8139, |
|
"eval_samples_per_second": 8.063, |
|
"eval_steps_per_second": 1.024, |
|
"eval_wer": 0.6638730918886561, |
|
"step": 1344 |
|
}, |
|
{ |
|
"epoch": 168.68, |
|
"learning_rate": 6.329113924050633e-05, |
|
"loss": 0.6164, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 168.9, |
|
"eval_loss": 2.1438896656036377, |
|
"eval_runtime": 7.7254, |
|
"eval_samples_per_second": 8.155, |
|
"eval_steps_per_second": 1.036, |
|
"eval_wer": 0.6542951212211913, |
|
"step": 1352 |
|
}, |
|
{ |
|
"epoch": 169.9, |
|
"learning_rate": 6.0759493670886084e-05, |
|
"loss": 0.5963, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 169.9, |
|
"eval_loss": 2.1238651275634766, |
|
"eval_runtime": 7.8668, |
|
"eval_samples_per_second": 8.008, |
|
"eval_steps_per_second": 1.017, |
|
"eval_wer": 0.6513020053876085, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 170.9, |
|
"eval_loss": 2.1392319202423096, |
|
"eval_runtime": 7.7989, |
|
"eval_samples_per_second": 8.078, |
|
"eval_steps_per_second": 1.026, |
|
"eval_wer": 0.6593834181382819, |
|
"step": 1368 |
|
}, |
|
{ |
|
"epoch": 171.23, |
|
"learning_rate": 5.822784810126583e-05, |
|
"loss": 0.6782, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 171.9, |
|
"eval_loss": 2.129168748855591, |
|
"eval_runtime": 7.8187, |
|
"eval_samples_per_second": 8.058, |
|
"eval_steps_per_second": 1.023, |
|
"eval_wer": 0.6578868602214906, |
|
"step": 1376 |
|
}, |
|
{ |
|
"epoch": 172.45, |
|
"learning_rate": 5.569620253164557e-05, |
|
"loss": 0.5783, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 172.9, |
|
"eval_loss": 2.1256933212280273, |
|
"eval_runtime": 7.8785, |
|
"eval_samples_per_second": 7.996, |
|
"eval_steps_per_second": 1.015, |
|
"eval_wer": 0.6596827297216402, |
|
"step": 1384 |
|
}, |
|
{ |
|
"epoch": 173.68, |
|
"learning_rate": 5.3164556962025316e-05, |
|
"loss": 0.6087, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 173.9, |
|
"eval_loss": 2.125344753265381, |
|
"eval_runtime": 7.7825, |
|
"eval_samples_per_second": 8.095, |
|
"eval_steps_per_second": 1.028, |
|
"eval_wer": 0.6593834181382819, |
|
"step": 1392 |
|
}, |
|
{ |
|
"epoch": 174.9, |
|
"learning_rate": 5.0632911392405066e-05, |
|
"loss": 0.6045, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 174.9, |
|
"eval_loss": 2.133322238922119, |
|
"eval_runtime": 7.8569, |
|
"eval_samples_per_second": 8.018, |
|
"eval_steps_per_second": 1.018, |
|
"eval_wer": 0.6560909907213409, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 175.9, |
|
"eval_loss": 2.1041812896728516, |
|
"eval_runtime": 7.7837, |
|
"eval_samples_per_second": 8.094, |
|
"eval_steps_per_second": 1.028, |
|
"eval_wer": 0.650703382220892, |
|
"step": 1408 |
|
}, |
|
{ |
|
"epoch": 176.23, |
|
"learning_rate": 4.810126582278481e-05, |
|
"loss": 0.6299, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 176.9, |
|
"eval_loss": 2.111027956008911, |
|
"eval_runtime": 7.8074, |
|
"eval_samples_per_second": 8.069, |
|
"eval_steps_per_second": 1.025, |
|
"eval_wer": 0.6569889254714157, |
|
"step": 1416 |
|
}, |
|
{ |
|
"epoch": 177.45, |
|
"learning_rate": 4.556962025316456e-05, |
|
"loss": 0.6401, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 177.9, |
|
"eval_loss": 2.116133451461792, |
|
"eval_runtime": 7.7361, |
|
"eval_samples_per_second": 8.144, |
|
"eval_steps_per_second": 1.034, |
|
"eval_wer": 0.6611792876384316, |
|
"step": 1424 |
|
}, |
|
{ |
|
"epoch": 178.68, |
|
"learning_rate": 4.3037974683544305e-05, |
|
"loss": 0.622, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 178.9, |
|
"eval_loss": 2.168447494506836, |
|
"eval_runtime": 7.8421, |
|
"eval_samples_per_second": 8.034, |
|
"eval_steps_per_second": 1.02, |
|
"eval_wer": 0.6483088895540258, |
|
"step": 1432 |
|
}, |
|
{ |
|
"epoch": 179.9, |
|
"learning_rate": 4.050632911392405e-05, |
|
"loss": 0.599, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 179.9, |
|
"eval_loss": 2.190560817718506, |
|
"eval_runtime": 7.8024, |
|
"eval_samples_per_second": 8.074, |
|
"eval_steps_per_second": 1.025, |
|
"eval_wer": 0.6551930559712661, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 180.9, |
|
"eval_loss": 2.2257890701293945, |
|
"eval_runtime": 7.7768, |
|
"eval_samples_per_second": 8.101, |
|
"eval_steps_per_second": 1.029, |
|
"eval_wer": 0.6492068243041006, |
|
"step": 1448 |
|
}, |
|
{ |
|
"epoch": 181.23, |
|
"learning_rate": 3.79746835443038e-05, |
|
"loss": 0.6516, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 181.9, |
|
"eval_loss": 2.203824758529663, |
|
"eval_runtime": 7.8475, |
|
"eval_samples_per_second": 8.028, |
|
"eval_steps_per_second": 1.019, |
|
"eval_wer": 0.6536964980544747, |
|
"step": 1456 |
|
}, |
|
{ |
|
"epoch": 182.45, |
|
"learning_rate": 3.5443037974683544e-05, |
|
"loss": 0.5907, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 182.9, |
|
"eval_loss": 2.194875955581665, |
|
"eval_runtime": 7.95, |
|
"eval_samples_per_second": 7.925, |
|
"eval_steps_per_second": 1.006, |
|
"eval_wer": 0.6533971864711164, |
|
"step": 1464 |
|
}, |
|
{ |
|
"epoch": 183.68, |
|
"learning_rate": 3.291139240506329e-05, |
|
"loss": 0.5979, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 183.9, |
|
"eval_loss": 2.196150541305542, |
|
"eval_runtime": 7.8608, |
|
"eval_samples_per_second": 8.014, |
|
"eval_steps_per_second": 1.018, |
|
"eval_wer": 0.6530978748877582, |
|
"step": 1472 |
|
}, |
|
{ |
|
"epoch": 184.9, |
|
"learning_rate": 3.0379746835443042e-05, |
|
"loss": 0.6064, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 184.9, |
|
"eval_loss": 2.1942551136016846, |
|
"eval_runtime": 7.781, |
|
"eval_samples_per_second": 8.097, |
|
"eval_steps_per_second": 1.028, |
|
"eval_wer": 0.6498054474708171, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 185.9, |
|
"eval_loss": 2.1707868576049805, |
|
"eval_runtime": 7.9327, |
|
"eval_samples_per_second": 7.942, |
|
"eval_steps_per_second": 1.008, |
|
"eval_wer": 0.6524992517210416, |
|
"step": 1488 |
|
}, |
|
{ |
|
"epoch": 186.23, |
|
"learning_rate": 2.7848101265822786e-05, |
|
"loss": 0.6363, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 186.9, |
|
"eval_loss": 2.1659958362579346, |
|
"eval_runtime": 7.8392, |
|
"eval_samples_per_second": 8.037, |
|
"eval_steps_per_second": 1.021, |
|
"eval_wer": 0.6560909907213409, |
|
"step": 1496 |
|
}, |
|
{ |
|
"epoch": 187.45, |
|
"learning_rate": 2.5316455696202533e-05, |
|
"loss": 0.6257, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 187.9, |
|
"eval_loss": 2.174118757247925, |
|
"eval_runtime": 7.8051, |
|
"eval_samples_per_second": 8.072, |
|
"eval_steps_per_second": 1.025, |
|
"eval_wer": 0.657288237054774, |
|
"step": 1504 |
|
}, |
|
{ |
|
"epoch": 188.68, |
|
"learning_rate": 2.278481012658228e-05, |
|
"loss": 0.6128, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 188.9, |
|
"eval_loss": 2.1725566387176514, |
|
"eval_runtime": 7.7709, |
|
"eval_samples_per_second": 8.107, |
|
"eval_steps_per_second": 1.029, |
|
"eval_wer": 0.6563903023046992, |
|
"step": 1512 |
|
}, |
|
{ |
|
"epoch": 189.9, |
|
"learning_rate": 2.0253164556962025e-05, |
|
"loss": 0.602, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 189.9, |
|
"eval_loss": 2.168877124786377, |
|
"eval_runtime": 7.9172, |
|
"eval_samples_per_second": 7.957, |
|
"eval_steps_per_second": 1.01, |
|
"eval_wer": 0.6554923675546244, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 190.9, |
|
"eval_loss": 2.170177936553955, |
|
"eval_runtime": 7.8092, |
|
"eval_samples_per_second": 8.067, |
|
"eval_steps_per_second": 1.024, |
|
"eval_wer": 0.6566896138880575, |
|
"step": 1528 |
|
}, |
|
{ |
|
"epoch": 191.23, |
|
"learning_rate": 1.7721518987341772e-05, |
|
"loss": 0.645, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 191.9, |
|
"eval_loss": 2.1751959323883057, |
|
"eval_runtime": 7.9351, |
|
"eval_samples_per_second": 7.939, |
|
"eval_steps_per_second": 1.008, |
|
"eval_wer": 0.6590841065549237, |
|
"step": 1536 |
|
}, |
|
{ |
|
"epoch": 192.45, |
|
"learning_rate": 1.5189873417721521e-05, |
|
"loss": 0.5916, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 192.9, |
|
"eval_loss": 2.190694570541382, |
|
"eval_runtime": 7.7664, |
|
"eval_samples_per_second": 8.112, |
|
"eval_steps_per_second": 1.03, |
|
"eval_wer": 0.6560909907213409, |
|
"step": 1544 |
|
}, |
|
{ |
|
"epoch": 193.68, |
|
"learning_rate": 1.2658227848101267e-05, |
|
"loss": 0.5853, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 193.9, |
|
"eval_loss": 2.1865837574005127, |
|
"eval_runtime": 7.7908, |
|
"eval_samples_per_second": 8.086, |
|
"eval_steps_per_second": 1.027, |
|
"eval_wer": 0.6545944328045495, |
|
"step": 1552 |
|
}, |
|
{ |
|
"epoch": 194.9, |
|
"learning_rate": 1.0126582278481012e-05, |
|
"loss": 0.5735, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 194.9, |
|
"eval_loss": 2.1829705238342285, |
|
"eval_runtime": 7.8222, |
|
"eval_samples_per_second": 8.054, |
|
"eval_steps_per_second": 1.023, |
|
"eval_wer": 0.6554923675546244, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 195.9, |
|
"eval_loss": 2.176032781600952, |
|
"eval_runtime": 7.8048, |
|
"eval_samples_per_second": 8.072, |
|
"eval_steps_per_second": 1.025, |
|
"eval_wer": 0.6563903023046992, |
|
"step": 1568 |
|
}, |
|
{ |
|
"epoch": 196.23, |
|
"learning_rate": 7.5949367088607605e-06, |
|
"loss": 0.6294, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 196.9, |
|
"eval_loss": 2.1679303646087646, |
|
"eval_runtime": 7.925, |
|
"eval_samples_per_second": 7.949, |
|
"eval_steps_per_second": 1.009, |
|
"eval_wer": 0.6578868602214906, |
|
"step": 1576 |
|
}, |
|
{ |
|
"epoch": 197.45, |
|
"learning_rate": 5.063291139240506e-06, |
|
"loss": 0.6149, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 197.9, |
|
"eval_loss": 2.1631834506988525, |
|
"eval_runtime": 7.8984, |
|
"eval_samples_per_second": 7.976, |
|
"eval_steps_per_second": 1.013, |
|
"eval_wer": 0.6575875486381323, |
|
"step": 1584 |
|
}, |
|
{ |
|
"epoch": 198.68, |
|
"learning_rate": 2.531645569620253e-06, |
|
"loss": 0.5761, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 198.9, |
|
"eval_loss": 2.1613857746124268, |
|
"eval_runtime": 7.8089, |
|
"eval_samples_per_second": 8.068, |
|
"eval_steps_per_second": 1.024, |
|
"eval_wer": 0.657288237054774, |
|
"step": 1592 |
|
}, |
|
{ |
|
"epoch": 199.9, |
|
"learning_rate": 0.0, |
|
"loss": 0.6111, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 199.9, |
|
"eval_loss": 2.1611926555633545, |
|
"eval_runtime": 7.7531, |
|
"eval_samples_per_second": 8.126, |
|
"eval_steps_per_second": 1.032, |
|
"eval_wer": 0.6575875486381323, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 199.9, |
|
"step": 1600, |
|
"total_flos": 1.1848602213269387e+19, |
|
"train_loss": 0.7314322146773339, |
|
"train_runtime": 18544.7872, |
|
"train_samples_per_second": 6.072, |
|
"train_steps_per_second": 0.086 |
|
} |
|
], |
|
"max_steps": 1600, |
|
"num_train_epochs": 200, |
|
"total_flos": 1.1848602213269387e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|