|
{ |
|
"best_metric": 0.17350260416666666, |
|
"best_model_checkpoint": "/cluster/home/torstefl/Master/saved_model/W2V/single/NB/30.05/checkpoint-3344", |
|
"epoch": 40.0, |
|
"eval_steps": 500, |
|
"global_step": 3520, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 3.9170310497283936, |
|
"learning_rate": 5.8e-06, |
|
"loss": 0.9591, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_rundkast_loss": 0.5365053415298462, |
|
"eval_rundkast_runtime": 35.6704, |
|
"eval_rundkast_samples_per_second": 37.706, |
|
"eval_rundkast_steps_per_second": 1.205, |
|
"eval_rundkast_wer": 0.23027362392618517, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_nb_samtale_loss": 0.67843097448349, |
|
"eval_nb_samtale_runtime": 39.2213, |
|
"eval_nb_samtale_samples_per_second": 13.59, |
|
"eval_nb_samtale_steps_per_second": 0.433, |
|
"eval_nb_samtale_wer": 0.24666341145833334, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_bigbrother_loss": 2.74082612991333, |
|
"eval_bigbrother_runtime": 41.8079, |
|
"eval_bigbrother_samples_per_second": 32.769, |
|
"eval_bigbrother_steps_per_second": 1.029, |
|
"eval_bigbrother_wer": 0.6191089986304681, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 4.4212822914123535, |
|
"learning_rate": 1.1666666666666668e-05, |
|
"loss": 0.8123, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_rundkast_loss": 0.4252748191356659, |
|
"eval_rundkast_runtime": 33.6441, |
|
"eval_rundkast_samples_per_second": 39.977, |
|
"eval_rundkast_steps_per_second": 1.278, |
|
"eval_rundkast_wer": 0.2241489023226217, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_nb_samtale_loss": 0.5754267573356628, |
|
"eval_nb_samtale_runtime": 38.5834, |
|
"eval_nb_samtale_samples_per_second": 13.814, |
|
"eval_nb_samtale_steps_per_second": 0.441, |
|
"eval_nb_samtale_wer": 0.23527018229166666, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_bigbrother_loss": 2.240948438644409, |
|
"eval_bigbrother_runtime": 41.5488, |
|
"eval_bigbrother_samples_per_second": 32.973, |
|
"eval_bigbrother_steps_per_second": 1.035, |
|
"eval_bigbrother_wer": 0.6037219044550068, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 3.1233577728271484, |
|
"learning_rate": 1.7533333333333334e-05, |
|
"loss": 0.7517, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_rundkast_loss": 0.419950932264328, |
|
"eval_rundkast_runtime": 33.5339, |
|
"eval_rundkast_samples_per_second": 40.109, |
|
"eval_rundkast_steps_per_second": 1.282, |
|
"eval_rundkast_wer": 0.2205695195672924, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_nb_samtale_loss": 0.5531216263771057, |
|
"eval_nb_samtale_runtime": 38.5893, |
|
"eval_nb_samtale_samples_per_second": 13.812, |
|
"eval_nb_samtale_steps_per_second": 0.441, |
|
"eval_nb_samtale_wer": 0.22932942708333334, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_bigbrother_loss": 2.1344377994537354, |
|
"eval_bigbrother_runtime": 41.2938, |
|
"eval_bigbrother_samples_per_second": 33.177, |
|
"eval_bigbrother_steps_per_second": 1.041, |
|
"eval_bigbrother_wer": 0.5964714412309675, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 3.1834347248077393, |
|
"learning_rate": 2.3400000000000003e-05, |
|
"loss": 0.7074, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_rundkast_loss": 0.4088253676891327, |
|
"eval_rundkast_runtime": 33.3755, |
|
"eval_rundkast_samples_per_second": 40.299, |
|
"eval_rundkast_steps_per_second": 1.288, |
|
"eval_rundkast_wer": 0.22009226853324848, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_nb_samtale_loss": 0.5255401134490967, |
|
"eval_nb_samtale_runtime": 38.2361, |
|
"eval_nb_samtale_samples_per_second": 13.94, |
|
"eval_nb_samtale_steps_per_second": 0.445, |
|
"eval_nb_samtale_wer": 0.22696940104166666, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_bigbrother_loss": 2.036323070526123, |
|
"eval_bigbrother_runtime": 41.2821, |
|
"eval_bigbrother_samples_per_second": 33.186, |
|
"eval_bigbrother_steps_per_second": 1.042, |
|
"eval_bigbrother_wer": 0.5905099492467575, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"grad_norm": 1.5792173147201538, |
|
"learning_rate": 2.926666666666667e-05, |
|
"loss": 0.6874, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_rundkast_loss": 0.40116986632347107, |
|
"eval_rundkast_runtime": 33.3522, |
|
"eval_rundkast_samples_per_second": 40.327, |
|
"eval_rundkast_steps_per_second": 1.289, |
|
"eval_rundkast_wer": 0.21985364301622654, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_nb_samtale_loss": 0.5132110714912415, |
|
"eval_nb_samtale_runtime": 38.393, |
|
"eval_nb_samtale_samples_per_second": 13.883, |
|
"eval_nb_samtale_steps_per_second": 0.443, |
|
"eval_nb_samtale_wer": 0.224365234375, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_bigbrother_loss": 1.9978336095809937, |
|
"eval_bigbrother_runtime": 41.45, |
|
"eval_bigbrother_samples_per_second": 33.052, |
|
"eval_bigbrother_steps_per_second": 1.037, |
|
"eval_bigbrother_wer": 0.5905099492467575, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"grad_norm": 2.002620220184326, |
|
"learning_rate": 3.513333333333334e-05, |
|
"loss": 0.6548, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_rundkast_loss": 0.4032253921031952, |
|
"eval_rundkast_runtime": 33.4235, |
|
"eval_rundkast_samples_per_second": 40.241, |
|
"eval_rundkast_steps_per_second": 1.287, |
|
"eval_rundkast_wer": 0.217467387846007, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_nb_samtale_loss": 0.5047765374183655, |
|
"eval_nb_samtale_runtime": 38.2135, |
|
"eval_nb_samtale_samples_per_second": 13.948, |
|
"eval_nb_samtale_steps_per_second": 0.445, |
|
"eval_nb_samtale_wer": 0.220947265625, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_bigbrother_loss": 1.9670743942260742, |
|
"eval_bigbrother_runtime": 41.4322, |
|
"eval_bigbrother_samples_per_second": 33.066, |
|
"eval_bigbrother_steps_per_second": 1.038, |
|
"eval_bigbrother_wer": 0.5801176186256344, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"grad_norm": 1.4368079900741577, |
|
"learning_rate": 4.1e-05, |
|
"loss": 0.6439, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_rundkast_loss": 0.39594346284866333, |
|
"eval_rundkast_runtime": 33.337, |
|
"eval_rundkast_samples_per_second": 40.346, |
|
"eval_rundkast_steps_per_second": 1.29, |
|
"eval_rundkast_wer": 0.22009226853324848, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_nb_samtale_loss": 0.48273980617523193, |
|
"eval_nb_samtale_runtime": 38.0773, |
|
"eval_nb_samtale_samples_per_second": 13.998, |
|
"eval_nb_samtale_steps_per_second": 0.446, |
|
"eval_nb_samtale_wer": 0.21964518229166666, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_bigbrother_loss": 1.921434998512268, |
|
"eval_bigbrother_runtime": 41.5595, |
|
"eval_bigbrother_samples_per_second": 32.965, |
|
"eval_bigbrother_steps_per_second": 1.035, |
|
"eval_bigbrother_wer": 0.5865624748247805, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"grad_norm": 1.1675041913986206, |
|
"learning_rate": 4.686666666666667e-05, |
|
"loss": 0.6158, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_rundkast_loss": 0.40993189811706543, |
|
"eval_rundkast_runtime": 33.4667, |
|
"eval_rundkast_samples_per_second": 40.189, |
|
"eval_rundkast_steps_per_second": 1.285, |
|
"eval_rundkast_wer": 0.22128539611835826, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_nb_samtale_loss": 0.494111031293869, |
|
"eval_nb_samtale_runtime": 38.0186, |
|
"eval_nb_samtale_samples_per_second": 14.019, |
|
"eval_nb_samtale_steps_per_second": 0.447, |
|
"eval_nb_samtale_wer": 0.21712239583333334, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_bigbrother_loss": 1.9780834913253784, |
|
"eval_bigbrother_runtime": 41.757, |
|
"eval_bigbrother_samples_per_second": 32.809, |
|
"eval_bigbrother_steps_per_second": 1.03, |
|
"eval_bigbrother_wer": 0.5809232256505277, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"grad_norm": 2.030735731124878, |
|
"learning_rate": 5.273333333333333e-05, |
|
"loss": 0.6041, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_rundkast_loss": 0.3950729966163635, |
|
"eval_rundkast_runtime": 33.5538, |
|
"eval_rundkast_samples_per_second": 40.085, |
|
"eval_rundkast_steps_per_second": 1.282, |
|
"eval_rundkast_wer": 0.21762647152402165, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_nb_samtale_loss": 0.45818740129470825, |
|
"eval_nb_samtale_runtime": 37.9662, |
|
"eval_nb_samtale_samples_per_second": 14.039, |
|
"eval_nb_samtale_steps_per_second": 0.448, |
|
"eval_nb_samtale_wer": 0.21183268229166666, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_bigbrother_loss": 1.8719202280044556, |
|
"eval_bigbrother_runtime": 42.0553, |
|
"eval_bigbrother_samples_per_second": 32.576, |
|
"eval_bigbrother_steps_per_second": 1.022, |
|
"eval_bigbrother_wer": 0.5723032304841699, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"grad_norm": 1.995400071144104, |
|
"learning_rate": 5.86e-05, |
|
"loss": 0.586, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_rundkast_loss": 0.39550164341926575, |
|
"eval_rundkast_runtime": 33.4182, |
|
"eval_rundkast_samples_per_second": 40.248, |
|
"eval_rundkast_steps_per_second": 1.287, |
|
"eval_rundkast_wer": 0.21762647152402165, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_nb_samtale_loss": 0.46760231256484985, |
|
"eval_nb_samtale_runtime": 38.1072, |
|
"eval_nb_samtale_samples_per_second": 13.987, |
|
"eval_nb_samtale_steps_per_second": 0.446, |
|
"eval_nb_samtale_wer": 0.208251953125, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_bigbrother_loss": 1.8893344402313232, |
|
"eval_bigbrother_runtime": 41.5152, |
|
"eval_bigbrother_samples_per_second": 33.0, |
|
"eval_bigbrother_steps_per_second": 1.036, |
|
"eval_bigbrother_wer": 0.5765729477161041, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"grad_norm": 1.3565304279327393, |
|
"learning_rate": 6.446666666666667e-05, |
|
"loss": 0.5779, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_rundkast_loss": 0.39460697770118713, |
|
"eval_rundkast_runtime": 33.627, |
|
"eval_rundkast_samples_per_second": 39.998, |
|
"eval_rundkast_steps_per_second": 1.279, |
|
"eval_rundkast_wer": 0.21810372255806554, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_nb_samtale_loss": 0.45650386810302734, |
|
"eval_nb_samtale_runtime": 37.9631, |
|
"eval_nb_samtale_samples_per_second": 14.04, |
|
"eval_nb_samtale_steps_per_second": 0.448, |
|
"eval_nb_samtale_wer": 0.20613606770833334, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_bigbrother_loss": 1.9142467975616455, |
|
"eval_bigbrother_runtime": 41.4099, |
|
"eval_bigbrother_samples_per_second": 33.084, |
|
"eval_bigbrother_steps_per_second": 1.038, |
|
"eval_bigbrother_wer": 0.5729477161040845, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"grad_norm": 1.7810662984848022, |
|
"learning_rate": 7.033333333333334e-05, |
|
"loss": 0.5375, |
|
"step": 1056 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_rundkast_loss": 0.40589994192123413, |
|
"eval_rundkast_runtime": 33.6087, |
|
"eval_rundkast_samples_per_second": 40.019, |
|
"eval_rundkast_steps_per_second": 1.279, |
|
"eval_rundkast_wer": 0.2184218899140948, |
|
"step": 1056 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_nb_samtale_loss": 0.46365875005722046, |
|
"eval_nb_samtale_runtime": 38.2581, |
|
"eval_nb_samtale_samples_per_second": 13.932, |
|
"eval_nb_samtale_steps_per_second": 0.444, |
|
"eval_nb_samtale_wer": 0.201171875, |
|
"step": 1056 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_bigbrother_loss": 1.9760468006134033, |
|
"eval_bigbrother_runtime": 42.201, |
|
"eval_bigbrother_samples_per_second": 32.464, |
|
"eval_bigbrother_steps_per_second": 1.019, |
|
"eval_bigbrother_wer": 0.5706920164343833, |
|
"step": 1056 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"grad_norm": 1.9388726949691772, |
|
"learning_rate": 7.620000000000001e-05, |
|
"loss": 0.5238, |
|
"step": 1144 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_rundkast_loss": 0.4222950339317322, |
|
"eval_rundkast_runtime": 33.5989, |
|
"eval_rundkast_samples_per_second": 40.031, |
|
"eval_rundkast_steps_per_second": 1.28, |
|
"eval_rundkast_wer": 0.21945593382118994, |
|
"step": 1144 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_nb_samtale_loss": 0.47446364164352417, |
|
"eval_nb_samtale_runtime": 37.9325, |
|
"eval_nb_samtale_samples_per_second": 14.051, |
|
"eval_nb_samtale_steps_per_second": 0.448, |
|
"eval_nb_samtale_wer": 0.20345052083333334, |
|
"step": 1144 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_bigbrother_loss": 1.956613540649414, |
|
"eval_bigbrother_runtime": 41.6064, |
|
"eval_bigbrother_samples_per_second": 32.928, |
|
"eval_bigbrother_steps_per_second": 1.033, |
|
"eval_bigbrother_wer": 0.564408281640216, |
|
"step": 1144 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"grad_norm": 1.7794352769851685, |
|
"learning_rate": 8.206666666666666e-05, |
|
"loss": 0.5002, |
|
"step": 1232 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_rundkast_loss": 0.40369686484336853, |
|
"eval_rundkast_runtime": 33.4225, |
|
"eval_rundkast_samples_per_second": 40.242, |
|
"eval_rundkast_steps_per_second": 1.287, |
|
"eval_rundkast_wer": 0.2170696786509704, |
|
"step": 1232 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_nb_samtale_loss": 0.4533463418483734, |
|
"eval_nb_samtale_runtime": 38.1602, |
|
"eval_nb_samtale_samples_per_second": 13.967, |
|
"eval_nb_samtale_steps_per_second": 0.445, |
|
"eval_nb_samtale_wer": 0.19881184895833334, |
|
"step": 1232 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_bigbrother_loss": 1.9202650785446167, |
|
"eval_bigbrother_runtime": 41.8647, |
|
"eval_bigbrother_samples_per_second": 32.724, |
|
"eval_bigbrother_steps_per_second": 1.027, |
|
"eval_bigbrother_wer": 0.5650527672601305, |
|
"step": 1232 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"grad_norm": 2.007899284362793, |
|
"learning_rate": 8.793333333333333e-05, |
|
"loss": 0.4939, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_rundkast_loss": 0.39848214387893677, |
|
"eval_rundkast_runtime": 33.4922, |
|
"eval_rundkast_samples_per_second": 40.159, |
|
"eval_rundkast_steps_per_second": 1.284, |
|
"eval_rundkast_wer": 0.21794463888005092, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_nb_samtale_loss": 0.43362265825271606, |
|
"eval_nb_samtale_runtime": 37.9431, |
|
"eval_nb_samtale_samples_per_second": 14.047, |
|
"eval_nb_samtale_steps_per_second": 0.448, |
|
"eval_nb_samtale_wer": 0.20182291666666666, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_bigbrother_loss": 1.824406385421753, |
|
"eval_bigbrother_runtime": 41.6014, |
|
"eval_bigbrother_samples_per_second": 32.932, |
|
"eval_bigbrother_steps_per_second": 1.034, |
|
"eval_bigbrother_wer": 0.5605413679207283, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"grad_norm": 1.890428066253662, |
|
"learning_rate": 9.38e-05, |
|
"loss": 0.493, |
|
"step": 1408 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_rundkast_loss": 0.42038509249687195, |
|
"eval_rundkast_runtime": 33.8742, |
|
"eval_rundkast_samples_per_second": 39.706, |
|
"eval_rundkast_steps_per_second": 1.269, |
|
"eval_rundkast_wer": 0.22009226853324848, |
|
"step": 1408 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_nb_samtale_loss": 0.4756682515144348, |
|
"eval_nb_samtale_runtime": 38.2223, |
|
"eval_nb_samtale_samples_per_second": 13.945, |
|
"eval_nb_samtale_steps_per_second": 0.445, |
|
"eval_nb_samtale_wer": 0.20003255208333334, |
|
"step": 1408 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_bigbrother_loss": 2.017906665802002, |
|
"eval_bigbrother_runtime": 41.7123, |
|
"eval_bigbrother_samples_per_second": 32.844, |
|
"eval_bigbrother_steps_per_second": 1.031, |
|
"eval_bigbrother_wer": 0.5624748247804721, |
|
"step": 1408 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"grad_norm": 1.8507376909255981, |
|
"learning_rate": 9.966666666666667e-05, |
|
"loss": 0.4734, |
|
"step": 1496 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_rundkast_loss": 0.4344000518321991, |
|
"eval_rundkast_runtime": 34.0756, |
|
"eval_rundkast_samples_per_second": 39.471, |
|
"eval_rundkast_steps_per_second": 1.262, |
|
"eval_rundkast_wer": 0.21699013681196308, |
|
"step": 1496 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_nb_samtale_loss": 0.4668411314487457, |
|
"eval_nb_samtale_runtime": 37.9422, |
|
"eval_nb_samtale_samples_per_second": 14.048, |
|
"eval_nb_samtale_steps_per_second": 0.448, |
|
"eval_nb_samtale_wer": 0.19694010416666666, |
|
"step": 1496 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_bigbrother_loss": 1.9858715534210205, |
|
"eval_bigbrother_runtime": 41.6236, |
|
"eval_bigbrother_samples_per_second": 32.914, |
|
"eval_bigbrother_steps_per_second": 1.033, |
|
"eval_bigbrother_wer": 0.5611052928381536, |
|
"step": 1496 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"grad_norm": 1.6648415327072144, |
|
"learning_rate": 9.589108910891089e-05, |
|
"loss": 0.4444, |
|
"step": 1584 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_rundkast_loss": 0.459636926651001, |
|
"eval_rundkast_runtime": 33.6234, |
|
"eval_rundkast_samples_per_second": 40.002, |
|
"eval_rundkast_steps_per_second": 1.279, |
|
"eval_rundkast_wer": 0.2184218899140948, |
|
"step": 1584 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_nb_samtale_loss": 0.4778790771961212, |
|
"eval_nb_samtale_runtime": 37.93, |
|
"eval_nb_samtale_samples_per_second": 14.052, |
|
"eval_nb_samtale_steps_per_second": 0.448, |
|
"eval_nb_samtale_wer": 0.19482421875, |
|
"step": 1584 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_bigbrother_loss": 2.19545316696167, |
|
"eval_bigbrother_runtime": 41.4756, |
|
"eval_bigbrother_samples_per_second": 33.031, |
|
"eval_bigbrother_steps_per_second": 1.037, |
|
"eval_bigbrother_wer": 0.5647305244501732, |
|
"step": 1584 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"grad_norm": 2.001049041748047, |
|
"learning_rate": 9.153465346534654e-05, |
|
"loss": 0.4535, |
|
"step": 1672 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_rundkast_loss": 0.4358086585998535, |
|
"eval_rundkast_runtime": 33.8712, |
|
"eval_rundkast_samples_per_second": 39.709, |
|
"eval_rundkast_steps_per_second": 1.27, |
|
"eval_rundkast_wer": 0.21317212853961184, |
|
"step": 1672 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_nb_samtale_loss": 0.4770139157772064, |
|
"eval_nb_samtale_runtime": 38.0225, |
|
"eval_nb_samtale_samples_per_second": 14.018, |
|
"eval_nb_samtale_steps_per_second": 0.447, |
|
"eval_nb_samtale_wer": 0.19514973958333334, |
|
"step": 1672 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_bigbrother_loss": 2.057053804397583, |
|
"eval_bigbrother_runtime": 41.9595, |
|
"eval_bigbrother_samples_per_second": 32.651, |
|
"eval_bigbrother_steps_per_second": 1.025, |
|
"eval_bigbrother_wer": 0.5624748247804721, |
|
"step": 1672 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"grad_norm": 1.6764415502548218, |
|
"learning_rate": 8.717821782178219e-05, |
|
"loss": 0.4231, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_rundkast_loss": 0.41307970881462097, |
|
"eval_rundkast_runtime": 34.1234, |
|
"eval_rundkast_samples_per_second": 39.416, |
|
"eval_rundkast_steps_per_second": 1.26, |
|
"eval_rundkast_wer": 0.21603563474387527, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_nb_samtale_loss": 0.43975648283958435, |
|
"eval_nb_samtale_runtime": 37.7496, |
|
"eval_nb_samtale_samples_per_second": 14.119, |
|
"eval_nb_samtale_steps_per_second": 0.45, |
|
"eval_nb_samtale_wer": 0.18717447916666666, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_bigbrother_loss": 1.9533370733261108, |
|
"eval_bigbrother_runtime": 41.7155, |
|
"eval_bigbrother_samples_per_second": 32.842, |
|
"eval_bigbrother_steps_per_second": 1.031, |
|
"eval_bigbrother_wer": 0.5524852976717957, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"grad_norm": 1.599152684211731, |
|
"learning_rate": 8.282178217821782e-05, |
|
"loss": 0.4086, |
|
"step": 1848 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_rundkast_loss": 0.4404522776603699, |
|
"eval_rundkast_runtime": 33.7971, |
|
"eval_rundkast_samples_per_second": 39.796, |
|
"eval_rundkast_steps_per_second": 1.272, |
|
"eval_rundkast_wer": 0.2138880050906777, |
|
"step": 1848 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_nb_samtale_loss": 0.4675068259239197, |
|
"eval_nb_samtale_runtime": 37.8694, |
|
"eval_nb_samtale_samples_per_second": 14.075, |
|
"eval_nb_samtale_steps_per_second": 0.449, |
|
"eval_nb_samtale_wer": 0.19132486979166666, |
|
"step": 1848 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_bigbrother_loss": 2.153977632522583, |
|
"eval_bigbrother_runtime": 41.6263, |
|
"eval_bigbrother_samples_per_second": 32.912, |
|
"eval_bigbrother_steps_per_second": 1.033, |
|
"eval_bigbrother_wer": 0.5508740836220092, |
|
"step": 1848 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"grad_norm": 2.0951850414276123, |
|
"learning_rate": 7.846534653465347e-05, |
|
"loss": 0.3963, |
|
"step": 1936 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_rundkast_loss": 0.455175518989563, |
|
"eval_rundkast_runtime": 33.5764, |
|
"eval_rundkast_samples_per_second": 40.058, |
|
"eval_rundkast_steps_per_second": 1.281, |
|
"eval_rundkast_wer": 0.21643334393891187, |
|
"step": 1936 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_nb_samtale_loss": 0.4630807340145111, |
|
"eval_nb_samtale_runtime": 37.7668, |
|
"eval_nb_samtale_samples_per_second": 14.113, |
|
"eval_nb_samtale_steps_per_second": 0.45, |
|
"eval_nb_samtale_wer": 0.18977864583333334, |
|
"step": 1936 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_bigbrother_loss": 2.1877636909484863, |
|
"eval_bigbrother_runtime": 41.6656, |
|
"eval_bigbrother_samples_per_second": 32.881, |
|
"eval_bigbrother_steps_per_second": 1.032, |
|
"eval_bigbrother_wer": 0.5527269797792637, |
|
"step": 1936 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"grad_norm": 1.5935795307159424, |
|
"learning_rate": 7.410891089108911e-05, |
|
"loss": 0.3858, |
|
"step": 2024 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_rundkast_loss": 0.46708443760871887, |
|
"eval_rundkast_runtime": 33.6671, |
|
"eval_rundkast_samples_per_second": 39.95, |
|
"eval_rundkast_steps_per_second": 1.277, |
|
"eval_rundkast_wer": 0.2152402163538021, |
|
"step": 2024 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_nb_samtale_loss": 0.4683707356452942, |
|
"eval_nb_samtale_runtime": 37.9855, |
|
"eval_nb_samtale_samples_per_second": 14.032, |
|
"eval_nb_samtale_steps_per_second": 0.448, |
|
"eval_nb_samtale_wer": 0.185791015625, |
|
"step": 2024 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_bigbrother_loss": 2.148423910140991, |
|
"eval_bigbrother_runtime": 41.6837, |
|
"eval_bigbrother_samples_per_second": 32.867, |
|
"eval_bigbrother_steps_per_second": 1.032, |
|
"eval_bigbrother_wer": 0.5511157657294772, |
|
"step": 2024 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"grad_norm": 2.292874813079834, |
|
"learning_rate": 6.975247524752476e-05, |
|
"loss": 0.3763, |
|
"step": 2112 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_rundkast_loss": 0.47619959712028503, |
|
"eval_rundkast_runtime": 33.4679, |
|
"eval_rundkast_samples_per_second": 40.188, |
|
"eval_rundkast_steps_per_second": 1.285, |
|
"eval_rundkast_wer": 0.21810372255806554, |
|
"step": 2112 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_nb_samtale_loss": 0.47193431854248047, |
|
"eval_nb_samtale_runtime": 37.7044, |
|
"eval_nb_samtale_samples_per_second": 14.136, |
|
"eval_nb_samtale_steps_per_second": 0.451, |
|
"eval_nb_samtale_wer": 0.18709309895833334, |
|
"step": 2112 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_bigbrother_loss": 2.1468334197998047, |
|
"eval_bigbrother_runtime": 41.3571, |
|
"eval_bigbrother_samples_per_second": 33.126, |
|
"eval_bigbrother_steps_per_second": 1.04, |
|
"eval_bigbrother_wer": 0.559735760895835, |
|
"step": 2112 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"grad_norm": 2.466032028198242, |
|
"learning_rate": 6.53960396039604e-05, |
|
"loss": 0.3741, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_rundkast_loss": 0.4504246711730957, |
|
"eval_rundkast_runtime": 33.5467, |
|
"eval_rundkast_samples_per_second": 40.093, |
|
"eval_rundkast_steps_per_second": 1.282, |
|
"eval_rundkast_wer": 0.2134107540566338, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_nb_samtale_loss": 0.4489670693874359, |
|
"eval_nb_samtale_runtime": 37.6649, |
|
"eval_nb_samtale_samples_per_second": 14.151, |
|
"eval_nb_samtale_steps_per_second": 0.451, |
|
"eval_nb_samtale_wer": 0.185546875, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_bigbrother_loss": 2.1423308849334717, |
|
"eval_bigbrother_runtime": 41.5117, |
|
"eval_bigbrother_samples_per_second": 33.003, |
|
"eval_bigbrother_steps_per_second": 1.036, |
|
"eval_bigbrother_wer": 0.553290904696689, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"grad_norm": 1.5329217910766602, |
|
"learning_rate": 6.103960396039604e-05, |
|
"loss": 0.3723, |
|
"step": 2288 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_rundkast_loss": 0.4836730659008026, |
|
"eval_rundkast_runtime": 33.6087, |
|
"eval_rundkast_samples_per_second": 40.019, |
|
"eval_rundkast_steps_per_second": 1.279, |
|
"eval_rundkast_wer": 0.2152402163538021, |
|
"step": 2288 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_nb_samtale_loss": 0.4578171968460083, |
|
"eval_nb_samtale_runtime": 37.9306, |
|
"eval_nb_samtale_samples_per_second": 14.052, |
|
"eval_nb_samtale_steps_per_second": 0.448, |
|
"eval_nb_samtale_wer": 0.18424479166666666, |
|
"step": 2288 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_bigbrother_loss": 2.1947550773620605, |
|
"eval_bigbrother_runtime": 41.8164, |
|
"eval_bigbrother_samples_per_second": 32.762, |
|
"eval_bigbrother_steps_per_second": 1.028, |
|
"eval_bigbrother_wer": 0.549101748167244, |
|
"step": 2288 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"grad_norm": 1.3041456937789917, |
|
"learning_rate": 5.668316831683168e-05, |
|
"loss": 0.3684, |
|
"step": 2376 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_rundkast_loss": 0.4755971133708954, |
|
"eval_rundkast_runtime": 33.7598, |
|
"eval_rundkast_samples_per_second": 39.84, |
|
"eval_rundkast_steps_per_second": 1.274, |
|
"eval_rundkast_wer": 0.21197900095450206, |
|
"step": 2376 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_nb_samtale_loss": 0.4704474210739136, |
|
"eval_nb_samtale_runtime": 37.5334, |
|
"eval_nb_samtale_samples_per_second": 14.201, |
|
"eval_nb_samtale_steps_per_second": 0.453, |
|
"eval_nb_samtale_wer": 0.180419921875, |
|
"step": 2376 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_bigbrother_loss": 2.2346787452697754, |
|
"eval_bigbrother_runtime": 41.523, |
|
"eval_bigbrother_samples_per_second": 32.994, |
|
"eval_bigbrother_steps_per_second": 1.036, |
|
"eval_bigbrother_wer": 0.549101748167244, |
|
"step": 2376 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"grad_norm": 1.4035800695419312, |
|
"learning_rate": 5.232673267326733e-05, |
|
"loss": 0.3536, |
|
"step": 2464 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_rundkast_loss": 0.4591149687767029, |
|
"eval_rundkast_runtime": 33.8672, |
|
"eval_rundkast_samples_per_second": 39.714, |
|
"eval_rundkast_steps_per_second": 1.27, |
|
"eval_rundkast_wer": 0.2085587018771874, |
|
"step": 2464 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_nb_samtale_loss": 0.4490753710269928, |
|
"eval_nb_samtale_runtime": 37.5138, |
|
"eval_nb_samtale_samples_per_second": 14.208, |
|
"eval_nb_samtale_steps_per_second": 0.453, |
|
"eval_nb_samtale_wer": 0.18074544270833334, |
|
"step": 2464 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_bigbrother_loss": 2.1479501724243164, |
|
"eval_bigbrother_runtime": 41.5368, |
|
"eval_bigbrother_samples_per_second": 32.983, |
|
"eval_bigbrother_steps_per_second": 1.035, |
|
"eval_bigbrother_wer": 0.5502295980020946, |
|
"step": 2464 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"grad_norm": 2.3787569999694824, |
|
"learning_rate": 4.797029702970297e-05, |
|
"loss": 0.3422, |
|
"step": 2552 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"eval_rundkast_loss": 0.45249199867248535, |
|
"eval_rundkast_runtime": 33.8412, |
|
"eval_rundkast_samples_per_second": 39.745, |
|
"eval_rundkast_steps_per_second": 1.271, |
|
"eval_rundkast_wer": 0.2116608335984728, |
|
"step": 2552 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"eval_nb_samtale_loss": 0.42878830432891846, |
|
"eval_nb_samtale_runtime": 37.8412, |
|
"eval_nb_samtale_samples_per_second": 14.085, |
|
"eval_nb_samtale_steps_per_second": 0.449, |
|
"eval_nb_samtale_wer": 0.179931640625, |
|
"step": 2552 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"eval_bigbrother_loss": 2.074361562728882, |
|
"eval_bigbrother_runtime": 41.584, |
|
"eval_bigbrother_samples_per_second": 32.945, |
|
"eval_bigbrother_steps_per_second": 1.034, |
|
"eval_bigbrother_wer": 0.5470877306050109, |
|
"step": 2552 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"grad_norm": 1.0397918224334717, |
|
"learning_rate": 4.3613861386138617e-05, |
|
"loss": 0.3271, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_rundkast_loss": 0.4873400926589966, |
|
"eval_rundkast_runtime": 33.7561, |
|
"eval_rundkast_samples_per_second": 39.845, |
|
"eval_rundkast_steps_per_second": 1.274, |
|
"eval_rundkast_wer": 0.21174037543748012, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_nb_samtale_loss": 0.4706071615219116, |
|
"eval_nb_samtale_runtime": 37.8498, |
|
"eval_nb_samtale_samples_per_second": 14.082, |
|
"eval_nb_samtale_steps_per_second": 0.449, |
|
"eval_nb_samtale_wer": 0.17960611979166666, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_bigbrother_loss": 2.284547805786133, |
|
"eval_bigbrother_runtime": 43.2872, |
|
"eval_bigbrother_samples_per_second": 31.649, |
|
"eval_bigbrother_steps_per_second": 0.993, |
|
"eval_bigbrother_wer": 0.5479738983323934, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"grad_norm": 1.832560420036316, |
|
"learning_rate": 3.925742574257426e-05, |
|
"loss": 0.3351, |
|
"step": 2728 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"eval_rundkast_loss": 0.4675346612930298, |
|
"eval_rundkast_runtime": 33.6595, |
|
"eval_rundkast_samples_per_second": 39.959, |
|
"eval_rundkast_steps_per_second": 1.277, |
|
"eval_rundkast_wer": 0.21094495704740693, |
|
"step": 2728 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"eval_nb_samtale_loss": 0.44606560468673706, |
|
"eval_nb_samtale_runtime": 37.8399, |
|
"eval_nb_samtale_samples_per_second": 14.086, |
|
"eval_nb_samtale_steps_per_second": 0.449, |
|
"eval_nb_samtale_wer": 0.1767578125, |
|
"step": 2728 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"eval_bigbrother_loss": 2.2436790466308594, |
|
"eval_bigbrother_runtime": 41.8909, |
|
"eval_bigbrother_samples_per_second": 32.704, |
|
"eval_bigbrother_steps_per_second": 1.026, |
|
"eval_bigbrother_wer": 0.5497462337871586, |
|
"step": 2728 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"grad_norm": 2.598604917526245, |
|
"learning_rate": 3.49009900990099e-05, |
|
"loss": 0.3266, |
|
"step": 2816 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_rundkast_loss": 0.47241419553756714, |
|
"eval_rundkast_runtime": 33.6802, |
|
"eval_rundkast_samples_per_second": 39.934, |
|
"eval_rundkast_steps_per_second": 1.277, |
|
"eval_rundkast_wer": 0.2099109131403118, |
|
"step": 2816 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_nb_samtale_loss": 0.46288925409317017, |
|
"eval_nb_samtale_runtime": 37.3999, |
|
"eval_nb_samtale_samples_per_second": 14.251, |
|
"eval_nb_samtale_steps_per_second": 0.455, |
|
"eval_nb_samtale_wer": 0.17952473958333334, |
|
"step": 2816 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_bigbrother_loss": 2.2838997840881348, |
|
"eval_bigbrother_runtime": 41.6372, |
|
"eval_bigbrother_samples_per_second": 32.903, |
|
"eval_bigbrother_steps_per_second": 1.033, |
|
"eval_bigbrother_wer": 0.5497462337871586, |
|
"step": 2816 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"grad_norm": 2.8319473266601562, |
|
"learning_rate": 3.054455445544554e-05, |
|
"loss": 0.3164, |
|
"step": 2904 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"eval_rundkast_loss": 0.49051862955093384, |
|
"eval_rundkast_runtime": 33.5657, |
|
"eval_rundkast_samples_per_second": 40.071, |
|
"eval_rundkast_steps_per_second": 1.281, |
|
"eval_rundkast_wer": 0.21102449888641425, |
|
"step": 2904 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"eval_nb_samtale_loss": 0.47550272941589355, |
|
"eval_nb_samtale_runtime": 37.719, |
|
"eval_nb_samtale_samples_per_second": 14.131, |
|
"eval_nb_samtale_steps_per_second": 0.451, |
|
"eval_nb_samtale_wer": 0.17708333333333334, |
|
"step": 2904 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"eval_bigbrother_loss": 2.360140085220337, |
|
"eval_bigbrother_runtime": 41.7548, |
|
"eval_bigbrother_samples_per_second": 32.811, |
|
"eval_bigbrother_steps_per_second": 1.03, |
|
"eval_bigbrother_wer": 0.549343430274712, |
|
"step": 2904 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"grad_norm": 1.165239930152893, |
|
"learning_rate": 2.6188118811881192e-05, |
|
"loss": 0.325, |
|
"step": 2992 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_rundkast_loss": 0.4772399365901947, |
|
"eval_rundkast_runtime": 34.1144, |
|
"eval_rundkast_samples_per_second": 39.426, |
|
"eval_rundkast_steps_per_second": 1.26, |
|
"eval_rundkast_wer": 0.21142220808145085, |
|
"step": 2992 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_nb_samtale_loss": 0.45948928594589233, |
|
"eval_nb_samtale_runtime": 37.8345, |
|
"eval_nb_samtale_samples_per_second": 14.088, |
|
"eval_nb_samtale_steps_per_second": 0.449, |
|
"eval_nb_samtale_wer": 0.17692057291666666, |
|
"step": 2992 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_bigbrother_loss": 2.2873830795288086, |
|
"eval_bigbrother_runtime": 41.8269, |
|
"eval_bigbrother_samples_per_second": 32.754, |
|
"eval_bigbrother_steps_per_second": 1.028, |
|
"eval_bigbrother_wer": 0.5465238056875856, |
|
"step": 2992 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"grad_norm": 1.3796650171279907, |
|
"learning_rate": 2.1831683168316834e-05, |
|
"loss": 0.3195, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"eval_rundkast_loss": 0.48370474576950073, |
|
"eval_rundkast_runtime": 33.7199, |
|
"eval_rundkast_samples_per_second": 39.887, |
|
"eval_rundkast_steps_per_second": 1.275, |
|
"eval_rundkast_wer": 0.21006999681832644, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"eval_nb_samtale_loss": 0.4662381708621979, |
|
"eval_nb_samtale_runtime": 38.5158, |
|
"eval_nb_samtale_samples_per_second": 13.838, |
|
"eval_nb_samtale_steps_per_second": 0.441, |
|
"eval_nb_samtale_wer": 0.17635091145833334, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"eval_bigbrother_loss": 2.249934434890747, |
|
"eval_bigbrother_runtime": 41.5219, |
|
"eval_bigbrother_samples_per_second": 32.995, |
|
"eval_bigbrother_steps_per_second": 1.036, |
|
"eval_bigbrother_wer": 0.5461210021751389, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"grad_norm": 2.232914686203003, |
|
"learning_rate": 1.7475247524752476e-05, |
|
"loss": 0.3169, |
|
"step": 3168 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"eval_rundkast_loss": 0.4873496890068054, |
|
"eval_rundkast_runtime": 33.9486, |
|
"eval_rundkast_samples_per_second": 39.619, |
|
"eval_rundkast_steps_per_second": 1.267, |
|
"eval_rundkast_wer": 0.20935412026726058, |
|
"step": 3168 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"eval_nb_samtale_loss": 0.4695265591144562, |
|
"eval_nb_samtale_runtime": 37.5912, |
|
"eval_nb_samtale_samples_per_second": 14.179, |
|
"eval_nb_samtale_steps_per_second": 0.452, |
|
"eval_nb_samtale_wer": 0.175048828125, |
|
"step": 3168 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"eval_bigbrother_loss": 2.2812929153442383, |
|
"eval_bigbrother_runtime": 42.1676, |
|
"eval_bigbrother_samples_per_second": 32.489, |
|
"eval_bigbrother_steps_per_second": 1.02, |
|
"eval_bigbrother_wer": 0.545879320067671, |
|
"step": 3168 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"grad_norm": 1.227518916130066, |
|
"learning_rate": 1.311881188118812e-05, |
|
"loss": 0.3205, |
|
"step": 3256 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"eval_rundkast_loss": 0.4848962128162384, |
|
"eval_rundkast_runtime": 34.1615, |
|
"eval_rundkast_samples_per_second": 39.372, |
|
"eval_rundkast_steps_per_second": 1.259, |
|
"eval_rundkast_wer": 0.21046770601336304, |
|
"step": 3256 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"eval_nb_samtale_loss": 0.45955517888069153, |
|
"eval_nb_samtale_runtime": 37.7962, |
|
"eval_nb_samtale_samples_per_second": 14.102, |
|
"eval_nb_samtale_steps_per_second": 0.45, |
|
"eval_nb_samtale_wer": 0.17521158854166666, |
|
"step": 3256 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"eval_bigbrother_loss": 2.2612295150756836, |
|
"eval_bigbrother_runtime": 41.8835, |
|
"eval_bigbrother_samples_per_second": 32.71, |
|
"eval_bigbrother_steps_per_second": 1.027, |
|
"eval_bigbrother_wer": 0.5478127769274148, |
|
"step": 3256 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"grad_norm": 1.3873769044876099, |
|
"learning_rate": 8.762376237623762e-06, |
|
"loss": 0.3164, |
|
"step": 3344 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"eval_rundkast_loss": 0.4849892556667328, |
|
"eval_rundkast_runtime": 33.7856, |
|
"eval_rundkast_samples_per_second": 39.81, |
|
"eval_rundkast_steps_per_second": 1.273, |
|
"eval_rundkast_wer": 0.20999045497931912, |
|
"step": 3344 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"eval_nb_samtale_loss": 0.45760098099708557, |
|
"eval_nb_samtale_runtime": 37.6736, |
|
"eval_nb_samtale_samples_per_second": 14.148, |
|
"eval_nb_samtale_steps_per_second": 0.451, |
|
"eval_nb_samtale_wer": 0.17350260416666666, |
|
"step": 3344 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"eval_bigbrother_loss": 2.2668609619140625, |
|
"eval_bigbrother_runtime": 42.1887, |
|
"eval_bigbrother_samples_per_second": 32.473, |
|
"eval_bigbrother_steps_per_second": 1.019, |
|
"eval_bigbrother_wer": 0.5437041811004591, |
|
"step": 3344 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"grad_norm": 1.6516066789627075, |
|
"learning_rate": 4.405940594059406e-06, |
|
"loss": 0.3145, |
|
"step": 3432 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"eval_rundkast_loss": 0.48802104592323303, |
|
"eval_rundkast_runtime": 34.1337, |
|
"eval_rundkast_samples_per_second": 39.404, |
|
"eval_rundkast_steps_per_second": 1.26, |
|
"eval_rundkast_wer": 0.20911549475023863, |
|
"step": 3432 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"eval_nb_samtale_loss": 0.4593607187271118, |
|
"eval_nb_samtale_runtime": 37.7724, |
|
"eval_nb_samtale_samples_per_second": 14.111, |
|
"eval_nb_samtale_steps_per_second": 0.45, |
|
"eval_nb_samtale_wer": 0.17390950520833334, |
|
"step": 3432 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"eval_bigbrother_loss": 2.2844576835632324, |
|
"eval_bigbrother_runtime": 41.9423, |
|
"eval_bigbrother_samples_per_second": 32.664, |
|
"eval_bigbrother_steps_per_second": 1.025, |
|
"eval_bigbrother_wer": 0.5430596954805446, |
|
"step": 3432 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"grad_norm": 1.1558164358139038, |
|
"learning_rate": 4.950495049504951e-08, |
|
"loss": 0.3122, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_rundkast_loss": 0.4869055449962616, |
|
"eval_rundkast_runtime": 33.9078, |
|
"eval_rundkast_samples_per_second": 39.666, |
|
"eval_rundkast_steps_per_second": 1.268, |
|
"eval_rundkast_wer": 0.20895641107222399, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_nb_samtale_loss": 0.4583870470523834, |
|
"eval_nb_samtale_runtime": 37.9869, |
|
"eval_nb_samtale_samples_per_second": 14.031, |
|
"eval_nb_samtale_steps_per_second": 0.448, |
|
"eval_nb_samtale_wer": 0.174072265625, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_bigbrother_loss": 2.280332565307617, |
|
"eval_bigbrother_runtime": 41.7851, |
|
"eval_bigbrother_samples_per_second": 32.787, |
|
"eval_bigbrother_steps_per_second": 1.029, |
|
"eval_bigbrother_wer": 0.5434624989929913, |
|
"step": 3520 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 3520, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 40, |
|
"save_steps": 500, |
|
"total_flos": 4.807600562770979e+19, |
|
"train_batch_size": 48, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|