|
{ |
|
"best_metric": 0.5104931592941284, |
|
"best_model_checkpoint": "ai-light-dance_singing3_ft_wav2vec2-large-xlsr-53-v2/checkpoint-1872", |
|
"epoch": 49.996539792387544, |
|
"global_step": 3600, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 0.6213, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.9000000000000002e-06, |
|
"loss": 0.639, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 2.9e-06, |
|
"loss": 0.6243, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 3.900000000000001e-06, |
|
"loss": 0.6178, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.9000000000000005e-06, |
|
"loss": 0.5815, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 4.9887323943661975e-06, |
|
"loss": 0.5894, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.974647887323944e-06, |
|
"loss": 0.6154, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.526593804359436, |
|
"eval_runtime": 128.5544, |
|
"eval_samples_per_second": 4.138, |
|
"eval_steps_per_second": 0.521, |
|
"eval_wer": 0.2550620299543614, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 4.96056338028169e-06, |
|
"loss": 0.6172, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 4.946478873239437e-06, |
|
"loss": 0.6082, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 4.932394366197184e-06, |
|
"loss": 0.5951, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 4.91830985915493e-06, |
|
"loss": 0.6139, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 4.904225352112677e-06, |
|
"loss": 0.5957, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 4.890140845070423e-06, |
|
"loss": 0.6248, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 4.876056338028169e-06, |
|
"loss": 0.5958, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.5272348523139954, |
|
"eval_runtime": 129.1192, |
|
"eval_samples_per_second": 4.12, |
|
"eval_steps_per_second": 0.519, |
|
"eval_wer": 0.2585974159542328, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 4.861971830985916e-06, |
|
"loss": 0.6125, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 4.847887323943662e-06, |
|
"loss": 0.6002, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.833802816901409e-06, |
|
"loss": 0.5946, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 4.819718309859156e-06, |
|
"loss": 0.586, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 4.805633802816902e-06, |
|
"loss": 0.5792, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 4.7915492957746485e-06, |
|
"loss": 0.5761, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 4.777464788732395e-06, |
|
"loss": 0.5825, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.5249173641204834, |
|
"eval_runtime": 127.7312, |
|
"eval_samples_per_second": 4.165, |
|
"eval_steps_per_second": 0.525, |
|
"eval_wer": 0.25866169569968506, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 4.763380281690141e-06, |
|
"loss": 0.5987, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 4.749295774647888e-06, |
|
"loss": 0.5777, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 4.735211267605635e-06, |
|
"loss": 0.61, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 4.721126760563381e-06, |
|
"loss": 0.5833, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 4.707042253521127e-06, |
|
"loss": 0.5848, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 4.692957746478873e-06, |
|
"loss": 0.5808, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 4.6788732394366195e-06, |
|
"loss": 0.5717, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 0.5236164927482605, |
|
"eval_runtime": 127.5564, |
|
"eval_samples_per_second": 4.171, |
|
"eval_steps_per_second": 0.525, |
|
"eval_wer": 0.2570547020633798, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 4.664788732394367e-06, |
|
"loss": 0.6055, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 4.650704225352113e-06, |
|
"loss": 0.5704, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 4.6366197183098594e-06, |
|
"loss": 0.5835, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 4.622535211267606e-06, |
|
"loss": 0.5677, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"learning_rate": 4.608450704225352e-06, |
|
"loss": 0.5799, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 4.5943661971830986e-06, |
|
"loss": 0.576, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 4.580281690140845e-06, |
|
"loss": 0.5975, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 4.566197183098592e-06, |
|
"loss": 0.5831, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 0.5202789306640625, |
|
"eval_runtime": 129.1161, |
|
"eval_samples_per_second": 4.12, |
|
"eval_steps_per_second": 0.519, |
|
"eval_wer": 0.25898309442694606, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 4.5521126760563385e-06, |
|
"loss": 0.58, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 5.28, |
|
"learning_rate": 4.538028169014085e-06, |
|
"loss": 0.5869, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 5.42, |
|
"learning_rate": 4.523943661971831e-06, |
|
"loss": 0.5603, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 5.55, |
|
"learning_rate": 4.509859154929578e-06, |
|
"loss": 0.5657, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 4.495774647887324e-06, |
|
"loss": 0.5846, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 5.83, |
|
"learning_rate": 4.48169014084507e-06, |
|
"loss": 0.5951, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 5.97, |
|
"learning_rate": 4.467605633802818e-06, |
|
"loss": 0.5652, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 0.5126761198043823, |
|
"eval_runtime": 126.6507, |
|
"eval_samples_per_second": 4.201, |
|
"eval_steps_per_second": 0.529, |
|
"eval_wer": 0.2575046602815453, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 6.11, |
|
"learning_rate": 4.453521126760564e-06, |
|
"loss": 0.5965, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 4.43943661971831e-06, |
|
"loss": 0.5641, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 6.39, |
|
"learning_rate": 4.425352112676057e-06, |
|
"loss": 0.5547, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 4.411267605633803e-06, |
|
"loss": 0.5849, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 6.66, |
|
"learning_rate": 4.3971830985915495e-06, |
|
"loss": 0.5665, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 6.8, |
|
"learning_rate": 4.383098591549296e-06, |
|
"loss": 0.5666, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 6.94, |
|
"learning_rate": 4.369014084507043e-06, |
|
"loss": 0.5665, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 0.5228902101516724, |
|
"eval_runtime": 126.6048, |
|
"eval_samples_per_second": 4.202, |
|
"eval_steps_per_second": 0.529, |
|
"eval_wer": 0.25866169569968506, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 7.08, |
|
"learning_rate": 4.3549295774647895e-06, |
|
"loss": 0.5989, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 7.22, |
|
"learning_rate": 4.340845070422536e-06, |
|
"loss": 0.5625, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 7.36, |
|
"learning_rate": 4.326760563380282e-06, |
|
"loss": 0.5827, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 4.312676056338029e-06, |
|
"loss": 0.5864, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 7.64, |
|
"learning_rate": 4.298591549295775e-06, |
|
"loss": 0.5555, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 7.78, |
|
"learning_rate": 4.284507042253521e-06, |
|
"loss": 0.5606, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 7.91, |
|
"learning_rate": 4.270422535211268e-06, |
|
"loss": 0.5625, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 0.5248023867607117, |
|
"eval_runtime": 127.7741, |
|
"eval_samples_per_second": 4.164, |
|
"eval_steps_per_second": 0.524, |
|
"eval_wer": 0.25474063122710033, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"learning_rate": 4.256338028169014e-06, |
|
"loss": 0.5798, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 8.19, |
|
"learning_rate": 4.2422535211267605e-06, |
|
"loss": 0.5645, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 8.33, |
|
"learning_rate": 4.228169014084507e-06, |
|
"loss": 0.5717, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 8.47, |
|
"learning_rate": 4.214084507042254e-06, |
|
"loss": 0.5617, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 8.61, |
|
"learning_rate": 4.2000000000000004e-06, |
|
"loss": 0.564, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"learning_rate": 4.185915492957747e-06, |
|
"loss": 0.5758, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 8.89, |
|
"learning_rate": 4.171830985915493e-06, |
|
"loss": 0.5661, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 0.5213830471038818, |
|
"eval_runtime": 127.0111, |
|
"eval_samples_per_second": 4.189, |
|
"eval_steps_per_second": 0.528, |
|
"eval_wer": 0.25576910715433565, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 4.1577464788732396e-06, |
|
"loss": 0.6023, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 9.17, |
|
"learning_rate": 4.143661971830986e-06, |
|
"loss": 0.5642, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 9.3, |
|
"learning_rate": 4.129577464788732e-06, |
|
"loss": 0.5741, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 9.44, |
|
"learning_rate": 4.1154929577464795e-06, |
|
"loss": 0.5746, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 9.58, |
|
"learning_rate": 4.101408450704226e-06, |
|
"loss": 0.5749, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 9.72, |
|
"learning_rate": 4.087323943661972e-06, |
|
"loss": 0.5623, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"learning_rate": 4.073239436619719e-06, |
|
"loss": 0.5628, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 4.059154929577465e-06, |
|
"loss": 0.5583, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 0.5196737051010132, |
|
"eval_runtime": 128.0441, |
|
"eval_samples_per_second": 4.155, |
|
"eval_steps_per_second": 0.523, |
|
"eval_wer": 0.2582117374815196, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 10.14, |
|
"learning_rate": 4.045070422535211e-06, |
|
"loss": 0.5656, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 10.28, |
|
"learning_rate": 4.030985915492958e-06, |
|
"loss": 0.5687, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 10.42, |
|
"learning_rate": 4.016901408450705e-06, |
|
"loss": 0.5569, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 10.55, |
|
"learning_rate": 4.002816901408451e-06, |
|
"loss": 0.5771, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 10.69, |
|
"learning_rate": 3.988732394366198e-06, |
|
"loss": 0.568, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 10.83, |
|
"learning_rate": 3.974647887323944e-06, |
|
"loss": 0.5699, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 10.97, |
|
"learning_rate": 3.9605633802816905e-06, |
|
"loss": 0.5605, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_loss": 0.5212803483009338, |
|
"eval_runtime": 127.3424, |
|
"eval_samples_per_second": 4.178, |
|
"eval_steps_per_second": 0.526, |
|
"eval_wer": 0.26110432602686895, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 11.11, |
|
"learning_rate": 3.946478873239437e-06, |
|
"loss": 0.5823, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 11.25, |
|
"learning_rate": 3.932394366197183e-06, |
|
"loss": 0.5739, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 11.39, |
|
"learning_rate": 3.9183098591549305e-06, |
|
"loss": 0.5539, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 11.53, |
|
"learning_rate": 3.904225352112677e-06, |
|
"loss": 0.5672, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 11.66, |
|
"learning_rate": 3.890140845070423e-06, |
|
"loss": 0.557, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 11.8, |
|
"learning_rate": 3.87605633802817e-06, |
|
"loss": 0.5711, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 11.94, |
|
"learning_rate": 3.861971830985915e-06, |
|
"loss": 0.5784, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_loss": 0.5327615141868591, |
|
"eval_runtime": 128.9244, |
|
"eval_samples_per_second": 4.126, |
|
"eval_steps_per_second": 0.52, |
|
"eval_wer": 0.258340296972424, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 12.08, |
|
"learning_rate": 3.847887323943662e-06, |
|
"loss": 0.5593, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 12.22, |
|
"learning_rate": 3.833802816901409e-06, |
|
"loss": 0.5503, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 12.36, |
|
"learning_rate": 3.819718309859155e-06, |
|
"loss": 0.57, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 12.5, |
|
"learning_rate": 3.8056338028169015e-06, |
|
"loss": 0.556, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 12.64, |
|
"learning_rate": 3.791549295774648e-06, |
|
"loss": 0.5469, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 12.78, |
|
"learning_rate": 3.7774647887323947e-06, |
|
"loss": 0.5575, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 12.91, |
|
"learning_rate": 3.763380281690141e-06, |
|
"loss": 0.5636, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_loss": 0.5246398448944092, |
|
"eval_runtime": 127.8438, |
|
"eval_samples_per_second": 4.161, |
|
"eval_steps_per_second": 0.524, |
|
"eval_wer": 0.2585974159542328, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 13.06, |
|
"learning_rate": 3.7492957746478874e-06, |
|
"loss": 0.5818, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 13.19, |
|
"learning_rate": 3.7352112676056342e-06, |
|
"loss": 0.5387, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 13.33, |
|
"learning_rate": 3.7211267605633806e-06, |
|
"loss": 0.5792, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 13.47, |
|
"learning_rate": 3.707042253521127e-06, |
|
"loss": 0.5543, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 13.61, |
|
"learning_rate": 3.6929577464788733e-06, |
|
"loss": 0.5429, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 13.75, |
|
"learning_rate": 3.67887323943662e-06, |
|
"loss": 0.5571, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 13.89, |
|
"learning_rate": 3.6647887323943665e-06, |
|
"loss": 0.5581, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_loss": 0.522983968257904, |
|
"eval_runtime": 128.2074, |
|
"eval_samples_per_second": 4.15, |
|
"eval_steps_per_second": 0.523, |
|
"eval_wer": 0.25461207173619593, |
|
"step": 1008 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.650704225352113e-06, |
|
"loss": 0.5776, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 14.17, |
|
"learning_rate": 3.6366197183098597e-06, |
|
"loss": 0.5523, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 14.3, |
|
"learning_rate": 3.622535211267606e-06, |
|
"loss": 0.5603, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 14.44, |
|
"learning_rate": 3.6084507042253524e-06, |
|
"loss": 0.5341, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 14.58, |
|
"learning_rate": 3.594366197183099e-06, |
|
"loss": 0.5589, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 14.72, |
|
"learning_rate": 3.5802816901408456e-06, |
|
"loss": 0.5508, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 14.86, |
|
"learning_rate": 3.566197183098592e-06, |
|
"loss": 0.5483, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 3.5521126760563384e-06, |
|
"loss": 0.567, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_loss": 0.5205003619194031, |
|
"eval_runtime": 128.349, |
|
"eval_samples_per_second": 4.145, |
|
"eval_steps_per_second": 0.522, |
|
"eval_wer": 0.2571832615542842, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 15.14, |
|
"learning_rate": 3.538028169014085e-06, |
|
"loss": 0.5702, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 15.28, |
|
"learning_rate": 3.5239436619718315e-06, |
|
"loss": 0.5812, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 15.42, |
|
"learning_rate": 3.509859154929578e-06, |
|
"loss": 0.5577, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 15.55, |
|
"learning_rate": 3.4957746478873243e-06, |
|
"loss": 0.5728, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 15.69, |
|
"learning_rate": 3.481690140845071e-06, |
|
"loss": 0.5317, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 15.83, |
|
"learning_rate": 3.4676056338028175e-06, |
|
"loss": 0.5601, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 15.97, |
|
"learning_rate": 3.4535211267605634e-06, |
|
"loss": 0.5586, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_loss": 0.5258896946907043, |
|
"eval_runtime": 126.5534, |
|
"eval_samples_per_second": 4.204, |
|
"eval_steps_per_second": 0.529, |
|
"eval_wer": 0.25564054766343125, |
|
"step": 1152 |
|
}, |
|
{ |
|
"epoch": 16.11, |
|
"learning_rate": 3.4394366197183098e-06, |
|
"loss": 0.5722, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 16.25, |
|
"learning_rate": 3.4253521126760566e-06, |
|
"loss": 0.5569, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 16.39, |
|
"learning_rate": 3.411267605633803e-06, |
|
"loss": 0.5507, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 16.53, |
|
"learning_rate": 3.3971830985915493e-06, |
|
"loss": 0.57, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 16.66, |
|
"learning_rate": 3.3830985915492957e-06, |
|
"loss": 0.557, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 16.8, |
|
"learning_rate": 3.3690140845070425e-06, |
|
"loss": 0.5366, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 16.94, |
|
"learning_rate": 3.354929577464789e-06, |
|
"loss": 0.5358, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_loss": 0.5334306359291077, |
|
"eval_runtime": 128.1209, |
|
"eval_samples_per_second": 4.152, |
|
"eval_steps_per_second": 0.523, |
|
"eval_wer": 0.26046152857234683, |
|
"step": 1224 |
|
}, |
|
{ |
|
"epoch": 17.08, |
|
"learning_rate": 3.3408450704225353e-06, |
|
"loss": 0.5884, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 17.22, |
|
"learning_rate": 3.326760563380282e-06, |
|
"loss": 0.5394, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 17.36, |
|
"learning_rate": 3.3126760563380284e-06, |
|
"loss": 0.5633, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 17.5, |
|
"learning_rate": 3.298591549295775e-06, |
|
"loss": 0.5456, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 17.64, |
|
"learning_rate": 3.284507042253521e-06, |
|
"loss": 0.5509, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 17.78, |
|
"learning_rate": 3.270422535211268e-06, |
|
"loss": 0.5746, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 17.91, |
|
"learning_rate": 3.2563380281690144e-06, |
|
"loss": 0.5526, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_loss": 0.5180938839912415, |
|
"eval_runtime": 127.6299, |
|
"eval_samples_per_second": 4.168, |
|
"eval_steps_per_second": 0.525, |
|
"eval_wer": 0.25557626791797905, |
|
"step": 1296 |
|
}, |
|
{ |
|
"epoch": 18.06, |
|
"learning_rate": 3.2422535211267607e-06, |
|
"loss": 0.5768, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 18.19, |
|
"learning_rate": 3.2281690140845075e-06, |
|
"loss": 0.5491, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 18.33, |
|
"learning_rate": 3.214084507042254e-06, |
|
"loss": 0.5738, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 18.47, |
|
"learning_rate": 3.2000000000000003e-06, |
|
"loss": 0.5564, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 18.61, |
|
"learning_rate": 3.1859154929577467e-06, |
|
"loss": 0.5605, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 18.75, |
|
"learning_rate": 3.1718309859154935e-06, |
|
"loss": 0.5443, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 18.89, |
|
"learning_rate": 3.15774647887324e-06, |
|
"loss": 0.5483, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_loss": 0.5131182074546814, |
|
"eval_runtime": 127.3437, |
|
"eval_samples_per_second": 4.178, |
|
"eval_steps_per_second": 0.526, |
|
"eval_wer": 0.2562190653725011, |
|
"step": 1368 |
|
}, |
|
{ |
|
"epoch": 19.03, |
|
"learning_rate": 3.143661971830986e-06, |
|
"loss": 0.5656, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 19.17, |
|
"learning_rate": 3.129577464788733e-06, |
|
"loss": 0.5534, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 19.3, |
|
"learning_rate": 3.1154929577464794e-06, |
|
"loss": 0.5568, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 19.44, |
|
"learning_rate": 3.1014084507042258e-06, |
|
"loss": 0.5492, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 19.58, |
|
"learning_rate": 3.087323943661972e-06, |
|
"loss": 0.5537, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 19.72, |
|
"learning_rate": 3.073239436619719e-06, |
|
"loss": 0.5541, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 19.86, |
|
"learning_rate": 3.0591549295774653e-06, |
|
"loss": 0.5536, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"learning_rate": 3.0450704225352117e-06, |
|
"loss": 0.5487, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_loss": 0.5178970098495483, |
|
"eval_runtime": 128.7396, |
|
"eval_samples_per_second": 4.132, |
|
"eval_steps_per_second": 0.52, |
|
"eval_wer": 0.2560905058815967, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 20.14, |
|
"learning_rate": 3.0309859154929576e-06, |
|
"loss": 0.551, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 20.28, |
|
"learning_rate": 3.0169014084507044e-06, |
|
"loss": 0.554, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 20.42, |
|
"learning_rate": 3.002816901408451e-06, |
|
"loss": 0.5524, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 20.55, |
|
"learning_rate": 2.988732394366197e-06, |
|
"loss": 0.5676, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 20.69, |
|
"learning_rate": 2.9746478873239436e-06, |
|
"loss": 0.5488, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 20.83, |
|
"learning_rate": 2.9605633802816903e-06, |
|
"loss": 0.5691, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 20.97, |
|
"learning_rate": 2.9464788732394367e-06, |
|
"loss": 0.5489, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_loss": 0.525915265083313, |
|
"eval_runtime": 128.1257, |
|
"eval_samples_per_second": 4.152, |
|
"eval_steps_per_second": 0.523, |
|
"eval_wer": 0.2596258918814682, |
|
"step": 1512 |
|
}, |
|
{ |
|
"epoch": 21.11, |
|
"learning_rate": 2.932394366197183e-06, |
|
"loss": 0.5522, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 21.25, |
|
"learning_rate": 2.91830985915493e-06, |
|
"loss": 0.5544, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 21.39, |
|
"learning_rate": 2.9042253521126763e-06, |
|
"loss": 0.5715, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 21.53, |
|
"learning_rate": 2.8901408450704226e-06, |
|
"loss": 0.5547, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 21.66, |
|
"learning_rate": 2.876056338028169e-06, |
|
"loss": 0.5256, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 21.8, |
|
"learning_rate": 2.861971830985916e-06, |
|
"loss": 0.5507, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 21.94, |
|
"learning_rate": 2.847887323943662e-06, |
|
"loss": 0.5582, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_loss": 0.5199052691459656, |
|
"eval_runtime": 128.2296, |
|
"eval_samples_per_second": 4.149, |
|
"eval_steps_per_second": 0.523, |
|
"eval_wer": 0.2550620299543614, |
|
"step": 1584 |
|
}, |
|
{ |
|
"epoch": 22.08, |
|
"learning_rate": 2.8338028169014086e-06, |
|
"loss": 0.5706, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 22.22, |
|
"learning_rate": 2.8197183098591554e-06, |
|
"loss": 0.5487, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 22.36, |
|
"learning_rate": 2.8056338028169017e-06, |
|
"loss": 0.5494, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 22.5, |
|
"learning_rate": 2.791549295774648e-06, |
|
"loss": 0.5536, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 22.64, |
|
"learning_rate": 2.7774647887323945e-06, |
|
"loss": 0.5469, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 22.78, |
|
"learning_rate": 2.7633802816901413e-06, |
|
"loss": 0.556, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 22.91, |
|
"learning_rate": 2.7492957746478877e-06, |
|
"loss": 0.5351, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_loss": 0.5282993316650391, |
|
"eval_runtime": 129.7167, |
|
"eval_samples_per_second": 4.101, |
|
"eval_steps_per_second": 0.517, |
|
"eval_wer": 0.25345503631805616, |
|
"step": 1656 |
|
}, |
|
{ |
|
"epoch": 23.06, |
|
"learning_rate": 2.735211267605634e-06, |
|
"loss": 0.5505, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 23.19, |
|
"learning_rate": 2.721126760563381e-06, |
|
"loss": 0.5448, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 23.33, |
|
"learning_rate": 2.7070422535211272e-06, |
|
"loss": 0.5539, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 23.47, |
|
"learning_rate": 2.6929577464788736e-06, |
|
"loss": 0.5401, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 23.61, |
|
"learning_rate": 2.67887323943662e-06, |
|
"loss": 0.5626, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 23.75, |
|
"learning_rate": 2.6647887323943668e-06, |
|
"loss": 0.5588, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 23.89, |
|
"learning_rate": 2.650704225352113e-06, |
|
"loss": 0.5572, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_loss": 0.5120282173156738, |
|
"eval_runtime": 127.888, |
|
"eval_samples_per_second": 4.16, |
|
"eval_steps_per_second": 0.524, |
|
"eval_wer": 0.25326219708169956, |
|
"step": 1728 |
|
}, |
|
{ |
|
"epoch": 24.03, |
|
"learning_rate": 2.6366197183098595e-06, |
|
"loss": 0.5649, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 24.17, |
|
"learning_rate": 2.6225352112676055e-06, |
|
"loss": 0.5481, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 24.3, |
|
"learning_rate": 2.6084507042253523e-06, |
|
"loss": 0.5478, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 24.44, |
|
"learning_rate": 2.5943661971830986e-06, |
|
"loss": 0.5404, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 24.58, |
|
"learning_rate": 2.580281690140845e-06, |
|
"loss": 0.5593, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 24.72, |
|
"learning_rate": 2.5661971830985914e-06, |
|
"loss": 0.5302, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 24.86, |
|
"learning_rate": 2.552112676056338e-06, |
|
"loss": 0.5557, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"learning_rate": 2.5380281690140846e-06, |
|
"loss": 0.5467, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_loss": 0.5175544619560242, |
|
"eval_runtime": 129.516, |
|
"eval_samples_per_second": 4.108, |
|
"eval_steps_per_second": 0.517, |
|
"eval_wer": 0.25782605900880634, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 25.14, |
|
"learning_rate": 2.523943661971831e-06, |
|
"loss": 0.5543, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 25.28, |
|
"learning_rate": 2.5098591549295777e-06, |
|
"loss": 0.55, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 25.42, |
|
"learning_rate": 2.495774647887324e-06, |
|
"loss": 0.5348, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 25.55, |
|
"learning_rate": 2.4816901408450705e-06, |
|
"loss": 0.5401, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 25.69, |
|
"learning_rate": 2.467605633802817e-06, |
|
"loss": 0.5341, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 25.83, |
|
"learning_rate": 2.4535211267605637e-06, |
|
"loss": 0.552, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 25.97, |
|
"learning_rate": 2.43943661971831e-06, |
|
"loss": 0.5424, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_loss": 0.5104931592941284, |
|
"eval_runtime": 127.2359, |
|
"eval_samples_per_second": 4.181, |
|
"eval_steps_per_second": 0.527, |
|
"eval_wer": 0.2551905894452658, |
|
"step": 1872 |
|
}, |
|
{ |
|
"epoch": 26.11, |
|
"learning_rate": 2.4253521126760564e-06, |
|
"loss": 0.5556, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 26.25, |
|
"learning_rate": 2.4112676056338032e-06, |
|
"loss": 0.546, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 26.39, |
|
"learning_rate": 2.3971830985915496e-06, |
|
"loss": 0.5392, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 26.53, |
|
"learning_rate": 2.383098591549296e-06, |
|
"loss": 0.5465, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 26.66, |
|
"learning_rate": 2.3690140845070423e-06, |
|
"loss": 0.5393, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 26.8, |
|
"learning_rate": 2.354929577464789e-06, |
|
"loss": 0.5469, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 26.94, |
|
"learning_rate": 2.3408450704225355e-06, |
|
"loss": 0.5344, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_loss": 0.5212472677230835, |
|
"eval_runtime": 128.0249, |
|
"eval_samples_per_second": 4.155, |
|
"eval_steps_per_second": 0.523, |
|
"eval_wer": 0.2540978337725783, |
|
"step": 1944 |
|
}, |
|
{ |
|
"epoch": 27.08, |
|
"learning_rate": 2.326760563380282e-06, |
|
"loss": 0.5647, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 27.22, |
|
"learning_rate": 2.3126760563380287e-06, |
|
"loss": 0.5383, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 27.36, |
|
"learning_rate": 2.2985915492957746e-06, |
|
"loss": 0.5333, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 27.5, |
|
"learning_rate": 2.284507042253521e-06, |
|
"loss": 0.5551, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 27.64, |
|
"learning_rate": 2.270422535211268e-06, |
|
"loss": 0.5477, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 27.78, |
|
"learning_rate": 2.256338028169014e-06, |
|
"loss": 0.5594, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 27.91, |
|
"learning_rate": 2.2422535211267606e-06, |
|
"loss": 0.5444, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_loss": 0.5154798626899719, |
|
"eval_runtime": 128.0303, |
|
"eval_samples_per_second": 4.155, |
|
"eval_steps_per_second": 0.523, |
|
"eval_wer": 0.25557626791797905, |
|
"step": 2016 |
|
}, |
|
{ |
|
"epoch": 28.06, |
|
"learning_rate": 2.2281690140845074e-06, |
|
"loss": 0.569, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 28.19, |
|
"learning_rate": 2.2140845070422537e-06, |
|
"loss": 0.5493, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 28.33, |
|
"learning_rate": 2.2e-06, |
|
"loss": 0.5485, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 28.47, |
|
"learning_rate": 2.1859154929577465e-06, |
|
"loss": 0.5367, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 28.61, |
|
"learning_rate": 2.1718309859154933e-06, |
|
"loss": 0.5487, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 28.75, |
|
"learning_rate": 2.1577464788732397e-06, |
|
"loss": 0.5487, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 28.89, |
|
"learning_rate": 2.143661971830986e-06, |
|
"loss": 0.5276, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"eval_loss": 0.523108959197998, |
|
"eval_runtime": 128.0146, |
|
"eval_samples_per_second": 4.156, |
|
"eval_steps_per_second": 0.523, |
|
"eval_wer": 0.2551263096998136, |
|
"step": 2088 |
|
}, |
|
{ |
|
"epoch": 29.03, |
|
"learning_rate": 2.129577464788733e-06, |
|
"loss": 0.546, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 29.17, |
|
"learning_rate": 2.115492957746479e-06, |
|
"loss": 0.5308, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 29.3, |
|
"learning_rate": 2.1014084507042256e-06, |
|
"loss": 0.5303, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 29.44, |
|
"learning_rate": 2.087323943661972e-06, |
|
"loss": 0.5453, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 29.58, |
|
"learning_rate": 2.0732394366197183e-06, |
|
"loss": 0.5396, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 29.72, |
|
"learning_rate": 2.0591549295774647e-06, |
|
"loss": 0.5352, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 29.86, |
|
"learning_rate": 2.0450704225352115e-06, |
|
"loss": 0.54, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"learning_rate": 2.030985915492958e-06, |
|
"loss": 0.5501, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_loss": 0.5224459171295166, |
|
"eval_runtime": 128.6755, |
|
"eval_samples_per_second": 4.134, |
|
"eval_steps_per_second": 0.521, |
|
"eval_wer": 0.25570482740888345, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 30.14, |
|
"learning_rate": 2.0169014084507043e-06, |
|
"loss": 0.554, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 30.28, |
|
"learning_rate": 2.002816901408451e-06, |
|
"loss": 0.555, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 30.42, |
|
"learning_rate": 1.9887323943661974e-06, |
|
"loss": 0.5414, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 30.55, |
|
"learning_rate": 1.974647887323944e-06, |
|
"loss": 0.5469, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 30.69, |
|
"learning_rate": 1.96056338028169e-06, |
|
"loss": 0.5356, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 30.83, |
|
"learning_rate": 1.946478873239437e-06, |
|
"loss": 0.5483, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 30.97, |
|
"learning_rate": 1.9323943661971834e-06, |
|
"loss": 0.5335, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"eval_loss": 0.5278783440589905, |
|
"eval_runtime": 128.5943, |
|
"eval_samples_per_second": 4.137, |
|
"eval_steps_per_second": 0.521, |
|
"eval_wer": 0.2549977502089092, |
|
"step": 2232 |
|
}, |
|
{ |
|
"epoch": 31.11, |
|
"learning_rate": 1.9183098591549297e-06, |
|
"loss": 0.5579, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 31.25, |
|
"learning_rate": 1.9042253521126763e-06, |
|
"loss": 0.5416, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 31.39, |
|
"learning_rate": 1.8901408450704225e-06, |
|
"loss": 0.5325, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 31.53, |
|
"learning_rate": 1.876056338028169e-06, |
|
"loss": 0.5376, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 31.66, |
|
"learning_rate": 1.8619718309859157e-06, |
|
"loss": 0.5382, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 31.8, |
|
"learning_rate": 1.847887323943662e-06, |
|
"loss": 0.5517, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 31.94, |
|
"learning_rate": 1.8338028169014086e-06, |
|
"loss": 0.5315, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_loss": 0.515052855014801, |
|
"eval_runtime": 128.5881, |
|
"eval_samples_per_second": 4.137, |
|
"eval_steps_per_second": 0.521, |
|
"eval_wer": 0.25448351224529153, |
|
"step": 2304 |
|
}, |
|
{ |
|
"epoch": 32.08, |
|
"learning_rate": 1.819718309859155e-06, |
|
"loss": 0.5712, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 32.22, |
|
"learning_rate": 1.8056338028169016e-06, |
|
"loss": 0.5349, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 32.36, |
|
"learning_rate": 1.791549295774648e-06, |
|
"loss": 0.5408, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 32.5, |
|
"learning_rate": 1.7774647887323945e-06, |
|
"loss": 0.5405, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 32.64, |
|
"learning_rate": 1.7633802816901411e-06, |
|
"loss": 0.5309, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 32.78, |
|
"learning_rate": 1.7492957746478875e-06, |
|
"loss": 0.5463, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 32.91, |
|
"learning_rate": 1.735211267605634e-06, |
|
"loss": 0.5344, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"eval_loss": 0.5204472541809082, |
|
"eval_runtime": 129.9429, |
|
"eval_samples_per_second": 4.094, |
|
"eval_steps_per_second": 0.516, |
|
"eval_wer": 0.2528122388635341, |
|
"step": 2376 |
|
}, |
|
{ |
|
"epoch": 33.06, |
|
"learning_rate": 1.7211267605633805e-06, |
|
"loss": 0.5626, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 33.19, |
|
"learning_rate": 1.707042253521127e-06, |
|
"loss": 0.5423, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 33.33, |
|
"learning_rate": 1.6929577464788734e-06, |
|
"loss": 0.5505, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 33.47, |
|
"learning_rate": 1.6788732394366198e-06, |
|
"loss": 0.5354, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 33.61, |
|
"learning_rate": 1.6647887323943662e-06, |
|
"loss": 0.5366, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 33.75, |
|
"learning_rate": 1.6507042253521128e-06, |
|
"loss": 0.5445, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 33.89, |
|
"learning_rate": 1.6366197183098591e-06, |
|
"loss": 0.5249, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_loss": 0.515282154083252, |
|
"eval_runtime": 128.9615, |
|
"eval_samples_per_second": 4.125, |
|
"eval_steps_per_second": 0.52, |
|
"eval_wer": 0.2542906730089349, |
|
"step": 2448 |
|
}, |
|
{ |
|
"epoch": 34.03, |
|
"learning_rate": 1.6225352112676057e-06, |
|
"loss": 0.5576, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 34.17, |
|
"learning_rate": 1.6084507042253523e-06, |
|
"loss": 0.5375, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 34.3, |
|
"learning_rate": 1.5943661971830987e-06, |
|
"loss": 0.5439, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 34.44, |
|
"learning_rate": 1.5802816901408453e-06, |
|
"loss": 0.5376, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 34.58, |
|
"learning_rate": 1.5661971830985917e-06, |
|
"loss": 0.5327, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 34.72, |
|
"learning_rate": 1.5521126760563382e-06, |
|
"loss": 0.5283, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 34.86, |
|
"learning_rate": 1.5380281690140846e-06, |
|
"loss": 0.5375, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"learning_rate": 1.5239436619718312e-06, |
|
"loss": 0.5478, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"eval_loss": 0.5154463648796082, |
|
"eval_runtime": 129.3389, |
|
"eval_samples_per_second": 4.113, |
|
"eval_steps_per_second": 0.518, |
|
"eval_wer": 0.2543549527543871, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 35.14, |
|
"learning_rate": 1.5098591549295778e-06, |
|
"loss": 0.526, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 35.28, |
|
"learning_rate": 1.4957746478873242e-06, |
|
"loss": 0.5461, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 35.42, |
|
"learning_rate": 1.4816901408450707e-06, |
|
"loss": 0.5357, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 35.55, |
|
"learning_rate": 1.467605633802817e-06, |
|
"loss": 0.5379, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 35.69, |
|
"learning_rate": 1.4535211267605635e-06, |
|
"loss": 0.5354, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 35.83, |
|
"learning_rate": 1.4394366197183099e-06, |
|
"loss": 0.5374, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 35.97, |
|
"learning_rate": 1.4253521126760565e-06, |
|
"loss": 0.5346, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"eval_loss": 0.5122731924057007, |
|
"eval_runtime": 127.7459, |
|
"eval_samples_per_second": 4.165, |
|
"eval_steps_per_second": 0.524, |
|
"eval_wer": 0.25339075657260396, |
|
"step": 2592 |
|
}, |
|
{ |
|
"epoch": 36.11, |
|
"learning_rate": 1.4112676056338028e-06, |
|
"loss": 0.5522, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 36.25, |
|
"learning_rate": 1.3971830985915494e-06, |
|
"loss": 0.5397, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 36.39, |
|
"learning_rate": 1.3830985915492958e-06, |
|
"loss": 0.519, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 36.53, |
|
"learning_rate": 1.3690140845070424e-06, |
|
"loss": 0.5558, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 36.66, |
|
"learning_rate": 1.354929577464789e-06, |
|
"loss": 0.5407, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 36.8, |
|
"learning_rate": 1.3408450704225353e-06, |
|
"loss": 0.533, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 36.94, |
|
"learning_rate": 1.326760563380282e-06, |
|
"loss": 0.5436, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"eval_loss": 0.5210054516792297, |
|
"eval_runtime": 129.2771, |
|
"eval_samples_per_second": 4.115, |
|
"eval_steps_per_second": 0.518, |
|
"eval_wer": 0.25654046409976217, |
|
"step": 2664 |
|
}, |
|
{ |
|
"epoch": 37.08, |
|
"learning_rate": 1.3126760563380283e-06, |
|
"loss": 0.564, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 37.22, |
|
"learning_rate": 1.298591549295775e-06, |
|
"loss": 0.5498, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 37.36, |
|
"learning_rate": 1.2845070422535213e-06, |
|
"loss": 0.5368, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 37.5, |
|
"learning_rate": 1.2704225352112676e-06, |
|
"loss": 0.5166, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 37.64, |
|
"learning_rate": 1.256338028169014e-06, |
|
"loss": 0.5314, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 37.78, |
|
"learning_rate": 1.2422535211267606e-06, |
|
"loss": 0.5411, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 37.91, |
|
"learning_rate": 1.2281690140845072e-06, |
|
"loss": 0.5299, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"eval_loss": 0.5182300806045532, |
|
"eval_runtime": 129.021, |
|
"eval_samples_per_second": 4.123, |
|
"eval_steps_per_second": 0.519, |
|
"eval_wer": 0.253712155299865, |
|
"step": 2736 |
|
}, |
|
{ |
|
"epoch": 38.06, |
|
"learning_rate": 1.2140845070422536e-06, |
|
"loss": 0.5536, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 38.19, |
|
"learning_rate": 1.2000000000000002e-06, |
|
"loss": 0.5458, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 38.33, |
|
"learning_rate": 1.1859154929577465e-06, |
|
"loss": 0.5247, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 38.47, |
|
"learning_rate": 1.1718309859154931e-06, |
|
"loss": 0.5405, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 38.61, |
|
"learning_rate": 1.1577464788732395e-06, |
|
"loss": 0.547, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 38.75, |
|
"learning_rate": 1.143661971830986e-06, |
|
"loss": 0.5321, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 38.89, |
|
"learning_rate": 1.1295774647887325e-06, |
|
"loss": 0.5248, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"eval_loss": 0.5239952206611633, |
|
"eval_runtime": 129.8341, |
|
"eval_samples_per_second": 4.098, |
|
"eval_steps_per_second": 0.516, |
|
"eval_wer": 0.2528765186089863, |
|
"step": 2808 |
|
}, |
|
{ |
|
"epoch": 39.03, |
|
"learning_rate": 1.115492957746479e-06, |
|
"loss": 0.5514, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 39.17, |
|
"learning_rate": 1.1014084507042254e-06, |
|
"loss": 0.5394, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 39.3, |
|
"learning_rate": 1.0873239436619718e-06, |
|
"loss": 0.5387, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 39.44, |
|
"learning_rate": 1.0732394366197184e-06, |
|
"loss": 0.513, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 39.58, |
|
"learning_rate": 1.059154929577465e-06, |
|
"loss": 0.5425, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 39.72, |
|
"learning_rate": 1.0450704225352113e-06, |
|
"loss": 0.5397, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 39.86, |
|
"learning_rate": 1.030985915492958e-06, |
|
"loss": 0.5407, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"learning_rate": 1.0169014084507043e-06, |
|
"loss": 0.5295, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_loss": 0.5250394940376282, |
|
"eval_runtime": 128.5995, |
|
"eval_samples_per_second": 4.137, |
|
"eval_steps_per_second": 0.521, |
|
"eval_wer": 0.2562833451179533, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 40.14, |
|
"learning_rate": 1.0028169014084507e-06, |
|
"loss": 0.5626, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 40.28, |
|
"learning_rate": 9.887323943661973e-07, |
|
"loss": 0.5362, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 40.42, |
|
"learning_rate": 9.746478873239436e-07, |
|
"loss": 0.5354, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 40.55, |
|
"learning_rate": 9.605633802816902e-07, |
|
"loss": 0.5444, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 40.69, |
|
"learning_rate": 9.464788732394367e-07, |
|
"loss": 0.5232, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 40.83, |
|
"learning_rate": 9.323943661971832e-07, |
|
"loss": 0.5378, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 40.97, |
|
"learning_rate": 9.183098591549297e-07, |
|
"loss": 0.5343, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 41.0, |
|
"eval_loss": 0.5179165601730347, |
|
"eval_runtime": 129.1365, |
|
"eval_samples_per_second": 4.12, |
|
"eval_steps_per_second": 0.519, |
|
"eval_wer": 0.2536478755544128, |
|
"step": 2952 |
|
}, |
|
{ |
|
"epoch": 41.11, |
|
"learning_rate": 9.04225352112676e-07, |
|
"loss": 0.5501, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 41.25, |
|
"learning_rate": 8.901408450704225e-07, |
|
"loss": 0.5334, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 41.39, |
|
"learning_rate": 8.76056338028169e-07, |
|
"loss": 0.5379, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 41.53, |
|
"learning_rate": 8.619718309859156e-07, |
|
"loss": 0.54, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 41.66, |
|
"learning_rate": 8.478873239436621e-07, |
|
"loss": 0.5446, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 41.8, |
|
"learning_rate": 8.338028169014086e-07, |
|
"loss": 0.5478, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 41.94, |
|
"learning_rate": 8.19718309859155e-07, |
|
"loss": 0.5255, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 42.0, |
|
"eval_loss": 0.5213496685028076, |
|
"eval_runtime": 130.2226, |
|
"eval_samples_per_second": 4.085, |
|
"eval_steps_per_second": 0.515, |
|
"eval_wer": 0.2560262261361445, |
|
"step": 3024 |
|
}, |
|
{ |
|
"epoch": 42.08, |
|
"learning_rate": 8.056338028169015e-07, |
|
"loss": 0.5457, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 42.22, |
|
"learning_rate": 7.915492957746479e-07, |
|
"loss": 0.5311, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 42.36, |
|
"learning_rate": 7.774647887323944e-07, |
|
"loss": 0.5355, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 42.5, |
|
"learning_rate": 7.633802816901409e-07, |
|
"loss": 0.5426, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 42.64, |
|
"learning_rate": 7.492957746478873e-07, |
|
"loss": 0.5349, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 42.78, |
|
"learning_rate": 7.352112676056339e-07, |
|
"loss": 0.5379, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 42.91, |
|
"learning_rate": 7.211267605633804e-07, |
|
"loss": 0.525, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 43.0, |
|
"eval_loss": 0.522071123123169, |
|
"eval_runtime": 128.4426, |
|
"eval_samples_per_second": 4.142, |
|
"eval_steps_per_second": 0.522, |
|
"eval_wer": 0.2553191489361702, |
|
"step": 3096 |
|
}, |
|
{ |
|
"epoch": 43.06, |
|
"learning_rate": 7.070422535211269e-07, |
|
"loss": 0.5591, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 43.19, |
|
"learning_rate": 6.929577464788733e-07, |
|
"loss": 0.5328, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 43.33, |
|
"learning_rate": 6.788732394366197e-07, |
|
"loss": 0.5427, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 43.47, |
|
"learning_rate": 6.647887323943662e-07, |
|
"loss": 0.5322, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 43.61, |
|
"learning_rate": 6.507042253521127e-07, |
|
"loss": 0.5293, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 43.75, |
|
"learning_rate": 6.366197183098592e-07, |
|
"loss": 0.5299, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 43.89, |
|
"learning_rate": 6.225352112676057e-07, |
|
"loss": 0.5345, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 44.0, |
|
"eval_loss": 0.5230171084403992, |
|
"eval_runtime": 129.6367, |
|
"eval_samples_per_second": 4.104, |
|
"eval_steps_per_second": 0.517, |
|
"eval_wer": 0.25306935784534296, |
|
"step": 3168 |
|
}, |
|
{ |
|
"epoch": 44.03, |
|
"learning_rate": 6.084507042253521e-07, |
|
"loss": 0.5468, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 44.17, |
|
"learning_rate": 5.943661971830986e-07, |
|
"loss": 0.5317, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 44.3, |
|
"learning_rate": 5.802816901408451e-07, |
|
"loss": 0.5357, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 44.44, |
|
"learning_rate": 5.661971830985916e-07, |
|
"loss": 0.5407, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 44.58, |
|
"learning_rate": 5.521126760563381e-07, |
|
"loss": 0.5388, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 44.72, |
|
"learning_rate": 5.380281690140846e-07, |
|
"loss": 0.5096, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 44.86, |
|
"learning_rate": 5.23943661971831e-07, |
|
"loss": 0.5349, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"learning_rate": 5.098591549295775e-07, |
|
"loss": 0.5485, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"eval_loss": 0.52115797996521, |
|
"eval_runtime": 128.2514, |
|
"eval_samples_per_second": 4.148, |
|
"eval_steps_per_second": 0.522, |
|
"eval_wer": 0.253712155299865, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 45.14, |
|
"learning_rate": 4.95774647887324e-07, |
|
"loss": 0.536, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 45.28, |
|
"learning_rate": 4.816901408450705e-07, |
|
"loss": 0.5245, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 45.42, |
|
"learning_rate": 4.6760563380281696e-07, |
|
"loss": 0.5197, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 45.55, |
|
"learning_rate": 4.535211267605634e-07, |
|
"loss": 0.5556, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 45.69, |
|
"learning_rate": 4.3943661971830987e-07, |
|
"loss": 0.5254, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 45.83, |
|
"learning_rate": 4.253521126760564e-07, |
|
"loss": 0.5375, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 45.97, |
|
"learning_rate": 4.1126760563380283e-07, |
|
"loss": 0.5471, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 46.0, |
|
"eval_loss": 0.5214569568634033, |
|
"eval_runtime": 128.944, |
|
"eval_samples_per_second": 4.126, |
|
"eval_steps_per_second": 0.52, |
|
"eval_wer": 0.25319791733624736, |
|
"step": 3312 |
|
}, |
|
{ |
|
"epoch": 46.11, |
|
"learning_rate": 3.971830985915493e-07, |
|
"loss": 0.5549, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 46.25, |
|
"learning_rate": 3.830985915492958e-07, |
|
"loss": 0.5379, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 46.39, |
|
"learning_rate": 3.690140845070423e-07, |
|
"loss": 0.521, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 46.53, |
|
"learning_rate": 3.5492957746478875e-07, |
|
"loss": 0.5369, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 46.66, |
|
"learning_rate": 3.4084507042253523e-07, |
|
"loss": 0.5292, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 46.8, |
|
"learning_rate": 3.267605633802817e-07, |
|
"loss": 0.5373, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 46.94, |
|
"learning_rate": 3.126760563380282e-07, |
|
"loss": 0.5375, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 47.0, |
|
"eval_loss": 0.5215560793876648, |
|
"eval_runtime": 128.748, |
|
"eval_samples_per_second": 4.132, |
|
"eval_steps_per_second": 0.52, |
|
"eval_wer": 0.2543549527543871, |
|
"step": 3384 |
|
}, |
|
{ |
|
"epoch": 47.08, |
|
"learning_rate": 2.985915492957747e-07, |
|
"loss": 0.5496, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 47.22, |
|
"learning_rate": 2.8450704225352116e-07, |
|
"loss": 0.5405, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 47.36, |
|
"learning_rate": 2.7042253521126764e-07, |
|
"loss": 0.5353, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 47.5, |
|
"learning_rate": 2.563380281690141e-07, |
|
"loss": 0.5449, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 47.64, |
|
"learning_rate": 2.422535211267606e-07, |
|
"loss": 0.5415, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 47.78, |
|
"learning_rate": 2.2816901408450706e-07, |
|
"loss": 0.5322, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 47.91, |
|
"learning_rate": 2.1408450704225354e-07, |
|
"loss": 0.5229, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 48.0, |
|
"eval_loss": 0.5208937525749207, |
|
"eval_runtime": 128.6533, |
|
"eval_samples_per_second": 4.135, |
|
"eval_steps_per_second": 0.521, |
|
"eval_wer": 0.2551263096998136, |
|
"step": 3456 |
|
}, |
|
{ |
|
"epoch": 48.06, |
|
"learning_rate": 2.0000000000000002e-07, |
|
"loss": 0.5553, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 48.19, |
|
"learning_rate": 1.8591549295774647e-07, |
|
"loss": 0.5291, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 48.33, |
|
"learning_rate": 1.7183098591549298e-07, |
|
"loss": 0.5476, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 48.47, |
|
"learning_rate": 1.5774647887323943e-07, |
|
"loss": 0.5228, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 48.61, |
|
"learning_rate": 1.4366197183098591e-07, |
|
"loss": 0.5252, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 48.75, |
|
"learning_rate": 1.295774647887324e-07, |
|
"loss": 0.5434, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 48.89, |
|
"learning_rate": 1.1549295774647888e-07, |
|
"loss": 0.5218, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 49.0, |
|
"eval_loss": 0.5215969085693359, |
|
"eval_runtime": 129.6533, |
|
"eval_samples_per_second": 4.103, |
|
"eval_steps_per_second": 0.517, |
|
"eval_wer": 0.2536478755544128, |
|
"step": 3528 |
|
}, |
|
{ |
|
"epoch": 49.03, |
|
"learning_rate": 1.0281690140845071e-07, |
|
"loss": 0.5535, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 49.17, |
|
"learning_rate": 8.873239436619719e-08, |
|
"loss": 0.529, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 49.3, |
|
"learning_rate": 7.464788732394367e-08, |
|
"loss": 0.5316, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 49.44, |
|
"learning_rate": 6.056338028169015e-08, |
|
"loss": 0.5191, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 49.58, |
|
"learning_rate": 4.647887323943662e-08, |
|
"loss": 0.5505, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 49.72, |
|
"learning_rate": 3.23943661971831e-08, |
|
"loss": 0.5328, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 49.86, |
|
"learning_rate": 1.830985915492958e-08, |
|
"loss": 0.5271, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"learning_rate": 4.225352112676057e-09, |
|
"loss": 0.5292, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"eval_loss": 0.5207646489143372, |
|
"eval_runtime": 128.939, |
|
"eval_samples_per_second": 4.126, |
|
"eval_steps_per_second": 0.52, |
|
"eval_wer": 0.25448351224529153, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"step": 3600, |
|
"total_flos": 1.3601359172236221e+20, |
|
"train_loss": 0.5537465457121531, |
|
"train_runtime": 95750.0723, |
|
"train_samples_per_second": 2.411, |
|
"train_steps_per_second": 0.038 |
|
} |
|
], |
|
"max_steps": 3600, |
|
"num_train_epochs": 50, |
|
"total_flos": 1.3601359172236221e+20, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|