|
{ |
|
"best_metric": 1.076346755027771, |
|
"best_model_checkpoint": "ai-light-dance_stepmania_ft_wav2vec2-large-xlsr-53-v1/checkpoint-1350", |
|
"epoch": 9.997091815538013, |
|
"global_step": 1500, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 1.2075, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 1.1818, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.16e-05, |
|
"loss": 1.1731, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.5600000000000003e-05, |
|
"loss": 1.1906, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.9600000000000002e-05, |
|
"loss": 1.1936, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 2.36e-05, |
|
"loss": 1.1728, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 2.76e-05, |
|
"loss": 1.1748, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 3.16e-05, |
|
"loss": 1.1953, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.5600000000000005e-05, |
|
"loss": 1.1795, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.96e-05, |
|
"loss": 1.1763, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.9742857142857146e-05, |
|
"loss": 1.1647, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.945714285714286e-05, |
|
"loss": 1.203, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 3.917142857142858e-05, |
|
"loss": 1.164, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 3.8885714285714294e-05, |
|
"loss": 1.185, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 3.86e-05, |
|
"loss": 1.1632, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 1.2007131576538086, |
|
"eval_runtime": 347.7999, |
|
"eval_samples_per_second": 7.688, |
|
"eval_steps_per_second": 0.771, |
|
"eval_wer": 0.9874654499394933, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 3.831428571428572e-05, |
|
"loss": 1.239, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 3.8028571428571435e-05, |
|
"loss": 1.1681, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 3.7742857142857145e-05, |
|
"loss": 1.1626, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 3.745714285714286e-05, |
|
"loss": 1.1778, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 3.717142857142858e-05, |
|
"loss": 1.1747, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 3.6885714285714286e-05, |
|
"loss": 1.1547, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 3.66e-05, |
|
"loss": 1.1504, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 3.631428571428572e-05, |
|
"loss": 1.1893, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 3.602857142857143e-05, |
|
"loss": 1.1549, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 3.574285714285714e-05, |
|
"loss": 1.1574, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 3.545714285714286e-05, |
|
"loss": 1.1671, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.5171428571428575e-05, |
|
"loss": 1.1832, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 3.4885714285714285e-05, |
|
"loss": 1.1662, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 3.46e-05, |
|
"loss": 1.175, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 3.431428571428572e-05, |
|
"loss": 1.1615, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 1.1911734342575073, |
|
"eval_runtime": 346.7276, |
|
"eval_samples_per_second": 7.712, |
|
"eval_steps_per_second": 0.773, |
|
"eval_wer": 0.9874701586359848, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 3.402857142857143e-05, |
|
"loss": 1.2418, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 3.374285714285714e-05, |
|
"loss": 1.1563, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 3.345714285714286e-05, |
|
"loss": 1.1407, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 3.3171428571428574e-05, |
|
"loss": 1.1596, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 3.288571428571429e-05, |
|
"loss": 1.1629, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 3.26e-05, |
|
"loss": 1.1473, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 3.2314285714285716e-05, |
|
"loss": 1.1506, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 3.202857142857143e-05, |
|
"loss": 1.1651, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 3.174285714285715e-05, |
|
"loss": 1.1321, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 3.1457142857142864e-05, |
|
"loss": 1.1463, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 3.117142857142857e-05, |
|
"loss": 1.1402, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 3.088571428571429e-05, |
|
"loss": 1.1679, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 3.0600000000000005e-05, |
|
"loss": 1.1486, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 3.0314285714285718e-05, |
|
"loss": 1.1389, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 3.002857142857143e-05, |
|
"loss": 1.1487, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 1.1941583156585693, |
|
"eval_runtime": 346.4981, |
|
"eval_samples_per_second": 7.717, |
|
"eval_steps_per_second": 0.773, |
|
"eval_wer": 0.9874701586359848, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 2.9742857142857143e-05, |
|
"loss": 1.2107, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 2.945714285714286e-05, |
|
"loss": 1.1357, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 2.9171428571428575e-05, |
|
"loss": 1.1376, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 2.888571428571429e-05, |
|
"loss": 1.1519, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 2.86e-05, |
|
"loss": 1.1355, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 2.8314285714285717e-05, |
|
"loss": 1.133, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 2.8028571428571433e-05, |
|
"loss": 1.1274, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 2.7742857142857145e-05, |
|
"loss": 1.1595, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 2.7457142857142858e-05, |
|
"loss": 1.1319, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 2.7171428571428574e-05, |
|
"loss": 1.1248, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 2.6885714285714287e-05, |
|
"loss": 1.1312, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 2.6600000000000003e-05, |
|
"loss": 1.1549, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 2.6314285714285715e-05, |
|
"loss": 1.1301, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 2.6028571428571428e-05, |
|
"loss": 1.1261, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 2.5742857142857144e-05, |
|
"loss": 1.1207, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 1.1753019094467163, |
|
"eval_runtime": 353.8634, |
|
"eval_samples_per_second": 7.557, |
|
"eval_steps_per_second": 0.757, |
|
"eval_wer": 0.9874701586359848, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 2.545714285714286e-05, |
|
"loss": 1.185, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"learning_rate": 2.5171428571428576e-05, |
|
"loss": 1.1175, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"learning_rate": 2.491428571428572e-05, |
|
"loss": 1.1091, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"learning_rate": 2.462857142857143e-05, |
|
"loss": 1.1214, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 2.4342857142857144e-05, |
|
"loss": 1.1384, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 2.405714285714286e-05, |
|
"loss": 1.1156, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 2.3771428571428573e-05, |
|
"loss": 1.104, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 2.348571428571429e-05, |
|
"loss": 1.1295, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 2.32e-05, |
|
"loss": 1.1031, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 2.2914285714285714e-05, |
|
"loss": 1.1016, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 2.262857142857143e-05, |
|
"loss": 1.0909, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"learning_rate": 2.2342857142857146e-05, |
|
"loss": 1.1106, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 2.2057142857142856e-05, |
|
"loss": 1.0826, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 2.177142857142857e-05, |
|
"loss": 1.0738, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 2.1485714285714288e-05, |
|
"loss": 1.0638, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 1.134511947631836, |
|
"eval_runtime": 349.9878, |
|
"eval_samples_per_second": 7.64, |
|
"eval_steps_per_second": 0.766, |
|
"eval_wer": 0.8213991420754992, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"learning_rate": 2.1200000000000004e-05, |
|
"loss": 1.1294, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 5.13, |
|
"learning_rate": 2.091428571428572e-05, |
|
"loss": 1.0545, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 5.2, |
|
"learning_rate": 2.062857142857143e-05, |
|
"loss": 1.0526, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 5.27, |
|
"learning_rate": 2.0342857142857145e-05, |
|
"loss": 1.0592, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 5.33, |
|
"learning_rate": 2.005714285714286e-05, |
|
"loss": 1.0525, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 5.4, |
|
"learning_rate": 1.9771428571428574e-05, |
|
"loss": 1.0453, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 5.47, |
|
"learning_rate": 1.9485714285714286e-05, |
|
"loss": 1.0356, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"learning_rate": 1.9200000000000003e-05, |
|
"loss": 1.0539, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 5.6, |
|
"learning_rate": 1.8914285714285715e-05, |
|
"loss": 1.0441, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 5.66, |
|
"learning_rate": 1.8628571428571428e-05, |
|
"loss": 1.0297, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 5.73, |
|
"learning_rate": 1.8342857142857144e-05, |
|
"loss": 1.0192, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"learning_rate": 1.8057142857142857e-05, |
|
"loss": 1.0391, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 5.86, |
|
"learning_rate": 1.7771428571428573e-05, |
|
"loss": 1.017, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 5.93, |
|
"learning_rate": 1.748571428571429e-05, |
|
"loss": 1.0152, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 1.72e-05, |
|
"loss": 1.0174, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 1.154058814048767, |
|
"eval_runtime": 347.8027, |
|
"eval_samples_per_second": 7.688, |
|
"eval_steps_per_second": 0.771, |
|
"eval_wer": 0.7665004496805149, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"learning_rate": 1.6914285714285717e-05, |
|
"loss": 1.0812, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 6.13, |
|
"learning_rate": 1.662857142857143e-05, |
|
"loss": 1.0024, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 6.2, |
|
"learning_rate": 1.6342857142857146e-05, |
|
"loss": 1.006, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 6.27, |
|
"learning_rate": 1.605714285714286e-05, |
|
"loss": 0.9978, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 6.33, |
|
"learning_rate": 1.577142857142857e-05, |
|
"loss": 1.0089, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 6.4, |
|
"learning_rate": 1.5485714285714287e-05, |
|
"loss": 0.9975, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 6.47, |
|
"learning_rate": 1.5200000000000002e-05, |
|
"loss": 0.993, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 1.4914285714285715e-05, |
|
"loss": 1.0142, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 1.462857142857143e-05, |
|
"loss": 0.9909, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 6.66, |
|
"learning_rate": 1.4342857142857145e-05, |
|
"loss": 0.9961, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 6.73, |
|
"learning_rate": 1.4057142857142858e-05, |
|
"loss": 0.9873, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 6.8, |
|
"learning_rate": 1.3771428571428574e-05, |
|
"loss": 1.0009, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 6.86, |
|
"learning_rate": 1.3485714285714286e-05, |
|
"loss": 0.9802, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 6.93, |
|
"learning_rate": 1.3200000000000002e-05, |
|
"loss": 0.9912, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 1.2914285714285715e-05, |
|
"loss": 0.9946, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 1.079897403717041, |
|
"eval_runtime": 353.1403, |
|
"eval_samples_per_second": 7.572, |
|
"eval_steps_per_second": 0.759, |
|
"eval_wer": 0.7716470549457793, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 7.07, |
|
"learning_rate": 1.262857142857143e-05, |
|
"loss": 1.0446, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 7.13, |
|
"learning_rate": 1.2342857142857144e-05, |
|
"loss": 0.9733, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 7.2, |
|
"learning_rate": 1.2057142857142858e-05, |
|
"loss": 0.9724, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 7.27, |
|
"learning_rate": 1.177142857142857e-05, |
|
"loss": 0.9857, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 7.33, |
|
"learning_rate": 1.1485714285714287e-05, |
|
"loss": 0.99, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 7.4, |
|
"learning_rate": 1.1200000000000001e-05, |
|
"loss": 0.9732, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 7.47, |
|
"learning_rate": 1.0914285714285716e-05, |
|
"loss": 0.9712, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 7.53, |
|
"learning_rate": 1.062857142857143e-05, |
|
"loss": 0.9932, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 7.6, |
|
"learning_rate": 1.0342857142857143e-05, |
|
"loss": 0.9754, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 7.66, |
|
"learning_rate": 1.0057142857142859e-05, |
|
"loss": 0.9697, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 7.73, |
|
"learning_rate": 9.771428571428571e-06, |
|
"loss": 0.9716, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 7.8, |
|
"learning_rate": 9.485714285714287e-06, |
|
"loss": 0.9877, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 7.86, |
|
"learning_rate": 9.200000000000002e-06, |
|
"loss": 0.9635, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 7.93, |
|
"learning_rate": 8.914285714285716e-06, |
|
"loss": 0.9608, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 8.628571428571429e-06, |
|
"loss": 0.9694, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 1.0847891569137573, |
|
"eval_runtime": 351.0567, |
|
"eval_samples_per_second": 7.617, |
|
"eval_steps_per_second": 0.763, |
|
"eval_wer": 0.7417656670104015, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 8.07, |
|
"learning_rate": 8.342857142857143e-06, |
|
"loss": 1.0234, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 8.13, |
|
"learning_rate": 8.057142857142857e-06, |
|
"loss": 0.9587, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 8.2, |
|
"learning_rate": 7.771428571428572e-06, |
|
"loss": 0.9652, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 8.27, |
|
"learning_rate": 7.485714285714286e-06, |
|
"loss": 0.9725, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 8.33, |
|
"learning_rate": 7.2000000000000005e-06, |
|
"loss": 0.9743, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 8.4, |
|
"learning_rate": 6.914285714285715e-06, |
|
"loss": 0.9559, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 8.47, |
|
"learning_rate": 6.628571428571428e-06, |
|
"loss": 0.9502, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 8.53, |
|
"learning_rate": 6.342857142857143e-06, |
|
"loss": 0.9843, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 8.6, |
|
"learning_rate": 6.057142857142858e-06, |
|
"loss": 0.9639, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 8.66, |
|
"learning_rate": 5.771428571428572e-06, |
|
"loss": 0.9642, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 8.73, |
|
"learning_rate": 5.485714285714287e-06, |
|
"loss": 0.9511, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 8.8, |
|
"learning_rate": 5.2e-06, |
|
"loss": 0.9808, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 8.86, |
|
"learning_rate": 4.9142857142857145e-06, |
|
"loss": 0.9569, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 8.93, |
|
"learning_rate": 4.628571428571429e-06, |
|
"loss": 0.9536, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 4.342857142857143e-06, |
|
"loss": 0.9566, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 1.076346755027771, |
|
"eval_runtime": 348.8389, |
|
"eval_samples_per_second": 7.665, |
|
"eval_steps_per_second": 0.768, |
|
"eval_wer": 0.7344389352695493, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 9.07, |
|
"learning_rate": 4.057142857142858e-06, |
|
"loss": 1.0231, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 9.13, |
|
"learning_rate": 3.771428571428572e-06, |
|
"loss": 0.9566, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 9.2, |
|
"learning_rate": 3.4857142857142863e-06, |
|
"loss": 0.9549, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 9.27, |
|
"learning_rate": 3.2000000000000003e-06, |
|
"loss": 0.9602, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 9.33, |
|
"learning_rate": 2.9142857142857146e-06, |
|
"loss": 0.9586, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 9.4, |
|
"learning_rate": 2.6285714285714286e-06, |
|
"loss": 0.9533, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 9.47, |
|
"learning_rate": 2.342857142857143e-06, |
|
"loss": 0.9522, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 9.53, |
|
"learning_rate": 2.0571428571428573e-06, |
|
"loss": 0.9786, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 9.6, |
|
"learning_rate": 1.7714285714285714e-06, |
|
"loss": 0.9473, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 9.66, |
|
"learning_rate": 1.4857142857142858e-06, |
|
"loss": 0.9531, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 9.73, |
|
"learning_rate": 1.2000000000000002e-06, |
|
"loss": 0.9488, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 9.8, |
|
"learning_rate": 9.142857142857144e-07, |
|
"loss": 0.9673, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"learning_rate": 6.285714285714287e-07, |
|
"loss": 0.9485, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 9.93, |
|
"learning_rate": 3.428571428571429e-07, |
|
"loss": 0.9524, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 5.714285714285715e-08, |
|
"loss": 0.9466, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 1.0791033506393433, |
|
"eval_runtime": 348.3672, |
|
"eval_samples_per_second": 7.676, |
|
"eval_steps_per_second": 0.769, |
|
"eval_wer": 0.7239809203618163, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"step": 1500, |
|
"total_flos": 7.288767136199884e+19, |
|
"train_loss": 1.0715523401896159, |
|
"train_runtime": 76023.7044, |
|
"train_samples_per_second": 3.165, |
|
"train_steps_per_second": 0.02 |
|
} |
|
], |
|
"max_steps": 1500, |
|
"num_train_epochs": 10, |
|
"total_flos": 7.288767136199884e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|