|
{ |
|
"best_metric": 0.21432486176490784, |
|
"best_model_checkpoint": "ai-light-dance_singing2_ft_wav2vec2-large-xlsr-53-5gram-v4-1/checkpoint-1960", |
|
"epoch": 9.998662505572893, |
|
"global_step": 2800, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 3e-06, |
|
"loss": 0.2316, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 5.7000000000000005e-06, |
|
"loss": 0.2616, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.7e-06, |
|
"loss": 0.2223, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.1700000000000001e-05, |
|
"loss": 0.2346, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.47e-05, |
|
"loss": 0.2567, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.77e-05, |
|
"loss": 0.2465, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 2.07e-05, |
|
"loss": 0.2245, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 2.37e-05, |
|
"loss": 0.2238, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 2.6700000000000002e-05, |
|
"loss": 0.209, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 2.97e-05, |
|
"loss": 0.2728, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 2.9900000000000002e-05, |
|
"loss": 0.2248, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 2.978888888888889e-05, |
|
"loss": 0.224, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 2.9677777777777778e-05, |
|
"loss": 0.2273, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 2.9566666666666666e-05, |
|
"loss": 0.2335, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 2.9455555555555557e-05, |
|
"loss": 0.2913, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 2.9344444444444442e-05, |
|
"loss": 0.2403, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 2.9233333333333334e-05, |
|
"loss": 0.2676, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.9122222222222225e-05, |
|
"loss": 0.232, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.9022222222222223e-05, |
|
"loss": 0.2172, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.8911111111111114e-05, |
|
"loss": 0.2588, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 2.88e-05, |
|
"loss": 0.2297, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.868888888888889e-05, |
|
"loss": 0.2399, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 2.8577777777777778e-05, |
|
"loss": 0.2487, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.8466666666666666e-05, |
|
"loss": 0.2488, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.8355555555555558e-05, |
|
"loss": 0.346, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.8244444444444446e-05, |
|
"loss": 0.2245, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.8133333333333334e-05, |
|
"loss": 0.2521, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.8022222222222222e-05, |
|
"loss": 0.2609, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.23127533495426178, |
|
"eval_runtime": 229.0439, |
|
"eval_samples_per_second": 8.706, |
|
"eval_steps_per_second": 1.091, |
|
"eval_wer": 0.1375581006774615, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.7911111111111113e-05, |
|
"loss": 0.2157, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 2.78e-05, |
|
"loss": 0.1941, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 2.768888888888889e-05, |
|
"loss": 0.191, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 2.7577777777777778e-05, |
|
"loss": 0.193, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 2.746666666666667e-05, |
|
"loss": 0.2677, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 2.7355555555555554e-05, |
|
"loss": 0.2045, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 2.7244444444444445e-05, |
|
"loss": 0.2135, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 2.7133333333333337e-05, |
|
"loss": 0.1998, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 2.702222222222222e-05, |
|
"loss": 0.2092, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 2.6911111111111113e-05, |
|
"loss": 0.2336, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 2.68e-05, |
|
"loss": 0.2006, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 2.668888888888889e-05, |
|
"loss": 0.2057, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 2.6577777777777777e-05, |
|
"loss": 0.2161, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 2.646666666666667e-05, |
|
"loss": 0.2043, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 2.6355555555555557e-05, |
|
"loss": 0.2585, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 2.6244444444444445e-05, |
|
"loss": 0.2092, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 2.6133333333333333e-05, |
|
"loss": 0.2153, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 2.6022222222222224e-05, |
|
"loss": 0.2008, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 2.5911111111111112e-05, |
|
"loss": 0.211, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 2.58e-05, |
|
"loss": 0.2817, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 2.5688888888888892e-05, |
|
"loss": 0.1859, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.5577777777777776e-05, |
|
"loss": 0.2119, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 2.5466666666666668e-05, |
|
"loss": 0.2129, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 2.5355555555555556e-05, |
|
"loss": 0.2149, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 2.5244444444444444e-05, |
|
"loss": 0.2483, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 2.5133333333333332e-05, |
|
"loss": 0.1944, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 2.5022222222222224e-05, |
|
"loss": 0.2044, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 2.4911111111111115e-05, |
|
"loss": 0.2297, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.22399240732192993, |
|
"eval_runtime": 227.0775, |
|
"eval_samples_per_second": 8.781, |
|
"eval_steps_per_second": 1.101, |
|
"eval_wer": 0.13971504086034572, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 2.48e-05, |
|
"loss": 0.1975, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 2.468888888888889e-05, |
|
"loss": 0.2018, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 2.457777777777778e-05, |
|
"loss": 0.1865, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 2.4466666666666667e-05, |
|
"loss": 0.1752, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 2.4355555555555555e-05, |
|
"loss": 0.228, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 2.4244444444444447e-05, |
|
"loss": 0.1761, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 2.413333333333333e-05, |
|
"loss": 0.1735, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 2.4022222222222223e-05, |
|
"loss": 0.1717, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 2.391111111111111e-05, |
|
"loss": 0.1869, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 2.38e-05, |
|
"loss": 0.2187, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 2.368888888888889e-05, |
|
"loss": 0.1723, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 2.357777777777778e-05, |
|
"loss": 0.1709, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 2.346666666666667e-05, |
|
"loss": 0.187, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 2.3355555555555555e-05, |
|
"loss": 0.1776, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 2.3244444444444446e-05, |
|
"loss": 0.2241, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 2.3133333333333334e-05, |
|
"loss": 0.1707, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 2.3022222222222222e-05, |
|
"loss": 0.1744, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 2.291111111111111e-05, |
|
"loss": 0.1816, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.2800000000000002e-05, |
|
"loss": 0.161, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.2688888888888887e-05, |
|
"loss": 0.2259, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 2.2577777777777778e-05, |
|
"loss": 0.1744, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 2.246666666666667e-05, |
|
"loss": 0.1715, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 2.2355555555555554e-05, |
|
"loss": 0.181, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 2.2244444444444446e-05, |
|
"loss": 0.1685, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 2.2133333333333334e-05, |
|
"loss": 0.2364, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 2.2022222222222225e-05, |
|
"loss": 0.1696, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 2.191111111111111e-05, |
|
"loss": 0.1709, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 2.18e-05, |
|
"loss": 0.1951, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.22800597548484802, |
|
"eval_runtime": 227.1154, |
|
"eval_samples_per_second": 8.78, |
|
"eval_steps_per_second": 1.101, |
|
"eval_wer": 0.13613026703527054, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 2.168888888888889e-05, |
|
"loss": 0.1683, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 2.1577777777777777e-05, |
|
"loss": 0.1568, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 2.1466666666666666e-05, |
|
"loss": 0.1518, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 2.1355555555555557e-05, |
|
"loss": 0.1473, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 2.1244444444444445e-05, |
|
"loss": 0.2078, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 2.1133333333333333e-05, |
|
"loss": 0.1523, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 2.1022222222222225e-05, |
|
"loss": 0.1508, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 2.091111111111111e-05, |
|
"loss": 0.1533, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 2.08e-05, |
|
"loss": 0.1565, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 2.068888888888889e-05, |
|
"loss": 0.1926, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 2.057777777777778e-05, |
|
"loss": 0.1554, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 2.0466666666666665e-05, |
|
"loss": 0.149, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 2.0355555555555556e-05, |
|
"loss": 0.1495, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 2.0244444444444448e-05, |
|
"loss": 0.1648, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 2.0133333333333333e-05, |
|
"loss": 0.207, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 2.0022222222222224e-05, |
|
"loss": 0.165, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 1.9911111111111112e-05, |
|
"loss": 0.1531, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 1.98e-05, |
|
"loss": 0.1623, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 1.9688888888888888e-05, |
|
"loss": 0.1747, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 1.957777777777778e-05, |
|
"loss": 0.2086, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 1.9466666666666664e-05, |
|
"loss": 0.1519, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 1.9355555555555556e-05, |
|
"loss": 0.148, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 1.9244444444444444e-05, |
|
"loss": 0.1493, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 1.9133333333333335e-05, |
|
"loss": 0.159, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 1.9022222222222223e-05, |
|
"loss": 0.1949, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 1.891111111111111e-05, |
|
"loss": 0.1625, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 1.8800000000000003e-05, |
|
"loss": 0.1647, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 1.8688888888888888e-05, |
|
"loss": 0.1816, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 0.22146841883659363, |
|
"eval_runtime": 228.3052, |
|
"eval_samples_per_second": 8.734, |
|
"eval_steps_per_second": 1.095, |
|
"eval_wer": 0.12823161284442688, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 1.857777777777778e-05, |
|
"loss": 0.15, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 1.8466666666666667e-05, |
|
"loss": 0.1464, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"learning_rate": 1.8355555555555555e-05, |
|
"loss": 0.1358, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 1.8244444444444443e-05, |
|
"loss": 0.1377, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 4.18, |
|
"learning_rate": 1.8133333333333335e-05, |
|
"loss": 0.1948, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 1.802222222222222e-05, |
|
"loss": 0.1356, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 1.791111111111111e-05, |
|
"loss": 0.1405, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 4.29, |
|
"learning_rate": 1.7800000000000002e-05, |
|
"loss": 0.1448, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 4.32, |
|
"learning_rate": 1.768888888888889e-05, |
|
"loss": 0.1386, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 1.757777777777778e-05, |
|
"loss": 0.1811, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"learning_rate": 1.7466666666666667e-05, |
|
"loss": 0.1312, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 4.43, |
|
"learning_rate": 1.7355555555555558e-05, |
|
"loss": 0.1325, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 1.7244444444444443e-05, |
|
"loss": 0.1331, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"learning_rate": 1.7133333333333334e-05, |
|
"loss": 0.1412, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 1.7022222222222222e-05, |
|
"loss": 0.1777, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 1.691111111111111e-05, |
|
"loss": 0.1395, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 1.6800000000000002e-05, |
|
"loss": 0.1472, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 1.668888888888889e-05, |
|
"loss": 0.1491, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 1.6577777777777778e-05, |
|
"loss": 0.1387, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 4.71, |
|
"learning_rate": 1.6466666666666666e-05, |
|
"loss": 0.1861, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 1.6355555555555557e-05, |
|
"loss": 0.1318, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 4.78, |
|
"learning_rate": 1.6244444444444446e-05, |
|
"loss": 0.1521, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"learning_rate": 1.6133333333333334e-05, |
|
"loss": 0.1331, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 1.602222222222222e-05, |
|
"loss": 0.1474, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 1.5911111111111113e-05, |
|
"loss": 0.1687, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 1.5799999999999998e-05, |
|
"loss": 0.1391, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 1.568888888888889e-05, |
|
"loss": 0.1325, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 1.557777777777778e-05, |
|
"loss": 0.1634, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 0.21799303591251373, |
|
"eval_runtime": 229.2501, |
|
"eval_samples_per_second": 8.698, |
|
"eval_steps_per_second": 1.091, |
|
"eval_wer": 0.12397849135704955, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 1.5466666666666665e-05, |
|
"loss": 0.1304, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"learning_rate": 1.5355555555555557e-05, |
|
"loss": 0.1323, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 5.11, |
|
"learning_rate": 1.5244444444444445e-05, |
|
"loss": 0.1232, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 1.5133333333333335e-05, |
|
"loss": 0.1254, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 5.18, |
|
"learning_rate": 1.5022222222222221e-05, |
|
"loss": 0.149, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"learning_rate": 1.4911111111111113e-05, |
|
"loss": 0.1146, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 5.25, |
|
"learning_rate": 1.48e-05, |
|
"loss": 0.1235, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 5.29, |
|
"learning_rate": 1.4688888888888889e-05, |
|
"loss": 0.1352, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 5.32, |
|
"learning_rate": 1.4577777777777778e-05, |
|
"loss": 0.1492, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 5.36, |
|
"learning_rate": 1.4466666666666667e-05, |
|
"loss": 0.1664, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 5.39, |
|
"learning_rate": 1.4355555555555556e-05, |
|
"loss": 0.1277, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 5.43, |
|
"learning_rate": 1.4244444444444444e-05, |
|
"loss": 0.1562, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 5.46, |
|
"learning_rate": 1.4133333333333332e-05, |
|
"loss": 0.1252, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"learning_rate": 1.4022222222222222e-05, |
|
"loss": 0.1244, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"learning_rate": 1.3911111111111112e-05, |
|
"loss": 0.1595, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 5.57, |
|
"learning_rate": 1.3800000000000002e-05, |
|
"loss": 0.1239, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 5.61, |
|
"learning_rate": 1.368888888888889e-05, |
|
"loss": 0.1336, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 5.64, |
|
"learning_rate": 1.3577777777777778e-05, |
|
"loss": 0.1298, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 5.68, |
|
"learning_rate": 1.3466666666666668e-05, |
|
"loss": 0.1151, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 5.71, |
|
"learning_rate": 1.3355555555555556e-05, |
|
"loss": 0.1664, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 5.75, |
|
"learning_rate": 1.3244444444444444e-05, |
|
"loss": 0.1318, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 5.78, |
|
"learning_rate": 1.3133333333333334e-05, |
|
"loss": 0.1269, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 1.3022222222222222e-05, |
|
"loss": 0.1271, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 5.86, |
|
"learning_rate": 1.2911111111111111e-05, |
|
"loss": 0.1254, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 5.89, |
|
"learning_rate": 1.2800000000000001e-05, |
|
"loss": 0.1712, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 5.93, |
|
"learning_rate": 1.268888888888889e-05, |
|
"loss": 0.1386, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 5.96, |
|
"learning_rate": 1.2577777777777779e-05, |
|
"loss": 0.1192, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 1.2466666666666667e-05, |
|
"loss": 0.1338, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 0.222590833902359, |
|
"eval_runtime": 226.2306, |
|
"eval_samples_per_second": 8.814, |
|
"eval_steps_per_second": 1.105, |
|
"eval_wer": 0.12406962967463621, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 1.2355555555555557e-05, |
|
"loss": 0.1161, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"learning_rate": 1.2244444444444445e-05, |
|
"loss": 0.1118, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 6.11, |
|
"learning_rate": 1.2133333333333333e-05, |
|
"loss": 0.1105, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 6.14, |
|
"learning_rate": 1.2022222222222223e-05, |
|
"loss": 0.1132, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 6.18, |
|
"learning_rate": 1.191111111111111e-05, |
|
"loss": 0.1466, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 6.21, |
|
"learning_rate": 1.1799999999999999e-05, |
|
"loss": 0.1097, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 1.1688888888888889e-05, |
|
"loss": 0.1105, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 6.29, |
|
"learning_rate": 1.1577777777777778e-05, |
|
"loss": 0.1051, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 6.32, |
|
"learning_rate": 1.1466666666666668e-05, |
|
"loss": 0.1101, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 6.36, |
|
"learning_rate": 1.1355555555555556e-05, |
|
"loss": 0.1425, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 6.39, |
|
"learning_rate": 1.1244444444444444e-05, |
|
"loss": 0.1096, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 6.43, |
|
"learning_rate": 1.1133333333333334e-05, |
|
"loss": 0.1161, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 6.46, |
|
"learning_rate": 1.1022222222222222e-05, |
|
"loss": 0.1147, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"learning_rate": 1.0911111111111112e-05, |
|
"loss": 0.1159, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 1.08e-05, |
|
"loss": 0.1622, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 6.57, |
|
"learning_rate": 1.0688888888888888e-05, |
|
"loss": 0.116, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 6.61, |
|
"learning_rate": 1.0577777777777778e-05, |
|
"loss": 0.1077, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 6.64, |
|
"learning_rate": 1.0466666666666668e-05, |
|
"loss": 0.1114, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 6.68, |
|
"learning_rate": 1.0355555555555556e-05, |
|
"loss": 0.1118, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 6.71, |
|
"learning_rate": 1.0244444444444445e-05, |
|
"loss": 0.1554, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 6.75, |
|
"learning_rate": 1.0133333333333333e-05, |
|
"loss": 0.1217, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 6.78, |
|
"learning_rate": 1.0022222222222223e-05, |
|
"loss": 0.1124, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 6.82, |
|
"learning_rate": 9.911111111111111e-06, |
|
"loss": 0.1149, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 6.86, |
|
"learning_rate": 9.8e-06, |
|
"loss": 0.1197, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 6.89, |
|
"learning_rate": 9.688888888888889e-06, |
|
"loss": 0.1492, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 6.93, |
|
"learning_rate": 9.577777777777777e-06, |
|
"loss": 0.1187, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 6.96, |
|
"learning_rate": 9.466666666666667e-06, |
|
"loss": 0.1254, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 9.355555555555555e-06, |
|
"loss": 0.1411, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 0.21432486176490784, |
|
"eval_runtime": 225.8473, |
|
"eval_samples_per_second": 8.829, |
|
"eval_steps_per_second": 1.107, |
|
"eval_wer": 0.12109244463347206, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 9.244444444444445e-06, |
|
"loss": 0.1057, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 7.07, |
|
"learning_rate": 9.133333333333335e-06, |
|
"loss": 0.1064, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 7.11, |
|
"learning_rate": 9.022222222222223e-06, |
|
"loss": 0.1072, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 7.14, |
|
"learning_rate": 8.911111111111112e-06, |
|
"loss": 0.103, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 7.18, |
|
"learning_rate": 8.8e-06, |
|
"loss": 0.151, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 7.21, |
|
"learning_rate": 8.688888888888889e-06, |
|
"loss": 0.1021, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 7.25, |
|
"learning_rate": 8.577777777777778e-06, |
|
"loss": 0.097, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 7.29, |
|
"learning_rate": 8.466666666666666e-06, |
|
"loss": 0.1157, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 7.32, |
|
"learning_rate": 8.355555555555554e-06, |
|
"loss": 0.1021, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 7.36, |
|
"learning_rate": 8.244444444444444e-06, |
|
"loss": 0.1403, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 7.39, |
|
"learning_rate": 8.133333333333334e-06, |
|
"loss": 0.1057, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 7.43, |
|
"learning_rate": 8.022222222222224e-06, |
|
"loss": 0.1137, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 7.46, |
|
"learning_rate": 7.911111111111112e-06, |
|
"loss": 0.1129, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 7.8e-06, |
|
"loss": 0.108, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 7.53, |
|
"learning_rate": 7.68888888888889e-06, |
|
"loss": 0.1322, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 7.57, |
|
"learning_rate": 7.577777777777778e-06, |
|
"loss": 0.1021, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 7.61, |
|
"learning_rate": 7.466666666666667e-06, |
|
"loss": 0.104, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 7.64, |
|
"learning_rate": 7.3555555555555555e-06, |
|
"loss": 0.1065, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 7.68, |
|
"learning_rate": 7.244444444444445e-06, |
|
"loss": 0.0996, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 7.71, |
|
"learning_rate": 7.133333333333333e-06, |
|
"loss": 0.1334, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 7.75, |
|
"learning_rate": 7.022222222222222e-06, |
|
"loss": 0.1045, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 7.78, |
|
"learning_rate": 6.911111111111111e-06, |
|
"loss": 0.1156, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 7.82, |
|
"learning_rate": 6.8e-06, |
|
"loss": 0.1101, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 7.86, |
|
"learning_rate": 6.688888888888889e-06, |
|
"loss": 0.1093, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 7.89, |
|
"learning_rate": 6.577777777777778e-06, |
|
"loss": 0.1501, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 7.93, |
|
"learning_rate": 6.466666666666667e-06, |
|
"loss": 0.1033, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 7.96, |
|
"learning_rate": 6.355555555555556e-06, |
|
"loss": 0.0964, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 6.244444444444445e-06, |
|
"loss": 0.1143, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 0.2181498110294342, |
|
"eval_runtime": 226.8788, |
|
"eval_samples_per_second": 8.789, |
|
"eval_steps_per_second": 1.102, |
|
"eval_wer": 0.11738615305161466, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 6.133333333333334e-06, |
|
"loss": 0.1048, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 8.07, |
|
"learning_rate": 6.0222222222222225e-06, |
|
"loss": 0.0961, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 8.11, |
|
"learning_rate": 5.922222222222223e-06, |
|
"loss": 0.1179, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 8.14, |
|
"learning_rate": 5.8111111111111116e-06, |
|
"loss": 0.1054, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 8.18, |
|
"learning_rate": 5.7000000000000005e-06, |
|
"loss": 0.1232, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 8.21, |
|
"learning_rate": 5.5888888888888886e-06, |
|
"loss": 0.0988, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 8.25, |
|
"learning_rate": 5.4777777777777775e-06, |
|
"loss": 0.0887, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 8.29, |
|
"learning_rate": 5.366666666666666e-06, |
|
"loss": 0.0884, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 8.32, |
|
"learning_rate": 5.255555555555556e-06, |
|
"loss": 0.1029, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 8.36, |
|
"learning_rate": 5.144444444444445e-06, |
|
"loss": 0.1396, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 8.39, |
|
"learning_rate": 5.033333333333333e-06, |
|
"loss": 0.088, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 8.43, |
|
"learning_rate": 4.922222222222222e-06, |
|
"loss": 0.1014, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 8.46, |
|
"learning_rate": 4.811111111111111e-06, |
|
"loss": 0.1011, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"learning_rate": 4.700000000000001e-06, |
|
"loss": 0.098, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 8.53, |
|
"learning_rate": 4.588888888888889e-06, |
|
"loss": 0.1367, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 8.57, |
|
"learning_rate": 4.477777777777778e-06, |
|
"loss": 0.0964, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 8.61, |
|
"learning_rate": 4.366666666666667e-06, |
|
"loss": 0.0978, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 8.64, |
|
"learning_rate": 4.2555555555555556e-06, |
|
"loss": 0.0931, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 8.68, |
|
"learning_rate": 4.1444444444444445e-06, |
|
"loss": 0.1048, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 8.71, |
|
"learning_rate": 4.033333333333333e-06, |
|
"loss": 0.1188, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"learning_rate": 3.922222222222222e-06, |
|
"loss": 0.0968, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 8.78, |
|
"learning_rate": 3.8111111111111112e-06, |
|
"loss": 0.1026, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 8.82, |
|
"learning_rate": 3.7e-06, |
|
"loss": 0.0963, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 8.86, |
|
"learning_rate": 3.588888888888889e-06, |
|
"loss": 0.0951, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 8.89, |
|
"learning_rate": 3.477777777777778e-06, |
|
"loss": 0.1337, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 8.93, |
|
"learning_rate": 3.366666666666667e-06, |
|
"loss": 0.0955, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 8.96, |
|
"learning_rate": 3.2555555555555554e-06, |
|
"loss": 0.101, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 3.1444444444444447e-06, |
|
"loss": 0.1127, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 0.2215089648962021, |
|
"eval_runtime": 228.2548, |
|
"eval_samples_per_second": 8.736, |
|
"eval_steps_per_second": 1.095, |
|
"eval_wer": 0.1166570465109214, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 3.0333333333333332e-06, |
|
"loss": 0.1108, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 9.07, |
|
"learning_rate": 2.922222222222222e-06, |
|
"loss": 0.0895, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 9.11, |
|
"learning_rate": 2.811111111111111e-06, |
|
"loss": 0.0901, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 9.14, |
|
"learning_rate": 2.7e-06, |
|
"loss": 0.0921, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 9.18, |
|
"learning_rate": 2.588888888888889e-06, |
|
"loss": 0.1313, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 9.21, |
|
"learning_rate": 2.477777777777778e-06, |
|
"loss": 0.0962, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 9.25, |
|
"learning_rate": 2.3666666666666667e-06, |
|
"loss": 0.0891, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 9.29, |
|
"learning_rate": 2.2555555555555557e-06, |
|
"loss": 0.1005, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 9.32, |
|
"learning_rate": 2.1444444444444446e-06, |
|
"loss": 0.1062, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 9.36, |
|
"learning_rate": 2.0333333333333335e-06, |
|
"loss": 0.1204, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 9.39, |
|
"learning_rate": 1.9222222222222224e-06, |
|
"loss": 0.0982, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 9.43, |
|
"learning_rate": 1.8111111111111113e-06, |
|
"loss": 0.0996, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 9.46, |
|
"learning_rate": 1.7e-06, |
|
"loss": 0.0861, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"learning_rate": 1.588888888888889e-06, |
|
"loss": 0.0915, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 9.53, |
|
"learning_rate": 1.4777777777777779e-06, |
|
"loss": 0.1318, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"learning_rate": 1.3666666666666668e-06, |
|
"loss": 0.0862, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 9.61, |
|
"learning_rate": 1.2555555555555555e-06, |
|
"loss": 0.1022, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 9.64, |
|
"learning_rate": 1.1444444444444444e-06, |
|
"loss": 0.0935, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 9.68, |
|
"learning_rate": 1.0333333333333333e-06, |
|
"loss": 0.0979, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 9.71, |
|
"learning_rate": 9.222222222222222e-07, |
|
"loss": 0.1253, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 9.75, |
|
"learning_rate": 8.111111111111112e-07, |
|
"loss": 0.0903, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 9.78, |
|
"learning_rate": 7.000000000000001e-07, |
|
"loss": 0.0983, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 9.82, |
|
"learning_rate": 5.888888888888889e-07, |
|
"loss": 0.0989, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"learning_rate": 4.777777777777778e-07, |
|
"loss": 0.0955, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 9.89, |
|
"learning_rate": 3.666666666666667e-07, |
|
"loss": 0.1178, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 9.93, |
|
"learning_rate": 2.555555555555556e-07, |
|
"loss": 0.0863, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 9.96, |
|
"learning_rate": 1.4444444444444445e-07, |
|
"loss": 0.0966, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 3.3333333333333334e-08, |
|
"loss": 0.105, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 0.2196090817451477, |
|
"eval_runtime": 231.6237, |
|
"eval_samples_per_second": 8.609, |
|
"eval_steps_per_second": 1.079, |
|
"eval_wer": 0.11595831940942371, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"step": 2800, |
|
"total_flos": 5.435692193960383e+19, |
|
"train_loss": 0.15355599720563207, |
|
"train_runtime": 33608.852, |
|
"train_samples_per_second": 5.337, |
|
"train_steps_per_second": 0.083 |
|
} |
|
], |
|
"max_steps": 2800, |
|
"num_train_epochs": 10, |
|
"total_flos": 5.435692193960383e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|