|
{ |
|
"best_metric": 23.029642758418838, |
|
"best_model_checkpoint": "./whisper-small-fa-aug/checkpoint-30000", |
|
"epoch": 18.01205, |
|
"global_step": 32000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 4.2000000000000006e-07, |
|
"loss": 3.8, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 9.200000000000001e-07, |
|
"loss": 3.0497, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.42e-06, |
|
"loss": 2.0102, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.9200000000000003e-06, |
|
"loss": 1.5948, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.42e-06, |
|
"loss": 1.4094, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.92e-06, |
|
"loss": 1.2611, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 3.4200000000000007e-06, |
|
"loss": 1.1366, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 3.920000000000001e-06, |
|
"loss": 1.0817, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.42e-06, |
|
"loss": 1.0324, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.92e-06, |
|
"loss": 0.9176, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 5.420000000000001e-06, |
|
"loss": 0.7975, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 5.92e-06, |
|
"loss": 0.7079, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 6.42e-06, |
|
"loss": 0.5832, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 6.92e-06, |
|
"loss": 0.4469, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 7.420000000000001e-06, |
|
"loss": 0.3913, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 7.92e-06, |
|
"loss": 0.4189, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.42e-06, |
|
"loss": 0.3756, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.920000000000001e-06, |
|
"loss": 0.3876, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.42e-06, |
|
"loss": 0.3802, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.920000000000002e-06, |
|
"loss": 0.4061, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.994683544303798e-06, |
|
"loss": 0.3885, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.988354430379748e-06, |
|
"loss": 0.3449, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.982025316455697e-06, |
|
"loss": 0.3292, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.975696202531647e-06, |
|
"loss": 0.3439, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.969367088607596e-06, |
|
"loss": 0.3031, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.963037974683545e-06, |
|
"loss": 0.3159, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.956708860759495e-06, |
|
"loss": 0.3304, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.950379746835444e-06, |
|
"loss": 0.2918, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.944050632911392e-06, |
|
"loss": 0.2785, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.937721518987341e-06, |
|
"loss": 0.2618, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.931392405063291e-06, |
|
"loss": 0.2862, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.92506329113924e-06, |
|
"loss": 0.2962, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.91873417721519e-06, |
|
"loss": 0.2935, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.912405063291141e-06, |
|
"loss": 0.2827, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.90607594936709e-06, |
|
"loss": 0.2702, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.89974683544304e-06, |
|
"loss": 0.2987, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.89341772151899e-06, |
|
"loss": 0.3067, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.887088607594939e-06, |
|
"loss": 0.2962, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.880759493670888e-06, |
|
"loss": 0.2807, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.874430379746836e-06, |
|
"loss": 0.2533, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"eval_loss": 0.3274388611316681, |
|
"eval_runtime": 1481.9106, |
|
"eval_samples_per_second": 7.025, |
|
"eval_steps_per_second": 0.439, |
|
"eval_wer": 40.657571667832, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.868101265822785e-06, |
|
"loss": 0.21, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.861772151898735e-06, |
|
"loss": 0.2463, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.855443037974684e-06, |
|
"loss": 0.2688, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.849113924050633e-06, |
|
"loss": 0.2293, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.842784810126583e-06, |
|
"loss": 0.2283, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.836455696202532e-06, |
|
"loss": 0.228, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.830126582278482e-06, |
|
"loss": 0.1919, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.823797468354431e-06, |
|
"loss": 0.2173, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.81746835443038e-06, |
|
"loss": 0.2066, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.81113924050633e-06, |
|
"loss": 0.1826, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.80481012658228e-06, |
|
"loss": 0.1855, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.798481012658229e-06, |
|
"loss": 0.1702, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.792151898734178e-06, |
|
"loss": 0.1714, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.785822784810128e-06, |
|
"loss": 0.1686, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.779493670886077e-06, |
|
"loss": 0.1762, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.773164556962027e-06, |
|
"loss": 0.1722, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.766835443037976e-06, |
|
"loss": 0.2051, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.760506329113924e-06, |
|
"loss": 0.2071, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.754177215189873e-06, |
|
"loss": 0.1664, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.747848101265823e-06, |
|
"loss": 0.1871, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.741518987341772e-06, |
|
"loss": 0.1893, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.735189873417721e-06, |
|
"loss": 0.1756, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.728860759493671e-06, |
|
"loss": 0.1687, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.72253164556962e-06, |
|
"loss": 0.1685, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.71620253164557e-06, |
|
"loss": 0.1805, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.70987341772152e-06, |
|
"loss": 0.1868, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.70354430379747e-06, |
|
"loss": 0.212, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.69721518987342e-06, |
|
"loss": 0.2019, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.69088607594937e-06, |
|
"loss": 0.2076, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.684556962025317e-06, |
|
"loss": 0.1971, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.678227848101266e-06, |
|
"loss": 0.2424, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.671898734177216e-06, |
|
"loss": 0.2346, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.665569620253165e-06, |
|
"loss": 0.2436, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.659240506329115e-06, |
|
"loss": 0.2613, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.652911392405064e-06, |
|
"loss": 0.2781, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.646582278481013e-06, |
|
"loss": 0.2465, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.640253164556963e-06, |
|
"loss": 0.2386, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.633924050632912e-06, |
|
"loss": 0.2298, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.627594936708862e-06, |
|
"loss": 0.264, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.621265822784811e-06, |
|
"loss": 0.2077, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"eval_loss": 0.25397753715515137, |
|
"eval_runtime": 1477.9319, |
|
"eval_samples_per_second": 7.044, |
|
"eval_steps_per_second": 0.44, |
|
"eval_wer": 35.12252231743606, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.61493670886076e-06, |
|
"loss": 0.192, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.60860759493671e-06, |
|
"loss": 0.1903, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.60227848101266e-06, |
|
"loss": 0.2135, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.595949367088609e-06, |
|
"loss": 0.1917, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.589620253164558e-06, |
|
"loss": 0.1579, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.583291139240508e-06, |
|
"loss": 0.1779, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.576962025316457e-06, |
|
"loss": 0.1504, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.570632911392405e-06, |
|
"loss": 0.1574, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.564303797468354e-06, |
|
"loss": 0.1757, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.557974683544304e-06, |
|
"loss": 0.1996, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.551645569620253e-06, |
|
"loss": 0.1716, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.545316455696203e-06, |
|
"loss": 0.1509, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.538987341772152e-06, |
|
"loss": 0.1402, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.532658227848101e-06, |
|
"loss": 0.1612, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.526329113924051e-06, |
|
"loss": 0.1357, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.52e-06, |
|
"loss": 0.1426, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.51367088607595e-06, |
|
"loss": 0.1644, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.5073417721519e-06, |
|
"loss": 0.1196, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.501012658227849e-06, |
|
"loss": 0.1332, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.494683544303798e-06, |
|
"loss": 0.1096, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.488354430379747e-06, |
|
"loss": 0.1263, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.482025316455697e-06, |
|
"loss": 0.1467, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.475696202531646e-06, |
|
"loss": 0.1372, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.469367088607596e-06, |
|
"loss": 0.1256, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.463037974683545e-06, |
|
"loss": 0.1319, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.456708860759495e-06, |
|
"loss": 0.1385, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.450379746835444e-06, |
|
"loss": 0.147, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.444050632911393e-06, |
|
"loss": 0.1527, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.437721518987343e-06, |
|
"loss": 0.1357, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.431392405063292e-06, |
|
"loss": 0.122, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.425063291139242e-06, |
|
"loss": 0.0918, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.418734177215191e-06, |
|
"loss": 0.1218, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.41240506329114e-06, |
|
"loss": 0.139, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.40607594936709e-06, |
|
"loss": 0.1205, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.39974683544304e-06, |
|
"loss": 0.1158, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.393417721518989e-06, |
|
"loss": 0.1156, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.387088607594937e-06, |
|
"loss": 0.1044, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.380759493670886e-06, |
|
"loss": 0.112, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.374430379746835e-06, |
|
"loss": 0.1023, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.368101265822785e-06, |
|
"loss": 0.0966, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"eval_loss": 0.2638459801673889, |
|
"eval_runtime": 1500.936, |
|
"eval_samples_per_second": 6.936, |
|
"eval_steps_per_second": 0.434, |
|
"eval_wer": 37.93069806013106, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.361772151898734e-06, |
|
"loss": 0.0942, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.355443037974684e-06, |
|
"loss": 0.0826, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.349113924050633e-06, |
|
"loss": 0.086, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.342784810126583e-06, |
|
"loss": 0.0956, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.336455696202532e-06, |
|
"loss": 0.0935, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.330126582278481e-06, |
|
"loss": 0.0923, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.323797468354431e-06, |
|
"loss": 0.1163, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.31746835443038e-06, |
|
"loss": 0.1099, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.31113924050633e-06, |
|
"loss": 0.0879, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.30481012658228e-06, |
|
"loss": 0.0994, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.298481012658229e-06, |
|
"loss": 0.1069, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.292151898734178e-06, |
|
"loss": 0.0957, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.285822784810127e-06, |
|
"loss": 0.0866, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.279493670886077e-06, |
|
"loss": 0.0965, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.273164556962026e-06, |
|
"loss": 0.0952, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.266835443037976e-06, |
|
"loss": 0.1178, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.260506329113925e-06, |
|
"loss": 0.1154, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.254177215189875e-06, |
|
"loss": 0.1075, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.247848101265824e-06, |
|
"loss": 0.1161, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.241518987341773e-06, |
|
"loss": 0.1153, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 9.235189873417723e-06, |
|
"loss": 0.1343, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 9.228860759493672e-06, |
|
"loss": 0.1374, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 9.222531645569622e-06, |
|
"loss": 0.1436, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 9.216202531645571e-06, |
|
"loss": 0.1561, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 9.20987341772152e-06, |
|
"loss": 0.1709, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 9.20354430379747e-06, |
|
"loss": 0.142, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 9.197215189873418e-06, |
|
"loss": 0.1389, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 9.190886075949367e-06, |
|
"loss": 0.1398, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.184556962025317e-06, |
|
"loss": 0.1526, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.178227848101266e-06, |
|
"loss": 0.1273, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.171898734177215e-06, |
|
"loss": 0.1119, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.165569620253165e-06, |
|
"loss": 0.1088, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.159493670886076e-06, |
|
"loss": 0.1362, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.153164556962026e-06, |
|
"loss": 0.1247, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.146835443037975e-06, |
|
"loss": 0.0974, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.140506329113925e-06, |
|
"loss": 0.1218, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.134177215189874e-06, |
|
"loss": 0.0846, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.127848101265823e-06, |
|
"loss": 0.0964, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.121518987341773e-06, |
|
"loss": 0.1166, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.115189873417722e-06, |
|
"loss": 0.1212, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"eval_loss": 0.22517943382263184, |
|
"eval_runtime": 1537.556, |
|
"eval_samples_per_second": 6.771, |
|
"eval_steps_per_second": 0.423, |
|
"eval_wer": 69.04990325045935, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.108860759493672e-06, |
|
"loss": 0.1076, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.102531645569621e-06, |
|
"loss": 0.086, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.09620253164557e-06, |
|
"loss": 0.0773, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.08987341772152e-06, |
|
"loss": 0.098, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 9.08354430379747e-06, |
|
"loss": 0.0747, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 9.077215189873419e-06, |
|
"loss": 0.0895, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 9.070886075949368e-06, |
|
"loss": 0.0957, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 9.064556962025318e-06, |
|
"loss": 0.0789, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 9.058227848101267e-06, |
|
"loss": 0.0718, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 9.051898734177215e-06, |
|
"loss": 0.0737, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 9.045569620253164e-06, |
|
"loss": 0.0731, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 9.039240506329114e-06, |
|
"loss": 0.087, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 9.032911392405063e-06, |
|
"loss": 0.0875, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 9.026582278481013e-06, |
|
"loss": 0.0768, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 9.020253164556962e-06, |
|
"loss": 0.0842, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 9.013924050632911e-06, |
|
"loss": 0.092, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 9.007594936708861e-06, |
|
"loss": 0.0949, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 9.00126582278481e-06, |
|
"loss": 0.0989, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 8.99493670886076e-06, |
|
"loss": 0.0908, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 8.98860759493671e-06, |
|
"loss": 0.0783, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.982278481012659e-06, |
|
"loss": 0.0651, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.975949367088608e-06, |
|
"loss": 0.0791, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.969620253164557e-06, |
|
"loss": 0.0829, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.963291139240507e-06, |
|
"loss": 0.0797, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.956962025316456e-06, |
|
"loss": 0.0743, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.950632911392406e-06, |
|
"loss": 0.0814, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.944303797468355e-06, |
|
"loss": 0.0594, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.937974683544305e-06, |
|
"loss": 0.0709, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.931645569620254e-06, |
|
"loss": 0.0621, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.925316455696203e-06, |
|
"loss": 0.0599, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.918987341772153e-06, |
|
"loss": 0.0625, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.912658227848102e-06, |
|
"loss": 0.0541, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.906329113924052e-06, |
|
"loss": 0.059, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.900000000000001e-06, |
|
"loss": 0.0576, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.89367088607595e-06, |
|
"loss": 0.0623, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.8873417721519e-06, |
|
"loss": 0.0592, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.88101265822785e-06, |
|
"loss": 0.0692, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.874683544303799e-06, |
|
"loss": 0.077, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.868354430379748e-06, |
|
"loss": 0.0662, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.862025316455696e-06, |
|
"loss": 0.0605, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"eval_loss": 0.2468421906232834, |
|
"eval_runtime": 1442.9116, |
|
"eval_samples_per_second": 7.215, |
|
"eval_steps_per_second": 0.451, |
|
"eval_wer": 33.46883689165677, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.855696202531645e-06, |
|
"loss": 0.0769, |
|
"step": 5025 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.849367088607595e-06, |
|
"loss": 0.0601, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.843037974683544e-06, |
|
"loss": 0.0698, |
|
"step": 5075 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.836708860759494e-06, |
|
"loss": 0.0593, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.830379746835443e-06, |
|
"loss": 0.0521, |
|
"step": 5125 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.824050632911393e-06, |
|
"loss": 0.0687, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.817721518987342e-06, |
|
"loss": 0.0818, |
|
"step": 5175 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.811392405063291e-06, |
|
"loss": 0.0706, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.805063291139241e-06, |
|
"loss": 0.0734, |
|
"step": 5225 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.79873417721519e-06, |
|
"loss": 0.0795, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 8.79240506329114e-06, |
|
"loss": 0.0905, |
|
"step": 5275 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 8.78607594936709e-06, |
|
"loss": 0.086, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 8.779746835443039e-06, |
|
"loss": 0.105, |
|
"step": 5325 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 8.773417721518988e-06, |
|
"loss": 0.0981, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 8.767088607594937e-06, |
|
"loss": 0.1131, |
|
"step": 5375 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 8.760759493670887e-06, |
|
"loss": 0.0916, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 8.754430379746836e-06, |
|
"loss": 0.0974, |
|
"step": 5425 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 8.748101265822786e-06, |
|
"loss": 0.0899, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 8.741772151898735e-06, |
|
"loss": 0.1104, |
|
"step": 5475 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 8.735443037974685e-06, |
|
"loss": 0.0818, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 8.729113924050634e-06, |
|
"loss": 0.0764, |
|
"step": 5525 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 8.722784810126583e-06, |
|
"loss": 0.0709, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 8.716455696202533e-06, |
|
"loss": 0.0891, |
|
"step": 5575 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 8.710126582278482e-06, |
|
"loss": 0.0869, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 8.703797468354432e-06, |
|
"loss": 0.0631, |
|
"step": 5625 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 8.697468354430381e-06, |
|
"loss": 0.0795, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 8.69113924050633e-06, |
|
"loss": 0.0603, |
|
"step": 5675 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 8.68481012658228e-06, |
|
"loss": 0.064, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 8.67848101265823e-06, |
|
"loss": 0.0738, |
|
"step": 5725 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 8.672151898734177e-06, |
|
"loss": 0.0796, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 8.665822784810127e-06, |
|
"loss": 0.0723, |
|
"step": 5775 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 8.659493670886076e-06, |
|
"loss": 0.0592, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 8.653164556962025e-06, |
|
"loss": 0.051, |
|
"step": 5825 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 8.646835443037975e-06, |
|
"loss": 0.0716, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 8.640506329113924e-06, |
|
"loss": 0.0598, |
|
"step": 5875 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 8.634177215189874e-06, |
|
"loss": 0.0524, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 8.627848101265823e-06, |
|
"loss": 0.0623, |
|
"step": 5925 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 8.621518987341773e-06, |
|
"loss": 0.054, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 8.615189873417722e-06, |
|
"loss": 0.057, |
|
"step": 5975 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 8.608860759493671e-06, |
|
"loss": 0.0479, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"eval_loss": 0.24428367614746094, |
|
"eval_runtime": 1457.5568, |
|
"eval_samples_per_second": 7.143, |
|
"eval_steps_per_second": 0.447, |
|
"eval_wer": 36.51929299663409, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 8.602531645569621e-06, |
|
"loss": 0.0501, |
|
"step": 6025 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 8.59620253164557e-06, |
|
"loss": 0.0527, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 8.58987341772152e-06, |
|
"loss": 0.0558, |
|
"step": 6075 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 8.58354430379747e-06, |
|
"loss": 0.0548, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 8.577215189873419e-06, |
|
"loss": 0.0503, |
|
"step": 6125 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 8.570886075949368e-06, |
|
"loss": 0.0641, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 8.564556962025316e-06, |
|
"loss": 0.0748, |
|
"step": 6175 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 8.558227848101267e-06, |
|
"loss": 0.0775, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 8.551898734177216e-06, |
|
"loss": 0.0662, |
|
"step": 6225 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 8.545569620253166e-06, |
|
"loss": 0.0591, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 8.539240506329115e-06, |
|
"loss": 0.0421, |
|
"step": 6275 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 8.532911392405065e-06, |
|
"loss": 0.0565, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 8.526582278481014e-06, |
|
"loss": 0.0574, |
|
"step": 6325 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 8.520253164556963e-06, |
|
"loss": 0.0466, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 8.513924050632913e-06, |
|
"loss": 0.0486, |
|
"step": 6375 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 8.507594936708862e-06, |
|
"loss": 0.0502, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 8.501265822784812e-06, |
|
"loss": 0.0415, |
|
"step": 6425 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 8.494936708860761e-06, |
|
"loss": 0.0449, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 8.488607594936709e-06, |
|
"loss": 0.0525, |
|
"step": 6475 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 8.482278481012658e-06, |
|
"loss": 0.0401, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 8.475949367088608e-06, |
|
"loss": 0.0502, |
|
"step": 6525 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 8.469620253164557e-06, |
|
"loss": 0.0453, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 8.463291139240507e-06, |
|
"loss": 0.0442, |
|
"step": 6575 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 8.456962025316456e-06, |
|
"loss": 0.0423, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 8.450632911392405e-06, |
|
"loss": 0.0518, |
|
"step": 6625 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 8.444303797468355e-06, |
|
"loss": 0.0479, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 8.437974683544304e-06, |
|
"loss": 0.0521, |
|
"step": 6675 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 8.431645569620254e-06, |
|
"loss": 0.0436, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 8.425316455696203e-06, |
|
"loss": 0.0539, |
|
"step": 6725 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 8.418987341772153e-06, |
|
"loss": 0.0446, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 8.412658227848102e-06, |
|
"loss": 0.0521, |
|
"step": 6775 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 8.406329113924051e-06, |
|
"loss": 0.0391, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 8.400000000000001e-06, |
|
"loss": 0.0513, |
|
"step": 6825 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 8.39367088607595e-06, |
|
"loss": 0.0449, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 8.3873417721519e-06, |
|
"loss": 0.0463, |
|
"step": 6875 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 8.38101265822785e-06, |
|
"loss": 0.0495, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 8.374683544303797e-06, |
|
"loss": 0.0628, |
|
"step": 6925 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 8.368354430379746e-06, |
|
"loss": 0.0594, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 8.362025316455696e-06, |
|
"loss": 0.0573, |
|
"step": 6975 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 8.355696202531645e-06, |
|
"loss": 0.057, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"eval_loss": 0.25129273533821106, |
|
"eval_runtime": 1479.2043, |
|
"eval_samples_per_second": 7.038, |
|
"eval_steps_per_second": 0.44, |
|
"eval_wer": 57.010683100538216, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 8.349367088607596e-06, |
|
"loss": 0.0534, |
|
"step": 7025 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 8.343037974683546e-06, |
|
"loss": 0.0663, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 8.336708860759495e-06, |
|
"loss": 0.0698, |
|
"step": 7075 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 8.330379746835445e-06, |
|
"loss": 0.0633, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 8.324050632911394e-06, |
|
"loss": 0.0808, |
|
"step": 7125 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 8.317721518987343e-06, |
|
"loss": 0.0699, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 8.311392405063293e-06, |
|
"loss": 0.0736, |
|
"step": 7175 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 8.305063291139242e-06, |
|
"loss": 0.069, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 8.29873417721519e-06, |
|
"loss": 0.0718, |
|
"step": 7225 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 8.29240506329114e-06, |
|
"loss": 0.0674, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 8.286075949367089e-06, |
|
"loss": 0.0568, |
|
"step": 7275 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 8.279746835443038e-06, |
|
"loss": 0.0497, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 8.273417721518988e-06, |
|
"loss": 0.0563, |
|
"step": 7325 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 8.267088607594937e-06, |
|
"loss": 0.0649, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 8.260759493670887e-06, |
|
"loss": 0.042, |
|
"step": 7375 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 8.254430379746836e-06, |
|
"loss": 0.0559, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 8.248101265822785e-06, |
|
"loss": 0.0476, |
|
"step": 7425 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 8.241772151898735e-06, |
|
"loss": 0.0427, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 8.235443037974684e-06, |
|
"loss": 0.0533, |
|
"step": 7475 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 8.229113924050634e-06, |
|
"loss": 0.0633, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 8.222784810126583e-06, |
|
"loss": 0.055, |
|
"step": 7525 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 8.216455696202533e-06, |
|
"loss": 0.043, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 8.210126582278482e-06, |
|
"loss": 0.0421, |
|
"step": 7575 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 8.203797468354431e-06, |
|
"loss": 0.0519, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.197468354430381e-06, |
|
"loss": 0.0398, |
|
"step": 7625 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.19113924050633e-06, |
|
"loss": 0.041, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.184810126582278e-06, |
|
"loss": 0.0477, |
|
"step": 7675 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.178481012658227e-06, |
|
"loss": 0.046, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.172151898734177e-06, |
|
"loss": 0.0465, |
|
"step": 7725 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.165822784810126e-06, |
|
"loss": 0.0359, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.159493670886076e-06, |
|
"loss": 0.0362, |
|
"step": 7775 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.153164556962025e-06, |
|
"loss": 0.04, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.146835443037975e-06, |
|
"loss": 0.0433, |
|
"step": 7825 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.140506329113926e-06, |
|
"loss": 0.0399, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.134177215189875e-06, |
|
"loss": 0.0396, |
|
"step": 7875 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.127848101265825e-06, |
|
"loss": 0.0472, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.121518987341774e-06, |
|
"loss": 0.0533, |
|
"step": 7925 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.115189873417722e-06, |
|
"loss": 0.0564, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.108860759493671e-06, |
|
"loss": 0.0508, |
|
"step": 7975 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 8.10253164556962e-06, |
|
"loss": 0.0375, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"eval_loss": 0.252152681350708, |
|
"eval_runtime": 1456.2607, |
|
"eval_samples_per_second": 7.149, |
|
"eval_steps_per_second": 0.447, |
|
"eval_wer": 44.38120945056017, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.09620253164557e-06, |
|
"loss": 0.0296, |
|
"step": 8025 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.08987341772152e-06, |
|
"loss": 0.0395, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.083544303797469e-06, |
|
"loss": 0.0409, |
|
"step": 8075 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.077215189873418e-06, |
|
"loss": 0.0343, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.070886075949368e-06, |
|
"loss": 0.0321, |
|
"step": 8125 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.064556962025317e-06, |
|
"loss": 0.0333, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.058227848101267e-06, |
|
"loss": 0.0327, |
|
"step": 8175 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.051898734177216e-06, |
|
"loss": 0.0318, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.045569620253165e-06, |
|
"loss": 0.0405, |
|
"step": 8225 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.039240506329115e-06, |
|
"loss": 0.0382, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.032911392405064e-06, |
|
"loss": 0.0317, |
|
"step": 8275 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.026582278481014e-06, |
|
"loss": 0.0285, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.020253164556963e-06, |
|
"loss": 0.0285, |
|
"step": 8325 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.013924050632913e-06, |
|
"loss": 0.0269, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.007594936708862e-06, |
|
"loss": 0.0334, |
|
"step": 8375 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.00126582278481e-06, |
|
"loss": 0.0406, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 7.99493670886076e-06, |
|
"loss": 0.038, |
|
"step": 8425 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 7.988607594936709e-06, |
|
"loss": 0.0315, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 7.982278481012658e-06, |
|
"loss": 0.0329, |
|
"step": 8475 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 7.975949367088607e-06, |
|
"loss": 0.0373, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 7.969620253164557e-06, |
|
"loss": 0.0514, |
|
"step": 8525 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 7.963291139240506e-06, |
|
"loss": 0.0349, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 7.956962025316456e-06, |
|
"loss": 0.0385, |
|
"step": 8575 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 7.950632911392405e-06, |
|
"loss": 0.0367, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 7.944303797468355e-06, |
|
"loss": 0.0317, |
|
"step": 8625 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 7.937974683544304e-06, |
|
"loss": 0.0368, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 7.931645569620255e-06, |
|
"loss": 0.045, |
|
"step": 8675 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 7.925316455696203e-06, |
|
"loss": 0.0456, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 7.918987341772152e-06, |
|
"loss": 0.0421, |
|
"step": 8725 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 7.912658227848102e-06, |
|
"loss": 0.043, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 7.906329113924051e-06, |
|
"loss": 0.0428, |
|
"step": 8775 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 7.9e-06, |
|
"loss": 0.0459, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 7.89367088607595e-06, |
|
"loss": 0.0467, |
|
"step": 8825 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 7.8873417721519e-06, |
|
"loss": 0.0462, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 7.881012658227849e-06, |
|
"loss": 0.0578, |
|
"step": 8875 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 7.874683544303798e-06, |
|
"loss": 0.0422, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 7.868354430379748e-06, |
|
"loss": 0.0546, |
|
"step": 8925 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 7.862025316455697e-06, |
|
"loss": 0.0515, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 7.855696202531647e-06, |
|
"loss": 0.0541, |
|
"step": 8975 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 7.849367088607596e-06, |
|
"loss": 0.0479, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"eval_loss": 0.2375001162290573, |
|
"eval_runtime": 1456.1088, |
|
"eval_samples_per_second": 7.15, |
|
"eval_steps_per_second": 0.447, |
|
"eval_wer": 31.58262735979447, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 7.843037974683545e-06, |
|
"loss": 0.0434, |
|
"step": 9025 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 7.836708860759495e-06, |
|
"loss": 0.033, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 7.830379746835444e-06, |
|
"loss": 0.0494, |
|
"step": 9075 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 7.824050632911394e-06, |
|
"loss": 0.0522, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 7.817721518987343e-06, |
|
"loss": 0.0338, |
|
"step": 9125 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 7.811392405063291e-06, |
|
"loss": 0.0395, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 7.80506329113924e-06, |
|
"loss": 0.0332, |
|
"step": 9175 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 7.79873417721519e-06, |
|
"loss": 0.0381, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 7.79240506329114e-06, |
|
"loss": 0.0409, |
|
"step": 9225 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 7.786075949367089e-06, |
|
"loss": 0.0529, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 7.779746835443038e-06, |
|
"loss": 0.0498, |
|
"step": 9275 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 7.773417721518987e-06, |
|
"loss": 0.0379, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 7.767088607594937e-06, |
|
"loss": 0.0316, |
|
"step": 9325 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 7.760759493670886e-06, |
|
"loss": 0.0398, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 7.754430379746836e-06, |
|
"loss": 0.0427, |
|
"step": 9375 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 7.748101265822785e-06, |
|
"loss": 0.0363, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 7.741772151898735e-06, |
|
"loss": 0.0393, |
|
"step": 9425 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 7.735443037974684e-06, |
|
"loss": 0.0377, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 7.729113924050633e-06, |
|
"loss": 0.0313, |
|
"step": 9475 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 7.722784810126583e-06, |
|
"loss": 0.0359, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 7.716455696202532e-06, |
|
"loss": 0.0289, |
|
"step": 9525 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 7.710126582278482e-06, |
|
"loss": 0.0354, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 7.703797468354431e-06, |
|
"loss": 0.0309, |
|
"step": 9575 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 7.69746835443038e-06, |
|
"loss": 0.0321, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 7.69113924050633e-06, |
|
"loss": 0.0392, |
|
"step": 9625 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 7.68481012658228e-06, |
|
"loss": 0.0362, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 7.678481012658229e-06, |
|
"loss": 0.0358, |
|
"step": 9675 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 7.672151898734178e-06, |
|
"loss": 0.045, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 7.665822784810128e-06, |
|
"loss": 0.0401, |
|
"step": 9725 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 7.659493670886077e-06, |
|
"loss": 0.0357, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 7.653164556962027e-06, |
|
"loss": 0.0256, |
|
"step": 9775 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 7.646835443037976e-06, |
|
"loss": 0.0358, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 7.640506329113925e-06, |
|
"loss": 0.0361, |
|
"step": 9825 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 7.634177215189875e-06, |
|
"loss": 0.0365, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 7.6278481012658234e-06, |
|
"loss": 0.0344, |
|
"step": 9875 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 7.621518987341773e-06, |
|
"loss": 0.0349, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 7.615189873417722e-06, |
|
"loss": 0.0289, |
|
"step": 9925 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 7.608860759493672e-06, |
|
"loss": 0.0259, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 7.602531645569621e-06, |
|
"loss": 0.0289, |
|
"step": 9975 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 7.59620253164557e-06, |
|
"loss": 0.0269, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"eval_loss": 0.2615405321121216, |
|
"eval_runtime": 1420.7197, |
|
"eval_samples_per_second": 7.328, |
|
"eval_steps_per_second": 0.458, |
|
"eval_wer": 36.23310948145498, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 7.589873417721519e-06, |
|
"loss": 0.0261, |
|
"step": 10025 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 7.583544303797469e-06, |
|
"loss": 0.0278, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 7.577215189873418e-06, |
|
"loss": 0.0248, |
|
"step": 10075 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 7.5708860759493674e-06, |
|
"loss": 0.0263, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 7.564556962025317e-06, |
|
"loss": 0.0253, |
|
"step": 10125 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 7.558227848101266e-06, |
|
"loss": 0.0248, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 7.551898734177216e-06, |
|
"loss": 0.0281, |
|
"step": 10175 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 7.545569620253165e-06, |
|
"loss": 0.0276, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 7.539240506329114e-06, |
|
"loss": 0.0285, |
|
"step": 10225 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 7.532911392405063e-06, |
|
"loss": 0.0265, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 7.526582278481013e-06, |
|
"loss": 0.0288, |
|
"step": 10275 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 7.520253164556963e-06, |
|
"loss": 0.0247, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 7.513924050632912e-06, |
|
"loss": 0.0281, |
|
"step": 10325 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 7.507594936708862e-06, |
|
"loss": 0.0262, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 7.501265822784811e-06, |
|
"loss": 0.0243, |
|
"step": 10375 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 7.494936708860761e-06, |
|
"loss": 0.0331, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 7.48860759493671e-06, |
|
"loss": 0.0346, |
|
"step": 10425 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 7.4822784810126594e-06, |
|
"loss": 0.0348, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 7.475949367088609e-06, |
|
"loss": 0.0272, |
|
"step": 10475 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 7.4696202531645574e-06, |
|
"loss": 0.0326, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 7.463291139240507e-06, |
|
"loss": 0.033, |
|
"step": 10525 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 7.456962025316456e-06, |
|
"loss": 0.0333, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 7.450632911392406e-06, |
|
"loss": 0.0371, |
|
"step": 10575 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 7.444303797468355e-06, |
|
"loss": 0.0359, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 7.437974683544305e-06, |
|
"loss": 0.0438, |
|
"step": 10625 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 7.431645569620254e-06, |
|
"loss": 0.0382, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 7.4253164556962034e-06, |
|
"loss": 0.0337, |
|
"step": 10675 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 7.418987341772153e-06, |
|
"loss": 0.0422, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.4126582278481014e-06, |
|
"loss": 0.0467, |
|
"step": 10725 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.406329113924051e-06, |
|
"loss": 0.0437, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.4e-06, |
|
"loss": 0.0328, |
|
"step": 10775 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.39367088607595e-06, |
|
"loss": 0.0295, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.387341772151899e-06, |
|
"loss": 0.0431, |
|
"step": 10825 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.381012658227849e-06, |
|
"loss": 0.0456, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.374683544303798e-06, |
|
"loss": 0.0267, |
|
"step": 10875 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.3683544303797474e-06, |
|
"loss": 0.0334, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.362025316455697e-06, |
|
"loss": 0.0296, |
|
"step": 10925 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.3556962025316454e-06, |
|
"loss": 0.0261, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.349367088607595e-06, |
|
"loss": 0.0299, |
|
"step": 10975 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.343037974683544e-06, |
|
"loss": 0.0379, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"eval_loss": 0.23971551656723022, |
|
"eval_runtime": 1435.6772, |
|
"eval_samples_per_second": 7.252, |
|
"eval_steps_per_second": 0.453, |
|
"eval_wer": 31.7712483129807, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.336708860759494e-06, |
|
"loss": 0.0457, |
|
"step": 11025 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.330379746835443e-06, |
|
"loss": 0.0275, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.324050632911393e-06, |
|
"loss": 0.0254, |
|
"step": 11075 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 7.317721518987342e-06, |
|
"loss": 0.035, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.311392405063292e-06, |
|
"loss": 0.0273, |
|
"step": 11125 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.305063291139242e-06, |
|
"loss": 0.0227, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.298734177215191e-06, |
|
"loss": 0.0263, |
|
"step": 11175 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.2924050632911406e-06, |
|
"loss": 0.0294, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.28607594936709e-06, |
|
"loss": 0.026, |
|
"step": 11225 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.279746835443039e-06, |
|
"loss": 0.0251, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.273417721518988e-06, |
|
"loss": 0.024, |
|
"step": 11275 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.2670886075949374e-06, |
|
"loss": 0.0344, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.260759493670887e-06, |
|
"loss": 0.0338, |
|
"step": 11325 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.254430379746836e-06, |
|
"loss": 0.0266, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.248101265822786e-06, |
|
"loss": 0.0271, |
|
"step": 11375 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.241772151898735e-06, |
|
"loss": 0.0338, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.2354430379746846e-06, |
|
"loss": 0.0321, |
|
"step": 11425 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.229113924050634e-06, |
|
"loss": 0.0385, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.222784810126583e-06, |
|
"loss": 0.0284, |
|
"step": 11475 |
|
}, |
|
{ |
|
"epoch": 6.02, |
|
"learning_rate": 7.216455696202532e-06, |
|
"loss": 0.0268, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.2101265822784814e-06, |
|
"loss": 0.0191, |
|
"step": 11525 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.203797468354431e-06, |
|
"loss": 0.0243, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.19746835443038e-06, |
|
"loss": 0.0228, |
|
"step": 11575 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.19113924050633e-06, |
|
"loss": 0.0187, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.184810126582279e-06, |
|
"loss": 0.0255, |
|
"step": 11625 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.1784810126582286e-06, |
|
"loss": 0.0218, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.172151898734178e-06, |
|
"loss": 0.0249, |
|
"step": 11675 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.165822784810127e-06, |
|
"loss": 0.0254, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.159493670886076e-06, |
|
"loss": 0.0323, |
|
"step": 11725 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.1531645569620254e-06, |
|
"loss": 0.021, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.146835443037975e-06, |
|
"loss": 0.0242, |
|
"step": 11775 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.140506329113924e-06, |
|
"loss": 0.0233, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.134177215189874e-06, |
|
"loss": 0.0237, |
|
"step": 11825 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.127848101265823e-06, |
|
"loss": 0.0257, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.121518987341773e-06, |
|
"loss": 0.0256, |
|
"step": 11875 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 7.115189873417722e-06, |
|
"loss": 0.0264, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 7.108860759493671e-06, |
|
"loss": 0.0254, |
|
"step": 11925 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 7.102531645569622e-06, |
|
"loss": 0.025, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 7.096202531645571e-06, |
|
"loss": 0.0219, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 7.08987341772152e-06, |
|
"loss": 0.0241, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"eval_loss": 0.26222535967826843, |
|
"eval_runtime": 1460.2321, |
|
"eval_samples_per_second": 7.13, |
|
"eval_steps_per_second": 0.446, |
|
"eval_wer": 32.87370526350022, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 7.083544303797469e-06, |
|
"loss": 0.0337, |
|
"step": 12025 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 7.0772151898734186e-06, |
|
"loss": 0.0179, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 7.070886075949368e-06, |
|
"loss": 0.0174, |
|
"step": 12075 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 7.0645569620253174e-06, |
|
"loss": 0.0211, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 7.058227848101267e-06, |
|
"loss": 0.0224, |
|
"step": 12125 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 7.051898734177216e-06, |
|
"loss": 0.0226, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 7.045569620253166e-06, |
|
"loss": 0.0271, |
|
"step": 12175 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 7.039240506329114e-06, |
|
"loss": 0.0264, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 7.032911392405064e-06, |
|
"loss": 0.031, |
|
"step": 12225 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 7.026582278481013e-06, |
|
"loss": 0.022, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 7.0202531645569626e-06, |
|
"loss": 0.0266, |
|
"step": 12275 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 7.013924050632912e-06, |
|
"loss": 0.0269, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 7.0075949367088614e-06, |
|
"loss": 0.028, |
|
"step": 12325 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 7.001265822784811e-06, |
|
"loss": 0.0361, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 6.99493670886076e-06, |
|
"loss": 0.0383, |
|
"step": 12375 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 6.98860759493671e-06, |
|
"loss": 0.0312, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 6.982278481012658e-06, |
|
"loss": 0.0275, |
|
"step": 12425 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 6.975949367088608e-06, |
|
"loss": 0.0407, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 6.969620253164557e-06, |
|
"loss": 0.0436, |
|
"step": 12475 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 6.9632911392405066e-06, |
|
"loss": 0.0359, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 6.956962025316456e-06, |
|
"loss": 0.0281, |
|
"step": 12525 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 6.9506329113924054e-06, |
|
"loss": 0.0283, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 6.944303797468355e-06, |
|
"loss": 0.0248, |
|
"step": 12575 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 6.937974683544304e-06, |
|
"loss": 0.0295, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 6.931645569620254e-06, |
|
"loss": 0.0287, |
|
"step": 12625 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 6.925316455696202e-06, |
|
"loss": 0.0261, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 6.918987341772152e-06, |
|
"loss": 0.0259, |
|
"step": 12675 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 6.912658227848101e-06, |
|
"loss": 0.0265, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 6.906329113924051e-06, |
|
"loss": 0.0239, |
|
"step": 12725 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 6.9e-06, |
|
"loss": 0.0342, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 6.89367088607595e-06, |
|
"loss": 0.0253, |
|
"step": 12775 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 6.8873417721519e-06, |
|
"loss": 0.0317, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 6.881012658227849e-06, |
|
"loss": 0.0233, |
|
"step": 12825 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 6.8746835443037986e-06, |
|
"loss": 0.0198, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 6.868354430379748e-06, |
|
"loss": 0.018, |
|
"step": 12875 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 6.862025316455697e-06, |
|
"loss": 0.0182, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 6.855949367088608e-06, |
|
"loss": 0.0279, |
|
"step": 12925 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 6.8496202531645574e-06, |
|
"loss": 0.0201, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 6.843291139240507e-06, |
|
"loss": 0.0222, |
|
"step": 12975 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 6.836962025316456e-06, |
|
"loss": 0.0236, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"eval_loss": 0.2616036534309387, |
|
"eval_runtime": 1403.0214, |
|
"eval_samples_per_second": 7.42, |
|
"eval_steps_per_second": 0.464, |
|
"eval_wer": 27.36142051090262, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 6.830632911392405e-06, |
|
"loss": 0.0159, |
|
"step": 13025 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 6.824303797468354e-06, |
|
"loss": 0.0186, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 6.817974683544304e-06, |
|
"loss": 0.0246, |
|
"step": 13075 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 6.811645569620253e-06, |
|
"loss": 0.0198, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 6.805316455696203e-06, |
|
"loss": 0.0287, |
|
"step": 13125 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 6.798987341772152e-06, |
|
"loss": 0.0255, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 6.792658227848102e-06, |
|
"loss": 0.035, |
|
"step": 13175 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 6.786329113924052e-06, |
|
"loss": 0.0358, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 6.780000000000001e-06, |
|
"loss": 0.0283, |
|
"step": 13225 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 6.773670886075951e-06, |
|
"loss": 0.0203, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 6.7673417721519e-06, |
|
"loss": 0.0193, |
|
"step": 13275 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 6.7610126582278494e-06, |
|
"loss": 0.0174, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 6.754683544303798e-06, |
|
"loss": 0.0279, |
|
"step": 13325 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 6.7483544303797474e-06, |
|
"loss": 0.023, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 6.742025316455697e-06, |
|
"loss": 0.0266, |
|
"step": 13375 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 6.735696202531646e-06, |
|
"loss": 0.024, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 6.729367088607596e-06, |
|
"loss": 0.0164, |
|
"step": 13425 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 6.723037974683545e-06, |
|
"loss": 0.019, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 6.716708860759495e-06, |
|
"loss": 0.0209, |
|
"step": 13475 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 6.710379746835444e-06, |
|
"loss": 0.0241, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 6.7040506329113934e-06, |
|
"loss": 0.0233, |
|
"step": 13525 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 6.697721518987342e-06, |
|
"loss": 0.0204, |
|
"step": 13550 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 6.6913924050632914e-06, |
|
"loss": 0.0131, |
|
"step": 13575 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 6.685063291139241e-06, |
|
"loss": 0.0268, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 6.67873417721519e-06, |
|
"loss": 0.0156, |
|
"step": 13625 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 6.67240506329114e-06, |
|
"loss": 0.0205, |
|
"step": 13650 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 6.666075949367089e-06, |
|
"loss": 0.0217, |
|
"step": 13675 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 6.659746835443039e-06, |
|
"loss": 0.0258, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 6.653417721518988e-06, |
|
"loss": 0.0211, |
|
"step": 13725 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 6.647088607594937e-06, |
|
"loss": 0.0188, |
|
"step": 13750 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 6.640759493670886e-06, |
|
"loss": 0.0234, |
|
"step": 13775 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 6.6344303797468355e-06, |
|
"loss": 0.0202, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 6.628101265822785e-06, |
|
"loss": 0.0251, |
|
"step": 13825 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 6.621772151898734e-06, |
|
"loss": 0.017, |
|
"step": 13850 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 6.615443037974684e-06, |
|
"loss": 0.0254, |
|
"step": 13875 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 6.609113924050633e-06, |
|
"loss": 0.0289, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 6.602784810126583e-06, |
|
"loss": 0.0238, |
|
"step": 13925 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 6.596455696202532e-06, |
|
"loss": 0.0218, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 6.590126582278481e-06, |
|
"loss": 0.0289, |
|
"step": 13975 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 6.583797468354432e-06, |
|
"loss": 0.0271, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"eval_loss": 0.26684999465942383, |
|
"eval_runtime": 1448.0228, |
|
"eval_samples_per_second": 7.19, |
|
"eval_steps_per_second": 0.45, |
|
"eval_wer": 42.464105107400115, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 6.577468354430381e-06, |
|
"loss": 0.0292, |
|
"step": 14025 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 6.57113924050633e-06, |
|
"loss": 0.0192, |
|
"step": 14050 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 6.564810126582279e-06, |
|
"loss": 0.027, |
|
"step": 14075 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 6.558481012658229e-06, |
|
"loss": 0.03, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 6.552151898734178e-06, |
|
"loss": 0.0375, |
|
"step": 14125 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 6.5458227848101274e-06, |
|
"loss": 0.0338, |
|
"step": 14150 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 6.539493670886077e-06, |
|
"loss": 0.0305, |
|
"step": 14175 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 6.533164556962026e-06, |
|
"loss": 0.0266, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 6.526835443037976e-06, |
|
"loss": 0.0378, |
|
"step": 14225 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 6.520506329113925e-06, |
|
"loss": 0.0256, |
|
"step": 14250 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 6.514177215189874e-06, |
|
"loss": 0.0203, |
|
"step": 14275 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 6.507848101265823e-06, |
|
"loss": 0.0212, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 6.501518987341773e-06, |
|
"loss": 0.0204, |
|
"step": 14325 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 6.495189873417722e-06, |
|
"loss": 0.0259, |
|
"step": 14350 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 6.4888607594936714e-06, |
|
"loss": 0.0214, |
|
"step": 14375 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 6.482531645569621e-06, |
|
"loss": 0.021, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 6.47620253164557e-06, |
|
"loss": 0.0241, |
|
"step": 14425 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 6.46987341772152e-06, |
|
"loss": 0.0216, |
|
"step": 14450 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 6.463544303797469e-06, |
|
"loss": 0.0209, |
|
"step": 14475 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 6.457215189873418e-06, |
|
"loss": 0.0292, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 6.450886075949367e-06, |
|
"loss": 0.0286, |
|
"step": 14525 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 6.444556962025317e-06, |
|
"loss": 0.0218, |
|
"step": 14550 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 6.438227848101266e-06, |
|
"loss": 0.0209, |
|
"step": 14575 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 6.4318987341772154e-06, |
|
"loss": 0.0244, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 6.425569620253165e-06, |
|
"loss": 0.0212, |
|
"step": 14625 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 6.419240506329114e-06, |
|
"loss": 0.0155, |
|
"step": 14650 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 6.412911392405064e-06, |
|
"loss": 0.0209, |
|
"step": 14675 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 6.406582278481013e-06, |
|
"loss": 0.0203, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 6.400253164556962e-06, |
|
"loss": 0.0192, |
|
"step": 14725 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 6.393924050632911e-06, |
|
"loss": 0.0216, |
|
"step": 14750 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 6.387594936708861e-06, |
|
"loss": 0.012, |
|
"step": 14775 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 6.38126582278481e-06, |
|
"loss": 0.0241, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 6.37493670886076e-06, |
|
"loss": 0.0192, |
|
"step": 14825 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 6.36860759493671e-06, |
|
"loss": 0.0202, |
|
"step": 14850 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 6.362278481012659e-06, |
|
"loss": 0.0257, |
|
"step": 14875 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 6.3559493670886086e-06, |
|
"loss": 0.0218, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 6.349620253164558e-06, |
|
"loss": 0.0232, |
|
"step": 14925 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 6.3432911392405074e-06, |
|
"loss": 0.03, |
|
"step": 14950 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 6.337215189873418e-06, |
|
"loss": 0.0252, |
|
"step": 14975 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 6.3308860759493675e-06, |
|
"loss": 0.0176, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"eval_loss": 0.2596448063850403, |
|
"eval_runtime": 1417.6481, |
|
"eval_samples_per_second": 7.344, |
|
"eval_steps_per_second": 0.459, |
|
"eval_wer": 28.52241499861786, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 6.324556962025317e-06, |
|
"loss": 0.0166, |
|
"step": 15025 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 6.318227848101266e-06, |
|
"loss": 0.017, |
|
"step": 15050 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 6.311898734177216e-06, |
|
"loss": 0.0291, |
|
"step": 15075 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 6.305569620253164e-06, |
|
"loss": 0.0149, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 6.299240506329114e-06, |
|
"loss": 0.0223, |
|
"step": 15125 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 6.292911392405063e-06, |
|
"loss": 0.0212, |
|
"step": 15150 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 6.286582278481013e-06, |
|
"loss": 0.0125, |
|
"step": 15175 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 6.280253164556962e-06, |
|
"loss": 0.0169, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 6.273924050632912e-06, |
|
"loss": 0.0244, |
|
"step": 15225 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 6.267594936708862e-06, |
|
"loss": 0.015, |
|
"step": 15250 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 6.261265822784811e-06, |
|
"loss": 0.0171, |
|
"step": 15275 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 6.254936708860761e-06, |
|
"loss": 0.017, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 6.24860759493671e-06, |
|
"loss": 0.0162, |
|
"step": 15325 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 6.2422784810126594e-06, |
|
"loss": 0.0133, |
|
"step": 15350 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 6.235949367088608e-06, |
|
"loss": 0.0139, |
|
"step": 15375 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 6.2296202531645575e-06, |
|
"loss": 0.0177, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 6.223291139240507e-06, |
|
"loss": 0.0207, |
|
"step": 15425 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 6.216962025316456e-06, |
|
"loss": 0.0192, |
|
"step": 15450 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 6.210632911392406e-06, |
|
"loss": 0.0176, |
|
"step": 15475 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 6.204303797468355e-06, |
|
"loss": 0.0133, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 6.197974683544305e-06, |
|
"loss": 0.0199, |
|
"step": 15525 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 6.191645569620254e-06, |
|
"loss": 0.0135, |
|
"step": 15550 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 6.1853164556962035e-06, |
|
"loss": 0.0121, |
|
"step": 15575 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 6.178987341772152e-06, |
|
"loss": 0.0152, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 6.1726582278481015e-06, |
|
"loss": 0.0201, |
|
"step": 15625 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 6.166329113924051e-06, |
|
"loss": 0.0226, |
|
"step": 15650 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 6.16e-06, |
|
"loss": 0.0222, |
|
"step": 15675 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 6.15367088607595e-06, |
|
"loss": 0.0184, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 6.147341772151899e-06, |
|
"loss": 0.0219, |
|
"step": 15725 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 6.141012658227849e-06, |
|
"loss": 0.0219, |
|
"step": 15750 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 6.134683544303798e-06, |
|
"loss": 0.0208, |
|
"step": 15775 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 6.1283544303797475e-06, |
|
"loss": 0.0143, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 6.122025316455696e-06, |
|
"loss": 0.0254, |
|
"step": 15825 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 6.1156962025316455e-06, |
|
"loss": 0.0287, |
|
"step": 15850 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 6.109367088607595e-06, |
|
"loss": 0.0252, |
|
"step": 15875 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 6.103037974683544e-06, |
|
"loss": 0.0275, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 6.096708860759494e-06, |
|
"loss": 0.0193, |
|
"step": 15925 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 6.090379746835443e-06, |
|
"loss": 0.0255, |
|
"step": 15950 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 6.084050632911393e-06, |
|
"loss": 0.0248, |
|
"step": 15975 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 6.077721518987342e-06, |
|
"loss": 0.0328, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"eval_loss": 0.2512108087539673, |
|
"eval_runtime": 1474.2883, |
|
"eval_samples_per_second": 7.062, |
|
"eval_steps_per_second": 0.442, |
|
"eval_wer": 35.66236849379665, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 6.0713924050632915e-06, |
|
"loss": 0.0181, |
|
"step": 16025 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 6.065063291139242e-06, |
|
"loss": 0.0134, |
|
"step": 16050 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 6.058734177215191e-06, |
|
"loss": 0.0203, |
|
"step": 16075 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 6.052405063291141e-06, |
|
"loss": 0.0224, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 6.046075949367089e-06, |
|
"loss": 0.0164, |
|
"step": 16125 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 6.039746835443039e-06, |
|
"loss": 0.019, |
|
"step": 16150 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 6.033417721518988e-06, |
|
"loss": 0.0152, |
|
"step": 16175 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 6.0270886075949374e-06, |
|
"loss": 0.0197, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 6.020759493670887e-06, |
|
"loss": 0.0188, |
|
"step": 16225 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 6.014430379746836e-06, |
|
"loss": 0.0245, |
|
"step": 16250 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 6.008101265822786e-06, |
|
"loss": 0.0202, |
|
"step": 16275 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 6.001772151898735e-06, |
|
"loss": 0.0184, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 5.995443037974685e-06, |
|
"loss": 0.0158, |
|
"step": 16325 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 5.989113924050633e-06, |
|
"loss": 0.0175, |
|
"step": 16350 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.982784810126583e-06, |
|
"loss": 0.014, |
|
"step": 16375 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.976455696202532e-06, |
|
"loss": 0.0149, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.9701265822784815e-06, |
|
"loss": 0.0188, |
|
"step": 16425 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.963797468354431e-06, |
|
"loss": 0.0215, |
|
"step": 16450 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.95746835443038e-06, |
|
"loss": 0.017, |
|
"step": 16475 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.95113924050633e-06, |
|
"loss": 0.0095, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.944810126582279e-06, |
|
"loss": 0.0112, |
|
"step": 16525 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.938481012658229e-06, |
|
"loss": 0.0198, |
|
"step": 16550 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.932151898734177e-06, |
|
"loss": 0.0252, |
|
"step": 16575 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.925822784810127e-06, |
|
"loss": 0.0226, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.919493670886076e-06, |
|
"loss": 0.0167, |
|
"step": 16625 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.9131645569620255e-06, |
|
"loss": 0.0259, |
|
"step": 16650 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.906835443037975e-06, |
|
"loss": 0.0229, |
|
"step": 16675 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.900506329113924e-06, |
|
"loss": 0.0238, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.894177215189874e-06, |
|
"loss": 0.02, |
|
"step": 16725 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 5.887848101265823e-06, |
|
"loss": 0.0182, |
|
"step": 16750 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.881518987341773e-06, |
|
"loss": 0.0157, |
|
"step": 16775 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.875189873417721e-06, |
|
"loss": 0.0145, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.868860759493671e-06, |
|
"loss": 0.0166, |
|
"step": 16825 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.86253164556962e-06, |
|
"loss": 0.0108, |
|
"step": 16850 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.85620253164557e-06, |
|
"loss": 0.014, |
|
"step": 16875 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.84987341772152e-06, |
|
"loss": 0.0167, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.843544303797469e-06, |
|
"loss": 0.0121, |
|
"step": 16925 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.837215189873419e-06, |
|
"loss": 0.0149, |
|
"step": 16950 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.831139240506329e-06, |
|
"loss": 0.0143, |
|
"step": 16975 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.824810126582279e-06, |
|
"loss": 0.0139, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"eval_loss": 0.2688957154750824, |
|
"eval_runtime": 1423.1196, |
|
"eval_samples_per_second": 7.316, |
|
"eval_steps_per_second": 0.457, |
|
"eval_wer": 25.920746678807788, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.818481012658228e-06, |
|
"loss": 0.0169, |
|
"step": 17025 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.8121518987341775e-06, |
|
"loss": 0.013, |
|
"step": 17050 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.805822784810127e-06, |
|
"loss": 0.0123, |
|
"step": 17075 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.799493670886076e-06, |
|
"loss": 0.0161, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.793164556962026e-06, |
|
"loss": 0.0114, |
|
"step": 17125 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 5.786835443037974e-06, |
|
"loss": 0.0194, |
|
"step": 17150 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 5.780506329113924e-06, |
|
"loss": 0.0179, |
|
"step": 17175 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 5.774177215189873e-06, |
|
"loss": 0.0165, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 5.767848101265823e-06, |
|
"loss": 0.0165, |
|
"step": 17225 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 5.761518987341773e-06, |
|
"loss": 0.0125, |
|
"step": 17250 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 5.755189873417722e-06, |
|
"loss": 0.0172, |
|
"step": 17275 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 5.748860759493672e-06, |
|
"loss": 0.0158, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 5.742531645569621e-06, |
|
"loss": 0.0174, |
|
"step": 17325 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 5.736202531645571e-06, |
|
"loss": 0.0162, |
|
"step": 17350 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 5.72987341772152e-06, |
|
"loss": 0.0203, |
|
"step": 17375 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 5.7235443037974695e-06, |
|
"loss": 0.0157, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 5.717215189873419e-06, |
|
"loss": 0.018, |
|
"step": 17425 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 5.7108860759493675e-06, |
|
"loss": 0.0199, |
|
"step": 17450 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 5.704556962025317e-06, |
|
"loss": 0.0153, |
|
"step": 17475 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 5.698227848101266e-06, |
|
"loss": 0.0234, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 5.691898734177216e-06, |
|
"loss": 0.0168, |
|
"step": 17525 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 5.685569620253165e-06, |
|
"loss": 0.0214, |
|
"step": 17550 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 5.679240506329115e-06, |
|
"loss": 0.0185, |
|
"step": 17575 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 5.672911392405064e-06, |
|
"loss": 0.0228, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 5.6665822784810135e-06, |
|
"loss": 0.024, |
|
"step": 17625 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 5.660253164556963e-06, |
|
"loss": 0.0235, |
|
"step": 17650 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 5.6539240506329115e-06, |
|
"loss": 0.0262, |
|
"step": 17675 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 5.647594936708861e-06, |
|
"loss": 0.0214, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 5.64126582278481e-06, |
|
"loss": 0.0279, |
|
"step": 17725 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 5.63493670886076e-06, |
|
"loss": 0.0248, |
|
"step": 17750 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 5.628607594936709e-06, |
|
"loss": 0.0155, |
|
"step": 17775 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 5.622278481012659e-06, |
|
"loss": 0.017, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 5.615949367088608e-06, |
|
"loss": 0.0205, |
|
"step": 17825 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 5.6096202531645575e-06, |
|
"loss": 0.021, |
|
"step": 17850 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 5.603291139240507e-06, |
|
"loss": 0.0178, |
|
"step": 17875 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 5.5969620253164555e-06, |
|
"loss": 0.0169, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 5.590632911392405e-06, |
|
"loss": 0.0158, |
|
"step": 17925 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 5.584303797468354e-06, |
|
"loss": 0.0195, |
|
"step": 17950 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 5.577974683544304e-06, |
|
"loss": 0.0195, |
|
"step": 17975 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 5.571645569620253e-06, |
|
"loss": 0.0187, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"eval_loss": 0.2605174481868744, |
|
"eval_runtime": 1553.7854, |
|
"eval_samples_per_second": 6.7, |
|
"eval_steps_per_second": 0.419, |
|
"eval_wer": 25.330493178750874, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 5.565316455696203e-06, |
|
"loss": 0.0214, |
|
"step": 18025 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 5.558987341772152e-06, |
|
"loss": 0.0193, |
|
"step": 18050 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 5.552658227848102e-06, |
|
"loss": 0.0168, |
|
"step": 18075 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 5.546329113924052e-06, |
|
"loss": 0.0147, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 5.540000000000001e-06, |
|
"loss": 0.0192, |
|
"step": 18125 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 5.533670886075951e-06, |
|
"loss": 0.014, |
|
"step": 18150 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 5.527341772151899e-06, |
|
"loss": 0.0163, |
|
"step": 18175 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 5.521012658227849e-06, |
|
"loss": 0.0139, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 5.514683544303798e-06, |
|
"loss": 0.0168, |
|
"step": 18225 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 5.5083544303797475e-06, |
|
"loss": 0.0118, |
|
"step": 18250 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 5.502025316455697e-06, |
|
"loss": 0.0119, |
|
"step": 18275 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 5.495696202531646e-06, |
|
"loss": 0.0136, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 5.489367088607596e-06, |
|
"loss": 0.0132, |
|
"step": 18325 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 5.483037974683545e-06, |
|
"loss": 0.0148, |
|
"step": 18350 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 5.476708860759495e-06, |
|
"loss": 0.015, |
|
"step": 18375 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 5.470379746835443e-06, |
|
"loss": 0.0154, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 5.464050632911393e-06, |
|
"loss": 0.0139, |
|
"step": 18425 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 5.457721518987342e-06, |
|
"loss": 0.0239, |
|
"step": 18450 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 5.4513924050632915e-06, |
|
"loss": 0.0203, |
|
"step": 18475 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 5.445063291139241e-06, |
|
"loss": 0.0199, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 5.43873417721519e-06, |
|
"loss": 0.013, |
|
"step": 18525 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 5.43240506329114e-06, |
|
"loss": 0.0128, |
|
"step": 18550 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 5.426075949367089e-06, |
|
"loss": 0.0161, |
|
"step": 18575 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 5.419746835443039e-06, |
|
"loss": 0.0129, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 5.413417721518987e-06, |
|
"loss": 0.0124, |
|
"step": 18625 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 5.407088607594937e-06, |
|
"loss": 0.0161, |
|
"step": 18650 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 5.400759493670886e-06, |
|
"loss": 0.0104, |
|
"step": 18675 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 5.3944303797468355e-06, |
|
"loss": 0.0109, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 5.388101265822785e-06, |
|
"loss": 0.0128, |
|
"step": 18725 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 5.381772151898734e-06, |
|
"loss": 0.014, |
|
"step": 18750 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 5.375443037974684e-06, |
|
"loss": 0.0159, |
|
"step": 18775 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 5.369113924050633e-06, |
|
"loss": 0.017, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 5.362784810126583e-06, |
|
"loss": 0.012, |
|
"step": 18825 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 5.356455696202531e-06, |
|
"loss": 0.0165, |
|
"step": 18850 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 5.350126582278481e-06, |
|
"loss": 0.0153, |
|
"step": 18875 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 5.343797468354432e-06, |
|
"loss": 0.0182, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"learning_rate": 5.33746835443038e-06, |
|
"loss": 0.0194, |
|
"step": 18925 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"learning_rate": 5.33113924050633e-06, |
|
"loss": 0.0144, |
|
"step": 18950 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"learning_rate": 5.324810126582279e-06, |
|
"loss": 0.0157, |
|
"step": 18975 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"learning_rate": 5.318481012658229e-06, |
|
"loss": 0.0111, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"eval_loss": 0.2696678638458252, |
|
"eval_runtime": 1463.5986, |
|
"eval_samples_per_second": 7.113, |
|
"eval_steps_per_second": 0.445, |
|
"eval_wer": 24.79877721588969, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"learning_rate": 5.312151898734178e-06, |
|
"loss": 0.0176, |
|
"step": 19025 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"learning_rate": 5.3058227848101275e-06, |
|
"loss": 0.0144, |
|
"step": 19050 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"learning_rate": 5.299493670886077e-06, |
|
"loss": 0.0129, |
|
"step": 19075 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"learning_rate": 5.293164556962026e-06, |
|
"loss": 0.0162, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"learning_rate": 5.286835443037976e-06, |
|
"loss": 0.0128, |
|
"step": 19125 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"learning_rate": 5.280506329113924e-06, |
|
"loss": 0.0152, |
|
"step": 19150 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"learning_rate": 5.274177215189874e-06, |
|
"loss": 0.0183, |
|
"step": 19175 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"learning_rate": 5.267848101265823e-06, |
|
"loss": 0.0198, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"learning_rate": 5.261518987341773e-06, |
|
"loss": 0.0179, |
|
"step": 19225 |
|
}, |
|
{ |
|
"epoch": 10.04, |
|
"learning_rate": 5.255443037974683e-06, |
|
"loss": 0.019, |
|
"step": 19250 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 5.249113924050633e-06, |
|
"loss": 0.0159, |
|
"step": 19275 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 5.242784810126583e-06, |
|
"loss": 0.0156, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 5.236455696202532e-06, |
|
"loss": 0.015, |
|
"step": 19325 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 5.230126582278482e-06, |
|
"loss": 0.0162, |
|
"step": 19350 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 5.223797468354431e-06, |
|
"loss": 0.0262, |
|
"step": 19375 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 5.217468354430381e-06, |
|
"loss": 0.0207, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 5.21113924050633e-06, |
|
"loss": 0.0222, |
|
"step": 19425 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 5.2048101265822795e-06, |
|
"loss": 0.016, |
|
"step": 19450 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.198481012658229e-06, |
|
"loss": 0.0176, |
|
"step": 19475 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.192151898734178e-06, |
|
"loss": 0.0209, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.185822784810127e-06, |
|
"loss": 0.0212, |
|
"step": 19525 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.179493670886076e-06, |
|
"loss": 0.0143, |
|
"step": 19550 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.173164556962026e-06, |
|
"loss": 0.0224, |
|
"step": 19575 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.166835443037975e-06, |
|
"loss": 0.0171, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.160506329113925e-06, |
|
"loss": 0.0166, |
|
"step": 19625 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.154177215189874e-06, |
|
"loss": 0.0135, |
|
"step": 19650 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.1478481012658235e-06, |
|
"loss": 0.0164, |
|
"step": 19675 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.141518987341773e-06, |
|
"loss": 0.0154, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.135189873417722e-06, |
|
"loss": 0.0183, |
|
"step": 19725 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.128860759493671e-06, |
|
"loss": 0.0149, |
|
"step": 19750 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.12253164556962e-06, |
|
"loss": 0.0176, |
|
"step": 19775 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.11620253164557e-06, |
|
"loss": 0.0135, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.109873417721519e-06, |
|
"loss": 0.0121, |
|
"step": 19825 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.103544303797469e-06, |
|
"loss": 0.0158, |
|
"step": 19850 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 5.097215189873418e-06, |
|
"loss": 0.0191, |
|
"step": 19875 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 5.0908860759493675e-06, |
|
"loss": 0.0127, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 5.084556962025317e-06, |
|
"loss": 0.0149, |
|
"step": 19925 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 5.0782278481012655e-06, |
|
"loss": 0.0152, |
|
"step": 19950 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 5.071898734177215e-06, |
|
"loss": 0.0139, |
|
"step": 19975 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 5.065569620253164e-06, |
|
"loss": 0.015, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"eval_loss": 0.2623814344406128, |
|
"eval_runtime": 1462.1621, |
|
"eval_samples_per_second": 7.12, |
|
"eval_steps_per_second": 0.445, |
|
"eval_wer": 24.392266540919362, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 5.059240506329114e-06, |
|
"loss": 0.0094, |
|
"step": 20025 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 5.052911392405063e-06, |
|
"loss": 0.0108, |
|
"step": 20050 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 5.046582278481013e-06, |
|
"loss": 0.0155, |
|
"step": 20075 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 5.040253164556962e-06, |
|
"loss": 0.0103, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 5.033924050632912e-06, |
|
"loss": 0.0109, |
|
"step": 20125 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 5.027594936708862e-06, |
|
"loss": 0.0188, |
|
"step": 20150 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 5.021265822784811e-06, |
|
"loss": 0.0178, |
|
"step": 20175 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 5.014936708860761e-06, |
|
"loss": 0.0262, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 5.00860759493671e-06, |
|
"loss": 0.0194, |
|
"step": 20225 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 5.002278481012659e-06, |
|
"loss": 0.0107, |
|
"step": 20250 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 4.995949367088608e-06, |
|
"loss": 0.0147, |
|
"step": 20275 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 4.9896202531645575e-06, |
|
"loss": 0.0084, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 4.983291139240507e-06, |
|
"loss": 0.0214, |
|
"step": 20325 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 4.976962025316456e-06, |
|
"loss": 0.014, |
|
"step": 20350 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 4.970632911392406e-06, |
|
"loss": 0.0129, |
|
"step": 20375 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 4.964303797468355e-06, |
|
"loss": 0.0145, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 4.957974683544305e-06, |
|
"loss": 0.0109, |
|
"step": 20425 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 4.951645569620254e-06, |
|
"loss": 0.0117, |
|
"step": 20450 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 4.945316455696203e-06, |
|
"loss": 0.0137, |
|
"step": 20475 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 4.938987341772152e-06, |
|
"loss": 0.0086, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 4.9326582278481015e-06, |
|
"loss": 0.0128, |
|
"step": 20525 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 4.926329113924051e-06, |
|
"loss": 0.0101, |
|
"step": 20550 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 4.92e-06, |
|
"loss": 0.0132, |
|
"step": 20575 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 4.91367088607595e-06, |
|
"loss": 0.0147, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 4.907341772151899e-06, |
|
"loss": 0.0077, |
|
"step": 20625 |
|
}, |
|
{ |
|
"epoch": 11.03, |
|
"learning_rate": 4.901012658227849e-06, |
|
"loss": 0.0125, |
|
"step": 20650 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 4.894683544303798e-06, |
|
"loss": 0.0115, |
|
"step": 20675 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 4.888354430379747e-06, |
|
"loss": 0.0136, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 4.882025316455696e-06, |
|
"loss": 0.0098, |
|
"step": 20725 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 4.875696202531646e-06, |
|
"loss": 0.0154, |
|
"step": 20750 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 4.869367088607596e-06, |
|
"loss": 0.0126, |
|
"step": 20775 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 4.863037974683545e-06, |
|
"loss": 0.0116, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 4.856708860759495e-06, |
|
"loss": 0.0133, |
|
"step": 20825 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 4.850379746835443e-06, |
|
"loss": 0.0106, |
|
"step": 20850 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 4.844050632911393e-06, |
|
"loss": 0.0151, |
|
"step": 20875 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 4.837721518987342e-06, |
|
"loss": 0.0148, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 4.8313924050632915e-06, |
|
"loss": 0.0173, |
|
"step": 20925 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 4.825063291139241e-06, |
|
"loss": 0.0126, |
|
"step": 20950 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 4.81873417721519e-06, |
|
"loss": 0.0163, |
|
"step": 20975 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 4.81240506329114e-06, |
|
"loss": 0.0203, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"eval_loss": 0.27788785099983215, |
|
"eval_runtime": 1546.852, |
|
"eval_samples_per_second": 6.73, |
|
"eval_steps_per_second": 0.421, |
|
"eval_wer": 25.284963983154196, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"learning_rate": 4.806075949367089e-06, |
|
"loss": 0.0176, |
|
"step": 21025 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"learning_rate": 4.799746835443039e-06, |
|
"loss": 0.0157, |
|
"step": 21050 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"learning_rate": 4.793417721518987e-06, |
|
"loss": 0.0192, |
|
"step": 21075 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"learning_rate": 4.787088607594937e-06, |
|
"loss": 0.0124, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"learning_rate": 4.780759493670886e-06, |
|
"loss": 0.0194, |
|
"step": 21125 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"learning_rate": 4.774430379746836e-06, |
|
"loss": 0.0138, |
|
"step": 21150 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"learning_rate": 4.768101265822786e-06, |
|
"loss": 0.0191, |
|
"step": 21175 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"learning_rate": 4.761772151898735e-06, |
|
"loss": 0.0124, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 4.755443037974684e-06, |
|
"loss": 0.0175, |
|
"step": 21225 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 4.749113924050633e-06, |
|
"loss": 0.0118, |
|
"step": 21250 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 4.742784810126583e-06, |
|
"loss": 0.0172, |
|
"step": 21275 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 4.736455696202532e-06, |
|
"loss": 0.0136, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 4.7301265822784815e-06, |
|
"loss": 0.018, |
|
"step": 21325 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 4.723797468354431e-06, |
|
"loss": 0.0184, |
|
"step": 21350 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 4.71746835443038e-06, |
|
"loss": 0.0166, |
|
"step": 21375 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 4.71113924050633e-06, |
|
"loss": 0.0107, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 4.704810126582279e-06, |
|
"loss": 0.0125, |
|
"step": 21425 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 4.698481012658228e-06, |
|
"loss": 0.0098, |
|
"step": 21450 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 4.692151898734177e-06, |
|
"loss": 0.0103, |
|
"step": 21475 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 4.685822784810127e-06, |
|
"loss": 0.0211, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 4.679493670886076e-06, |
|
"loss": 0.0125, |
|
"step": 21525 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 4.6731645569620255e-06, |
|
"loss": 0.0083, |
|
"step": 21550 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 4.666835443037975e-06, |
|
"loss": 0.0156, |
|
"step": 21575 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 4.660506329113924e-06, |
|
"loss": 0.0137, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 4.654177215189874e-06, |
|
"loss": 0.0137, |
|
"step": 21625 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 4.647848101265823e-06, |
|
"loss": 0.0092, |
|
"step": 21650 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 4.641518987341773e-06, |
|
"loss": 0.0101, |
|
"step": 21675 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 4.635189873417722e-06, |
|
"loss": 0.0118, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 4.6288607594936715e-06, |
|
"loss": 0.0119, |
|
"step": 21725 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 4.622531645569621e-06, |
|
"loss": 0.009, |
|
"step": 21750 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 4.61620253164557e-06, |
|
"loss": 0.014, |
|
"step": 21775 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 4.609873417721519e-06, |
|
"loss": 0.0108, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 4.603544303797468e-06, |
|
"loss": 0.0117, |
|
"step": 21825 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 4.597215189873418e-06, |
|
"loss": 0.0147, |
|
"step": 21850 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 4.590886075949367e-06, |
|
"loss": 0.0133, |
|
"step": 21875 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 4.584556962025317e-06, |
|
"loss": 0.0131, |
|
"step": 21900 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 4.578227848101266e-06, |
|
"loss": 0.0156, |
|
"step": 21925 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 4.5718987341772155e-06, |
|
"loss": 0.0211, |
|
"step": 21950 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 4.565569620253165e-06, |
|
"loss": 0.0127, |
|
"step": 21975 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 4.559240506329114e-06, |
|
"loss": 0.011, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"eval_loss": 0.26152506470680237, |
|
"eval_runtime": 1497.9019, |
|
"eval_samples_per_second": 6.95, |
|
"eval_steps_per_second": 0.435, |
|
"eval_wer": 26.236198962584755, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 4.552911392405064e-06, |
|
"loss": 0.0164, |
|
"step": 22025 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 4.546582278481013e-06, |
|
"loss": 0.01, |
|
"step": 22050 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 4.540253164556963e-06, |
|
"loss": 0.0127, |
|
"step": 22075 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 4.533924050632912e-06, |
|
"loss": 0.0089, |
|
"step": 22100 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 4.5275949367088615e-06, |
|
"loss": 0.0077, |
|
"step": 22125 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 4.521265822784811e-06, |
|
"loss": 0.0098, |
|
"step": 22150 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 4.5149367088607595e-06, |
|
"loss": 0.0084, |
|
"step": 22175 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 4.508607594936709e-06, |
|
"loss": 0.0096, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 4.502278481012658e-06, |
|
"loss": 0.0097, |
|
"step": 22225 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 4.495949367088608e-06, |
|
"loss": 0.0114, |
|
"step": 22250 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 4.489620253164557e-06, |
|
"loss": 0.0109, |
|
"step": 22275 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 4.483291139240507e-06, |
|
"loss": 0.0096, |
|
"step": 22300 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 4.476962025316456e-06, |
|
"loss": 0.0117, |
|
"step": 22325 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 4.4706329113924055e-06, |
|
"loss": 0.0105, |
|
"step": 22350 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 4.464303797468355e-06, |
|
"loss": 0.0103, |
|
"step": 22375 |
|
}, |
|
{ |
|
"epoch": 12.03, |
|
"learning_rate": 4.457974683544304e-06, |
|
"loss": 0.0085, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 12.04, |
|
"learning_rate": 4.451645569620254e-06, |
|
"loss": 0.0129, |
|
"step": 22425 |
|
}, |
|
{ |
|
"epoch": 12.04, |
|
"learning_rate": 4.445316455696203e-06, |
|
"loss": 0.0188, |
|
"step": 22450 |
|
}, |
|
{ |
|
"epoch": 12.04, |
|
"learning_rate": 4.438987341772153e-06, |
|
"loss": 0.013, |
|
"step": 22475 |
|
}, |
|
{ |
|
"epoch": 12.04, |
|
"learning_rate": 4.432658227848102e-06, |
|
"loss": 0.0095, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 12.04, |
|
"learning_rate": 4.4263291139240515e-06, |
|
"loss": 0.0111, |
|
"step": 22525 |
|
}, |
|
{ |
|
"epoch": 12.04, |
|
"learning_rate": 4.42e-06, |
|
"loss": 0.0115, |
|
"step": 22550 |
|
}, |
|
{ |
|
"epoch": 12.04, |
|
"learning_rate": 4.4136708860759495e-06, |
|
"loss": 0.0096, |
|
"step": 22575 |
|
}, |
|
{ |
|
"epoch": 12.04, |
|
"learning_rate": 4.407341772151899e-06, |
|
"loss": 0.0108, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 12.04, |
|
"learning_rate": 4.401012658227848e-06, |
|
"loss": 0.0118, |
|
"step": 22625 |
|
}, |
|
{ |
|
"epoch": 12.04, |
|
"learning_rate": 4.394683544303798e-06, |
|
"loss": 0.0136, |
|
"step": 22650 |
|
}, |
|
{ |
|
"epoch": 12.04, |
|
"learning_rate": 4.388354430379747e-06, |
|
"loss": 0.0176, |
|
"step": 22675 |
|
}, |
|
{ |
|
"epoch": 12.04, |
|
"learning_rate": 4.382025316455697e-06, |
|
"loss": 0.011, |
|
"step": 22700 |
|
}, |
|
{ |
|
"epoch": 12.04, |
|
"learning_rate": 4.375696202531646e-06, |
|
"loss": 0.0125, |
|
"step": 22725 |
|
}, |
|
{ |
|
"epoch": 12.04, |
|
"learning_rate": 4.3693670886075955e-06, |
|
"loss": 0.0164, |
|
"step": 22750 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"learning_rate": 4.363037974683544e-06, |
|
"loss": 0.0192, |
|
"step": 22775 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"learning_rate": 4.356708860759494e-06, |
|
"loss": 0.0145, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"learning_rate": 4.350379746835444e-06, |
|
"loss": 0.0174, |
|
"step": 22825 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"learning_rate": 4.344050632911393e-06, |
|
"loss": 0.0142, |
|
"step": 22850 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"learning_rate": 4.337721518987343e-06, |
|
"loss": 0.0228, |
|
"step": 22875 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"learning_rate": 4.331392405063292e-06, |
|
"loss": 0.0201, |
|
"step": 22900 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"learning_rate": 4.325063291139241e-06, |
|
"loss": 0.0188, |
|
"step": 22925 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"learning_rate": 4.31873417721519e-06, |
|
"loss": 0.0185, |
|
"step": 22950 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 4.3124050632911395e-06, |
|
"loss": 0.0186, |
|
"step": 22975 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 4.306075949367089e-06, |
|
"loss": 0.0191, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"eval_loss": 0.25467297434806824, |
|
"eval_runtime": 1486.1182, |
|
"eval_samples_per_second": 7.005, |
|
"eval_steps_per_second": 0.438, |
|
"eval_wer": 23.982503780549276, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 4.299746835443038e-06, |
|
"loss": 0.0145, |
|
"step": 23025 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 4.293417721518988e-06, |
|
"loss": 0.0106, |
|
"step": 23050 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 4.287088607594937e-06, |
|
"loss": 0.0135, |
|
"step": 23075 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 4.280759493670887e-06, |
|
"loss": 0.0173, |
|
"step": 23100 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 4.274430379746836e-06, |
|
"loss": 0.0126, |
|
"step": 23125 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 4.268101265822785e-06, |
|
"loss": 0.0166, |
|
"step": 23150 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 4.261772151898734e-06, |
|
"loss": 0.0097, |
|
"step": 23175 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 4.255443037974684e-06, |
|
"loss": 0.0094, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 4.249113924050634e-06, |
|
"loss": 0.012, |
|
"step": 23225 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 4.242784810126583e-06, |
|
"loss": 0.0153, |
|
"step": 23250 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 4.236455696202532e-06, |
|
"loss": 0.0169, |
|
"step": 23275 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 4.230379746835443e-06, |
|
"loss": 0.0097, |
|
"step": 23300 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 4.224050632911393e-06, |
|
"loss": 0.0098, |
|
"step": 23325 |
|
}, |
|
{ |
|
"epoch": 13.01, |
|
"learning_rate": 4.217721518987342e-06, |
|
"loss": 0.0129, |
|
"step": 23350 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 4.2113924050632915e-06, |
|
"loss": 0.0157, |
|
"step": 23375 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 4.205063291139241e-06, |
|
"loss": 0.0088, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 4.19873417721519e-06, |
|
"loss": 0.0105, |
|
"step": 23425 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 4.19240506329114e-06, |
|
"loss": 0.0131, |
|
"step": 23450 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 4.186075949367089e-06, |
|
"loss": 0.0133, |
|
"step": 23475 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 4.179746835443038e-06, |
|
"loss": 0.0054, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 4.173417721518987e-06, |
|
"loss": 0.01, |
|
"step": 23525 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 4.167088607594937e-06, |
|
"loss": 0.013, |
|
"step": 23550 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 4.160759493670886e-06, |
|
"loss": 0.0141, |
|
"step": 23575 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 4.154430379746836e-06, |
|
"loss": 0.0126, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 4.148101265822786e-06, |
|
"loss": 0.015, |
|
"step": 23625 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 4.141772151898734e-06, |
|
"loss": 0.0164, |
|
"step": 23650 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 4.135443037974684e-06, |
|
"loss": 0.0135, |
|
"step": 23675 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 4.129113924050633e-06, |
|
"loss": 0.0151, |
|
"step": 23700 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 4.122784810126583e-06, |
|
"loss": 0.0157, |
|
"step": 23725 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 4.116455696202532e-06, |
|
"loss": 0.0087, |
|
"step": 23750 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 4.1101265822784815e-06, |
|
"loss": 0.0097, |
|
"step": 23775 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 4.103797468354431e-06, |
|
"loss": 0.0141, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 4.09746835443038e-06, |
|
"loss": 0.0079, |
|
"step": 23825 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 4.09113924050633e-06, |
|
"loss": 0.0116, |
|
"step": 23850 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 4.084810126582278e-06, |
|
"loss": 0.0093, |
|
"step": 23875 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 4.078481012658228e-06, |
|
"loss": 0.0149, |
|
"step": 23900 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 4.072151898734177e-06, |
|
"loss": 0.0125, |
|
"step": 23925 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 4.065822784810127e-06, |
|
"loss": 0.0106, |
|
"step": 23950 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 4.059493670886076e-06, |
|
"loss": 0.0124, |
|
"step": 23975 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 4.0531645569620255e-06, |
|
"loss": 0.0074, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"eval_loss": 0.27036768198013306, |
|
"eval_runtime": 1458.9195, |
|
"eval_samples_per_second": 7.136, |
|
"eval_steps_per_second": 0.446, |
|
"eval_wer": 24.153238264036816, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 4.046835443037975e-06, |
|
"loss": 0.0128, |
|
"step": 24025 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 4.040506329113924e-06, |
|
"loss": 0.012, |
|
"step": 24050 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 4.034177215189874e-06, |
|
"loss": 0.0098, |
|
"step": 24075 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 4.027848101265823e-06, |
|
"loss": 0.0091, |
|
"step": 24100 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 4.021518987341773e-06, |
|
"loss": 0.0077, |
|
"step": 24125 |
|
}, |
|
{ |
|
"epoch": 13.03, |
|
"learning_rate": 4.015189873417722e-06, |
|
"loss": 0.0135, |
|
"step": 24150 |
|
}, |
|
{ |
|
"epoch": 13.04, |
|
"learning_rate": 4.0088607594936715e-06, |
|
"loss": 0.0124, |
|
"step": 24175 |
|
}, |
|
{ |
|
"epoch": 13.04, |
|
"learning_rate": 4.002531645569621e-06, |
|
"loss": 0.0152, |
|
"step": 24200 |
|
}, |
|
{ |
|
"epoch": 13.04, |
|
"learning_rate": 3.99620253164557e-06, |
|
"loss": 0.0127, |
|
"step": 24225 |
|
}, |
|
{ |
|
"epoch": 13.04, |
|
"learning_rate": 3.989873417721519e-06, |
|
"loss": 0.0066, |
|
"step": 24250 |
|
}, |
|
{ |
|
"epoch": 13.04, |
|
"learning_rate": 3.983544303797468e-06, |
|
"loss": 0.0163, |
|
"step": 24275 |
|
}, |
|
{ |
|
"epoch": 13.04, |
|
"learning_rate": 3.977215189873418e-06, |
|
"loss": 0.0095, |
|
"step": 24300 |
|
}, |
|
{ |
|
"epoch": 13.04, |
|
"learning_rate": 3.970886075949367e-06, |
|
"loss": 0.0103, |
|
"step": 24325 |
|
}, |
|
{ |
|
"epoch": 13.04, |
|
"learning_rate": 3.964556962025317e-06, |
|
"loss": 0.0112, |
|
"step": 24350 |
|
}, |
|
{ |
|
"epoch": 13.04, |
|
"learning_rate": 3.958227848101266e-06, |
|
"loss": 0.01, |
|
"step": 24375 |
|
}, |
|
{ |
|
"epoch": 13.04, |
|
"learning_rate": 3.9518987341772155e-06, |
|
"loss": 0.0131, |
|
"step": 24400 |
|
}, |
|
{ |
|
"epoch": 13.04, |
|
"learning_rate": 3.945569620253165e-06, |
|
"loss": 0.0143, |
|
"step": 24425 |
|
}, |
|
{ |
|
"epoch": 13.04, |
|
"learning_rate": 3.939240506329114e-06, |
|
"loss": 0.0151, |
|
"step": 24450 |
|
}, |
|
{ |
|
"epoch": 13.04, |
|
"learning_rate": 3.932911392405064e-06, |
|
"loss": 0.0129, |
|
"step": 24475 |
|
}, |
|
{ |
|
"epoch": 13.04, |
|
"learning_rate": 3.926582278481013e-06, |
|
"loss": 0.0125, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"learning_rate": 3.920506329113925e-06, |
|
"loss": 0.0165, |
|
"step": 24525 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"learning_rate": 3.914177215189874e-06, |
|
"loss": 0.0114, |
|
"step": 24550 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"learning_rate": 3.9078481012658235e-06, |
|
"loss": 0.0111, |
|
"step": 24575 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"learning_rate": 3.901518987341772e-06, |
|
"loss": 0.0134, |
|
"step": 24600 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"learning_rate": 3.8951898734177215e-06, |
|
"loss": 0.0138, |
|
"step": 24625 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"learning_rate": 3.888860759493671e-06, |
|
"loss": 0.0151, |
|
"step": 24650 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"learning_rate": 3.88253164556962e-06, |
|
"loss": 0.0101, |
|
"step": 24675 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"learning_rate": 3.87620253164557e-06, |
|
"loss": 0.0108, |
|
"step": 24700 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 3.869873417721519e-06, |
|
"loss": 0.0151, |
|
"step": 24725 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 3.863544303797469e-06, |
|
"loss": 0.0132, |
|
"step": 24750 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 3.857215189873418e-06, |
|
"loss": 0.012, |
|
"step": 24775 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 3.8508860759493675e-06, |
|
"loss": 0.017, |
|
"step": 24800 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 3.844556962025317e-06, |
|
"loss": 0.0141, |
|
"step": 24825 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 3.838227848101266e-06, |
|
"loss": 0.0166, |
|
"step": 24850 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 3.831898734177216e-06, |
|
"loss": 0.0112, |
|
"step": 24875 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 3.825569620253165e-06, |
|
"loss": 0.0097, |
|
"step": 24900 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 3.819240506329115e-06, |
|
"loss": 0.0138, |
|
"step": 24925 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 3.8129113924050636e-06, |
|
"loss": 0.0069, |
|
"step": 24950 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 3.806582278481013e-06, |
|
"loss": 0.0111, |
|
"step": 24975 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 3.8002531645569625e-06, |
|
"loss": 0.0149, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"eval_loss": 0.2640444040298462, |
|
"eval_runtime": 1478.4977, |
|
"eval_samples_per_second": 7.042, |
|
"eval_steps_per_second": 0.44, |
|
"eval_wer": 24.063805915543345, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 3.7939240506329115e-06, |
|
"loss": 0.0176, |
|
"step": 25025 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 3.787594936708861e-06, |
|
"loss": 0.0104, |
|
"step": 25050 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 3.7812658227848103e-06, |
|
"loss": 0.0084, |
|
"step": 25075 |
|
}, |
|
{ |
|
"epoch": 14.01, |
|
"learning_rate": 3.7749367088607598e-06, |
|
"loss": 0.0118, |
|
"step": 25100 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 3.768607594936709e-06, |
|
"loss": 0.0099, |
|
"step": 25125 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 3.762278481012658e-06, |
|
"loss": 0.0089, |
|
"step": 25150 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 3.7559493670886076e-06, |
|
"loss": 0.0119, |
|
"step": 25175 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 3.749620253164557e-06, |
|
"loss": 0.0118, |
|
"step": 25200 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 3.743291139240507e-06, |
|
"loss": 0.0108, |
|
"step": 25225 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 3.7369620253164563e-06, |
|
"loss": 0.0101, |
|
"step": 25250 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 3.7306329113924058e-06, |
|
"loss": 0.0095, |
|
"step": 25275 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 3.7243037974683548e-06, |
|
"loss": 0.0106, |
|
"step": 25300 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 3.717974683544304e-06, |
|
"loss": 0.0135, |
|
"step": 25325 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 3.7116455696202536e-06, |
|
"loss": 0.0115, |
|
"step": 25350 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 3.705316455696203e-06, |
|
"loss": 0.0114, |
|
"step": 25375 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 3.698987341772152e-06, |
|
"loss": 0.0119, |
|
"step": 25400 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 3.6926582278481015e-06, |
|
"loss": 0.0123, |
|
"step": 25425 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 3.686329113924051e-06, |
|
"loss": 0.0138, |
|
"step": 25450 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 3.6800000000000003e-06, |
|
"loss": 0.0133, |
|
"step": 25475 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 3.6736708860759498e-06, |
|
"loss": 0.0106, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.6673417721518988e-06, |
|
"loss": 0.0099, |
|
"step": 25525 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.661012658227848e-06, |
|
"loss": 0.0104, |
|
"step": 25550 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.6546835443037976e-06, |
|
"loss": 0.0143, |
|
"step": 25575 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.648354430379747e-06, |
|
"loss": 0.0096, |
|
"step": 25600 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.642025316455696e-06, |
|
"loss": 0.0094, |
|
"step": 25625 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.6356962025316463e-06, |
|
"loss": 0.0162, |
|
"step": 25650 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.6293670886075953e-06, |
|
"loss": 0.008, |
|
"step": 25675 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.6230379746835448e-06, |
|
"loss": 0.0094, |
|
"step": 25700 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.616708860759494e-06, |
|
"loss": 0.0122, |
|
"step": 25725 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.6103797468354436e-06, |
|
"loss": 0.0128, |
|
"step": 25750 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.6040506329113926e-06, |
|
"loss": 0.0117, |
|
"step": 25775 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.597721518987342e-06, |
|
"loss": 0.0067, |
|
"step": 25800 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.5913924050632915e-06, |
|
"loss": 0.0078, |
|
"step": 25825 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.585063291139241e-06, |
|
"loss": 0.0071, |
|
"step": 25850 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.57873417721519e-06, |
|
"loss": 0.0093, |
|
"step": 25875 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 3.5724050632911393e-06, |
|
"loss": 0.0064, |
|
"step": 25900 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 3.5660759493670888e-06, |
|
"loss": 0.0067, |
|
"step": 25925 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 3.559746835443038e-06, |
|
"loss": 0.0076, |
|
"step": 25950 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 3.5534177215189876e-06, |
|
"loss": 0.0099, |
|
"step": 25975 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 3.5470886075949366e-06, |
|
"loss": 0.0111, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"eval_loss": 0.2734878659248352, |
|
"eval_runtime": 1456.7487, |
|
"eval_samples_per_second": 7.147, |
|
"eval_steps_per_second": 0.447, |
|
"eval_wer": 24.359745686921737, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 3.540759493670886e-06, |
|
"loss": 0.0134, |
|
"step": 26025 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 3.534430379746836e-06, |
|
"loss": 0.008, |
|
"step": 26050 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 3.5281012658227853e-06, |
|
"loss": 0.0138, |
|
"step": 26075 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 3.5217721518987348e-06, |
|
"loss": 0.0075, |
|
"step": 26100 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 3.515443037974684e-06, |
|
"loss": 0.0097, |
|
"step": 26125 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 3.509113924050633e-06, |
|
"loss": 0.0139, |
|
"step": 26150 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 3.5027848101265826e-06, |
|
"loss": 0.0099, |
|
"step": 26175 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 3.496455696202532e-06, |
|
"loss": 0.0148, |
|
"step": 26200 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 3.4901265822784815e-06, |
|
"loss": 0.0119, |
|
"step": 26225 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 3.4837974683544305e-06, |
|
"loss": 0.0086, |
|
"step": 26250 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 3.47746835443038e-06, |
|
"loss": 0.0091, |
|
"step": 26275 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 3.4711392405063293e-06, |
|
"loss": 0.0105, |
|
"step": 26300 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 3.465063291139241e-06, |
|
"loss": 0.0138, |
|
"step": 26325 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 3.45873417721519e-06, |
|
"loss": 0.0185, |
|
"step": 26350 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 3.4524050632911392e-06, |
|
"loss": 0.015, |
|
"step": 26375 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 3.4460759493670886e-06, |
|
"loss": 0.0177, |
|
"step": 26400 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 3.439746835443038e-06, |
|
"loss": 0.014, |
|
"step": 26425 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 3.433417721518988e-06, |
|
"loss": 0.0127, |
|
"step": 26450 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 3.4270886075949374e-06, |
|
"loss": 0.0135, |
|
"step": 26475 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 3.4207594936708864e-06, |
|
"loss": 0.0171, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 3.4144303797468358e-06, |
|
"loss": 0.0114, |
|
"step": 26525 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 3.4081012658227852e-06, |
|
"loss": 0.0099, |
|
"step": 26550 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 3.4017721518987346e-06, |
|
"loss": 0.0126, |
|
"step": 26575 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 3.395443037974684e-06, |
|
"loss": 0.0116, |
|
"step": 26600 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 3.389113924050633e-06, |
|
"loss": 0.0087, |
|
"step": 26625 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 3.3827848101265825e-06, |
|
"loss": 0.0073, |
|
"step": 26650 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 3.376455696202532e-06, |
|
"loss": 0.0102, |
|
"step": 26675 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 3.3701265822784814e-06, |
|
"loss": 0.0106, |
|
"step": 26700 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 3.3637974683544304e-06, |
|
"loss": 0.0094, |
|
"step": 26725 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 3.35746835443038e-06, |
|
"loss": 0.0113, |
|
"step": 26750 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 3.3511392405063292e-06, |
|
"loss": 0.0169, |
|
"step": 26775 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 3.3448101265822786e-06, |
|
"loss": 0.0084, |
|
"step": 26800 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 3.338481012658228e-06, |
|
"loss": 0.0153, |
|
"step": 26825 |
|
}, |
|
{ |
|
"epoch": 15.01, |
|
"learning_rate": 3.332151898734177e-06, |
|
"loss": 0.012, |
|
"step": 26850 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 3.325822784810127e-06, |
|
"loss": 0.0106, |
|
"step": 26875 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 3.3194936708860764e-06, |
|
"loss": 0.0099, |
|
"step": 26900 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 3.3131645569620258e-06, |
|
"loss": 0.008, |
|
"step": 26925 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 3.306835443037975e-06, |
|
"loss": 0.0076, |
|
"step": 26950 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 3.3005063291139246e-06, |
|
"loss": 0.0088, |
|
"step": 26975 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 3.2941772151898736e-06, |
|
"loss": 0.009, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"eval_loss": 0.2693029046058655, |
|
"eval_runtime": 1479.1374, |
|
"eval_samples_per_second": 7.039, |
|
"eval_steps_per_second": 0.44, |
|
"eval_wer": 24.28819980812696, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 3.287848101265823e-06, |
|
"loss": 0.0115, |
|
"step": 27025 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 3.2815189873417725e-06, |
|
"loss": 0.0099, |
|
"step": 27050 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 3.275189873417722e-06, |
|
"loss": 0.0073, |
|
"step": 27075 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 3.268860759493671e-06, |
|
"loss": 0.0061, |
|
"step": 27100 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 3.2625316455696204e-06, |
|
"loss": 0.007, |
|
"step": 27125 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 3.2562025316455698e-06, |
|
"loss": 0.0109, |
|
"step": 27150 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 3.2498734177215192e-06, |
|
"loss": 0.0123, |
|
"step": 27175 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 3.2435443037974686e-06, |
|
"loss": 0.0121, |
|
"step": 27200 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 3.2372151898734176e-06, |
|
"loss": 0.0129, |
|
"step": 27225 |
|
}, |
|
{ |
|
"epoch": 15.02, |
|
"learning_rate": 3.230886075949367e-06, |
|
"loss": 0.0128, |
|
"step": 27250 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 3.224556962025317e-06, |
|
"loss": 0.0094, |
|
"step": 27275 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 3.2182278481012664e-06, |
|
"loss": 0.0081, |
|
"step": 27300 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 3.2118987341772158e-06, |
|
"loss": 0.0087, |
|
"step": 27325 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 3.2055696202531648e-06, |
|
"loss": 0.0131, |
|
"step": 27350 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 3.199240506329114e-06, |
|
"loss": 0.0073, |
|
"step": 27375 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 3.1929113924050636e-06, |
|
"loss": 0.0086, |
|
"step": 27400 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 3.186582278481013e-06, |
|
"loss": 0.0108, |
|
"step": 27425 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 3.1802531645569625e-06, |
|
"loss": 0.01, |
|
"step": 27450 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 3.1739240506329115e-06, |
|
"loss": 0.0123, |
|
"step": 27475 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 3.167594936708861e-06, |
|
"loss": 0.0081, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 3.1612658227848104e-06, |
|
"loss": 0.0123, |
|
"step": 27525 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 3.1549367088607598e-06, |
|
"loss": 0.0096, |
|
"step": 27550 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 3.1486075949367088e-06, |
|
"loss": 0.0054, |
|
"step": 27575 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 3.1422784810126582e-06, |
|
"loss": 0.006, |
|
"step": 27600 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 3.1359493670886076e-06, |
|
"loss": 0.0089, |
|
"step": 27625 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 3.129620253164557e-06, |
|
"loss": 0.0109, |
|
"step": 27650 |
|
}, |
|
{ |
|
"epoch": 15.04, |
|
"learning_rate": 3.1232911392405065e-06, |
|
"loss": 0.0089, |
|
"step": 27675 |
|
}, |
|
{ |
|
"epoch": 15.04, |
|
"learning_rate": 3.1169620253164563e-06, |
|
"loss": 0.0094, |
|
"step": 27700 |
|
}, |
|
{ |
|
"epoch": 15.04, |
|
"learning_rate": 3.1106329113924054e-06, |
|
"loss": 0.0102, |
|
"step": 27725 |
|
}, |
|
{ |
|
"epoch": 15.04, |
|
"learning_rate": 3.1043037974683548e-06, |
|
"loss": 0.0097, |
|
"step": 27750 |
|
}, |
|
{ |
|
"epoch": 15.04, |
|
"learning_rate": 3.097974683544304e-06, |
|
"loss": 0.0078, |
|
"step": 27775 |
|
}, |
|
{ |
|
"epoch": 15.04, |
|
"learning_rate": 3.0916455696202536e-06, |
|
"loss": 0.0073, |
|
"step": 27800 |
|
}, |
|
{ |
|
"epoch": 15.04, |
|
"learning_rate": 3.085316455696203e-06, |
|
"loss": 0.0105, |
|
"step": 27825 |
|
}, |
|
{ |
|
"epoch": 15.04, |
|
"learning_rate": 3.078987341772152e-06, |
|
"loss": 0.0083, |
|
"step": 27850 |
|
}, |
|
{ |
|
"epoch": 15.04, |
|
"learning_rate": 3.0726582278481015e-06, |
|
"loss": 0.0138, |
|
"step": 27875 |
|
}, |
|
{ |
|
"epoch": 15.04, |
|
"learning_rate": 3.066329113924051e-06, |
|
"loss": 0.0049, |
|
"step": 27900 |
|
}, |
|
{ |
|
"epoch": 15.04, |
|
"learning_rate": 3.0600000000000003e-06, |
|
"loss": 0.0107, |
|
"step": 27925 |
|
}, |
|
{ |
|
"epoch": 15.04, |
|
"learning_rate": 3.0536708860759494e-06, |
|
"loss": 0.0105, |
|
"step": 27950 |
|
}, |
|
{ |
|
"epoch": 15.04, |
|
"learning_rate": 3.0473417721518988e-06, |
|
"loss": 0.0134, |
|
"step": 27975 |
|
}, |
|
{ |
|
"epoch": 15.04, |
|
"learning_rate": 3.041012658227848e-06, |
|
"loss": 0.0137, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 15.04, |
|
"eval_loss": 0.2780563235282898, |
|
"eval_runtime": 1465.9542, |
|
"eval_samples_per_second": 7.102, |
|
"eval_steps_per_second": 0.444, |
|
"eval_wer": 25.010162766874256, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"learning_rate": 3.0346835443037976e-06, |
|
"loss": 0.009, |
|
"step": 28025 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"learning_rate": 3.028354430379747e-06, |
|
"loss": 0.0081, |
|
"step": 28050 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"learning_rate": 3.022025316455696e-06, |
|
"loss": 0.0109, |
|
"step": 28075 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"learning_rate": 3.015696202531646e-06, |
|
"loss": 0.0132, |
|
"step": 28100 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"learning_rate": 3.0093670886075953e-06, |
|
"loss": 0.0142, |
|
"step": 28125 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"learning_rate": 3.0030379746835448e-06, |
|
"loss": 0.0152, |
|
"step": 28150 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"learning_rate": 2.996708860759494e-06, |
|
"loss": 0.0074, |
|
"step": 28175 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"learning_rate": 2.990379746835443e-06, |
|
"loss": 0.0129, |
|
"step": 28200 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 2.9840506329113926e-06, |
|
"loss": 0.0118, |
|
"step": 28225 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 2.977721518987342e-06, |
|
"loss": 0.0153, |
|
"step": 28250 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 2.9713924050632915e-06, |
|
"loss": 0.0121, |
|
"step": 28275 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 2.965063291139241e-06, |
|
"loss": 0.0097, |
|
"step": 28300 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 2.95873417721519e-06, |
|
"loss": 0.0083, |
|
"step": 28325 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 2.9524050632911393e-06, |
|
"loss": 0.01, |
|
"step": 28350 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 2.9460759493670888e-06, |
|
"loss": 0.0084, |
|
"step": 28375 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 2.939746835443038e-06, |
|
"loss": 0.0124, |
|
"step": 28400 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 2.933417721518987e-06, |
|
"loss": 0.0071, |
|
"step": 28425 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 2.9270886075949366e-06, |
|
"loss": 0.0113, |
|
"step": 28450 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 2.920759493670886e-06, |
|
"loss": 0.0085, |
|
"step": 28475 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 2.914430379746836e-06, |
|
"loss": 0.0133, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 2.9081012658227853e-06, |
|
"loss": 0.014, |
|
"step": 28525 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 2.9017721518987348e-06, |
|
"loss": 0.0105, |
|
"step": 28550 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 2.8954430379746838e-06, |
|
"loss": 0.01, |
|
"step": 28575 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 2.889113924050633e-06, |
|
"loss": 0.0135, |
|
"step": 28600 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 2.8827848101265826e-06, |
|
"loss": 0.0118, |
|
"step": 28625 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 2.876455696202532e-06, |
|
"loss": 0.0111, |
|
"step": 28650 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 2.8701265822784815e-06, |
|
"loss": 0.008, |
|
"step": 28675 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 2.8637974683544305e-06, |
|
"loss": 0.0103, |
|
"step": 28700 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 2.85746835443038e-06, |
|
"loss": 0.0082, |
|
"step": 28725 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 2.8511392405063293e-06, |
|
"loss": 0.0057, |
|
"step": 28750 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 2.8448101265822788e-06, |
|
"loss": 0.0062, |
|
"step": 28775 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 2.8384810126582278e-06, |
|
"loss": 0.0092, |
|
"step": 28800 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 2.832151898734177e-06, |
|
"loss": 0.0084, |
|
"step": 28825 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 2.8258227848101266e-06, |
|
"loss": 0.0099, |
|
"step": 28850 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 2.819493670886076e-06, |
|
"loss": 0.0117, |
|
"step": 28875 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 2.8131645569620255e-06, |
|
"loss": 0.0117, |
|
"step": 28900 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 2.8068354430379753e-06, |
|
"loss": 0.0127, |
|
"step": 28925 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 2.8005063291139243e-06, |
|
"loss": 0.0123, |
|
"step": 28950 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 2.7941772151898738e-06, |
|
"loss": 0.0128, |
|
"step": 28975 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"learning_rate": 2.787848101265823e-06, |
|
"loss": 0.0073, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 16.02, |
|
"eval_loss": 0.2685891091823578, |
|
"eval_runtime": 1456.562, |
|
"eval_samples_per_second": 7.148, |
|
"eval_steps_per_second": 0.447, |
|
"eval_wer": 23.278427291500677, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 2.7815189873417726e-06, |
|
"loss": 0.0068, |
|
"step": 29025 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 2.7751898734177216e-06, |
|
"loss": 0.0087, |
|
"step": 29050 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 2.768860759493671e-06, |
|
"loss": 0.0084, |
|
"step": 29075 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 2.7625316455696205e-06, |
|
"loss": 0.0039, |
|
"step": 29100 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 2.75620253164557e-06, |
|
"loss": 0.0066, |
|
"step": 29125 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 2.7498734177215193e-06, |
|
"loss": 0.0108, |
|
"step": 29150 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 2.7435443037974683e-06, |
|
"loss": 0.0052, |
|
"step": 29175 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 2.7372151898734178e-06, |
|
"loss": 0.0061, |
|
"step": 29200 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 2.730886075949367e-06, |
|
"loss": 0.0083, |
|
"step": 29225 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 2.7245569620253166e-06, |
|
"loss": 0.0062, |
|
"step": 29250 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 2.7182278481012656e-06, |
|
"loss": 0.0078, |
|
"step": 29275 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 2.711898734177215e-06, |
|
"loss": 0.0127, |
|
"step": 29300 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 2.705569620253165e-06, |
|
"loss": 0.0092, |
|
"step": 29325 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 2.6992405063291143e-06, |
|
"loss": 0.0087, |
|
"step": 29350 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 2.6929113924050638e-06, |
|
"loss": 0.0062, |
|
"step": 29375 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 2.686582278481013e-06, |
|
"loss": 0.0065, |
|
"step": 29400 |
|
}, |
|
{ |
|
"epoch": 16.04, |
|
"learning_rate": 2.680253164556962e-06, |
|
"loss": 0.0057, |
|
"step": 29425 |
|
}, |
|
{ |
|
"epoch": 16.04, |
|
"learning_rate": 2.6739240506329116e-06, |
|
"loss": 0.0071, |
|
"step": 29450 |
|
}, |
|
{ |
|
"epoch": 16.04, |
|
"learning_rate": 2.667594936708861e-06, |
|
"loss": 0.01, |
|
"step": 29475 |
|
}, |
|
{ |
|
"epoch": 16.04, |
|
"learning_rate": 2.6612658227848105e-06, |
|
"loss": 0.009, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 16.04, |
|
"learning_rate": 2.65493670886076e-06, |
|
"loss": 0.0091, |
|
"step": 29525 |
|
}, |
|
{ |
|
"epoch": 16.04, |
|
"learning_rate": 2.648607594936709e-06, |
|
"loss": 0.0104, |
|
"step": 29550 |
|
}, |
|
{ |
|
"epoch": 16.04, |
|
"learning_rate": 2.6422784810126583e-06, |
|
"loss": 0.0071, |
|
"step": 29575 |
|
}, |
|
{ |
|
"epoch": 16.04, |
|
"learning_rate": 2.6359493670886078e-06, |
|
"loss": 0.0081, |
|
"step": 29600 |
|
}, |
|
{ |
|
"epoch": 16.04, |
|
"learning_rate": 2.629620253164557e-06, |
|
"loss": 0.008, |
|
"step": 29625 |
|
}, |
|
{ |
|
"epoch": 16.04, |
|
"learning_rate": 2.623291139240506e-06, |
|
"loss": 0.0067, |
|
"step": 29650 |
|
}, |
|
{ |
|
"epoch": 16.04, |
|
"learning_rate": 2.6169620253164556e-06, |
|
"loss": 0.0066, |
|
"step": 29675 |
|
}, |
|
{ |
|
"epoch": 16.04, |
|
"learning_rate": 2.610632911392405e-06, |
|
"loss": 0.0121, |
|
"step": 29700 |
|
}, |
|
{ |
|
"epoch": 16.04, |
|
"learning_rate": 2.6043037974683545e-06, |
|
"loss": 0.0075, |
|
"step": 29725 |
|
}, |
|
{ |
|
"epoch": 16.04, |
|
"learning_rate": 2.5979746835443043e-06, |
|
"loss": 0.0111, |
|
"step": 29750 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 2.5916455696202538e-06, |
|
"loss": 0.0095, |
|
"step": 29775 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 2.5853164556962028e-06, |
|
"loss": 0.0077, |
|
"step": 29800 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 2.578987341772152e-06, |
|
"loss": 0.0047, |
|
"step": 29825 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 2.5726582278481016e-06, |
|
"loss": 0.0092, |
|
"step": 29850 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 2.566329113924051e-06, |
|
"loss": 0.0191, |
|
"step": 29875 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 2.56e-06, |
|
"loss": 0.0102, |
|
"step": 29900 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 2.5536708860759495e-06, |
|
"loss": 0.0146, |
|
"step": 29925 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 2.547341772151899e-06, |
|
"loss": 0.0096, |
|
"step": 29950 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 2.5410126582278483e-06, |
|
"loss": 0.0159, |
|
"step": 29975 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 2.5346835443037978e-06, |
|
"loss": 0.0094, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"eval_loss": 0.2636851966381073, |
|
"eval_runtime": 1508.7196, |
|
"eval_samples_per_second": 6.901, |
|
"eval_steps_per_second": 0.431, |
|
"eval_wer": 23.029642758418838, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 2.5283544303797468e-06, |
|
"loss": 0.0078, |
|
"step": 30025 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 2.522025316455696e-06, |
|
"loss": 0.01, |
|
"step": 30050 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 2.5156962025316456e-06, |
|
"loss": 0.0058, |
|
"step": 30075 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 2.509367088607595e-06, |
|
"loss": 0.0092, |
|
"step": 30100 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 2.503037974683544e-06, |
|
"loss": 0.0084, |
|
"step": 30125 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 2.496708860759494e-06, |
|
"loss": 0.0069, |
|
"step": 30150 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 2.4903797468354433e-06, |
|
"loss": 0.0059, |
|
"step": 30175 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 2.4840506329113923e-06, |
|
"loss": 0.0097, |
|
"step": 30200 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 2.4777215189873418e-06, |
|
"loss": 0.0126, |
|
"step": 30225 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 2.4713924050632916e-06, |
|
"loss": 0.0096, |
|
"step": 30250 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 2.4650632911392406e-06, |
|
"loss": 0.0094, |
|
"step": 30275 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 2.45873417721519e-06, |
|
"loss": 0.0111, |
|
"step": 30300 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 2.4524050632911395e-06, |
|
"loss": 0.0062, |
|
"step": 30325 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 2.446075949367089e-06, |
|
"loss": 0.0078, |
|
"step": 30350 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 2.4397468354430383e-06, |
|
"loss": 0.0043, |
|
"step": 30375 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 2.4336708860759494e-06, |
|
"loss": 0.0102, |
|
"step": 30400 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 2.4273417721518988e-06, |
|
"loss": 0.0101, |
|
"step": 30425 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 2.4210126582278482e-06, |
|
"loss": 0.0066, |
|
"step": 30450 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 2.4146835443037976e-06, |
|
"loss": 0.0078, |
|
"step": 30475 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 2.408354430379747e-06, |
|
"loss": 0.0074, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 2.4020253164556965e-06, |
|
"loss": 0.0072, |
|
"step": 30525 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 2.395696202531646e-06, |
|
"loss": 0.0109, |
|
"step": 30550 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 2.389367088607595e-06, |
|
"loss": 0.0113, |
|
"step": 30575 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 2.3830379746835444e-06, |
|
"loss": 0.0109, |
|
"step": 30600 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 2.3767088607594938e-06, |
|
"loss": 0.0089, |
|
"step": 30625 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 2.3703797468354432e-06, |
|
"loss": 0.0062, |
|
"step": 30650 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 2.3640506329113926e-06, |
|
"loss": 0.0101, |
|
"step": 30675 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 2.357721518987342e-06, |
|
"loss": 0.0096, |
|
"step": 30700 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 2.3513924050632915e-06, |
|
"loss": 0.0093, |
|
"step": 30725 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 2.3450632911392405e-06, |
|
"loss": 0.0113, |
|
"step": 30750 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 2.33873417721519e-06, |
|
"loss": 0.0094, |
|
"step": 30775 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 2.3324050632911394e-06, |
|
"loss": 0.0086, |
|
"step": 30800 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 2.3260759493670888e-06, |
|
"loss": 0.0096, |
|
"step": 30825 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 2.319746835443038e-06, |
|
"loss": 0.0086, |
|
"step": 30850 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 2.3134177215189876e-06, |
|
"loss": 0.0064, |
|
"step": 30875 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 2.307088607594937e-06, |
|
"loss": 0.0107, |
|
"step": 30900 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 2.3007594936708865e-06, |
|
"loss": 0.0073, |
|
"step": 30925 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 2.2944303797468355e-06, |
|
"loss": 0.005, |
|
"step": 30950 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 2.288101265822785e-06, |
|
"loss": 0.0055, |
|
"step": 30975 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 2.2817721518987344e-06, |
|
"loss": 0.006, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"eval_loss": 0.2709895372390747, |
|
"eval_runtime": 1461.5544, |
|
"eval_samples_per_second": 7.123, |
|
"eval_steps_per_second": 0.445, |
|
"eval_wer": 23.29468771849949, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 2.2754430379746838e-06, |
|
"loss": 0.0094, |
|
"step": 31025 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 2.2691139240506328e-06, |
|
"loss": 0.0088, |
|
"step": 31050 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 2.2627848101265826e-06, |
|
"loss": 0.0057, |
|
"step": 31075 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 2.256455696202532e-06, |
|
"loss": 0.0066, |
|
"step": 31100 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 2.250126582278481e-06, |
|
"loss": 0.008, |
|
"step": 31125 |
|
}, |
|
{ |
|
"epoch": 17.03, |
|
"learning_rate": 2.2437974683544305e-06, |
|
"loss": 0.0048, |
|
"step": 31150 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 2.23746835443038e-06, |
|
"loss": 0.0074, |
|
"step": 31175 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 2.2311392405063294e-06, |
|
"loss": 0.0061, |
|
"step": 31200 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 2.2248101265822788e-06, |
|
"loss": 0.0088, |
|
"step": 31225 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 2.2184810126582278e-06, |
|
"loss": 0.0087, |
|
"step": 31250 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 2.2121518987341776e-06, |
|
"loss": 0.0089, |
|
"step": 31275 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 2.2058227848101266e-06, |
|
"loss": 0.0061, |
|
"step": 31300 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 2.199493670886076e-06, |
|
"loss": 0.0097, |
|
"step": 31325 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 2.1931645569620255e-06, |
|
"loss": 0.0063, |
|
"step": 31350 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 2.186835443037975e-06, |
|
"loss": 0.0084, |
|
"step": 31375 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 2.1805063291139243e-06, |
|
"loss": 0.0089, |
|
"step": 31400 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 2.1741772151898734e-06, |
|
"loss": 0.0077, |
|
"step": 31425 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 2.1678481012658228e-06, |
|
"loss": 0.0153, |
|
"step": 31450 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 2.1615189873417726e-06, |
|
"loss": 0.0088, |
|
"step": 31475 |
|
}, |
|
{ |
|
"epoch": 17.04, |
|
"learning_rate": 2.1551898734177216e-06, |
|
"loss": 0.0095, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"learning_rate": 2.148860759493671e-06, |
|
"loss": 0.0084, |
|
"step": 31525 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"learning_rate": 2.1425316455696205e-06, |
|
"loss": 0.0054, |
|
"step": 31550 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"learning_rate": 2.13620253164557e-06, |
|
"loss": 0.013, |
|
"step": 31575 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"learning_rate": 2.129873417721519e-06, |
|
"loss": 0.0074, |
|
"step": 31600 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"learning_rate": 2.1235443037974684e-06, |
|
"loss": 0.0083, |
|
"step": 31625 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"learning_rate": 2.1172151898734178e-06, |
|
"loss": 0.0118, |
|
"step": 31650 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"learning_rate": 2.110886075949367e-06, |
|
"loss": 0.0095, |
|
"step": 31675 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"learning_rate": 2.1045569620253166e-06, |
|
"loss": 0.0075, |
|
"step": 31700 |
|
}, |
|
{ |
|
"epoch": 18.01, |
|
"learning_rate": 2.098227848101266e-06, |
|
"loss": 0.0149, |
|
"step": 31725 |
|
}, |
|
{ |
|
"epoch": 18.01, |
|
"learning_rate": 2.0918987341772155e-06, |
|
"loss": 0.0095, |
|
"step": 31750 |
|
}, |
|
{ |
|
"epoch": 18.01, |
|
"learning_rate": 2.085569620253165e-06, |
|
"loss": 0.0098, |
|
"step": 31775 |
|
}, |
|
{ |
|
"epoch": 18.01, |
|
"learning_rate": 2.079240506329114e-06, |
|
"loss": 0.0058, |
|
"step": 31800 |
|
}, |
|
{ |
|
"epoch": 18.01, |
|
"learning_rate": 2.0729113924050633e-06, |
|
"loss": 0.0072, |
|
"step": 31825 |
|
}, |
|
{ |
|
"epoch": 18.01, |
|
"learning_rate": 2.0665822784810128e-06, |
|
"loss": 0.0103, |
|
"step": 31850 |
|
}, |
|
{ |
|
"epoch": 18.01, |
|
"learning_rate": 2.060253164556962e-06, |
|
"loss": 0.0085, |
|
"step": 31875 |
|
}, |
|
{ |
|
"epoch": 18.01, |
|
"learning_rate": 2.0539240506329116e-06, |
|
"loss": 0.0064, |
|
"step": 31900 |
|
}, |
|
{ |
|
"epoch": 18.01, |
|
"learning_rate": 2.047594936708861e-06, |
|
"loss": 0.0077, |
|
"step": 31925 |
|
}, |
|
{ |
|
"epoch": 18.01, |
|
"learning_rate": 2.0412658227848105e-06, |
|
"loss": 0.0086, |
|
"step": 31950 |
|
}, |
|
{ |
|
"epoch": 18.01, |
|
"learning_rate": 2.0349367088607595e-06, |
|
"loss": 0.0136, |
|
"step": 31975 |
|
}, |
|
{ |
|
"epoch": 18.01, |
|
"learning_rate": 2.028607594936709e-06, |
|
"loss": 0.0085, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 18.01, |
|
"eval_loss": 0.26488059759140015, |
|
"eval_runtime": 1466.953, |
|
"eval_samples_per_second": 7.097, |
|
"eval_steps_per_second": 0.444, |
|
"eval_wer": 23.089806338314446, |
|
"step": 32000 |
|
} |
|
], |
|
"max_steps": 40000, |
|
"num_train_epochs": 9223372036854775807, |
|
"total_flos": 1.477037795917824e+20, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|