|
{ |
|
"best_metric": 12.021334704238024, |
|
"best_model_checkpoint": "./checkpoint-32500", |
|
"epoch": 1.2588, |
|
"global_step": 40000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 4.4e-07, |
|
"loss": 1.8084, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 9.400000000000001e-07, |
|
"loss": 1.6156, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.44e-06, |
|
"loss": 1.1067, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.94e-06, |
|
"loss": 0.6228, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.4400000000000004e-06, |
|
"loss": 0.45, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.9400000000000002e-06, |
|
"loss": 0.365, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 3.44e-06, |
|
"loss": 0.3328, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 3.94e-06, |
|
"loss": 0.3055, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.440000000000001e-06, |
|
"loss": 0.3007, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.94e-06, |
|
"loss": 0.2964, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 5.4400000000000004e-06, |
|
"loss": 0.2899, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 5.94e-06, |
|
"loss": 0.2651, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 6.440000000000001e-06, |
|
"loss": 0.2554, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 6.9400000000000005e-06, |
|
"loss": 0.2672, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 7.440000000000001e-06, |
|
"loss": 0.2646, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 7.94e-06, |
|
"loss": 0.2632, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.44e-06, |
|
"loss": 0.2372, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.94e-06, |
|
"loss": 0.2129, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.440000000000001e-06, |
|
"loss": 0.2041, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.940000000000001e-06, |
|
"loss": 0.2196, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.994430379746836e-06, |
|
"loss": 0.2256, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.988101265822786e-06, |
|
"loss": 0.2331, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.981772151898735e-06, |
|
"loss": 0.2268, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.975443037974685e-06, |
|
"loss": 0.2169, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.969113924050634e-06, |
|
"loss": 0.234, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.962784810126583e-06, |
|
"loss": 0.3795, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.956455696202533e-06, |
|
"loss": 0.3479, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.950126582278482e-06, |
|
"loss": 0.3889, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.943797468354432e-06, |
|
"loss": 0.3687, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.937468354430381e-06, |
|
"loss": 0.3991, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.93113924050633e-06, |
|
"loss": 0.3388, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.92481012658228e-06, |
|
"loss": 0.2639, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.91848101265823e-06, |
|
"loss": 0.2188, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.912151898734179e-06, |
|
"loss": 0.2138, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.905822784810127e-06, |
|
"loss": 0.2176, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.899493670886076e-06, |
|
"loss": 0.2054, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.893164556962025e-06, |
|
"loss": 0.196, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.886835443037975e-06, |
|
"loss": 0.1664, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.880506329113924e-06, |
|
"loss": 0.1501, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.874177215189874e-06, |
|
"loss": 0.154, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.867848101265823e-06, |
|
"loss": 0.1684, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.861518987341773e-06, |
|
"loss": 0.1611, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.855189873417722e-06, |
|
"loss": 0.1397, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.848860759493671e-06, |
|
"loss": 0.1083, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.842531645569621e-06, |
|
"loss": 0.0957, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.83620253164557e-06, |
|
"loss": 0.1514, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.82987341772152e-06, |
|
"loss": 0.1429, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.82354430379747e-06, |
|
"loss": 0.148, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.817215189873419e-06, |
|
"loss": 0.1594, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.810886075949368e-06, |
|
"loss": 0.1557, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.804556962025317e-06, |
|
"loss": 0.214, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.798227848101267e-06, |
|
"loss": 0.2759, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.791898734177216e-06, |
|
"loss": 0.3147, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.785569620253166e-06, |
|
"loss": 0.3295, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.779240506329115e-06, |
|
"loss": 0.2397, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.772911392405065e-06, |
|
"loss": 0.2159, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.766582278481014e-06, |
|
"loss": 0.2256, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.760253164556963e-06, |
|
"loss": 0.3192, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.753924050632913e-06, |
|
"loss": 0.4899, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.747594936708862e-06, |
|
"loss": 0.2394, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.741265822784812e-06, |
|
"loss": 0.1659, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.734936708860761e-06, |
|
"loss": 0.2089, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.72860759493671e-06, |
|
"loss": 0.2285, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.722278481012658e-06, |
|
"loss": 0.2665, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.715949367088608e-06, |
|
"loss": 0.4278, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.709620253164557e-06, |
|
"loss": 0.289, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.703291139240507e-06, |
|
"loss": 0.2451, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.696962025316456e-06, |
|
"loss": 0.305, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.690632911392405e-06, |
|
"loss": 0.2836, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.684303797468355e-06, |
|
"loss": 0.2507, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.677974683544304e-06, |
|
"loss": 0.2964, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.671645569620254e-06, |
|
"loss": 0.269, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.665316455696203e-06, |
|
"loss": 0.2306, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.658987341772153e-06, |
|
"loss": 0.2127, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.652658227848102e-06, |
|
"loss": 0.2387, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.646329113924051e-06, |
|
"loss": 0.2131, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.640000000000001e-06, |
|
"loss": 0.1797, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.63367088607595e-06, |
|
"loss": 0.2154, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.6273417721519e-06, |
|
"loss": 0.2188, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.62101265822785e-06, |
|
"loss": 0.2392, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.614683544303799e-06, |
|
"loss": 0.2277, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.608354430379746e-06, |
|
"loss": 0.291, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.602025316455696e-06, |
|
"loss": 0.31, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.595696202531647e-06, |
|
"loss": 0.2412, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.589367088607596e-06, |
|
"loss": 0.3096, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.583037974683546e-06, |
|
"loss": 0.1602, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.576708860759495e-06, |
|
"loss": 0.1801, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.570379746835445e-06, |
|
"loss": 0.21, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.564050632911394e-06, |
|
"loss": 0.2148, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.557721518987343e-06, |
|
"loss": 0.1847, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.551392405063293e-06, |
|
"loss": 0.2079, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.545063291139242e-06, |
|
"loss": 0.2032, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.538734177215192e-06, |
|
"loss": 0.1833, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.53240506329114e-06, |
|
"loss": 0.2486, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.526075949367089e-06, |
|
"loss": 0.2633, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.519746835443038e-06, |
|
"loss": 0.2043, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.513417721518988e-06, |
|
"loss": 0.1735, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.507088607594937e-06, |
|
"loss": 0.1864, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.500759493670887e-06, |
|
"loss": 0.1834, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.494430379746836e-06, |
|
"loss": 0.1577, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"eval_loss": 0.4077220559120178, |
|
"eval_runtime": 1902.577, |
|
"eval_samples_per_second": 8.596, |
|
"eval_steps_per_second": 0.538, |
|
"eval_wer": 16.23493879124763, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.488101265822785e-06, |
|
"loss": 0.1771, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.481772151898735e-06, |
|
"loss": 0.1817, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.475443037974684e-06, |
|
"loss": 0.2135, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.469113924050634e-06, |
|
"loss": 0.2543, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.462784810126583e-06, |
|
"loss": 0.2893, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.456455696202533e-06, |
|
"loss": 0.3303, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.450126582278482e-06, |
|
"loss": 0.2511, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.443797468354431e-06, |
|
"loss": 0.2964, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.437468354430381e-06, |
|
"loss": 0.2452, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.43113924050633e-06, |
|
"loss": 0.243, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.42481012658228e-06, |
|
"loss": 0.2023, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.418481012658227e-06, |
|
"loss": 0.1834, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.412151898734177e-06, |
|
"loss": 0.1653, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.405822784810126e-06, |
|
"loss": 0.2025, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.399493670886076e-06, |
|
"loss": 0.2024, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.393164556962025e-06, |
|
"loss": 0.1786, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.386835443037976e-06, |
|
"loss": 0.1876, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.380506329113926e-06, |
|
"loss": 0.2348, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.374177215189875e-06, |
|
"loss": 0.2407, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.367848101265825e-06, |
|
"loss": 0.1743, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.361518987341774e-06, |
|
"loss": 0.2515, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.355189873417723e-06, |
|
"loss": 0.3183, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.348860759493671e-06, |
|
"loss": 0.3268, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.34253164556962e-06, |
|
"loss": 0.2201, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.33620253164557e-06, |
|
"loss": 0.4741, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.32987341772152e-06, |
|
"loss": 0.8504, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.323544303797469e-06, |
|
"loss": 0.408, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.317215189873418e-06, |
|
"loss": 0.2727, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.310886075949368e-06, |
|
"loss": 0.2284, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.304556962025317e-06, |
|
"loss": 0.2393, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.298227848101267e-06, |
|
"loss": 0.2901, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.291898734177216e-06, |
|
"loss": 0.3287, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.285569620253165e-06, |
|
"loss": 0.2224, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.279240506329115e-06, |
|
"loss": 0.2181, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.272911392405064e-06, |
|
"loss": 0.2189, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.266582278481014e-06, |
|
"loss": 0.2534, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.260253164556963e-06, |
|
"loss": 0.2462, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.253924050632913e-06, |
|
"loss": 0.2229, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.247594936708862e-06, |
|
"loss": 0.1657, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.241265822784811e-06, |
|
"loss": 0.1312, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.23493670886076e-06, |
|
"loss": 0.2281, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.228607594936709e-06, |
|
"loss": 0.2682, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.222278481012658e-06, |
|
"loss": 0.2952, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.215949367088607e-06, |
|
"loss": 0.3846, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.209620253164557e-06, |
|
"loss": 0.3287, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.203291139240506e-06, |
|
"loss": 0.2258, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.196962025316456e-06, |
|
"loss": 0.2325, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.190632911392405e-06, |
|
"loss": 0.2739, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.184303797468355e-06, |
|
"loss": 0.3693, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.177974683544306e-06, |
|
"loss": 0.4055, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.171645569620255e-06, |
|
"loss": 0.3022, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.165316455696205e-06, |
|
"loss": 0.2857, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.158987341772152e-06, |
|
"loss": 0.224, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.152658227848102e-06, |
|
"loss": 0.2307, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.146329113924051e-06, |
|
"loss": 0.2087, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.14e-06, |
|
"loss": 0.2431, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.13367088607595e-06, |
|
"loss": 0.2078, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.1273417721519e-06, |
|
"loss": 0.2152, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.121012658227849e-06, |
|
"loss": 0.2451, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.114683544303798e-06, |
|
"loss": 0.2844, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.108354430379748e-06, |
|
"loss": 0.2504, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.102025316455697e-06, |
|
"loss": 0.2734, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.095696202531647e-06, |
|
"loss": 0.2819, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.089367088607596e-06, |
|
"loss": 0.2597, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.083037974683545e-06, |
|
"loss": 0.2528, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.076708860759495e-06, |
|
"loss": 0.2319, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.070379746835444e-06, |
|
"loss": 0.2465, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.064050632911394e-06, |
|
"loss": 0.2709, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.057721518987343e-06, |
|
"loss": 0.2312, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.051392405063293e-06, |
|
"loss": 0.2508, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.04506329113924e-06, |
|
"loss": 0.2236, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.03873417721519e-06, |
|
"loss": 0.2377, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.03240506329114e-06, |
|
"loss": 0.2631, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.026075949367089e-06, |
|
"loss": 0.2328, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.019746835443038e-06, |
|
"loss": 0.207, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.013417721518987e-06, |
|
"loss": 0.2147, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.007088607594937e-06, |
|
"loss": 0.2124, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.000759493670886e-06, |
|
"loss": 0.2379, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.994430379746836e-06, |
|
"loss": 0.2744, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.988101265822785e-06, |
|
"loss": 0.2381, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.981772151898735e-06, |
|
"loss": 0.2005, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.975443037974684e-06, |
|
"loss": 0.2449, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 8.969113924050633e-06, |
|
"loss": 0.2596, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.962784810126583e-06, |
|
"loss": 0.2093, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.956455696202532e-06, |
|
"loss": 0.2161, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.950126582278482e-06, |
|
"loss": 0.1944, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.943797468354431e-06, |
|
"loss": 0.2318, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.93746835443038e-06, |
|
"loss": 0.2558, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.93113924050633e-06, |
|
"loss": 0.2348, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.92481012658228e-06, |
|
"loss": 0.1844, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.918481012658229e-06, |
|
"loss": 0.2447, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.912151898734178e-06, |
|
"loss": 0.2061, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.905822784810128e-06, |
|
"loss": 0.2153, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.899493670886077e-06, |
|
"loss": 0.2427, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.893164556962027e-06, |
|
"loss": 0.2292, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.886835443037976e-06, |
|
"loss": 0.2054, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.880506329113925e-06, |
|
"loss": 0.1916, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.874177215189875e-06, |
|
"loss": 0.192, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.867848101265824e-06, |
|
"loss": 0.2317, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 8.861518987341772e-06, |
|
"loss": 0.2244, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"eval_loss": 0.3697776198387146, |
|
"eval_runtime": 1717.5108, |
|
"eval_samples_per_second": 9.522, |
|
"eval_steps_per_second": 0.596, |
|
"eval_wer": 14.732512932557915, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.855189873417721e-06, |
|
"loss": 0.164, |
|
"step": 5025 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.848860759493671e-06, |
|
"loss": 0.147, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.84253164556962e-06, |
|
"loss": 0.135, |
|
"step": 5075 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.83620253164557e-06, |
|
"loss": 0.1959, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.829873417721519e-06, |
|
"loss": 0.7075, |
|
"step": 5125 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.823544303797469e-06, |
|
"loss": 0.8746, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.817215189873418e-06, |
|
"loss": 0.8571, |
|
"step": 5175 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.810886075949367e-06, |
|
"loss": 0.8078, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.804556962025317e-06, |
|
"loss": 0.8197, |
|
"step": 5225 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.798227848101266e-06, |
|
"loss": 0.7873, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.791898734177216e-06, |
|
"loss": 0.7204, |
|
"step": 5275 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.785569620253165e-06, |
|
"loss": 0.6956, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.779240506329115e-06, |
|
"loss": 0.4503, |
|
"step": 5325 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.772911392405064e-06, |
|
"loss": 0.2521, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 8.766582278481013e-06, |
|
"loss": 0.1978, |
|
"step": 5375 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.760253164556963e-06, |
|
"loss": 0.1572, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.753924050632912e-06, |
|
"loss": 0.179, |
|
"step": 5425 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.747594936708862e-06, |
|
"loss": 0.1696, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.741265822784811e-06, |
|
"loss": 0.1883, |
|
"step": 5475 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.73493670886076e-06, |
|
"loss": 0.181, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.72860759493671e-06, |
|
"loss": 0.1869, |
|
"step": 5525 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.72227848101266e-06, |
|
"loss": 0.1858, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.715949367088609e-06, |
|
"loss": 0.1944, |
|
"step": 5575 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.709620253164558e-06, |
|
"loss": 0.3527, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.703291139240508e-06, |
|
"loss": 0.341, |
|
"step": 5625 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.696962025316457e-06, |
|
"loss": 0.3151, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.690632911392407e-06, |
|
"loss": 0.2082, |
|
"step": 5675 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.684303797468356e-06, |
|
"loss": 0.1809, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.677974683544305e-06, |
|
"loss": 0.1774, |
|
"step": 5725 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.671645569620253e-06, |
|
"loss": 0.1789, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.665316455696203e-06, |
|
"loss": 0.1613, |
|
"step": 5775 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8.658987341772152e-06, |
|
"loss": 0.1406, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.652658227848101e-06, |
|
"loss": 0.1585, |
|
"step": 5825 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.646329113924051e-06, |
|
"loss": 0.1674, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.64e-06, |
|
"loss": 0.1433, |
|
"step": 5875 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.63367088607595e-06, |
|
"loss": 0.1531, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.627341772151899e-06, |
|
"loss": 0.1324, |
|
"step": 5925 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.621012658227849e-06, |
|
"loss": 0.3431, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.614683544303798e-06, |
|
"loss": 0.4039, |
|
"step": 5975 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.608354430379747e-06, |
|
"loss": 0.3622, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.602025316455697e-06, |
|
"loss": 0.3706, |
|
"step": 6025 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.595696202531646e-06, |
|
"loss": 0.3716, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.589367088607596e-06, |
|
"loss": 0.3677, |
|
"step": 6075 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.583037974683545e-06, |
|
"loss": 0.3563, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.576708860759495e-06, |
|
"loss": 0.3571, |
|
"step": 6125 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.570379746835444e-06, |
|
"loss": 0.3203, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.564050632911393e-06, |
|
"loss": 0.3303, |
|
"step": 6175 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 8.557721518987341e-06, |
|
"loss": 0.2475, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.551392405063292e-06, |
|
"loss": 0.1953, |
|
"step": 6225 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.545063291139242e-06, |
|
"loss": 0.1947, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.538734177215191e-06, |
|
"loss": 0.2221, |
|
"step": 6275 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.53240506329114e-06, |
|
"loss": 0.2616, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.52607594936709e-06, |
|
"loss": 0.2586, |
|
"step": 6325 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.51974683544304e-06, |
|
"loss": 0.2577, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.513417721518989e-06, |
|
"loss": 0.2456, |
|
"step": 6375 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.507088607594938e-06, |
|
"loss": 0.2667, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.500759493670888e-06, |
|
"loss": 0.258, |
|
"step": 6425 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.494683544303797e-06, |
|
"loss": 0.2539, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.488354430379747e-06, |
|
"loss": 0.2646, |
|
"step": 6475 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.482025316455696e-06, |
|
"loss": 0.2601, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.475696202531646e-06, |
|
"loss": 0.2475, |
|
"step": 6525 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.469367088607595e-06, |
|
"loss": 0.261, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 8.463037974683545e-06, |
|
"loss": 0.2716, |
|
"step": 6575 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.456708860759494e-06, |
|
"loss": 0.2633, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.450379746835443e-06, |
|
"loss": 0.2679, |
|
"step": 6625 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.444050632911393e-06, |
|
"loss": 0.2526, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.437721518987342e-06, |
|
"loss": 0.2612, |
|
"step": 6675 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.431392405063292e-06, |
|
"loss": 0.2454, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.425063291139241e-06, |
|
"loss": 0.2346, |
|
"step": 6725 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.41873417721519e-06, |
|
"loss": 0.2649, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.41240506329114e-06, |
|
"loss": 0.2524, |
|
"step": 6775 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.40607594936709e-06, |
|
"loss": 0.2596, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.399746835443039e-06, |
|
"loss": 0.2497, |
|
"step": 6825 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.393417721518988e-06, |
|
"loss": 0.2679, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.387088607594938e-06, |
|
"loss": 0.2383, |
|
"step": 6875 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.380759493670887e-06, |
|
"loss": 0.2566, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.374430379746837e-06, |
|
"loss": 0.2567, |
|
"step": 6925 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.368101265822786e-06, |
|
"loss": 0.2606, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.361772151898735e-06, |
|
"loss": 0.2688, |
|
"step": 6975 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 8.355443037974685e-06, |
|
"loss": 0.2554, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.349113924050634e-06, |
|
"loss": 0.2799, |
|
"step": 7025 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.342784810126584e-06, |
|
"loss": 0.3009, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.336455696202531e-06, |
|
"loss": 0.2902, |
|
"step": 7075 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.330126582278481e-06, |
|
"loss": 0.2932, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.32379746835443e-06, |
|
"loss": 0.297, |
|
"step": 7125 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.31746835443038e-06, |
|
"loss": 0.3124, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.31113924050633e-06, |
|
"loss": 0.2929, |
|
"step": 7175 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.304810126582279e-06, |
|
"loss": 0.31, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.298481012658228e-06, |
|
"loss": 0.3089, |
|
"step": 7225 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.292151898734177e-06, |
|
"loss": 0.3032, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.285822784810127e-06, |
|
"loss": 0.2995, |
|
"step": 7275 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.279493670886076e-06, |
|
"loss": 0.2988, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.273164556962026e-06, |
|
"loss": 0.2754, |
|
"step": 7325 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.266835443037975e-06, |
|
"loss": 0.2953, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.260506329113925e-06, |
|
"loss": 0.2818, |
|
"step": 7375 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 8.254177215189874e-06, |
|
"loss": 0.3161, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.247848101265823e-06, |
|
"loss": 0.299, |
|
"step": 7425 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.241518987341773e-06, |
|
"loss": 0.3071, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.235189873417722e-06, |
|
"loss": 0.2662, |
|
"step": 7475 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.228860759493672e-06, |
|
"loss": 0.3231, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"eval_loss": 0.3433650732040405, |
|
"eval_runtime": 1815.8431, |
|
"eval_samples_per_second": 9.006, |
|
"eval_steps_per_second": 0.563, |
|
"eval_wer": 13.744818944189186, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.222531645569621e-06, |
|
"loss": 0.2243, |
|
"step": 7525 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.21620253164557e-06, |
|
"loss": 0.307, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.20987341772152e-06, |
|
"loss": 0.2827, |
|
"step": 7575 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.20354430379747e-06, |
|
"loss": 0.1606, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.197215189873419e-06, |
|
"loss": 0.2046, |
|
"step": 7625 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.190886075949368e-06, |
|
"loss": 0.1987, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.184556962025318e-06, |
|
"loss": 0.2046, |
|
"step": 7675 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.178227848101267e-06, |
|
"loss": 0.2173, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.171898734177217e-06, |
|
"loss": 0.1797, |
|
"step": 7725 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.165569620253166e-06, |
|
"loss": 0.1281, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 8.159240506329115e-06, |
|
"loss": 0.1671, |
|
"step": 7775 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.152911392405065e-06, |
|
"loss": 0.2503, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.146582278481013e-06, |
|
"loss": 0.4373, |
|
"step": 7825 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.140253164556962e-06, |
|
"loss": 0.3136, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.133924050632911e-06, |
|
"loss": 0.2205, |
|
"step": 7875 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.127594936708861e-06, |
|
"loss": 0.2451, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.12126582278481e-06, |
|
"loss": 0.2599, |
|
"step": 7925 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.11493670886076e-06, |
|
"loss": 0.2688, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.108607594936709e-06, |
|
"loss": 0.2625, |
|
"step": 7975 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.102278481012659e-06, |
|
"loss": 0.2571, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.095949367088608e-06, |
|
"loss": 0.2074, |
|
"step": 8025 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.089620253164557e-06, |
|
"loss": 0.2484, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.083291139240507e-06, |
|
"loss": 0.2571, |
|
"step": 8075 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.076962025316456e-06, |
|
"loss": 0.3578, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.070632911392406e-06, |
|
"loss": 0.3629, |
|
"step": 8125 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.064303797468355e-06, |
|
"loss": 0.2216, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.057974683544305e-06, |
|
"loss": 0.2079, |
|
"step": 8175 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.051645569620254e-06, |
|
"loss": 0.2035, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.045316455696203e-06, |
|
"loss": 0.1913, |
|
"step": 8225 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.038987341772151e-06, |
|
"loss": 0.2097, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.032658227848102e-06, |
|
"loss": 0.2059, |
|
"step": 8275 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.026329113924052e-06, |
|
"loss": 0.1805, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.020000000000001e-06, |
|
"loss": 0.1696, |
|
"step": 8325 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.01367088607595e-06, |
|
"loss": 0.1879, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.0073417721519e-06, |
|
"loss": 0.1906, |
|
"step": 8375 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.00101265822785e-06, |
|
"loss": 0.1809, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 7.994683544303799e-06, |
|
"loss": 0.1999, |
|
"step": 8425 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 7.988354430379748e-06, |
|
"loss": 0.1747, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 7.982025316455698e-06, |
|
"loss": 0.2052, |
|
"step": 8475 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 7.975696202531647e-06, |
|
"loss": 0.2929, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 7.969367088607597e-06, |
|
"loss": 0.225, |
|
"step": 8525 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 7.963037974683544e-06, |
|
"loss": 0.1506, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 7.956962025316456e-06, |
|
"loss": 0.1231, |
|
"step": 8575 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 7.950632911392405e-06, |
|
"loss": 0.1401, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.944303797468355e-06, |
|
"loss": 0.1842, |
|
"step": 8625 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.937974683544304e-06, |
|
"loss": 0.2131, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.931645569620255e-06, |
|
"loss": 0.1935, |
|
"step": 8675 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.925316455696203e-06, |
|
"loss": 0.1902, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.918987341772152e-06, |
|
"loss": 0.2444, |
|
"step": 8725 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.912658227848102e-06, |
|
"loss": 0.239, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.906329113924051e-06, |
|
"loss": 0.2845, |
|
"step": 8775 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.9e-06, |
|
"loss": 0.27, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.89367088607595e-06, |
|
"loss": 0.2615, |
|
"step": 8825 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.8873417721519e-06, |
|
"loss": 0.2501, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.881012658227849e-06, |
|
"loss": 0.2664, |
|
"step": 8875 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.874683544303798e-06, |
|
"loss": 0.2494, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.868354430379748e-06, |
|
"loss": 0.2512, |
|
"step": 8925 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.862025316455697e-06, |
|
"loss": 0.2585, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 7.855696202531647e-06, |
|
"loss": 0.2491, |
|
"step": 8975 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.849367088607596e-06, |
|
"loss": 0.2659, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.843037974683545e-06, |
|
"loss": 0.2599, |
|
"step": 9025 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.836708860759495e-06, |
|
"loss": 0.2199, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.830379746835444e-06, |
|
"loss": 0.2295, |
|
"step": 9075 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.824050632911394e-06, |
|
"loss": 0.2556, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.817721518987343e-06, |
|
"loss": 0.229, |
|
"step": 9125 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.811392405063291e-06, |
|
"loss": 0.2288, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.80506329113924e-06, |
|
"loss": 0.2515, |
|
"step": 9175 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.79873417721519e-06, |
|
"loss": 0.2397, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.79240506329114e-06, |
|
"loss": 0.2346, |
|
"step": 9225 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.786075949367089e-06, |
|
"loss": 0.258, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.779746835443038e-06, |
|
"loss": 0.2444, |
|
"step": 9275 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.773417721518987e-06, |
|
"loss": 0.2145, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.767088607594937e-06, |
|
"loss": 0.2256, |
|
"step": 9325 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.760759493670886e-06, |
|
"loss": 0.2627, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.754430379746836e-06, |
|
"loss": 0.2778, |
|
"step": 9375 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 7.748101265822785e-06, |
|
"loss": 0.2904, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.741772151898735e-06, |
|
"loss": 0.2759, |
|
"step": 9425 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.735443037974684e-06, |
|
"loss": 0.2773, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.729113924050633e-06, |
|
"loss": 0.2887, |
|
"step": 9475 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.722784810126583e-06, |
|
"loss": 0.2687, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.716455696202532e-06, |
|
"loss": 0.2852, |
|
"step": 9525 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.710126582278482e-06, |
|
"loss": 0.291, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.703797468354431e-06, |
|
"loss": 0.274, |
|
"step": 9575 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.69746835443038e-06, |
|
"loss": 0.3004, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.69113924050633e-06, |
|
"loss": 0.2753, |
|
"step": 9625 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.68481012658228e-06, |
|
"loss": 0.27, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.678481012658229e-06, |
|
"loss": 0.2721, |
|
"step": 9675 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.672151898734178e-06, |
|
"loss": 0.2703, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.665822784810128e-06, |
|
"loss": 0.2456, |
|
"step": 9725 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.659493670886077e-06, |
|
"loss": 0.2432, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.653164556962027e-06, |
|
"loss": 0.2374, |
|
"step": 9775 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 7.646835443037976e-06, |
|
"loss": 0.2446, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.640506329113925e-06, |
|
"loss": 0.2506, |
|
"step": 9825 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.634177215189875e-06, |
|
"loss": 0.2451, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.6278481012658234e-06, |
|
"loss": 0.2221, |
|
"step": 9875 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.621518987341773e-06, |
|
"loss": 0.2322, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.615189873417722e-06, |
|
"loss": 0.25, |
|
"step": 9925 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.608860759493672e-06, |
|
"loss": 0.2233, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.602531645569621e-06, |
|
"loss": 0.2166, |
|
"step": 9975 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.59620253164557e-06, |
|
"loss": 0.2536, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"eval_loss": 0.3406411409378052, |
|
"eval_runtime": 1694.5787, |
|
"eval_samples_per_second": 9.651, |
|
"eval_steps_per_second": 0.604, |
|
"eval_wer": 13.498056099990361, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.589873417721519e-06, |
|
"loss": 0.2489, |
|
"step": 10025 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.583544303797469e-06, |
|
"loss": 0.2521, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.577215189873418e-06, |
|
"loss": 0.2289, |
|
"step": 10075 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.5708860759493674e-06, |
|
"loss": 0.2417, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.564556962025317e-06, |
|
"loss": 0.2268, |
|
"step": 10125 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.558227848101266e-06, |
|
"loss": 0.2526, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.551898734177216e-06, |
|
"loss": 0.2409, |
|
"step": 10175 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.545569620253165e-06, |
|
"loss": 0.226, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.539240506329114e-06, |
|
"loss": 0.2172, |
|
"step": 10225 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.532911392405063e-06, |
|
"loss": 0.2072, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.526582278481013e-06, |
|
"loss": 0.2313, |
|
"step": 10275 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.520253164556963e-06, |
|
"loss": 0.2484, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.513924050632912e-06, |
|
"loss": 0.2893, |
|
"step": 10325 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.507594936708862e-06, |
|
"loss": 0.2627, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.501265822784811e-06, |
|
"loss": 0.2304, |
|
"step": 10375 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.494936708860761e-06, |
|
"loss": 0.2412, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.48860759493671e-06, |
|
"loss": 0.2473, |
|
"step": 10425 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.4822784810126594e-06, |
|
"loss": 0.22, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.475949367088609e-06, |
|
"loss": 0.2303, |
|
"step": 10475 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.4696202531645574e-06, |
|
"loss": 0.2289, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.463291139240507e-06, |
|
"loss": 0.2221, |
|
"step": 10525 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.456962025316456e-06, |
|
"loss": 0.2397, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 7.450886075949367e-06, |
|
"loss": 0.2397, |
|
"step": 10575 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.444556962025316e-06, |
|
"loss": 0.2328, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.438227848101266e-06, |
|
"loss": 0.2466, |
|
"step": 10625 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.431898734177215e-06, |
|
"loss": 0.2494, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.425569620253165e-06, |
|
"loss": 0.2219, |
|
"step": 10675 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.419240506329115e-06, |
|
"loss": 0.2085, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.412911392405064e-06, |
|
"loss": 0.2473, |
|
"step": 10725 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.406582278481014e-06, |
|
"loss": 0.1858, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.400253164556963e-06, |
|
"loss": 0.2472, |
|
"step": 10775 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.393924050632913e-06, |
|
"loss": 0.2249, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.387594936708862e-06, |
|
"loss": 0.234, |
|
"step": 10825 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.3812658227848115e-06, |
|
"loss": 0.2385, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.37493670886076e-06, |
|
"loss": 0.2634, |
|
"step": 10875 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.3686075949367095e-06, |
|
"loss": 0.2334, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.362278481012659e-06, |
|
"loss": 0.2388, |
|
"step": 10925 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.355949367088608e-06, |
|
"loss": 0.2336, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 7.349620253164558e-06, |
|
"loss": 0.2477, |
|
"step": 10975 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.343291139240507e-06, |
|
"loss": 0.2536, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.336962025316457e-06, |
|
"loss": 0.2122, |
|
"step": 11025 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.330632911392406e-06, |
|
"loss": 0.2197, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.3243037974683555e-06, |
|
"loss": 0.2062, |
|
"step": 11075 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.317974683544304e-06, |
|
"loss": 0.1931, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.3116455696202535e-06, |
|
"loss": 0.2398, |
|
"step": 11125 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.305316455696203e-06, |
|
"loss": 0.2373, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.298987341772152e-06, |
|
"loss": 0.2451, |
|
"step": 11175 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.292658227848102e-06, |
|
"loss": 0.2365, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.286329113924051e-06, |
|
"loss": 0.231, |
|
"step": 11225 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.280000000000001e-06, |
|
"loss": 0.2325, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.27367088607595e-06, |
|
"loss": 0.1925, |
|
"step": 11275 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.2673417721518995e-06, |
|
"loss": 0.1746, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.261012658227848e-06, |
|
"loss": 0.1603, |
|
"step": 11325 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.2546835443037975e-06, |
|
"loss": 0.1629, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.248354430379747e-06, |
|
"loss": 0.1766, |
|
"step": 11375 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 7.242025316455696e-06, |
|
"loss": 0.1476, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.235696202531646e-06, |
|
"loss": 0.1618, |
|
"step": 11425 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.229367088607595e-06, |
|
"loss": 0.1549, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.223037974683545e-06, |
|
"loss": 0.1488, |
|
"step": 11475 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.216708860759494e-06, |
|
"loss": 0.1507, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.210379746835444e-06, |
|
"loss": 0.1584, |
|
"step": 11525 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.204050632911394e-06, |
|
"loss": 0.1541, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.197721518987343e-06, |
|
"loss": 0.1623, |
|
"step": 11575 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.191392405063293e-06, |
|
"loss": 0.1786, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.185063291139241e-06, |
|
"loss": 0.1671, |
|
"step": 11625 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.178734177215191e-06, |
|
"loss": 0.1486, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.17240506329114e-06, |
|
"loss": 0.1514, |
|
"step": 11675 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.1660759493670895e-06, |
|
"loss": 0.1546, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.159746835443039e-06, |
|
"loss": 0.1411, |
|
"step": 11725 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.153417721518988e-06, |
|
"loss": 0.163, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.147088607594938e-06, |
|
"loss": 0.1653, |
|
"step": 11775 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 7.140759493670887e-06, |
|
"loss": 0.1498, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.134430379746836e-06, |
|
"loss": 0.1613, |
|
"step": 11825 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.128101265822785e-06, |
|
"loss": 0.1665, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.121772151898735e-06, |
|
"loss": 0.1533, |
|
"step": 11875 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.115443037974684e-06, |
|
"loss": 0.1637, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.1091139240506335e-06, |
|
"loss": 0.1893, |
|
"step": 11925 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.102784810126583e-06, |
|
"loss": 0.197, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.096455696202532e-06, |
|
"loss": 0.156, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.090126582278482e-06, |
|
"loss": 0.1577, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.083797468354431e-06, |
|
"loss": 0.1614, |
|
"step": 12025 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.07746835443038e-06, |
|
"loss": 0.1458, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.071139240506329e-06, |
|
"loss": 0.1505, |
|
"step": 12075 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.064810126582279e-06, |
|
"loss": 0.1318, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.058481012658228e-06, |
|
"loss": 0.1314, |
|
"step": 12125 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.0521518987341775e-06, |
|
"loss": 0.1783, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.045822784810127e-06, |
|
"loss": 0.2108, |
|
"step": 12175 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.039493670886076e-06, |
|
"loss": 0.2441, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.033164556962026e-06, |
|
"loss": 0.2563, |
|
"step": 12225 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.026835443037975e-06, |
|
"loss": 0.2146, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.020506329113924e-06, |
|
"loss": 0.2431, |
|
"step": 12275 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.014177215189873e-06, |
|
"loss": 0.2101, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.007848101265823e-06, |
|
"loss": 0.2086, |
|
"step": 12325 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.001518987341773e-06, |
|
"loss": 0.2014, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 6.995189873417722e-06, |
|
"loss": 0.1805, |
|
"step": 12375 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 6.988860759493672e-06, |
|
"loss": 0.1986, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 6.982531645569621e-06, |
|
"loss": 0.1779, |
|
"step": 12425 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 6.976202531645571e-06, |
|
"loss": 0.1914, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 6.96987341772152e-06, |
|
"loss": 0.1614, |
|
"step": 12475 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 6.9635443037974694e-06, |
|
"loss": 0.2234, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"eval_loss": 0.3509575426578522, |
|
"eval_runtime": 1690.018, |
|
"eval_samples_per_second": 9.677, |
|
"eval_steps_per_second": 0.605, |
|
"eval_wer": 14.130385888249847, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 6.957215189873419e-06, |
|
"loss": 0.2149, |
|
"step": 12525 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 6.950886075949368e-06, |
|
"loss": 0.2528, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 6.944556962025317e-06, |
|
"loss": 0.2415, |
|
"step": 12575 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.938227848101266e-06, |
|
"loss": 0.205, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.931898734177216e-06, |
|
"loss": 0.2136, |
|
"step": 12625 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.925569620253165e-06, |
|
"loss": 0.2173, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.919240506329115e-06, |
|
"loss": 0.2126, |
|
"step": 12675 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.912911392405064e-06, |
|
"loss": 0.186, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.906835443037975e-06, |
|
"loss": 0.2234, |
|
"step": 12725 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.900506329113925e-06, |
|
"loss": 0.2118, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.894177215189874e-06, |
|
"loss": 0.2034, |
|
"step": 12775 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.887848101265824e-06, |
|
"loss": 0.2735, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.881518987341773e-06, |
|
"loss": 0.3176, |
|
"step": 12825 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.875189873417723e-06, |
|
"loss": 0.2868, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.868860759493672e-06, |
|
"loss": 0.2184, |
|
"step": 12875 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.8625316455696215e-06, |
|
"loss": 0.2127, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.856202531645571e-06, |
|
"loss": 0.2462, |
|
"step": 12925 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.8498734177215195e-06, |
|
"loss": 0.1899, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 6.843544303797469e-06, |
|
"loss": 0.2275, |
|
"step": 12975 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.837215189873418e-06, |
|
"loss": 0.1769, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.830886075949368e-06, |
|
"loss": 0.1739, |
|
"step": 13025 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.824556962025317e-06, |
|
"loss": 0.178, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.818227848101267e-06, |
|
"loss": 0.2253, |
|
"step": 13075 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.811898734177216e-06, |
|
"loss": 0.2199, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.8055696202531655e-06, |
|
"loss": 0.2224, |
|
"step": 13125 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.799240506329115e-06, |
|
"loss": 0.2148, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.7929113924050635e-06, |
|
"loss": 0.2615, |
|
"step": 13175 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.786582278481013e-06, |
|
"loss": 0.1635, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.780253164556962e-06, |
|
"loss": 0.177, |
|
"step": 13225 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.773924050632912e-06, |
|
"loss": 0.1738, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.767594936708861e-06, |
|
"loss": 0.2059, |
|
"step": 13275 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.761265822784811e-06, |
|
"loss": 0.2164, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.75493670886076e-06, |
|
"loss": 0.2346, |
|
"step": 13325 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.7486075949367095e-06, |
|
"loss": 0.2334, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 6.742278481012658e-06, |
|
"loss": 0.2114, |
|
"step": 13375 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.7359493670886075e-06, |
|
"loss": 0.1912, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.729620253164557e-06, |
|
"loss": 0.1892, |
|
"step": 13425 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.723291139240506e-06, |
|
"loss": 0.2073, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.716962025316456e-06, |
|
"loss": 0.2188, |
|
"step": 13475 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.710632911392405e-06, |
|
"loss": 0.253, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.704303797468355e-06, |
|
"loss": 0.2636, |
|
"step": 13525 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.697974683544304e-06, |
|
"loss": 0.1595, |
|
"step": 13550 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.691645569620254e-06, |
|
"loss": 0.1452, |
|
"step": 13575 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.685316455696204e-06, |
|
"loss": 0.2654, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.678987341772153e-06, |
|
"loss": 0.2343, |
|
"step": 13625 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.672658227848103e-06, |
|
"loss": 0.21, |
|
"step": 13650 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.666329113924051e-06, |
|
"loss": 0.193, |
|
"step": 13675 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.660000000000001e-06, |
|
"loss": 0.2393, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.65367088607595e-06, |
|
"loss": 0.1928, |
|
"step": 13725 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.6473417721518995e-06, |
|
"loss": 0.1911, |
|
"step": 13750 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.64126582278481e-06, |
|
"loss": 0.165, |
|
"step": 13775 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 6.6349367088607595e-06, |
|
"loss": 0.2005, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.628607594936709e-06, |
|
"loss": 0.2789, |
|
"step": 13825 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.622278481012658e-06, |
|
"loss": 0.2182, |
|
"step": 13850 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.615949367088608e-06, |
|
"loss": 0.1869, |
|
"step": 13875 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.609620253164557e-06, |
|
"loss": 0.1545, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.603291139240507e-06, |
|
"loss": 0.1962, |
|
"step": 13925 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.596962025316456e-06, |
|
"loss": 0.2219, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.590632911392406e-06, |
|
"loss": 0.1886, |
|
"step": 13975 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.584303797468356e-06, |
|
"loss": 0.2225, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.577974683544305e-06, |
|
"loss": 0.2407, |
|
"step": 14025 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.571645569620254e-06, |
|
"loss": 0.2198, |
|
"step": 14050 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.565316455696203e-06, |
|
"loss": 0.2021, |
|
"step": 14075 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.558987341772153e-06, |
|
"loss": 0.2058, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.552658227848102e-06, |
|
"loss": 0.1881, |
|
"step": 14125 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.5463291139240515e-06, |
|
"loss": 0.1933, |
|
"step": 14150 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.540000000000001e-06, |
|
"loss": 0.199, |
|
"step": 14175 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 6.53367088607595e-06, |
|
"loss": 0.1747, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.5273417721519e-06, |
|
"loss": 0.208, |
|
"step": 14225 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.521012658227849e-06, |
|
"loss": 0.2599, |
|
"step": 14250 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.514683544303798e-06, |
|
"loss": 0.2467, |
|
"step": 14275 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.508354430379747e-06, |
|
"loss": 0.2192, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.502025316455697e-06, |
|
"loss": 0.2267, |
|
"step": 14325 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.495696202531646e-06, |
|
"loss": 0.2067, |
|
"step": 14350 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.4893670886075955e-06, |
|
"loss": 0.2006, |
|
"step": 14375 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.483037974683545e-06, |
|
"loss": 0.1911, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.476708860759494e-06, |
|
"loss": 0.1925, |
|
"step": 14425 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.470379746835444e-06, |
|
"loss": 0.1888, |
|
"step": 14450 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.464050632911393e-06, |
|
"loss": 0.2042, |
|
"step": 14475 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.457721518987342e-06, |
|
"loss": 0.1605, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.451392405063291e-06, |
|
"loss": 0.159, |
|
"step": 14525 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.445063291139241e-06, |
|
"loss": 0.2122, |
|
"step": 14550 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.43873417721519e-06, |
|
"loss": 0.2133, |
|
"step": 14575 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 6.4324050632911395e-06, |
|
"loss": 0.2635, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.426075949367089e-06, |
|
"loss": 0.2274, |
|
"step": 14625 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.419746835443038e-06, |
|
"loss": 0.2142, |
|
"step": 14650 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.413417721518988e-06, |
|
"loss": 0.2181, |
|
"step": 14675 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.407088607594937e-06, |
|
"loss": 0.2092, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.400759493670886e-06, |
|
"loss": 0.2458, |
|
"step": 14725 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.394430379746835e-06, |
|
"loss": 0.238, |
|
"step": 14750 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.388101265822785e-06, |
|
"loss": 0.248, |
|
"step": 14775 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.381772151898735e-06, |
|
"loss": 0.2845, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.375443037974684e-06, |
|
"loss": 0.228, |
|
"step": 14825 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.369113924050634e-06, |
|
"loss": 0.2418, |
|
"step": 14850 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.362784810126583e-06, |
|
"loss": 0.217, |
|
"step": 14875 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.356455696202533e-06, |
|
"loss": 0.2518, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.350126582278482e-06, |
|
"loss": 0.2378, |
|
"step": 14925 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.3437974683544315e-06, |
|
"loss": 0.2615, |
|
"step": 14950 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.337468354430381e-06, |
|
"loss": 0.2259, |
|
"step": 14975 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.3311392405063295e-06, |
|
"loss": 0.1989, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"eval_loss": 0.3388429284095764, |
|
"eval_runtime": 1725.6593, |
|
"eval_samples_per_second": 9.477, |
|
"eval_steps_per_second": 0.593, |
|
"eval_wer": 13.639430646145936, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.324810126582279e-06, |
|
"loss": 0.2127, |
|
"step": 15025 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.318481012658228e-06, |
|
"loss": 0.2046, |
|
"step": 15050 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.312151898734178e-06, |
|
"loss": 0.2302, |
|
"step": 15075 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.305822784810127e-06, |
|
"loss": 0.2622, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.299493670886077e-06, |
|
"loss": 0.2398, |
|
"step": 15125 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.293164556962026e-06, |
|
"loss": 0.245, |
|
"step": 15150 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.2868354430379755e-06, |
|
"loss": 0.2151, |
|
"step": 15175 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.280506329113925e-06, |
|
"loss": 0.2445, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.2741772151898735e-06, |
|
"loss": 0.2183, |
|
"step": 15225 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.267848101265823e-06, |
|
"loss": 0.2083, |
|
"step": 15250 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.261518987341772e-06, |
|
"loss": 0.2186, |
|
"step": 15275 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.255189873417722e-06, |
|
"loss": 0.2258, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.248860759493671e-06, |
|
"loss": 0.2083, |
|
"step": 15325 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.242531645569621e-06, |
|
"loss": 0.2169, |
|
"step": 15350 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.23620253164557e-06, |
|
"loss": 0.2482, |
|
"step": 15375 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.2298734177215195e-06, |
|
"loss": 0.2308, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.223544303797469e-06, |
|
"loss": 0.2173, |
|
"step": 15425 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.2172151898734175e-06, |
|
"loss": 0.2267, |
|
"step": 15450 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.210886075949367e-06, |
|
"loss": 0.2102, |
|
"step": 15475 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.204556962025316e-06, |
|
"loss": 0.1675, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.198227848101266e-06, |
|
"loss": 0.2046, |
|
"step": 15525 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.191898734177215e-06, |
|
"loss": 0.2701, |
|
"step": 15550 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.185569620253165e-06, |
|
"loss": 0.2554, |
|
"step": 15575 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.179240506329114e-06, |
|
"loss": 0.223, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.172911392405064e-06, |
|
"loss": 0.2288, |
|
"step": 15625 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.166582278481014e-06, |
|
"loss": 0.2321, |
|
"step": 15650 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.160253164556963e-06, |
|
"loss": 0.1938, |
|
"step": 15675 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.153924050632913e-06, |
|
"loss": 0.2805, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.147594936708862e-06, |
|
"loss": 0.243, |
|
"step": 15725 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.141265822784811e-06, |
|
"loss": 0.212, |
|
"step": 15750 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.13493670886076e-06, |
|
"loss": 0.1963, |
|
"step": 15775 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.1286075949367095e-06, |
|
"loss": 0.2742, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.122278481012659e-06, |
|
"loss": 0.2601, |
|
"step": 15825 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.115949367088608e-06, |
|
"loss": 0.2595, |
|
"step": 15850 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.109620253164558e-06, |
|
"loss": 0.2254, |
|
"step": 15875 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.103291139240507e-06, |
|
"loss": 0.2145, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.096962025316457e-06, |
|
"loss": 0.221, |
|
"step": 15925 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.090632911392406e-06, |
|
"loss": 0.2184, |
|
"step": 15950 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.084303797468355e-06, |
|
"loss": 0.2136, |
|
"step": 15975 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.077974683544304e-06, |
|
"loss": 0.2079, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.0716455696202535e-06, |
|
"loss": 0.2321, |
|
"step": 16025 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.065316455696203e-06, |
|
"loss": 0.2126, |
|
"step": 16050 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.058987341772152e-06, |
|
"loss": 0.2191, |
|
"step": 16075 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.052658227848102e-06, |
|
"loss": 0.2271, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.046329113924051e-06, |
|
"loss": 0.2502, |
|
"step": 16125 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.040000000000001e-06, |
|
"loss": 0.2581, |
|
"step": 16150 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.03367088607595e-06, |
|
"loss": 0.2649, |
|
"step": 16175 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.027341772151899e-06, |
|
"loss": 0.2731, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.021012658227848e-06, |
|
"loss": 0.2318, |
|
"step": 16225 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.0146835443037975e-06, |
|
"loss": 0.219, |
|
"step": 16250 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.008354430379747e-06, |
|
"loss": 0.2797, |
|
"step": 16275 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.002025316455696e-06, |
|
"loss": 0.248, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.995696202531646e-06, |
|
"loss": 0.2376, |
|
"step": 16325 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.989367088607595e-06, |
|
"loss": 0.2607, |
|
"step": 16350 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.983037974683545e-06, |
|
"loss": 0.2447, |
|
"step": 16375 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.976708860759494e-06, |
|
"loss": 0.2473, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.970379746835444e-06, |
|
"loss": 0.2214, |
|
"step": 16425 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.964050632911394e-06, |
|
"loss": 0.2496, |
|
"step": 16450 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.957721518987342e-06, |
|
"loss": 0.2203, |
|
"step": 16475 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.951392405063292e-06, |
|
"loss": 0.2189, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.945063291139241e-06, |
|
"loss": 0.2163, |
|
"step": 16525 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.938734177215191e-06, |
|
"loss": 0.2501, |
|
"step": 16550 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.93240506329114e-06, |
|
"loss": 0.2611, |
|
"step": 16575 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 5.9260759493670895e-06, |
|
"loss": 0.2476, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.919746835443039e-06, |
|
"loss": 0.2346, |
|
"step": 16625 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.913417721518988e-06, |
|
"loss": 0.2674, |
|
"step": 16650 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.907088607594938e-06, |
|
"loss": 0.2245, |
|
"step": 16675 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.900759493670886e-06, |
|
"loss": 0.2261, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.894430379746836e-06, |
|
"loss": 0.2153, |
|
"step": 16725 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.888101265822785e-06, |
|
"loss": 0.2489, |
|
"step": 16750 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.881772151898735e-06, |
|
"loss": 0.2226, |
|
"step": 16775 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.875443037974684e-06, |
|
"loss": 0.2031, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.8691139240506335e-06, |
|
"loss": 0.2307, |
|
"step": 16825 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.862784810126583e-06, |
|
"loss": 0.2668, |
|
"step": 16850 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.856455696202532e-06, |
|
"loss": 0.2345, |
|
"step": 16875 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.850126582278482e-06, |
|
"loss": 0.2368, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.84379746835443e-06, |
|
"loss": 0.2362, |
|
"step": 16925 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.83746835443038e-06, |
|
"loss": 0.2231, |
|
"step": 16950 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.831139240506329e-06, |
|
"loss": 0.2308, |
|
"step": 16975 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 5.824810126582279e-06, |
|
"loss": 0.2385, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.818481012658228e-06, |
|
"loss": 0.2242, |
|
"step": 17025 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.8121518987341775e-06, |
|
"loss": 0.2207, |
|
"step": 17050 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.805822784810127e-06, |
|
"loss": 0.2465, |
|
"step": 17075 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.799493670886076e-06, |
|
"loss": 0.231, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.793164556962026e-06, |
|
"loss": 0.2151, |
|
"step": 17125 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.786835443037974e-06, |
|
"loss": 0.208, |
|
"step": 17150 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.780506329113924e-06, |
|
"loss": 0.1771, |
|
"step": 17175 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.774177215189873e-06, |
|
"loss": 0.1272, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.767848101265823e-06, |
|
"loss": 0.1704, |
|
"step": 17225 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.761518987341773e-06, |
|
"loss": 0.1651, |
|
"step": 17250 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.755189873417722e-06, |
|
"loss": 0.2129, |
|
"step": 17275 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.748860759493672e-06, |
|
"loss": 0.254, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.742531645569621e-06, |
|
"loss": 0.2359, |
|
"step": 17325 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.736202531645571e-06, |
|
"loss": 0.1956, |
|
"step": 17350 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.72987341772152e-06, |
|
"loss": 0.1671, |
|
"step": 17375 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 5.7235443037974695e-06, |
|
"loss": 0.1724, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.717215189873419e-06, |
|
"loss": 0.1629, |
|
"step": 17425 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.7108860759493675e-06, |
|
"loss": 0.1353, |
|
"step": 17450 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.704556962025317e-06, |
|
"loss": 0.16, |
|
"step": 17475 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.698227848101266e-06, |
|
"loss": 0.2449, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"eval_loss": 0.33941036462783813, |
|
"eval_runtime": 1701.5958, |
|
"eval_samples_per_second": 9.611, |
|
"eval_steps_per_second": 0.601, |
|
"eval_wer": 13.429296661632875, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.691898734177216e-06, |
|
"loss": 0.277, |
|
"step": 17525 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.685569620253165e-06, |
|
"loss": 0.474, |
|
"step": 17550 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.679240506329115e-06, |
|
"loss": 0.774, |
|
"step": 17575 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.672911392405064e-06, |
|
"loss": 0.7322, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.6665822784810135e-06, |
|
"loss": 0.2992, |
|
"step": 17625 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.660253164556963e-06, |
|
"loss": 0.1059, |
|
"step": 17650 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.6539240506329115e-06, |
|
"loss": 0.1396, |
|
"step": 17675 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.647594936708861e-06, |
|
"loss": 0.2348, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.64126582278481e-06, |
|
"loss": 0.264, |
|
"step": 17725 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.63493670886076e-06, |
|
"loss": 0.1593, |
|
"step": 17750 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.628607594936709e-06, |
|
"loss": 0.1322, |
|
"step": 17775 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.622278481012659e-06, |
|
"loss": 0.1866, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.615949367088608e-06, |
|
"loss": 0.2874, |
|
"step": 17825 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.6096202531645575e-06, |
|
"loss": 0.3162, |
|
"step": 17850 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.603291139240507e-06, |
|
"loss": 0.2407, |
|
"step": 17875 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.5969620253164555e-06, |
|
"loss": 0.1775, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.590632911392405e-06, |
|
"loss": 0.1833, |
|
"step": 17925 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.584303797468354e-06, |
|
"loss": 0.2227, |
|
"step": 17950 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.578227848101267e-06, |
|
"loss": 0.2246, |
|
"step": 17975 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.571898734177216e-06, |
|
"loss": 0.2236, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.565569620253165e-06, |
|
"loss": 0.1513, |
|
"step": 18025 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.559240506329114e-06, |
|
"loss": 0.1528, |
|
"step": 18050 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.5529113924050635e-06, |
|
"loss": 0.194, |
|
"step": 18075 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.546582278481013e-06, |
|
"loss": 0.2515, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.540253164556962e-06, |
|
"loss": 0.245, |
|
"step": 18125 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.533924050632912e-06, |
|
"loss": 0.1739, |
|
"step": 18150 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.527594936708861e-06, |
|
"loss": 0.1539, |
|
"step": 18175 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.521265822784811e-06, |
|
"loss": 0.1294, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.51493670886076e-06, |
|
"loss": 0.1334, |
|
"step": 18225 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.508607594936709e-06, |
|
"loss": 0.1466, |
|
"step": 18250 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.502278481012658e-06, |
|
"loss": 0.1302, |
|
"step": 18275 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.4959493670886075e-06, |
|
"loss": 0.1164, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.489620253164557e-06, |
|
"loss": 0.1098, |
|
"step": 18325 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.483291139240506e-06, |
|
"loss": 0.0942, |
|
"step": 18350 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.476962025316456e-06, |
|
"loss": 0.163, |
|
"step": 18375 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.470632911392405e-06, |
|
"loss": 0.2109, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.464303797468355e-06, |
|
"loss": 0.1966, |
|
"step": 18425 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.457974683544304e-06, |
|
"loss": 0.2097, |
|
"step": 18450 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.451645569620254e-06, |
|
"loss": 0.2255, |
|
"step": 18475 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.445316455696204e-06, |
|
"loss": 0.2523, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.438987341772153e-06, |
|
"loss": 0.234, |
|
"step": 18525 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.432658227848102e-06, |
|
"loss": 0.1941, |
|
"step": 18550 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.426329113924051e-06, |
|
"loss": 0.2105, |
|
"step": 18575 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.420000000000001e-06, |
|
"loss": 0.206, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.41367088607595e-06, |
|
"loss": 0.1914, |
|
"step": 18625 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.4073417721518995e-06, |
|
"loss": 0.2029, |
|
"step": 18650 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.401012658227849e-06, |
|
"loss": 0.1994, |
|
"step": 18675 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.394683544303798e-06, |
|
"loss": 0.2047, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.388354430379748e-06, |
|
"loss": 0.2319, |
|
"step": 18725 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.382025316455697e-06, |
|
"loss": 0.1913, |
|
"step": 18750 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.375696202531646e-06, |
|
"loss": 0.2612, |
|
"step": 18775 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.369367088607595e-06, |
|
"loss": 0.2278, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.363037974683545e-06, |
|
"loss": 0.2344, |
|
"step": 18825 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.356708860759494e-06, |
|
"loss": 0.221, |
|
"step": 18850 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.3503797468354435e-06, |
|
"loss": 0.2172, |
|
"step": 18875 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.344050632911393e-06, |
|
"loss": 0.2094, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.337721518987342e-06, |
|
"loss": 0.2207, |
|
"step": 18925 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.331392405063292e-06, |
|
"loss": 0.2068, |
|
"step": 18950 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.325063291139241e-06, |
|
"loss": 0.2051, |
|
"step": 18975 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.31873417721519e-06, |
|
"loss": 0.2141, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.312405063291139e-06, |
|
"loss": 0.1912, |
|
"step": 19025 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.306075949367089e-06, |
|
"loss": 0.1826, |
|
"step": 19050 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.299746835443038e-06, |
|
"loss": 0.1913, |
|
"step": 19075 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.2934177215189875e-06, |
|
"loss": 0.1998, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.287088607594937e-06, |
|
"loss": 0.1984, |
|
"step": 19125 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.280759493670886e-06, |
|
"loss": 0.2284, |
|
"step": 19150 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.274430379746836e-06, |
|
"loss": 0.2196, |
|
"step": 19175 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.268101265822785e-06, |
|
"loss": 0.2287, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.261772151898734e-06, |
|
"loss": 0.2265, |
|
"step": 19225 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.255443037974683e-06, |
|
"loss": 0.2352, |
|
"step": 19250 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.249113924050633e-06, |
|
"loss": 0.2401, |
|
"step": 19275 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.242784810126583e-06, |
|
"loss": 0.2415, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.236455696202532e-06, |
|
"loss": 0.2314, |
|
"step": 19325 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.230126582278482e-06, |
|
"loss": 0.215, |
|
"step": 19350 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.223797468354431e-06, |
|
"loss": 0.1987, |
|
"step": 19375 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.217468354430381e-06, |
|
"loss": 0.1909, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.21113924050633e-06, |
|
"loss": 0.2259, |
|
"step": 19425 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.2048101265822795e-06, |
|
"loss": 0.1908, |
|
"step": 19450 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.198481012658229e-06, |
|
"loss": 0.2072, |
|
"step": 19475 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.192151898734178e-06, |
|
"loss": 0.2124, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.185822784810127e-06, |
|
"loss": 0.187, |
|
"step": 19525 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.179493670886076e-06, |
|
"loss": 0.2084, |
|
"step": 19550 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.173164556962026e-06, |
|
"loss": 0.2162, |
|
"step": 19575 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.166835443037975e-06, |
|
"loss": 0.251, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.160506329113925e-06, |
|
"loss": 0.2351, |
|
"step": 19625 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.154177215189874e-06, |
|
"loss": 0.2399, |
|
"step": 19650 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.1478481012658235e-06, |
|
"loss": 0.2498, |
|
"step": 19675 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.141518987341773e-06, |
|
"loss": 0.2261, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.135189873417722e-06, |
|
"loss": 0.2545, |
|
"step": 19725 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.128860759493671e-06, |
|
"loss": 0.2289, |
|
"step": 19750 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.12253164556962e-06, |
|
"loss": 0.2363, |
|
"step": 19775 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.11620253164557e-06, |
|
"loss": 0.2268, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.109873417721519e-06, |
|
"loss": 0.2546, |
|
"step": 19825 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.103544303797469e-06, |
|
"loss": 0.2472, |
|
"step": 19850 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.097215189873418e-06, |
|
"loss": 0.2444, |
|
"step": 19875 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.0908860759493675e-06, |
|
"loss": 0.23, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.084556962025317e-06, |
|
"loss": 0.2227, |
|
"step": 19925 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.0782278481012655e-06, |
|
"loss": 0.2272, |
|
"step": 19950 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.071898734177215e-06, |
|
"loss": 0.2351, |
|
"step": 19975 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.065569620253164e-06, |
|
"loss": 0.2302, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"eval_loss": 0.31980159878730774, |
|
"eval_runtime": 1736.6344, |
|
"eval_samples_per_second": 9.417, |
|
"eval_steps_per_second": 0.589, |
|
"eval_wer": 12.502008161166984, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.059240506329114e-06, |
|
"loss": 0.2162, |
|
"step": 20025 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.052911392405063e-06, |
|
"loss": 0.2111, |
|
"step": 20050 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.046582278481013e-06, |
|
"loss": 0.2077, |
|
"step": 20075 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.040253164556962e-06, |
|
"loss": 0.2024, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.033924050632912e-06, |
|
"loss": 0.2179, |
|
"step": 20125 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.027594936708862e-06, |
|
"loss": 0.2215, |
|
"step": 20150 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.021265822784811e-06, |
|
"loss": 0.2073, |
|
"step": 20175 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.014936708860761e-06, |
|
"loss": 0.1948, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.00860759493671e-06, |
|
"loss": 0.2057, |
|
"step": 20225 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.002278481012659e-06, |
|
"loss": 0.2114, |
|
"step": 20250 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.995949367088608e-06, |
|
"loss": 0.2008, |
|
"step": 20275 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.9896202531645575e-06, |
|
"loss": 0.2053, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.983291139240507e-06, |
|
"loss": 0.2223, |
|
"step": 20325 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.976962025316456e-06, |
|
"loss": 0.2132, |
|
"step": 20350 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.970632911392406e-06, |
|
"loss": 0.2103, |
|
"step": 20375 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.964303797468355e-06, |
|
"loss": 0.2046, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.957974683544305e-06, |
|
"loss": 0.2284, |
|
"step": 20425 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.951645569620254e-06, |
|
"loss": 0.2437, |
|
"step": 20450 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.945316455696203e-06, |
|
"loss": 0.2214, |
|
"step": 20475 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.938987341772152e-06, |
|
"loss": 0.2152, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.9326582278481015e-06, |
|
"loss": 0.2505, |
|
"step": 20525 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.926329113924051e-06, |
|
"loss": 0.2742, |
|
"step": 20550 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.92e-06, |
|
"loss": 0.2768, |
|
"step": 20575 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.91367088607595e-06, |
|
"loss": 0.2845, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.907341772151899e-06, |
|
"loss": 0.2701, |
|
"step": 20625 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.901012658227849e-06, |
|
"loss": 0.2628, |
|
"step": 20650 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.894683544303798e-06, |
|
"loss": 0.2589, |
|
"step": 20675 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.888354430379747e-06, |
|
"loss": 0.28, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.882025316455696e-06, |
|
"loss": 0.2577, |
|
"step": 20725 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.875696202531646e-06, |
|
"loss": 0.2592, |
|
"step": 20750 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.869367088607596e-06, |
|
"loss": 0.272, |
|
"step": 20775 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.863037974683545e-06, |
|
"loss": 0.2644, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.856708860759495e-06, |
|
"loss": 0.2495, |
|
"step": 20825 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.850379746835443e-06, |
|
"loss": 0.2546, |
|
"step": 20850 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.844050632911393e-06, |
|
"loss": 0.2509, |
|
"step": 20875 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.837721518987342e-06, |
|
"loss": 0.2477, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.8313924050632915e-06, |
|
"loss": 0.225, |
|
"step": 20925 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.825063291139241e-06, |
|
"loss": 0.2302, |
|
"step": 20950 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.81873417721519e-06, |
|
"loss": 0.2139, |
|
"step": 20975 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.81240506329114e-06, |
|
"loss": 0.2063, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.806075949367089e-06, |
|
"loss": 0.2147, |
|
"step": 21025 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.799746835443039e-06, |
|
"loss": 0.1982, |
|
"step": 21050 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.793417721518987e-06, |
|
"loss": 0.2079, |
|
"step": 21075 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.787088607594937e-06, |
|
"loss": 0.228, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.780759493670886e-06, |
|
"loss": 0.212, |
|
"step": 21125 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.774430379746836e-06, |
|
"loss": 0.2448, |
|
"step": 21150 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.768101265822786e-06, |
|
"loss": 0.1985, |
|
"step": 21175 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.761772151898735e-06, |
|
"loss": 0.2294, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.755443037974684e-06, |
|
"loss": 0.2368, |
|
"step": 21225 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.749113924050633e-06, |
|
"loss": 0.2637, |
|
"step": 21250 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.742784810126583e-06, |
|
"loss": 0.2541, |
|
"step": 21275 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.736455696202532e-06, |
|
"loss": 0.2097, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.7301265822784815e-06, |
|
"loss": 0.211, |
|
"step": 21325 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.723797468354431e-06, |
|
"loss": 0.2524, |
|
"step": 21350 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.71746835443038e-06, |
|
"loss": 0.3004, |
|
"step": 21375 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.71113924050633e-06, |
|
"loss": 0.2308, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.704810126582279e-06, |
|
"loss": 0.2243, |
|
"step": 21425 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.698481012658228e-06, |
|
"loss": 0.226, |
|
"step": 21450 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.692151898734177e-06, |
|
"loss": 0.219, |
|
"step": 21475 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.685822784810127e-06, |
|
"loss": 0.1913, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.679493670886076e-06, |
|
"loss": 0.1959, |
|
"step": 21525 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.6731645569620255e-06, |
|
"loss": 0.2098, |
|
"step": 21550 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.666835443037975e-06, |
|
"loss": 0.1948, |
|
"step": 21575 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.660506329113924e-06, |
|
"loss": 0.2013, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.654177215189874e-06, |
|
"loss": 0.208, |
|
"step": 21625 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.647848101265823e-06, |
|
"loss": 0.2017, |
|
"step": 21650 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.641518987341773e-06, |
|
"loss": 0.2029, |
|
"step": 21675 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.635189873417722e-06, |
|
"loss": 0.2068, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.6288607594936715e-06, |
|
"loss": 0.2205, |
|
"step": 21725 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.622531645569621e-06, |
|
"loss": 0.2261, |
|
"step": 21750 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.61620253164557e-06, |
|
"loss": 0.2373, |
|
"step": 21775 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.609873417721519e-06, |
|
"loss": 0.211, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.603544303797468e-06, |
|
"loss": 0.2143, |
|
"step": 21825 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.597215189873418e-06, |
|
"loss": 0.2023, |
|
"step": 21850 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.590886075949367e-06, |
|
"loss": 0.186, |
|
"step": 21875 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.584556962025317e-06, |
|
"loss": 0.2535, |
|
"step": 21900 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.578227848101266e-06, |
|
"loss": 0.2313, |
|
"step": 21925 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.5718987341772155e-06, |
|
"loss": 0.2022, |
|
"step": 21950 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.565822784810127e-06, |
|
"loss": 0.2157, |
|
"step": 21975 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.559493670886076e-06, |
|
"loss": 0.2072, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.553164556962026e-06, |
|
"loss": 0.2091, |
|
"step": 22025 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.546835443037975e-06, |
|
"loss": 0.278, |
|
"step": 22050 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.540506329113925e-06, |
|
"loss": 0.227, |
|
"step": 22075 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.534177215189874e-06, |
|
"loss": 0.2057, |
|
"step": 22100 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.5278481012658235e-06, |
|
"loss": 0.2365, |
|
"step": 22125 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.521518987341773e-06, |
|
"loss": 0.2349, |
|
"step": 22150 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.5151898734177215e-06, |
|
"loss": 0.205, |
|
"step": 22175 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.508860759493671e-06, |
|
"loss": 0.2246, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.50253164556962e-06, |
|
"loss": 0.2049, |
|
"step": 22225 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.49620253164557e-06, |
|
"loss": 0.223, |
|
"step": 22250 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.489873417721519e-06, |
|
"loss": 0.2044, |
|
"step": 22275 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.483544303797469e-06, |
|
"loss": 0.1859, |
|
"step": 22300 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.477215189873418e-06, |
|
"loss": 0.2117, |
|
"step": 22325 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.4708860759493675e-06, |
|
"loss": 0.2022, |
|
"step": 22350 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.464556962025317e-06, |
|
"loss": 0.2307, |
|
"step": 22375 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.458227848101266e-06, |
|
"loss": 0.2187, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.451898734177216e-06, |
|
"loss": 0.1981, |
|
"step": 22425 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.445569620253165e-06, |
|
"loss": 0.206, |
|
"step": 22450 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.439240506329115e-06, |
|
"loss": 0.2381, |
|
"step": 22475 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.433164556962026e-06, |
|
"loss": 0.213, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"eval_loss": 0.31673741340637207, |
|
"eval_runtime": 2166.263, |
|
"eval_samples_per_second": 7.549, |
|
"eval_steps_per_second": 0.472, |
|
"eval_wer": 12.490441152845163, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.4268354430379755e-06, |
|
"loss": 0.2, |
|
"step": 22525 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.420506329113924e-06, |
|
"loss": 0.2257, |
|
"step": 22550 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.4141772151898735e-06, |
|
"loss": 0.1922, |
|
"step": 22575 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.407848101265823e-06, |
|
"loss": 0.184, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.401518987341772e-06, |
|
"loss": 0.1837, |
|
"step": 22625 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.395189873417722e-06, |
|
"loss": 0.1803, |
|
"step": 22650 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.388860759493671e-06, |
|
"loss": 0.231, |
|
"step": 22675 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.382531645569621e-06, |
|
"loss": 0.1974, |
|
"step": 22700 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.37620253164557e-06, |
|
"loss": 0.2205, |
|
"step": 22725 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.369873417721519e-06, |
|
"loss": 0.2233, |
|
"step": 22750 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.363544303797469e-06, |
|
"loss": 0.1911, |
|
"step": 22775 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.357215189873418e-06, |
|
"loss": 0.2081, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.350886075949368e-06, |
|
"loss": 0.2496, |
|
"step": 22825 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.344556962025317e-06, |
|
"loss": 0.2239, |
|
"step": 22850 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.338227848101267e-06, |
|
"loss": 0.224, |
|
"step": 22875 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.331898734177215e-06, |
|
"loss": 0.2181, |
|
"step": 22900 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.325569620253165e-06, |
|
"loss": 0.2152, |
|
"step": 22925 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.319240506329114e-06, |
|
"loss": 0.2059, |
|
"step": 22950 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.3129113924050635e-06, |
|
"loss": 0.2167, |
|
"step": 22975 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.306582278481013e-06, |
|
"loss": 0.251, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.300253164556962e-06, |
|
"loss": 0.2033, |
|
"step": 23025 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.293924050632912e-06, |
|
"loss": 0.1936, |
|
"step": 23050 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.287594936708861e-06, |
|
"loss": 0.206, |
|
"step": 23075 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.281265822784811e-06, |
|
"loss": 0.2192, |
|
"step": 23100 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.274936708860759e-06, |
|
"loss": 0.2187, |
|
"step": 23125 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.268607594936709e-06, |
|
"loss": 0.2277, |
|
"step": 23150 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.262278481012659e-06, |
|
"loss": 0.3065, |
|
"step": 23175 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.255949367088608e-06, |
|
"loss": 0.2418, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.249620253164558e-06, |
|
"loss": 0.1922, |
|
"step": 23225 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.243291139240507e-06, |
|
"loss": 0.193, |
|
"step": 23250 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.236962025316456e-06, |
|
"loss": 0.1954, |
|
"step": 23275 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.230632911392405e-06, |
|
"loss": 0.18, |
|
"step": 23300 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.224303797468355e-06, |
|
"loss": 0.1747, |
|
"step": 23325 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.217974683544304e-06, |
|
"loss": 0.2124, |
|
"step": 23350 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.2116455696202535e-06, |
|
"loss": 0.2162, |
|
"step": 23375 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.205316455696203e-06, |
|
"loss": 0.2529, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.198987341772152e-06, |
|
"loss": 0.2168, |
|
"step": 23425 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.192658227848102e-06, |
|
"loss": 0.1919, |
|
"step": 23450 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.186329113924051e-06, |
|
"loss": 0.221, |
|
"step": 23475 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.18e-06, |
|
"loss": 0.2093, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.173670886075949e-06, |
|
"loss": 0.1925, |
|
"step": 23525 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.167341772151899e-06, |
|
"loss": 0.1968, |
|
"step": 23550 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.161012658227848e-06, |
|
"loss": 0.229, |
|
"step": 23575 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.154683544303798e-06, |
|
"loss": 0.2405, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.148354430379748e-06, |
|
"loss": 0.2285, |
|
"step": 23625 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.142025316455696e-06, |
|
"loss": 0.16, |
|
"step": 23650 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.135696202531646e-06, |
|
"loss": 0.1483, |
|
"step": 23675 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.129367088607595e-06, |
|
"loss": 0.1521, |
|
"step": 23700 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.123037974683545e-06, |
|
"loss": 0.1457, |
|
"step": 23725 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.116708860759494e-06, |
|
"loss": 0.1415, |
|
"step": 23750 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.1103797468354435e-06, |
|
"loss": 0.1492, |
|
"step": 23775 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.104050632911393e-06, |
|
"loss": 0.1544, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.097721518987342e-06, |
|
"loss": 0.1265, |
|
"step": 23825 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.091392405063292e-06, |
|
"loss": 0.1914, |
|
"step": 23850 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.08506329113924e-06, |
|
"loss": 0.2038, |
|
"step": 23875 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.07873417721519e-06, |
|
"loss": 0.1847, |
|
"step": 23900 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.072405063291139e-06, |
|
"loss": 0.1615, |
|
"step": 23925 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.066075949367089e-06, |
|
"loss": 0.1045, |
|
"step": 23950 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.059746835443038e-06, |
|
"loss": 0.0857, |
|
"step": 23975 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.053417721518988e-06, |
|
"loss": 0.0625, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.047088607594937e-06, |
|
"loss": 0.1157, |
|
"step": 24025 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.040759493670886e-06, |
|
"loss": 0.2766, |
|
"step": 24050 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.034430379746836e-06, |
|
"loss": 0.3155, |
|
"step": 24075 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.028101265822785e-06, |
|
"loss": 0.3143, |
|
"step": 24100 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.021772151898735e-06, |
|
"loss": 0.3236, |
|
"step": 24125 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.015443037974684e-06, |
|
"loss": 0.2577, |
|
"step": 24150 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.0091139240506335e-06, |
|
"loss": 0.1949, |
|
"step": 24175 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 4.002784810126583e-06, |
|
"loss": 0.1884, |
|
"step": 24200 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.996455696202532e-06, |
|
"loss": 0.1997, |
|
"step": 24225 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.990126582278481e-06, |
|
"loss": 0.3531, |
|
"step": 24250 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.98379746835443e-06, |
|
"loss": 0.8159, |
|
"step": 24275 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.97746835443038e-06, |
|
"loss": 0.872, |
|
"step": 24300 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.971139240506329e-06, |
|
"loss": 0.8259, |
|
"step": 24325 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.964810126582279e-06, |
|
"loss": 0.8696, |
|
"step": 24350 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.958481012658228e-06, |
|
"loss": 0.6438, |
|
"step": 24375 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.9521518987341775e-06, |
|
"loss": 0.2854, |
|
"step": 24400 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.945822784810127e-06, |
|
"loss": 0.2291, |
|
"step": 24425 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.939493670886076e-06, |
|
"loss": 0.1981, |
|
"step": 24450 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.933164556962026e-06, |
|
"loss": 0.188, |
|
"step": 24475 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.926835443037975e-06, |
|
"loss": 0.2366, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.920506329113925e-06, |
|
"loss": 0.2776, |
|
"step": 24525 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.914177215189874e-06, |
|
"loss": 0.2613, |
|
"step": 24550 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.9078481012658235e-06, |
|
"loss": 0.2569, |
|
"step": 24575 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.901518987341772e-06, |
|
"loss": 0.2388, |
|
"step": 24600 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.8951898734177215e-06, |
|
"loss": 0.2275, |
|
"step": 24625 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.888860759493671e-06, |
|
"loss": 0.2533, |
|
"step": 24650 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.88253164556962e-06, |
|
"loss": 0.2547, |
|
"step": 24675 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.87620253164557e-06, |
|
"loss": 0.2273, |
|
"step": 24700 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.869873417721519e-06, |
|
"loss": 0.2231, |
|
"step": 24725 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.863544303797469e-06, |
|
"loss": 0.2101, |
|
"step": 24750 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.857215189873418e-06, |
|
"loss": 0.2015, |
|
"step": 24775 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.8508860759493675e-06, |
|
"loss": 0.2064, |
|
"step": 24800 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.844556962025317e-06, |
|
"loss": 0.1784, |
|
"step": 24825 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.838227848101266e-06, |
|
"loss": 0.1797, |
|
"step": 24850 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.831898734177216e-06, |
|
"loss": 0.1853, |
|
"step": 24875 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.825569620253165e-06, |
|
"loss": 0.2025, |
|
"step": 24900 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.819240506329115e-06, |
|
"loss": 0.245, |
|
"step": 24925 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.8129113924050636e-06, |
|
"loss": 0.2464, |
|
"step": 24950 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.806582278481013e-06, |
|
"loss": 0.2546, |
|
"step": 24975 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.8002531645569625e-06, |
|
"loss": 0.2395, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"eval_loss": 0.3145264983177185, |
|
"eval_runtime": 2219.7626, |
|
"eval_samples_per_second": 7.367, |
|
"eval_steps_per_second": 0.461, |
|
"eval_wer": 12.753269286379847, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.7941772151898735e-06, |
|
"loss": 0.2256, |
|
"step": 25025 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.787848101265823e-06, |
|
"loss": 0.2138, |
|
"step": 25050 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.7815189873417724e-06, |
|
"loss": 0.2107, |
|
"step": 25075 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.775189873417722e-06, |
|
"loss": 0.2126, |
|
"step": 25100 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.7688607594936712e-06, |
|
"loss": 0.2331, |
|
"step": 25125 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.7625316455696202e-06, |
|
"loss": 0.2397, |
|
"step": 25150 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.7562025316455697e-06, |
|
"loss": 0.2273, |
|
"step": 25175 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.749873417721519e-06, |
|
"loss": 0.2068, |
|
"step": 25200 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.743544303797469e-06, |
|
"loss": 0.2072, |
|
"step": 25225 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.7372151898734184e-06, |
|
"loss": 0.2108, |
|
"step": 25250 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.730886075949368e-06, |
|
"loss": 0.1854, |
|
"step": 25275 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.724556962025317e-06, |
|
"loss": 0.1906, |
|
"step": 25300 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.7182278481012662e-06, |
|
"loss": 0.2085, |
|
"step": 25325 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.7118987341772156e-06, |
|
"loss": 0.2045, |
|
"step": 25350 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.705569620253165e-06, |
|
"loss": 0.199, |
|
"step": 25375 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.699240506329114e-06, |
|
"loss": 0.2099, |
|
"step": 25400 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.6929113924050635e-06, |
|
"loss": 0.204, |
|
"step": 25425 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.686582278481013e-06, |
|
"loss": 0.2245, |
|
"step": 25450 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.6802531645569624e-06, |
|
"loss": 0.2248, |
|
"step": 25475 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.6739240506329114e-06, |
|
"loss": 0.2393, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.667594936708861e-06, |
|
"loss": 0.2569, |
|
"step": 25525 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.6612658227848102e-06, |
|
"loss": 0.2319, |
|
"step": 25550 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.6549367088607597e-06, |
|
"loss": 0.221, |
|
"step": 25575 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.648607594936709e-06, |
|
"loss": 0.2361, |
|
"step": 25600 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.642278481012659e-06, |
|
"loss": 0.2289, |
|
"step": 25625 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.635949367088608e-06, |
|
"loss": 0.2123, |
|
"step": 25650 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.6296202531645574e-06, |
|
"loss": 0.2074, |
|
"step": 25675 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.623291139240507e-06, |
|
"loss": 0.2168, |
|
"step": 25700 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.6169620253164562e-06, |
|
"loss": 0.1936, |
|
"step": 25725 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.6106329113924056e-06, |
|
"loss": 0.2586, |
|
"step": 25750 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.6043037974683546e-06, |
|
"loss": 0.2039, |
|
"step": 25775 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.597974683544304e-06, |
|
"loss": 0.2123, |
|
"step": 25800 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.5916455696202535e-06, |
|
"loss": 0.2128, |
|
"step": 25825 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.585316455696203e-06, |
|
"loss": 0.2318, |
|
"step": 25850 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.578987341772152e-06, |
|
"loss": 0.2294, |
|
"step": 25875 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.5726582278481014e-06, |
|
"loss": 0.2298, |
|
"step": 25900 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.566329113924051e-06, |
|
"loss": 0.2261, |
|
"step": 25925 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.5600000000000002e-06, |
|
"loss": 0.2242, |
|
"step": 25950 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.5536708860759496e-06, |
|
"loss": 0.1974, |
|
"step": 25975 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.5473417721518987e-06, |
|
"loss": 0.1803, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.541012658227848e-06, |
|
"loss": 0.2106, |
|
"step": 26025 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.534683544303798e-06, |
|
"loss": 0.2308, |
|
"step": 26050 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.5283544303797474e-06, |
|
"loss": 0.2136, |
|
"step": 26075 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.5220253164556968e-06, |
|
"loss": 0.1795, |
|
"step": 26100 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.5156962025316462e-06, |
|
"loss": 0.1774, |
|
"step": 26125 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.5093670886075952e-06, |
|
"loss": 0.1997, |
|
"step": 26150 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.5030379746835446e-06, |
|
"loss": 0.1859, |
|
"step": 26175 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.496708860759494e-06, |
|
"loss": 0.1825, |
|
"step": 26200 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.4903797468354435e-06, |
|
"loss": 0.1667, |
|
"step": 26225 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.4840506329113925e-06, |
|
"loss": 0.1726, |
|
"step": 26250 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.477721518987342e-06, |
|
"loss": 0.2175, |
|
"step": 26275 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.4713924050632914e-06, |
|
"loss": 0.2412, |
|
"step": 26300 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.465063291139241e-06, |
|
"loss": 0.1868, |
|
"step": 26325 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.45873417721519e-06, |
|
"loss": 0.1802, |
|
"step": 26350 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.4524050632911392e-06, |
|
"loss": 0.1682, |
|
"step": 26375 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.4460759493670886e-06, |
|
"loss": 0.1924, |
|
"step": 26400 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.439746835443038e-06, |
|
"loss": 0.1875, |
|
"step": 26425 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.433417721518988e-06, |
|
"loss": 0.1958, |
|
"step": 26450 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.4270886075949374e-06, |
|
"loss": 0.2163, |
|
"step": 26475 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.4207594936708864e-06, |
|
"loss": 0.1971, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.4144303797468358e-06, |
|
"loss": 0.2124, |
|
"step": 26525 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.4081012658227852e-06, |
|
"loss": 0.169, |
|
"step": 26550 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.4017721518987346e-06, |
|
"loss": 0.1761, |
|
"step": 26575 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.395443037974684e-06, |
|
"loss": 0.2105, |
|
"step": 26600 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.389113924050633e-06, |
|
"loss": 0.1994, |
|
"step": 26625 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.3827848101265825e-06, |
|
"loss": 0.1751, |
|
"step": 26650 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.376455696202532e-06, |
|
"loss": 0.2071, |
|
"step": 26675 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.3701265822784814e-06, |
|
"loss": 0.2421, |
|
"step": 26700 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.3637974683544304e-06, |
|
"loss": 0.2675, |
|
"step": 26725 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.35746835443038e-06, |
|
"loss": 0.2626, |
|
"step": 26750 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.3511392405063292e-06, |
|
"loss": 0.2882, |
|
"step": 26775 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.3448101265822786e-06, |
|
"loss": 0.2752, |
|
"step": 26800 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.338481012658228e-06, |
|
"loss": 0.328, |
|
"step": 26825 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.332151898734177e-06, |
|
"loss": 0.2159, |
|
"step": 26850 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.325822784810127e-06, |
|
"loss": 0.2358, |
|
"step": 26875 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.3194936708860764e-06, |
|
"loss": 0.2266, |
|
"step": 26900 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.3131645569620258e-06, |
|
"loss": 0.2128, |
|
"step": 26925 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.306835443037975e-06, |
|
"loss": 0.1739, |
|
"step": 26950 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.3005063291139246e-06, |
|
"loss": 0.1805, |
|
"step": 26975 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.2941772151898736e-06, |
|
"loss": 0.2028, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.287848101265823e-06, |
|
"loss": 0.1851, |
|
"step": 27025 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.2815189873417725e-06, |
|
"loss": 0.2103, |
|
"step": 27050 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.275443037974684e-06, |
|
"loss": 0.1957, |
|
"step": 27075 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.269113924050633e-06, |
|
"loss": 0.2069, |
|
"step": 27100 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.2627848101265824e-06, |
|
"loss": 0.1912, |
|
"step": 27125 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.256455696202532e-06, |
|
"loss": 0.188, |
|
"step": 27150 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.2501265822784812e-06, |
|
"loss": 0.1652, |
|
"step": 27175 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.2437974683544302e-06, |
|
"loss": 0.1732, |
|
"step": 27200 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.2374683544303797e-06, |
|
"loss": 0.1648, |
|
"step": 27225 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.2311392405063295e-06, |
|
"loss": 0.1493, |
|
"step": 27250 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.224810126582279e-06, |
|
"loss": 0.1463, |
|
"step": 27275 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.2184810126582284e-06, |
|
"loss": 0.1367, |
|
"step": 27300 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.212151898734178e-06, |
|
"loss": 0.1313, |
|
"step": 27325 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.205822784810127e-06, |
|
"loss": 0.1287, |
|
"step": 27350 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.1994936708860762e-06, |
|
"loss": 0.1516, |
|
"step": 27375 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.1931645569620257e-06, |
|
"loss": 0.1502, |
|
"step": 27400 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.186835443037975e-06, |
|
"loss": 0.1446, |
|
"step": 27425 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.1805063291139245e-06, |
|
"loss": 0.1328, |
|
"step": 27450 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.1741772151898735e-06, |
|
"loss": 0.1311, |
|
"step": 27475 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.167848101265823e-06, |
|
"loss": 0.1152, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"eval_loss": 0.3180542290210724, |
|
"eval_runtime": 2148.7465, |
|
"eval_samples_per_second": 7.611, |
|
"eval_steps_per_second": 0.476, |
|
"eval_wer": 12.608681682357098, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.1615189873417724e-06, |
|
"loss": 0.1177, |
|
"step": 27525 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.155189873417722e-06, |
|
"loss": 0.1236, |
|
"step": 27550 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.148860759493671e-06, |
|
"loss": 0.1243, |
|
"step": 27575 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.1425316455696202e-06, |
|
"loss": 0.1278, |
|
"step": 27600 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.1362025316455697e-06, |
|
"loss": 0.131, |
|
"step": 27625 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.129873417721519e-06, |
|
"loss": 0.1218, |
|
"step": 27650 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.123544303797469e-06, |
|
"loss": 0.1147, |
|
"step": 27675 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.1172151898734184e-06, |
|
"loss": 0.11, |
|
"step": 27700 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.1108860759493674e-06, |
|
"loss": 0.1051, |
|
"step": 27725 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.104556962025317e-06, |
|
"loss": 0.0955, |
|
"step": 27750 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.0982278481012662e-06, |
|
"loss": 0.0966, |
|
"step": 27775 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.0918987341772157e-06, |
|
"loss": 0.0875, |
|
"step": 27800 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.0855696202531647e-06, |
|
"loss": 0.0872, |
|
"step": 27825 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.079240506329114e-06, |
|
"loss": 0.0829, |
|
"step": 27850 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.0729113924050635e-06, |
|
"loss": 0.0771, |
|
"step": 27875 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.066582278481013e-06, |
|
"loss": 0.0808, |
|
"step": 27900 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.0602531645569624e-06, |
|
"loss": 0.076, |
|
"step": 27925 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.0539240506329114e-06, |
|
"loss": 0.0679, |
|
"step": 27950 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.047594936708861e-06, |
|
"loss": 0.0681, |
|
"step": 27975 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.0412658227848102e-06, |
|
"loss": 0.0785, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.0349367088607597e-06, |
|
"loss": 0.1437, |
|
"step": 28025 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.0286075949367087e-06, |
|
"loss": 0.1527, |
|
"step": 28050 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.022278481012659e-06, |
|
"loss": 0.1472, |
|
"step": 28075 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.015949367088608e-06, |
|
"loss": 0.1538, |
|
"step": 28100 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.0096202531645574e-06, |
|
"loss": 0.143, |
|
"step": 28125 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.003291139240507e-06, |
|
"loss": 0.1364, |
|
"step": 28150 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.9969620253164562e-06, |
|
"loss": 0.1453, |
|
"step": 28175 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.9906329113924052e-06, |
|
"loss": 0.1497, |
|
"step": 28200 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.9843037974683547e-06, |
|
"loss": 0.151, |
|
"step": 28225 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.977974683544304e-06, |
|
"loss": 0.1444, |
|
"step": 28250 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.9716455696202535e-06, |
|
"loss": 0.1444, |
|
"step": 28275 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.965316455696203e-06, |
|
"loss": 0.1409, |
|
"step": 28300 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.958987341772152e-06, |
|
"loss": 0.1551, |
|
"step": 28325 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.9526582278481014e-06, |
|
"loss": 0.1493, |
|
"step": 28350 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.946329113924051e-06, |
|
"loss": 0.1504, |
|
"step": 28375 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.9400000000000002e-06, |
|
"loss": 0.1803, |
|
"step": 28400 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.9336708860759492e-06, |
|
"loss": 0.2503, |
|
"step": 28425 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.9273417721518987e-06, |
|
"loss": 0.2533, |
|
"step": 28450 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.921012658227848e-06, |
|
"loss": 0.2584, |
|
"step": 28475 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.914683544303798e-06, |
|
"loss": 0.2749, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.9083544303797474e-06, |
|
"loss": 0.2801, |
|
"step": 28525 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.902025316455697e-06, |
|
"loss": 0.2621, |
|
"step": 28550 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.895696202531646e-06, |
|
"loss": 0.2584, |
|
"step": 28575 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.8893670886075952e-06, |
|
"loss": 0.2624, |
|
"step": 28600 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.8830379746835447e-06, |
|
"loss": 0.2688, |
|
"step": 28625 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.876708860759494e-06, |
|
"loss": 0.2735, |
|
"step": 28650 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.870379746835443e-06, |
|
"loss": 0.2642, |
|
"step": 28675 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.8640506329113925e-06, |
|
"loss": 0.2634, |
|
"step": 28700 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.857721518987342e-06, |
|
"loss": 0.2548, |
|
"step": 28725 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.8513924050632914e-06, |
|
"loss": 0.267, |
|
"step": 28750 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.845063291139241e-06, |
|
"loss": 0.2306, |
|
"step": 28775 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.83873417721519e-06, |
|
"loss": 0.2715, |
|
"step": 28800 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.8324050632911392e-06, |
|
"loss": 0.2587, |
|
"step": 28825 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.8260759493670887e-06, |
|
"loss": 0.2408, |
|
"step": 28850 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.819746835443038e-06, |
|
"loss": 0.2767, |
|
"step": 28875 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.813417721518988e-06, |
|
"loss": 0.2661, |
|
"step": 28900 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.8070886075949374e-06, |
|
"loss": 0.2586, |
|
"step": 28925 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.8007594936708864e-06, |
|
"loss": 0.2637, |
|
"step": 28950 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.794430379746836e-06, |
|
"loss": 0.2602, |
|
"step": 28975 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.7881012658227852e-06, |
|
"loss": 0.251, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.7817721518987346e-06, |
|
"loss": 0.2669, |
|
"step": 29025 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.7754430379746837e-06, |
|
"loss": 0.2533, |
|
"step": 29050 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.769113924050633e-06, |
|
"loss": 0.2138, |
|
"step": 29075 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.7627848101265825e-06, |
|
"loss": 0.2298, |
|
"step": 29100 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.756455696202532e-06, |
|
"loss": 0.2344, |
|
"step": 29125 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.7501265822784814e-06, |
|
"loss": 0.2142, |
|
"step": 29150 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.7437974683544304e-06, |
|
"loss": 0.2122, |
|
"step": 29175 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.73746835443038e-06, |
|
"loss": 0.22, |
|
"step": 29200 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.7311392405063292e-06, |
|
"loss": 0.2086, |
|
"step": 29225 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.7248101265822787e-06, |
|
"loss": 0.2316, |
|
"step": 29250 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.7184810126582277e-06, |
|
"loss": 0.2116, |
|
"step": 29275 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.712151898734177e-06, |
|
"loss": 0.2285, |
|
"step": 29300 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.705822784810127e-06, |
|
"loss": 0.2273, |
|
"step": 29325 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.6994936708860764e-06, |
|
"loss": 0.2369, |
|
"step": 29350 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.693164556962026e-06, |
|
"loss": 0.2101, |
|
"step": 29375 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 2.6868354430379752e-06, |
|
"loss": 0.2056, |
|
"step": 29400 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 2.6805063291139242e-06, |
|
"loss": 0.2298, |
|
"step": 29425 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 2.6744303797468357e-06, |
|
"loss": 0.2267, |
|
"step": 29450 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 2.668101265822785e-06, |
|
"loss": 0.222, |
|
"step": 29475 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 2.6617721518987345e-06, |
|
"loss": 0.2194, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 2.6554430379746835e-06, |
|
"loss": 0.2165, |
|
"step": 29525 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 2.649113924050633e-06, |
|
"loss": 0.2238, |
|
"step": 29550 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 2.6427848101265824e-06, |
|
"loss": 0.2139, |
|
"step": 29575 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 2.636455696202532e-06, |
|
"loss": 0.2318, |
|
"step": 29600 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 2.6301265822784812e-06, |
|
"loss": 0.2024, |
|
"step": 29625 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.6237974683544302e-06, |
|
"loss": 0.1976, |
|
"step": 29650 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.6174683544303797e-06, |
|
"loss": 0.1163, |
|
"step": 29675 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.611139240506329e-06, |
|
"loss": 0.1178, |
|
"step": 29700 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.604810126582279e-06, |
|
"loss": 0.1242, |
|
"step": 29725 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.5984810126582284e-06, |
|
"loss": 0.1142, |
|
"step": 29750 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.592151898734178e-06, |
|
"loss": 0.1088, |
|
"step": 29775 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.585822784810127e-06, |
|
"loss": 0.0997, |
|
"step": 29800 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.5794936708860762e-06, |
|
"loss": 0.0947, |
|
"step": 29825 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.5731645569620257e-06, |
|
"loss": 0.0991, |
|
"step": 29850 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.566835443037975e-06, |
|
"loss": 0.1105, |
|
"step": 29875 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.560506329113924e-06, |
|
"loss": 0.1121, |
|
"step": 29900 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.5541772151898735e-06, |
|
"loss": 0.1074, |
|
"step": 29925 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.547848101265823e-06, |
|
"loss": 0.0999, |
|
"step": 29950 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.5415189873417724e-06, |
|
"loss": 0.0994, |
|
"step": 29975 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.535189873417722e-06, |
|
"loss": 0.0901, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"eval_loss": 0.31337109208106995, |
|
"eval_runtime": 1548.9005, |
|
"eval_samples_per_second": 10.558, |
|
"eval_steps_per_second": 0.66, |
|
"eval_wer": 12.324004755325642, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.528860759493671e-06, |
|
"loss": 0.0925, |
|
"step": 30025 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.5225316455696202e-06, |
|
"loss": 0.09, |
|
"step": 30050 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.5162025316455697e-06, |
|
"loss": 0.0965, |
|
"step": 30075 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.509873417721519e-06, |
|
"loss": 0.0926, |
|
"step": 30100 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.503544303797469e-06, |
|
"loss": 0.0974, |
|
"step": 30125 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.497215189873418e-06, |
|
"loss": 0.102, |
|
"step": 30150 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.4908860759493674e-06, |
|
"loss": 0.0833, |
|
"step": 30175 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.4845569620253164e-06, |
|
"loss": 0.0784, |
|
"step": 30200 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.4782278481012662e-06, |
|
"loss": 0.0744, |
|
"step": 30225 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.4718987341772157e-06, |
|
"loss": 0.0754, |
|
"step": 30250 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.4655696202531647e-06, |
|
"loss": 0.0685, |
|
"step": 30275 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.459240506329114e-06, |
|
"loss": 0.0699, |
|
"step": 30300 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.4529113924050635e-06, |
|
"loss": 0.0617, |
|
"step": 30325 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.446582278481013e-06, |
|
"loss": 0.0603, |
|
"step": 30350 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.440253164556962e-06, |
|
"loss": 0.053, |
|
"step": 30375 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.4339240506329114e-06, |
|
"loss": 0.0623, |
|
"step": 30400 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.4275949367088612e-06, |
|
"loss": 0.054, |
|
"step": 30425 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.4212658227848102e-06, |
|
"loss": 0.0504, |
|
"step": 30450 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.4149367088607597e-06, |
|
"loss": 0.0503, |
|
"step": 30475 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.408607594936709e-06, |
|
"loss": 0.0507, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.4022784810126585e-06, |
|
"loss": 0.1057, |
|
"step": 30525 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.395949367088608e-06, |
|
"loss": 0.1041, |
|
"step": 30550 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.389620253164557e-06, |
|
"loss": 0.1031, |
|
"step": 30575 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.3832911392405064e-06, |
|
"loss": 0.111, |
|
"step": 30600 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.376962025316456e-06, |
|
"loss": 0.0969, |
|
"step": 30625 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.3706329113924052e-06, |
|
"loss": 0.1011, |
|
"step": 30650 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.3643037974683547e-06, |
|
"loss": 0.1093, |
|
"step": 30675 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.357974683544304e-06, |
|
"loss": 0.103, |
|
"step": 30700 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.3516455696202535e-06, |
|
"loss": 0.1173, |
|
"step": 30725 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.3453164556962025e-06, |
|
"loss": 0.1062, |
|
"step": 30750 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.338987341772152e-06, |
|
"loss": 0.1005, |
|
"step": 30775 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.3326582278481014e-06, |
|
"loss": 0.0985, |
|
"step": 30800 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.326329113924051e-06, |
|
"loss": 0.1086, |
|
"step": 30825 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.3200000000000002e-06, |
|
"loss": 0.1168, |
|
"step": 30850 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.3136708860759497e-06, |
|
"loss": 0.1019, |
|
"step": 30875 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.307341772151899e-06, |
|
"loss": 0.138, |
|
"step": 30900 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.3010126582278485e-06, |
|
"loss": 0.1795, |
|
"step": 30925 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.2946835443037975e-06, |
|
"loss": 0.1936, |
|
"step": 30950 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.288354430379747e-06, |
|
"loss": 0.1891, |
|
"step": 30975 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.2820253164556964e-06, |
|
"loss": 0.2029, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.275696202531646e-06, |
|
"loss": 0.2021, |
|
"step": 31025 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.2693670886075952e-06, |
|
"loss": 0.2064, |
|
"step": 31050 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.2630379746835447e-06, |
|
"loss": 0.2094, |
|
"step": 31075 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.256708860759494e-06, |
|
"loss": 0.1949, |
|
"step": 31100 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.250379746835443e-06, |
|
"loss": 0.1945, |
|
"step": 31125 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.2440506329113925e-06, |
|
"loss": 0.2024, |
|
"step": 31150 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.237721518987342e-06, |
|
"loss": 0.1952, |
|
"step": 31175 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.2313924050632914e-06, |
|
"loss": 0.1972, |
|
"step": 31200 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.2250632911392404e-06, |
|
"loss": 0.1927, |
|
"step": 31225 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.2187341772151902e-06, |
|
"loss": 0.197, |
|
"step": 31250 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.2124050632911397e-06, |
|
"loss": 0.1896, |
|
"step": 31275 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.2060759493670887e-06, |
|
"loss": 0.2026, |
|
"step": 31300 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.199746835443038e-06, |
|
"loss": 0.1929, |
|
"step": 31325 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.1934177215189875e-06, |
|
"loss": 0.1839, |
|
"step": 31350 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.187088607594937e-06, |
|
"loss": 0.1995, |
|
"step": 31375 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.1807594936708864e-06, |
|
"loss": 0.2043, |
|
"step": 31400 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.1744303797468354e-06, |
|
"loss": 0.2078, |
|
"step": 31425 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.168101265822785e-06, |
|
"loss": 0.2002, |
|
"step": 31450 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.1617721518987347e-06, |
|
"loss": 0.2004, |
|
"step": 31475 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.1556962025316457e-06, |
|
"loss": 0.1958, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.149367088607595e-06, |
|
"loss": 0.191, |
|
"step": 31525 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.1430379746835445e-06, |
|
"loss": 0.1838, |
|
"step": 31550 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.136708860759494e-06, |
|
"loss": 0.1753, |
|
"step": 31575 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.130379746835443e-06, |
|
"loss": 0.1735, |
|
"step": 31600 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.1240506329113924e-06, |
|
"loss": 0.1781, |
|
"step": 31625 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.1177215189873422e-06, |
|
"loss": 0.1658, |
|
"step": 31650 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.1113924050632913e-06, |
|
"loss": 0.1755, |
|
"step": 31675 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.1050632911392407e-06, |
|
"loss": 0.1673, |
|
"step": 31700 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.09873417721519e-06, |
|
"loss": 0.168, |
|
"step": 31725 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.0924050632911395e-06, |
|
"loss": 0.1696, |
|
"step": 31750 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.0860759493670885e-06, |
|
"loss": 0.1582, |
|
"step": 31775 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.079746835443038e-06, |
|
"loss": 0.1815, |
|
"step": 31800 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.0734177215189874e-06, |
|
"loss": 0.1802, |
|
"step": 31825 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 2.067088607594937e-06, |
|
"loss": 0.1752, |
|
"step": 31850 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 2.0607594936708863e-06, |
|
"loss": 0.1692, |
|
"step": 31875 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 2.0544303797468357e-06, |
|
"loss": 0.1591, |
|
"step": 31900 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 2.048101265822785e-06, |
|
"loss": 0.1799, |
|
"step": 31925 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 2.0417721518987345e-06, |
|
"loss": 0.1682, |
|
"step": 31950 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 2.0354430379746835e-06, |
|
"loss": 0.17, |
|
"step": 31975 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 2.029113924050633e-06, |
|
"loss": 0.1664, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 2.0227848101265824e-06, |
|
"loss": 0.1636, |
|
"step": 32025 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 2.016455696202532e-06, |
|
"loss": 0.1671, |
|
"step": 32050 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 2.0101265822784812e-06, |
|
"loss": 0.175, |
|
"step": 32075 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 2.0037974683544307e-06, |
|
"loss": 0.1744, |
|
"step": 32100 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.99746835443038e-06, |
|
"loss": 0.1606, |
|
"step": 32125 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.991139240506329e-06, |
|
"loss": 0.1518, |
|
"step": 32150 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.9848101265822785e-06, |
|
"loss": 0.1636, |
|
"step": 32175 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.978481012658228e-06, |
|
"loss": 0.1623, |
|
"step": 32200 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.9721518987341774e-06, |
|
"loss": 0.176, |
|
"step": 32225 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.965822784810127e-06, |
|
"loss": 0.1617, |
|
"step": 32250 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.9594936708860762e-06, |
|
"loss": 0.1604, |
|
"step": 32275 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.9531645569620257e-06, |
|
"loss": 0.1567, |
|
"step": 32300 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.946835443037975e-06, |
|
"loss": 0.1775, |
|
"step": 32325 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.940506329113924e-06, |
|
"loss": 0.1676, |
|
"step": 32350 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.9341772151898735e-06, |
|
"loss": 0.1675, |
|
"step": 32375 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.927848101265823e-06, |
|
"loss": 0.1486, |
|
"step": 32400 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.9215189873417724e-06, |
|
"loss": 0.1506, |
|
"step": 32425 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.9151898734177214e-06, |
|
"loss": 0.1697, |
|
"step": 32450 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.9088607594936712e-06, |
|
"loss": 0.1591, |
|
"step": 32475 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.9025316455696205e-06, |
|
"loss": 0.1595, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"eval_loss": 0.31068873405456543, |
|
"eval_runtime": 1554.1986, |
|
"eval_samples_per_second": 10.522, |
|
"eval_steps_per_second": 0.658, |
|
"eval_wer": 12.021334704238024, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.8962025316455699e-06, |
|
"loss": 0.1359, |
|
"step": 32525 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.889873417721519e-06, |
|
"loss": 0.1466, |
|
"step": 32550 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.8835443037974685e-06, |
|
"loss": 0.1388, |
|
"step": 32575 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.8772151898734177e-06, |
|
"loss": 0.1488, |
|
"step": 32600 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.8708860759493672e-06, |
|
"loss": 0.1568, |
|
"step": 32625 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.8645569620253164e-06, |
|
"loss": 0.1527, |
|
"step": 32650 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.858227848101266e-06, |
|
"loss": 0.1332, |
|
"step": 32675 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.8518987341772155e-06, |
|
"loss": 0.1423, |
|
"step": 32700 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.8455696202531647e-06, |
|
"loss": 0.1367, |
|
"step": 32725 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.839240506329114e-06, |
|
"loss": 0.1425, |
|
"step": 32750 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.8329113924050635e-06, |
|
"loss": 0.1429, |
|
"step": 32775 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.8265822784810127e-06, |
|
"loss": 0.1473, |
|
"step": 32800 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.8202531645569622e-06, |
|
"loss": 0.1524, |
|
"step": 32825 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.8139240506329114e-06, |
|
"loss": 0.1418, |
|
"step": 32850 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.8075949367088608e-06, |
|
"loss": 0.141, |
|
"step": 32875 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.8012658227848105e-06, |
|
"loss": 0.1445, |
|
"step": 32900 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.7949367088607597e-06, |
|
"loss": 0.156, |
|
"step": 32925 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.788607594936709e-06, |
|
"loss": 0.1569, |
|
"step": 32950 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.7822784810126583e-06, |
|
"loss": 0.1435, |
|
"step": 32975 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.7759493670886077e-06, |
|
"loss": 0.1669, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.769620253164557e-06, |
|
"loss": 0.1845, |
|
"step": 33025 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.7632911392405064e-06, |
|
"loss": 0.1805, |
|
"step": 33050 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.7569620253164558e-06, |
|
"loss": 0.1927, |
|
"step": 33075 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.7506329113924052e-06, |
|
"loss": 0.2009, |
|
"step": 33100 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.7443037974683547e-06, |
|
"loss": 0.1771, |
|
"step": 33125 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.7379746835443039e-06, |
|
"loss": 0.1882, |
|
"step": 33150 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.7316455696202533e-06, |
|
"loss": 0.1823, |
|
"step": 33175 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.7253164556962027e-06, |
|
"loss": 0.1779, |
|
"step": 33200 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.718987341772152e-06, |
|
"loss": 0.1825, |
|
"step": 33225 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.7126582278481014e-06, |
|
"loss": 0.18, |
|
"step": 33250 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.7063291139240506e-06, |
|
"loss": 0.1852, |
|
"step": 33275 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.7000000000000002e-06, |
|
"loss": 0.1737, |
|
"step": 33300 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.6936708860759497e-06, |
|
"loss": 0.1772, |
|
"step": 33325 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.6873417721518989e-06, |
|
"loss": 0.1788, |
|
"step": 33350 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.6810126582278483e-06, |
|
"loss": 0.1734, |
|
"step": 33375 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.6746835443037975e-06, |
|
"loss": 0.1544, |
|
"step": 33400 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.668354430379747e-06, |
|
"loss": 0.1398, |
|
"step": 33425 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.6620253164556962e-06, |
|
"loss": 0.1547, |
|
"step": 33450 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.6556962025316456e-06, |
|
"loss": 0.1763, |
|
"step": 33475 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.6493670886075952e-06, |
|
"loss": 0.1857, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.6430379746835445e-06, |
|
"loss": 0.1407, |
|
"step": 33525 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.636962025316456e-06, |
|
"loss": 0.1237, |
|
"step": 33550 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.6306329113924051e-06, |
|
"loss": 0.1212, |
|
"step": 33575 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.6243037974683545e-06, |
|
"loss": 0.19, |
|
"step": 33600 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.6179746835443038e-06, |
|
"loss": 0.2653, |
|
"step": 33625 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.6116455696202532e-06, |
|
"loss": 0.1257, |
|
"step": 33650 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.6053164556962026e-06, |
|
"loss": 0.1001, |
|
"step": 33675 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.598987341772152e-06, |
|
"loss": 0.117, |
|
"step": 33700 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.5926582278481015e-06, |
|
"loss": 0.138, |
|
"step": 33725 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.586329113924051e-06, |
|
"loss": 0.1708, |
|
"step": 33750 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.5800000000000001e-06, |
|
"loss": 0.216, |
|
"step": 33775 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.5736708860759495e-06, |
|
"loss": 0.1605, |
|
"step": 33800 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.5673417721518988e-06, |
|
"loss": 0.1447, |
|
"step": 33825 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.5610126582278482e-06, |
|
"loss": 0.1783, |
|
"step": 33850 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.5546835443037974e-06, |
|
"loss": 0.153, |
|
"step": 33875 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.548354430379747e-06, |
|
"loss": 0.139, |
|
"step": 33900 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.5420253164556965e-06, |
|
"loss": 0.163, |
|
"step": 33925 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.5356962025316457e-06, |
|
"loss": 0.1552, |
|
"step": 33950 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.5293670886075951e-06, |
|
"loss": 0.1278, |
|
"step": 33975 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.5230379746835443e-06, |
|
"loss": 0.1313, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.5167088607594938e-06, |
|
"loss": 0.1353, |
|
"step": 34025 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.5103797468354432e-06, |
|
"loss": 0.1055, |
|
"step": 34050 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.5040506329113924e-06, |
|
"loss": 0.1047, |
|
"step": 34075 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.497721518987342e-06, |
|
"loss": 0.1314, |
|
"step": 34100 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.4913924050632913e-06, |
|
"loss": 0.1246, |
|
"step": 34125 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.4850632911392407e-06, |
|
"loss": 0.1351, |
|
"step": 34150 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.4787341772151901e-06, |
|
"loss": 0.1268, |
|
"step": 34175 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.4724050632911393e-06, |
|
"loss": 0.1601, |
|
"step": 34200 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.4660759493670888e-06, |
|
"loss": 0.1769, |
|
"step": 34225 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.459746835443038e-06, |
|
"loss": 0.1573, |
|
"step": 34250 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.4534177215189874e-06, |
|
"loss": 0.2516, |
|
"step": 34275 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.4470886075949366e-06, |
|
"loss": 0.1174, |
|
"step": 34300 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.4407594936708863e-06, |
|
"loss": 0.1142, |
|
"step": 34325 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.4344303797468357e-06, |
|
"loss": 0.1244, |
|
"step": 34350 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.428101265822785e-06, |
|
"loss": 0.1121, |
|
"step": 34375 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.4217721518987343e-06, |
|
"loss": 0.1076, |
|
"step": 34400 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.4154430379746835e-06, |
|
"loss": 0.128, |
|
"step": 34425 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.409113924050633e-06, |
|
"loss": 0.1066, |
|
"step": 34450 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.4027848101265824e-06, |
|
"loss": 0.113, |
|
"step": 34475 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.3964556962025316e-06, |
|
"loss": 0.1314, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.3901265822784813e-06, |
|
"loss": 0.1461, |
|
"step": 34525 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.3837974683544305e-06, |
|
"loss": 0.1143, |
|
"step": 34550 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.37746835443038e-06, |
|
"loss": 0.1003, |
|
"step": 34575 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.3711392405063293e-06, |
|
"loss": 0.0972, |
|
"step": 34600 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.3648101265822785e-06, |
|
"loss": 0.1014, |
|
"step": 34625 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.358481012658228e-06, |
|
"loss": 0.1151, |
|
"step": 34650 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.3521518987341772e-06, |
|
"loss": 0.1655, |
|
"step": 34675 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.3458227848101266e-06, |
|
"loss": 0.1595, |
|
"step": 34700 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.3394936708860763e-06, |
|
"loss": 0.1445, |
|
"step": 34725 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.3331645569620255e-06, |
|
"loss": 0.1263, |
|
"step": 34750 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.326835443037975e-06, |
|
"loss": 0.1575, |
|
"step": 34775 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.3205063291139241e-06, |
|
"loss": 0.1678, |
|
"step": 34800 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.3141772151898735e-06, |
|
"loss": 0.1638, |
|
"step": 34825 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.3078481012658228e-06, |
|
"loss": 0.1369, |
|
"step": 34850 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.3015189873417722e-06, |
|
"loss": 0.1263, |
|
"step": 34875 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.2951898734177216e-06, |
|
"loss": 0.1483, |
|
"step": 34900 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.288860759493671e-06, |
|
"loss": 0.1304, |
|
"step": 34925 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.2825316455696205e-06, |
|
"loss": 0.1245, |
|
"step": 34950 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.2762025316455697e-06, |
|
"loss": 0.1244, |
|
"step": 34975 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.2698734177215191e-06, |
|
"loss": 0.1249, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"eval_loss": 0.3131045699119568, |
|
"eval_runtime": 1560.2837, |
|
"eval_samples_per_second": 10.481, |
|
"eval_steps_per_second": 0.656, |
|
"eval_wer": 12.086881084728336, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.2635443037974685e-06, |
|
"loss": 0.1714, |
|
"step": 35025 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.2572151898734178e-06, |
|
"loss": 0.1644, |
|
"step": 35050 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.2508860759493672e-06, |
|
"loss": 0.1527, |
|
"step": 35075 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.2445569620253166e-06, |
|
"loss": 0.1179, |
|
"step": 35100 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.2382278481012658e-06, |
|
"loss": 0.1221, |
|
"step": 35125 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.2318987341772153e-06, |
|
"loss": 0.1469, |
|
"step": 35150 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.2255696202531647e-06, |
|
"loss": 0.1441, |
|
"step": 35175 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.2192405063291141e-06, |
|
"loss": 0.1371, |
|
"step": 35200 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.2129113924050633e-06, |
|
"loss": 0.1726, |
|
"step": 35225 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.2065822784810128e-06, |
|
"loss": 0.1424, |
|
"step": 35250 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.200253164556962e-06, |
|
"loss": 0.1511, |
|
"step": 35275 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.1941772151898736e-06, |
|
"loss": 0.1476, |
|
"step": 35300 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.1878481012658228e-06, |
|
"loss": 0.1304, |
|
"step": 35325 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.1815189873417723e-06, |
|
"loss": 0.1661, |
|
"step": 35350 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.1751898734177217e-06, |
|
"loss": 0.1358, |
|
"step": 35375 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.168860759493671e-06, |
|
"loss": 0.134, |
|
"step": 35400 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.1625316455696203e-06, |
|
"loss": 0.1629, |
|
"step": 35425 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.1562025316455698e-06, |
|
"loss": 0.1624, |
|
"step": 35450 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.1498734177215192e-06, |
|
"loss": 0.2122, |
|
"step": 35475 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.1435443037974684e-06, |
|
"loss": 0.2037, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.1372151898734178e-06, |
|
"loss": 0.2228, |
|
"step": 35525 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.130886075949367e-06, |
|
"loss": 0.2026, |
|
"step": 35550 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.1245569620253167e-06, |
|
"loss": 0.2468, |
|
"step": 35575 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.118227848101266e-06, |
|
"loss": 0.1536, |
|
"step": 35600 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.1118987341772153e-06, |
|
"loss": 0.1815, |
|
"step": 35625 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.1055696202531646e-06, |
|
"loss": 0.1585, |
|
"step": 35650 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.099240506329114e-06, |
|
"loss": 0.165, |
|
"step": 35675 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.0929113924050634e-06, |
|
"loss": 0.1424, |
|
"step": 35700 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.0865822784810128e-06, |
|
"loss": 0.1299, |
|
"step": 35725 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.080253164556962e-06, |
|
"loss": 0.1581, |
|
"step": 35750 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.0739240506329115e-06, |
|
"loss": 0.1365, |
|
"step": 35775 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.067594936708861e-06, |
|
"loss": 0.1553, |
|
"step": 35800 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.0612658227848101e-06, |
|
"loss": 0.1383, |
|
"step": 35825 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.0549367088607596e-06, |
|
"loss": 0.1491, |
|
"step": 35850 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.048607594936709e-06, |
|
"loss": 0.1488, |
|
"step": 35875 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.0422784810126584e-06, |
|
"loss": 0.1449, |
|
"step": 35900 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.0359493670886076e-06, |
|
"loss": 0.1114, |
|
"step": 35925 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.029620253164557e-06, |
|
"loss": 0.104, |
|
"step": 35950 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.0232911392405063e-06, |
|
"loss": 0.0957, |
|
"step": 35975 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.016962025316456e-06, |
|
"loss": 0.0935, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.0106329113924051e-06, |
|
"loss": 0.0985, |
|
"step": 36025 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.0043037974683546e-06, |
|
"loss": 0.0959, |
|
"step": 36050 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.979746835443038e-07, |
|
"loss": 0.096, |
|
"step": 36075 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.916455696202532e-07, |
|
"loss": 0.0927, |
|
"step": 36100 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.853164556962026e-07, |
|
"loss": 0.0903, |
|
"step": 36125 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.78987341772152e-07, |
|
"loss": 0.0935, |
|
"step": 36150 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.726582278481013e-07, |
|
"loss": 0.0936, |
|
"step": 36175 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.663291139240507e-07, |
|
"loss": 0.097, |
|
"step": 36200 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.600000000000001e-07, |
|
"loss": 0.1039, |
|
"step": 36225 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.536708860759494e-07, |
|
"loss": 0.1032, |
|
"step": 36250 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.473417721518988e-07, |
|
"loss": 0.1014, |
|
"step": 36275 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.410126582278482e-07, |
|
"loss": 0.0888, |
|
"step": 36300 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.346835443037975e-07, |
|
"loss": 0.0972, |
|
"step": 36325 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.283544303797469e-07, |
|
"loss": 0.0952, |
|
"step": 36350 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.220253164556963e-07, |
|
"loss": 0.0856, |
|
"step": 36375 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.156962025316457e-07, |
|
"loss": 0.0984, |
|
"step": 36400 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.09367088607595e-07, |
|
"loss": 0.1005, |
|
"step": 36425 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.030379746835443e-07, |
|
"loss": 0.0964, |
|
"step": 36450 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.967088607594937e-07, |
|
"loss": 0.1046, |
|
"step": 36475 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.903797468354432e-07, |
|
"loss": 0.1037, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.840506329113925e-07, |
|
"loss": 0.0973, |
|
"step": 36525 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.777215189873418e-07, |
|
"loss": 0.1061, |
|
"step": 36550 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.713924050632912e-07, |
|
"loss": 0.1201, |
|
"step": 36575 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.650632911392407e-07, |
|
"loss": 0.1196, |
|
"step": 36600 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.5873417721519e-07, |
|
"loss": 0.1046, |
|
"step": 36625 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.524050632911393e-07, |
|
"loss": 0.0965, |
|
"step": 36650 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.460759493670887e-07, |
|
"loss": 0.1086, |
|
"step": 36675 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.39746835443038e-07, |
|
"loss": 0.0846, |
|
"step": 36700 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.334177215189874e-07, |
|
"loss": 0.0876, |
|
"step": 36725 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.270886075949367e-07, |
|
"loss": 0.0872, |
|
"step": 36750 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.207594936708862e-07, |
|
"loss": 0.0804, |
|
"step": 36775 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.144303797468355e-07, |
|
"loss": 0.1063, |
|
"step": 36800 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.081012658227849e-07, |
|
"loss": 0.1396, |
|
"step": 36825 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.017721518987342e-07, |
|
"loss": 0.1649, |
|
"step": 36850 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 7.954430379746836e-07, |
|
"loss": 0.1192, |
|
"step": 36875 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 7.891139240506329e-07, |
|
"loss": 0.1374, |
|
"step": 36900 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 7.827848101265824e-07, |
|
"loss": 0.1314, |
|
"step": 36925 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 7.764556962025317e-07, |
|
"loss": 0.1348, |
|
"step": 36950 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 7.70126582278481e-07, |
|
"loss": 0.1203, |
|
"step": 36975 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 7.637974683544304e-07, |
|
"loss": 0.1132, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 7.574683544303799e-07, |
|
"loss": 0.1667, |
|
"step": 37025 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 7.511392405063292e-07, |
|
"loss": 0.1485, |
|
"step": 37050 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 7.448101265822785e-07, |
|
"loss": 0.1569, |
|
"step": 37075 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 7.384810126582279e-07, |
|
"loss": 0.1577, |
|
"step": 37100 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 7.321518987341773e-07, |
|
"loss": 0.1666, |
|
"step": 37125 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 7.258227848101266e-07, |
|
"loss": 0.1511, |
|
"step": 37150 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 7.194936708860759e-07, |
|
"loss": 0.1388, |
|
"step": 37175 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 7.131645569620254e-07, |
|
"loss": 0.1284, |
|
"step": 37200 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 7.068354430379748e-07, |
|
"loss": 0.1492, |
|
"step": 37225 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 7.005063291139241e-07, |
|
"loss": 0.1512, |
|
"step": 37250 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 6.941772151898734e-07, |
|
"loss": 0.144, |
|
"step": 37275 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 6.878481012658228e-07, |
|
"loss": 0.129, |
|
"step": 37300 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 6.815189873417723e-07, |
|
"loss": 0.133, |
|
"step": 37325 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 6.751898734177216e-07, |
|
"loss": 0.1462, |
|
"step": 37350 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 6.688607594936709e-07, |
|
"loss": 0.1516, |
|
"step": 37375 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 6.625316455696203e-07, |
|
"loss": 0.1253, |
|
"step": 37400 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 6.562025316455697e-07, |
|
"loss": 0.1131, |
|
"step": 37425 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.498734177215191e-07, |
|
"loss": 0.1259, |
|
"step": 37450 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.435443037974684e-07, |
|
"loss": 0.1092, |
|
"step": 37475 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.372151898734178e-07, |
|
"loss": 0.1404, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"eval_loss": 0.31174880266189575, |
|
"eval_runtime": 1735.2822, |
|
"eval_samples_per_second": 9.424, |
|
"eval_steps_per_second": 0.59, |
|
"eval_wer": 12.463451466760917, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.308860759493671e-07, |
|
"loss": 0.1608, |
|
"step": 37525 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.245569620253165e-07, |
|
"loss": 0.1343, |
|
"step": 37550 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.182278481012658e-07, |
|
"loss": 0.1202, |
|
"step": 37575 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.118987341772153e-07, |
|
"loss": 0.1506, |
|
"step": 37600 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.055696202531646e-07, |
|
"loss": 0.1427, |
|
"step": 37625 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 5.992405063291139e-07, |
|
"loss": 0.1108, |
|
"step": 37650 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 5.929113924050633e-07, |
|
"loss": 0.1225, |
|
"step": 37675 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 5.865822784810127e-07, |
|
"loss": 0.115, |
|
"step": 37700 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 5.802531645569621e-07, |
|
"loss": 0.1452, |
|
"step": 37725 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 5.739240506329114e-07, |
|
"loss": 0.1421, |
|
"step": 37750 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 5.675949367088608e-07, |
|
"loss": 0.1196, |
|
"step": 37775 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 5.612658227848102e-07, |
|
"loss": 0.1063, |
|
"step": 37800 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 5.549367088607596e-07, |
|
"loss": 0.1419, |
|
"step": 37825 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 5.486075949367089e-07, |
|
"loss": 0.1116, |
|
"step": 37850 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 5.422784810126583e-07, |
|
"loss": 0.1343, |
|
"step": 37875 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 5.359493670886077e-07, |
|
"loss": 0.1362, |
|
"step": 37900 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 5.29620253164557e-07, |
|
"loss": 0.1233, |
|
"step": 37925 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 5.232911392405064e-07, |
|
"loss": 0.1181, |
|
"step": 37950 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 5.169620253164557e-07, |
|
"loss": 0.1173, |
|
"step": 37975 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 5.10632911392405e-07, |
|
"loss": 0.1099, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 5.043037974683545e-07, |
|
"loss": 0.1315, |
|
"step": 38025 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 4.979746835443038e-07, |
|
"loss": 0.1373, |
|
"step": 38050 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 4.916455696202532e-07, |
|
"loss": 0.1132, |
|
"step": 38075 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 4.853164556962025e-07, |
|
"loss": 0.0815, |
|
"step": 38100 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 4.78987341772152e-07, |
|
"loss": 0.117, |
|
"step": 38125 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 4.726582278481013e-07, |
|
"loss": 0.109, |
|
"step": 38150 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 4.663291139240507e-07, |
|
"loss": 0.1403, |
|
"step": 38175 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 4.6000000000000004e-07, |
|
"loss": 0.1778, |
|
"step": 38200 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 4.536708860759494e-07, |
|
"loss": 0.1603, |
|
"step": 38225 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 4.4734177215189874e-07, |
|
"loss": 0.1205, |
|
"step": 38250 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 4.4101265822784817e-07, |
|
"loss": 0.119, |
|
"step": 38275 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 4.346835443037975e-07, |
|
"loss": 0.1159, |
|
"step": 38300 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 4.283544303797469e-07, |
|
"loss": 0.1041, |
|
"step": 38325 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 4.2202531645569624e-07, |
|
"loss": 0.09, |
|
"step": 38350 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 4.156962025316456e-07, |
|
"loss": 0.1037, |
|
"step": 38375 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 4.09367088607595e-07, |
|
"loss": 0.1743, |
|
"step": 38400 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 4.030379746835443e-07, |
|
"loss": 0.1889, |
|
"step": 38425 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 3.967088607594937e-07, |
|
"loss": 0.3427, |
|
"step": 38450 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 3.9037974683544306e-07, |
|
"loss": 0.5569, |
|
"step": 38475 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 3.8405063291139243e-07, |
|
"loss": 0.5484, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 3.7772151898734176e-07, |
|
"loss": 0.1786, |
|
"step": 38525 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 3.713924050632912e-07, |
|
"loss": 0.0673, |
|
"step": 38550 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 3.650632911392405e-07, |
|
"loss": 0.0928, |
|
"step": 38575 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 3.5873417721518993e-07, |
|
"loss": 0.1669, |
|
"step": 38600 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 3.5240506329113925e-07, |
|
"loss": 0.1816, |
|
"step": 38625 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 3.4607594936708863e-07, |
|
"loss": 0.1014, |
|
"step": 38650 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 3.39746835443038e-07, |
|
"loss": 0.0872, |
|
"step": 38675 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 3.334177215189874e-07, |
|
"loss": 0.126, |
|
"step": 38700 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 3.270886075949367e-07, |
|
"loss": 0.1821, |
|
"step": 38725 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 3.2075949367088613e-07, |
|
"loss": 0.2322, |
|
"step": 38750 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 3.1443037974683545e-07, |
|
"loss": 0.1604, |
|
"step": 38775 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 3.081012658227848e-07, |
|
"loss": 0.1144, |
|
"step": 38800 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 3.017721518987342e-07, |
|
"loss": 0.1205, |
|
"step": 38825 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 2.954430379746836e-07, |
|
"loss": 0.1558, |
|
"step": 38850 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 2.891139240506329e-07, |
|
"loss": 0.1561, |
|
"step": 38875 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 2.8278481012658227e-07, |
|
"loss": 0.1406, |
|
"step": 38900 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 2.7645569620253165e-07, |
|
"loss": 0.1119, |
|
"step": 38925 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 2.70126582278481e-07, |
|
"loss": 0.1014, |
|
"step": 38950 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 2.637974683544304e-07, |
|
"loss": 0.1347, |
|
"step": 38975 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 2.5746835443037977e-07, |
|
"loss": 0.1729, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 2.5113924050632914e-07, |
|
"loss": 0.1548, |
|
"step": 39025 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 2.4481012658227847e-07, |
|
"loss": 0.1231, |
|
"step": 39050 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 2.3848101265822784e-07, |
|
"loss": 0.0956, |
|
"step": 39075 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 2.3215189873417722e-07, |
|
"loss": 0.0943, |
|
"step": 39100 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 2.258227848101266e-07, |
|
"loss": 0.0936, |
|
"step": 39125 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 2.1949367088607597e-07, |
|
"loss": 0.0946, |
|
"step": 39150 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 2.1316455696202534e-07, |
|
"loss": 0.0801, |
|
"step": 39175 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 2.068354430379747e-07, |
|
"loss": 0.0778, |
|
"step": 39200 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 2.0050632911392406e-07, |
|
"loss": 0.0658, |
|
"step": 39225 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.9417721518987344e-07, |
|
"loss": 0.0657, |
|
"step": 39250 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.8784810126582279e-07, |
|
"loss": 0.115, |
|
"step": 39275 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.8151898734177216e-07, |
|
"loss": 0.156, |
|
"step": 39300 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.7518987341772154e-07, |
|
"loss": 0.1441, |
|
"step": 39325 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.688607594936709e-07, |
|
"loss": 0.1533, |
|
"step": 39350 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.6253164556962026e-07, |
|
"loss": 0.1601, |
|
"step": 39375 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.5620253164556963e-07, |
|
"loss": 0.1398, |
|
"step": 39400 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.49873417721519e-07, |
|
"loss": 0.1524, |
|
"step": 39425 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.437974683544304e-07, |
|
"loss": 0.1411, |
|
"step": 39450 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.3746835443037975e-07, |
|
"loss": 0.1243, |
|
"step": 39475 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.3113924050632913e-07, |
|
"loss": 0.1273, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.248101265822785e-07, |
|
"loss": 0.1032, |
|
"step": 39525 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.1848101265822787e-07, |
|
"loss": 0.1165, |
|
"step": 39550 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.1215189873417723e-07, |
|
"loss": 0.1081, |
|
"step": 39575 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.058227848101266e-07, |
|
"loss": 0.1129, |
|
"step": 39600 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 9.949367088607595e-08, |
|
"loss": 0.1003, |
|
"step": 39625 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 9.316455696202531e-08, |
|
"loss": 0.1484, |
|
"step": 39650 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 8.683544303797469e-08, |
|
"loss": 0.1375, |
|
"step": 39675 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 8.050632911392405e-08, |
|
"loss": 0.161, |
|
"step": 39700 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 7.417721518987342e-08, |
|
"loss": 0.1496, |
|
"step": 39725 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 6.784810126582278e-08, |
|
"loss": 0.1301, |
|
"step": 39750 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 6.151898734177216e-08, |
|
"loss": 0.1403, |
|
"step": 39775 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 5.518987341772152e-08, |
|
"loss": 0.1288, |
|
"step": 39800 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 4.886075949367089e-08, |
|
"loss": 0.1333, |
|
"step": 39825 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 4.2531645569620256e-08, |
|
"loss": 0.1177, |
|
"step": 39850 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 3.6202531645569624e-08, |
|
"loss": 0.1393, |
|
"step": 39875 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 2.9873417721518986e-08, |
|
"loss": 0.1316, |
|
"step": 39900 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 2.3544303797468354e-08, |
|
"loss": 0.1175, |
|
"step": 39925 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.7215189873417722e-08, |
|
"loss": 0.1799, |
|
"step": 39950 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.088607594936709e-08, |
|
"loss": 0.2076, |
|
"step": 39975 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 4.556962025316456e-09, |
|
"loss": 0.1812, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"eval_loss": 0.3104100823402405, |
|
"eval_runtime": 2197.4372, |
|
"eval_samples_per_second": 7.442, |
|
"eval_steps_per_second": 0.466, |
|
"eval_wer": 12.141503068470264, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"step": 40000, |
|
"total_flos": 3.6938844665339904e+20, |
|
"train_loss": 0.2097677074164152, |
|
"train_runtime": 178630.0447, |
|
"train_samples_per_second": 7.166, |
|
"train_steps_per_second": 0.224 |
|
} |
|
], |
|
"max_steps": 40000, |
|
"num_train_epochs": 9223372036854775807, |
|
"total_flos": 3.6938844665339904e+20, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|