|
{ |
|
"best_metric": 8.527938593638197, |
|
"best_model_checkpoint": "data/model/whisper-large-zh-cv11/checkpoint-10000", |
|
"epoch": 10.770186335403727, |
|
"global_step": 13000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 5.7500000000000005e-08, |
|
"loss": 0.7131, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.2000000000000002e-07, |
|
"loss": 0.6928, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.8e-07, |
|
"loss": 0.6595, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 2.425e-07, |
|
"loss": 0.5153, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 3.0500000000000004e-07, |
|
"loss": 0.4906, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 3.6750000000000003e-07, |
|
"loss": 0.5071, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.275000000000001e-07, |
|
"loss": 0.4779, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.900000000000001e-07, |
|
"loss": 0.4359, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 5.525e-07, |
|
"loss": 0.3774, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 6.15e-07, |
|
"loss": 0.3551, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 6.775000000000001e-07, |
|
"loss": 0.4035, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 7.4e-07, |
|
"loss": 0.3458, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.000000000000001e-07, |
|
"loss": 0.3318, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.625e-07, |
|
"loss": 0.37, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.25e-07, |
|
"loss": 0.3916, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.875e-07, |
|
"loss": 0.326, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.0500000000000001e-06, |
|
"loss": 0.3782, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.1125000000000001e-06, |
|
"loss": 0.3511, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.175e-06, |
|
"loss": 0.3968, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.2375e-06, |
|
"loss": 0.3452, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.3e-06, |
|
"loss": 0.3163, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.3625000000000003e-06, |
|
"loss": 0.3828, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.425e-06, |
|
"loss": 0.3238, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.4875000000000002e-06, |
|
"loss": 0.304, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.5500000000000002e-06, |
|
"loss": 0.297, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.6125000000000002e-06, |
|
"loss": 0.3206, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.6750000000000003e-06, |
|
"loss": 0.3039, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.7375e-06, |
|
"loss": 0.3194, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.8000000000000001e-06, |
|
"loss": 0.4097, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.8625000000000002e-06, |
|
"loss": 0.3444, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.925e-06, |
|
"loss": 0.3444, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.9875000000000005e-06, |
|
"loss": 0.3229, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.05e-06, |
|
"loss": 0.326, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.11e-06, |
|
"loss": 0.319, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 2.1725000000000004e-06, |
|
"loss": 0.3635, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 2.235e-06, |
|
"loss": 0.3485, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.2975000000000004e-06, |
|
"loss": 0.347, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.3600000000000003e-06, |
|
"loss": 0.3167, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.4225e-06, |
|
"loss": 0.3109, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 2.4850000000000003e-06, |
|
"loss": 0.3314, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"eval_cer": 10.804657516424902, |
|
"eval_loss": 0.21102139353752136, |
|
"eval_runtime": 1318.6013, |
|
"eval_samples_per_second": 0.758, |
|
"eval_steps_per_second": 0.095, |
|
"eval_wer": 65.70135746606334, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 2.5475e-06, |
|
"loss": 0.3142, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.6100000000000004e-06, |
|
"loss": 0.3126, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.6725000000000002e-06, |
|
"loss": 0.3299, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.7350000000000005e-06, |
|
"loss": 0.3266, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.7975000000000003e-06, |
|
"loss": 0.3309, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.86e-06, |
|
"loss": 0.286, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.9225000000000004e-06, |
|
"loss": 0.3244, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.9850000000000002e-06, |
|
"loss": 0.3118, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 3.0475000000000005e-06, |
|
"loss": 0.2895, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 3.1100000000000003e-06, |
|
"loss": 0.2544, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 3.1725e-06, |
|
"loss": 0.3011, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 3.2350000000000004e-06, |
|
"loss": 0.2979, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 3.2975e-06, |
|
"loss": 0.2447, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 3.3600000000000004e-06, |
|
"loss": 0.2354, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 3.4225000000000003e-06, |
|
"loss": 0.2887, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 3.485e-06, |
|
"loss": 0.2693, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 3.5475000000000003e-06, |
|
"loss": 0.2511, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 3.61e-06, |
|
"loss": 0.2638, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 3.6725000000000004e-06, |
|
"loss": 0.2627, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 3.7350000000000002e-06, |
|
"loss": 0.2528, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 3.7975e-06, |
|
"loss": 0.2675, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 3.86e-06, |
|
"loss": 0.2385, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 3.9225e-06, |
|
"loss": 0.2715, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 3.985000000000001e-06, |
|
"loss": 0.2803, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 4.047500000000001e-06, |
|
"loss": 0.2713, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 4.1100000000000005e-06, |
|
"loss": 0.2667, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 4.1725e-06, |
|
"loss": 0.2927, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 4.235e-06, |
|
"loss": 0.3252, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 4.297500000000001e-06, |
|
"loss": 0.295, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 4.360000000000001e-06, |
|
"loss": 0.2235, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 4.4225000000000004e-06, |
|
"loss": 0.2847, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 4.485e-06, |
|
"loss": 0.2674, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 4.5475e-06, |
|
"loss": 0.2974, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 4.610000000000001e-06, |
|
"loss": 0.2653, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.672500000000001e-06, |
|
"loss": 0.3325, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.735e-06, |
|
"loss": 0.2757, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 4.7975e-06, |
|
"loss": 0.2582, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 4.86e-06, |
|
"loss": 0.3071, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 4.922500000000001e-06, |
|
"loss": 0.252, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 4.9850000000000006e-06, |
|
"loss": 0.2747, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"eval_cer": 9.419111429128993, |
|
"eval_loss": 0.2004988044500351, |
|
"eval_runtime": 1247.0192, |
|
"eval_samples_per_second": 0.802, |
|
"eval_steps_per_second": 0.1, |
|
"eval_wer": 58.190045248868785, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.994722222222223e-06, |
|
"loss": 0.2901, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.987777777777778e-06, |
|
"loss": 0.2592, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 4.980833333333333e-06, |
|
"loss": 0.2879, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 4.973888888888889e-06, |
|
"loss": 0.2954, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 4.966944444444445e-06, |
|
"loss": 0.2808, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 4.960000000000001e-06, |
|
"loss": 0.3092, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 4.953055555555556e-06, |
|
"loss": 0.2634, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 4.946111111111111e-06, |
|
"loss": 0.3078, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 4.9391666666666675e-06, |
|
"loss": 0.2763, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 4.932222222222223e-06, |
|
"loss": 0.3011, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 4.925277777777778e-06, |
|
"loss": 0.2745, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 4.918333333333334e-06, |
|
"loss": 0.247, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 4.9113888888888895e-06, |
|
"loss": 0.2612, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 4.904444444444445e-06, |
|
"loss": 0.29, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 4.897500000000001e-06, |
|
"loss": 0.2413, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 4.890555555555556e-06, |
|
"loss": 0.2996, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 4.8836111111111116e-06, |
|
"loss": 0.23, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 4.876666666666668e-06, |
|
"loss": 0.1839, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 4.869722222222222e-06, |
|
"loss": 0.1838, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 4.862777777777778e-06, |
|
"loss": 0.1913, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 4.8558333333333336e-06, |
|
"loss": 0.184, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 4.848888888888889e-06, |
|
"loss": 0.1738, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 4.841944444444445e-06, |
|
"loss": 0.1657, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 4.835e-06, |
|
"loss": 0.1812, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 4.828055555555556e-06, |
|
"loss": 0.2329, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 4.821111111111112e-06, |
|
"loss": 0.1668, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 4.814166666666667e-06, |
|
"loss": 0.2225, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 4.807222222222222e-06, |
|
"loss": 0.176, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 4.8002777777777785e-06, |
|
"loss": 0.191, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 4.793333333333334e-06, |
|
"loss": 0.1801, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 4.786388888888889e-06, |
|
"loss": 0.1818, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 4.779444444444445e-06, |
|
"loss": 0.1905, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 4.7725000000000005e-06, |
|
"loss": 0.184, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.765555555555556e-06, |
|
"loss": 0.2096, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 4.758611111111112e-06, |
|
"loss": 0.1834, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 4.751666666666667e-06, |
|
"loss": 0.1748, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 4.7447222222222225e-06, |
|
"loss": 0.1842, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 4.737777777777779e-06, |
|
"loss": 0.1677, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 4.730833333333334e-06, |
|
"loss": 0.2009, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 4.723888888888889e-06, |
|
"loss": 0.1989, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"eval_cer": 9.093865868730893, |
|
"eval_loss": 0.1983313262462616, |
|
"eval_runtime": 1253.8539, |
|
"eval_samples_per_second": 0.798, |
|
"eval_steps_per_second": 0.1, |
|
"eval_wer": 56.199095022624434, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 4.716944444444445e-06, |
|
"loss": 0.1982, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 4.71e-06, |
|
"loss": 0.209, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 4.703055555555556e-06, |
|
"loss": 0.2011, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 4.696111111111111e-06, |
|
"loss": 0.1636, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 4.6891666666666665e-06, |
|
"loss": 0.2001, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 4.682222222222223e-06, |
|
"loss": 0.175, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 4.675277777777778e-06, |
|
"loss": 0.1793, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 4.668333333333333e-06, |
|
"loss": 0.178, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 4.661388888888889e-06, |
|
"loss": 0.1952, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 4.654444444444445e-06, |
|
"loss": 0.2102, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 4.6475e-06, |
|
"loss": 0.1808, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.640555555555556e-06, |
|
"loss": 0.2351, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 4.633611111111111e-06, |
|
"loss": 0.2021, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 4.626666666666667e-06, |
|
"loss": 0.2086, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 4.619722222222223e-06, |
|
"loss": 0.1765, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 4.612777777777778e-06, |
|
"loss": 0.1985, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 4.605833333333333e-06, |
|
"loss": 0.182, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 4.5988888888888896e-06, |
|
"loss": 0.1857, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 4.591944444444445e-06, |
|
"loss": 0.2088, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 4.585e-06, |
|
"loss": 0.195, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 4.578055555555556e-06, |
|
"loss": 0.2044, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 4.571111111111112e-06, |
|
"loss": 0.1874, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 4.564166666666668e-06, |
|
"loss": 0.2111, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 4.557222222222222e-06, |
|
"loss": 0.1801, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 4.550277777777778e-06, |
|
"loss": 0.1896, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 4.543333333333334e-06, |
|
"loss": 0.129, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 4.536388888888889e-06, |
|
"loss": 0.1039, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 4.529444444444445e-06, |
|
"loss": 0.1133, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 4.5225e-06, |
|
"loss": 0.1216, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 4.515555555555556e-06, |
|
"loss": 0.1199, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 4.508611111111112e-06, |
|
"loss": 0.1221, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 4.501666666666667e-06, |
|
"loss": 0.1357, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 4.494722222222222e-06, |
|
"loss": 0.0955, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 4.4877777777777785e-06, |
|
"loss": 0.1242, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 4.480833333333334e-06, |
|
"loss": 0.1122, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 4.473888888888889e-06, |
|
"loss": 0.127, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 4.466944444444445e-06, |
|
"loss": 0.1101, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 4.4600000000000005e-06, |
|
"loss": 0.1317, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 4.453055555555556e-06, |
|
"loss": 0.1186, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 4.446111111111112e-06, |
|
"loss": 0.1142, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"eval_cer": 9.158914980810511, |
|
"eval_loss": 0.20759885013103485, |
|
"eval_runtime": 1254.2811, |
|
"eval_samples_per_second": 0.797, |
|
"eval_steps_per_second": 0.1, |
|
"eval_wer": 55.022624434389144, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 4.439166666666667e-06, |
|
"loss": 0.1187, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 4.4322222222222225e-06, |
|
"loss": 0.1318, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 4.425277777777779e-06, |
|
"loss": 0.1111, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 4.418333333333334e-06, |
|
"loss": 0.1136, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 4.411388888888889e-06, |
|
"loss": 0.1061, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 4.404444444444445e-06, |
|
"loss": 0.1045, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 4.3975e-06, |
|
"loss": 0.1176, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 4.390555555555556e-06, |
|
"loss": 0.1193, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 4.383611111111111e-06, |
|
"loss": 0.0952, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 4.3766666666666665e-06, |
|
"loss": 0.1183, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 4.369722222222223e-06, |
|
"loss": 0.1106, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 4.362777777777778e-06, |
|
"loss": 0.125, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"learning_rate": 4.355833333333333e-06, |
|
"loss": 0.1492, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 4.348888888888889e-06, |
|
"loss": 0.1215, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 4.341944444444445e-06, |
|
"loss": 0.1318, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 4.335e-06, |
|
"loss": 0.1189, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 4.328055555555556e-06, |
|
"loss": 0.1259, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 4.3211111111111114e-06, |
|
"loss": 0.1375, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 4.314166666666667e-06, |
|
"loss": 0.1247, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 4.307222222222223e-06, |
|
"loss": 0.1222, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 4.300277777777778e-06, |
|
"loss": 0.1225, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 4.2933333333333334e-06, |
|
"loss": 0.1156, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 4.28638888888889e-06, |
|
"loss": 0.1444, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 4.279444444444445e-06, |
|
"loss": 0.1316, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 4.2725e-06, |
|
"loss": 0.143, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 4.265555555555556e-06, |
|
"loss": 0.134, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"learning_rate": 4.258611111111112e-06, |
|
"loss": 0.1254, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 4.251666666666667e-06, |
|
"loss": 0.1423, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 4.244722222222222e-06, |
|
"loss": 0.125, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 4.2377777777777775e-06, |
|
"loss": 0.1502, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 4.230833333333334e-06, |
|
"loss": 0.1235, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"learning_rate": 4.223888888888889e-06, |
|
"loss": 0.1123, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 4.216944444444444e-06, |
|
"loss": 0.1301, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 4.21e-06, |
|
"loss": 0.0698, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 4.203055555555556e-06, |
|
"loss": 0.0619, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"learning_rate": 4.196111111111111e-06, |
|
"loss": 0.0714, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 4.08, |
|
"learning_rate": 4.189166666666667e-06, |
|
"loss": 0.0705, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 4.182222222222222e-06, |
|
"loss": 0.0749, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 4.12, |
|
"learning_rate": 4.1752777777777785e-06, |
|
"loss": 0.0783, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 4.168333333333334e-06, |
|
"loss": 0.0747, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"eval_cer": 9.048331490275158, |
|
"eval_loss": 0.21309606730937958, |
|
"eval_runtime": 1246.5514, |
|
"eval_samples_per_second": 0.802, |
|
"eval_steps_per_second": 0.1, |
|
"eval_wer": 56.380090497737555, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"learning_rate": 4.161666666666667e-06, |
|
"loss": 0.0658, |
|
"step": 5025 |
|
}, |
|
{ |
|
"epoch": 4.18, |
|
"learning_rate": 4.154722222222223e-06, |
|
"loss": 0.0705, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"learning_rate": 4.147777777777778e-06, |
|
"loss": 0.0631, |
|
"step": 5075 |
|
}, |
|
{ |
|
"epoch": 4.23, |
|
"learning_rate": 4.140833333333334e-06, |
|
"loss": 0.0783, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 4.133888888888889e-06, |
|
"loss": 0.0649, |
|
"step": 5125 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"learning_rate": 4.126944444444444e-06, |
|
"loss": 0.083, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 4.29, |
|
"learning_rate": 4.12e-06, |
|
"loss": 0.0911, |
|
"step": 5175 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 4.113055555555556e-06, |
|
"loss": 0.0618, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 4.106111111111111e-06, |
|
"loss": 0.067, |
|
"step": 5225 |
|
}, |
|
{ |
|
"epoch": 4.35, |
|
"learning_rate": 4.099166666666667e-06, |
|
"loss": 0.077, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 4.092222222222222e-06, |
|
"loss": 0.0656, |
|
"step": 5275 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"learning_rate": 4.085277777777778e-06, |
|
"loss": 0.0701, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 4.078333333333334e-06, |
|
"loss": 0.0786, |
|
"step": 5325 |
|
}, |
|
{ |
|
"epoch": 4.43, |
|
"learning_rate": 4.071388888888889e-06, |
|
"loss": 0.0738, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 4.45, |
|
"learning_rate": 4.064444444444444e-06, |
|
"loss": 0.062, |
|
"step": 5375 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 4.0575000000000005e-06, |
|
"loss": 0.1005, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"learning_rate": 4.050555555555556e-06, |
|
"loss": 0.0794, |
|
"step": 5425 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 4.043611111111111e-06, |
|
"loss": 0.0554, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 4.036666666666667e-06, |
|
"loss": 0.0663, |
|
"step": 5475 |
|
}, |
|
{ |
|
"epoch": 4.56, |
|
"learning_rate": 4.0297222222222226e-06, |
|
"loss": 0.0867, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"learning_rate": 4.022777777777778e-06, |
|
"loss": 0.0804, |
|
"step": 5525 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 4.015833333333334e-06, |
|
"loss": 0.0754, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"learning_rate": 4.008888888888889e-06, |
|
"loss": 0.0693, |
|
"step": 5575 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 4.0019444444444446e-06, |
|
"loss": 0.0714, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 3.995000000000001e-06, |
|
"loss": 0.0813, |
|
"step": 5625 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 3.988055555555556e-06, |
|
"loss": 0.0745, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"learning_rate": 3.981111111111111e-06, |
|
"loss": 0.0699, |
|
"step": 5675 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 3.974166666666667e-06, |
|
"loss": 0.0832, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"learning_rate": 3.967222222222222e-06, |
|
"loss": 0.067, |
|
"step": 5725 |
|
}, |
|
{ |
|
"epoch": 4.76, |
|
"learning_rate": 3.960277777777778e-06, |
|
"loss": 0.0814, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 4.78, |
|
"learning_rate": 3.953333333333333e-06, |
|
"loss": 0.0799, |
|
"step": 5775 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"learning_rate": 3.9463888888888895e-06, |
|
"loss": 0.0639, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 4.83, |
|
"learning_rate": 3.939444444444445e-06, |
|
"loss": 0.0741, |
|
"step": 5825 |
|
}, |
|
{ |
|
"epoch": 4.85, |
|
"learning_rate": 3.9325e-06, |
|
"loss": 0.077, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 3.925555555555556e-06, |
|
"loss": 0.0855, |
|
"step": 5875 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 3.9186111111111115e-06, |
|
"loss": 0.0836, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"learning_rate": 3.911666666666667e-06, |
|
"loss": 0.0812, |
|
"step": 5925 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 3.904722222222223e-06, |
|
"loss": 0.0724, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 3.897777777777778e-06, |
|
"loss": 0.0815, |
|
"step": 5975 |
|
}, |
|
{ |
|
"epoch": 4.97, |
|
"learning_rate": 3.8908333333333335e-06, |
|
"loss": 0.0709, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 4.97, |
|
"eval_cer": 8.976777466987576, |
|
"eval_loss": 0.21651260554790497, |
|
"eval_runtime": 1250.0161, |
|
"eval_samples_per_second": 0.8, |
|
"eval_steps_per_second": 0.1, |
|
"eval_wer": 54.660633484162894, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"learning_rate": 3.88388888888889e-06, |
|
"loss": 0.0564, |
|
"step": 6025 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 3.876944444444445e-06, |
|
"loss": 0.0456, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 3.87e-06, |
|
"loss": 0.0394, |
|
"step": 6075 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"learning_rate": 3.863055555555556e-06, |
|
"loss": 0.0348, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"learning_rate": 3.856111111111112e-06, |
|
"loss": 0.0488, |
|
"step": 6125 |
|
}, |
|
{ |
|
"epoch": 5.1, |
|
"learning_rate": 3.849166666666667e-06, |
|
"loss": 0.0372, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 5.12, |
|
"learning_rate": 3.842222222222223e-06, |
|
"loss": 0.0517, |
|
"step": 6175 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 3.835277777777778e-06, |
|
"loss": 0.0397, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 5.16, |
|
"learning_rate": 3.828333333333334e-06, |
|
"loss": 0.0342, |
|
"step": 6225 |
|
}, |
|
{ |
|
"epoch": 5.18, |
|
"learning_rate": 3.821666666666667e-06, |
|
"loss": 0.0489, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 5.2, |
|
"learning_rate": 3.814722222222222e-06, |
|
"loss": 0.0631, |
|
"step": 6275 |
|
}, |
|
{ |
|
"epoch": 5.22, |
|
"learning_rate": 3.8077777777777782e-06, |
|
"loss": 0.0477, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 5.24, |
|
"learning_rate": 3.8008333333333335e-06, |
|
"loss": 0.0341, |
|
"step": 6325 |
|
}, |
|
{ |
|
"epoch": 5.26, |
|
"learning_rate": 3.7938888888888892e-06, |
|
"loss": 0.0503, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 5.28, |
|
"learning_rate": 3.786944444444445e-06, |
|
"loss": 0.0356, |
|
"step": 6375 |
|
}, |
|
{ |
|
"epoch": 5.3, |
|
"learning_rate": 3.7800000000000002e-06, |
|
"loss": 0.0395, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 5.32, |
|
"learning_rate": 3.773055555555556e-06, |
|
"loss": 0.0404, |
|
"step": 6425 |
|
}, |
|
{ |
|
"epoch": 5.34, |
|
"learning_rate": 3.7661111111111117e-06, |
|
"loss": 0.0437, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 5.36, |
|
"learning_rate": 3.759166666666667e-06, |
|
"loss": 0.0395, |
|
"step": 6475 |
|
}, |
|
{ |
|
"epoch": 5.39, |
|
"learning_rate": 3.7522222222222227e-06, |
|
"loss": 0.0382, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 5.41, |
|
"learning_rate": 3.7452777777777784e-06, |
|
"loss": 0.0414, |
|
"step": 6525 |
|
}, |
|
{ |
|
"epoch": 5.43, |
|
"learning_rate": 3.7383333333333337e-06, |
|
"loss": 0.0387, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 5.45, |
|
"learning_rate": 3.7313888888888894e-06, |
|
"loss": 0.0355, |
|
"step": 6575 |
|
}, |
|
{ |
|
"epoch": 5.47, |
|
"learning_rate": 3.724444444444445e-06, |
|
"loss": 0.0373, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 5.49, |
|
"learning_rate": 3.7175000000000004e-06, |
|
"loss": 0.0347, |
|
"step": 6625 |
|
}, |
|
{ |
|
"epoch": 5.51, |
|
"learning_rate": 3.710555555555556e-06, |
|
"loss": 0.0464, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"learning_rate": 3.703611111111111e-06, |
|
"loss": 0.0368, |
|
"step": 6675 |
|
}, |
|
{ |
|
"epoch": 5.55, |
|
"learning_rate": 3.6966666666666667e-06, |
|
"loss": 0.0459, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 5.57, |
|
"learning_rate": 3.6897222222222224e-06, |
|
"loss": 0.0454, |
|
"step": 6725 |
|
}, |
|
{ |
|
"epoch": 5.59, |
|
"learning_rate": 3.6827777777777777e-06, |
|
"loss": 0.0336, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 5.61, |
|
"learning_rate": 3.6758333333333334e-06, |
|
"loss": 0.045, |
|
"step": 6775 |
|
}, |
|
{ |
|
"epoch": 5.63, |
|
"learning_rate": 3.668888888888889e-06, |
|
"loss": 0.0407, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 5.65, |
|
"learning_rate": 3.6619444444444444e-06, |
|
"loss": 0.0356, |
|
"step": 6825 |
|
}, |
|
{ |
|
"epoch": 5.67, |
|
"learning_rate": 3.655e-06, |
|
"loss": 0.0429, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 5.7, |
|
"learning_rate": 3.648055555555556e-06, |
|
"loss": 0.0411, |
|
"step": 6875 |
|
}, |
|
{ |
|
"epoch": 5.72, |
|
"learning_rate": 3.641111111111111e-06, |
|
"loss": 0.0429, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 5.74, |
|
"learning_rate": 3.634166666666667e-06, |
|
"loss": 0.043, |
|
"step": 6925 |
|
}, |
|
{ |
|
"epoch": 5.76, |
|
"learning_rate": 3.6272222222222226e-06, |
|
"loss": 0.0415, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 5.78, |
|
"learning_rate": 3.6202777777777783e-06, |
|
"loss": 0.0476, |
|
"step": 6975 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"learning_rate": 3.6133333333333336e-06, |
|
"loss": 0.0432, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"eval_cer": 8.950757822155728, |
|
"eval_loss": 0.222176656126976, |
|
"eval_runtime": 1244.1829, |
|
"eval_samples_per_second": 0.804, |
|
"eval_steps_per_second": 0.1, |
|
"eval_wer": 54.02714932126697, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 3.6063888888888893e-06, |
|
"loss": 0.0481, |
|
"step": 7025 |
|
}, |
|
{ |
|
"epoch": 5.84, |
|
"learning_rate": 3.599444444444445e-06, |
|
"loss": 0.0462, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 5.86, |
|
"learning_rate": 3.5925000000000003e-06, |
|
"loss": 0.0458, |
|
"step": 7075 |
|
}, |
|
{ |
|
"epoch": 5.88, |
|
"learning_rate": 3.585555555555556e-06, |
|
"loss": 0.0593, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 5.9, |
|
"learning_rate": 3.5786111111111118e-06, |
|
"loss": 0.0614, |
|
"step": 7125 |
|
}, |
|
{ |
|
"epoch": 5.92, |
|
"learning_rate": 3.571666666666667e-06, |
|
"loss": 0.0472, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 5.94, |
|
"learning_rate": 3.5647222222222228e-06, |
|
"loss": 0.0441, |
|
"step": 7175 |
|
}, |
|
{ |
|
"epoch": 5.96, |
|
"learning_rate": 3.5577777777777785e-06, |
|
"loss": 0.0381, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 5.99, |
|
"learning_rate": 3.5508333333333334e-06, |
|
"loss": 0.0373, |
|
"step": 7225 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 3.543888888888889e-06, |
|
"loss": 0.0277, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 3.5369444444444444e-06, |
|
"loss": 0.0262, |
|
"step": 7275 |
|
}, |
|
{ |
|
"epoch": 6.05, |
|
"learning_rate": 3.53e-06, |
|
"loss": 0.0189, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"learning_rate": 3.523055555555556e-06, |
|
"loss": 0.0199, |
|
"step": 7325 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"learning_rate": 3.516111111111111e-06, |
|
"loss": 0.0252, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 6.11, |
|
"learning_rate": 3.509166666666667e-06, |
|
"loss": 0.0197, |
|
"step": 7375 |
|
}, |
|
{ |
|
"epoch": 6.13, |
|
"learning_rate": 3.5022222222222225e-06, |
|
"loss": 0.0218, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 6.15, |
|
"learning_rate": 3.495277777777778e-06, |
|
"loss": 0.0174, |
|
"step": 7425 |
|
}, |
|
{ |
|
"epoch": 6.17, |
|
"learning_rate": 3.4883333333333335e-06, |
|
"loss": 0.0209, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 6.19, |
|
"learning_rate": 3.4813888888888892e-06, |
|
"loss": 0.0195, |
|
"step": 7475 |
|
}, |
|
{ |
|
"epoch": 6.21, |
|
"learning_rate": 3.4744444444444445e-06, |
|
"loss": 0.0241, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 6.23, |
|
"learning_rate": 3.4675000000000003e-06, |
|
"loss": 0.0285, |
|
"step": 7525 |
|
}, |
|
{ |
|
"epoch": 6.26, |
|
"learning_rate": 3.460555555555556e-06, |
|
"loss": 0.0215, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 6.28, |
|
"learning_rate": 3.4536111111111113e-06, |
|
"loss": 0.0268, |
|
"step": 7575 |
|
}, |
|
{ |
|
"epoch": 6.3, |
|
"learning_rate": 3.446666666666667e-06, |
|
"loss": 0.0247, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 6.32, |
|
"learning_rate": 3.4397222222222227e-06, |
|
"loss": 0.0207, |
|
"step": 7625 |
|
}, |
|
{ |
|
"epoch": 6.34, |
|
"learning_rate": 3.4327777777777784e-06, |
|
"loss": 0.0167, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 6.36, |
|
"learning_rate": 3.4258333333333337e-06, |
|
"loss": 0.0288, |
|
"step": 7675 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"learning_rate": 3.4188888888888894e-06, |
|
"loss": 0.0268, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 6.4, |
|
"learning_rate": 3.411944444444445e-06, |
|
"loss": 0.0265, |
|
"step": 7725 |
|
}, |
|
{ |
|
"epoch": 6.42, |
|
"learning_rate": 3.4050000000000004e-06, |
|
"loss": 0.0189, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 6.44, |
|
"learning_rate": 3.398055555555556e-06, |
|
"loss": 0.0311, |
|
"step": 7775 |
|
}, |
|
{ |
|
"epoch": 6.46, |
|
"learning_rate": 3.391111111111111e-06, |
|
"loss": 0.0164, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 6.48, |
|
"learning_rate": 3.3841666666666667e-06, |
|
"loss": 0.0153, |
|
"step": 7825 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"learning_rate": 3.3772222222222225e-06, |
|
"loss": 0.0233, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 6.52, |
|
"learning_rate": 3.3702777777777777e-06, |
|
"loss": 0.028, |
|
"step": 7875 |
|
}, |
|
{ |
|
"epoch": 6.54, |
|
"learning_rate": 3.3633333333333335e-06, |
|
"loss": 0.0269, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 6.57, |
|
"learning_rate": 3.356388888888889e-06, |
|
"loss": 0.0232, |
|
"step": 7925 |
|
}, |
|
{ |
|
"epoch": 6.59, |
|
"learning_rate": 3.3494444444444445e-06, |
|
"loss": 0.0248, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 6.61, |
|
"learning_rate": 3.3425e-06, |
|
"loss": 0.0279, |
|
"step": 7975 |
|
}, |
|
{ |
|
"epoch": 6.63, |
|
"learning_rate": 3.335555555555556e-06, |
|
"loss": 0.0261, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 6.63, |
|
"eval_cer": 9.035321667859234, |
|
"eval_loss": 0.22986744344234467, |
|
"eval_runtime": 1240.496, |
|
"eval_samples_per_second": 0.806, |
|
"eval_steps_per_second": 0.101, |
|
"eval_wer": 54.47963800904977, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"learning_rate": 3.328611111111111e-06, |
|
"loss": 0.0261, |
|
"step": 8025 |
|
}, |
|
{ |
|
"epoch": 6.67, |
|
"learning_rate": 3.321666666666667e-06, |
|
"loss": 0.0264, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 6.69, |
|
"learning_rate": 3.3147222222222226e-06, |
|
"loss": 0.0217, |
|
"step": 8075 |
|
}, |
|
{ |
|
"epoch": 6.71, |
|
"learning_rate": 3.307777777777778e-06, |
|
"loss": 0.0235, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 6.73, |
|
"learning_rate": 3.3008333333333336e-06, |
|
"loss": 0.0284, |
|
"step": 8125 |
|
}, |
|
{ |
|
"epoch": 6.75, |
|
"learning_rate": 3.2938888888888894e-06, |
|
"loss": 0.0233, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"learning_rate": 3.2869444444444446e-06, |
|
"loss": 0.0192, |
|
"step": 8175 |
|
}, |
|
{ |
|
"epoch": 6.79, |
|
"learning_rate": 3.2800000000000004e-06, |
|
"loss": 0.0234, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 6.81, |
|
"learning_rate": 3.273055555555556e-06, |
|
"loss": 0.023, |
|
"step": 8225 |
|
}, |
|
{ |
|
"epoch": 6.83, |
|
"learning_rate": 3.2661111111111114e-06, |
|
"loss": 0.0248, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 6.86, |
|
"learning_rate": 3.259166666666667e-06, |
|
"loss": 0.0298, |
|
"step": 8275 |
|
}, |
|
{ |
|
"epoch": 6.88, |
|
"learning_rate": 3.252222222222223e-06, |
|
"loss": 0.0283, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 6.9, |
|
"learning_rate": 3.245277777777778e-06, |
|
"loss": 0.0276, |
|
"step": 8325 |
|
}, |
|
{ |
|
"epoch": 6.92, |
|
"learning_rate": 3.2383333333333334e-06, |
|
"loss": 0.024, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 6.94, |
|
"learning_rate": 3.231388888888889e-06, |
|
"loss": 0.0241, |
|
"step": 8375 |
|
}, |
|
{ |
|
"epoch": 6.96, |
|
"learning_rate": 3.2244444444444444e-06, |
|
"loss": 0.0195, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 6.98, |
|
"learning_rate": 3.2175e-06, |
|
"loss": 0.0277, |
|
"step": 8425 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 3.210555555555556e-06, |
|
"loss": 0.0208, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 3.203611111111111e-06, |
|
"loss": 0.0135, |
|
"step": 8475 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 3.196666666666667e-06, |
|
"loss": 0.0098, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 7.06, |
|
"learning_rate": 3.1897222222222226e-06, |
|
"loss": 0.0109, |
|
"step": 8525 |
|
}, |
|
{ |
|
"epoch": 7.08, |
|
"learning_rate": 3.182777777777778e-06, |
|
"loss": 0.0139, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 7.1, |
|
"learning_rate": 3.1758333333333336e-06, |
|
"loss": 0.0115, |
|
"step": 8575 |
|
}, |
|
{ |
|
"epoch": 7.13, |
|
"learning_rate": 3.1688888888888893e-06, |
|
"loss": 0.0127, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 7.15, |
|
"learning_rate": 3.1619444444444446e-06, |
|
"loss": 0.0139, |
|
"step": 8625 |
|
}, |
|
{ |
|
"epoch": 7.17, |
|
"learning_rate": 3.1550000000000003e-06, |
|
"loss": 0.0147, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 7.19, |
|
"learning_rate": 3.148055555555556e-06, |
|
"loss": 0.0117, |
|
"step": 8675 |
|
}, |
|
{ |
|
"epoch": 7.21, |
|
"learning_rate": 3.1411111111111113e-06, |
|
"loss": 0.0117, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 7.23, |
|
"learning_rate": 3.134166666666667e-06, |
|
"loss": 0.0097, |
|
"step": 8725 |
|
}, |
|
{ |
|
"epoch": 7.25, |
|
"learning_rate": 3.1272222222222227e-06, |
|
"loss": 0.0128, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 7.27, |
|
"learning_rate": 3.120277777777778e-06, |
|
"loss": 0.0129, |
|
"step": 8775 |
|
}, |
|
{ |
|
"epoch": 7.29, |
|
"learning_rate": 3.1133333333333337e-06, |
|
"loss": 0.0113, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 7.31, |
|
"learning_rate": 3.1063888888888895e-06, |
|
"loss": 0.0115, |
|
"step": 8825 |
|
}, |
|
{ |
|
"epoch": 7.33, |
|
"learning_rate": 3.0994444444444447e-06, |
|
"loss": 0.0142, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 7.35, |
|
"learning_rate": 3.0925000000000005e-06, |
|
"loss": 0.0109, |
|
"step": 8875 |
|
}, |
|
{ |
|
"epoch": 7.37, |
|
"learning_rate": 3.085555555555556e-06, |
|
"loss": 0.0125, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 7.39, |
|
"learning_rate": 3.078611111111111e-06, |
|
"loss": 0.0155, |
|
"step": 8925 |
|
}, |
|
{ |
|
"epoch": 7.41, |
|
"learning_rate": 3.0716666666666668e-06, |
|
"loss": 0.0177, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 7.44, |
|
"learning_rate": 3.064722222222222e-06, |
|
"loss": 0.0133, |
|
"step": 8975 |
|
}, |
|
{ |
|
"epoch": 7.46, |
|
"learning_rate": 3.0577777777777778e-06, |
|
"loss": 0.0152, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 7.46, |
|
"eval_cer": 8.807649775580563, |
|
"eval_loss": 0.22896485030651093, |
|
"eval_runtime": 1251.4358, |
|
"eval_samples_per_second": 0.799, |
|
"eval_steps_per_second": 0.1, |
|
"eval_wer": 52.76018099547512, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 7.48, |
|
"learning_rate": 3.0508333333333335e-06, |
|
"loss": 0.0085, |
|
"step": 9025 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 3.043888888888889e-06, |
|
"loss": 0.0131, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 7.52, |
|
"learning_rate": 3.0369444444444445e-06, |
|
"loss": 0.0125, |
|
"step": 9075 |
|
}, |
|
{ |
|
"epoch": 7.54, |
|
"learning_rate": 3.0300000000000002e-06, |
|
"loss": 0.0109, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 7.56, |
|
"learning_rate": 3.023055555555556e-06, |
|
"loss": 0.0123, |
|
"step": 9125 |
|
}, |
|
{ |
|
"epoch": 7.58, |
|
"learning_rate": 3.0161111111111112e-06, |
|
"loss": 0.0151, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 7.6, |
|
"learning_rate": 3.009166666666667e-06, |
|
"loss": 0.0117, |
|
"step": 9175 |
|
}, |
|
{ |
|
"epoch": 7.62, |
|
"learning_rate": 3.0022222222222227e-06, |
|
"loss": 0.0143, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 7.64, |
|
"learning_rate": 2.995277777777778e-06, |
|
"loss": 0.0135, |
|
"step": 9225 |
|
}, |
|
{ |
|
"epoch": 7.66, |
|
"learning_rate": 2.9883333333333337e-06, |
|
"loss": 0.013, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 7.68, |
|
"learning_rate": 2.9813888888888894e-06, |
|
"loss": 0.0166, |
|
"step": 9275 |
|
}, |
|
{ |
|
"epoch": 7.7, |
|
"learning_rate": 2.9744444444444447e-06, |
|
"loss": 0.0115, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 7.73, |
|
"learning_rate": 2.9675000000000004e-06, |
|
"loss": 0.0123, |
|
"step": 9325 |
|
}, |
|
{ |
|
"epoch": 7.75, |
|
"learning_rate": 2.960555555555556e-06, |
|
"loss": 0.0164, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 7.77, |
|
"learning_rate": 2.9536111111111114e-06, |
|
"loss": 0.0167, |
|
"step": 9375 |
|
}, |
|
{ |
|
"epoch": 7.79, |
|
"learning_rate": 2.946666666666667e-06, |
|
"loss": 0.0137, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 7.81, |
|
"learning_rate": 2.939722222222223e-06, |
|
"loss": 0.0157, |
|
"step": 9425 |
|
}, |
|
{ |
|
"epoch": 7.83, |
|
"learning_rate": 2.932777777777778e-06, |
|
"loss": 0.0103, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 7.85, |
|
"learning_rate": 2.9258333333333334e-06, |
|
"loss": 0.0138, |
|
"step": 9475 |
|
}, |
|
{ |
|
"epoch": 7.87, |
|
"learning_rate": 2.9188888888888887e-06, |
|
"loss": 0.0116, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 7.89, |
|
"learning_rate": 2.9119444444444444e-06, |
|
"loss": 0.0106, |
|
"step": 9525 |
|
}, |
|
{ |
|
"epoch": 7.91, |
|
"learning_rate": 2.905e-06, |
|
"loss": 0.0142, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 7.93, |
|
"learning_rate": 2.8980555555555554e-06, |
|
"loss": 0.0126, |
|
"step": 9575 |
|
}, |
|
{ |
|
"epoch": 7.95, |
|
"learning_rate": 2.891111111111111e-06, |
|
"loss": 0.0139, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 7.97, |
|
"learning_rate": 2.884166666666667e-06, |
|
"loss": 0.0135, |
|
"step": 9625 |
|
}, |
|
{ |
|
"epoch": 7.99, |
|
"learning_rate": 2.877222222222222e-06, |
|
"loss": 0.0132, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 2.870277777777778e-06, |
|
"loss": 0.0071, |
|
"step": 9675 |
|
}, |
|
{ |
|
"epoch": 8.04, |
|
"learning_rate": 2.8633333333333336e-06, |
|
"loss": 0.0071, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"learning_rate": 2.8563888888888893e-06, |
|
"loss": 0.0061, |
|
"step": 9725 |
|
}, |
|
{ |
|
"epoch": 8.08, |
|
"learning_rate": 2.8494444444444446e-06, |
|
"loss": 0.0064, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 8.1, |
|
"learning_rate": 2.8425000000000003e-06, |
|
"loss": 0.008, |
|
"step": 9775 |
|
}, |
|
{ |
|
"epoch": 8.12, |
|
"learning_rate": 2.835555555555556e-06, |
|
"loss": 0.0058, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 8.14, |
|
"learning_rate": 2.8286111111111113e-06, |
|
"loss": 0.0067, |
|
"step": 9825 |
|
}, |
|
{ |
|
"epoch": 8.16, |
|
"learning_rate": 2.821666666666667e-06, |
|
"loss": 0.0059, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 8.18, |
|
"learning_rate": 2.8147222222222228e-06, |
|
"loss": 0.0051, |
|
"step": 9875 |
|
}, |
|
{ |
|
"epoch": 8.2, |
|
"learning_rate": 2.807777777777778e-06, |
|
"loss": 0.0074, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 8.22, |
|
"learning_rate": 2.8008333333333338e-06, |
|
"loss": 0.0063, |
|
"step": 9925 |
|
}, |
|
{ |
|
"epoch": 8.24, |
|
"learning_rate": 2.7938888888888895e-06, |
|
"loss": 0.0052, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 8.26, |
|
"learning_rate": 2.7869444444444448e-06, |
|
"loss": 0.0065, |
|
"step": 9975 |
|
}, |
|
{ |
|
"epoch": 8.28, |
|
"learning_rate": 2.7800000000000005e-06, |
|
"loss": 0.0054, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 8.28, |
|
"eval_cer": 8.527938593638197, |
|
"eval_loss": 0.2435399442911148, |
|
"eval_runtime": 1243.3236, |
|
"eval_samples_per_second": 0.804, |
|
"eval_steps_per_second": 0.101, |
|
"eval_wer": 51.67420814479639, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 8.31, |
|
"learning_rate": 2.773055555555556e-06, |
|
"loss": 0.0047, |
|
"step": 10025 |
|
}, |
|
{ |
|
"epoch": 8.33, |
|
"learning_rate": 2.766111111111111e-06, |
|
"loss": 0.0064, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 8.35, |
|
"learning_rate": 2.759166666666667e-06, |
|
"loss": 0.0068, |
|
"step": 10075 |
|
}, |
|
{ |
|
"epoch": 8.37, |
|
"learning_rate": 2.752222222222222e-06, |
|
"loss": 0.0057, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 8.39, |
|
"learning_rate": 2.745277777777778e-06, |
|
"loss": 0.0058, |
|
"step": 10125 |
|
}, |
|
{ |
|
"epoch": 8.41, |
|
"learning_rate": 2.7383333333333335e-06, |
|
"loss": 0.0064, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 8.43, |
|
"learning_rate": 2.731388888888889e-06, |
|
"loss": 0.0071, |
|
"step": 10175 |
|
}, |
|
{ |
|
"epoch": 8.45, |
|
"learning_rate": 2.7244444444444445e-06, |
|
"loss": 0.0059, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 8.47, |
|
"learning_rate": 2.7175000000000002e-06, |
|
"loss": 0.0082, |
|
"step": 10225 |
|
}, |
|
{ |
|
"epoch": 8.49, |
|
"learning_rate": 2.7105555555555555e-06, |
|
"loss": 0.0056, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 8.51, |
|
"learning_rate": 2.7036111111111112e-06, |
|
"loss": 0.0048, |
|
"step": 10275 |
|
}, |
|
{ |
|
"epoch": 8.53, |
|
"learning_rate": 2.696666666666667e-06, |
|
"loss": 0.0068, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 8.55, |
|
"learning_rate": 2.6897222222222223e-06, |
|
"loss": 0.0063, |
|
"step": 10325 |
|
}, |
|
{ |
|
"epoch": 8.57, |
|
"learning_rate": 2.682777777777778e-06, |
|
"loss": 0.0061, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 8.6, |
|
"learning_rate": 2.6758333333333337e-06, |
|
"loss": 0.0075, |
|
"step": 10375 |
|
}, |
|
{ |
|
"epoch": 8.62, |
|
"learning_rate": 2.6688888888888894e-06, |
|
"loss": 0.0051, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 8.64, |
|
"learning_rate": 2.6619444444444447e-06, |
|
"loss": 0.0071, |
|
"step": 10425 |
|
}, |
|
{ |
|
"epoch": 8.66, |
|
"learning_rate": 2.6550000000000004e-06, |
|
"loss": 0.0067, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 8.68, |
|
"learning_rate": 2.648055555555556e-06, |
|
"loss": 0.0076, |
|
"step": 10475 |
|
}, |
|
{ |
|
"epoch": 8.7, |
|
"learning_rate": 2.6411111111111114e-06, |
|
"loss": 0.0084, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 8.72, |
|
"learning_rate": 2.634166666666667e-06, |
|
"loss": 0.0066, |
|
"step": 10525 |
|
}, |
|
{ |
|
"epoch": 8.74, |
|
"learning_rate": 2.627222222222223e-06, |
|
"loss": 0.0078, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 8.76, |
|
"learning_rate": 2.620277777777778e-06, |
|
"loss": 0.0068, |
|
"step": 10575 |
|
}, |
|
{ |
|
"epoch": 8.78, |
|
"learning_rate": 2.6133333333333334e-06, |
|
"loss": 0.0077, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 8.8, |
|
"learning_rate": 2.6063888888888887e-06, |
|
"loss": 0.0057, |
|
"step": 10625 |
|
}, |
|
{ |
|
"epoch": 8.82, |
|
"learning_rate": 2.5994444444444444e-06, |
|
"loss": 0.0082, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 8.84, |
|
"learning_rate": 2.5925e-06, |
|
"loss": 0.0078, |
|
"step": 10675 |
|
}, |
|
{ |
|
"epoch": 8.86, |
|
"learning_rate": 2.5855555555555555e-06, |
|
"loss": 0.0069, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 8.89, |
|
"learning_rate": 2.578611111111111e-06, |
|
"loss": 0.0057, |
|
"step": 10725 |
|
}, |
|
{ |
|
"epoch": 8.91, |
|
"learning_rate": 2.571666666666667e-06, |
|
"loss": 0.0083, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 8.93, |
|
"learning_rate": 2.564722222222222e-06, |
|
"loss": 0.0069, |
|
"step": 10775 |
|
}, |
|
{ |
|
"epoch": 8.95, |
|
"learning_rate": 2.557777777777778e-06, |
|
"loss": 0.009, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 8.97, |
|
"learning_rate": 2.5508333333333336e-06, |
|
"loss": 0.0074, |
|
"step": 10825 |
|
}, |
|
{ |
|
"epoch": 8.99, |
|
"learning_rate": 2.543888888888889e-06, |
|
"loss": 0.0082, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 9.01, |
|
"learning_rate": 2.5369444444444446e-06, |
|
"loss": 0.0056, |
|
"step": 10875 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 2.5300000000000003e-06, |
|
"loss": 0.0036, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 9.05, |
|
"learning_rate": 2.5230555555555556e-06, |
|
"loss": 0.003, |
|
"step": 10925 |
|
}, |
|
{ |
|
"epoch": 9.07, |
|
"learning_rate": 2.5161111111111113e-06, |
|
"loss": 0.0051, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 9.09, |
|
"learning_rate": 2.509166666666667e-06, |
|
"loss": 0.0033, |
|
"step": 10975 |
|
}, |
|
{ |
|
"epoch": 9.11, |
|
"learning_rate": 2.5022222222222224e-06, |
|
"loss": 0.0028, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 9.11, |
|
"eval_cer": 8.983282378195538, |
|
"eval_loss": 0.2420540750026703, |
|
"eval_runtime": 1244.171, |
|
"eval_samples_per_second": 0.804, |
|
"eval_steps_per_second": 0.1, |
|
"eval_wer": 53.0316742081448, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 9.13, |
|
"learning_rate": 2.4952777777777777e-06, |
|
"loss": 0.0031, |
|
"step": 11025 |
|
}, |
|
{ |
|
"epoch": 9.15, |
|
"learning_rate": 2.4883333333333334e-06, |
|
"loss": 0.0026, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 9.18, |
|
"learning_rate": 2.481388888888889e-06, |
|
"loss": 0.0037, |
|
"step": 11075 |
|
}, |
|
{ |
|
"epoch": 9.2, |
|
"learning_rate": 2.474444444444445e-06, |
|
"loss": 0.0022, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 9.22, |
|
"learning_rate": 2.4675e-06, |
|
"loss": 0.003, |
|
"step": 11125 |
|
}, |
|
{ |
|
"epoch": 9.24, |
|
"learning_rate": 2.460555555555556e-06, |
|
"loss": 0.003, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 9.26, |
|
"learning_rate": 2.4536111111111115e-06, |
|
"loss": 0.0039, |
|
"step": 11175 |
|
}, |
|
{ |
|
"epoch": 9.28, |
|
"learning_rate": 2.446666666666667e-06, |
|
"loss": 0.0033, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 9.3, |
|
"learning_rate": 2.4397222222222225e-06, |
|
"loss": 0.0032, |
|
"step": 11225 |
|
}, |
|
{ |
|
"epoch": 9.32, |
|
"learning_rate": 2.4327777777777782e-06, |
|
"loss": 0.0028, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 9.34, |
|
"learning_rate": 2.4258333333333335e-06, |
|
"loss": 0.0036, |
|
"step": 11275 |
|
}, |
|
{ |
|
"epoch": 9.36, |
|
"learning_rate": 2.418888888888889e-06, |
|
"loss": 0.0038, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 9.38, |
|
"learning_rate": 2.4119444444444446e-06, |
|
"loss": 0.0035, |
|
"step": 11325 |
|
}, |
|
{ |
|
"epoch": 9.4, |
|
"learning_rate": 2.4050000000000003e-06, |
|
"loss": 0.0039, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 9.42, |
|
"learning_rate": 2.3980555555555556e-06, |
|
"loss": 0.0031, |
|
"step": 11375 |
|
}, |
|
{ |
|
"epoch": 9.44, |
|
"learning_rate": 2.3911111111111113e-06, |
|
"loss": 0.0039, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 9.47, |
|
"learning_rate": 2.384166666666667e-06, |
|
"loss": 0.0019, |
|
"step": 11425 |
|
}, |
|
{ |
|
"epoch": 9.49, |
|
"learning_rate": 2.3772222222222223e-06, |
|
"loss": 0.0045, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 9.51, |
|
"learning_rate": 2.370277777777778e-06, |
|
"loss": 0.0024, |
|
"step": 11475 |
|
}, |
|
{ |
|
"epoch": 9.53, |
|
"learning_rate": 2.3633333333333337e-06, |
|
"loss": 0.0027, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 9.55, |
|
"learning_rate": 2.356388888888889e-06, |
|
"loss": 0.004, |
|
"step": 11525 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"learning_rate": 2.3494444444444447e-06, |
|
"loss": 0.0027, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 9.59, |
|
"learning_rate": 2.3425000000000004e-06, |
|
"loss": 0.0034, |
|
"step": 11575 |
|
}, |
|
{ |
|
"epoch": 9.61, |
|
"learning_rate": 2.3355555555555557e-06, |
|
"loss": 0.0034, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 9.63, |
|
"learning_rate": 2.328611111111111e-06, |
|
"loss": 0.0022, |
|
"step": 11625 |
|
}, |
|
{ |
|
"epoch": 9.65, |
|
"learning_rate": 2.3216666666666667e-06, |
|
"loss": 0.0031, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 9.67, |
|
"learning_rate": 2.3147222222222225e-06, |
|
"loss": 0.0029, |
|
"step": 11675 |
|
}, |
|
{ |
|
"epoch": 9.69, |
|
"learning_rate": 2.3077777777777778e-06, |
|
"loss": 0.0035, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 9.71, |
|
"learning_rate": 2.3008333333333335e-06, |
|
"loss": 0.0037, |
|
"step": 11725 |
|
}, |
|
{ |
|
"epoch": 9.73, |
|
"learning_rate": 2.293888888888889e-06, |
|
"loss": 0.002, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 9.76, |
|
"learning_rate": 2.286944444444445e-06, |
|
"loss": 0.003, |
|
"step": 11775 |
|
}, |
|
{ |
|
"epoch": 9.78, |
|
"learning_rate": 2.28e-06, |
|
"loss": 0.002, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 9.8, |
|
"learning_rate": 2.273055555555556e-06, |
|
"loss": 0.003, |
|
"step": 11825 |
|
}, |
|
{ |
|
"epoch": 9.82, |
|
"learning_rate": 2.2661111111111116e-06, |
|
"loss": 0.0038, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 9.84, |
|
"learning_rate": 2.259166666666667e-06, |
|
"loss": 0.004, |
|
"step": 11875 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"learning_rate": 2.252222222222222e-06, |
|
"loss": 0.0041, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 9.88, |
|
"learning_rate": 2.245277777777778e-06, |
|
"loss": 0.0032, |
|
"step": 11925 |
|
}, |
|
{ |
|
"epoch": 9.9, |
|
"learning_rate": 2.2383333333333336e-06, |
|
"loss": 0.0042, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 9.92, |
|
"learning_rate": 2.231388888888889e-06, |
|
"loss": 0.0047, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 9.94, |
|
"learning_rate": 2.2244444444444447e-06, |
|
"loss": 0.0045, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 9.94, |
|
"eval_cer": 8.775125219540755, |
|
"eval_loss": 0.24615037441253662, |
|
"eval_runtime": 1241.5039, |
|
"eval_samples_per_second": 0.805, |
|
"eval_steps_per_second": 0.101, |
|
"eval_wer": 52.94117647058824, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 9.96, |
|
"learning_rate": 2.2175000000000004e-06, |
|
"loss": 0.0044, |
|
"step": 12025 |
|
}, |
|
{ |
|
"epoch": 9.98, |
|
"learning_rate": 2.2105555555555557e-06, |
|
"loss": 0.0032, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 2.2036111111111114e-06, |
|
"loss": 0.0027, |
|
"step": 12075 |
|
}, |
|
{ |
|
"epoch": 10.02, |
|
"learning_rate": 2.196666666666667e-06, |
|
"loss": 0.0029, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 10.05, |
|
"learning_rate": 2.1897222222222224e-06, |
|
"loss": 0.002, |
|
"step": 12125 |
|
}, |
|
{ |
|
"epoch": 10.07, |
|
"learning_rate": 2.1827777777777777e-06, |
|
"loss": 0.0016, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 10.09, |
|
"learning_rate": 2.1758333333333334e-06, |
|
"loss": 0.0014, |
|
"step": 12175 |
|
}, |
|
{ |
|
"epoch": 10.11, |
|
"learning_rate": 2.168888888888889e-06, |
|
"loss": 0.0014, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 10.13, |
|
"learning_rate": 2.1619444444444444e-06, |
|
"loss": 0.0019, |
|
"step": 12225 |
|
}, |
|
{ |
|
"epoch": 10.15, |
|
"learning_rate": 2.155e-06, |
|
"loss": 0.0017, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 10.17, |
|
"learning_rate": 2.148055555555556e-06, |
|
"loss": 0.0019, |
|
"step": 12275 |
|
}, |
|
{ |
|
"epoch": 10.19, |
|
"learning_rate": 2.141111111111111e-06, |
|
"loss": 0.0021, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 10.21, |
|
"learning_rate": 2.134166666666667e-06, |
|
"loss": 0.0019, |
|
"step": 12325 |
|
}, |
|
{ |
|
"epoch": 10.23, |
|
"learning_rate": 2.1272222222222226e-06, |
|
"loss": 0.0024, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 10.25, |
|
"learning_rate": 2.120277777777778e-06, |
|
"loss": 0.0018, |
|
"step": 12375 |
|
}, |
|
{ |
|
"epoch": 10.27, |
|
"learning_rate": 2.1133333333333336e-06, |
|
"loss": 0.002, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 10.29, |
|
"learning_rate": 2.106388888888889e-06, |
|
"loss": 0.0029, |
|
"step": 12425 |
|
}, |
|
{ |
|
"epoch": 10.31, |
|
"learning_rate": 2.0994444444444446e-06, |
|
"loss": 0.0022, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 10.34, |
|
"learning_rate": 2.0925000000000003e-06, |
|
"loss": 0.0022, |
|
"step": 12475 |
|
}, |
|
{ |
|
"epoch": 10.36, |
|
"learning_rate": 2.0855555555555556e-06, |
|
"loss": 0.0024, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 10.38, |
|
"learning_rate": 2.0786111111111113e-06, |
|
"loss": 0.0021, |
|
"step": 12525 |
|
}, |
|
{ |
|
"epoch": 10.4, |
|
"learning_rate": 2.071666666666667e-06, |
|
"loss": 0.0023, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 10.42, |
|
"learning_rate": 2.0647222222222223e-06, |
|
"loss": 0.0028, |
|
"step": 12575 |
|
}, |
|
{ |
|
"epoch": 10.44, |
|
"learning_rate": 2.057777777777778e-06, |
|
"loss": 0.0019, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 10.46, |
|
"learning_rate": 2.0508333333333337e-06, |
|
"loss": 0.0022, |
|
"step": 12625 |
|
}, |
|
{ |
|
"epoch": 10.48, |
|
"learning_rate": 2.043888888888889e-06, |
|
"loss": 0.0013, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 10.5, |
|
"learning_rate": 2.0369444444444448e-06, |
|
"loss": 0.002, |
|
"step": 12675 |
|
}, |
|
{ |
|
"epoch": 10.52, |
|
"learning_rate": 2.0300000000000005e-06, |
|
"loss": 0.0024, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 10.54, |
|
"learning_rate": 2.0230555555555558e-06, |
|
"loss": 0.0021, |
|
"step": 12725 |
|
}, |
|
{ |
|
"epoch": 10.56, |
|
"learning_rate": 2.016111111111111e-06, |
|
"loss": 0.0025, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 10.58, |
|
"learning_rate": 2.0091666666666668e-06, |
|
"loss": 0.0021, |
|
"step": 12775 |
|
}, |
|
{ |
|
"epoch": 10.6, |
|
"learning_rate": 2.0022222222222225e-06, |
|
"loss": 0.0018, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 10.63, |
|
"learning_rate": 1.9952777777777778e-06, |
|
"loss": 0.0023, |
|
"step": 12825 |
|
}, |
|
{ |
|
"epoch": 10.65, |
|
"learning_rate": 1.9883333333333335e-06, |
|
"loss": 0.0023, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 10.67, |
|
"learning_rate": 1.9813888888888892e-06, |
|
"loss": 0.0019, |
|
"step": 12875 |
|
}, |
|
{ |
|
"epoch": 10.69, |
|
"learning_rate": 1.9744444444444445e-06, |
|
"loss": 0.0022, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 10.71, |
|
"learning_rate": 1.9675000000000002e-06, |
|
"loss": 0.002, |
|
"step": 12925 |
|
}, |
|
{ |
|
"epoch": 10.73, |
|
"learning_rate": 1.960555555555556e-06, |
|
"loss": 0.0015, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 10.75, |
|
"learning_rate": 1.9536111111111112e-06, |
|
"loss": 0.0014, |
|
"step": 12975 |
|
}, |
|
{ |
|
"epoch": 10.77, |
|
"learning_rate": 1.9466666666666665e-06, |
|
"loss": 0.0016, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 10.77, |
|
"eval_cer": 8.95726273336369, |
|
"eval_loss": 0.2500666379928589, |
|
"eval_runtime": 1240.5149, |
|
"eval_samples_per_second": 0.806, |
|
"eval_steps_per_second": 0.101, |
|
"eval_wer": 52.307692307692314, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 10.77, |
|
"step": 13000, |
|
"total_flos": 8.83515156885504e+20, |
|
"train_loss": 0.10485561544820667, |
|
"train_runtime": 107666.9313, |
|
"train_samples_per_second": 5.944, |
|
"train_steps_per_second": 0.186 |
|
} |
|
], |
|
"max_steps": 20000, |
|
"num_train_epochs": 17, |
|
"total_flos": 8.83515156885504e+20, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|