|
{ |
|
"best_metric": 18.48141795311607, |
|
"best_model_checkpoint": "whisper-base-nl-3/checkpoint-29000", |
|
"epoch": 49.07306434023991, |
|
"global_step": 45000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 4.0000000000000003e-07, |
|
"loss": 1.3847, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 9.000000000000001e-07, |
|
"loss": 1.4601, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.4000000000000001e-06, |
|
"loss": 1.3556, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9000000000000002e-06, |
|
"loss": 1.1454, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.4000000000000003e-06, |
|
"loss": 1.2552, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.9e-06, |
|
"loss": 1.1695, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 3.4000000000000005e-06, |
|
"loss": 0.9503, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 3.900000000000001e-06, |
|
"loss": 0.9096, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.4e-06, |
|
"loss": 0.8853, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.9000000000000005e-06, |
|
"loss": 0.833, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 5.400000000000001e-06, |
|
"loss": 0.8597, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 5.9e-06, |
|
"loss": 0.8199, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 6.4000000000000006e-06, |
|
"loss": 0.7158, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 6.9e-06, |
|
"loss": 0.6896, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 7.4e-06, |
|
"loss": 0.662, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 7.9e-06, |
|
"loss": 0.8284, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.400000000000001e-06, |
|
"loss": 0.6544, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.900000000000001e-06, |
|
"loss": 0.7486, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.4e-06, |
|
"loss": 0.5652, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.9e-06, |
|
"loss": 0.6604, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.994202898550725e-06, |
|
"loss": 0.7287, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.986956521739132e-06, |
|
"loss": 0.6413, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.979710144927537e-06, |
|
"loss": 0.6762, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.972463768115943e-06, |
|
"loss": 0.6882, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.965217391304348e-06, |
|
"loss": 0.7015, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.957971014492755e-06, |
|
"loss": 0.6145, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.95072463768116e-06, |
|
"loss": 0.6674, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.943478260869565e-06, |
|
"loss": 0.5915, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.936231884057972e-06, |
|
"loss": 0.6632, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.928985507246378e-06, |
|
"loss": 0.6157, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.921739130434783e-06, |
|
"loss": 0.7222, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.91449275362319e-06, |
|
"loss": 0.6447, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.907246376811595e-06, |
|
"loss": 0.6848, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.9e-06, |
|
"loss": 0.6391, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.892753623188407e-06, |
|
"loss": 0.6543, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.885507246376812e-06, |
|
"loss": 0.5579, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.878260869565218e-06, |
|
"loss": 0.5699, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.871014492753625e-06, |
|
"loss": 0.5404, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.86376811594203e-06, |
|
"loss": 0.6237, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.856521739130435e-06, |
|
"loss": 0.5761, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"eval_cer": 10.115352452161936, |
|
"eval_loss": 0.5675458908081055, |
|
"eval_runtime": 3220.6897, |
|
"eval_samples_per_second": 0.771, |
|
"eval_steps_per_second": 0.385, |
|
"eval_wer": 28.15323041738136, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.84927536231884e-06, |
|
"loss": 0.6175, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.842028985507247e-06, |
|
"loss": 0.6934, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.834782608695654e-06, |
|
"loss": 0.5818, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.827536231884059e-06, |
|
"loss": 0.6175, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.820289855072465e-06, |
|
"loss": 0.4952, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.81304347826087e-06, |
|
"loss": 0.6194, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.805797101449275e-06, |
|
"loss": 0.5897, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.798550724637682e-06, |
|
"loss": 0.5356, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.791304347826089e-06, |
|
"loss": 0.5805, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.784057971014494e-06, |
|
"loss": 0.5446, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.776811594202899e-06, |
|
"loss": 0.6734, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.769565217391305e-06, |
|
"loss": 0.5457, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.76231884057971e-06, |
|
"loss": 0.4686, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.755072463768117e-06, |
|
"loss": 0.5954, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.747826086956522e-06, |
|
"loss": 0.4973, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.740579710144929e-06, |
|
"loss": 0.5727, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.733333333333334e-06, |
|
"loss": 0.5767, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.726086956521739e-06, |
|
"loss": 0.4814, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.718840579710146e-06, |
|
"loss": 0.5544, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.71159420289855e-06, |
|
"loss": 0.5281, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.704347826086957e-06, |
|
"loss": 0.612, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.697101449275364e-06, |
|
"loss": 0.5293, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.689855072463769e-06, |
|
"loss": 0.5881, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.682608695652174e-06, |
|
"loss": 0.5571, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.67536231884058e-06, |
|
"loss": 0.429, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.668115942028986e-06, |
|
"loss": 0.4661, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.660869565217392e-06, |
|
"loss": 0.5181, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.653623188405797e-06, |
|
"loss": 0.5188, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.646376811594204e-06, |
|
"loss": 0.4881, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.639130434782609e-06, |
|
"loss": 0.6319, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.631884057971014e-06, |
|
"loss": 0.6, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.624637681159421e-06, |
|
"loss": 0.5251, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.617391304347828e-06, |
|
"loss": 0.4999, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.610144927536233e-06, |
|
"loss": 0.5612, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.60289855072464e-06, |
|
"loss": 0.456, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.595652173913044e-06, |
|
"loss": 0.5589, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.58840579710145e-06, |
|
"loss": 0.396, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.581159420289856e-06, |
|
"loss": 0.4713, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.573913043478261e-06, |
|
"loss": 0.5178, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.566666666666668e-06, |
|
"loss": 0.48, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"eval_cer": 9.69108170593733, |
|
"eval_loss": 0.5239235162734985, |
|
"eval_runtime": 3297.9692, |
|
"eval_samples_per_second": 0.753, |
|
"eval_steps_per_second": 0.376, |
|
"eval_wer": 26.436439870402133, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.559420289855074e-06, |
|
"loss": 0.5154, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.55217391304348e-06, |
|
"loss": 0.5715, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.545217391304349e-06, |
|
"loss": 0.5522, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.537971014492754e-06, |
|
"loss": 0.5205, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.53072463768116e-06, |
|
"loss": 0.4494, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.523478260869566e-06, |
|
"loss": 0.5226, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.516231884057972e-06, |
|
"loss": 0.469, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.508985507246377e-06, |
|
"loss": 0.5067, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.501739130434784e-06, |
|
"loss": 0.5879, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.494492753623189e-06, |
|
"loss": 0.8063, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.487246376811596e-06, |
|
"loss": 0.5067, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.48e-06, |
|
"loss": 0.5124, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.472753623188406e-06, |
|
"loss": 0.4894, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.465507246376812e-06, |
|
"loss": 0.504, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.458260869565217e-06, |
|
"loss": 0.545, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.451014492753624e-06, |
|
"loss": 0.4962, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.44376811594203e-06, |
|
"loss": 0.4774, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.436521739130436e-06, |
|
"loss": 0.5429, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.42927536231884e-06, |
|
"loss": 0.445, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.422028985507246e-06, |
|
"loss": 0.4111, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.414782608695653e-06, |
|
"loss": 0.5188, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.40753623188406e-06, |
|
"loss": 0.4584, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.400289855072464e-06, |
|
"loss": 0.5017, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.393043478260871e-06, |
|
"loss": 0.593, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.385797101449276e-06, |
|
"loss": 0.5195, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.378550724637681e-06, |
|
"loss": 0.5083, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.371304347826088e-06, |
|
"loss": 0.5534, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.364057971014493e-06, |
|
"loss": 0.4893, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.3568115942029e-06, |
|
"loss": 0.4512, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.349565217391306e-06, |
|
"loss": 0.5876, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.342318840579711e-06, |
|
"loss": 0.4597, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.335072463768116e-06, |
|
"loss": 0.3634, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.327826086956523e-06, |
|
"loss": 0.4483, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.320579710144928e-06, |
|
"loss": 0.5484, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.313333333333335e-06, |
|
"loss": 0.4852, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.306086956521741e-06, |
|
"loss": 0.5576, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.298840579710146e-06, |
|
"loss": 0.5107, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.291594202898551e-06, |
|
"loss": 0.5286, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.284347826086956e-06, |
|
"loss": 0.5618, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.277101449275363e-06, |
|
"loss": 0.4094, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"eval_cer": 9.153180185941267, |
|
"eval_loss": 0.49246901273727417, |
|
"eval_runtime": 3242.5657, |
|
"eval_samples_per_second": 0.765, |
|
"eval_steps_per_second": 0.383, |
|
"eval_wer": 24.835525061940157, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.26985507246377e-06, |
|
"loss": 0.5152, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.262608695652175e-06, |
|
"loss": 0.5572, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.255362318840581e-06, |
|
"loss": 0.4801, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.248115942028986e-06, |
|
"loss": 0.4005, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.240869565217391e-06, |
|
"loss": 0.4693, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.233623188405798e-06, |
|
"loss": 0.4487, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.226376811594203e-06, |
|
"loss": 0.4477, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.21913043478261e-06, |
|
"loss": 0.4, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.211884057971015e-06, |
|
"loss": 0.4358, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.204637681159421e-06, |
|
"loss": 0.4952, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.197391304347826e-06, |
|
"loss": 0.4526, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.190144927536233e-06, |
|
"loss": 0.3588, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.182898550724638e-06, |
|
"loss": 0.5127, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.175652173913045e-06, |
|
"loss": 0.6166, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.16840579710145e-06, |
|
"loss": 0.5032, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.161159420289855e-06, |
|
"loss": 0.5632, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.153913043478262e-06, |
|
"loss": 0.5268, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.146666666666667e-06, |
|
"loss": 0.4672, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.139420289855073e-06, |
|
"loss": 0.4342, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.13217391304348e-06, |
|
"loss": 0.4504, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.124927536231885e-06, |
|
"loss": 0.4076, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.11768115942029e-06, |
|
"loss": 0.5036, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.110434782608695e-06, |
|
"loss": 0.3835, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.103188405797102e-06, |
|
"loss": 0.3652, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.095942028985508e-06, |
|
"loss": 0.4231, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.088695652173913e-06, |
|
"loss": 0.4644, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.08144927536232e-06, |
|
"loss": 0.5602, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.074202898550725e-06, |
|
"loss": 0.4637, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.06695652173913e-06, |
|
"loss": 0.4324, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.059710144927537e-06, |
|
"loss": 0.446, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.052463768115944e-06, |
|
"loss": 0.5127, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.045217391304349e-06, |
|
"loss": 0.4453, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.037971014492755e-06, |
|
"loss": 0.4382, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.03072463768116e-06, |
|
"loss": 0.5816, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.023478260869565e-06, |
|
"loss": 0.397, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.016231884057972e-06, |
|
"loss": 0.4512, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.008985507246377e-06, |
|
"loss": 0.4541, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.001739130434784e-06, |
|
"loss": 0.3829, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.99449275362319e-06, |
|
"loss": 0.5275, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.987246376811595e-06, |
|
"loss": 0.4792, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"eval_cer": 8.841433420237099, |
|
"eval_loss": 0.47019296884536743, |
|
"eval_runtime": 3179.8405, |
|
"eval_samples_per_second": 0.781, |
|
"eval_steps_per_second": 0.39, |
|
"eval_wer": 24.110539355822375, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.98e-06, |
|
"loss": 0.4638, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.972753623188405e-06, |
|
"loss": 0.5326, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.965507246376812e-06, |
|
"loss": 0.4799, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.958260869565219e-06, |
|
"loss": 0.4798, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.951014492753624e-06, |
|
"loss": 0.4938, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.94376811594203e-06, |
|
"loss": 0.4882, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.936521739130436e-06, |
|
"loss": 0.4136, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.92927536231884e-06, |
|
"loss": 0.5116, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.922028985507247e-06, |
|
"loss": 0.5381, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.914782608695652e-06, |
|
"loss": 0.3999, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.907536231884059e-06, |
|
"loss": 0.3925, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.900289855072464e-06, |
|
"loss": 0.469, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.89304347826087e-06, |
|
"loss": 0.4854, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.885797101449276e-06, |
|
"loss": 0.5086, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.878550724637682e-06, |
|
"loss": 0.4679, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.871594202898552e-06, |
|
"loss": 0.5175, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.864347826086957e-06, |
|
"loss": 0.4921, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.857101449275362e-06, |
|
"loss": 0.4375, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.849855072463769e-06, |
|
"loss": 0.6226, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.842608695652175e-06, |
|
"loss": 0.4383, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.83536231884058e-06, |
|
"loss": 0.4799, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.828115942028987e-06, |
|
"loss": 0.4063, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.820869565217392e-06, |
|
"loss": 0.4039, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.813623188405797e-06, |
|
"loss": 0.3688, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.806376811594204e-06, |
|
"loss": 0.4714, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.799130434782609e-06, |
|
"loss": 0.426, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.791884057971015e-06, |
|
"loss": 0.4623, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.784637681159422e-06, |
|
"loss": 0.4223, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 8.777391304347827e-06, |
|
"loss": 0.504, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 8.770144927536232e-06, |
|
"loss": 0.4396, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 8.762898550724639e-06, |
|
"loss": 0.4665, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 8.755652173913044e-06, |
|
"loss": 0.4356, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 8.74840579710145e-06, |
|
"loss": 0.4058, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 8.741159420289856e-06, |
|
"loss": 0.4825, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 8.733913043478262e-06, |
|
"loss": 0.5461, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 8.726666666666667e-06, |
|
"loss": 0.4718, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 8.719420289855072e-06, |
|
"loss": 0.5011, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 8.712173913043479e-06, |
|
"loss": 0.4281, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 8.704927536231886e-06, |
|
"loss": 0.5207, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 8.69768115942029e-06, |
|
"loss": 0.3444, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"eval_cer": 8.853116237886763, |
|
"eval_loss": 0.45443984866142273, |
|
"eval_runtime": 3270.5063, |
|
"eval_samples_per_second": 0.759, |
|
"eval_steps_per_second": 0.379, |
|
"eval_wer": 23.901658090337335, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 8.690434782608697e-06, |
|
"loss": 0.4848, |
|
"step": 5025 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 8.683188405797102e-06, |
|
"loss": 0.4012, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 8.675942028985507e-06, |
|
"loss": 0.3266, |
|
"step": 5075 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 8.668695652173914e-06, |
|
"loss": 0.4751, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 8.661449275362319e-06, |
|
"loss": 0.5552, |
|
"step": 5125 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 8.654202898550726e-06, |
|
"loss": 0.4079, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 8.64695652173913e-06, |
|
"loss": 0.4326, |
|
"step": 5175 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 8.639710144927538e-06, |
|
"loss": 0.4591, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 8.632463768115943e-06, |
|
"loss": 0.3962, |
|
"step": 5225 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 8.625217391304348e-06, |
|
"loss": 0.4825, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 8.617971014492754e-06, |
|
"loss": 0.4377, |
|
"step": 5275 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 8.610724637681161e-06, |
|
"loss": 0.437, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 8.603478260869566e-06, |
|
"loss": 0.3836, |
|
"step": 5325 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 8.596231884057971e-06, |
|
"loss": 0.4649, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 8.588985507246378e-06, |
|
"loss": 0.4631, |
|
"step": 5375 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 8.581739130434783e-06, |
|
"loss": 0.5619, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 8.57449275362319e-06, |
|
"loss": 0.5031, |
|
"step": 5425 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 8.567246376811596e-06, |
|
"loss": 0.3797, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 8.560000000000001e-06, |
|
"loss": 0.4993, |
|
"step": 5475 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.552753623188406e-06, |
|
"loss": 0.4146, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.545507246376811e-06, |
|
"loss": 0.4524, |
|
"step": 5525 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.538260869565218e-06, |
|
"loss": 0.4812, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.531014492753624e-06, |
|
"loss": 0.3781, |
|
"step": 5575 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.52376811594203e-06, |
|
"loss": 0.4343, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 8.516521739130436e-06, |
|
"loss": 0.4226, |
|
"step": 5625 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.509275362318841e-06, |
|
"loss": 0.4341, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.502028985507246e-06, |
|
"loss": 0.4008, |
|
"step": 5675 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.494782608695653e-06, |
|
"loss": 0.4199, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.487536231884058e-06, |
|
"loss": 0.4822, |
|
"step": 5725 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.480289855072465e-06, |
|
"loss": 0.3857, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 8.473043478260871e-06, |
|
"loss": 0.5646, |
|
"step": 5775 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 8.465797101449276e-06, |
|
"loss": 0.3835, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 8.458550724637681e-06, |
|
"loss": 0.4856, |
|
"step": 5825 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 8.451304347826088e-06, |
|
"loss": 0.4807, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 8.444057971014493e-06, |
|
"loss": 0.5281, |
|
"step": 5875 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 8.4368115942029e-06, |
|
"loss": 0.4234, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 8.429565217391305e-06, |
|
"loss": 0.6375, |
|
"step": 5925 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 8.422318840579711e-06, |
|
"loss": 0.4303, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 8.415072463768116e-06, |
|
"loss": 0.4196, |
|
"step": 5975 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 8.407826086956521e-06, |
|
"loss": 0.3943, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"eval_cer": 8.360224310098873, |
|
"eval_loss": 0.44455686211586, |
|
"eval_runtime": 3263.9495, |
|
"eval_samples_per_second": 0.76, |
|
"eval_steps_per_second": 0.38, |
|
"eval_wer": 22.735277301315037, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 8.400579710144928e-06, |
|
"loss": 0.4164, |
|
"step": 6025 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 8.393333333333335e-06, |
|
"loss": 0.4818, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 8.38608695652174e-06, |
|
"loss": 0.5536, |
|
"step": 6075 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 8.378840579710147e-06, |
|
"loss": 0.4731, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 8.371594202898552e-06, |
|
"loss": 0.4081, |
|
"step": 6125 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 8.364347826086957e-06, |
|
"loss": 0.3741, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 8.357101449275363e-06, |
|
"loss": 0.4173, |
|
"step": 6175 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 8.349855072463768e-06, |
|
"loss": 0.4954, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 8.342608695652175e-06, |
|
"loss": 0.461, |
|
"step": 6225 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 8.33536231884058e-06, |
|
"loss": 0.4501, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 8.328115942028987e-06, |
|
"loss": 0.3652, |
|
"step": 6275 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 8.320869565217392e-06, |
|
"loss": 0.4046, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 8.313623188405798e-06, |
|
"loss": 0.4184, |
|
"step": 6325 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 8.306376811594203e-06, |
|
"loss": 0.4287, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 8.29913043478261e-06, |
|
"loss": 0.3938, |
|
"step": 6375 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 8.291884057971015e-06, |
|
"loss": 0.4758, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 8.28463768115942e-06, |
|
"loss": 0.4716, |
|
"step": 6425 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 8.277391304347827e-06, |
|
"loss": 0.3819, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 8.270144927536232e-06, |
|
"loss": 0.3889, |
|
"step": 6475 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 8.262898550724639e-06, |
|
"loss": 0.4659, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 8.255652173913045e-06, |
|
"loss": 0.3274, |
|
"step": 6525 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 8.24840579710145e-06, |
|
"loss": 0.4144, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 8.241159420289855e-06, |
|
"loss": 0.3875, |
|
"step": 6575 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 8.23391304347826e-06, |
|
"loss": 0.452, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 8.226666666666667e-06, |
|
"loss": 0.4747, |
|
"step": 6625 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 8.219420289855074e-06, |
|
"loss": 0.4448, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 8.212173913043479e-06, |
|
"loss": 0.5215, |
|
"step": 6675 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 8.204927536231885e-06, |
|
"loss": 0.4261, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 8.19768115942029e-06, |
|
"loss": 0.4482, |
|
"step": 6725 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 8.190434782608695e-06, |
|
"loss": 0.39, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 8.183188405797102e-06, |
|
"loss": 0.6096, |
|
"step": 6775 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 8.175942028985507e-06, |
|
"loss": 0.3912, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 8.168695652173914e-06, |
|
"loss": 0.3141, |
|
"step": 6825 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 8.16144927536232e-06, |
|
"loss": 0.3593, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 8.154202898550725e-06, |
|
"loss": 0.4331, |
|
"step": 6875 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 8.14695652173913e-06, |
|
"loss": 0.4574, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 8.139710144927537e-06, |
|
"loss": 0.3546, |
|
"step": 6925 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 8.132463768115942e-06, |
|
"loss": 0.3894, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 8.125217391304349e-06, |
|
"loss": 0.393, |
|
"step": 6975 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 8.117971014492756e-06, |
|
"loss": 0.4925, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"eval_cer": 8.3723990358601, |
|
"eval_loss": 0.4348393380641937, |
|
"eval_runtime": 3254.2457, |
|
"eval_samples_per_second": 0.763, |
|
"eval_steps_per_second": 0.381, |
|
"eval_wer": 22.178768820278254, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 8.11072463768116e-06, |
|
"loss": 0.3986, |
|
"step": 7025 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 8.103478260869566e-06, |
|
"loss": 0.3943, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 8.09623188405797e-06, |
|
"loss": 0.4803, |
|
"step": 7075 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 8.088985507246377e-06, |
|
"loss": 0.4128, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 8.081739130434784e-06, |
|
"loss": 0.5511, |
|
"step": 7125 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 8.074492753623189e-06, |
|
"loss": 0.4377, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 8.067246376811596e-06, |
|
"loss": 0.453, |
|
"step": 7175 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 8.06e-06, |
|
"loss": 0.4019, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 8.052753623188406e-06, |
|
"loss": 0.4934, |
|
"step": 7225 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 8.045507246376812e-06, |
|
"loss": 0.3873, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 8.038260869565217e-06, |
|
"loss": 0.389, |
|
"step": 7275 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 8.031014492753624e-06, |
|
"loss": 0.4086, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 8.02376811594203e-06, |
|
"loss": 0.3844, |
|
"step": 7325 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 8.016521739130436e-06, |
|
"loss": 0.4577, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 8.009275362318841e-06, |
|
"loss": 0.383, |
|
"step": 7375 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 8.002028985507248e-06, |
|
"loss": 0.5191, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 7.994782608695653e-06, |
|
"loss": 0.3293, |
|
"step": 7425 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 7.98753623188406e-06, |
|
"loss": 0.4246, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 7.980289855072464e-06, |
|
"loss": 0.3811, |
|
"step": 7475 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 7.97304347826087e-06, |
|
"loss": 0.3891, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 7.965797101449276e-06, |
|
"loss": 0.4367, |
|
"step": 7525 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 7.958550724637681e-06, |
|
"loss": 0.4793, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 7.951304347826088e-06, |
|
"loss": 0.3993, |
|
"step": 7575 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 7.944347826086957e-06, |
|
"loss": 0.3095, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 7.937101449275362e-06, |
|
"loss": 0.4002, |
|
"step": 7625 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 7.929855072463769e-06, |
|
"loss": 0.4037, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 7.922608695652174e-06, |
|
"loss": 0.3694, |
|
"step": 7675 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 7.91536231884058e-06, |
|
"loss": 0.3969, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 7.908115942028987e-06, |
|
"loss": 0.4404, |
|
"step": 7725 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 7.900869565217392e-06, |
|
"loss": 0.4489, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 7.893623188405797e-06, |
|
"loss": 0.3542, |
|
"step": 7775 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 7.886376811594204e-06, |
|
"loss": 0.4793, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 7.879130434782609e-06, |
|
"loss": 0.3876, |
|
"step": 7825 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 7.871884057971016e-06, |
|
"loss": 0.3983, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 7.86463768115942e-06, |
|
"loss": 0.5017, |
|
"step": 7875 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 7.857391304347827e-06, |
|
"loss": 0.3704, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 7.850144927536232e-06, |
|
"loss": 0.4443, |
|
"step": 7925 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 7.842898550724637e-06, |
|
"loss": 0.3535, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 7.835652173913044e-06, |
|
"loss": 0.4255, |
|
"step": 7975 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 7.828405797101451e-06, |
|
"loss": 0.4455, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"eval_cer": 8.29885877318117, |
|
"eval_loss": 0.42695876955986023, |
|
"eval_runtime": 3202.8428, |
|
"eval_samples_per_second": 0.775, |
|
"eval_steps_per_second": 0.387, |
|
"eval_wer": 21.754907566228322, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 7.821159420289856e-06, |
|
"loss": 0.5201, |
|
"step": 8025 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 7.813913043478263e-06, |
|
"loss": 0.4287, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 7.806666666666668e-06, |
|
"loss": 0.3597, |
|
"step": 8075 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 7.799420289855073e-06, |
|
"loss": 0.2911, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 7.79217391304348e-06, |
|
"loss": 0.4192, |
|
"step": 8125 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 7.784927536231884e-06, |
|
"loss": 0.328, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 7.777681159420291e-06, |
|
"loss": 0.4364, |
|
"step": 8175 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 7.770434782608696e-06, |
|
"loss": 0.2877, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 7.763188405797101e-06, |
|
"loss": 0.3641, |
|
"step": 8225 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 7.755942028985508e-06, |
|
"loss": 0.3338, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 7.748695652173913e-06, |
|
"loss": 0.3425, |
|
"step": 8275 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 7.74144927536232e-06, |
|
"loss": 0.3633, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 7.734202898550726e-06, |
|
"loss": 0.4027, |
|
"step": 8325 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 7.726956521739131e-06, |
|
"loss": 0.4075, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 7.719710144927536e-06, |
|
"loss": 0.4272, |
|
"step": 8375 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 7.712463768115943e-06, |
|
"loss": 0.4463, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 7.705217391304348e-06, |
|
"loss": 0.4225, |
|
"step": 8425 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 7.697971014492755e-06, |
|
"loss": 0.4155, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 7.69072463768116e-06, |
|
"loss": 0.3319, |
|
"step": 8475 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 7.683478260869566e-06, |
|
"loss": 0.3675, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 7.676231884057971e-06, |
|
"loss": 0.4573, |
|
"step": 8525 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 7.668985507246376e-06, |
|
"loss": 0.3742, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 7.661739130434783e-06, |
|
"loss": 0.4487, |
|
"step": 8575 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 7.65449275362319e-06, |
|
"loss": 0.3993, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 7.647246376811595e-06, |
|
"loss": 0.3999, |
|
"step": 8625 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 7.640000000000001e-06, |
|
"loss": 0.4322, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 7.632753623188406e-06, |
|
"loss": 0.2906, |
|
"step": 8675 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 7.625507246376812e-06, |
|
"loss": 0.41, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 7.618260869565218e-06, |
|
"loss": 0.2934, |
|
"step": 8725 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 7.611014492753623e-06, |
|
"loss": 0.3668, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 7.60376811594203e-06, |
|
"loss": 0.3826, |
|
"step": 8775 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 7.596521739130436e-06, |
|
"loss": 0.4043, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 7.589275362318841e-06, |
|
"loss": 0.6248, |
|
"step": 8825 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 7.582028985507247e-06, |
|
"loss": 0.4668, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 7.574782608695653e-06, |
|
"loss": 0.4407, |
|
"step": 8875 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.567536231884058e-06, |
|
"loss": 0.3811, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.560289855072465e-06, |
|
"loss": 0.4717, |
|
"step": 8925 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.55304347826087e-06, |
|
"loss": 0.3531, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.545797101449276e-06, |
|
"loss": 0.4022, |
|
"step": 8975 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.5385507246376825e-06, |
|
"loss": 0.3987, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"eval_cer": 7.9417334841851535, |
|
"eval_loss": 0.41387349367141724, |
|
"eval_runtime": 3295.4388, |
|
"eval_samples_per_second": 0.753, |
|
"eval_steps_per_second": 0.377, |
|
"eval_wer": 20.842386125404992, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.5313043478260875e-06, |
|
"loss": 0.2728, |
|
"step": 9025 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.524057971014493e-06, |
|
"loss": 0.5125, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.516811594202899e-06, |
|
"loss": 0.3452, |
|
"step": 9075 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.509565217391305e-06, |
|
"loss": 0.3387, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.502318840579711e-06, |
|
"loss": 0.5818, |
|
"step": 9125 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.495072463768116e-06, |
|
"loss": 0.3494, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.487826086956523e-06, |
|
"loss": 0.2902, |
|
"step": 9175 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.4805797101449285e-06, |
|
"loss": 0.3691, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.4733333333333335e-06, |
|
"loss": 0.3344, |
|
"step": 9225 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.466086956521739e-06, |
|
"loss": 0.3612, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.458840579710146e-06, |
|
"loss": 0.4311, |
|
"step": 9275 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.451594202898551e-06, |
|
"loss": 0.4298, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.444347826086957e-06, |
|
"loss": 0.3657, |
|
"step": 9325 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.437101449275363e-06, |
|
"loss": 0.4382, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.429855072463769e-06, |
|
"loss": 0.3775, |
|
"step": 9375 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.4226086956521745e-06, |
|
"loss": 0.3566, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.4153623188405795e-06, |
|
"loss": 0.3423, |
|
"step": 9425 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.408115942028986e-06, |
|
"loss": 0.4855, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.400869565217392e-06, |
|
"loss": 0.3648, |
|
"step": 9475 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.393623188405797e-06, |
|
"loss": 0.3294, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.386376811594204e-06, |
|
"loss": 0.3991, |
|
"step": 9525 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.37913043478261e-06, |
|
"loss": 0.347, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.371884057971015e-06, |
|
"loss": 0.4011, |
|
"step": 9575 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.364637681159421e-06, |
|
"loss": 0.3433, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 7.357391304347826e-06, |
|
"loss": 0.4642, |
|
"step": 9625 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.350144927536232e-06, |
|
"loss": 0.3897, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.342898550724639e-06, |
|
"loss": 0.4562, |
|
"step": 9675 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.335652173913044e-06, |
|
"loss": 0.4942, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.32840579710145e-06, |
|
"loss": 0.3497, |
|
"step": 9725 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.3211594202898564e-06, |
|
"loss": 0.3713, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.3139130434782614e-06, |
|
"loss": 0.3558, |
|
"step": 9775 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.306666666666667e-06, |
|
"loss": 0.4417, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.299420289855072e-06, |
|
"loss": 0.4752, |
|
"step": 9825 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.292173913043479e-06, |
|
"loss": 0.3515, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.284927536231885e-06, |
|
"loss": 0.4731, |
|
"step": 9875 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.27768115942029e-06, |
|
"loss": 0.3111, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.2704347826086966e-06, |
|
"loss": 0.3776, |
|
"step": 9925 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 7.263188405797102e-06, |
|
"loss": 0.4209, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 7.2559420289855074e-06, |
|
"loss": 0.3707, |
|
"step": 9975 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 7.248695652173914e-06, |
|
"loss": 0.3373, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"eval_cer": 7.887131683801465, |
|
"eval_loss": 0.4116414785385132, |
|
"eval_runtime": 3350.806, |
|
"eval_samples_per_second": 0.741, |
|
"eval_steps_per_second": 0.37, |
|
"eval_wer": 21.214408233276156, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 7.241449275362319e-06, |
|
"loss": 0.3843, |
|
"step": 10025 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 7.234202898550725e-06, |
|
"loss": 0.3891, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 7.226956521739132e-06, |
|
"loss": 0.4058, |
|
"step": 10075 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 7.219710144927537e-06, |
|
"loss": 0.3908, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 7.2124637681159426e-06, |
|
"loss": 0.384, |
|
"step": 10125 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 7.205217391304348e-06, |
|
"loss": 0.4804, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 7.197971014492754e-06, |
|
"loss": 0.2675, |
|
"step": 10175 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 7.19072463768116e-06, |
|
"loss": 0.4003, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 7.183478260869566e-06, |
|
"loss": 0.4716, |
|
"step": 10225 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 7.176231884057972e-06, |
|
"loss": 0.3955, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 7.168985507246378e-06, |
|
"loss": 0.4133, |
|
"step": 10275 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 7.161739130434783e-06, |
|
"loss": 0.454, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 7.1544927536231885e-06, |
|
"loss": 0.4465, |
|
"step": 10325 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 7.147246376811595e-06, |
|
"loss": 0.3997, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 7.14e-06, |
|
"loss": 0.43, |
|
"step": 10375 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 7.132753623188406e-06, |
|
"loss": 0.373, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 7.125507246376813e-06, |
|
"loss": 0.4268, |
|
"step": 10425 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 7.118260869565218e-06, |
|
"loss": 0.331, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 7.111014492753624e-06, |
|
"loss": 0.4018, |
|
"step": 10475 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 7.103768115942029e-06, |
|
"loss": 0.3489, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 7.096521739130435e-06, |
|
"loss": 0.3914, |
|
"step": 10525 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 7.089275362318841e-06, |
|
"loss": 0.3368, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 7.082028985507246e-06, |
|
"loss": 0.3999, |
|
"step": 10575 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 7.074782608695653e-06, |
|
"loss": 0.334, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 7.067536231884059e-06, |
|
"loss": 0.4635, |
|
"step": 10625 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 7.060289855072464e-06, |
|
"loss": 0.475, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 7.0530434782608705e-06, |
|
"loss": 0.3621, |
|
"step": 10675 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 7.0457971014492755e-06, |
|
"loss": 0.4164, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 7.038550724637681e-06, |
|
"loss": 0.3648, |
|
"step": 10725 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 7.031304347826088e-06, |
|
"loss": 0.4358, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 7.024057971014493e-06, |
|
"loss": 0.3316, |
|
"step": 10775 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 7.016811594202899e-06, |
|
"loss": 0.4052, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 7.009565217391306e-06, |
|
"loss": 0.4474, |
|
"step": 10825 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 7.002318840579711e-06, |
|
"loss": 0.339, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 6.9950724637681165e-06, |
|
"loss": 0.3653, |
|
"step": 10875 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 6.9878260869565215e-06, |
|
"loss": 0.3337, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 6.980579710144928e-06, |
|
"loss": 0.3516, |
|
"step": 10925 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 6.973333333333334e-06, |
|
"loss": 0.3092, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 6.966086956521739e-06, |
|
"loss": 0.4312, |
|
"step": 10975 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 6.958840579710146e-06, |
|
"loss": 0.3808, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"eval_cer": 7.6264203846721434, |
|
"eval_loss": 0.4016057252883911, |
|
"eval_runtime": 3251.7215, |
|
"eval_samples_per_second": 0.763, |
|
"eval_steps_per_second": 0.382, |
|
"eval_wer": 20.509243377167905, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 6.951594202898552e-06, |
|
"loss": 0.6109, |
|
"step": 11025 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 6.944347826086957e-06, |
|
"loss": 0.4322, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 6.937101449275363e-06, |
|
"loss": 0.318, |
|
"step": 11075 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 6.929855072463769e-06, |
|
"loss": 0.3855, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 6.922608695652174e-06, |
|
"loss": 0.3817, |
|
"step": 11125 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 6.915362318840581e-06, |
|
"loss": 0.384, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 6.908115942028986e-06, |
|
"loss": 0.3451, |
|
"step": 11175 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 6.900869565217392e-06, |
|
"loss": 0.3518, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 6.8936231884057984e-06, |
|
"loss": 0.4567, |
|
"step": 11225 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 6.886666666666667e-06, |
|
"loss": 0.3848, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 6.879420289855073e-06, |
|
"loss": 0.4563, |
|
"step": 11275 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 6.872173913043478e-06, |
|
"loss": 0.3633, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 6.864927536231885e-06, |
|
"loss": 0.425, |
|
"step": 11325 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 6.8576811594202905e-06, |
|
"loss": 0.3508, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 6.8504347826086955e-06, |
|
"loss": 0.3351, |
|
"step": 11375 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 6.843188405797102e-06, |
|
"loss": 0.4659, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 6.835942028985508e-06, |
|
"loss": 0.346, |
|
"step": 11425 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 6.828695652173913e-06, |
|
"loss": 0.3896, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 6.82144927536232e-06, |
|
"loss": 0.4153, |
|
"step": 11475 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 6.814202898550725e-06, |
|
"loss": 0.3759, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 6.806956521739131e-06, |
|
"loss": 0.3328, |
|
"step": 11525 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 6.799710144927537e-06, |
|
"loss": 0.5639, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 6.792463768115942e-06, |
|
"loss": 0.3775, |
|
"step": 11575 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 6.785217391304348e-06, |
|
"loss": 0.4081, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 6.777971014492755e-06, |
|
"loss": 0.4266, |
|
"step": 11625 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 6.77072463768116e-06, |
|
"loss": 0.3683, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 6.763478260869566e-06, |
|
"loss": 0.4148, |
|
"step": 11675 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 6.756231884057971e-06, |
|
"loss": 0.3413, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 6.7489855072463775e-06, |
|
"loss": 0.3954, |
|
"step": 11725 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 6.741739130434783e-06, |
|
"loss": 0.3807, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 6.734492753623188e-06, |
|
"loss": 0.3475, |
|
"step": 11775 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 6.727246376811595e-06, |
|
"loss": 0.3428, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 6.720000000000001e-06, |
|
"loss": 0.3614, |
|
"step": 11825 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 6.712753623188406e-06, |
|
"loss": 0.3261, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 6.705507246376813e-06, |
|
"loss": 0.463, |
|
"step": 11875 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 6.6982608695652184e-06, |
|
"loss": 0.3045, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 6.6910144927536235e-06, |
|
"loss": 0.3677, |
|
"step": 11925 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 6.68376811594203e-06, |
|
"loss": 0.3405, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 6.676521739130435e-06, |
|
"loss": 0.3467, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 6.669275362318841e-06, |
|
"loss": 0.4214, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"eval_cer": 7.4152688277829695, |
|
"eval_loss": 0.39488446712493896, |
|
"eval_runtime": 3259.8037, |
|
"eval_samples_per_second": 0.761, |
|
"eval_steps_per_second": 0.381, |
|
"eval_wer": 20.093767867352774, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 6.662028985507248e-06, |
|
"loss": 0.3608, |
|
"step": 12025 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 6.654782608695653e-06, |
|
"loss": 0.3147, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 6.647536231884059e-06, |
|
"loss": 0.2764, |
|
"step": 12075 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 6.6402898550724644e-06, |
|
"loss": 0.3447, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 6.63304347826087e-06, |
|
"loss": 0.3727, |
|
"step": 12125 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 6.625797101449276e-06, |
|
"loss": 0.4042, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 6.618550724637681e-06, |
|
"loss": 0.3758, |
|
"step": 12175 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 6.611304347826088e-06, |
|
"loss": 0.3582, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 6.604057971014494e-06, |
|
"loss": 0.3571, |
|
"step": 12225 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 6.596811594202899e-06, |
|
"loss": 0.3676, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 6.5895652173913046e-06, |
|
"loss": 0.3224, |
|
"step": 12275 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 6.582318840579711e-06, |
|
"loss": 0.3025, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 6.575072463768116e-06, |
|
"loss": 0.3686, |
|
"step": 12325 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 6.567826086956522e-06, |
|
"loss": 0.3866, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 6.560579710144928e-06, |
|
"loss": 0.3662, |
|
"step": 12375 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 6.553333333333334e-06, |
|
"loss": 0.3507, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 6.54608695652174e-06, |
|
"loss": 0.3165, |
|
"step": 12425 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 6.538840579710145e-06, |
|
"loss": 0.3616, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 6.531594202898551e-06, |
|
"loss": 0.3466, |
|
"step": 12475 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 6.524347826086957e-06, |
|
"loss": 0.5202, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 6.517101449275362e-06, |
|
"loss": 0.3789, |
|
"step": 12525 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 6.509855072463769e-06, |
|
"loss": 0.5112, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 6.502608695652174e-06, |
|
"loss": 0.353, |
|
"step": 12575 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 6.49536231884058e-06, |
|
"loss": 0.3469, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 6.4881159420289865e-06, |
|
"loss": 0.4366, |
|
"step": 12625 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 6.4808695652173915e-06, |
|
"loss": 0.309, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 6.473623188405797e-06, |
|
"loss": 0.4358, |
|
"step": 12675 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 6.466376811594204e-06, |
|
"loss": 0.3286, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 6.459130434782609e-06, |
|
"loss": 0.2837, |
|
"step": 12725 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.451884057971015e-06, |
|
"loss": 0.3613, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.444637681159422e-06, |
|
"loss": 0.2637, |
|
"step": 12775 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.437391304347827e-06, |
|
"loss": 0.3415, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.4301449275362325e-06, |
|
"loss": 0.371, |
|
"step": 12825 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.4228985507246375e-06, |
|
"loss": 0.286, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.415652173913044e-06, |
|
"loss": 0.3646, |
|
"step": 12875 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.40840579710145e-06, |
|
"loss": 0.4352, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.401159420289855e-06, |
|
"loss": 0.2876, |
|
"step": 12925 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.393913043478262e-06, |
|
"loss": 0.3672, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.386666666666668e-06, |
|
"loss": 0.2878, |
|
"step": 12975 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.379420289855073e-06, |
|
"loss": 0.3029, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"eval_cer": 7.358084509813567, |
|
"eval_loss": 0.3902455270290375, |
|
"eval_runtime": 3231.646, |
|
"eval_samples_per_second": 0.768, |
|
"eval_steps_per_second": 0.384, |
|
"eval_wer": 19.7347055460263, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 6.372173913043479e-06, |
|
"loss": 0.3232, |
|
"step": 13025 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 6.364927536231884e-06, |
|
"loss": 0.3888, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 6.35768115942029e-06, |
|
"loss": 0.3928, |
|
"step": 13075 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 6.350434782608697e-06, |
|
"loss": 0.3615, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.343188405797102e-06, |
|
"loss": 0.4292, |
|
"step": 13125 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.335942028985508e-06, |
|
"loss": 0.3445, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.328695652173914e-06, |
|
"loss": 0.3983, |
|
"step": 13175 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 6.3214492753623195e-06, |
|
"loss": 0.3746, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 6.314202898550725e-06, |
|
"loss": 0.3714, |
|
"step": 13225 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 6.30695652173913e-06, |
|
"loss": 0.369, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 6.299710144927537e-06, |
|
"loss": 0.3991, |
|
"step": 13275 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 6.292463768115943e-06, |
|
"loss": 0.4264, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 6.285217391304348e-06, |
|
"loss": 0.2885, |
|
"step": 13325 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 6.277971014492754e-06, |
|
"loss": 0.3356, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 6.2707246376811605e-06, |
|
"loss": 0.3758, |
|
"step": 13375 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 6.2634782608695655e-06, |
|
"loss": 0.3424, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 6.256231884057971e-06, |
|
"loss": 0.3376, |
|
"step": 13425 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 6.248985507246378e-06, |
|
"loss": 0.3928, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 6.241739130434783e-06, |
|
"loss": 0.3247, |
|
"step": 13475 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 6.234492753623189e-06, |
|
"loss": 0.3002, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 6.227246376811594e-06, |
|
"loss": 0.2954, |
|
"step": 13525 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 6.220000000000001e-06, |
|
"loss": 0.3292, |
|
"step": 13550 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 6.213043478260871e-06, |
|
"loss": 0.3325, |
|
"step": 13575 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 6.205797101449276e-06, |
|
"loss": 0.3929, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 6.198550724637682e-06, |
|
"loss": 0.3311, |
|
"step": 13625 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 6.191304347826087e-06, |
|
"loss": 0.2803, |
|
"step": 13650 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 6.1840579710144935e-06, |
|
"loss": 0.3777, |
|
"step": 13675 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 6.176811594202899e-06, |
|
"loss": 0.387, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 6.169565217391304e-06, |
|
"loss": 0.3304, |
|
"step": 13725 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 6.162318840579711e-06, |
|
"loss": 0.4854, |
|
"step": 13750 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 6.155072463768117e-06, |
|
"loss": 0.3572, |
|
"step": 13775 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 6.147826086956522e-06, |
|
"loss": 0.3484, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 6.140579710144929e-06, |
|
"loss": 0.3671, |
|
"step": 13825 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 6.133333333333334e-06, |
|
"loss": 0.376, |
|
"step": 13850 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 6.1260869565217395e-06, |
|
"loss": 0.3435, |
|
"step": 13875 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 6.118840579710146e-06, |
|
"loss": 0.4171, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 6.111594202898551e-06, |
|
"loss": 0.3793, |
|
"step": 13925 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 6.104347826086957e-06, |
|
"loss": 0.4026, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 6.097101449275363e-06, |
|
"loss": 0.3086, |
|
"step": 13975 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 6.089855072463769e-06, |
|
"loss": 0.3549, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"eval_cer": 7.119509075704658, |
|
"eval_loss": 0.39083221554756165, |
|
"eval_runtime": 3180.5722, |
|
"eval_samples_per_second": 0.78, |
|
"eval_steps_per_second": 0.39, |
|
"eval_wer": 19.411473222793976, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 6.082608695652175e-06, |
|
"loss": 0.3889, |
|
"step": 14025 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 6.07536231884058e-06, |
|
"loss": 0.3197, |
|
"step": 14050 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 6.068115942028986e-06, |
|
"loss": 0.3279, |
|
"step": 14075 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 6.060869565217392e-06, |
|
"loss": 0.4178, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 6.053623188405797e-06, |
|
"loss": 0.3872, |
|
"step": 14125 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 6.046376811594203e-06, |
|
"loss": 0.3375, |
|
"step": 14150 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 6.03913043478261e-06, |
|
"loss": 0.4023, |
|
"step": 14175 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 6.031884057971015e-06, |
|
"loss": 0.4128, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 6.024637681159421e-06, |
|
"loss": 0.3603, |
|
"step": 14225 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 6.0173913043478264e-06, |
|
"loss": 0.3824, |
|
"step": 14250 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 6.010144927536232e-06, |
|
"loss": 0.4071, |
|
"step": 14275 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 6.002898550724638e-06, |
|
"loss": 0.2672, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 5.995652173913043e-06, |
|
"loss": 0.327, |
|
"step": 14325 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 5.98840579710145e-06, |
|
"loss": 0.3569, |
|
"step": 14350 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 5.981159420289856e-06, |
|
"loss": 0.3308, |
|
"step": 14375 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 5.973913043478261e-06, |
|
"loss": 0.3802, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 5.966666666666667e-06, |
|
"loss": 0.4012, |
|
"step": 14425 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 5.959420289855073e-06, |
|
"loss": 0.3191, |
|
"step": 14450 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 5.952173913043478e-06, |
|
"loss": 0.2991, |
|
"step": 14475 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 5.944927536231885e-06, |
|
"loss": 0.3889, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 5.93768115942029e-06, |
|
"loss": 0.3315, |
|
"step": 14525 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 5.930434782608696e-06, |
|
"loss": 0.3925, |
|
"step": 14550 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 5.9231884057971025e-06, |
|
"loss": 0.3661, |
|
"step": 14575 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 5.9159420289855076e-06, |
|
"loss": 0.4426, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 5.908695652173913e-06, |
|
"loss": 0.4046, |
|
"step": 14625 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 5.90144927536232e-06, |
|
"loss": 0.3361, |
|
"step": 14650 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 5.894202898550725e-06, |
|
"loss": 0.4759, |
|
"step": 14675 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 5.886956521739131e-06, |
|
"loss": 0.4326, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 5.879710144927536e-06, |
|
"loss": 0.2716, |
|
"step": 14725 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 5.872463768115943e-06, |
|
"loss": 0.3416, |
|
"step": 14750 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 5.8652173913043485e-06, |
|
"loss": 0.3715, |
|
"step": 14775 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 5.8579710144927535e-06, |
|
"loss": 0.3737, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 5.85072463768116e-06, |
|
"loss": 0.309, |
|
"step": 14825 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 5.843478260869566e-06, |
|
"loss": 0.3327, |
|
"step": 14850 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 5.836231884057971e-06, |
|
"loss": 0.401, |
|
"step": 14875 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 5.828985507246378e-06, |
|
"loss": 0.2827, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 5.821739130434783e-06, |
|
"loss": 0.3644, |
|
"step": 14925 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 5.814492753623189e-06, |
|
"loss": 0.3638, |
|
"step": 14950 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 5.807246376811595e-06, |
|
"loss": 0.4652, |
|
"step": 14975 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 5.8e-06, |
|
"loss": 0.3385, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"eval_cer": 7.779157853313001, |
|
"eval_loss": 0.39061304926872253, |
|
"eval_runtime": 3221.9453, |
|
"eval_samples_per_second": 0.77, |
|
"eval_steps_per_second": 0.385, |
|
"eval_wer": 20.20506956356013, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 5.792753623188406e-06, |
|
"loss": 0.4065, |
|
"step": 15025 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 5.785507246376812e-06, |
|
"loss": 0.4354, |
|
"step": 15050 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 5.778260869565218e-06, |
|
"loss": 0.3932, |
|
"step": 15075 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 5.771014492753624e-06, |
|
"loss": 0.3212, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 5.76376811594203e-06, |
|
"loss": 0.4244, |
|
"step": 15125 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 5.7565217391304355e-06, |
|
"loss": 0.306, |
|
"step": 15150 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 5.749275362318841e-06, |
|
"loss": 0.3492, |
|
"step": 15175 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 5.742028985507246e-06, |
|
"loss": 0.3573, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 5.734782608695652e-06, |
|
"loss": 0.3975, |
|
"step": 15225 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 5.727536231884059e-06, |
|
"loss": 0.3276, |
|
"step": 15250 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 5.720289855072464e-06, |
|
"loss": 0.3638, |
|
"step": 15275 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 5.71304347826087e-06, |
|
"loss": 0.3312, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 5.7057971014492765e-06, |
|
"loss": 0.3026, |
|
"step": 15325 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 5.6985507246376815e-06, |
|
"loss": 0.3493, |
|
"step": 15350 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 5.691304347826087e-06, |
|
"loss": 0.309, |
|
"step": 15375 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 5.684057971014492e-06, |
|
"loss": 0.2761, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 5.676811594202899e-06, |
|
"loss": 0.3507, |
|
"step": 15425 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 5.669565217391305e-06, |
|
"loss": 0.3359, |
|
"step": 15450 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 5.66231884057971e-06, |
|
"loss": 0.3132, |
|
"step": 15475 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 5.655072463768117e-06, |
|
"loss": 0.2976, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 5.6478260869565225e-06, |
|
"loss": 0.3893, |
|
"step": 15525 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 5.6405797101449275e-06, |
|
"loss": 0.3417, |
|
"step": 15550 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 5.633333333333334e-06, |
|
"loss": 0.3462, |
|
"step": 15575 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 5.626086956521739e-06, |
|
"loss": 0.3885, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 5.618840579710145e-06, |
|
"loss": 0.3844, |
|
"step": 15625 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 5.611594202898552e-06, |
|
"loss": 0.3112, |
|
"step": 15650 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 5.604347826086957e-06, |
|
"loss": 0.3266, |
|
"step": 15675 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 5.597101449275363e-06, |
|
"loss": 0.3538, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 5.589855072463769e-06, |
|
"loss": 0.351, |
|
"step": 15725 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 5.582608695652174e-06, |
|
"loss": 0.3373, |
|
"step": 15750 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 5.57536231884058e-06, |
|
"loss": 0.4156, |
|
"step": 15775 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 5.568115942028985e-06, |
|
"loss": 0.4124, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 5.560869565217392e-06, |
|
"loss": 0.3716, |
|
"step": 15825 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 5.553623188405798e-06, |
|
"loss": 0.5276, |
|
"step": 15850 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 5.546376811594203e-06, |
|
"loss": 0.3412, |
|
"step": 15875 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 5.5391304347826094e-06, |
|
"loss": 0.3293, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 5.531884057971015e-06, |
|
"loss": 0.3759, |
|
"step": 15925 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 5.52463768115942e-06, |
|
"loss": 0.3926, |
|
"step": 15950 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 5.517391304347827e-06, |
|
"loss": 0.3686, |
|
"step": 15975 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 5.510144927536233e-06, |
|
"loss": 0.3282, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"eval_cer": 7.108072212110778, |
|
"eval_loss": 0.39232152700424194, |
|
"eval_runtime": 3154.8615, |
|
"eval_samples_per_second": 0.787, |
|
"eval_steps_per_second": 0.393, |
|
"eval_wer": 19.265103868877453, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 5.502898550724638e-06, |
|
"loss": 0.3279, |
|
"step": 16025 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 5.4956521739130446e-06, |
|
"loss": 0.4442, |
|
"step": 16050 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 5.4884057971014496e-06, |
|
"loss": 0.4354, |
|
"step": 16075 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 5.481159420289855e-06, |
|
"loss": 0.362, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 5.473913043478261e-06, |
|
"loss": 0.3849, |
|
"step": 16125 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 5.466666666666667e-06, |
|
"loss": 0.3956, |
|
"step": 16150 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 5.459420289855073e-06, |
|
"loss": 0.2881, |
|
"step": 16175 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 5.452173913043479e-06, |
|
"loss": 0.3321, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 5.444927536231885e-06, |
|
"loss": 0.3512, |
|
"step": 16225 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 5.4376811594202905e-06, |
|
"loss": 0.3227, |
|
"step": 16250 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 5.4304347826086955e-06, |
|
"loss": 0.3204, |
|
"step": 16275 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 5.423188405797101e-06, |
|
"loss": 0.2884, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 5.415942028985508e-06, |
|
"loss": 0.3259, |
|
"step": 16325 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 5.408695652173913e-06, |
|
"loss": 0.3611, |
|
"step": 16350 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 5.401449275362319e-06, |
|
"loss": 0.3598, |
|
"step": 16375 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 5.394202898550726e-06, |
|
"loss": 0.3911, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 5.386956521739131e-06, |
|
"loss": 0.4117, |
|
"step": 16425 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 5.3797101449275365e-06, |
|
"loss": 0.4095, |
|
"step": 16450 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 5.3724637681159415e-06, |
|
"loss": 0.3033, |
|
"step": 16475 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 5.365217391304348e-06, |
|
"loss": 0.3534, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 5.357971014492754e-06, |
|
"loss": 0.2906, |
|
"step": 16525 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 5.350724637681159e-06, |
|
"loss": 0.2948, |
|
"step": 16550 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 5.343478260869566e-06, |
|
"loss": 0.2731, |
|
"step": 16575 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 5.336231884057972e-06, |
|
"loss": 0.3218, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 5.328985507246377e-06, |
|
"loss": 0.328, |
|
"step": 16625 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 5.321739130434783e-06, |
|
"loss": 0.274, |
|
"step": 16650 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 5.314492753623189e-06, |
|
"loss": 0.2936, |
|
"step": 16675 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 5.307246376811594e-06, |
|
"loss": 0.3601, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 5.3002898550724646e-06, |
|
"loss": 0.4228, |
|
"step": 16725 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 5.2930434782608696e-06, |
|
"loss": 0.3024, |
|
"step": 16750 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 5.285797101449276e-06, |
|
"loss": 0.3463, |
|
"step": 16775 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 5.278550724637682e-06, |
|
"loss": 0.3282, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 5.271304347826087e-06, |
|
"loss": 0.3138, |
|
"step": 16825 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 5.264057971014494e-06, |
|
"loss": 0.3608, |
|
"step": 16850 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 5.256811594202899e-06, |
|
"loss": 0.2716, |
|
"step": 16875 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 5.249565217391305e-06, |
|
"loss": 0.4661, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 5.242318840579711e-06, |
|
"loss": 0.2482, |
|
"step": 16925 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 5.235072463768116e-06, |
|
"loss": 0.34, |
|
"step": 16950 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 5.227826086956522e-06, |
|
"loss": 0.3228, |
|
"step": 16975 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 5.220579710144928e-06, |
|
"loss": 0.3196, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"eval_cer": 7.224900388607408, |
|
"eval_loss": 0.39229413866996765, |
|
"eval_runtime": 3162.0207, |
|
"eval_samples_per_second": 0.785, |
|
"eval_steps_per_second": 0.392, |
|
"eval_wer": 19.335239184295787, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 5.213333333333334e-06, |
|
"loss": 0.2583, |
|
"step": 17025 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 5.20608695652174e-06, |
|
"loss": 0.248, |
|
"step": 17050 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 5.198840579710145e-06, |
|
"loss": 0.2901, |
|
"step": 17075 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 5.191594202898551e-06, |
|
"loss": 0.2492, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 5.184347826086957e-06, |
|
"loss": 0.3115, |
|
"step": 17125 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 5.177101449275362e-06, |
|
"loss": 0.3328, |
|
"step": 17150 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 5.169855072463768e-06, |
|
"loss": 0.2782, |
|
"step": 17175 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 5.162608695652175e-06, |
|
"loss": 0.2778, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 5.15536231884058e-06, |
|
"loss": 0.2721, |
|
"step": 17225 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 5.148115942028986e-06, |
|
"loss": 0.361, |
|
"step": 17250 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 5.140869565217391e-06, |
|
"loss": 0.3612, |
|
"step": 17275 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 5.1336231884057975e-06, |
|
"loss": 0.2824, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 5.126376811594203e-06, |
|
"loss": 0.2909, |
|
"step": 17325 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 5.119130434782608e-06, |
|
"loss": 0.2773, |
|
"step": 17350 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 5.111884057971015e-06, |
|
"loss": 0.3971, |
|
"step": 17375 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 5.104637681159421e-06, |
|
"loss": 0.2334, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 5.097391304347826e-06, |
|
"loss": 0.3899, |
|
"step": 17425 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 5.090144927536233e-06, |
|
"loss": 0.321, |
|
"step": 17450 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 5.082898550724638e-06, |
|
"loss": 0.2295, |
|
"step": 17475 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 5.0756521739130435e-06, |
|
"loss": 0.278, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 5.06840579710145e-06, |
|
"loss": 0.2696, |
|
"step": 17525 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 5.061159420289855e-06, |
|
"loss": 0.2604, |
|
"step": 17550 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 5.053913043478261e-06, |
|
"loss": 0.3015, |
|
"step": 17575 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 5.046666666666668e-06, |
|
"loss": 0.1998, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 5.039420289855073e-06, |
|
"loss": 0.3426, |
|
"step": 17625 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 5.032173913043479e-06, |
|
"loss": 0.3037, |
|
"step": 17650 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 5.024927536231885e-06, |
|
"loss": 0.3063, |
|
"step": 17675 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 5.01768115942029e-06, |
|
"loss": 0.2932, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 5.010434782608696e-06, |
|
"loss": 0.3028, |
|
"step": 17725 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 5.003188405797101e-06, |
|
"loss": 0.2644, |
|
"step": 17750 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 4.995942028985508e-06, |
|
"loss": 0.2902, |
|
"step": 17775 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 4.988695652173913e-06, |
|
"loss": 0.2816, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 4.98144927536232e-06, |
|
"loss": 0.3323, |
|
"step": 17825 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 4.9742028985507255e-06, |
|
"loss": 0.2923, |
|
"step": 17850 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 4.9669565217391305e-06, |
|
"loss": 0.3281, |
|
"step": 17875 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 4.959710144927536e-06, |
|
"loss": 0.3877, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 4.952463768115943e-06, |
|
"loss": 0.372, |
|
"step": 17925 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 4.945217391304348e-06, |
|
"loss": 0.2854, |
|
"step": 17950 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 4.937971014492754e-06, |
|
"loss": 0.3001, |
|
"step": 17975 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 4.93072463768116e-06, |
|
"loss": 0.3251, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"eval_cer": 7.176078508534606, |
|
"eval_loss": 0.39806175231933594, |
|
"eval_runtime": 3134.0346, |
|
"eval_samples_per_second": 0.792, |
|
"eval_steps_per_second": 0.396, |
|
"eval_wer": 19.483133218982278, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 4.923478260869566e-06, |
|
"loss": 0.3021, |
|
"step": 18025 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 4.9162318840579714e-06, |
|
"loss": 0.2846, |
|
"step": 18050 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 4.908985507246377e-06, |
|
"loss": 0.284, |
|
"step": 18075 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 4.901739130434783e-06, |
|
"loss": 0.2423, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 4.894492753623189e-06, |
|
"loss": 0.2551, |
|
"step": 18125 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 4.887246376811595e-06, |
|
"loss": 0.4317, |
|
"step": 18150 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 4.880000000000001e-06, |
|
"loss": 0.4729, |
|
"step": 18175 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 4.872753623188406e-06, |
|
"loss": 0.2887, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 4.865507246376812e-06, |
|
"loss": 0.4263, |
|
"step": 18225 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 4.8582608695652174e-06, |
|
"loss": 0.3566, |
|
"step": 18250 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 4.851014492753623e-06, |
|
"loss": 0.3274, |
|
"step": 18275 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 4.843768115942029e-06, |
|
"loss": 0.2818, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 4.836521739130435e-06, |
|
"loss": 0.29, |
|
"step": 18325 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 4.829275362318841e-06, |
|
"loss": 0.3009, |
|
"step": 18350 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 4.822028985507247e-06, |
|
"loss": 0.3663, |
|
"step": 18375 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 4.8147826086956526e-06, |
|
"loss": 0.2181, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 4.807536231884058e-06, |
|
"loss": 0.4402, |
|
"step": 18425 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 4.800289855072464e-06, |
|
"loss": 0.3224, |
|
"step": 18450 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 4.79304347826087e-06, |
|
"loss": 0.4503, |
|
"step": 18475 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 4.785797101449276e-06, |
|
"loss": 0.3597, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 4.778550724637682e-06, |
|
"loss": 0.3584, |
|
"step": 18525 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 4.771304347826088e-06, |
|
"loss": 0.3391, |
|
"step": 18550 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 4.764057971014493e-06, |
|
"loss": 0.3192, |
|
"step": 18575 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 4.756811594202899e-06, |
|
"loss": 0.3671, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 4.749565217391305e-06, |
|
"loss": 0.3506, |
|
"step": 18625 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 4.74231884057971e-06, |
|
"loss": 0.3515, |
|
"step": 18650 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 4.735072463768116e-06, |
|
"loss": 0.2615, |
|
"step": 18675 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 4.727826086956522e-06, |
|
"loss": 0.2509, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 4.720579710144928e-06, |
|
"loss": 0.3, |
|
"step": 18725 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 4.713333333333334e-06, |
|
"loss": 0.302, |
|
"step": 18750 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 4.7060869565217395e-06, |
|
"loss": 0.2971, |
|
"step": 18775 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 4.698840579710145e-06, |
|
"loss": 0.4138, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 4.691594202898551e-06, |
|
"loss": 0.296, |
|
"step": 18825 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 4.684347826086957e-06, |
|
"loss": 0.2533, |
|
"step": 18850 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 4.677101449275362e-06, |
|
"loss": 0.2477, |
|
"step": 18875 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 4.669855072463769e-06, |
|
"loss": 0.2952, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 4.662608695652175e-06, |
|
"loss": 0.2471, |
|
"step": 18925 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 4.65536231884058e-06, |
|
"loss": 0.2664, |
|
"step": 18950 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 4.6481159420289855e-06, |
|
"loss": 0.2127, |
|
"step": 18975 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 4.640869565217392e-06, |
|
"loss": 0.4162, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"eval_cer": 7.059004377982194, |
|
"eval_loss": 0.39584028720855713, |
|
"eval_runtime": 3153.6429, |
|
"eval_samples_per_second": 0.787, |
|
"eval_steps_per_second": 0.394, |
|
"eval_wer": 19.057747284162378, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 4.633623188405797e-06, |
|
"loss": 0.286, |
|
"step": 19025 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 4.626376811594203e-06, |
|
"loss": 0.2912, |
|
"step": 19050 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 4.619130434782609e-06, |
|
"loss": 0.2905, |
|
"step": 19075 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 4.611884057971015e-06, |
|
"loss": 0.333, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 4.604637681159421e-06, |
|
"loss": 0.2647, |
|
"step": 19125 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 4.5973913043478265e-06, |
|
"loss": 0.2746, |
|
"step": 19150 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 4.590144927536232e-06, |
|
"loss": 0.3317, |
|
"step": 19175 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 4.582898550724638e-06, |
|
"loss": 0.2549, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 4.575652173913044e-06, |
|
"loss": 0.312, |
|
"step": 19225 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 4.56840579710145e-06, |
|
"loss": 0.3177, |
|
"step": 19250 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 4.561159420289856e-06, |
|
"loss": 0.3063, |
|
"step": 19275 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 4.553913043478262e-06, |
|
"loss": 0.3964, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 4.546666666666667e-06, |
|
"loss": 0.3377, |
|
"step": 19325 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 4.5394202898550725e-06, |
|
"loss": 0.2753, |
|
"step": 19350 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 4.532173913043479e-06, |
|
"loss": 0.2735, |
|
"step": 19375 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 4.524927536231884e-06, |
|
"loss": 0.292, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 4.51768115942029e-06, |
|
"loss": 0.2894, |
|
"step": 19425 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 4.510434782608696e-06, |
|
"loss": 0.367, |
|
"step": 19450 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 4.503188405797102e-06, |
|
"loss": 0.2483, |
|
"step": 19475 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 4.495942028985508e-06, |
|
"loss": 0.4151, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 4.4886956521739134e-06, |
|
"loss": 0.269, |
|
"step": 19525 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 4.481449275362319e-06, |
|
"loss": 0.3127, |
|
"step": 19550 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 4.474202898550725e-06, |
|
"loss": 0.2615, |
|
"step": 19575 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 4.466956521739131e-06, |
|
"loss": 0.2606, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 4.459710144927537e-06, |
|
"loss": 0.2855, |
|
"step": 19625 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 4.452463768115942e-06, |
|
"loss": 0.2419, |
|
"step": 19650 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 4.4452173913043486e-06, |
|
"loss": 0.3292, |
|
"step": 19675 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 4.4379710144927544e-06, |
|
"loss": 0.2821, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 4.4307246376811594e-06, |
|
"loss": 0.3189, |
|
"step": 19725 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 4.423478260869565e-06, |
|
"loss": 0.286, |
|
"step": 19750 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 4.416231884057971e-06, |
|
"loss": 0.2853, |
|
"step": 19775 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 4.408985507246377e-06, |
|
"loss": 0.2835, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 4.401739130434783e-06, |
|
"loss": 0.2626, |
|
"step": 19825 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.394492753623189e-06, |
|
"loss": 0.3301, |
|
"step": 19850 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.3872463768115946e-06, |
|
"loss": 0.3094, |
|
"step": 19875 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.380289855072464e-06, |
|
"loss": 0.2322, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.37304347826087e-06, |
|
"loss": 0.3167, |
|
"step": 19925 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 4.365797101449276e-06, |
|
"loss": 0.3029, |
|
"step": 19950 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 4.358550724637682e-06, |
|
"loss": 0.2892, |
|
"step": 19975 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 4.3513043478260875e-06, |
|
"loss": 0.2851, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"eval_cer": 7.0167002803876235, |
|
"eval_loss": 0.39532455801963806, |
|
"eval_runtime": 3151.9912, |
|
"eval_samples_per_second": 0.787, |
|
"eval_steps_per_second": 0.394, |
|
"eval_wer": 19.209453020773775, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 4.344057971014493e-06, |
|
"loss": 0.2749, |
|
"step": 20025 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 4.336811594202899e-06, |
|
"loss": 0.3344, |
|
"step": 20050 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 4.329565217391305e-06, |
|
"loss": 0.2773, |
|
"step": 20075 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 4.322318840579711e-06, |
|
"loss": 0.2361, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 4.315072463768116e-06, |
|
"loss": 0.2254, |
|
"step": 20125 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 4.307826086956522e-06, |
|
"loss": 0.2396, |
|
"step": 20150 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 4.3005797101449284e-06, |
|
"loss": 0.2621, |
|
"step": 20175 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 4.2933333333333334e-06, |
|
"loss": 0.31, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 4.286086956521739e-06, |
|
"loss": 0.3039, |
|
"step": 20225 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 4.278840579710145e-06, |
|
"loss": 0.2772, |
|
"step": 20250 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 4.271594202898551e-06, |
|
"loss": 0.2535, |
|
"step": 20275 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 4.264347826086957e-06, |
|
"loss": 0.2903, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 4.257101449275363e-06, |
|
"loss": 0.3347, |
|
"step": 20325 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 4.2498550724637686e-06, |
|
"loss": 0.3133, |
|
"step": 20350 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 4.2426086956521744e-06, |
|
"loss": 0.2631, |
|
"step": 20375 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 4.23536231884058e-06, |
|
"loss": 0.2973, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 4.228115942028986e-06, |
|
"loss": 0.2731, |
|
"step": 20425 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 4.220869565217391e-06, |
|
"loss": 0.3088, |
|
"step": 20450 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 4.213623188405798e-06, |
|
"loss": 0.3291, |
|
"step": 20475 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 4.206376811594204e-06, |
|
"loss": 0.2447, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 4.199130434782609e-06, |
|
"loss": 0.3862, |
|
"step": 20525 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 4.1918840579710146e-06, |
|
"loss": 0.3494, |
|
"step": 20550 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 4.18463768115942e-06, |
|
"loss": 0.314, |
|
"step": 20575 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 4.177391304347826e-06, |
|
"loss": 0.3158, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 4.170144927536232e-06, |
|
"loss": 0.2839, |
|
"step": 20625 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 4.162898550724638e-06, |
|
"loss": 0.2518, |
|
"step": 20650 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 4.155652173913044e-06, |
|
"loss": 0.3692, |
|
"step": 20675 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 4.14840579710145e-06, |
|
"loss": 0.2712, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 4.1411594202898555e-06, |
|
"loss": 0.2804, |
|
"step": 20725 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 4.1339130434782605e-06, |
|
"loss": 0.3514, |
|
"step": 20750 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 4.126666666666667e-06, |
|
"loss": 0.2892, |
|
"step": 20775 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 4.119420289855073e-06, |
|
"loss": 0.2898, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 4.112173913043478e-06, |
|
"loss": 0.3071, |
|
"step": 20825 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 4.104927536231885e-06, |
|
"loss": 0.2719, |
|
"step": 20850 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 4.097681159420291e-06, |
|
"loss": 0.3628, |
|
"step": 20875 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 4.090434782608696e-06, |
|
"loss": 0.3201, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 4.0831884057971015e-06, |
|
"loss": 0.3166, |
|
"step": 20925 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 4.075942028985508e-06, |
|
"loss": 0.3418, |
|
"step": 20950 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 4.068695652173913e-06, |
|
"loss": 0.288, |
|
"step": 20975 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 4.061449275362319e-06, |
|
"loss": 0.2982, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"eval_cer": 6.842564808893699, |
|
"eval_loss": 0.39293238520622253, |
|
"eval_runtime": 3128.5672, |
|
"eval_samples_per_second": 0.793, |
|
"eval_steps_per_second": 0.397, |
|
"eval_wer": 18.80998665904326, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 4.054202898550725e-06, |
|
"loss": 0.2191, |
|
"step": 21025 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 4.046956521739131e-06, |
|
"loss": 0.3047, |
|
"step": 21050 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 4.039710144927537e-06, |
|
"loss": 0.2752, |
|
"step": 21075 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 4.0324637681159425e-06, |
|
"loss": 0.2675, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 4.025217391304348e-06, |
|
"loss": 0.2971, |
|
"step": 21125 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 4.017971014492754e-06, |
|
"loss": 0.3199, |
|
"step": 21150 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 4.01072463768116e-06, |
|
"loss": 0.2402, |
|
"step": 21175 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 4.003478260869565e-06, |
|
"loss": 0.2752, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 3.996231884057971e-06, |
|
"loss": 0.3148, |
|
"step": 21225 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 3.988985507246378e-06, |
|
"loss": 0.2471, |
|
"step": 21250 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 3.981739130434783e-06, |
|
"loss": 0.2807, |
|
"step": 21275 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 3.9744927536231885e-06, |
|
"loss": 0.2242, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 3.967246376811594e-06, |
|
"loss": 0.4252, |
|
"step": 21325 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 3.96e-06, |
|
"loss": 0.2453, |
|
"step": 21350 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 3.952753623188406e-06, |
|
"loss": 0.2282, |
|
"step": 21375 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 3.945507246376812e-06, |
|
"loss": 0.353, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 3.938260869565218e-06, |
|
"loss": 0.3021, |
|
"step": 21425 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 3.931014492753624e-06, |
|
"loss": 0.2803, |
|
"step": 21450 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 3.9237681159420295e-06, |
|
"loss": 0.2523, |
|
"step": 21475 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 3.916521739130435e-06, |
|
"loss": 0.2748, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 3.90927536231884e-06, |
|
"loss": 0.3106, |
|
"step": 21525 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 3.902028985507247e-06, |
|
"loss": 0.3489, |
|
"step": 21550 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 3.894782608695653e-06, |
|
"loss": 0.3188, |
|
"step": 21575 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 3.887536231884058e-06, |
|
"loss": 0.3188, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 3.880289855072465e-06, |
|
"loss": 0.3178, |
|
"step": 21625 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 3.87304347826087e-06, |
|
"loss": 0.2279, |
|
"step": 21650 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 3.8657971014492755e-06, |
|
"loss": 0.2803, |
|
"step": 21675 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 3.858550724637681e-06, |
|
"loss": 0.21, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 3.851304347826087e-06, |
|
"loss": 0.2753, |
|
"step": 21725 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 3.844057971014493e-06, |
|
"loss": 0.4256, |
|
"step": 21750 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 3.836811594202899e-06, |
|
"loss": 0.2934, |
|
"step": 21775 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 3.829565217391305e-06, |
|
"loss": 0.3364, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 3.822318840579711e-06, |
|
"loss": 0.3053, |
|
"step": 21825 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 3.8150724637681164e-06, |
|
"loss": 0.3143, |
|
"step": 21850 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 3.807826086956522e-06, |
|
"loss": 0.2758, |
|
"step": 21875 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 3.8005797101449277e-06, |
|
"loss": 0.2616, |
|
"step": 21900 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 3.793333333333334e-06, |
|
"loss": 0.2949, |
|
"step": 21925 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 3.7860869565217394e-06, |
|
"loss": 0.358, |
|
"step": 21950 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 3.7788405797101453e-06, |
|
"loss": 0.2856, |
|
"step": 21975 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 3.7715942028985507e-06, |
|
"loss": 0.3642, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"eval_cer": 6.886713561906636, |
|
"eval_loss": 0.3953614830970764, |
|
"eval_runtime": 3140.8887, |
|
"eval_samples_per_second": 0.79, |
|
"eval_steps_per_second": 0.395, |
|
"eval_wer": 18.69716028206594, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 3.764347826086957e-06, |
|
"loss": 0.283, |
|
"step": 22025 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 3.757101449275363e-06, |
|
"loss": 0.3267, |
|
"step": 22050 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 3.7498550724637683e-06, |
|
"loss": 0.3271, |
|
"step": 22075 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 3.742608695652174e-06, |
|
"loss": 0.3372, |
|
"step": 22100 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 3.7353623188405804e-06, |
|
"loss": 0.2551, |
|
"step": 22125 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 3.728115942028986e-06, |
|
"loss": 0.3063, |
|
"step": 22150 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 3.7208695652173917e-06, |
|
"loss": 0.2902, |
|
"step": 22175 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 3.713623188405797e-06, |
|
"loss": 0.255, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 3.7063768115942034e-06, |
|
"loss": 0.3041, |
|
"step": 22225 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 3.699130434782609e-06, |
|
"loss": 0.3338, |
|
"step": 22250 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 3.6918840579710147e-06, |
|
"loss": 0.2509, |
|
"step": 22275 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 3.68463768115942e-06, |
|
"loss": 0.2996, |
|
"step": 22300 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 3.6773913043478264e-06, |
|
"loss": 0.3299, |
|
"step": 22325 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 3.6701449275362323e-06, |
|
"loss": 0.234, |
|
"step": 22350 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 3.6628985507246377e-06, |
|
"loss": 0.304, |
|
"step": 22375 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 3.6556521739130435e-06, |
|
"loss": 0.3406, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 3.64840579710145e-06, |
|
"loss": 0.2437, |
|
"step": 22425 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 3.6411594202898552e-06, |
|
"loss": 0.4036, |
|
"step": 22450 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 3.633913043478261e-06, |
|
"loss": 0.431, |
|
"step": 22475 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 3.6266666666666674e-06, |
|
"loss": 0.2864, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 3.619420289855073e-06, |
|
"loss": 0.2791, |
|
"step": 22525 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 3.6121739130434787e-06, |
|
"loss": 0.2975, |
|
"step": 22550 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 3.604927536231884e-06, |
|
"loss": 0.3252, |
|
"step": 22575 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 3.5976811594202904e-06, |
|
"loss": 0.2389, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 3.5904347826086962e-06, |
|
"loss": 0.3063, |
|
"step": 22625 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 3.5831884057971017e-06, |
|
"loss": 0.2614, |
|
"step": 22650 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 3.5759420289855075e-06, |
|
"loss": 0.2833, |
|
"step": 22675 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 3.5686956521739134e-06, |
|
"loss": 0.3931, |
|
"step": 22700 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 3.5614492753623192e-06, |
|
"loss": 0.2804, |
|
"step": 22725 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 3.554202898550725e-06, |
|
"loss": 0.2688, |
|
"step": 22750 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 3.5469565217391305e-06, |
|
"loss": 0.2059, |
|
"step": 22775 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 3.5397101449275368e-06, |
|
"loss": 0.3133, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 3.5324637681159422e-06, |
|
"loss": 0.2725, |
|
"step": 22825 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 3.525217391304348e-06, |
|
"loss": 0.2712, |
|
"step": 22850 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 3.5179710144927535e-06, |
|
"loss": 0.264, |
|
"step": 22875 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 3.5107246376811598e-06, |
|
"loss": 0.2404, |
|
"step": 22900 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 3.5034782608695656e-06, |
|
"loss": 0.3211, |
|
"step": 22925 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 3.496231884057971e-06, |
|
"loss": 0.2726, |
|
"step": 22950 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 3.488985507246377e-06, |
|
"loss": 0.399, |
|
"step": 22975 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 3.481739130434783e-06, |
|
"loss": 0.2297, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"eval_cer": 6.938363913620936, |
|
"eval_loss": 0.3916235566139221, |
|
"eval_runtime": 3180.568, |
|
"eval_samples_per_second": 0.78, |
|
"eval_steps_per_second": 0.39, |
|
"eval_wer": 18.73299028016009, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 3.4744927536231886e-06, |
|
"loss": 0.3484, |
|
"step": 23025 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 3.4672463768115945e-06, |
|
"loss": 0.2585, |
|
"step": 23050 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 3.46e-06, |
|
"loss": 0.2843, |
|
"step": 23075 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 3.452753623188406e-06, |
|
"loss": 0.3526, |
|
"step": 23100 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 3.445507246376812e-06, |
|
"loss": 0.3063, |
|
"step": 23125 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 3.4382608695652175e-06, |
|
"loss": 0.3701, |
|
"step": 23150 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 3.4310144927536233e-06, |
|
"loss": 0.2438, |
|
"step": 23175 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 3.4237681159420296e-06, |
|
"loss": 0.3001, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 3.416521739130435e-06, |
|
"loss": 0.3103, |
|
"step": 23225 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 3.409275362318841e-06, |
|
"loss": 0.2769, |
|
"step": 23250 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 3.4020289855072467e-06, |
|
"loss": 0.4483, |
|
"step": 23275 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 3.3947826086956526e-06, |
|
"loss": 0.2777, |
|
"step": 23300 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 3.387536231884058e-06, |
|
"loss": 0.2413, |
|
"step": 23325 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 3.380289855072464e-06, |
|
"loss": 0.2911, |
|
"step": 23350 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 3.37304347826087e-06, |
|
"loss": 0.3544, |
|
"step": 23375 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 3.3657971014492756e-06, |
|
"loss": 0.349, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 3.3585507246376814e-06, |
|
"loss": 0.3243, |
|
"step": 23425 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 3.351304347826087e-06, |
|
"loss": 0.2601, |
|
"step": 23450 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 3.344057971014493e-06, |
|
"loss": 0.268, |
|
"step": 23475 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 3.336811594202899e-06, |
|
"loss": 0.294, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 3.3295652173913044e-06, |
|
"loss": 0.3095, |
|
"step": 23525 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 3.3223188405797103e-06, |
|
"loss": 0.2588, |
|
"step": 23550 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 3.3150724637681166e-06, |
|
"loss": 0.4027, |
|
"step": 23575 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 3.307826086956522e-06, |
|
"loss": 0.2887, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 3.300579710144928e-06, |
|
"loss": 0.2459, |
|
"step": 23625 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 3.2933333333333333e-06, |
|
"loss": 0.2288, |
|
"step": 23650 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 3.2860869565217396e-06, |
|
"loss": 0.2998, |
|
"step": 23675 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 3.2788405797101454e-06, |
|
"loss": 0.3197, |
|
"step": 23700 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 3.271594202898551e-06, |
|
"loss": 0.2931, |
|
"step": 23725 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 3.2643478260869567e-06, |
|
"loss": 0.2451, |
|
"step": 23750 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 3.2571014492753626e-06, |
|
"loss": 0.251, |
|
"step": 23775 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 3.2498550724637684e-06, |
|
"loss": 0.291, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 3.2426086956521743e-06, |
|
"loss": 0.2916, |
|
"step": 23825 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 3.2353623188405797e-06, |
|
"loss": 0.3365, |
|
"step": 23850 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 3.228115942028986e-06, |
|
"loss": 0.2967, |
|
"step": 23875 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 3.2208695652173914e-06, |
|
"loss": 0.2435, |
|
"step": 23900 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 3.2136231884057973e-06, |
|
"loss": 0.3885, |
|
"step": 23925 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 3.2063768115942027e-06, |
|
"loss": 0.2373, |
|
"step": 23950 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 3.199130434782609e-06, |
|
"loss": 0.3457, |
|
"step": 23975 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 3.191884057971015e-06, |
|
"loss": 0.2313, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"eval_cer": 6.778493777362389, |
|
"eval_loss": 0.3930225968360901, |
|
"eval_runtime": 3191.3103, |
|
"eval_samples_per_second": 0.778, |
|
"eval_steps_per_second": 0.389, |
|
"eval_wer": 18.60339241471317, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 3.1846376811594202e-06, |
|
"loss": 0.2984, |
|
"step": 24025 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 3.1773913043478265e-06, |
|
"loss": 0.2425, |
|
"step": 24050 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 3.1701449275362324e-06, |
|
"loss": 0.3009, |
|
"step": 24075 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 3.162898550724638e-06, |
|
"loss": 0.251, |
|
"step": 24100 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 3.1556521739130437e-06, |
|
"loss": 0.2794, |
|
"step": 24125 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 3.14840579710145e-06, |
|
"loss": 0.326, |
|
"step": 24150 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 3.1414492753623194e-06, |
|
"loss": 0.2487, |
|
"step": 24175 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 3.134202898550725e-06, |
|
"loss": 0.2793, |
|
"step": 24200 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 3.1269565217391307e-06, |
|
"loss": 0.3093, |
|
"step": 24225 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 3.119710144927536e-06, |
|
"loss": 0.2536, |
|
"step": 24250 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 3.1124637681159424e-06, |
|
"loss": 0.3143, |
|
"step": 24275 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 3.1052173913043483e-06, |
|
"loss": 0.2409, |
|
"step": 24300 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 3.0979710144927537e-06, |
|
"loss": 0.3879, |
|
"step": 24325 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 3.0907246376811596e-06, |
|
"loss": 0.3521, |
|
"step": 24350 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 3.083478260869566e-06, |
|
"loss": 0.2628, |
|
"step": 24375 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 3.0762318840579713e-06, |
|
"loss": 0.2127, |
|
"step": 24400 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 3.068985507246377e-06, |
|
"loss": 0.3071, |
|
"step": 24425 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 3.0617391304347826e-06, |
|
"loss": 0.2488, |
|
"step": 24450 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 3.054492753623189e-06, |
|
"loss": 0.2462, |
|
"step": 24475 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 3.0472463768115947e-06, |
|
"loss": 0.2651, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 3.04e-06, |
|
"loss": 0.2555, |
|
"step": 24525 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 3.032753623188406e-06, |
|
"loss": 0.2511, |
|
"step": 24550 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 3.025507246376812e-06, |
|
"loss": 0.2761, |
|
"step": 24575 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 3.0182608695652177e-06, |
|
"loss": 0.3315, |
|
"step": 24600 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 3.0110144927536235e-06, |
|
"loss": 0.4222, |
|
"step": 24625 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 3.003768115942029e-06, |
|
"loss": 0.252, |
|
"step": 24650 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 2.9965217391304352e-06, |
|
"loss": 0.3424, |
|
"step": 24675 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 2.9892753623188407e-06, |
|
"loss": 0.3086, |
|
"step": 24700 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 2.9820289855072465e-06, |
|
"loss": 0.3246, |
|
"step": 24725 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 2.974782608695652e-06, |
|
"loss": 0.4331, |
|
"step": 24750 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 2.9675362318840582e-06, |
|
"loss": 0.2933, |
|
"step": 24775 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 2.960289855072464e-06, |
|
"loss": 0.3234, |
|
"step": 24800 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 2.9530434782608695e-06, |
|
"loss": 0.2495, |
|
"step": 24825 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 2.9457971014492754e-06, |
|
"loss": 0.2506, |
|
"step": 24850 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 2.9385507246376817e-06, |
|
"loss": 0.2862, |
|
"step": 24875 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 2.931304347826087e-06, |
|
"loss": 0.2804, |
|
"step": 24900 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 2.924057971014493e-06, |
|
"loss": 0.2659, |
|
"step": 24925 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 2.9168115942028992e-06, |
|
"loss": 0.2801, |
|
"step": 24950 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 2.9095652173913046e-06, |
|
"loss": 0.3455, |
|
"step": 24975 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 2.902608695652174e-06, |
|
"loss": 0.2833, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"eval_cer": 6.855231442766491, |
|
"eval_loss": 0.39096301794052124, |
|
"eval_runtime": 3187.6224, |
|
"eval_samples_per_second": 0.779, |
|
"eval_steps_per_second": 0.389, |
|
"eval_wer": 18.598056032018295, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 2.89536231884058e-06, |
|
"loss": 0.2299, |
|
"step": 25025 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 2.8881159420289854e-06, |
|
"loss": 0.2482, |
|
"step": 25050 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 2.8808695652173917e-06, |
|
"loss": 0.3388, |
|
"step": 25075 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 2.8736231884057975e-06, |
|
"loss": 0.4429, |
|
"step": 25100 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 2.866376811594203e-06, |
|
"loss": 0.2245, |
|
"step": 25125 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 2.859130434782609e-06, |
|
"loss": 0.2648, |
|
"step": 25150 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 2.851884057971015e-06, |
|
"loss": 0.1995, |
|
"step": 25175 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 2.8446376811594205e-06, |
|
"loss": 0.2652, |
|
"step": 25200 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 2.8373913043478264e-06, |
|
"loss": 0.2587, |
|
"step": 25225 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 2.830144927536232e-06, |
|
"loss": 0.2889, |
|
"step": 25250 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 2.822898550724638e-06, |
|
"loss": 0.2257, |
|
"step": 25275 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 2.815652173913044e-06, |
|
"loss": 0.1742, |
|
"step": 25300 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 2.8084057971014494e-06, |
|
"loss": 0.1856, |
|
"step": 25325 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 2.8011594202898552e-06, |
|
"loss": 0.1801, |
|
"step": 25350 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 2.7939130434782615e-06, |
|
"loss": 0.2185, |
|
"step": 25375 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 2.786666666666667e-06, |
|
"loss": 0.2565, |
|
"step": 25400 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 2.779420289855073e-06, |
|
"loss": 0.2026, |
|
"step": 25425 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 2.7721739130434782e-06, |
|
"loss": 0.2373, |
|
"step": 25450 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 2.7649275362318845e-06, |
|
"loss": 0.2406, |
|
"step": 25475 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 2.75768115942029e-06, |
|
"loss": 0.4514, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 2.750434782608696e-06, |
|
"loss": 0.1876, |
|
"step": 25525 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 2.7431884057971012e-06, |
|
"loss": 0.2703, |
|
"step": 25550 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 2.7359420289855075e-06, |
|
"loss": 0.1749, |
|
"step": 25575 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 2.7286956521739134e-06, |
|
"loss": 0.2088, |
|
"step": 25600 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 2.721449275362319e-06, |
|
"loss": 0.2087, |
|
"step": 25625 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 2.7142028985507246e-06, |
|
"loss": 0.3172, |
|
"step": 25650 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 2.706956521739131e-06, |
|
"loss": 0.1898, |
|
"step": 25675 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 2.6997101449275364e-06, |
|
"loss": 0.1685, |
|
"step": 25700 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 2.692463768115942e-06, |
|
"loss": 0.1918, |
|
"step": 25725 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 2.6852173913043476e-06, |
|
"loss": 0.2734, |
|
"step": 25750 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 2.677971014492754e-06, |
|
"loss": 0.1485, |
|
"step": 25775 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 2.6707246376811598e-06, |
|
"loss": 0.1542, |
|
"step": 25800 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 2.663478260869565e-06, |
|
"loss": 0.1694, |
|
"step": 25825 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 2.6562318840579715e-06, |
|
"loss": 0.1685, |
|
"step": 25850 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 2.6489855072463773e-06, |
|
"loss": 0.1719, |
|
"step": 25875 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 2.6417391304347828e-06, |
|
"loss": 0.1573, |
|
"step": 25900 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 2.6344927536231886e-06, |
|
"loss": 0.2429, |
|
"step": 25925 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 2.6272463768115945e-06, |
|
"loss": 0.1751, |
|
"step": 25950 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 2.6202898550724644e-06, |
|
"loss": 0.2976, |
|
"step": 25975 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 2.61304347826087e-06, |
|
"loss": 0.2509, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"eval_cer": 6.816493678980766, |
|
"eval_loss": 0.3948805630207062, |
|
"eval_runtime": 3174.2689, |
|
"eval_samples_per_second": 0.782, |
|
"eval_steps_per_second": 0.391, |
|
"eval_wer": 18.518010291595196, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 2.6057971014492757e-06, |
|
"loss": 0.377, |
|
"step": 26025 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 2.598550724637681e-06, |
|
"loss": 0.1902, |
|
"step": 26050 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 2.5913043478260874e-06, |
|
"loss": 0.2086, |
|
"step": 26075 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 2.5840579710144932e-06, |
|
"loss": 0.269, |
|
"step": 26100 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 2.5768115942028987e-06, |
|
"loss": 0.2287, |
|
"step": 26125 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 2.5695652173913045e-06, |
|
"loss": 0.2673, |
|
"step": 26150 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 2.562318840579711e-06, |
|
"loss": 0.1238, |
|
"step": 26175 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 2.5550724637681162e-06, |
|
"loss": 0.2047, |
|
"step": 26200 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 2.547826086956522e-06, |
|
"loss": 0.3, |
|
"step": 26225 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 2.5405797101449275e-06, |
|
"loss": 0.201, |
|
"step": 26250 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 2.5333333333333338e-06, |
|
"loss": 0.1864, |
|
"step": 26275 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 2.5260869565217392e-06, |
|
"loss": 0.2565, |
|
"step": 26300 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 2.518840579710145e-06, |
|
"loss": 0.189, |
|
"step": 26325 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 2.511594202898551e-06, |
|
"loss": 0.2006, |
|
"step": 26350 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 2.5043478260869568e-06, |
|
"loss": 0.1992, |
|
"step": 26375 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 2.4971014492753626e-06, |
|
"loss": 0.1742, |
|
"step": 26400 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 2.489855072463768e-06, |
|
"loss": 0.1909, |
|
"step": 26425 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 2.4826086956521743e-06, |
|
"loss": 0.1935, |
|
"step": 26450 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 2.4753623188405798e-06, |
|
"loss": 0.2335, |
|
"step": 26475 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 2.4681159420289856e-06, |
|
"loss": 0.2635, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 2.4608695652173915e-06, |
|
"loss": 0.2079, |
|
"step": 26525 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 2.4536231884057973e-06, |
|
"loss": 0.2006, |
|
"step": 26550 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 2.446376811594203e-06, |
|
"loss": 0.244, |
|
"step": 26575 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 2.439130434782609e-06, |
|
"loss": 0.32, |
|
"step": 26600 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 2.4318840579710145e-06, |
|
"loss": 0.1757, |
|
"step": 26625 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 2.4246376811594203e-06, |
|
"loss": 0.151, |
|
"step": 26650 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 2.417391304347826e-06, |
|
"loss": 0.1985, |
|
"step": 26675 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 2.410144927536232e-06, |
|
"loss": 0.2241, |
|
"step": 26700 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 2.402898550724638e-06, |
|
"loss": 0.1411, |
|
"step": 26725 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 2.3956521739130437e-06, |
|
"loss": 0.2348, |
|
"step": 26750 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 2.3884057971014496e-06, |
|
"loss": 0.2529, |
|
"step": 26775 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 2.3811594202898555e-06, |
|
"loss": 0.1819, |
|
"step": 26800 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 2.3739130434782613e-06, |
|
"loss": 0.2473, |
|
"step": 26825 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 2.3666666666666667e-06, |
|
"loss": 0.2902, |
|
"step": 26850 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 2.3594202898550726e-06, |
|
"loss": 0.1368, |
|
"step": 26875 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 2.3521739130434784e-06, |
|
"loss": 0.2857, |
|
"step": 26900 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 2.3449275362318843e-06, |
|
"loss": 0.1613, |
|
"step": 26925 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 2.33768115942029e-06, |
|
"loss": 0.2573, |
|
"step": 26950 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 2.330434782608696e-06, |
|
"loss": 0.2029, |
|
"step": 26975 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 2.3231884057971014e-06, |
|
"loss": 0.2085, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"eval_cer": 6.811328643809336, |
|
"eval_loss": 0.3984606862068176, |
|
"eval_runtime": 3176.0872, |
|
"eval_samples_per_second": 0.781, |
|
"eval_steps_per_second": 0.391, |
|
"eval_wer": 18.613302839717935, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 2.3159420289855077e-06, |
|
"loss": 0.235, |
|
"step": 27025 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 2.308695652173913e-06, |
|
"loss": 0.1992, |
|
"step": 27050 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 2.301449275362319e-06, |
|
"loss": 0.2457, |
|
"step": 27075 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 2.294202898550725e-06, |
|
"loss": 0.1871, |
|
"step": 27100 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 2.2869565217391307e-06, |
|
"loss": 0.1464, |
|
"step": 27125 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 2.279710144927536e-06, |
|
"loss": 0.2012, |
|
"step": 27150 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 2.2724637681159424e-06, |
|
"loss": 0.1757, |
|
"step": 27175 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 2.265217391304348e-06, |
|
"loss": 0.1795, |
|
"step": 27200 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 2.2579710144927537e-06, |
|
"loss": 0.1706, |
|
"step": 27225 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 2.2507246376811596e-06, |
|
"loss": 0.2044, |
|
"step": 27250 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 2.2434782608695654e-06, |
|
"loss": 0.2155, |
|
"step": 27275 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 2.2362318840579713e-06, |
|
"loss": 0.1811, |
|
"step": 27300 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 2.228985507246377e-06, |
|
"loss": 0.2311, |
|
"step": 27325 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 2.2217391304347825e-06, |
|
"loss": 0.3311, |
|
"step": 27350 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 2.2144927536231884e-06, |
|
"loss": 0.1692, |
|
"step": 27375 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 2.2072463768115943e-06, |
|
"loss": 0.2069, |
|
"step": 27400 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 2.2e-06, |
|
"loss": 0.1755, |
|
"step": 27425 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 2.192753623188406e-06, |
|
"loss": 0.2029, |
|
"step": 27450 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 2.185507246376812e-06, |
|
"loss": 0.2542, |
|
"step": 27475 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 2.1782608695652177e-06, |
|
"loss": 0.1882, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 2.1710144927536235e-06, |
|
"loss": 0.2374, |
|
"step": 27525 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 2.1637681159420294e-06, |
|
"loss": 0.1967, |
|
"step": 27550 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 2.156521739130435e-06, |
|
"loss": 0.1869, |
|
"step": 27575 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 2.1492753623188407e-06, |
|
"loss": 0.2861, |
|
"step": 27600 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 2.1420289855072465e-06, |
|
"loss": 0.2392, |
|
"step": 27625 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 2.1347826086956524e-06, |
|
"loss": 0.18, |
|
"step": 27650 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 2.1275362318840582e-06, |
|
"loss": 0.2496, |
|
"step": 27675 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 2.120289855072464e-06, |
|
"loss": 0.1998, |
|
"step": 27700 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 2.1130434782608695e-06, |
|
"loss": 0.259, |
|
"step": 27725 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 2.105797101449276e-06, |
|
"loss": 0.3129, |
|
"step": 27750 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 2.0985507246376812e-06, |
|
"loss": 0.1907, |
|
"step": 27775 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 2.091304347826087e-06, |
|
"loss": 0.1343, |
|
"step": 27800 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 2.084057971014493e-06, |
|
"loss": 0.1973, |
|
"step": 27825 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 2.0768115942028988e-06, |
|
"loss": 0.1658, |
|
"step": 27850 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 2.0695652173913046e-06, |
|
"loss": 0.1508, |
|
"step": 27875 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 2.0623188405797105e-06, |
|
"loss": 0.1947, |
|
"step": 27900 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 2.055072463768116e-06, |
|
"loss": 0.1812, |
|
"step": 27925 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 2.0478260869565218e-06, |
|
"loss": 0.2154, |
|
"step": 27950 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 2.0405797101449276e-06, |
|
"loss": 0.1456, |
|
"step": 27975 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 2.0333333333333335e-06, |
|
"loss": 0.2055, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"eval_cer": 6.862364110384181, |
|
"eval_loss": 0.39950060844421387, |
|
"eval_runtime": 3191.9428, |
|
"eval_samples_per_second": 0.778, |
|
"eval_steps_per_second": 0.389, |
|
"eval_wer": 18.76119687440442, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 2.0260869565217393e-06, |
|
"loss": 0.1855, |
|
"step": 28025 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 2.018840579710145e-06, |
|
"loss": 0.2535, |
|
"step": 28050 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 2.0115942028985506e-06, |
|
"loss": 0.2101, |
|
"step": 28075 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 2.004347826086957e-06, |
|
"loss": 0.2977, |
|
"step": 28100 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 1.9971014492753623e-06, |
|
"loss": 0.1915, |
|
"step": 28125 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 1.989855072463768e-06, |
|
"loss": 0.1589, |
|
"step": 28150 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 1.982608695652174e-06, |
|
"loss": 0.175, |
|
"step": 28175 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 1.97536231884058e-06, |
|
"loss": 0.2892, |
|
"step": 28200 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 1.9681159420289853e-06, |
|
"loss": 0.223, |
|
"step": 28225 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 1.9608695652173916e-06, |
|
"loss": 0.1731, |
|
"step": 28250 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 1.9536231884057975e-06, |
|
"loss": 0.2097, |
|
"step": 28275 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 1.946376811594203e-06, |
|
"loss": 0.1909, |
|
"step": 28300 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 1.939130434782609e-06, |
|
"loss": 0.2079, |
|
"step": 28325 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 1.9318840579710146e-06, |
|
"loss": 0.1848, |
|
"step": 28350 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 1.9246376811594205e-06, |
|
"loss": 0.1813, |
|
"step": 28375 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 1.9173913043478263e-06, |
|
"loss": 0.1679, |
|
"step": 28400 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 1.910144927536232e-06, |
|
"loss": 0.1787, |
|
"step": 28425 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 1.9028985507246378e-06, |
|
"loss": 0.1718, |
|
"step": 28450 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 1.8956521739130437e-06, |
|
"loss": 0.2553, |
|
"step": 28475 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 1.8884057971014493e-06, |
|
"loss": 0.2249, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 1.8811594202898554e-06, |
|
"loss": 0.1657, |
|
"step": 28525 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 1.873913043478261e-06, |
|
"loss": 0.1738, |
|
"step": 28550 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 1.8666666666666669e-06, |
|
"loss": 0.1935, |
|
"step": 28575 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 1.8594202898550725e-06, |
|
"loss": 0.2179, |
|
"step": 28600 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 1.8521739130434786e-06, |
|
"loss": 0.2696, |
|
"step": 28625 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 1.8449275362318842e-06, |
|
"loss": 0.1657, |
|
"step": 28650 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 1.83768115942029e-06, |
|
"loss": 0.2302, |
|
"step": 28675 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 1.8304347826086957e-06, |
|
"loss": 0.2888, |
|
"step": 28700 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 1.8231884057971016e-06, |
|
"loss": 0.193, |
|
"step": 28725 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 1.8159420289855072e-06, |
|
"loss": 0.2122, |
|
"step": 28750 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 1.8086956521739133e-06, |
|
"loss": 0.1602, |
|
"step": 28775 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 1.801449275362319e-06, |
|
"loss": 0.2419, |
|
"step": 28800 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 1.7942028985507248e-06, |
|
"loss": 0.4138, |
|
"step": 28825 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 1.7869565217391304e-06, |
|
"loss": 0.2473, |
|
"step": 28850 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 1.7797101449275365e-06, |
|
"loss": 0.1792, |
|
"step": 28875 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 1.7724637681159421e-06, |
|
"loss": 0.1822, |
|
"step": 28900 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 1.765217391304348e-06, |
|
"loss": 0.1505, |
|
"step": 28925 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 1.7579710144927536e-06, |
|
"loss": 0.1418, |
|
"step": 28950 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 1.7507246376811595e-06, |
|
"loss": 0.1262, |
|
"step": 28975 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 1.7434782608695653e-06, |
|
"loss": 0.175, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"eval_cer": 6.772713857051503, |
|
"eval_loss": 0.40093937516212463, |
|
"eval_runtime": 3157.1944, |
|
"eval_samples_per_second": 0.786, |
|
"eval_steps_per_second": 0.393, |
|
"eval_wer": 18.48141795311607, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 1.7362318840579712e-06, |
|
"loss": 0.2065, |
|
"step": 29025 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 1.7289855072463768e-06, |
|
"loss": 0.1723, |
|
"step": 29050 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 1.7217391304347827e-06, |
|
"loss": 0.1419, |
|
"step": 29075 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 1.7144927536231887e-06, |
|
"loss": 0.2607, |
|
"step": 29100 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 1.7072463768115944e-06, |
|
"loss": 0.1884, |
|
"step": 29125 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 1.7000000000000002e-06, |
|
"loss": 0.1824, |
|
"step": 29150 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 1.6927536231884059e-06, |
|
"loss": 0.1657, |
|
"step": 29175 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 1.685507246376812e-06, |
|
"loss": 0.248, |
|
"step": 29200 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 1.6782608695652176e-06, |
|
"loss": 0.2124, |
|
"step": 29225 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 1.6710144927536234e-06, |
|
"loss": 0.1964, |
|
"step": 29250 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 1.663768115942029e-06, |
|
"loss": 0.3202, |
|
"step": 29275 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 1.656521739130435e-06, |
|
"loss": 0.194, |
|
"step": 29300 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 1.6492753623188406e-06, |
|
"loss": 0.2738, |
|
"step": 29325 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 1.6420289855072467e-06, |
|
"loss": 0.1667, |
|
"step": 29350 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 1.6347826086956523e-06, |
|
"loss": 0.2819, |
|
"step": 29375 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 1.6275362318840581e-06, |
|
"loss": 0.1894, |
|
"step": 29400 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 1.6202898550724638e-06, |
|
"loss": 0.1791, |
|
"step": 29425 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 1.6130434782608699e-06, |
|
"loss": 0.1853, |
|
"step": 29450 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 1.6057971014492755e-06, |
|
"loss": 0.2618, |
|
"step": 29475 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 1.5985507246376814e-06, |
|
"loss": 0.2052, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 1.591304347826087e-06, |
|
"loss": 0.2185, |
|
"step": 29525 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 1.5840579710144928e-06, |
|
"loss": 0.1305, |
|
"step": 29550 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 1.5768115942028985e-06, |
|
"loss": 0.233, |
|
"step": 29575 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 1.5695652173913046e-06, |
|
"loss": 0.2164, |
|
"step": 29600 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 1.5623188405797102e-06, |
|
"loss": 0.1524, |
|
"step": 29625 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 1.555072463768116e-06, |
|
"loss": 0.1961, |
|
"step": 29650 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 1.5478260869565217e-06, |
|
"loss": 0.2242, |
|
"step": 29675 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 1.5405797101449278e-06, |
|
"loss": 0.1997, |
|
"step": 29700 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 1.5333333333333334e-06, |
|
"loss": 0.2001, |
|
"step": 29725 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 1.5260869565217393e-06, |
|
"loss": 0.1888, |
|
"step": 29750 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 1.518840579710145e-06, |
|
"loss": 0.2124, |
|
"step": 29775 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 1.5115942028985508e-06, |
|
"loss": 0.2145, |
|
"step": 29800 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 1.5043478260869566e-06, |
|
"loss": 0.2512, |
|
"step": 29825 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 1.4971014492753625e-06, |
|
"loss": 0.1625, |
|
"step": 29850 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 1.4898550724637683e-06, |
|
"loss": 0.1556, |
|
"step": 29875 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 1.482608695652174e-06, |
|
"loss": 0.2134, |
|
"step": 29900 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 1.47536231884058e-06, |
|
"loss": 0.1959, |
|
"step": 29925 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 1.4681159420289857e-06, |
|
"loss": 0.197, |
|
"step": 29950 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 1.4608695652173915e-06, |
|
"loss": 0.1825, |
|
"step": 29975 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 1.4536231884057972e-06, |
|
"loss": 0.1701, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"eval_cer": 7.013625854690344, |
|
"eval_loss": 0.3998452425003052, |
|
"eval_runtime": 3192.2868, |
|
"eval_samples_per_second": 0.777, |
|
"eval_steps_per_second": 0.389, |
|
"eval_wer": 18.834381551362682, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 32.74, |
|
"learning_rate": 1.446376811594203e-06, |
|
"loss": 1.1855, |
|
"step": 30025 |
|
}, |
|
{ |
|
"epoch": 32.77, |
|
"learning_rate": 1.4391304347826089e-06, |
|
"loss": 1.0754, |
|
"step": 30050 |
|
}, |
|
{ |
|
"epoch": 32.8, |
|
"learning_rate": 1.4318840579710147e-06, |
|
"loss": 0.9609, |
|
"step": 30075 |
|
}, |
|
{ |
|
"epoch": 32.82, |
|
"learning_rate": 1.4246376811594204e-06, |
|
"loss": 0.9363, |
|
"step": 30100 |
|
}, |
|
{ |
|
"epoch": 32.85, |
|
"learning_rate": 1.4173913043478262e-06, |
|
"loss": 0.9573, |
|
"step": 30125 |
|
}, |
|
{ |
|
"epoch": 32.88, |
|
"learning_rate": 1.4101449275362319e-06, |
|
"loss": 0.945, |
|
"step": 30150 |
|
}, |
|
{ |
|
"epoch": 32.91, |
|
"learning_rate": 1.402898550724638e-06, |
|
"loss": 0.8939, |
|
"step": 30175 |
|
}, |
|
{ |
|
"epoch": 32.93, |
|
"learning_rate": 1.3956521739130436e-06, |
|
"loss": 0.9062, |
|
"step": 30200 |
|
}, |
|
{ |
|
"epoch": 32.96, |
|
"learning_rate": 1.3884057971014494e-06, |
|
"loss": 0.9019, |
|
"step": 30225 |
|
}, |
|
{ |
|
"epoch": 32.99, |
|
"learning_rate": 1.381159420289855e-06, |
|
"loss": 0.8691, |
|
"step": 30250 |
|
}, |
|
{ |
|
"epoch": 33.02, |
|
"learning_rate": 1.3739130434782611e-06, |
|
"loss": 0.8198, |
|
"step": 30275 |
|
}, |
|
{ |
|
"epoch": 33.04, |
|
"learning_rate": 1.3666666666666668e-06, |
|
"loss": 0.8642, |
|
"step": 30300 |
|
}, |
|
{ |
|
"epoch": 33.07, |
|
"learning_rate": 1.3594202898550726e-06, |
|
"loss": 0.8705, |
|
"step": 30325 |
|
}, |
|
{ |
|
"epoch": 33.1, |
|
"learning_rate": 1.3521739130434783e-06, |
|
"loss": 0.7469, |
|
"step": 30350 |
|
}, |
|
{ |
|
"epoch": 33.12, |
|
"learning_rate": 1.3449275362318841e-06, |
|
"loss": 0.7704, |
|
"step": 30375 |
|
}, |
|
{ |
|
"epoch": 33.15, |
|
"learning_rate": 1.3376811594202898e-06, |
|
"loss": 0.8126, |
|
"step": 30400 |
|
}, |
|
{ |
|
"epoch": 33.18, |
|
"learning_rate": 1.3304347826086958e-06, |
|
"loss": 0.8152, |
|
"step": 30425 |
|
}, |
|
{ |
|
"epoch": 33.21, |
|
"learning_rate": 1.3231884057971015e-06, |
|
"loss": 0.7214, |
|
"step": 30450 |
|
}, |
|
{ |
|
"epoch": 33.23, |
|
"learning_rate": 1.3159420289855073e-06, |
|
"loss": 0.8131, |
|
"step": 30475 |
|
}, |
|
{ |
|
"epoch": 33.26, |
|
"learning_rate": 1.308695652173913e-06, |
|
"loss": 0.7234, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 33.29, |
|
"learning_rate": 1.301449275362319e-06, |
|
"loss": 0.7552, |
|
"step": 30525 |
|
}, |
|
{ |
|
"epoch": 33.32, |
|
"learning_rate": 1.2942028985507247e-06, |
|
"loss": 0.7036, |
|
"step": 30550 |
|
}, |
|
{ |
|
"epoch": 33.34, |
|
"learning_rate": 1.2869565217391305e-06, |
|
"loss": 0.7777, |
|
"step": 30575 |
|
}, |
|
{ |
|
"epoch": 33.37, |
|
"learning_rate": 1.2797101449275362e-06, |
|
"loss": 0.6986, |
|
"step": 30600 |
|
}, |
|
{ |
|
"epoch": 33.4, |
|
"learning_rate": 1.272463768115942e-06, |
|
"loss": 0.7156, |
|
"step": 30625 |
|
}, |
|
{ |
|
"epoch": 33.42, |
|
"learning_rate": 1.265217391304348e-06, |
|
"loss": 0.7919, |
|
"step": 30650 |
|
}, |
|
{ |
|
"epoch": 33.45, |
|
"learning_rate": 1.2579710144927537e-06, |
|
"loss": 0.6559, |
|
"step": 30675 |
|
}, |
|
{ |
|
"epoch": 33.48, |
|
"learning_rate": 1.2507246376811596e-06, |
|
"loss": 0.6646, |
|
"step": 30700 |
|
}, |
|
{ |
|
"epoch": 33.51, |
|
"learning_rate": 1.2434782608695652e-06, |
|
"loss": 0.6952, |
|
"step": 30725 |
|
}, |
|
{ |
|
"epoch": 33.53, |
|
"learning_rate": 1.236231884057971e-06, |
|
"loss": 0.7332, |
|
"step": 30750 |
|
}, |
|
{ |
|
"epoch": 33.56, |
|
"learning_rate": 1.228985507246377e-06, |
|
"loss": 0.7673, |
|
"step": 30775 |
|
}, |
|
{ |
|
"epoch": 33.59, |
|
"learning_rate": 1.2217391304347826e-06, |
|
"loss": 0.7079, |
|
"step": 30800 |
|
}, |
|
{ |
|
"epoch": 33.62, |
|
"learning_rate": 1.2144927536231884e-06, |
|
"loss": 0.7293, |
|
"step": 30825 |
|
}, |
|
{ |
|
"epoch": 33.64, |
|
"learning_rate": 1.2072463768115943e-06, |
|
"loss": 0.6716, |
|
"step": 30850 |
|
}, |
|
{ |
|
"epoch": 33.67, |
|
"learning_rate": 1.2000000000000002e-06, |
|
"loss": 0.7037, |
|
"step": 30875 |
|
}, |
|
{ |
|
"epoch": 33.7, |
|
"learning_rate": 1.1927536231884058e-06, |
|
"loss": 0.7235, |
|
"step": 30900 |
|
}, |
|
{ |
|
"epoch": 33.72, |
|
"learning_rate": 1.1855072463768117e-06, |
|
"loss": 0.638, |
|
"step": 30925 |
|
}, |
|
{ |
|
"epoch": 33.75, |
|
"learning_rate": 1.1782608695652175e-06, |
|
"loss": 0.7433, |
|
"step": 30950 |
|
}, |
|
{ |
|
"epoch": 33.78, |
|
"learning_rate": 1.1710144927536232e-06, |
|
"loss": 0.7318, |
|
"step": 30975 |
|
}, |
|
{ |
|
"epoch": 33.81, |
|
"learning_rate": 1.163768115942029e-06, |
|
"loss": 0.6832, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 33.81, |
|
"eval_cer": 7.850853989513844, |
|
"eval_loss": 0.542524516582489, |
|
"eval_runtime": 3197.0599, |
|
"eval_samples_per_second": 0.677, |
|
"eval_steps_per_second": 0.17, |
|
"eval_wer": 24.621636222704176, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 33.83, |
|
"learning_rate": 1.1565217391304349e-06, |
|
"loss": 0.673, |
|
"step": 31025 |
|
}, |
|
{ |
|
"epoch": 33.86, |
|
"learning_rate": 1.1492753623188407e-06, |
|
"loss": 0.6608, |
|
"step": 31050 |
|
}, |
|
{ |
|
"epoch": 33.89, |
|
"learning_rate": 1.1420289855072466e-06, |
|
"loss": 0.6553, |
|
"step": 31075 |
|
}, |
|
{ |
|
"epoch": 33.91, |
|
"learning_rate": 1.1347826086956524e-06, |
|
"loss": 0.6837, |
|
"step": 31100 |
|
}, |
|
{ |
|
"epoch": 33.94, |
|
"learning_rate": 1.127536231884058e-06, |
|
"loss": 0.6683, |
|
"step": 31125 |
|
}, |
|
{ |
|
"epoch": 33.97, |
|
"learning_rate": 1.120289855072464e-06, |
|
"loss": 0.7043, |
|
"step": 31150 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"learning_rate": 1.1130434782608698e-06, |
|
"loss": 0.647, |
|
"step": 31175 |
|
}, |
|
{ |
|
"epoch": 34.02, |
|
"learning_rate": 1.1057971014492754e-06, |
|
"loss": 0.691, |
|
"step": 31200 |
|
}, |
|
{ |
|
"epoch": 34.05, |
|
"learning_rate": 1.0985507246376813e-06, |
|
"loss": 0.6857, |
|
"step": 31225 |
|
}, |
|
{ |
|
"epoch": 34.08, |
|
"learning_rate": 1.0913043478260871e-06, |
|
"loss": 0.5965, |
|
"step": 31250 |
|
}, |
|
{ |
|
"epoch": 34.11, |
|
"learning_rate": 1.0840579710144928e-06, |
|
"loss": 0.6514, |
|
"step": 31275 |
|
}, |
|
{ |
|
"epoch": 34.13, |
|
"learning_rate": 1.0768115942028986e-06, |
|
"loss": 0.5753, |
|
"step": 31300 |
|
}, |
|
{ |
|
"epoch": 34.16, |
|
"learning_rate": 1.0695652173913045e-06, |
|
"loss": 0.6485, |
|
"step": 31325 |
|
}, |
|
{ |
|
"epoch": 34.19, |
|
"learning_rate": 1.0623188405797103e-06, |
|
"loss": 0.6007, |
|
"step": 31350 |
|
}, |
|
{ |
|
"epoch": 34.21, |
|
"learning_rate": 1.055072463768116e-06, |
|
"loss": 0.5861, |
|
"step": 31375 |
|
}, |
|
{ |
|
"epoch": 34.24, |
|
"learning_rate": 1.0478260869565218e-06, |
|
"loss": 0.6009, |
|
"step": 31400 |
|
}, |
|
{ |
|
"epoch": 34.27, |
|
"learning_rate": 1.0405797101449277e-06, |
|
"loss": 0.5646, |
|
"step": 31425 |
|
}, |
|
{ |
|
"epoch": 34.3, |
|
"learning_rate": 1.0333333333333333e-06, |
|
"loss": 0.5896, |
|
"step": 31450 |
|
}, |
|
{ |
|
"epoch": 34.32, |
|
"learning_rate": 1.0260869565217392e-06, |
|
"loss": 0.6017, |
|
"step": 31475 |
|
}, |
|
{ |
|
"epoch": 34.35, |
|
"learning_rate": 1.018840579710145e-06, |
|
"loss": 0.6172, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 34.38, |
|
"learning_rate": 1.0115942028985509e-06, |
|
"loss": 0.5782, |
|
"step": 31525 |
|
}, |
|
{ |
|
"epoch": 34.41, |
|
"learning_rate": 1.0043478260869565e-06, |
|
"loss": 0.6157, |
|
"step": 31550 |
|
}, |
|
{ |
|
"epoch": 34.43, |
|
"learning_rate": 9.971014492753624e-07, |
|
"loss": 0.6407, |
|
"step": 31575 |
|
}, |
|
{ |
|
"epoch": 34.46, |
|
"learning_rate": 9.898550724637682e-07, |
|
"loss": 0.646, |
|
"step": 31600 |
|
}, |
|
{ |
|
"epoch": 34.49, |
|
"learning_rate": 9.826086956521739e-07, |
|
"loss": 0.5987, |
|
"step": 31625 |
|
}, |
|
{ |
|
"epoch": 34.51, |
|
"learning_rate": 9.753623188405797e-07, |
|
"loss": 0.5978, |
|
"step": 31650 |
|
}, |
|
{ |
|
"epoch": 34.54, |
|
"learning_rate": 9.681159420289856e-07, |
|
"loss": 0.6151, |
|
"step": 31675 |
|
}, |
|
{ |
|
"epoch": 34.57, |
|
"learning_rate": 9.608695652173912e-07, |
|
"loss": 0.6037, |
|
"step": 31700 |
|
}, |
|
{ |
|
"epoch": 34.6, |
|
"learning_rate": 9.536231884057971e-07, |
|
"loss": 0.5898, |
|
"step": 31725 |
|
}, |
|
{ |
|
"epoch": 34.62, |
|
"learning_rate": 9.463768115942029e-07, |
|
"loss": 0.5873, |
|
"step": 31750 |
|
}, |
|
{ |
|
"epoch": 34.65, |
|
"learning_rate": 9.391304347826087e-07, |
|
"loss": 0.6409, |
|
"step": 31775 |
|
}, |
|
{ |
|
"epoch": 34.68, |
|
"learning_rate": 9.318840579710145e-07, |
|
"loss": 0.6163, |
|
"step": 31800 |
|
}, |
|
{ |
|
"epoch": 34.71, |
|
"learning_rate": 9.246376811594204e-07, |
|
"loss": 0.549, |
|
"step": 31825 |
|
}, |
|
{ |
|
"epoch": 34.73, |
|
"learning_rate": 9.173913043478262e-07, |
|
"loss": 0.5967, |
|
"step": 31850 |
|
}, |
|
{ |
|
"epoch": 34.76, |
|
"learning_rate": 9.10144927536232e-07, |
|
"loss": 0.5369, |
|
"step": 31875 |
|
}, |
|
{ |
|
"epoch": 34.79, |
|
"learning_rate": 9.028985507246379e-07, |
|
"loss": 0.5848, |
|
"step": 31900 |
|
}, |
|
{ |
|
"epoch": 34.81, |
|
"learning_rate": 8.956521739130436e-07, |
|
"loss": 0.569, |
|
"step": 31925 |
|
}, |
|
{ |
|
"epoch": 34.84, |
|
"learning_rate": 8.884057971014493e-07, |
|
"loss": 0.5782, |
|
"step": 31950 |
|
}, |
|
{ |
|
"epoch": 34.87, |
|
"learning_rate": 8.811594202898552e-07, |
|
"loss": 0.5425, |
|
"step": 31975 |
|
}, |
|
{ |
|
"epoch": 34.9, |
|
"learning_rate": 8.73913043478261e-07, |
|
"loss": 0.5676, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 34.9, |
|
"eval_cer": 7.377586234087609, |
|
"eval_loss": 0.5141463875770569, |
|
"eval_runtime": 3170.6638, |
|
"eval_samples_per_second": 0.683, |
|
"eval_steps_per_second": 0.171, |
|
"eval_wer": 23.67895647544257, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 34.92, |
|
"learning_rate": 8.666666666666668e-07, |
|
"loss": 0.5505, |
|
"step": 32025 |
|
}, |
|
{ |
|
"epoch": 34.95, |
|
"learning_rate": 8.594202898550726e-07, |
|
"loss": 0.5515, |
|
"step": 32050 |
|
}, |
|
{ |
|
"epoch": 34.98, |
|
"learning_rate": 8.521739130434783e-07, |
|
"loss": 0.5501, |
|
"step": 32075 |
|
}, |
|
{ |
|
"epoch": 35.01, |
|
"learning_rate": 8.449275362318842e-07, |
|
"loss": 0.5561, |
|
"step": 32100 |
|
}, |
|
{ |
|
"epoch": 35.03, |
|
"learning_rate": 8.376811594202899e-07, |
|
"loss": 0.5846, |
|
"step": 32125 |
|
}, |
|
{ |
|
"epoch": 35.06, |
|
"learning_rate": 8.304347826086958e-07, |
|
"loss": 0.5215, |
|
"step": 32150 |
|
}, |
|
{ |
|
"epoch": 35.09, |
|
"learning_rate": 8.231884057971015e-07, |
|
"loss": 0.5006, |
|
"step": 32175 |
|
}, |
|
{ |
|
"epoch": 35.11, |
|
"learning_rate": 8.159420289855074e-07, |
|
"loss": 0.5434, |
|
"step": 32200 |
|
}, |
|
{ |
|
"epoch": 35.14, |
|
"learning_rate": 8.086956521739131e-07, |
|
"loss": 0.5447, |
|
"step": 32225 |
|
}, |
|
{ |
|
"epoch": 35.17, |
|
"learning_rate": 8.014492753623189e-07, |
|
"loss": 0.5672, |
|
"step": 32250 |
|
}, |
|
{ |
|
"epoch": 35.2, |
|
"learning_rate": 7.942028985507247e-07, |
|
"loss": 0.5617, |
|
"step": 32275 |
|
}, |
|
{ |
|
"epoch": 35.22, |
|
"learning_rate": 7.869565217391305e-07, |
|
"loss": 0.5237, |
|
"step": 32300 |
|
}, |
|
{ |
|
"epoch": 35.25, |
|
"learning_rate": 7.797101449275363e-07, |
|
"loss": 0.6142, |
|
"step": 32325 |
|
}, |
|
{ |
|
"epoch": 35.28, |
|
"learning_rate": 7.724637681159421e-07, |
|
"loss": 0.5075, |
|
"step": 32350 |
|
}, |
|
{ |
|
"epoch": 35.31, |
|
"learning_rate": 7.652173913043478e-07, |
|
"loss": 0.5431, |
|
"step": 32375 |
|
}, |
|
{ |
|
"epoch": 35.33, |
|
"learning_rate": 7.579710144927537e-07, |
|
"loss": 0.4816, |
|
"step": 32400 |
|
}, |
|
{ |
|
"epoch": 35.36, |
|
"learning_rate": 7.507246376811594e-07, |
|
"loss": 0.5244, |
|
"step": 32425 |
|
}, |
|
{ |
|
"epoch": 35.39, |
|
"learning_rate": 7.434782608695653e-07, |
|
"loss": 0.5152, |
|
"step": 32450 |
|
}, |
|
{ |
|
"epoch": 35.41, |
|
"learning_rate": 7.36231884057971e-07, |
|
"loss": 0.5653, |
|
"step": 32475 |
|
}, |
|
{ |
|
"epoch": 35.44, |
|
"learning_rate": 7.289855072463769e-07, |
|
"loss": 0.4795, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 35.47, |
|
"learning_rate": 7.217391304347826e-07, |
|
"loss": 0.5132, |
|
"step": 32525 |
|
}, |
|
{ |
|
"epoch": 35.5, |
|
"learning_rate": 7.144927536231884e-07, |
|
"loss": 0.5668, |
|
"step": 32550 |
|
}, |
|
{ |
|
"epoch": 35.52, |
|
"learning_rate": 7.072463768115942e-07, |
|
"loss": 0.5509, |
|
"step": 32575 |
|
}, |
|
{ |
|
"epoch": 35.55, |
|
"learning_rate": 7.000000000000001e-07, |
|
"loss": 0.532, |
|
"step": 32600 |
|
}, |
|
{ |
|
"epoch": 35.58, |
|
"learning_rate": 6.927536231884059e-07, |
|
"loss": 0.5321, |
|
"step": 32625 |
|
}, |
|
{ |
|
"epoch": 35.61, |
|
"learning_rate": 6.855072463768117e-07, |
|
"loss": 0.5079, |
|
"step": 32650 |
|
}, |
|
{ |
|
"epoch": 35.63, |
|
"learning_rate": 6.782608695652175e-07, |
|
"loss": 0.5015, |
|
"step": 32675 |
|
}, |
|
{ |
|
"epoch": 35.66, |
|
"learning_rate": 6.710144927536233e-07, |
|
"loss": 0.5693, |
|
"step": 32700 |
|
}, |
|
{ |
|
"epoch": 35.69, |
|
"learning_rate": 6.637681159420291e-07, |
|
"loss": 0.5046, |
|
"step": 32725 |
|
}, |
|
{ |
|
"epoch": 35.71, |
|
"learning_rate": 6.565217391304349e-07, |
|
"loss": 0.5433, |
|
"step": 32750 |
|
}, |
|
{ |
|
"epoch": 35.74, |
|
"learning_rate": 6.492753623188406e-07, |
|
"loss": 0.5056, |
|
"step": 32775 |
|
}, |
|
{ |
|
"epoch": 35.77, |
|
"learning_rate": 6.420289855072465e-07, |
|
"loss": 0.5413, |
|
"step": 32800 |
|
}, |
|
{ |
|
"epoch": 35.8, |
|
"learning_rate": 6.347826086956522e-07, |
|
"loss": 0.5512, |
|
"step": 32825 |
|
}, |
|
{ |
|
"epoch": 35.82, |
|
"learning_rate": 6.275362318840581e-07, |
|
"loss": 0.5224, |
|
"step": 32850 |
|
}, |
|
{ |
|
"epoch": 35.85, |
|
"learning_rate": 6.202898550724638e-07, |
|
"loss": 0.4961, |
|
"step": 32875 |
|
}, |
|
{ |
|
"epoch": 35.88, |
|
"learning_rate": 6.130434782608696e-07, |
|
"loss": 0.5242, |
|
"step": 32900 |
|
}, |
|
{ |
|
"epoch": 35.91, |
|
"learning_rate": 6.057971014492754e-07, |
|
"loss": 0.471, |
|
"step": 32925 |
|
}, |
|
{ |
|
"epoch": 35.93, |
|
"learning_rate": 5.985507246376812e-07, |
|
"loss": 0.547, |
|
"step": 32950 |
|
}, |
|
{ |
|
"epoch": 35.96, |
|
"learning_rate": 5.91304347826087e-07, |
|
"loss": 0.5817, |
|
"step": 32975 |
|
}, |
|
{ |
|
"epoch": 35.99, |
|
"learning_rate": 5.840579710144928e-07, |
|
"loss": 0.4863, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 35.99, |
|
"eval_cer": 7.244107809276393, |
|
"eval_loss": 0.5002937316894531, |
|
"eval_runtime": 3172.2622, |
|
"eval_samples_per_second": 0.683, |
|
"eval_steps_per_second": 0.171, |
|
"eval_wer": 23.0541571080715, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 36.01, |
|
"learning_rate": 5.768115942028985e-07, |
|
"loss": 0.4949, |
|
"step": 33025 |
|
}, |
|
{ |
|
"epoch": 36.04, |
|
"learning_rate": 5.695652173913044e-07, |
|
"loss": 0.4664, |
|
"step": 33050 |
|
}, |
|
{ |
|
"epoch": 36.07, |
|
"learning_rate": 5.623188405797101e-07, |
|
"loss": 0.4838, |
|
"step": 33075 |
|
}, |
|
{ |
|
"epoch": 36.1, |
|
"learning_rate": 5.55072463768116e-07, |
|
"loss": 0.4896, |
|
"step": 33100 |
|
}, |
|
{ |
|
"epoch": 36.12, |
|
"learning_rate": 5.478260869565217e-07, |
|
"loss": 0.5072, |
|
"step": 33125 |
|
}, |
|
{ |
|
"epoch": 36.15, |
|
"learning_rate": 5.405797101449276e-07, |
|
"loss": 0.4928, |
|
"step": 33150 |
|
}, |
|
{ |
|
"epoch": 36.18, |
|
"learning_rate": 5.333333333333335e-07, |
|
"loss": 0.5154, |
|
"step": 33175 |
|
}, |
|
{ |
|
"epoch": 36.21, |
|
"learning_rate": 5.260869565217392e-07, |
|
"loss": 0.5093, |
|
"step": 33200 |
|
}, |
|
{ |
|
"epoch": 36.23, |
|
"learning_rate": 5.188405797101449e-07, |
|
"loss": 0.4988, |
|
"step": 33225 |
|
}, |
|
{ |
|
"epoch": 36.26, |
|
"learning_rate": 5.115942028985508e-07, |
|
"loss": 0.4689, |
|
"step": 33250 |
|
}, |
|
{ |
|
"epoch": 36.29, |
|
"learning_rate": 5.043478260869565e-07, |
|
"loss": 0.5219, |
|
"step": 33275 |
|
}, |
|
{ |
|
"epoch": 36.31, |
|
"learning_rate": 4.971014492753624e-07, |
|
"loss": 0.4721, |
|
"step": 33300 |
|
}, |
|
{ |
|
"epoch": 36.34, |
|
"learning_rate": 4.898550724637682e-07, |
|
"loss": 0.5071, |
|
"step": 33325 |
|
}, |
|
{ |
|
"epoch": 36.37, |
|
"learning_rate": 4.82608695652174e-07, |
|
"loss": 0.5119, |
|
"step": 33350 |
|
}, |
|
{ |
|
"epoch": 36.4, |
|
"learning_rate": 4.7536231884057975e-07, |
|
"loss": 0.4618, |
|
"step": 33375 |
|
}, |
|
{ |
|
"epoch": 36.42, |
|
"learning_rate": 4.6811594202898556e-07, |
|
"loss": 0.4778, |
|
"step": 33400 |
|
}, |
|
{ |
|
"epoch": 36.45, |
|
"learning_rate": 4.6086956521739136e-07, |
|
"loss": 0.5451, |
|
"step": 33425 |
|
}, |
|
{ |
|
"epoch": 36.48, |
|
"learning_rate": 4.536231884057971e-07, |
|
"loss": 0.4701, |
|
"step": 33450 |
|
}, |
|
{ |
|
"epoch": 36.5, |
|
"learning_rate": 4.463768115942029e-07, |
|
"loss": 0.4444, |
|
"step": 33475 |
|
}, |
|
{ |
|
"epoch": 36.53, |
|
"learning_rate": 4.391304347826087e-07, |
|
"loss": 0.4971, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 36.56, |
|
"learning_rate": 4.318840579710145e-07, |
|
"loss": 0.5379, |
|
"step": 33525 |
|
}, |
|
{ |
|
"epoch": 36.59, |
|
"learning_rate": 4.246376811594203e-07, |
|
"loss": 0.4883, |
|
"step": 33550 |
|
}, |
|
{ |
|
"epoch": 36.61, |
|
"learning_rate": 4.1739130434782616e-07, |
|
"loss": 0.5602, |
|
"step": 33575 |
|
}, |
|
{ |
|
"epoch": 36.64, |
|
"learning_rate": 4.1014492753623197e-07, |
|
"loss": 0.4486, |
|
"step": 33600 |
|
}, |
|
{ |
|
"epoch": 36.67, |
|
"learning_rate": 4.028985507246377e-07, |
|
"loss": 0.4654, |
|
"step": 33625 |
|
}, |
|
{ |
|
"epoch": 36.7, |
|
"learning_rate": 3.956521739130435e-07, |
|
"loss": 0.4993, |
|
"step": 33650 |
|
}, |
|
{ |
|
"epoch": 36.72, |
|
"learning_rate": 3.884057971014493e-07, |
|
"loss": 0.4552, |
|
"step": 33675 |
|
}, |
|
{ |
|
"epoch": 36.75, |
|
"learning_rate": 3.811594202898551e-07, |
|
"loss": 0.5494, |
|
"step": 33700 |
|
}, |
|
{ |
|
"epoch": 36.78, |
|
"learning_rate": 3.739130434782609e-07, |
|
"loss": 0.5392, |
|
"step": 33725 |
|
}, |
|
{ |
|
"epoch": 36.8, |
|
"learning_rate": 3.666666666666667e-07, |
|
"loss": 0.4993, |
|
"step": 33750 |
|
}, |
|
{ |
|
"epoch": 36.83, |
|
"learning_rate": 3.5942028985507247e-07, |
|
"loss": 0.4896, |
|
"step": 33775 |
|
}, |
|
{ |
|
"epoch": 36.86, |
|
"learning_rate": 3.5217391304347827e-07, |
|
"loss": 0.4804, |
|
"step": 33800 |
|
}, |
|
{ |
|
"epoch": 36.89, |
|
"learning_rate": 3.4492753623188407e-07, |
|
"loss": 0.5006, |
|
"step": 33825 |
|
}, |
|
{ |
|
"epoch": 36.91, |
|
"learning_rate": 3.3768115942028987e-07, |
|
"loss": 0.4965, |
|
"step": 33850 |
|
}, |
|
{ |
|
"epoch": 36.94, |
|
"learning_rate": 3.304347826086957e-07, |
|
"loss": 0.5459, |
|
"step": 33875 |
|
}, |
|
{ |
|
"epoch": 36.97, |
|
"learning_rate": 3.231884057971015e-07, |
|
"loss": 0.4653, |
|
"step": 33900 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"learning_rate": 3.159420289855072e-07, |
|
"loss": 0.5432, |
|
"step": 33925 |
|
}, |
|
{ |
|
"epoch": 37.02, |
|
"learning_rate": 3.086956521739131e-07, |
|
"loss": 0.5017, |
|
"step": 33950 |
|
}, |
|
{ |
|
"epoch": 37.05, |
|
"learning_rate": 3.014492753623189e-07, |
|
"loss": 0.473, |
|
"step": 33975 |
|
}, |
|
{ |
|
"epoch": 37.08, |
|
"learning_rate": 2.942028985507247e-07, |
|
"loss": 0.5007, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 37.08, |
|
"eval_cer": 7.154500918291419, |
|
"eval_loss": 0.4948059916496277, |
|
"eval_runtime": 3148.2863, |
|
"eval_samples_per_second": 0.688, |
|
"eval_steps_per_second": 0.172, |
|
"eval_wer": 22.92340335575199, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 37.1, |
|
"learning_rate": 2.8695652173913043e-07, |
|
"loss": 0.4543, |
|
"step": 34025 |
|
}, |
|
{ |
|
"epoch": 37.13, |
|
"learning_rate": 2.7971014492753623e-07, |
|
"loss": 0.483, |
|
"step": 34050 |
|
}, |
|
{ |
|
"epoch": 37.16, |
|
"learning_rate": 2.724637681159421e-07, |
|
"loss": 0.5062, |
|
"step": 34075 |
|
}, |
|
{ |
|
"epoch": 37.19, |
|
"learning_rate": 2.6521739130434783e-07, |
|
"loss": 0.4614, |
|
"step": 34100 |
|
}, |
|
{ |
|
"epoch": 37.21, |
|
"learning_rate": 2.5797101449275363e-07, |
|
"loss": 0.4805, |
|
"step": 34125 |
|
}, |
|
{ |
|
"epoch": 37.24, |
|
"learning_rate": 2.5072463768115944e-07, |
|
"loss": 0.4723, |
|
"step": 34150 |
|
}, |
|
{ |
|
"epoch": 37.27, |
|
"learning_rate": 2.4347826086956524e-07, |
|
"loss": 0.4788, |
|
"step": 34175 |
|
}, |
|
{ |
|
"epoch": 37.3, |
|
"learning_rate": 2.3623188405797104e-07, |
|
"loss": 0.4736, |
|
"step": 34200 |
|
}, |
|
{ |
|
"epoch": 37.32, |
|
"learning_rate": 2.289855072463768e-07, |
|
"loss": 0.4979, |
|
"step": 34225 |
|
}, |
|
{ |
|
"epoch": 37.35, |
|
"learning_rate": 2.2173913043478261e-07, |
|
"loss": 0.4598, |
|
"step": 34250 |
|
}, |
|
{ |
|
"epoch": 37.38, |
|
"learning_rate": 2.1449275362318844e-07, |
|
"loss": 0.4501, |
|
"step": 34275 |
|
}, |
|
{ |
|
"epoch": 37.4, |
|
"learning_rate": 2.0724637681159422e-07, |
|
"loss": 0.4515, |
|
"step": 34300 |
|
}, |
|
{ |
|
"epoch": 37.43, |
|
"learning_rate": 2.0000000000000002e-07, |
|
"loss": 0.4789, |
|
"step": 34325 |
|
}, |
|
{ |
|
"epoch": 37.46, |
|
"learning_rate": 1.9275362318840582e-07, |
|
"loss": 0.4697, |
|
"step": 34350 |
|
}, |
|
{ |
|
"epoch": 37.49, |
|
"learning_rate": 1.855072463768116e-07, |
|
"loss": 0.468, |
|
"step": 34375 |
|
}, |
|
{ |
|
"epoch": 37.51, |
|
"learning_rate": 1.782608695652174e-07, |
|
"loss": 0.4904, |
|
"step": 34400 |
|
}, |
|
{ |
|
"epoch": 37.54, |
|
"learning_rate": 1.710144927536232e-07, |
|
"loss": 0.4827, |
|
"step": 34425 |
|
}, |
|
{ |
|
"epoch": 37.57, |
|
"learning_rate": 1.6376811594202902e-07, |
|
"loss": 0.4441, |
|
"step": 34450 |
|
}, |
|
{ |
|
"epoch": 37.6, |
|
"learning_rate": 1.565217391304348e-07, |
|
"loss": 0.4665, |
|
"step": 34475 |
|
}, |
|
{ |
|
"epoch": 37.62, |
|
"learning_rate": 1.492753623188406e-07, |
|
"loss": 0.4904, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 37.65, |
|
"learning_rate": 1.420289855072464e-07, |
|
"loss": 0.4366, |
|
"step": 34525 |
|
}, |
|
{ |
|
"epoch": 37.68, |
|
"learning_rate": 1.3478260869565218e-07, |
|
"loss": 0.5191, |
|
"step": 34550 |
|
}, |
|
{ |
|
"epoch": 37.7, |
|
"learning_rate": 1.2753623188405798e-07, |
|
"loss": 0.4901, |
|
"step": 34575 |
|
}, |
|
{ |
|
"epoch": 37.73, |
|
"learning_rate": 1.2028985507246378e-07, |
|
"loss": 0.5176, |
|
"step": 34600 |
|
}, |
|
{ |
|
"epoch": 37.76, |
|
"learning_rate": 1.1304347826086958e-07, |
|
"loss": 0.4337, |
|
"step": 34625 |
|
}, |
|
{ |
|
"epoch": 37.79, |
|
"learning_rate": 1.0579710144927537e-07, |
|
"loss": 0.4485, |
|
"step": 34650 |
|
}, |
|
{ |
|
"epoch": 37.81, |
|
"learning_rate": 9.855072463768116e-08, |
|
"loss": 0.5057, |
|
"step": 34675 |
|
}, |
|
{ |
|
"epoch": 37.84, |
|
"learning_rate": 9.130434782608697e-08, |
|
"loss": 0.4539, |
|
"step": 34700 |
|
}, |
|
{ |
|
"epoch": 37.87, |
|
"learning_rate": 8.405797101449276e-08, |
|
"loss": 0.4845, |
|
"step": 34725 |
|
}, |
|
{ |
|
"epoch": 37.9, |
|
"learning_rate": 7.681159420289856e-08, |
|
"loss": 0.4679, |
|
"step": 34750 |
|
}, |
|
{ |
|
"epoch": 37.92, |
|
"learning_rate": 6.956521739130436e-08, |
|
"loss": 0.5126, |
|
"step": 34775 |
|
}, |
|
{ |
|
"epoch": 37.95, |
|
"learning_rate": 6.231884057971015e-08, |
|
"loss": 0.5234, |
|
"step": 34800 |
|
}, |
|
{ |
|
"epoch": 37.98, |
|
"learning_rate": 5.507246376811595e-08, |
|
"loss": 0.4573, |
|
"step": 34825 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"learning_rate": 4.782608695652174e-08, |
|
"loss": 0.4481, |
|
"step": 34850 |
|
}, |
|
{ |
|
"epoch": 38.03, |
|
"learning_rate": 4.057971014492754e-08, |
|
"loss": 0.4473, |
|
"step": 34875 |
|
}, |
|
{ |
|
"epoch": 38.06, |
|
"learning_rate": 3.333333333333334e-08, |
|
"loss": 0.4978, |
|
"step": 34900 |
|
}, |
|
{ |
|
"epoch": 38.09, |
|
"learning_rate": 2.6086956521739135e-08, |
|
"loss": 0.4456, |
|
"step": 34925 |
|
}, |
|
{ |
|
"epoch": 38.11, |
|
"learning_rate": 1.884057971014493e-08, |
|
"loss": 0.47, |
|
"step": 34950 |
|
}, |
|
{ |
|
"epoch": 38.14, |
|
"learning_rate": 1.1594202898550725e-08, |
|
"loss": 0.4272, |
|
"step": 34975 |
|
}, |
|
{ |
|
"epoch": 38.17, |
|
"learning_rate": 4.3478260869565225e-09, |
|
"loss": 0.4519, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 38.17, |
|
"eval_cer": 7.125683342150651, |
|
"eval_loss": 0.49221470952033997, |
|
"eval_runtime": 3147.7575, |
|
"eval_samples_per_second": 0.688, |
|
"eval_steps_per_second": 0.172, |
|
"eval_wer": 22.824750824061823, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 38.2, |
|
"learning_rate": 2.244943820224719e-06, |
|
"loss": 0.5112, |
|
"step": 35025 |
|
}, |
|
{ |
|
"epoch": 38.22, |
|
"learning_rate": 2.2393258426966294e-06, |
|
"loss": 0.4629, |
|
"step": 35050 |
|
}, |
|
{ |
|
"epoch": 38.25, |
|
"learning_rate": 2.2337078651685396e-06, |
|
"loss": 0.4797, |
|
"step": 35075 |
|
}, |
|
{ |
|
"epoch": 38.28, |
|
"learning_rate": 2.22808988764045e-06, |
|
"loss": 0.4799, |
|
"step": 35100 |
|
}, |
|
{ |
|
"epoch": 38.3, |
|
"learning_rate": 2.22247191011236e-06, |
|
"loss": 0.4919, |
|
"step": 35125 |
|
}, |
|
{ |
|
"epoch": 38.33, |
|
"learning_rate": 2.21685393258427e-06, |
|
"loss": 0.4669, |
|
"step": 35150 |
|
}, |
|
{ |
|
"epoch": 38.36, |
|
"learning_rate": 2.21123595505618e-06, |
|
"loss": 0.4348, |
|
"step": 35175 |
|
}, |
|
{ |
|
"epoch": 38.39, |
|
"learning_rate": 2.20561797752809e-06, |
|
"loss": 0.5559, |
|
"step": 35200 |
|
}, |
|
{ |
|
"epoch": 38.41, |
|
"learning_rate": 2.2e-06, |
|
"loss": 0.4651, |
|
"step": 35225 |
|
}, |
|
{ |
|
"epoch": 38.44, |
|
"learning_rate": 2.1943820224719103e-06, |
|
"loss": 0.4023, |
|
"step": 35250 |
|
}, |
|
{ |
|
"epoch": 38.47, |
|
"learning_rate": 2.1887640449438206e-06, |
|
"loss": 0.4687, |
|
"step": 35275 |
|
}, |
|
{ |
|
"epoch": 38.5, |
|
"learning_rate": 2.183146067415731e-06, |
|
"loss": 0.4999, |
|
"step": 35300 |
|
}, |
|
{ |
|
"epoch": 38.52, |
|
"learning_rate": 2.1775280898876406e-06, |
|
"loss": 0.431, |
|
"step": 35325 |
|
}, |
|
{ |
|
"epoch": 38.55, |
|
"learning_rate": 2.171910112359551e-06, |
|
"loss": 0.4887, |
|
"step": 35350 |
|
}, |
|
{ |
|
"epoch": 38.58, |
|
"learning_rate": 2.1662921348314606e-06, |
|
"loss": 0.4401, |
|
"step": 35375 |
|
}, |
|
{ |
|
"epoch": 38.6, |
|
"learning_rate": 2.160674157303371e-06, |
|
"loss": 0.4975, |
|
"step": 35400 |
|
}, |
|
{ |
|
"epoch": 38.63, |
|
"learning_rate": 2.155056179775281e-06, |
|
"loss": 0.4447, |
|
"step": 35425 |
|
}, |
|
{ |
|
"epoch": 38.66, |
|
"learning_rate": 2.1494382022471913e-06, |
|
"loss": 0.5491, |
|
"step": 35450 |
|
}, |
|
{ |
|
"epoch": 38.69, |
|
"learning_rate": 2.143820224719101e-06, |
|
"loss": 0.4185, |
|
"step": 35475 |
|
}, |
|
{ |
|
"epoch": 38.71, |
|
"learning_rate": 2.1382022471910113e-06, |
|
"loss": 0.4195, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 38.74, |
|
"learning_rate": 2.1325842696629216e-06, |
|
"loss": 0.5616, |
|
"step": 35525 |
|
}, |
|
{ |
|
"epoch": 38.77, |
|
"learning_rate": 2.126966292134832e-06, |
|
"loss": 0.4316, |
|
"step": 35550 |
|
}, |
|
{ |
|
"epoch": 38.79, |
|
"learning_rate": 2.1213483146067416e-06, |
|
"loss": 0.4601, |
|
"step": 35575 |
|
}, |
|
{ |
|
"epoch": 38.82, |
|
"learning_rate": 2.1159550561797753e-06, |
|
"loss": 0.4268, |
|
"step": 35600 |
|
}, |
|
{ |
|
"epoch": 38.85, |
|
"learning_rate": 2.1103370786516855e-06, |
|
"loss": 0.4646, |
|
"step": 35625 |
|
}, |
|
{ |
|
"epoch": 38.88, |
|
"learning_rate": 2.1047191011235957e-06, |
|
"loss": 0.4052, |
|
"step": 35650 |
|
}, |
|
{ |
|
"epoch": 38.9, |
|
"learning_rate": 2.099101123595506e-06, |
|
"loss": 0.4437, |
|
"step": 35675 |
|
}, |
|
{ |
|
"epoch": 38.93, |
|
"learning_rate": 2.0934831460674158e-06, |
|
"loss": 0.4204, |
|
"step": 35700 |
|
}, |
|
{ |
|
"epoch": 38.96, |
|
"learning_rate": 2.087865168539326e-06, |
|
"loss": 0.4344, |
|
"step": 35725 |
|
}, |
|
{ |
|
"epoch": 38.99, |
|
"learning_rate": 2.082247191011236e-06, |
|
"loss": 0.5028, |
|
"step": 35750 |
|
}, |
|
{ |
|
"epoch": 39.01, |
|
"learning_rate": 2.076629213483146e-06, |
|
"loss": 0.3716, |
|
"step": 35775 |
|
}, |
|
{ |
|
"epoch": 39.04, |
|
"learning_rate": 2.0710112359550563e-06, |
|
"loss": 0.4251, |
|
"step": 35800 |
|
}, |
|
{ |
|
"epoch": 39.07, |
|
"learning_rate": 2.0653932584269665e-06, |
|
"loss": 0.4155, |
|
"step": 35825 |
|
}, |
|
{ |
|
"epoch": 39.09, |
|
"learning_rate": 2.0597752808988767e-06, |
|
"loss": 0.4249, |
|
"step": 35850 |
|
}, |
|
{ |
|
"epoch": 39.12, |
|
"learning_rate": 2.054157303370787e-06, |
|
"loss": 0.3721, |
|
"step": 35875 |
|
}, |
|
{ |
|
"epoch": 39.15, |
|
"learning_rate": 2.0485393258426968e-06, |
|
"loss": 0.4026, |
|
"step": 35900 |
|
}, |
|
{ |
|
"epoch": 39.18, |
|
"learning_rate": 2.042921348314607e-06, |
|
"loss": 0.4425, |
|
"step": 35925 |
|
}, |
|
{ |
|
"epoch": 39.2, |
|
"learning_rate": 2.037303370786517e-06, |
|
"loss": 0.399, |
|
"step": 35950 |
|
}, |
|
{ |
|
"epoch": 39.23, |
|
"learning_rate": 2.031685393258427e-06, |
|
"loss": 0.4291, |
|
"step": 35975 |
|
}, |
|
{ |
|
"epoch": 39.26, |
|
"learning_rate": 2.0260674157303372e-06, |
|
"loss": 0.3674, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 39.26, |
|
"eval_cer": 7.010413037587582, |
|
"eval_loss": 0.47536423802375793, |
|
"eval_runtime": 3145.4396, |
|
"eval_samples_per_second": 0.689, |
|
"eval_steps_per_second": 0.172, |
|
"eval_wer": 22.664244720915118, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 39.29, |
|
"learning_rate": 2.0204494382022475e-06, |
|
"loss": 0.3957, |
|
"step": 36025 |
|
}, |
|
{ |
|
"epoch": 39.31, |
|
"learning_rate": 2.0148314606741577e-06, |
|
"loss": 0.3944, |
|
"step": 36050 |
|
}, |
|
{ |
|
"epoch": 39.34, |
|
"learning_rate": 2.0092134831460675e-06, |
|
"loss": 0.4103, |
|
"step": 36075 |
|
}, |
|
{ |
|
"epoch": 39.37, |
|
"learning_rate": 2.0035955056179777e-06, |
|
"loss": 0.3942, |
|
"step": 36100 |
|
}, |
|
{ |
|
"epoch": 39.39, |
|
"learning_rate": 1.997977528089888e-06, |
|
"loss": 0.426, |
|
"step": 36125 |
|
}, |
|
{ |
|
"epoch": 39.42, |
|
"learning_rate": 1.9923595505617978e-06, |
|
"loss": 0.3685, |
|
"step": 36150 |
|
}, |
|
{ |
|
"epoch": 39.45, |
|
"learning_rate": 1.986741573033708e-06, |
|
"loss": 0.4251, |
|
"step": 36175 |
|
}, |
|
{ |
|
"epoch": 39.48, |
|
"learning_rate": 1.9811235955056182e-06, |
|
"loss": 0.3852, |
|
"step": 36200 |
|
}, |
|
{ |
|
"epoch": 39.5, |
|
"learning_rate": 1.9755056179775285e-06, |
|
"loss": 0.3743, |
|
"step": 36225 |
|
}, |
|
{ |
|
"epoch": 39.53, |
|
"learning_rate": 1.9698876404494383e-06, |
|
"loss": 0.4061, |
|
"step": 36250 |
|
}, |
|
{ |
|
"epoch": 39.56, |
|
"learning_rate": 1.9642696629213485e-06, |
|
"loss": 0.4062, |
|
"step": 36275 |
|
}, |
|
{ |
|
"epoch": 39.59, |
|
"learning_rate": 1.9586516853932587e-06, |
|
"loss": 0.3523, |
|
"step": 36300 |
|
}, |
|
{ |
|
"epoch": 39.61, |
|
"learning_rate": 1.9530337078651685e-06, |
|
"loss": 0.4023, |
|
"step": 36325 |
|
}, |
|
{ |
|
"epoch": 39.64, |
|
"learning_rate": 1.9474157303370787e-06, |
|
"loss": 0.3799, |
|
"step": 36350 |
|
}, |
|
{ |
|
"epoch": 39.67, |
|
"learning_rate": 1.941797752808989e-06, |
|
"loss": 0.3439, |
|
"step": 36375 |
|
}, |
|
{ |
|
"epoch": 39.69, |
|
"learning_rate": 1.9361797752808988e-06, |
|
"loss": 0.4106, |
|
"step": 36400 |
|
}, |
|
{ |
|
"epoch": 39.72, |
|
"learning_rate": 1.930561797752809e-06, |
|
"loss": 0.407, |
|
"step": 36425 |
|
}, |
|
{ |
|
"epoch": 39.75, |
|
"learning_rate": 1.9249438202247192e-06, |
|
"loss": 0.3814, |
|
"step": 36450 |
|
}, |
|
{ |
|
"epoch": 39.78, |
|
"learning_rate": 1.9193258426966295e-06, |
|
"loss": 0.3561, |
|
"step": 36475 |
|
}, |
|
{ |
|
"epoch": 39.8, |
|
"learning_rate": 1.9137078651685397e-06, |
|
"loss": 0.3886, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 39.83, |
|
"learning_rate": 1.9080898876404495e-06, |
|
"loss": 0.3777, |
|
"step": 36525 |
|
}, |
|
{ |
|
"epoch": 39.86, |
|
"learning_rate": 1.9024719101123597e-06, |
|
"loss": 0.39, |
|
"step": 36550 |
|
}, |
|
{ |
|
"epoch": 39.89, |
|
"learning_rate": 1.8968539325842697e-06, |
|
"loss": 0.3945, |
|
"step": 36575 |
|
}, |
|
{ |
|
"epoch": 39.91, |
|
"learning_rate": 1.8912359550561798e-06, |
|
"loss": 0.3531, |
|
"step": 36600 |
|
}, |
|
{ |
|
"epoch": 39.94, |
|
"learning_rate": 1.88561797752809e-06, |
|
"loss": 0.4433, |
|
"step": 36625 |
|
}, |
|
{ |
|
"epoch": 39.97, |
|
"learning_rate": 1.8800000000000002e-06, |
|
"loss": 0.4188, |
|
"step": 36650 |
|
}, |
|
{ |
|
"epoch": 39.99, |
|
"learning_rate": 1.8743820224719102e-06, |
|
"loss": 0.3897, |
|
"step": 36675 |
|
}, |
|
{ |
|
"epoch": 40.02, |
|
"learning_rate": 1.8687640449438205e-06, |
|
"loss": 0.3282, |
|
"step": 36700 |
|
}, |
|
{ |
|
"epoch": 40.05, |
|
"learning_rate": 1.8631460674157307e-06, |
|
"loss": 0.341, |
|
"step": 36725 |
|
}, |
|
{ |
|
"epoch": 40.08, |
|
"learning_rate": 1.8575280898876405e-06, |
|
"loss": 0.3239, |
|
"step": 36750 |
|
}, |
|
{ |
|
"epoch": 40.1, |
|
"learning_rate": 1.8519101123595507e-06, |
|
"loss": 0.3518, |
|
"step": 36775 |
|
}, |
|
{ |
|
"epoch": 40.13, |
|
"learning_rate": 1.8462921348314607e-06, |
|
"loss": 0.3387, |
|
"step": 36800 |
|
}, |
|
{ |
|
"epoch": 40.16, |
|
"learning_rate": 1.840674157303371e-06, |
|
"loss": 0.3201, |
|
"step": 36825 |
|
}, |
|
{ |
|
"epoch": 40.19, |
|
"learning_rate": 1.8350561797752812e-06, |
|
"loss": 0.3128, |
|
"step": 36850 |
|
}, |
|
{ |
|
"epoch": 40.21, |
|
"learning_rate": 1.8294382022471912e-06, |
|
"loss": 0.3749, |
|
"step": 36875 |
|
}, |
|
{ |
|
"epoch": 40.24, |
|
"learning_rate": 1.8238202247191012e-06, |
|
"loss": 0.3756, |
|
"step": 36900 |
|
}, |
|
{ |
|
"epoch": 40.27, |
|
"learning_rate": 1.8182022471910112e-06, |
|
"loss": 0.3545, |
|
"step": 36925 |
|
}, |
|
{ |
|
"epoch": 40.29, |
|
"learning_rate": 1.8125842696629215e-06, |
|
"loss": 0.3746, |
|
"step": 36950 |
|
}, |
|
{ |
|
"epoch": 40.32, |
|
"learning_rate": 1.8069662921348317e-06, |
|
"loss": 0.3308, |
|
"step": 36975 |
|
}, |
|
{ |
|
"epoch": 40.35, |
|
"learning_rate": 1.8013483146067417e-06, |
|
"loss": 0.3481, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 40.35, |
|
"eval_cer": 7.031058465270519, |
|
"eval_loss": 0.46790996193885803, |
|
"eval_runtime": 3161.6657, |
|
"eval_samples_per_second": 0.685, |
|
"eval_steps_per_second": 0.171, |
|
"eval_wer": 22.631360543685062, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 40.38, |
|
"learning_rate": 1.795730337078652e-06, |
|
"loss": 0.3225, |
|
"step": 37025 |
|
}, |
|
{ |
|
"epoch": 40.4, |
|
"learning_rate": 1.7901123595505622e-06, |
|
"loss": 0.299, |
|
"step": 37050 |
|
}, |
|
{ |
|
"epoch": 40.43, |
|
"learning_rate": 1.784494382022472e-06, |
|
"loss": 0.3152, |
|
"step": 37075 |
|
}, |
|
{ |
|
"epoch": 40.46, |
|
"learning_rate": 1.778876404494382e-06, |
|
"loss": 0.3268, |
|
"step": 37100 |
|
}, |
|
{ |
|
"epoch": 40.49, |
|
"learning_rate": 1.7732584269662922e-06, |
|
"loss": 0.3411, |
|
"step": 37125 |
|
}, |
|
{ |
|
"epoch": 40.51, |
|
"learning_rate": 1.7676404494382024e-06, |
|
"loss": 0.3846, |
|
"step": 37150 |
|
}, |
|
{ |
|
"epoch": 40.54, |
|
"learning_rate": 1.7620224719101125e-06, |
|
"loss": 0.3168, |
|
"step": 37175 |
|
}, |
|
{ |
|
"epoch": 40.57, |
|
"learning_rate": 1.7564044943820227e-06, |
|
"loss": 0.3377, |
|
"step": 37200 |
|
}, |
|
{ |
|
"epoch": 40.59, |
|
"learning_rate": 1.750786516853933e-06, |
|
"loss": 0.3618, |
|
"step": 37225 |
|
}, |
|
{ |
|
"epoch": 40.62, |
|
"learning_rate": 1.7451685393258427e-06, |
|
"loss": 0.3215, |
|
"step": 37250 |
|
}, |
|
{ |
|
"epoch": 40.65, |
|
"learning_rate": 1.739550561797753e-06, |
|
"loss": 0.3859, |
|
"step": 37275 |
|
}, |
|
{ |
|
"epoch": 40.68, |
|
"learning_rate": 1.733932584269663e-06, |
|
"loss": 0.3484, |
|
"step": 37300 |
|
}, |
|
{ |
|
"epoch": 40.7, |
|
"learning_rate": 1.7283146067415732e-06, |
|
"loss": 0.375, |
|
"step": 37325 |
|
}, |
|
{ |
|
"epoch": 40.73, |
|
"learning_rate": 1.7226966292134834e-06, |
|
"loss": 0.327, |
|
"step": 37350 |
|
}, |
|
{ |
|
"epoch": 40.76, |
|
"learning_rate": 1.7170786516853934e-06, |
|
"loss": 0.2798, |
|
"step": 37375 |
|
}, |
|
{ |
|
"epoch": 40.79, |
|
"learning_rate": 1.7114606741573035e-06, |
|
"loss": 0.3273, |
|
"step": 37400 |
|
}, |
|
{ |
|
"epoch": 40.81, |
|
"learning_rate": 1.7058426966292135e-06, |
|
"loss": 0.3762, |
|
"step": 37425 |
|
}, |
|
{ |
|
"epoch": 40.84, |
|
"learning_rate": 1.7002247191011237e-06, |
|
"loss": 0.2886, |
|
"step": 37450 |
|
}, |
|
{ |
|
"epoch": 40.87, |
|
"learning_rate": 1.6948314606741576e-06, |
|
"loss": 0.3368, |
|
"step": 37475 |
|
}, |
|
{ |
|
"epoch": 40.89, |
|
"learning_rate": 1.6892134831460674e-06, |
|
"loss": 0.3812, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 40.92, |
|
"learning_rate": 1.6835955056179776e-06, |
|
"loss": 0.3301, |
|
"step": 37525 |
|
}, |
|
{ |
|
"epoch": 40.95, |
|
"learning_rate": 1.6779775280898876e-06, |
|
"loss": 0.3193, |
|
"step": 37550 |
|
}, |
|
{ |
|
"epoch": 40.98, |
|
"learning_rate": 1.6723595505617979e-06, |
|
"loss": 0.3263, |
|
"step": 37575 |
|
}, |
|
{ |
|
"epoch": 41.0, |
|
"learning_rate": 1.666741573033708e-06, |
|
"loss": 0.3422, |
|
"step": 37600 |
|
}, |
|
{ |
|
"epoch": 41.03, |
|
"learning_rate": 1.6611235955056181e-06, |
|
"loss": 0.3138, |
|
"step": 37625 |
|
}, |
|
{ |
|
"epoch": 41.06, |
|
"learning_rate": 1.6555056179775283e-06, |
|
"loss": 0.2919, |
|
"step": 37650 |
|
}, |
|
{ |
|
"epoch": 41.09, |
|
"learning_rate": 1.6498876404494381e-06, |
|
"loss": 0.2828, |
|
"step": 37675 |
|
}, |
|
{ |
|
"epoch": 41.11, |
|
"learning_rate": 1.6442696629213484e-06, |
|
"loss": 0.2607, |
|
"step": 37700 |
|
}, |
|
{ |
|
"epoch": 41.14, |
|
"learning_rate": 1.6386516853932586e-06, |
|
"loss": 0.2524, |
|
"step": 37725 |
|
}, |
|
{ |
|
"epoch": 41.17, |
|
"learning_rate": 1.6330337078651686e-06, |
|
"loss": 0.3627, |
|
"step": 37750 |
|
}, |
|
{ |
|
"epoch": 41.19, |
|
"learning_rate": 1.6274157303370788e-06, |
|
"loss": 0.2641, |
|
"step": 37775 |
|
}, |
|
{ |
|
"epoch": 41.22, |
|
"learning_rate": 1.621797752808989e-06, |
|
"loss": 0.2859, |
|
"step": 37800 |
|
}, |
|
{ |
|
"epoch": 41.25, |
|
"learning_rate": 1.616179775280899e-06, |
|
"loss": 0.277, |
|
"step": 37825 |
|
}, |
|
{ |
|
"epoch": 41.28, |
|
"learning_rate": 1.610561797752809e-06, |
|
"loss": 0.2923, |
|
"step": 37850 |
|
}, |
|
{ |
|
"epoch": 41.3, |
|
"learning_rate": 1.6049438202247191e-06, |
|
"loss": 0.3439, |
|
"step": 37875 |
|
}, |
|
{ |
|
"epoch": 41.33, |
|
"learning_rate": 1.5993258426966293e-06, |
|
"loss": 0.2374, |
|
"step": 37900 |
|
}, |
|
{ |
|
"epoch": 41.36, |
|
"learning_rate": 1.5937078651685396e-06, |
|
"loss": 0.3332, |
|
"step": 37925 |
|
}, |
|
{ |
|
"epoch": 41.38, |
|
"learning_rate": 1.5880898876404496e-06, |
|
"loss": 0.2999, |
|
"step": 37950 |
|
}, |
|
{ |
|
"epoch": 41.41, |
|
"learning_rate": 1.5824719101123598e-06, |
|
"loss": 0.3052, |
|
"step": 37975 |
|
}, |
|
{ |
|
"epoch": 41.44, |
|
"learning_rate": 1.5768539325842696e-06, |
|
"loss": 0.2992, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 41.44, |
|
"eval_cer": 6.946469560180704, |
|
"eval_loss": 0.4622255265712738, |
|
"eval_runtime": 3167.1342, |
|
"eval_samples_per_second": 0.684, |
|
"eval_steps_per_second": 0.171, |
|
"eval_wer": 22.25945615834514, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 41.47, |
|
"learning_rate": 1.5712359550561799e-06, |
|
"loss": 0.2935, |
|
"step": 38025 |
|
}, |
|
{ |
|
"epoch": 41.49, |
|
"learning_rate": 1.5656179775280899e-06, |
|
"loss": 0.3086, |
|
"step": 38050 |
|
}, |
|
{ |
|
"epoch": 41.52, |
|
"learning_rate": 1.56e-06, |
|
"loss": 0.301, |
|
"step": 38075 |
|
}, |
|
{ |
|
"epoch": 41.55, |
|
"learning_rate": 1.5543820224719103e-06, |
|
"loss": 0.2848, |
|
"step": 38100 |
|
}, |
|
{ |
|
"epoch": 41.58, |
|
"learning_rate": 1.5487640449438203e-06, |
|
"loss": 0.2642, |
|
"step": 38125 |
|
}, |
|
{ |
|
"epoch": 41.6, |
|
"learning_rate": 1.5431460674157306e-06, |
|
"loss": 0.3192, |
|
"step": 38150 |
|
}, |
|
{ |
|
"epoch": 41.63, |
|
"learning_rate": 1.5375280898876404e-06, |
|
"loss": 0.2712, |
|
"step": 38175 |
|
}, |
|
{ |
|
"epoch": 41.66, |
|
"learning_rate": 1.5319101123595506e-06, |
|
"loss": 0.2649, |
|
"step": 38200 |
|
}, |
|
{ |
|
"epoch": 41.68, |
|
"learning_rate": 1.5262921348314608e-06, |
|
"loss": 0.2872, |
|
"step": 38225 |
|
}, |
|
{ |
|
"epoch": 41.71, |
|
"learning_rate": 1.5206741573033708e-06, |
|
"loss": 0.3132, |
|
"step": 38250 |
|
}, |
|
{ |
|
"epoch": 41.74, |
|
"learning_rate": 1.515056179775281e-06, |
|
"loss": 0.2831, |
|
"step": 38275 |
|
}, |
|
{ |
|
"epoch": 41.77, |
|
"learning_rate": 1.5094382022471913e-06, |
|
"loss": 0.2998, |
|
"step": 38300 |
|
}, |
|
{ |
|
"epoch": 41.79, |
|
"learning_rate": 1.5038202247191013e-06, |
|
"loss": 0.3153, |
|
"step": 38325 |
|
}, |
|
{ |
|
"epoch": 41.82, |
|
"learning_rate": 1.4982022471910113e-06, |
|
"loss": 0.2693, |
|
"step": 38350 |
|
}, |
|
{ |
|
"epoch": 41.85, |
|
"learning_rate": 1.4925842696629214e-06, |
|
"loss": 0.3361, |
|
"step": 38375 |
|
}, |
|
{ |
|
"epoch": 41.88, |
|
"learning_rate": 1.4869662921348316e-06, |
|
"loss": 0.3173, |
|
"step": 38400 |
|
}, |
|
{ |
|
"epoch": 41.9, |
|
"learning_rate": 1.4813483146067418e-06, |
|
"loss": 0.2723, |
|
"step": 38425 |
|
}, |
|
{ |
|
"epoch": 41.93, |
|
"learning_rate": 1.4757303370786518e-06, |
|
"loss": 0.3004, |
|
"step": 38450 |
|
}, |
|
{ |
|
"epoch": 41.96, |
|
"learning_rate": 1.470112359550562e-06, |
|
"loss": 0.3207, |
|
"step": 38475 |
|
}, |
|
{ |
|
"epoch": 41.98, |
|
"learning_rate": 1.4644943820224719e-06, |
|
"loss": 0.2761, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 42.01, |
|
"learning_rate": 1.458876404494382e-06, |
|
"loss": 0.3081, |
|
"step": 38525 |
|
}, |
|
{ |
|
"epoch": 42.04, |
|
"learning_rate": 1.4532584269662923e-06, |
|
"loss": 0.2535, |
|
"step": 38550 |
|
}, |
|
{ |
|
"epoch": 42.07, |
|
"learning_rate": 1.4476404494382023e-06, |
|
"loss": 0.2461, |
|
"step": 38575 |
|
}, |
|
{ |
|
"epoch": 42.09, |
|
"learning_rate": 1.4420224719101126e-06, |
|
"loss": 0.2822, |
|
"step": 38600 |
|
}, |
|
{ |
|
"epoch": 42.12, |
|
"learning_rate": 1.4364044943820226e-06, |
|
"loss": 0.2379, |
|
"step": 38625 |
|
}, |
|
{ |
|
"epoch": 42.15, |
|
"learning_rate": 1.4307865168539328e-06, |
|
"loss": 0.2468, |
|
"step": 38650 |
|
}, |
|
{ |
|
"epoch": 42.18, |
|
"learning_rate": 1.4251685393258426e-06, |
|
"loss": 0.2701, |
|
"step": 38675 |
|
}, |
|
{ |
|
"epoch": 42.2, |
|
"learning_rate": 1.4195505617977528e-06, |
|
"loss": 0.258, |
|
"step": 38700 |
|
}, |
|
{ |
|
"epoch": 42.23, |
|
"learning_rate": 1.413932584269663e-06, |
|
"loss": 0.2449, |
|
"step": 38725 |
|
}, |
|
{ |
|
"epoch": 42.26, |
|
"learning_rate": 1.408314606741573e-06, |
|
"loss": 0.2615, |
|
"step": 38750 |
|
}, |
|
{ |
|
"epoch": 42.28, |
|
"learning_rate": 1.4026966292134833e-06, |
|
"loss": 0.2735, |
|
"step": 38775 |
|
}, |
|
{ |
|
"epoch": 42.31, |
|
"learning_rate": 1.3970786516853935e-06, |
|
"loss": 0.2429, |
|
"step": 38800 |
|
}, |
|
{ |
|
"epoch": 42.34, |
|
"learning_rate": 1.3914606741573033e-06, |
|
"loss": 0.2835, |
|
"step": 38825 |
|
}, |
|
{ |
|
"epoch": 42.37, |
|
"learning_rate": 1.3858426966292136e-06, |
|
"loss": 0.2713, |
|
"step": 38850 |
|
}, |
|
{ |
|
"epoch": 42.39, |
|
"learning_rate": 1.3802247191011236e-06, |
|
"loss": 0.2523, |
|
"step": 38875 |
|
}, |
|
{ |
|
"epoch": 42.42, |
|
"learning_rate": 1.3746067415730338e-06, |
|
"loss": 0.3313, |
|
"step": 38900 |
|
}, |
|
{ |
|
"epoch": 42.45, |
|
"learning_rate": 1.368988764044944e-06, |
|
"loss": 0.243, |
|
"step": 38925 |
|
}, |
|
{ |
|
"epoch": 42.48, |
|
"learning_rate": 1.363370786516854e-06, |
|
"loss": 0.2526, |
|
"step": 38950 |
|
}, |
|
{ |
|
"epoch": 42.5, |
|
"learning_rate": 1.3577528089887643e-06, |
|
"loss": 0.2498, |
|
"step": 38975 |
|
}, |
|
{ |
|
"epoch": 42.53, |
|
"learning_rate": 1.352134831460674e-06, |
|
"loss": 0.2505, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 42.53, |
|
"eval_cer": 6.9198025494235775, |
|
"eval_loss": 0.4640973210334778, |
|
"eval_runtime": 3164.4611, |
|
"eval_samples_per_second": 0.684, |
|
"eval_steps_per_second": 0.171, |
|
"eval_wer": 22.19368780388503, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 42.56, |
|
"learning_rate": 1.3465168539325843e-06, |
|
"loss": 0.2524, |
|
"step": 39025 |
|
}, |
|
{ |
|
"epoch": 42.58, |
|
"learning_rate": 1.3408988764044945e-06, |
|
"loss": 0.2544, |
|
"step": 39050 |
|
}, |
|
{ |
|
"epoch": 42.61, |
|
"learning_rate": 1.3352808988764046e-06, |
|
"loss": 0.2662, |
|
"step": 39075 |
|
}, |
|
{ |
|
"epoch": 42.64, |
|
"learning_rate": 1.3296629213483148e-06, |
|
"loss": 0.2585, |
|
"step": 39100 |
|
}, |
|
{ |
|
"epoch": 42.67, |
|
"learning_rate": 1.324044943820225e-06, |
|
"loss": 0.2696, |
|
"step": 39125 |
|
}, |
|
{ |
|
"epoch": 42.69, |
|
"learning_rate": 1.318426966292135e-06, |
|
"loss": 0.2485, |
|
"step": 39150 |
|
}, |
|
{ |
|
"epoch": 42.72, |
|
"learning_rate": 1.312808988764045e-06, |
|
"loss": 0.2799, |
|
"step": 39175 |
|
}, |
|
{ |
|
"epoch": 42.75, |
|
"learning_rate": 1.307191011235955e-06, |
|
"loss": 0.2432, |
|
"step": 39200 |
|
}, |
|
{ |
|
"epoch": 42.78, |
|
"learning_rate": 1.3015730337078653e-06, |
|
"loss": 0.2511, |
|
"step": 39225 |
|
}, |
|
{ |
|
"epoch": 42.8, |
|
"learning_rate": 1.2959550561797753e-06, |
|
"loss": 0.3178, |
|
"step": 39250 |
|
}, |
|
{ |
|
"epoch": 42.83, |
|
"learning_rate": 1.2903370786516855e-06, |
|
"loss": 0.2375, |
|
"step": 39275 |
|
}, |
|
{ |
|
"epoch": 42.86, |
|
"learning_rate": 1.2847191011235958e-06, |
|
"loss": 0.2477, |
|
"step": 39300 |
|
}, |
|
{ |
|
"epoch": 42.88, |
|
"learning_rate": 1.2791011235955056e-06, |
|
"loss": 0.2577, |
|
"step": 39325 |
|
}, |
|
{ |
|
"epoch": 42.91, |
|
"learning_rate": 1.2734831460674158e-06, |
|
"loss": 0.2973, |
|
"step": 39350 |
|
}, |
|
{ |
|
"epoch": 42.94, |
|
"learning_rate": 1.2678651685393258e-06, |
|
"loss": 0.3231, |
|
"step": 39375 |
|
}, |
|
{ |
|
"epoch": 42.97, |
|
"learning_rate": 1.262247191011236e-06, |
|
"loss": 0.2593, |
|
"step": 39400 |
|
}, |
|
{ |
|
"epoch": 42.99, |
|
"learning_rate": 1.2566292134831463e-06, |
|
"loss": 0.2557, |
|
"step": 39425 |
|
}, |
|
{ |
|
"epoch": 43.02, |
|
"learning_rate": 1.2510112359550563e-06, |
|
"loss": 0.248, |
|
"step": 39450 |
|
}, |
|
{ |
|
"epoch": 43.05, |
|
"learning_rate": 1.2453932584269663e-06, |
|
"loss": 0.2346, |
|
"step": 39475 |
|
}, |
|
{ |
|
"epoch": 43.08, |
|
"learning_rate": 1.2397752808988765e-06, |
|
"loss": 0.2541, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 43.1, |
|
"learning_rate": 1.2341573033707868e-06, |
|
"loss": 0.2447, |
|
"step": 39525 |
|
}, |
|
{ |
|
"epoch": 43.13, |
|
"learning_rate": 1.2285393258426968e-06, |
|
"loss": 0.2534, |
|
"step": 39550 |
|
}, |
|
{ |
|
"epoch": 43.16, |
|
"learning_rate": 1.2229213483146068e-06, |
|
"loss": 0.2688, |
|
"step": 39575 |
|
}, |
|
{ |
|
"epoch": 43.18, |
|
"learning_rate": 1.217303370786517e-06, |
|
"loss": 0.2358, |
|
"step": 39600 |
|
}, |
|
{ |
|
"epoch": 43.21, |
|
"learning_rate": 1.211685393258427e-06, |
|
"loss": 0.2607, |
|
"step": 39625 |
|
}, |
|
{ |
|
"epoch": 43.24, |
|
"learning_rate": 1.2060674157303373e-06, |
|
"loss": 0.2362, |
|
"step": 39650 |
|
}, |
|
{ |
|
"epoch": 43.27, |
|
"learning_rate": 1.2004494382022473e-06, |
|
"loss": 0.2014, |
|
"step": 39675 |
|
}, |
|
{ |
|
"epoch": 43.29, |
|
"learning_rate": 1.1948314606741573e-06, |
|
"loss": 0.243, |
|
"step": 39700 |
|
}, |
|
{ |
|
"epoch": 43.32, |
|
"learning_rate": 1.1892134831460675e-06, |
|
"loss": 0.2331, |
|
"step": 39725 |
|
}, |
|
{ |
|
"epoch": 43.35, |
|
"learning_rate": 1.1835955056179778e-06, |
|
"loss": 0.2281, |
|
"step": 39750 |
|
}, |
|
{ |
|
"epoch": 43.38, |
|
"learning_rate": 1.1779775280898878e-06, |
|
"loss": 0.2671, |
|
"step": 39775 |
|
}, |
|
{ |
|
"epoch": 43.4, |
|
"learning_rate": 1.1723595505617978e-06, |
|
"loss": 0.2365, |
|
"step": 39800 |
|
}, |
|
{ |
|
"epoch": 43.43, |
|
"learning_rate": 1.166741573033708e-06, |
|
"loss": 0.2356, |
|
"step": 39825 |
|
}, |
|
{ |
|
"epoch": 43.46, |
|
"learning_rate": 1.161123595505618e-06, |
|
"loss": 0.2366, |
|
"step": 39850 |
|
}, |
|
{ |
|
"epoch": 43.48, |
|
"learning_rate": 1.155505617977528e-06, |
|
"loss": 0.2345, |
|
"step": 39875 |
|
}, |
|
{ |
|
"epoch": 43.51, |
|
"learning_rate": 1.1498876404494383e-06, |
|
"loss": 0.2239, |
|
"step": 39900 |
|
}, |
|
{ |
|
"epoch": 43.54, |
|
"learning_rate": 1.1442696629213485e-06, |
|
"loss": 0.2391, |
|
"step": 39925 |
|
}, |
|
{ |
|
"epoch": 43.57, |
|
"learning_rate": 1.1386516853932585e-06, |
|
"loss": 0.2517, |
|
"step": 39950 |
|
}, |
|
{ |
|
"epoch": 43.59, |
|
"learning_rate": 1.1330337078651685e-06, |
|
"loss": 0.2091, |
|
"step": 39975 |
|
}, |
|
{ |
|
"epoch": 43.62, |
|
"learning_rate": 1.1274157303370788e-06, |
|
"loss": 0.2477, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 43.62, |
|
"eval_cer": 7.2008097595524525, |
|
"eval_loss": 0.46780213713645935, |
|
"eval_runtime": 3174.0364, |
|
"eval_samples_per_second": 0.682, |
|
"eval_steps_per_second": 0.171, |
|
"eval_wer": 22.827882650464684, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 43.65, |
|
"learning_rate": 1.121797752808989e-06, |
|
"loss": 0.2545, |
|
"step": 40025 |
|
}, |
|
{ |
|
"epoch": 43.68, |
|
"learning_rate": 1.116179775280899e-06, |
|
"loss": 0.2254, |
|
"step": 40050 |
|
}, |
|
{ |
|
"epoch": 43.7, |
|
"learning_rate": 1.110561797752809e-06, |
|
"loss": 0.2051, |
|
"step": 40075 |
|
}, |
|
{ |
|
"epoch": 43.73, |
|
"learning_rate": 1.1049438202247193e-06, |
|
"loss": 0.2168, |
|
"step": 40100 |
|
}, |
|
{ |
|
"epoch": 43.76, |
|
"learning_rate": 1.0993258426966293e-06, |
|
"loss": 0.2304, |
|
"step": 40125 |
|
}, |
|
{ |
|
"epoch": 43.78, |
|
"learning_rate": 1.0937078651685395e-06, |
|
"loss": 0.2346, |
|
"step": 40150 |
|
}, |
|
{ |
|
"epoch": 43.81, |
|
"learning_rate": 1.0880898876404495e-06, |
|
"loss": 0.2231, |
|
"step": 40175 |
|
}, |
|
{ |
|
"epoch": 43.84, |
|
"learning_rate": 1.0824719101123595e-06, |
|
"loss": 0.2233, |
|
"step": 40200 |
|
}, |
|
{ |
|
"epoch": 43.87, |
|
"learning_rate": 1.0768539325842698e-06, |
|
"loss": 0.2394, |
|
"step": 40225 |
|
}, |
|
{ |
|
"epoch": 43.89, |
|
"learning_rate": 1.07123595505618e-06, |
|
"loss": 0.2243, |
|
"step": 40250 |
|
}, |
|
{ |
|
"epoch": 43.92, |
|
"learning_rate": 1.06561797752809e-06, |
|
"loss": 0.2352, |
|
"step": 40275 |
|
}, |
|
{ |
|
"epoch": 43.95, |
|
"learning_rate": 1.06e-06, |
|
"loss": 0.2776, |
|
"step": 40300 |
|
}, |
|
{ |
|
"epoch": 43.97, |
|
"learning_rate": 1.0543820224719102e-06, |
|
"loss": 0.2271, |
|
"step": 40325 |
|
}, |
|
{ |
|
"epoch": 44.0, |
|
"learning_rate": 1.0487640449438203e-06, |
|
"loss": 0.2344, |
|
"step": 40350 |
|
}, |
|
{ |
|
"epoch": 44.03, |
|
"learning_rate": 1.0431460674157303e-06, |
|
"loss": 0.1906, |
|
"step": 40375 |
|
}, |
|
{ |
|
"epoch": 44.06, |
|
"learning_rate": 1.0375280898876405e-06, |
|
"loss": 0.2142, |
|
"step": 40400 |
|
}, |
|
{ |
|
"epoch": 44.08, |
|
"learning_rate": 1.0319101123595507e-06, |
|
"loss": 0.1885, |
|
"step": 40425 |
|
}, |
|
{ |
|
"epoch": 44.11, |
|
"learning_rate": 1.0262921348314607e-06, |
|
"loss": 0.2192, |
|
"step": 40450 |
|
}, |
|
{ |
|
"epoch": 44.14, |
|
"learning_rate": 1.0206741573033708e-06, |
|
"loss": 0.2242, |
|
"step": 40475 |
|
}, |
|
{ |
|
"epoch": 44.17, |
|
"learning_rate": 1.015056179775281e-06, |
|
"loss": 0.2627, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 44.19, |
|
"learning_rate": 1.0094382022471912e-06, |
|
"loss": 0.1955, |
|
"step": 40525 |
|
}, |
|
{ |
|
"epoch": 44.22, |
|
"learning_rate": 1.0038202247191012e-06, |
|
"loss": 0.2384, |
|
"step": 40550 |
|
}, |
|
{ |
|
"epoch": 44.25, |
|
"learning_rate": 9.982022471910113e-07, |
|
"loss": 0.2457, |
|
"step": 40575 |
|
}, |
|
{ |
|
"epoch": 44.27, |
|
"learning_rate": 9.925842696629215e-07, |
|
"loss": 0.1971, |
|
"step": 40600 |
|
}, |
|
{ |
|
"epoch": 44.3, |
|
"learning_rate": 9.869662921348315e-07, |
|
"loss": 0.2013, |
|
"step": 40625 |
|
}, |
|
{ |
|
"epoch": 44.33, |
|
"learning_rate": 9.813483146067417e-07, |
|
"loss": 0.203, |
|
"step": 40650 |
|
}, |
|
{ |
|
"epoch": 44.36, |
|
"learning_rate": 9.757303370786517e-07, |
|
"loss": 0.2259, |
|
"step": 40675 |
|
}, |
|
{ |
|
"epoch": 44.38, |
|
"learning_rate": 9.701123595505618e-07, |
|
"loss": 0.201, |
|
"step": 40700 |
|
}, |
|
{ |
|
"epoch": 44.41, |
|
"learning_rate": 9.64494382022472e-07, |
|
"loss": 0.2531, |
|
"step": 40725 |
|
}, |
|
{ |
|
"epoch": 44.44, |
|
"learning_rate": 9.588764044943822e-07, |
|
"loss": 0.1835, |
|
"step": 40750 |
|
}, |
|
{ |
|
"epoch": 44.47, |
|
"learning_rate": 9.532584269662921e-07, |
|
"loss": 0.2384, |
|
"step": 40775 |
|
}, |
|
{ |
|
"epoch": 44.49, |
|
"learning_rate": 9.476404494382022e-07, |
|
"loss": 0.2023, |
|
"step": 40800 |
|
}, |
|
{ |
|
"epoch": 44.52, |
|
"learning_rate": 9.420224719101125e-07, |
|
"loss": 0.2448, |
|
"step": 40825 |
|
}, |
|
{ |
|
"epoch": 44.55, |
|
"learning_rate": 9.364044943820226e-07, |
|
"loss": 0.2104, |
|
"step": 40850 |
|
}, |
|
{ |
|
"epoch": 44.57, |
|
"learning_rate": 9.307865168539326e-07, |
|
"loss": 0.2211, |
|
"step": 40875 |
|
}, |
|
{ |
|
"epoch": 44.6, |
|
"learning_rate": 9.251685393258427e-07, |
|
"loss": 0.2498, |
|
"step": 40900 |
|
}, |
|
{ |
|
"epoch": 44.63, |
|
"learning_rate": 9.19550561797753e-07, |
|
"loss": 0.2139, |
|
"step": 40925 |
|
}, |
|
{ |
|
"epoch": 44.66, |
|
"learning_rate": 9.13932584269663e-07, |
|
"loss": 0.2349, |
|
"step": 40950 |
|
}, |
|
{ |
|
"epoch": 44.68, |
|
"learning_rate": 9.083146067415731e-07, |
|
"loss": 0.2201, |
|
"step": 40975 |
|
}, |
|
{ |
|
"epoch": 44.71, |
|
"learning_rate": 9.026966292134832e-07, |
|
"loss": 0.1994, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 44.71, |
|
"eval_cer": 7.11794130676955, |
|
"eval_loss": 0.46892717480659485, |
|
"eval_runtime": 3141.0329, |
|
"eval_samples_per_second": 0.69, |
|
"eval_steps_per_second": 0.173, |
|
"eval_wer": 22.380814431456063, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 44.74, |
|
"learning_rate": 8.970786516853932e-07, |
|
"loss": 0.1841, |
|
"step": 41025 |
|
}, |
|
{ |
|
"epoch": 44.77, |
|
"learning_rate": 8.914606741573035e-07, |
|
"loss": 0.204, |
|
"step": 41050 |
|
}, |
|
{ |
|
"epoch": 44.79, |
|
"learning_rate": 8.858426966292136e-07, |
|
"loss": 0.2144, |
|
"step": 41075 |
|
}, |
|
{ |
|
"epoch": 44.82, |
|
"learning_rate": 8.802247191011237e-07, |
|
"loss": 0.2586, |
|
"step": 41100 |
|
}, |
|
{ |
|
"epoch": 44.85, |
|
"learning_rate": 8.746067415730337e-07, |
|
"loss": 0.219, |
|
"step": 41125 |
|
}, |
|
{ |
|
"epoch": 44.87, |
|
"learning_rate": 8.689887640449439e-07, |
|
"loss": 0.2091, |
|
"step": 41150 |
|
}, |
|
{ |
|
"epoch": 44.9, |
|
"learning_rate": 8.633707865168541e-07, |
|
"loss": 0.1841, |
|
"step": 41175 |
|
}, |
|
{ |
|
"epoch": 44.93, |
|
"learning_rate": 8.577528089887641e-07, |
|
"loss": 0.202, |
|
"step": 41200 |
|
}, |
|
{ |
|
"epoch": 44.96, |
|
"learning_rate": 8.521348314606742e-07, |
|
"loss": 0.2057, |
|
"step": 41225 |
|
}, |
|
{ |
|
"epoch": 44.98, |
|
"learning_rate": 8.465168539325843e-07, |
|
"loss": 0.2353, |
|
"step": 41250 |
|
}, |
|
{ |
|
"epoch": 45.01, |
|
"learning_rate": 8.408988764044944e-07, |
|
"loss": 0.181, |
|
"step": 41275 |
|
}, |
|
{ |
|
"epoch": 45.04, |
|
"learning_rate": 8.352808988764046e-07, |
|
"loss": 0.2289, |
|
"step": 41300 |
|
}, |
|
{ |
|
"epoch": 45.07, |
|
"learning_rate": 8.296629213483147e-07, |
|
"loss": 0.2127, |
|
"step": 41325 |
|
}, |
|
{ |
|
"epoch": 45.09, |
|
"learning_rate": 8.240449438202248e-07, |
|
"loss": 0.2119, |
|
"step": 41350 |
|
}, |
|
{ |
|
"epoch": 45.12, |
|
"learning_rate": 8.184269662921348e-07, |
|
"loss": 0.2223, |
|
"step": 41375 |
|
}, |
|
{ |
|
"epoch": 45.15, |
|
"learning_rate": 8.12808988764045e-07, |
|
"loss": 0.2207, |
|
"step": 41400 |
|
}, |
|
{ |
|
"epoch": 45.17, |
|
"learning_rate": 8.071910112359552e-07, |
|
"loss": 0.213, |
|
"step": 41425 |
|
}, |
|
{ |
|
"epoch": 45.2, |
|
"learning_rate": 8.015730337078652e-07, |
|
"loss": 0.2092, |
|
"step": 41450 |
|
}, |
|
{ |
|
"epoch": 45.23, |
|
"learning_rate": 7.959550561797753e-07, |
|
"loss": 0.1926, |
|
"step": 41475 |
|
}, |
|
{ |
|
"epoch": 45.26, |
|
"learning_rate": 7.903370786516855e-07, |
|
"loss": 0.2111, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 45.28, |
|
"learning_rate": 7.847191011235955e-07, |
|
"loss": 0.1974, |
|
"step": 41525 |
|
}, |
|
{ |
|
"epoch": 45.31, |
|
"learning_rate": 7.791011235955057e-07, |
|
"loss": 0.1944, |
|
"step": 41550 |
|
}, |
|
{ |
|
"epoch": 45.34, |
|
"learning_rate": 7.734831460674158e-07, |
|
"loss": 0.159, |
|
"step": 41575 |
|
}, |
|
{ |
|
"epoch": 45.37, |
|
"learning_rate": 7.678651685393259e-07, |
|
"loss": 0.2071, |
|
"step": 41600 |
|
}, |
|
{ |
|
"epoch": 45.39, |
|
"learning_rate": 7.62247191011236e-07, |
|
"loss": 0.2017, |
|
"step": 41625 |
|
}, |
|
{ |
|
"epoch": 45.42, |
|
"learning_rate": 7.566292134831462e-07, |
|
"loss": 0.2139, |
|
"step": 41650 |
|
}, |
|
{ |
|
"epoch": 45.45, |
|
"learning_rate": 7.510112359550563e-07, |
|
"loss": 0.1659, |
|
"step": 41675 |
|
}, |
|
{ |
|
"epoch": 45.47, |
|
"learning_rate": 7.453932584269663e-07, |
|
"loss": 0.2089, |
|
"step": 41700 |
|
}, |
|
{ |
|
"epoch": 45.5, |
|
"learning_rate": 7.397752808988764e-07, |
|
"loss": 0.1634, |
|
"step": 41725 |
|
}, |
|
{ |
|
"epoch": 45.53, |
|
"learning_rate": 7.341573033707866e-07, |
|
"loss": 0.2303, |
|
"step": 41750 |
|
}, |
|
{ |
|
"epoch": 45.56, |
|
"learning_rate": 7.285393258426966e-07, |
|
"loss": 0.2094, |
|
"step": 41775 |
|
}, |
|
{ |
|
"epoch": 45.58, |
|
"learning_rate": 7.229213483146068e-07, |
|
"loss": 0.1817, |
|
"step": 41800 |
|
}, |
|
{ |
|
"epoch": 45.61, |
|
"learning_rate": 7.173033707865169e-07, |
|
"loss": 0.1848, |
|
"step": 41825 |
|
}, |
|
{ |
|
"epoch": 45.64, |
|
"learning_rate": 7.116853932584271e-07, |
|
"loss": 0.1971, |
|
"step": 41850 |
|
}, |
|
{ |
|
"epoch": 45.67, |
|
"learning_rate": 7.060674157303371e-07, |
|
"loss": 0.2081, |
|
"step": 41875 |
|
}, |
|
{ |
|
"epoch": 45.69, |
|
"learning_rate": 7.004494382022473e-07, |
|
"loss": 0.2245, |
|
"step": 41900 |
|
}, |
|
{ |
|
"epoch": 45.72, |
|
"learning_rate": 6.948314606741574e-07, |
|
"loss": 0.1896, |
|
"step": 41925 |
|
}, |
|
{ |
|
"epoch": 45.75, |
|
"learning_rate": 6.892134831460674e-07, |
|
"loss": 0.2193, |
|
"step": 41950 |
|
}, |
|
{ |
|
"epoch": 45.77, |
|
"learning_rate": 6.835955056179776e-07, |
|
"loss": 0.1998, |
|
"step": 41975 |
|
}, |
|
{ |
|
"epoch": 45.8, |
|
"learning_rate": 6.779775280898877e-07, |
|
"loss": 0.1865, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 45.8, |
|
"eval_cer": 7.135145829838664, |
|
"eval_loss": 0.47173011302948, |
|
"eval_runtime": 3187.2, |
|
"eval_samples_per_second": 0.68, |
|
"eval_steps_per_second": 0.17, |
|
"eval_wer": 22.566375145825667, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 45.83, |
|
"learning_rate": 6.723595505617977e-07, |
|
"loss": 0.2185, |
|
"step": 42025 |
|
}, |
|
{ |
|
"epoch": 45.86, |
|
"learning_rate": 6.667415730337079e-07, |
|
"loss": 0.2142, |
|
"step": 42050 |
|
}, |
|
{ |
|
"epoch": 45.88, |
|
"learning_rate": 6.61123595505618e-07, |
|
"loss": 0.203, |
|
"step": 42075 |
|
}, |
|
{ |
|
"epoch": 45.91, |
|
"learning_rate": 6.555056179775282e-07, |
|
"loss": 0.1912, |
|
"step": 42100 |
|
}, |
|
{ |
|
"epoch": 45.94, |
|
"learning_rate": 6.498876404494382e-07, |
|
"loss": 0.1853, |
|
"step": 42125 |
|
}, |
|
{ |
|
"epoch": 45.97, |
|
"learning_rate": 6.442696629213484e-07, |
|
"loss": 0.2056, |
|
"step": 42150 |
|
}, |
|
{ |
|
"epoch": 45.99, |
|
"learning_rate": 6.386516853932585e-07, |
|
"loss": 0.1564, |
|
"step": 42175 |
|
}, |
|
{ |
|
"epoch": 46.02, |
|
"learning_rate": 6.330337078651686e-07, |
|
"loss": 0.1701, |
|
"step": 42200 |
|
}, |
|
{ |
|
"epoch": 46.05, |
|
"learning_rate": 6.274157303370787e-07, |
|
"loss": 0.2156, |
|
"step": 42225 |
|
}, |
|
{ |
|
"epoch": 46.07, |
|
"learning_rate": 6.217977528089888e-07, |
|
"loss": 0.1844, |
|
"step": 42250 |
|
}, |
|
{ |
|
"epoch": 46.1, |
|
"learning_rate": 6.161797752808989e-07, |
|
"loss": 0.1641, |
|
"step": 42275 |
|
}, |
|
{ |
|
"epoch": 46.13, |
|
"learning_rate": 6.10561797752809e-07, |
|
"loss": 0.1754, |
|
"step": 42300 |
|
}, |
|
{ |
|
"epoch": 46.16, |
|
"learning_rate": 6.049438202247192e-07, |
|
"loss": 0.1857, |
|
"step": 42325 |
|
}, |
|
{ |
|
"epoch": 46.18, |
|
"learning_rate": 5.993258426966293e-07, |
|
"loss": 0.159, |
|
"step": 42350 |
|
}, |
|
{ |
|
"epoch": 46.21, |
|
"learning_rate": 5.937078651685393e-07, |
|
"loss": 0.2097, |
|
"step": 42375 |
|
}, |
|
{ |
|
"epoch": 46.24, |
|
"learning_rate": 5.880898876404495e-07, |
|
"loss": 0.1941, |
|
"step": 42400 |
|
}, |
|
{ |
|
"epoch": 46.26, |
|
"learning_rate": 5.824719101123595e-07, |
|
"loss": 0.1817, |
|
"step": 42425 |
|
}, |
|
{ |
|
"epoch": 46.29, |
|
"learning_rate": 5.768539325842698e-07, |
|
"loss": 0.2022, |
|
"step": 42450 |
|
}, |
|
{ |
|
"epoch": 46.32, |
|
"learning_rate": 5.712359550561798e-07, |
|
"loss": 0.1958, |
|
"step": 42475 |
|
}, |
|
{ |
|
"epoch": 46.35, |
|
"learning_rate": 5.656179775280899e-07, |
|
"loss": 0.176, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 46.37, |
|
"learning_rate": 5.6e-07, |
|
"loss": 0.2181, |
|
"step": 42525 |
|
}, |
|
{ |
|
"epoch": 46.4, |
|
"learning_rate": 5.543820224719102e-07, |
|
"loss": 0.2047, |
|
"step": 42550 |
|
}, |
|
{ |
|
"epoch": 46.43, |
|
"learning_rate": 5.487640449438203e-07, |
|
"loss": 0.1942, |
|
"step": 42575 |
|
}, |
|
{ |
|
"epoch": 46.46, |
|
"learning_rate": 5.431460674157304e-07, |
|
"loss": 0.1655, |
|
"step": 42600 |
|
}, |
|
{ |
|
"epoch": 46.48, |
|
"learning_rate": 5.375280898876404e-07, |
|
"loss": 0.1958, |
|
"step": 42625 |
|
}, |
|
{ |
|
"epoch": 46.51, |
|
"learning_rate": 5.319101123595506e-07, |
|
"loss": 0.2192, |
|
"step": 42650 |
|
}, |
|
{ |
|
"epoch": 46.54, |
|
"learning_rate": 5.262921348314607e-07, |
|
"loss": 0.1425, |
|
"step": 42675 |
|
}, |
|
{ |
|
"epoch": 46.56, |
|
"learning_rate": 5.206741573033709e-07, |
|
"loss": 0.1848, |
|
"step": 42700 |
|
}, |
|
{ |
|
"epoch": 46.59, |
|
"learning_rate": 5.150561797752809e-07, |
|
"loss": 0.1621, |
|
"step": 42725 |
|
}, |
|
{ |
|
"epoch": 46.62, |
|
"learning_rate": 5.09438202247191e-07, |
|
"loss": 0.177, |
|
"step": 42750 |
|
}, |
|
{ |
|
"epoch": 46.65, |
|
"learning_rate": 5.038202247191012e-07, |
|
"loss": 0.1937, |
|
"step": 42775 |
|
}, |
|
{ |
|
"epoch": 46.67, |
|
"learning_rate": 4.982022471910113e-07, |
|
"loss": 0.175, |
|
"step": 42800 |
|
}, |
|
{ |
|
"epoch": 46.7, |
|
"learning_rate": 4.925842696629214e-07, |
|
"loss": 0.1738, |
|
"step": 42825 |
|
}, |
|
{ |
|
"epoch": 46.73, |
|
"learning_rate": 4.869662921348315e-07, |
|
"loss": 0.1732, |
|
"step": 42850 |
|
}, |
|
{ |
|
"epoch": 46.76, |
|
"learning_rate": 4.813483146067415e-07, |
|
"loss": 0.1871, |
|
"step": 42875 |
|
}, |
|
{ |
|
"epoch": 46.78, |
|
"learning_rate": 4.7573033707865176e-07, |
|
"loss": 0.1949, |
|
"step": 42900 |
|
}, |
|
{ |
|
"epoch": 46.81, |
|
"learning_rate": 4.7011235955056183e-07, |
|
"loss": 0.1999, |
|
"step": 42925 |
|
}, |
|
{ |
|
"epoch": 46.84, |
|
"learning_rate": 4.6449438202247196e-07, |
|
"loss": 0.1805, |
|
"step": 42950 |
|
}, |
|
{ |
|
"epoch": 46.86, |
|
"learning_rate": 4.588764044943821e-07, |
|
"loss": 0.1668, |
|
"step": 42975 |
|
}, |
|
{ |
|
"epoch": 46.89, |
|
"learning_rate": 4.5325842696629215e-07, |
|
"loss": 0.2307, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 46.89, |
|
"eval_cer": 7.136436169068848, |
|
"eval_loss": 0.47538912296295166, |
|
"eval_runtime": 3233.0256, |
|
"eval_samples_per_second": 0.67, |
|
"eval_steps_per_second": 0.168, |
|
"eval_wer": 22.372201908848194, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 46.92, |
|
"learning_rate": 4.476404494382023e-07, |
|
"loss": 0.1901, |
|
"step": 43025 |
|
}, |
|
{ |
|
"epoch": 46.95, |
|
"learning_rate": 4.420224719101124e-07, |
|
"loss": 0.197, |
|
"step": 43050 |
|
}, |
|
{ |
|
"epoch": 46.97, |
|
"learning_rate": 4.364044943820225e-07, |
|
"loss": 0.2022, |
|
"step": 43075 |
|
}, |
|
{ |
|
"epoch": 47.0, |
|
"learning_rate": 4.3078651685393264e-07, |
|
"loss": 0.1763, |
|
"step": 43100 |
|
}, |
|
{ |
|
"epoch": 47.03, |
|
"learning_rate": 4.251685393258427e-07, |
|
"loss": 0.2063, |
|
"step": 43125 |
|
}, |
|
{ |
|
"epoch": 47.06, |
|
"learning_rate": 4.195505617977529e-07, |
|
"loss": 0.1824, |
|
"step": 43150 |
|
}, |
|
{ |
|
"epoch": 47.08, |
|
"learning_rate": 4.1393258426966295e-07, |
|
"loss": 0.2059, |
|
"step": 43175 |
|
}, |
|
{ |
|
"epoch": 47.11, |
|
"learning_rate": 4.0831460674157307e-07, |
|
"loss": 0.194, |
|
"step": 43200 |
|
}, |
|
{ |
|
"epoch": 47.14, |
|
"learning_rate": 4.026966292134832e-07, |
|
"loss": 0.155, |
|
"step": 43225 |
|
}, |
|
{ |
|
"epoch": 47.16, |
|
"learning_rate": 3.9707865168539326e-07, |
|
"loss": 0.1933, |
|
"step": 43250 |
|
}, |
|
{ |
|
"epoch": 47.19, |
|
"learning_rate": 3.9146067415730344e-07, |
|
"loss": 0.1939, |
|
"step": 43275 |
|
}, |
|
{ |
|
"epoch": 47.22, |
|
"learning_rate": 3.858426966292135e-07, |
|
"loss": 0.1826, |
|
"step": 43300 |
|
}, |
|
{ |
|
"epoch": 47.25, |
|
"learning_rate": 3.802247191011237e-07, |
|
"loss": 0.191, |
|
"step": 43325 |
|
}, |
|
{ |
|
"epoch": 47.27, |
|
"learning_rate": 3.7460674157303375e-07, |
|
"loss": 0.1771, |
|
"step": 43350 |
|
}, |
|
{ |
|
"epoch": 47.3, |
|
"learning_rate": 3.689887640449438e-07, |
|
"loss": 0.1942, |
|
"step": 43375 |
|
}, |
|
{ |
|
"epoch": 47.33, |
|
"learning_rate": 3.63370786516854e-07, |
|
"loss": 0.1873, |
|
"step": 43400 |
|
}, |
|
{ |
|
"epoch": 47.36, |
|
"learning_rate": 3.5775280898876406e-07, |
|
"loss": 0.1559, |
|
"step": 43425 |
|
}, |
|
{ |
|
"epoch": 47.38, |
|
"learning_rate": 3.5213483146067424e-07, |
|
"loss": 0.1719, |
|
"step": 43450 |
|
}, |
|
{ |
|
"epoch": 47.41, |
|
"learning_rate": 3.465168539325843e-07, |
|
"loss": 0.1971, |
|
"step": 43475 |
|
}, |
|
{ |
|
"epoch": 47.44, |
|
"learning_rate": 3.408988764044944e-07, |
|
"loss": 0.1812, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 47.46, |
|
"learning_rate": 3.3528089887640455e-07, |
|
"loss": 0.1667, |
|
"step": 43525 |
|
}, |
|
{ |
|
"epoch": 47.49, |
|
"learning_rate": 3.296629213483146e-07, |
|
"loss": 0.1625, |
|
"step": 43550 |
|
}, |
|
{ |
|
"epoch": 47.52, |
|
"learning_rate": 3.240449438202248e-07, |
|
"loss": 0.1383, |
|
"step": 43575 |
|
}, |
|
{ |
|
"epoch": 47.55, |
|
"learning_rate": 3.1842696629213487e-07, |
|
"loss": 0.1887, |
|
"step": 43600 |
|
}, |
|
{ |
|
"epoch": 47.57, |
|
"learning_rate": 3.1280898876404494e-07, |
|
"loss": 0.2228, |
|
"step": 43625 |
|
}, |
|
{ |
|
"epoch": 47.6, |
|
"learning_rate": 3.0719101123595506e-07, |
|
"loss": 0.18, |
|
"step": 43650 |
|
}, |
|
{ |
|
"epoch": 47.63, |
|
"learning_rate": 3.015730337078652e-07, |
|
"loss": 0.1823, |
|
"step": 43675 |
|
}, |
|
{ |
|
"epoch": 47.66, |
|
"learning_rate": 2.959550561797753e-07, |
|
"loss": 0.1498, |
|
"step": 43700 |
|
}, |
|
{ |
|
"epoch": 47.68, |
|
"learning_rate": 2.903370786516854e-07, |
|
"loss": 0.1725, |
|
"step": 43725 |
|
}, |
|
{ |
|
"epoch": 47.71, |
|
"learning_rate": 2.8471910112359555e-07, |
|
"loss": 0.1457, |
|
"step": 43750 |
|
}, |
|
{ |
|
"epoch": 47.74, |
|
"learning_rate": 2.791011235955056e-07, |
|
"loss": 0.1575, |
|
"step": 43775 |
|
}, |
|
{ |
|
"epoch": 47.76, |
|
"learning_rate": 2.7348314606741574e-07, |
|
"loss": 0.1722, |
|
"step": 43800 |
|
}, |
|
{ |
|
"epoch": 47.79, |
|
"learning_rate": 2.6786516853932586e-07, |
|
"loss": 0.1841, |
|
"step": 43825 |
|
}, |
|
{ |
|
"epoch": 47.82, |
|
"learning_rate": 2.62247191011236e-07, |
|
"loss": 0.1811, |
|
"step": 43850 |
|
}, |
|
{ |
|
"epoch": 47.85, |
|
"learning_rate": 2.566292134831461e-07, |
|
"loss": 0.1808, |
|
"step": 43875 |
|
}, |
|
{ |
|
"epoch": 47.87, |
|
"learning_rate": 2.5101123595505617e-07, |
|
"loss": 0.1564, |
|
"step": 43900 |
|
}, |
|
{ |
|
"epoch": 47.9, |
|
"learning_rate": 2.453932584269663e-07, |
|
"loss": 0.1803, |
|
"step": 43925 |
|
}, |
|
{ |
|
"epoch": 47.93, |
|
"learning_rate": 2.397752808988764e-07, |
|
"loss": 0.2052, |
|
"step": 43950 |
|
}, |
|
{ |
|
"epoch": 47.96, |
|
"learning_rate": 2.3415730337078654e-07, |
|
"loss": 0.2026, |
|
"step": 43975 |
|
}, |
|
{ |
|
"epoch": 47.98, |
|
"learning_rate": 2.2853932584269666e-07, |
|
"loss": 0.1705, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 47.98, |
|
"eval_cer": 7.082958776529017, |
|
"eval_loss": 0.47585317492485046, |
|
"eval_runtime": 3126.347, |
|
"eval_samples_per_second": 0.693, |
|
"eval_steps_per_second": 0.173, |
|
"eval_wer": 22.386295127661075, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 48.01, |
|
"learning_rate": 2.2292134831460676e-07, |
|
"loss": 0.1972, |
|
"step": 44025 |
|
}, |
|
{ |
|
"epoch": 48.04, |
|
"learning_rate": 2.1752808988764047e-07, |
|
"loss": 0.156, |
|
"step": 44050 |
|
}, |
|
{ |
|
"epoch": 48.06, |
|
"learning_rate": 2.119101123595506e-07, |
|
"loss": 0.1913, |
|
"step": 44075 |
|
}, |
|
{ |
|
"epoch": 48.09, |
|
"learning_rate": 2.062921348314607e-07, |
|
"loss": 0.1576, |
|
"step": 44100 |
|
}, |
|
{ |
|
"epoch": 48.12, |
|
"learning_rate": 2.0067415730337078e-07, |
|
"loss": 0.2132, |
|
"step": 44125 |
|
}, |
|
{ |
|
"epoch": 48.15, |
|
"learning_rate": 1.950561797752809e-07, |
|
"loss": 0.1657, |
|
"step": 44150 |
|
}, |
|
{ |
|
"epoch": 48.17, |
|
"learning_rate": 1.8943820224719102e-07, |
|
"loss": 0.1648, |
|
"step": 44175 |
|
}, |
|
{ |
|
"epoch": 48.2, |
|
"learning_rate": 1.8382022471910115e-07, |
|
"loss": 0.1761, |
|
"step": 44200 |
|
}, |
|
{ |
|
"epoch": 48.23, |
|
"learning_rate": 1.7820224719101127e-07, |
|
"loss": 0.1714, |
|
"step": 44225 |
|
}, |
|
{ |
|
"epoch": 48.26, |
|
"learning_rate": 1.7258426966292134e-07, |
|
"loss": 0.1919, |
|
"step": 44250 |
|
}, |
|
{ |
|
"epoch": 48.28, |
|
"learning_rate": 1.6696629213483146e-07, |
|
"loss": 0.1635, |
|
"step": 44275 |
|
}, |
|
{ |
|
"epoch": 48.31, |
|
"learning_rate": 1.6134831460674158e-07, |
|
"loss": 0.1621, |
|
"step": 44300 |
|
}, |
|
{ |
|
"epoch": 48.34, |
|
"learning_rate": 1.557303370786517e-07, |
|
"loss": 0.1812, |
|
"step": 44325 |
|
}, |
|
{ |
|
"epoch": 48.36, |
|
"learning_rate": 1.501123595505618e-07, |
|
"loss": 0.1473, |
|
"step": 44350 |
|
}, |
|
{ |
|
"epoch": 48.39, |
|
"learning_rate": 1.4449438202247192e-07, |
|
"loss": 0.1599, |
|
"step": 44375 |
|
}, |
|
{ |
|
"epoch": 48.42, |
|
"learning_rate": 1.3887640449438202e-07, |
|
"loss": 0.1755, |
|
"step": 44400 |
|
}, |
|
{ |
|
"epoch": 48.45, |
|
"learning_rate": 1.3325842696629214e-07, |
|
"loss": 0.1338, |
|
"step": 44425 |
|
}, |
|
{ |
|
"epoch": 48.47, |
|
"learning_rate": 1.2764044943820226e-07, |
|
"loss": 0.1643, |
|
"step": 44450 |
|
}, |
|
{ |
|
"epoch": 48.5, |
|
"learning_rate": 1.2202247191011236e-07, |
|
"loss": 0.1683, |
|
"step": 44475 |
|
}, |
|
{ |
|
"epoch": 48.53, |
|
"learning_rate": 1.1640449438202248e-07, |
|
"loss": 0.2074, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 48.56, |
|
"learning_rate": 1.1078651685393259e-07, |
|
"loss": 0.1888, |
|
"step": 44525 |
|
}, |
|
{ |
|
"epoch": 48.58, |
|
"learning_rate": 1.0516853932584271e-07, |
|
"loss": 0.1798, |
|
"step": 44550 |
|
}, |
|
{ |
|
"epoch": 48.61, |
|
"learning_rate": 9.955056179775281e-08, |
|
"loss": 0.1636, |
|
"step": 44575 |
|
}, |
|
{ |
|
"epoch": 48.64, |
|
"learning_rate": 9.393258426966293e-08, |
|
"loss": 0.1752, |
|
"step": 44600 |
|
}, |
|
{ |
|
"epoch": 48.66, |
|
"learning_rate": 8.831460674157305e-08, |
|
"loss": 0.1478, |
|
"step": 44625 |
|
}, |
|
{ |
|
"epoch": 48.69, |
|
"learning_rate": 8.269662921348315e-08, |
|
"loss": 0.1745, |
|
"step": 44650 |
|
}, |
|
{ |
|
"epoch": 48.72, |
|
"learning_rate": 7.707865168539327e-08, |
|
"loss": 0.2115, |
|
"step": 44675 |
|
}, |
|
{ |
|
"epoch": 48.75, |
|
"learning_rate": 7.146067415730338e-08, |
|
"loss": 0.1916, |
|
"step": 44700 |
|
}, |
|
{ |
|
"epoch": 48.77, |
|
"learning_rate": 6.584269662921349e-08, |
|
"loss": 0.1798, |
|
"step": 44725 |
|
}, |
|
{ |
|
"epoch": 48.8, |
|
"learning_rate": 6.02247191011236e-08, |
|
"loss": 0.1878, |
|
"step": 44750 |
|
}, |
|
{ |
|
"epoch": 48.83, |
|
"learning_rate": 5.460674157303371e-08, |
|
"loss": 0.1734, |
|
"step": 44775 |
|
}, |
|
{ |
|
"epoch": 48.85, |
|
"learning_rate": 4.8988764044943827e-08, |
|
"loss": 0.1588, |
|
"step": 44800 |
|
}, |
|
{ |
|
"epoch": 48.88, |
|
"learning_rate": 4.3370786516853935e-08, |
|
"loss": 0.1622, |
|
"step": 44825 |
|
}, |
|
{ |
|
"epoch": 48.91, |
|
"learning_rate": 3.775280898876405e-08, |
|
"loss": 0.1726, |
|
"step": 44850 |
|
}, |
|
{ |
|
"epoch": 48.94, |
|
"learning_rate": 3.213483146067416e-08, |
|
"loss": 0.1523, |
|
"step": 44875 |
|
}, |
|
{ |
|
"epoch": 48.96, |
|
"learning_rate": 2.6516853932584272e-08, |
|
"loss": 0.1753, |
|
"step": 44900 |
|
}, |
|
{ |
|
"epoch": 48.99, |
|
"learning_rate": 2.0898876404494384e-08, |
|
"loss": 0.1672, |
|
"step": 44925 |
|
}, |
|
{ |
|
"epoch": 49.02, |
|
"learning_rate": 1.5280898876404497e-08, |
|
"loss": 0.1675, |
|
"step": 44950 |
|
}, |
|
{ |
|
"epoch": 49.05, |
|
"learning_rate": 9.662921348314607e-09, |
|
"loss": 0.1766, |
|
"step": 44975 |
|
}, |
|
{ |
|
"epoch": 49.07, |
|
"learning_rate": 4.044943820224719e-09, |
|
"loss": 0.2007, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 49.07, |
|
"eval_cer": 7.1186581618974305, |
|
"eval_loss": 0.47667962312698364, |
|
"eval_runtime": 3154.6082, |
|
"eval_samples_per_second": 0.687, |
|
"eval_steps_per_second": 0.172, |
|
"eval_wer": 22.484947659351242, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 49.07, |
|
"step": 45000, |
|
"total_flos": 5.83519441453056e+18, |
|
"train_loss": 0.05880288976563348, |
|
"train_runtime": 37826.6777, |
|
"train_samples_per_second": 4.759, |
|
"train_steps_per_second": 1.19 |
|
}, |
|
{ |
|
"epoch": 49.07, |
|
"eval_cer": 9.956974355224656, |
|
"eval_loss": 0.7910374999046326, |
|
"eval_runtime": 3295.3661, |
|
"eval_samples_per_second": 0.657, |
|
"eval_steps_per_second": 0.164, |
|
"eval_wer": 31.400474471700036, |
|
"step": 45000 |
|
} |
|
], |
|
"max_steps": 45000, |
|
"num_train_epochs": 50, |
|
"total_flos": 5.83519441453056e+18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|