|
{ |
|
"best_metric": 0.1731826514005661, |
|
"best_model_checkpoint": "ai-light-dance_singing5_ft_wav2vec2-large-xlsr-53-5gram-v4-2-1/checkpoint-4500", |
|
"epoch": 49.995024875621894, |
|
"global_step": 5000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.4263, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.4214, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.16e-05, |
|
"loss": 0.4338, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.5600000000000003e-05, |
|
"loss": 0.4153, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.9600000000000002e-05, |
|
"loss": 0.4313, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 2.36e-05, |
|
"loss": 0.4245, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.7200000000000004e-05, |
|
"loss": 0.4128, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.1200000000000006e-05, |
|
"loss": 0.4334, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 3.52e-05, |
|
"loss": 0.4148, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 3.9200000000000004e-05, |
|
"loss": 0.4351, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.19478967785835266, |
|
"eval_runtime": 174.2645, |
|
"eval_samples_per_second": 4.206, |
|
"eval_steps_per_second": 0.528, |
|
"eval_wer": 0.09026203644936077, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 3.9942857142857144e-05, |
|
"loss": 0.4425, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 3.986122448979592e-05, |
|
"loss": 0.4134, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 3.97795918367347e-05, |
|
"loss": 0.4416, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 3.969795918367348e-05, |
|
"loss": 0.4184, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 3.961632653061225e-05, |
|
"loss": 0.4459, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 3.953469387755102e-05, |
|
"loss": 0.4293, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 3.9453061224489796e-05, |
|
"loss": 0.4196, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.9371428571428576e-05, |
|
"loss": 0.4546, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 3.928979591836735e-05, |
|
"loss": 0.4158, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 3.920816326530612e-05, |
|
"loss": 0.4381, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.19607539474964142, |
|
"eval_runtime": 172.6746, |
|
"eval_samples_per_second": 4.245, |
|
"eval_steps_per_second": 0.533, |
|
"eval_wer": 0.09302747302565963, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 3.91265306122449e-05, |
|
"loss": 0.4603, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 3.904489795918368e-05, |
|
"loss": 0.4327, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 3.8963265306122455e-05, |
|
"loss": 0.4309, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 3.888163265306123e-05, |
|
"loss": 0.4115, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 3.88e-05, |
|
"loss": 0.4582, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 3.8718367346938774e-05, |
|
"loss": 0.4328, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 3.8636734693877554e-05, |
|
"loss": 0.4252, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 3.855510204081633e-05, |
|
"loss": 0.4233, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 3.847346938775511e-05, |
|
"loss": 0.4039, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 3.839183673469388e-05, |
|
"loss": 0.441, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.1947716921567917, |
|
"eval_runtime": 172.0103, |
|
"eval_samples_per_second": 4.261, |
|
"eval_steps_per_second": 0.535, |
|
"eval_wer": 0.09565690452443558, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 3.831020408163266e-05, |
|
"loss": 0.433, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 3.822857142857143e-05, |
|
"loss": 0.4229, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 3.8146938775510206e-05, |
|
"loss": 0.4524, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 3.806530612244898e-05, |
|
"loss": 0.4359, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 3.798367346938776e-05, |
|
"loss": 0.4407, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 3.790204081632653e-05, |
|
"loss": 0.4552, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 3.782040816326531e-05, |
|
"loss": 0.4351, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 3.7738775510204086e-05, |
|
"loss": 0.4294, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 3.765714285714286e-05, |
|
"loss": 0.4109, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 3.757551020408164e-05, |
|
"loss": 0.453, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 0.19707180559635162, |
|
"eval_runtime": 173.4219, |
|
"eval_samples_per_second": 4.227, |
|
"eval_steps_per_second": 0.53, |
|
"eval_wer": 0.0904887115785656, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 3.749387755102041e-05, |
|
"loss": 0.4397, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"learning_rate": 3.7412244897959185e-05, |
|
"loss": 0.4265, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 3.7330612244897965e-05, |
|
"loss": 0.4401, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 3.724897959183674e-05, |
|
"loss": 0.4003, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"learning_rate": 3.716734693877552e-05, |
|
"loss": 0.4415, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 3.708571428571429e-05, |
|
"loss": 0.4275, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"learning_rate": 3.7004081632653064e-05, |
|
"loss": 0.4113, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"learning_rate": 3.692244897959184e-05, |
|
"loss": 0.4374, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"learning_rate": 3.684081632653062e-05, |
|
"loss": 0.4135, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 3.675918367346939e-05, |
|
"loss": 0.4324, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 0.18227937817573547, |
|
"eval_runtime": 172.0791, |
|
"eval_samples_per_second": 4.26, |
|
"eval_steps_per_second": 0.535, |
|
"eval_wer": 0.08785928007978965, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 5.1, |
|
"learning_rate": 3.667755102040816e-05, |
|
"loss": 0.4497, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 5.2, |
|
"learning_rate": 3.659591836734694e-05, |
|
"loss": 0.4102, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 5.3, |
|
"learning_rate": 3.6514285714285716e-05, |
|
"loss": 0.4298, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 5.4, |
|
"learning_rate": 3.6432653061224496e-05, |
|
"loss": 0.3982, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"learning_rate": 3.635102040816327e-05, |
|
"loss": 0.4349, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 5.6, |
|
"learning_rate": 3.626938775510204e-05, |
|
"loss": 0.4187, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 5.7, |
|
"learning_rate": 3.6187755102040815e-05, |
|
"loss": 0.4205, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"learning_rate": 3.6106122448979595e-05, |
|
"loss": 0.4462, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 5.9, |
|
"learning_rate": 3.602448979591837e-05, |
|
"loss": 0.4108, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 3.594285714285715e-05, |
|
"loss": 0.4561, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 0.193389430642128, |
|
"eval_runtime": 172.3669, |
|
"eval_samples_per_second": 4.253, |
|
"eval_steps_per_second": 0.534, |
|
"eval_wer": 0.08931000090670052, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 6.1, |
|
"learning_rate": 3.586122448979592e-05, |
|
"loss": 0.4437, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 6.2, |
|
"learning_rate": 3.5779591836734694e-05, |
|
"loss": 0.41, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 6.3, |
|
"learning_rate": 3.5697959183673474e-05, |
|
"loss": 0.4372, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 6.4, |
|
"learning_rate": 3.561632653061225e-05, |
|
"loss": 0.4153, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"learning_rate": 3.553469387755102e-05, |
|
"loss": 0.4252, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 3.54530612244898e-05, |
|
"loss": 0.4175, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 6.7, |
|
"learning_rate": 3.5371428571428574e-05, |
|
"loss": 0.42, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 6.8, |
|
"learning_rate": 3.5289795918367353e-05, |
|
"loss": 0.4283, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 6.9, |
|
"learning_rate": 3.5208163265306127e-05, |
|
"loss": 0.4215, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 3.51265306122449e-05, |
|
"loss": 0.4231, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 0.2088068276643753, |
|
"eval_runtime": 171.8962, |
|
"eval_samples_per_second": 4.264, |
|
"eval_steps_per_second": 0.535, |
|
"eval_wer": 0.09774231571311996, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 7.1, |
|
"learning_rate": 3.504489795918367e-05, |
|
"loss": 0.4597, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 7.2, |
|
"learning_rate": 3.496326530612245e-05, |
|
"loss": 0.4119, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 7.3, |
|
"learning_rate": 3.4881632653061226e-05, |
|
"loss": 0.4308, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 7.4, |
|
"learning_rate": 3.4800000000000006e-05, |
|
"loss": 0.428, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 3.471836734693878e-05, |
|
"loss": 0.4125, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 7.6, |
|
"learning_rate": 3.463673469387755e-05, |
|
"loss": 0.4147, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 7.7, |
|
"learning_rate": 3.455510204081633e-05, |
|
"loss": 0.4269, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 7.8, |
|
"learning_rate": 3.4473469387755105e-05, |
|
"loss": 0.4295, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 7.9, |
|
"learning_rate": 3.439183673469388e-05, |
|
"loss": 0.4085, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 3.431020408163265e-05, |
|
"loss": 0.4339, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 0.19241395592689514, |
|
"eval_runtime": 171.9467, |
|
"eval_samples_per_second": 4.263, |
|
"eval_steps_per_second": 0.535, |
|
"eval_wer": 0.08559252878774141, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 8.1, |
|
"learning_rate": 3.422857142857143e-05, |
|
"loss": 0.4337, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 8.2, |
|
"learning_rate": 3.414693877551021e-05, |
|
"loss": 0.408, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 8.3, |
|
"learning_rate": 3.4065306122448984e-05, |
|
"loss": 0.4102, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 8.4, |
|
"learning_rate": 3.398367346938776e-05, |
|
"loss": 0.4129, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"learning_rate": 3.390204081632653e-05, |
|
"loss": 0.4156, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 8.6, |
|
"learning_rate": 3.382040816326531e-05, |
|
"loss": 0.4061, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 8.7, |
|
"learning_rate": 3.373877551020408e-05, |
|
"loss": 0.401, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 8.8, |
|
"learning_rate": 3.3657142857142856e-05, |
|
"loss": 0.4274, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 8.9, |
|
"learning_rate": 3.3575510204081636e-05, |
|
"loss": 0.4187, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 3.3493877551020416e-05, |
|
"loss": 0.4195, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 0.18351006507873535, |
|
"eval_runtime": 171.8748, |
|
"eval_samples_per_second": 4.265, |
|
"eval_steps_per_second": 0.535, |
|
"eval_wer": 0.08464049324508115, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 9.1, |
|
"learning_rate": 3.341224489795919e-05, |
|
"loss": 0.4452, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 9.2, |
|
"learning_rate": 3.333061224489796e-05, |
|
"loss": 0.4055, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 9.3, |
|
"learning_rate": 3.3248979591836735e-05, |
|
"loss": 0.4114, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 9.4, |
|
"learning_rate": 3.316734693877551e-05, |
|
"loss": 0.3962, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"learning_rate": 3.308571428571429e-05, |
|
"loss": 0.4307, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 9.6, |
|
"learning_rate": 3.300408163265306e-05, |
|
"loss": 0.417, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 9.7, |
|
"learning_rate": 3.292244897959184e-05, |
|
"loss": 0.402, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 9.8, |
|
"learning_rate": 3.2840816326530615e-05, |
|
"loss": 0.4107, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 9.9, |
|
"learning_rate": 3.2759183673469394e-05, |
|
"loss": 0.3873, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 3.267755102040817e-05, |
|
"loss": 0.4162, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 0.18690763413906097, |
|
"eval_runtime": 172.8304, |
|
"eval_samples_per_second": 4.241, |
|
"eval_steps_per_second": 0.532, |
|
"eval_wer": 0.09080605675945236, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 10.1, |
|
"learning_rate": 3.259591836734694e-05, |
|
"loss": 0.4141, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 10.2, |
|
"learning_rate": 3.2514285714285714e-05, |
|
"loss": 0.4026, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 10.3, |
|
"learning_rate": 3.2432653061224494e-05, |
|
"loss": 0.4267, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 10.4, |
|
"learning_rate": 3.235102040816327e-05, |
|
"loss": 0.4001, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 10.5, |
|
"learning_rate": 3.2269387755102047e-05, |
|
"loss": 0.4012, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 10.6, |
|
"learning_rate": 3.218775510204082e-05, |
|
"loss": 0.4095, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 10.7, |
|
"learning_rate": 3.210612244897959e-05, |
|
"loss": 0.4019, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 10.8, |
|
"learning_rate": 3.202448979591837e-05, |
|
"loss": 0.3964, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 10.9, |
|
"learning_rate": 3.1942857142857146e-05, |
|
"loss": 0.3982, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 3.186122448979592e-05, |
|
"loss": 0.411, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_loss": 0.19660010933876038, |
|
"eval_runtime": 171.7745, |
|
"eval_samples_per_second": 4.267, |
|
"eval_steps_per_second": 0.536, |
|
"eval_wer": 0.09502221416266207, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 11.1, |
|
"learning_rate": 3.17795918367347e-05, |
|
"loss": 0.427, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 11.2, |
|
"learning_rate": 3.169795918367347e-05, |
|
"loss": 0.4162, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 11.3, |
|
"learning_rate": 3.161632653061225e-05, |
|
"loss": 0.4195, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 11.4, |
|
"learning_rate": 3.1534693877551025e-05, |
|
"loss": 0.4145, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 11.5, |
|
"learning_rate": 3.14530612244898e-05, |
|
"loss": 0.4077, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 11.6, |
|
"learning_rate": 3.137142857142857e-05, |
|
"loss": 0.3989, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 11.7, |
|
"learning_rate": 3.128979591836735e-05, |
|
"loss": 0.3979, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 11.8, |
|
"learning_rate": 3.1208163265306124e-05, |
|
"loss": 0.4033, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 11.9, |
|
"learning_rate": 3.1126530612244904e-05, |
|
"loss": 0.3981, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"learning_rate": 3.104489795918368e-05, |
|
"loss": 0.4034, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_loss": 0.1890147626399994, |
|
"eval_runtime": 172.2662, |
|
"eval_samples_per_second": 4.255, |
|
"eval_steps_per_second": 0.534, |
|
"eval_wer": 0.0879046151056306, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 12.1, |
|
"learning_rate": 3.096326530612245e-05, |
|
"loss": 0.4258, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 12.2, |
|
"learning_rate": 3.088163265306123e-05, |
|
"loss": 0.4016, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 12.3, |
|
"learning_rate": 3.080816326530613e-05, |
|
"loss": 0.398, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 12.4, |
|
"learning_rate": 3.07265306122449e-05, |
|
"loss": 0.3907, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 12.5, |
|
"learning_rate": 3.0644897959183674e-05, |
|
"loss": 0.3942, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 12.6, |
|
"learning_rate": 3.056326530612245e-05, |
|
"loss": 0.3862, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 12.7, |
|
"learning_rate": 3.0481632653061227e-05, |
|
"loss": 0.3868, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 12.8, |
|
"learning_rate": 3.0400000000000004e-05, |
|
"loss": 0.4104, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 12.9, |
|
"learning_rate": 3.0318367346938777e-05, |
|
"loss": 0.382, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"learning_rate": 3.0236734693877553e-05, |
|
"loss": 0.4155, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_loss": 0.18443575501441956, |
|
"eval_runtime": 172.6893, |
|
"eval_samples_per_second": 4.245, |
|
"eval_steps_per_second": 0.533, |
|
"eval_wer": 0.09148608214706683, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 13.1, |
|
"learning_rate": 3.015510204081633e-05, |
|
"loss": 0.4157, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 13.2, |
|
"learning_rate": 3.0073469387755106e-05, |
|
"loss": 0.3981, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 13.3, |
|
"learning_rate": 2.999183673469388e-05, |
|
"loss": 0.4027, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 13.4, |
|
"learning_rate": 2.9910204081632656e-05, |
|
"loss": 0.3876, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 13.5, |
|
"learning_rate": 2.982857142857143e-05, |
|
"loss": 0.4097, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 13.6, |
|
"learning_rate": 2.974693877551021e-05, |
|
"loss": 0.3976, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 13.7, |
|
"learning_rate": 2.9665306122448982e-05, |
|
"loss": 0.3894, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 13.8, |
|
"learning_rate": 2.958367346938776e-05, |
|
"loss": 0.3877, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 13.9, |
|
"learning_rate": 2.950204081632653e-05, |
|
"loss": 0.3986, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"learning_rate": 2.942040816326531e-05, |
|
"loss": 0.4123, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_loss": 0.18490195274353027, |
|
"eval_runtime": 174.0302, |
|
"eval_samples_per_second": 4.212, |
|
"eval_steps_per_second": 0.529, |
|
"eval_wer": 0.0890833257774957, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 14.1, |
|
"learning_rate": 2.9338775510204085e-05, |
|
"loss": 0.4107, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 14.2, |
|
"learning_rate": 2.925714285714286e-05, |
|
"loss": 0.3911, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 14.3, |
|
"learning_rate": 2.9175510204081634e-05, |
|
"loss": 0.4185, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 14.4, |
|
"learning_rate": 2.9093877551020407e-05, |
|
"loss": 0.4008, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 14.5, |
|
"learning_rate": 2.9012244897959187e-05, |
|
"loss": 0.4142, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 14.6, |
|
"learning_rate": 2.893061224489796e-05, |
|
"loss": 0.3922, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 14.7, |
|
"learning_rate": 2.8848979591836737e-05, |
|
"loss": 0.3963, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 14.8, |
|
"learning_rate": 2.876734693877551e-05, |
|
"loss": 0.4187, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 14.9, |
|
"learning_rate": 2.868571428571429e-05, |
|
"loss": 0.3959, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 2.8604081632653063e-05, |
|
"loss": 0.4002, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_loss": 0.19010967016220093, |
|
"eval_runtime": 175.0636, |
|
"eval_samples_per_second": 4.187, |
|
"eval_steps_per_second": 0.526, |
|
"eval_wer": 0.09017136639767885, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 15.1, |
|
"learning_rate": 2.852244897959184e-05, |
|
"loss": 0.4113, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 15.2, |
|
"learning_rate": 2.8440816326530613e-05, |
|
"loss": 0.3919, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 15.3, |
|
"learning_rate": 2.835918367346939e-05, |
|
"loss": 0.4063, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 15.4, |
|
"learning_rate": 2.8277551020408166e-05, |
|
"loss": 0.3823, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 15.5, |
|
"learning_rate": 2.8195918367346942e-05, |
|
"loss": 0.396, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 15.6, |
|
"learning_rate": 2.8114285714285715e-05, |
|
"loss": 0.4025, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 15.7, |
|
"learning_rate": 2.803265306122449e-05, |
|
"loss": 0.3798, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 15.8, |
|
"learning_rate": 2.7951020408163268e-05, |
|
"loss": 0.3986, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 15.9, |
|
"learning_rate": 2.7869387755102045e-05, |
|
"loss": 0.3852, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"learning_rate": 2.7787755102040818e-05, |
|
"loss": 0.3983, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_loss": 0.18791833519935608, |
|
"eval_runtime": 173.6527, |
|
"eval_samples_per_second": 4.221, |
|
"eval_steps_per_second": 0.53, |
|
"eval_wer": 0.08654456433040167, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 16.1, |
|
"learning_rate": 2.7706122448979594e-05, |
|
"loss": 0.4246, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 16.2, |
|
"learning_rate": 2.7624489795918367e-05, |
|
"loss": 0.3979, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 16.3, |
|
"learning_rate": 2.7542857142857147e-05, |
|
"loss": 0.3959, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 16.4, |
|
"learning_rate": 2.746122448979592e-05, |
|
"loss": 0.3961, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 16.5, |
|
"learning_rate": 2.7379591836734697e-05, |
|
"loss": 0.376, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 16.6, |
|
"learning_rate": 2.729795918367347e-05, |
|
"loss": 0.3787, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 16.7, |
|
"learning_rate": 2.721632653061225e-05, |
|
"loss": 0.3862, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 16.8, |
|
"learning_rate": 2.7134693877551023e-05, |
|
"loss": 0.38, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 16.9, |
|
"learning_rate": 2.70530612244898e-05, |
|
"loss": 0.3677, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 2.6971428571428573e-05, |
|
"loss": 0.3907, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_loss": 0.1863466203212738, |
|
"eval_runtime": 172.3785, |
|
"eval_samples_per_second": 4.252, |
|
"eval_steps_per_second": 0.534, |
|
"eval_wer": 0.08559252878774141, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 17.1, |
|
"learning_rate": 2.6889795918367346e-05, |
|
"loss": 0.3982, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 17.2, |
|
"learning_rate": 2.6808163265306126e-05, |
|
"loss": 0.3787, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 17.3, |
|
"learning_rate": 2.6726530612244902e-05, |
|
"loss": 0.3862, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 17.4, |
|
"learning_rate": 2.6644897959183675e-05, |
|
"loss": 0.385, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 17.5, |
|
"learning_rate": 2.656326530612245e-05, |
|
"loss": 0.3927, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 17.6, |
|
"learning_rate": 2.6481632653061225e-05, |
|
"loss": 0.3803, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 17.7, |
|
"learning_rate": 2.6400000000000005e-05, |
|
"loss": 0.3765, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 17.8, |
|
"learning_rate": 2.6318367346938778e-05, |
|
"loss": 0.3928, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 17.9, |
|
"learning_rate": 2.623673469387755e-05, |
|
"loss": 0.3714, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"learning_rate": 2.6155102040816327e-05, |
|
"loss": 0.3969, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_loss": 0.17734822630882263, |
|
"eval_runtime": 172.508, |
|
"eval_samples_per_second": 4.249, |
|
"eval_steps_per_second": 0.533, |
|
"eval_wer": 0.08364312267657993, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 18.1, |
|
"learning_rate": 2.6073469387755107e-05, |
|
"loss": 0.396, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 18.2, |
|
"learning_rate": 2.599183673469388e-05, |
|
"loss": 0.3882, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 18.3, |
|
"learning_rate": 2.5910204081632654e-05, |
|
"loss": 0.395, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 18.4, |
|
"learning_rate": 2.582857142857143e-05, |
|
"loss": 0.3791, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 18.5, |
|
"learning_rate": 2.5746938775510203e-05, |
|
"loss": 0.3962, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 18.6, |
|
"learning_rate": 2.5665306122448983e-05, |
|
"loss": 0.3628, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 18.7, |
|
"learning_rate": 2.5583673469387756e-05, |
|
"loss": 0.3692, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 18.8, |
|
"learning_rate": 2.5502040816326533e-05, |
|
"loss": 0.3933, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 18.9, |
|
"learning_rate": 2.5420408163265306e-05, |
|
"loss": 0.3793, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"learning_rate": 2.5338775510204086e-05, |
|
"loss": 0.3721, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_loss": 0.18335653841495514, |
|
"eval_runtime": 174.5785, |
|
"eval_samples_per_second": 4.199, |
|
"eval_steps_per_second": 0.527, |
|
"eval_wer": 0.08899265572581376, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 19.1, |
|
"learning_rate": 2.525714285714286e-05, |
|
"loss": 0.391, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 19.2, |
|
"learning_rate": 2.5175510204081635e-05, |
|
"loss": 0.3779, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 19.3, |
|
"learning_rate": 2.509387755102041e-05, |
|
"loss": 0.3687, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 19.4, |
|
"learning_rate": 2.5012244897959185e-05, |
|
"loss": 0.38, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 19.5, |
|
"learning_rate": 2.493061224489796e-05, |
|
"loss": 0.3705, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 19.6, |
|
"learning_rate": 2.4848979591836738e-05, |
|
"loss": 0.372, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 19.7, |
|
"learning_rate": 2.476734693877551e-05, |
|
"loss": 0.3737, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 19.8, |
|
"learning_rate": 2.4685714285714288e-05, |
|
"loss": 0.3953, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 19.9, |
|
"learning_rate": 2.4604081632653064e-05, |
|
"loss": 0.3882, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"learning_rate": 2.452244897959184e-05, |
|
"loss": 0.3987, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_loss": 0.18172703683376312, |
|
"eval_runtime": 173.4894, |
|
"eval_samples_per_second": 4.225, |
|
"eval_steps_per_second": 0.53, |
|
"eval_wer": 0.08518451355517273, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 20.1, |
|
"learning_rate": 2.4440816326530614e-05, |
|
"loss": 0.3965, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 20.2, |
|
"learning_rate": 2.435918367346939e-05, |
|
"loss": 0.3672, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 20.3, |
|
"learning_rate": 2.4277551020408163e-05, |
|
"loss": 0.3942, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 20.4, |
|
"learning_rate": 2.4195918367346943e-05, |
|
"loss": 0.3792, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 20.5, |
|
"learning_rate": 2.4114285714285716e-05, |
|
"loss": 0.4021, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 20.6, |
|
"learning_rate": 2.4032653061224493e-05, |
|
"loss": 0.3825, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 20.7, |
|
"learning_rate": 2.3951020408163266e-05, |
|
"loss": 0.3852, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 20.8, |
|
"learning_rate": 2.3869387755102046e-05, |
|
"loss": 0.3779, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 20.9, |
|
"learning_rate": 2.378775510204082e-05, |
|
"loss": 0.3765, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"learning_rate": 2.3706122448979595e-05, |
|
"loss": 0.3863, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_loss": 0.1898309588432312, |
|
"eval_runtime": 172.6948, |
|
"eval_samples_per_second": 4.244, |
|
"eval_steps_per_second": 0.533, |
|
"eval_wer": 0.09139541209538489, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 21.1, |
|
"learning_rate": 2.362448979591837e-05, |
|
"loss": 0.4055, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 21.2, |
|
"learning_rate": 2.354285714285714e-05, |
|
"loss": 0.3747, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 21.3, |
|
"learning_rate": 2.346122448979592e-05, |
|
"loss": 0.4122, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 21.4, |
|
"learning_rate": 2.3379591836734698e-05, |
|
"loss": 0.3865, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 21.5, |
|
"learning_rate": 2.329795918367347e-05, |
|
"loss": 0.38, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 21.6, |
|
"learning_rate": 2.3216326530612244e-05, |
|
"loss": 0.3725, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 21.7, |
|
"learning_rate": 2.3142857142857145e-05, |
|
"loss": 0.3789, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 21.8, |
|
"learning_rate": 2.3061224489795922e-05, |
|
"loss": 0.3971, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 21.9, |
|
"learning_rate": 2.2979591836734695e-05, |
|
"loss": 0.3869, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"learning_rate": 2.289795918367347e-05, |
|
"loss": 0.4052, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_loss": 0.18824028968811035, |
|
"eval_runtime": 175.5898, |
|
"eval_samples_per_second": 4.175, |
|
"eval_steps_per_second": 0.524, |
|
"eval_wer": 0.08568319883942334, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 22.1, |
|
"learning_rate": 2.2816326530612245e-05, |
|
"loss": 0.4021, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 22.2, |
|
"learning_rate": 2.2734693877551024e-05, |
|
"loss": 0.38, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 22.3, |
|
"learning_rate": 2.2653061224489798e-05, |
|
"loss": 0.3941, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 22.4, |
|
"learning_rate": 2.2571428571428574e-05, |
|
"loss": 0.3721, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 22.5, |
|
"learning_rate": 2.2489795918367347e-05, |
|
"loss": 0.3775, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 22.6, |
|
"learning_rate": 2.2408163265306124e-05, |
|
"loss": 0.3874, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 22.7, |
|
"learning_rate": 2.23265306122449e-05, |
|
"loss": 0.3834, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 22.8, |
|
"learning_rate": 2.2244897959183677e-05, |
|
"loss": 0.4009, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 22.9, |
|
"learning_rate": 2.216326530612245e-05, |
|
"loss": 0.3777, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"learning_rate": 2.2081632653061223e-05, |
|
"loss": 0.3811, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_loss": 0.18739941716194153, |
|
"eval_runtime": 174.0925, |
|
"eval_samples_per_second": 4.21, |
|
"eval_steps_per_second": 0.528, |
|
"eval_wer": 0.08563786381358238, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 23.1, |
|
"learning_rate": 2.2000000000000003e-05, |
|
"loss": 0.3961, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 23.2, |
|
"learning_rate": 2.191836734693878e-05, |
|
"loss": 0.3867, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 23.3, |
|
"learning_rate": 2.1836734693877552e-05, |
|
"loss": 0.3862, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 23.4, |
|
"learning_rate": 2.1755102040816326e-05, |
|
"loss": 0.3683, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 23.5, |
|
"learning_rate": 2.1673469387755102e-05, |
|
"loss": 0.3824, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 23.6, |
|
"learning_rate": 2.1591836734693882e-05, |
|
"loss": 0.3796, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 23.7, |
|
"learning_rate": 2.1510204081632655e-05, |
|
"loss": 0.3945, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 23.8, |
|
"learning_rate": 2.1428571428571428e-05, |
|
"loss": 0.3754, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 23.9, |
|
"learning_rate": 2.1346938775510205e-05, |
|
"loss": 0.3678, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"learning_rate": 2.1265306122448985e-05, |
|
"loss": 0.3791, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_loss": 0.1931917518377304, |
|
"eval_runtime": 174.6983, |
|
"eval_samples_per_second": 4.196, |
|
"eval_steps_per_second": 0.527, |
|
"eval_wer": 0.08853930546740411, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 24.1, |
|
"learning_rate": 2.1183673469387758e-05, |
|
"loss": 0.3891, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 24.2, |
|
"learning_rate": 2.110204081632653e-05, |
|
"loss": 0.3672, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 24.3, |
|
"learning_rate": 2.1020408163265307e-05, |
|
"loss": 0.3785, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 24.4, |
|
"learning_rate": 2.093877551020408e-05, |
|
"loss": 0.355, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 24.5, |
|
"learning_rate": 2.085714285714286e-05, |
|
"loss": 0.3851, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 24.6, |
|
"learning_rate": 2.0775510204081633e-05, |
|
"loss": 0.3816, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 24.7, |
|
"learning_rate": 2.069387755102041e-05, |
|
"loss": 0.3727, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 24.8, |
|
"learning_rate": 2.0612244897959183e-05, |
|
"loss": 0.375, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 24.9, |
|
"learning_rate": 2.0530612244897963e-05, |
|
"loss": 0.3676, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"learning_rate": 2.0448979591836736e-05, |
|
"loss": 0.3919, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_loss": 0.18465474247932434, |
|
"eval_runtime": 175.2344, |
|
"eval_samples_per_second": 4.183, |
|
"eval_steps_per_second": 0.525, |
|
"eval_wer": 0.08151237646205459, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 25.1, |
|
"learning_rate": 2.0367346938775512e-05, |
|
"loss": 0.3866, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 25.2, |
|
"learning_rate": 2.0285714285714286e-05, |
|
"loss": 0.3527, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 25.3, |
|
"learning_rate": 2.0204081632653062e-05, |
|
"loss": 0.3673, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 25.4, |
|
"learning_rate": 2.012244897959184e-05, |
|
"loss": 0.3688, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 25.5, |
|
"learning_rate": 2.0040816326530615e-05, |
|
"loss": 0.384, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 25.6, |
|
"learning_rate": 1.9959183673469388e-05, |
|
"loss": 0.3903, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 25.7, |
|
"learning_rate": 1.9877551020408165e-05, |
|
"loss": 0.3658, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 25.8, |
|
"learning_rate": 1.979591836734694e-05, |
|
"loss": 0.3686, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 25.9, |
|
"learning_rate": 1.9714285714285718e-05, |
|
"loss": 0.3763, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"learning_rate": 1.963265306122449e-05, |
|
"loss": 0.3891, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_loss": 0.18502917885780334, |
|
"eval_runtime": 174.4839, |
|
"eval_samples_per_second": 4.201, |
|
"eval_steps_per_second": 0.527, |
|
"eval_wer": 0.08522984858101369, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 26.1, |
|
"learning_rate": 1.9551020408163267e-05, |
|
"loss": 0.3809, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 26.2, |
|
"learning_rate": 1.9469387755102044e-05, |
|
"loss": 0.367, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 26.3, |
|
"learning_rate": 1.9387755102040817e-05, |
|
"loss": 0.3714, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 26.4, |
|
"learning_rate": 1.9306122448979593e-05, |
|
"loss": 0.3652, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 26.5, |
|
"learning_rate": 1.922448979591837e-05, |
|
"loss": 0.39, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 26.6, |
|
"learning_rate": 1.9142857142857146e-05, |
|
"loss": 0.3837, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 26.7, |
|
"learning_rate": 1.906122448979592e-05, |
|
"loss": 0.3672, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 26.8, |
|
"learning_rate": 1.8979591836734696e-05, |
|
"loss": 0.3622, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 26.9, |
|
"learning_rate": 1.8897959183673473e-05, |
|
"loss": 0.3715, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"learning_rate": 1.8816326530612246e-05, |
|
"loss": 0.3719, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_loss": 0.17737819254398346, |
|
"eval_runtime": 173.6719, |
|
"eval_samples_per_second": 4.221, |
|
"eval_steps_per_second": 0.53, |
|
"eval_wer": 0.08196572672046423, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 27.1, |
|
"learning_rate": 1.8734693877551022e-05, |
|
"loss": 0.39, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 27.2, |
|
"learning_rate": 1.8653061224489795e-05, |
|
"loss": 0.3679, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 27.3, |
|
"learning_rate": 1.8571428571428575e-05, |
|
"loss": 0.3691, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 27.4, |
|
"learning_rate": 1.8489795918367348e-05, |
|
"loss": 0.3667, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 27.5, |
|
"learning_rate": 1.8408163265306125e-05, |
|
"loss": 0.3668, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 27.6, |
|
"learning_rate": 1.8326530612244898e-05, |
|
"loss": 0.3675, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 27.7, |
|
"learning_rate": 1.8244897959183674e-05, |
|
"loss": 0.3618, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 27.8, |
|
"learning_rate": 1.816326530612245e-05, |
|
"loss": 0.3733, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 27.9, |
|
"learning_rate": 1.8081632653061224e-05, |
|
"loss": 0.353, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"learning_rate": 1.8e-05, |
|
"loss": 0.3791, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_loss": 0.17557118833065033, |
|
"eval_runtime": 173.548, |
|
"eval_samples_per_second": 4.224, |
|
"eval_steps_per_second": 0.53, |
|
"eval_wer": 0.0825097470305558, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 28.1, |
|
"learning_rate": 1.7918367346938777e-05, |
|
"loss": 0.3865, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 28.2, |
|
"learning_rate": 1.7836734693877553e-05, |
|
"loss": 0.3771, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 28.3, |
|
"learning_rate": 1.7755102040816327e-05, |
|
"loss": 0.3693, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 28.4, |
|
"learning_rate": 1.7673469387755103e-05, |
|
"loss": 0.3523, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 28.5, |
|
"learning_rate": 1.759183673469388e-05, |
|
"loss": 0.386, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 28.6, |
|
"learning_rate": 1.7510204081632653e-05, |
|
"loss": 0.3812, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 28.7, |
|
"learning_rate": 1.742857142857143e-05, |
|
"loss": 0.3656, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 28.8, |
|
"learning_rate": 1.7346938775510206e-05, |
|
"loss": 0.37, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 28.9, |
|
"learning_rate": 1.7265306122448982e-05, |
|
"loss": 0.3688, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"learning_rate": 1.7183673469387755e-05, |
|
"loss": 0.3537, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"eval_loss": 0.17965510487556458, |
|
"eval_runtime": 173.5786, |
|
"eval_samples_per_second": 4.223, |
|
"eval_steps_per_second": 0.53, |
|
"eval_wer": 0.08441381811587632, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 29.1, |
|
"learning_rate": 1.7102040816326532e-05, |
|
"loss": 0.3828, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 29.2, |
|
"learning_rate": 1.7020408163265308e-05, |
|
"loss": 0.3535, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 29.3, |
|
"learning_rate": 1.6938775510204085e-05, |
|
"loss": 0.3866, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 29.4, |
|
"learning_rate": 1.6857142857142858e-05, |
|
"loss": 0.3625, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 29.5, |
|
"learning_rate": 1.6775510204081634e-05, |
|
"loss": 0.3625, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 29.6, |
|
"learning_rate": 1.669387755102041e-05, |
|
"loss": 0.3592, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 29.7, |
|
"learning_rate": 1.6612244897959184e-05, |
|
"loss": 0.3695, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 29.8, |
|
"learning_rate": 1.653061224489796e-05, |
|
"loss": 0.3562, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 29.9, |
|
"learning_rate": 1.6448979591836737e-05, |
|
"loss": 0.3638, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"learning_rate": 1.6367346938775513e-05, |
|
"loss": 0.361, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_loss": 0.18183788657188416, |
|
"eval_runtime": 173.8338, |
|
"eval_samples_per_second": 4.217, |
|
"eval_steps_per_second": 0.529, |
|
"eval_wer": 0.08341644754737511, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 30.1, |
|
"learning_rate": 1.6285714285714287e-05, |
|
"loss": 0.385, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 30.2, |
|
"learning_rate": 1.6204081632653063e-05, |
|
"loss": 0.345, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 30.3, |
|
"learning_rate": 1.612244897959184e-05, |
|
"loss": 0.3558, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 30.4, |
|
"learning_rate": 1.6040816326530613e-05, |
|
"loss": 0.345, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 30.5, |
|
"learning_rate": 1.595918367346939e-05, |
|
"loss": 0.3792, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 30.6, |
|
"learning_rate": 1.5877551020408162e-05, |
|
"loss": 0.368, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 30.7, |
|
"learning_rate": 1.5795918367346942e-05, |
|
"loss": 0.3566, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 30.8, |
|
"learning_rate": 1.5714285714285715e-05, |
|
"loss": 0.3678, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 30.9, |
|
"learning_rate": 1.5632653061224492e-05, |
|
"loss": 0.3627, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"learning_rate": 1.5551020408163265e-05, |
|
"loss": 0.3619, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"eval_loss": 0.1747395247220993, |
|
"eval_runtime": 173.6135, |
|
"eval_samples_per_second": 4.222, |
|
"eval_steps_per_second": 0.53, |
|
"eval_wer": 0.08377912775410282, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 31.1, |
|
"learning_rate": 1.546938775510204e-05, |
|
"loss": 0.3646, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 31.2, |
|
"learning_rate": 1.5387755102040818e-05, |
|
"loss": 0.3582, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 31.3, |
|
"learning_rate": 1.530612244897959e-05, |
|
"loss": 0.3569, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 31.4, |
|
"learning_rate": 1.522448979591837e-05, |
|
"loss": 0.3525, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 31.5, |
|
"learning_rate": 1.5142857142857144e-05, |
|
"loss": 0.3697, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 31.6, |
|
"learning_rate": 1.506122448979592e-05, |
|
"loss": 0.3526, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 31.7, |
|
"learning_rate": 1.4979591836734695e-05, |
|
"loss": 0.3664, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 31.8, |
|
"learning_rate": 1.4897959183673472e-05, |
|
"loss": 0.3573, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 31.9, |
|
"learning_rate": 1.4816326530612247e-05, |
|
"loss": 0.3572, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"learning_rate": 1.4734693877551021e-05, |
|
"loss": 0.3626, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_loss": 0.1773301512002945, |
|
"eval_runtime": 175.1211, |
|
"eval_samples_per_second": 4.186, |
|
"eval_steps_per_second": 0.525, |
|
"eval_wer": 0.08436848309003536, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 32.1, |
|
"learning_rate": 1.4653061224489798e-05, |
|
"loss": 0.3623, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 32.2, |
|
"learning_rate": 1.4571428571428573e-05, |
|
"loss": 0.3579, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 32.3, |
|
"learning_rate": 1.448979591836735e-05, |
|
"loss": 0.3705, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 32.4, |
|
"learning_rate": 1.4408163265306122e-05, |
|
"loss": 0.3537, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 32.5, |
|
"learning_rate": 1.43265306122449e-05, |
|
"loss": 0.3715, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 32.6, |
|
"learning_rate": 1.4244897959183674e-05, |
|
"loss": 0.3779, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 32.7, |
|
"learning_rate": 1.4163265306122452e-05, |
|
"loss": 0.3489, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 32.8, |
|
"learning_rate": 1.4081632653061225e-05, |
|
"loss": 0.3595, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 32.9, |
|
"learning_rate": 1.4e-05, |
|
"loss": 0.3554, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"learning_rate": 1.3918367346938776e-05, |
|
"loss": 0.3632, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"eval_loss": 0.17745888233184814, |
|
"eval_runtime": 174.6122, |
|
"eval_samples_per_second": 4.198, |
|
"eval_steps_per_second": 0.527, |
|
"eval_wer": 0.0825097470305558, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 33.1, |
|
"learning_rate": 1.3836734693877551e-05, |
|
"loss": 0.3818, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 33.2, |
|
"learning_rate": 1.3755102040816328e-05, |
|
"loss": 0.3592, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 33.3, |
|
"learning_rate": 1.3673469387755102e-05, |
|
"loss": 0.3509, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 33.4, |
|
"learning_rate": 1.3591836734693879e-05, |
|
"loss": 0.3365, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 33.5, |
|
"learning_rate": 1.3510204081632654e-05, |
|
"loss": 0.3472, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 33.6, |
|
"learning_rate": 1.3428571428571429e-05, |
|
"loss": 0.3491, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 33.7, |
|
"learning_rate": 1.3346938775510205e-05, |
|
"loss": 0.3476, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 33.8, |
|
"learning_rate": 1.326530612244898e-05, |
|
"loss": 0.3571, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 33.9, |
|
"learning_rate": 1.3183673469387756e-05, |
|
"loss": 0.3618, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"learning_rate": 1.3102040816326531e-05, |
|
"loss": 0.3666, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_loss": 0.18347617983818054, |
|
"eval_runtime": 173.8457, |
|
"eval_samples_per_second": 4.216, |
|
"eval_steps_per_second": 0.529, |
|
"eval_wer": 0.0858645389427872, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 34.1, |
|
"learning_rate": 1.3020408163265308e-05, |
|
"loss": 0.3738, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 34.2, |
|
"learning_rate": 1.2938775510204082e-05, |
|
"loss": 0.3535, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 34.3, |
|
"learning_rate": 1.2857142857142859e-05, |
|
"loss": 0.3574, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 34.4, |
|
"learning_rate": 1.2775510204081634e-05, |
|
"loss": 0.3466, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 34.5, |
|
"learning_rate": 1.2693877551020409e-05, |
|
"loss": 0.3633, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 34.6, |
|
"learning_rate": 1.2612244897959185e-05, |
|
"loss": 0.363, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 34.7, |
|
"learning_rate": 1.253061224489796e-05, |
|
"loss": 0.3387, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 34.8, |
|
"learning_rate": 1.2448979591836736e-05, |
|
"loss": 0.3424, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 34.9, |
|
"learning_rate": 1.2367346938775511e-05, |
|
"loss": 0.3612, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"learning_rate": 1.2285714285714288e-05, |
|
"loss": 0.3581, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"eval_loss": 0.18585096299648285, |
|
"eval_runtime": 174.888, |
|
"eval_samples_per_second": 4.191, |
|
"eval_steps_per_second": 0.526, |
|
"eval_wer": 0.08681657448544745, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 35.1, |
|
"learning_rate": 1.2204081632653062e-05, |
|
"loss": 0.3667, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 35.2, |
|
"learning_rate": 1.2122448979591839e-05, |
|
"loss": 0.3579, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 35.3, |
|
"learning_rate": 1.2040816326530614e-05, |
|
"loss": 0.3585, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 35.4, |
|
"learning_rate": 1.1959183673469389e-05, |
|
"loss": 0.3461, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 35.5, |
|
"learning_rate": 1.1877551020408165e-05, |
|
"loss": 0.348, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 35.6, |
|
"learning_rate": 1.179591836734694e-05, |
|
"loss": 0.3588, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 35.7, |
|
"learning_rate": 1.1714285714285716e-05, |
|
"loss": 0.343, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 35.8, |
|
"learning_rate": 1.1632653061224491e-05, |
|
"loss": 0.3504, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 35.9, |
|
"learning_rate": 1.1551020408163268e-05, |
|
"loss": 0.3509, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"learning_rate": 1.146938775510204e-05, |
|
"loss": 0.3665, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"eval_loss": 0.174110546708107, |
|
"eval_runtime": 174.6383, |
|
"eval_samples_per_second": 4.197, |
|
"eval_steps_per_second": 0.527, |
|
"eval_wer": 0.08491250340012693, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 36.1, |
|
"learning_rate": 1.1387755102040819e-05, |
|
"loss": 0.3689, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 36.2, |
|
"learning_rate": 1.1306122448979592e-05, |
|
"loss": 0.353, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 36.3, |
|
"learning_rate": 1.1224489795918367e-05, |
|
"loss": 0.3637, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 36.4, |
|
"learning_rate": 1.1142857142857143e-05, |
|
"loss": 0.3424, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 36.5, |
|
"learning_rate": 1.1061224489795918e-05, |
|
"loss": 0.3602, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 36.6, |
|
"learning_rate": 1.0979591836734695e-05, |
|
"loss": 0.3533, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 36.7, |
|
"learning_rate": 1.089795918367347e-05, |
|
"loss": 0.3363, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 36.8, |
|
"learning_rate": 1.0816326530612246e-05, |
|
"loss": 0.3413, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 36.9, |
|
"learning_rate": 1.073469387755102e-05, |
|
"loss": 0.3523, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"learning_rate": 1.0653061224489796e-05, |
|
"loss": 0.3495, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"eval_loss": 0.17902912199497223, |
|
"eval_runtime": 175.0822, |
|
"eval_samples_per_second": 4.187, |
|
"eval_steps_per_second": 0.525, |
|
"eval_wer": 0.0836884577024209, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 37.1, |
|
"learning_rate": 1.0571428571428572e-05, |
|
"loss": 0.3567, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 37.2, |
|
"learning_rate": 1.0489795918367347e-05, |
|
"loss": 0.3464, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 37.3, |
|
"learning_rate": 1.0408163265306123e-05, |
|
"loss": 0.365, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 37.4, |
|
"learning_rate": 1.0326530612244898e-05, |
|
"loss": 0.3564, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 37.5, |
|
"learning_rate": 1.0244897959183675e-05, |
|
"loss": 0.3404, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 37.6, |
|
"learning_rate": 1.016326530612245e-05, |
|
"loss": 0.3512, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 37.7, |
|
"learning_rate": 1.0081632653061226e-05, |
|
"loss": 0.3394, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 37.8, |
|
"learning_rate": 1e-05, |
|
"loss": 0.3573, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 37.9, |
|
"learning_rate": 9.918367346938776e-06, |
|
"loss": 0.3508, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"learning_rate": 9.836734693877552e-06, |
|
"loss": 0.3509, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"eval_loss": 0.17817473411560059, |
|
"eval_runtime": 174.2852, |
|
"eval_samples_per_second": 4.206, |
|
"eval_steps_per_second": 0.528, |
|
"eval_wer": 0.08409647293498958, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 38.1, |
|
"learning_rate": 9.755102040816327e-06, |
|
"loss": 0.3641, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 38.2, |
|
"learning_rate": 9.673469387755103e-06, |
|
"loss": 0.343, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 38.3, |
|
"learning_rate": 9.591836734693878e-06, |
|
"loss": 0.3438, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 38.4, |
|
"learning_rate": 9.510204081632653e-06, |
|
"loss": 0.3464, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 38.5, |
|
"learning_rate": 9.42857142857143e-06, |
|
"loss": 0.3496, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 38.6, |
|
"learning_rate": 9.346938775510204e-06, |
|
"loss": 0.3483, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 38.7, |
|
"learning_rate": 9.26530612244898e-06, |
|
"loss": 0.3457, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 38.8, |
|
"learning_rate": 9.183673469387756e-06, |
|
"loss": 0.3621, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 38.9, |
|
"learning_rate": 9.102040816326532e-06, |
|
"loss": 0.346, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"learning_rate": 9.020408163265307e-06, |
|
"loss": 0.3621, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"eval_loss": 0.17591050267219543, |
|
"eval_runtime": 174.746, |
|
"eval_samples_per_second": 4.195, |
|
"eval_steps_per_second": 0.526, |
|
"eval_wer": 0.08414180796083054, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 39.1, |
|
"learning_rate": 8.938775510204082e-06, |
|
"loss": 0.3704, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 39.2, |
|
"learning_rate": 8.857142857142858e-06, |
|
"loss": 0.3372, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 39.3, |
|
"learning_rate": 8.775510204081633e-06, |
|
"loss": 0.3465, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 39.4, |
|
"learning_rate": 8.69387755102041e-06, |
|
"loss": 0.3438, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 39.5, |
|
"learning_rate": 8.612244897959184e-06, |
|
"loss": 0.3522, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 39.6, |
|
"learning_rate": 8.530612244897961e-06, |
|
"loss": 0.3497, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 39.7, |
|
"learning_rate": 8.448979591836736e-06, |
|
"loss": 0.3516, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 39.8, |
|
"learning_rate": 8.36734693877551e-06, |
|
"loss": 0.3427, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 39.9, |
|
"learning_rate": 8.285714285714287e-06, |
|
"loss": 0.3504, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"learning_rate": 8.204081632653062e-06, |
|
"loss": 0.3415, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_loss": 0.1795545220375061, |
|
"eval_runtime": 174.0239, |
|
"eval_samples_per_second": 4.212, |
|
"eval_steps_per_second": 0.529, |
|
"eval_wer": 0.0850938435034908, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 40.1, |
|
"learning_rate": 8.122448979591837e-06, |
|
"loss": 0.3596, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 40.2, |
|
"learning_rate": 8.040816326530613e-06, |
|
"loss": 0.3393, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 40.3, |
|
"learning_rate": 7.959183673469388e-06, |
|
"loss": 0.3521, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 40.4, |
|
"learning_rate": 7.877551020408164e-06, |
|
"loss": 0.3497, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 40.5, |
|
"learning_rate": 7.79591836734694e-06, |
|
"loss": 0.3487, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 40.6, |
|
"learning_rate": 7.714285714285716e-06, |
|
"loss": 0.3422, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 40.7, |
|
"learning_rate": 7.63265306122449e-06, |
|
"loss": 0.3522, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 40.8, |
|
"learning_rate": 7.551020408163265e-06, |
|
"loss": 0.3465, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 40.9, |
|
"learning_rate": 7.469387755102041e-06, |
|
"loss": 0.3483, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 41.0, |
|
"learning_rate": 7.387755102040817e-06, |
|
"loss": 0.3508, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 41.0, |
|
"eval_loss": 0.1777496188879013, |
|
"eval_runtime": 174.0585, |
|
"eval_samples_per_second": 4.211, |
|
"eval_steps_per_second": 0.529, |
|
"eval_wer": 0.0821470668238281, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 41.1, |
|
"learning_rate": 7.306122448979592e-06, |
|
"loss": 0.3683, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 41.2, |
|
"learning_rate": 7.224489795918368e-06, |
|
"loss": 0.3355, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 41.3, |
|
"learning_rate": 7.1428571428571436e-06, |
|
"loss": 0.3538, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 41.4, |
|
"learning_rate": 7.061224489795919e-06, |
|
"loss": 0.3286, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 41.5, |
|
"learning_rate": 6.979591836734695e-06, |
|
"loss": 0.3405, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 41.6, |
|
"learning_rate": 6.8979591836734705e-06, |
|
"loss": 0.3466, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 41.7, |
|
"learning_rate": 6.816326530612245e-06, |
|
"loss": 0.3461, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 41.8, |
|
"learning_rate": 6.734693877551021e-06, |
|
"loss": 0.3576, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 41.9, |
|
"learning_rate": 6.653061224489797e-06, |
|
"loss": 0.3473, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 42.0, |
|
"learning_rate": 6.571428571428572e-06, |
|
"loss": 0.3493, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 42.0, |
|
"eval_loss": 0.17577549815177917, |
|
"eval_runtime": 173.4214, |
|
"eval_samples_per_second": 4.227, |
|
"eval_steps_per_second": 0.53, |
|
"eval_wer": 0.08291776226312449, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 42.1, |
|
"learning_rate": 6.489795918367348e-06, |
|
"loss": 0.368, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 42.2, |
|
"learning_rate": 6.408163265306124e-06, |
|
"loss": 0.3288, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 42.3, |
|
"learning_rate": 6.326530612244899e-06, |
|
"loss": 0.3571, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 42.4, |
|
"learning_rate": 6.244897959183675e-06, |
|
"loss": 0.3475, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 42.5, |
|
"learning_rate": 6.163265306122449e-06, |
|
"loss": 0.3545, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 42.6, |
|
"learning_rate": 6.0816326530612245e-06, |
|
"loss": 0.3412, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 42.7, |
|
"learning_rate": 6e-06, |
|
"loss": 0.3403, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 42.8, |
|
"learning_rate": 5.918367346938776e-06, |
|
"loss": 0.3511, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 42.9, |
|
"learning_rate": 5.8367346938775515e-06, |
|
"loss": 0.3519, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 43.0, |
|
"learning_rate": 5.755102040816327e-06, |
|
"loss": 0.359, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 43.0, |
|
"eval_loss": 0.17879725992679596, |
|
"eval_runtime": 175.0692, |
|
"eval_samples_per_second": 4.187, |
|
"eval_steps_per_second": 0.526, |
|
"eval_wer": 0.08482183334844501, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 43.1, |
|
"learning_rate": 5.673469387755103e-06, |
|
"loss": 0.355, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 43.2, |
|
"learning_rate": 5.591836734693878e-06, |
|
"loss": 0.3469, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 43.3, |
|
"learning_rate": 5.510204081632653e-06, |
|
"loss": 0.3599, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 43.4, |
|
"learning_rate": 5.428571428571429e-06, |
|
"loss": 0.3491, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 43.5, |
|
"learning_rate": 5.3469387755102045e-06, |
|
"loss": 0.3427, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 43.6, |
|
"learning_rate": 5.26530612244898e-06, |
|
"loss": 0.3431, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 43.7, |
|
"learning_rate": 5.183673469387756e-06, |
|
"loss": 0.3469, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 43.8, |
|
"learning_rate": 5.1020408163265315e-06, |
|
"loss": 0.3463, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 43.9, |
|
"learning_rate": 5.020408163265307e-06, |
|
"loss": 0.3388, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 44.0, |
|
"learning_rate": 4.938775510204082e-06, |
|
"loss": 0.3438, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 44.0, |
|
"eval_loss": 0.17821075022220612, |
|
"eval_runtime": 173.9712, |
|
"eval_samples_per_second": 4.213, |
|
"eval_steps_per_second": 0.529, |
|
"eval_wer": 0.083552452624898, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 44.1, |
|
"learning_rate": 4.857142857142858e-06, |
|
"loss": 0.3613, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 44.2, |
|
"learning_rate": 4.783673469387755e-06, |
|
"loss": 0.3289, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 44.3, |
|
"learning_rate": 4.702040816326531e-06, |
|
"loss": 0.3552, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 44.4, |
|
"learning_rate": 4.620408163265307e-06, |
|
"loss": 0.3328, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 44.5, |
|
"learning_rate": 4.538775510204082e-06, |
|
"loss": 0.3394, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 44.6, |
|
"learning_rate": 4.457142857142858e-06, |
|
"loss": 0.3394, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 44.7, |
|
"learning_rate": 4.375510204081633e-06, |
|
"loss": 0.3396, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 44.8, |
|
"learning_rate": 4.2938775510204085e-06, |
|
"loss": 0.3487, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 44.9, |
|
"learning_rate": 4.212244897959184e-06, |
|
"loss": 0.3349, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"learning_rate": 4.13061224489796e-06, |
|
"loss": 0.3642, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"eval_loss": 0.1731826514005661, |
|
"eval_runtime": 174.3109, |
|
"eval_samples_per_second": 4.205, |
|
"eval_steps_per_second": 0.528, |
|
"eval_wer": 0.08305376734064739, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 45.1, |
|
"learning_rate": 4.0489795918367354e-06, |
|
"loss": 0.3719, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 45.2, |
|
"learning_rate": 3.96734693877551e-06, |
|
"loss": 0.3443, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 45.3, |
|
"learning_rate": 3.885714285714286e-06, |
|
"loss": 0.3526, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 45.4, |
|
"learning_rate": 3.8040816326530616e-06, |
|
"loss": 0.3381, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 45.5, |
|
"learning_rate": 3.722448979591837e-06, |
|
"loss": 0.3517, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 45.6, |
|
"learning_rate": 3.6408163265306124e-06, |
|
"loss": 0.3514, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 45.7, |
|
"learning_rate": 3.559183673469388e-06, |
|
"loss": 0.3462, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 45.8, |
|
"learning_rate": 3.4775510204081637e-06, |
|
"loss": 0.3432, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 45.9, |
|
"learning_rate": 3.3959183673469394e-06, |
|
"loss": 0.332, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 46.0, |
|
"learning_rate": 3.314285714285714e-06, |
|
"loss": 0.3456, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 46.0, |
|
"eval_loss": 0.17676028609275818, |
|
"eval_runtime": 173.8583, |
|
"eval_samples_per_second": 4.216, |
|
"eval_steps_per_second": 0.529, |
|
"eval_wer": 0.08228307190135098, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 46.1, |
|
"learning_rate": 3.23265306122449e-06, |
|
"loss": 0.3592, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 46.2, |
|
"learning_rate": 3.1510204081632655e-06, |
|
"loss": 0.3565, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 46.3, |
|
"learning_rate": 3.069387755102041e-06, |
|
"loss": 0.3337, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 46.4, |
|
"learning_rate": 2.9877551020408164e-06, |
|
"loss": 0.332, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 46.5, |
|
"learning_rate": 2.906122448979592e-06, |
|
"loss": 0.3523, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 46.6, |
|
"learning_rate": 2.8244897959183677e-06, |
|
"loss": 0.3483, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 46.7, |
|
"learning_rate": 2.7428571428571433e-06, |
|
"loss": 0.3291, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 46.8, |
|
"learning_rate": 2.661224489795919e-06, |
|
"loss": 0.3492, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 46.9, |
|
"learning_rate": 2.579591836734694e-06, |
|
"loss": 0.3457, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 47.0, |
|
"learning_rate": 2.4979591836734694e-06, |
|
"loss": 0.3532, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 47.0, |
|
"eval_loss": 0.17354446649551392, |
|
"eval_runtime": 175.9111, |
|
"eval_samples_per_second": 4.167, |
|
"eval_steps_per_second": 0.523, |
|
"eval_wer": 0.08341644754737511, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 47.1, |
|
"learning_rate": 2.416326530612245e-06, |
|
"loss": 0.3615, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 47.2, |
|
"learning_rate": 2.3346938775510208e-06, |
|
"loss": 0.3242, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 47.3, |
|
"learning_rate": 2.2530612244897964e-06, |
|
"loss": 0.3462, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 47.4, |
|
"learning_rate": 2.1714285714285716e-06, |
|
"loss": 0.35, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 47.5, |
|
"learning_rate": 2.0897959183673473e-06, |
|
"loss": 0.3429, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 47.6, |
|
"learning_rate": 2.0081632653061225e-06, |
|
"loss": 0.334, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 47.7, |
|
"learning_rate": 1.926530612244898e-06, |
|
"loss": 0.3312, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 47.8, |
|
"learning_rate": 1.8448979591836736e-06, |
|
"loss": 0.3404, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 47.9, |
|
"learning_rate": 1.7632653061224493e-06, |
|
"loss": 0.33, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 48.0, |
|
"learning_rate": 1.6816326530612245e-06, |
|
"loss": 0.3448, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 48.0, |
|
"eval_loss": 0.17552991211414337, |
|
"eval_runtime": 173.8979, |
|
"eval_samples_per_second": 4.215, |
|
"eval_steps_per_second": 0.529, |
|
"eval_wer": 0.08273642215976063, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 48.1, |
|
"learning_rate": 1.6000000000000001e-06, |
|
"loss": 0.3642, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 48.2, |
|
"learning_rate": 1.5183673469387756e-06, |
|
"loss": 0.354, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 48.3, |
|
"learning_rate": 1.4367346938775512e-06, |
|
"loss": 0.3402, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 48.4, |
|
"learning_rate": 1.3551020408163265e-06, |
|
"loss": 0.3363, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 48.5, |
|
"learning_rate": 1.2734693877551021e-06, |
|
"loss": 0.3399, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 48.6, |
|
"learning_rate": 1.1918367346938776e-06, |
|
"loss": 0.3422, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 48.7, |
|
"learning_rate": 1.1102040816326532e-06, |
|
"loss": 0.3302, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 48.8, |
|
"learning_rate": 1.0285714285714286e-06, |
|
"loss": 0.3348, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 48.9, |
|
"learning_rate": 9.469387755102041e-07, |
|
"loss": 0.3351, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 49.0, |
|
"learning_rate": 8.653061224489796e-07, |
|
"loss": 0.3487, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 49.0, |
|
"eval_loss": 0.17674094438552856, |
|
"eval_runtime": 175.6308, |
|
"eval_samples_per_second": 4.174, |
|
"eval_steps_per_second": 0.524, |
|
"eval_wer": 0.08328044246985221, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 49.1, |
|
"learning_rate": 7.836734693877551e-07, |
|
"loss": 0.3531, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 49.2, |
|
"learning_rate": 7.020408163265306e-07, |
|
"loss": 0.3288, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 49.3, |
|
"learning_rate": 6.204081632653062e-07, |
|
"loss": 0.3413, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 49.4, |
|
"learning_rate": 5.387755102040817e-07, |
|
"loss": 0.3395, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 49.5, |
|
"learning_rate": 4.571428571428572e-07, |
|
"loss": 0.3483, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 49.6, |
|
"learning_rate": 3.755102040816327e-07, |
|
"loss": 0.3372, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 49.7, |
|
"learning_rate": 2.938775510204082e-07, |
|
"loss": 0.3417, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 49.8, |
|
"learning_rate": 2.122448979591837e-07, |
|
"loss": 0.3511, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 49.9, |
|
"learning_rate": 1.306122448979592e-07, |
|
"loss": 0.3413, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"learning_rate": 4.897959183673469e-08, |
|
"loss": 0.3427, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"eval_loss": 0.17737863957881927, |
|
"eval_runtime": 174.0664, |
|
"eval_samples_per_second": 4.211, |
|
"eval_steps_per_second": 0.529, |
|
"eval_wer": 0.083552452624898, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"step": 5000, |
|
"total_flos": 1.8766164792490646e+20, |
|
"train_loss": 0.37988694682121277, |
|
"train_runtime": 130186.5803, |
|
"train_samples_per_second": 2.47, |
|
"train_steps_per_second": 0.038 |
|
} |
|
], |
|
"max_steps": 5000, |
|
"num_train_epochs": 50, |
|
"total_flos": 1.8766164792490646e+20, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|