|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 149.99521531100478, |
|
"global_step": 15600, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 7e-08, |
|
"loss": 29.6957, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.7000000000000001e-07, |
|
"loss": 21.6242, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 2.7e-07, |
|
"loss": 24.7339, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 3.7e-07, |
|
"loss": 24.5014, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.7000000000000005e-07, |
|
"loss": 19.6193, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 5.6e-07, |
|
"loss": 24.3139, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 6.6e-07, |
|
"loss": 18.7014, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 7.6e-07, |
|
"loss": 16.5435, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 8.6e-07, |
|
"loss": 15.7272, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.600000000000001e-07, |
|
"loss": 9.5116, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.06e-06, |
|
"loss": 6.6373, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.1600000000000001e-06, |
|
"loss": 4.5028, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.26e-06, |
|
"loss": 4.1949, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.3600000000000001e-06, |
|
"loss": 3.9797, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.46e-06, |
|
"loss": 3.7875, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.56e-06, |
|
"loss": 3.7195, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 1.6600000000000002e-06, |
|
"loss": 3.6404, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 1.76e-06, |
|
"loss": 3.6115, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.8600000000000002e-06, |
|
"loss": 3.6285, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.9600000000000003e-06, |
|
"loss": 3.5665, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 2.06e-06, |
|
"loss": 3.7519, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 2.16e-06, |
|
"loss": 3.5291, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 2.2600000000000004e-06, |
|
"loss": 3.5265, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 2.3600000000000003e-06, |
|
"loss": 3.5154, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 2.46e-06, |
|
"loss": 3.4915, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 2.56e-06, |
|
"loss": 3.4865, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 2.6600000000000004e-06, |
|
"loss": 3.4817, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.7600000000000003e-06, |
|
"loss": 3.4582, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 2.86e-06, |
|
"loss": 3.4687, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 2.96e-06, |
|
"loss": 3.4268, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 3.0600000000000003e-06, |
|
"loss": 3.4107, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 3.1600000000000002e-06, |
|
"loss": 3.5738, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 3.2600000000000006e-06, |
|
"loss": 3.3898, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 3.3600000000000004e-06, |
|
"loss": 3.4021, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 3.46e-06, |
|
"loss": 3.3818, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 3.5600000000000002e-06, |
|
"loss": 3.369, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 3.66e-06, |
|
"loss": 3.3569, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 3.7600000000000004e-06, |
|
"loss": 3.328, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 3.86e-06, |
|
"loss": 3.3149, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 3.96e-06, |
|
"loss": 3.3187, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 4.060000000000001e-06, |
|
"loss": 3.3071, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 4.16e-06, |
|
"loss": 3.4383, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"learning_rate": 4.26e-06, |
|
"loss": 3.2655, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 4.23, |
|
"learning_rate": 4.360000000000001e-06, |
|
"loss": 3.2137, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 4.4600000000000005e-06, |
|
"loss": 3.2058, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"learning_rate": 4.56e-06, |
|
"loss": 3.1397, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 4.66e-06, |
|
"loss": 3.0939, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 4.76e-06, |
|
"loss": 3.0008, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 4.71, |
|
"learning_rate": 4.86e-06, |
|
"loss": 2.967, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"learning_rate": 4.960000000000001e-06, |
|
"loss": 2.9038, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"eval_loss": 3.0125303268432617, |
|
"eval_runtime": 63.0507, |
|
"eval_samples_per_second": 17.637, |
|
"eval_steps_per_second": 4.409, |
|
"eval_wer": 1.0, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"learning_rate": 5.060000000000001e-06, |
|
"loss": 2.798, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 5.1600000000000006e-06, |
|
"loss": 2.7026, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 5.1, |
|
"learning_rate": 5.2600000000000005e-06, |
|
"loss": 2.7382, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 5.19, |
|
"learning_rate": 5.36e-06, |
|
"loss": 2.523, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 5.29, |
|
"learning_rate": 5.460000000000001e-06, |
|
"loss": 2.4356, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 5.38, |
|
"learning_rate": 5.560000000000001e-06, |
|
"loss": 2.3135, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 5.48, |
|
"learning_rate": 5.66e-06, |
|
"loss": 2.2726, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 5.57, |
|
"learning_rate": 5.76e-06, |
|
"loss": 2.1982, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 5.67, |
|
"learning_rate": 5.86e-06, |
|
"loss": 2.1517, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 5.77, |
|
"learning_rate": 5.9600000000000005e-06, |
|
"loss": 2.1003, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 5.86, |
|
"learning_rate": 6.0600000000000004e-06, |
|
"loss": 2.0039, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 5.96, |
|
"learning_rate": 6.16e-06, |
|
"loss": 2.002, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 6.06, |
|
"learning_rate": 6.26e-06, |
|
"loss": 2.0369, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 6.15, |
|
"learning_rate": 6.360000000000001e-06, |
|
"loss": 1.9028, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 6.460000000000001e-06, |
|
"loss": 1.9208, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 6.34, |
|
"learning_rate": 6.560000000000001e-06, |
|
"loss": 1.7979, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 6.44, |
|
"learning_rate": 6.660000000000001e-06, |
|
"loss": 1.8423, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 6.54, |
|
"learning_rate": 6.760000000000001e-06, |
|
"loss": 1.789, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 6.63, |
|
"learning_rate": 6.860000000000001e-06, |
|
"loss": 1.7757, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 6.73, |
|
"learning_rate": 6.96e-06, |
|
"loss": 1.7959, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 6.82, |
|
"learning_rate": 7.06e-06, |
|
"loss": 1.7051, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 6.92, |
|
"learning_rate": 7.16e-06, |
|
"loss": 1.733, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 7.260000000000001e-06, |
|
"loss": 1.8638, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 7.11, |
|
"learning_rate": 7.360000000000001e-06, |
|
"loss": 1.651, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 7.21, |
|
"learning_rate": 7.4600000000000006e-06, |
|
"loss": 1.7302, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 7.31, |
|
"learning_rate": 7.5600000000000005e-06, |
|
"loss": 1.6475, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 7.4, |
|
"learning_rate": 7.660000000000001e-06, |
|
"loss": 1.6387, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 7.76e-06, |
|
"loss": 1.6625, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 7.59, |
|
"learning_rate": 7.860000000000001e-06, |
|
"loss": 1.5618, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 7.69, |
|
"learning_rate": 7.960000000000002e-06, |
|
"loss": 1.6026, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 7.78, |
|
"learning_rate": 8.06e-06, |
|
"loss": 1.5912, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 7.88, |
|
"learning_rate": 8.16e-06, |
|
"loss": 1.5853, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 7.98, |
|
"learning_rate": 8.26e-06, |
|
"loss": 1.6174, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 8.08, |
|
"learning_rate": 8.36e-06, |
|
"loss": 1.5932, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 8.17, |
|
"learning_rate": 8.46e-06, |
|
"loss": 1.5812, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 8.27, |
|
"learning_rate": 8.560000000000001e-06, |
|
"loss": 1.5598, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 8.36, |
|
"learning_rate": 8.66e-06, |
|
"loss": 1.4793, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 8.46, |
|
"learning_rate": 8.76e-06, |
|
"loss": 1.5743, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 8.56, |
|
"learning_rate": 8.860000000000002e-06, |
|
"loss": 1.5608, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 8.65, |
|
"learning_rate": 8.96e-06, |
|
"loss": 1.5068, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"learning_rate": 9.060000000000001e-06, |
|
"loss": 1.5273, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 8.84, |
|
"learning_rate": 9.16e-06, |
|
"loss": 1.3943, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 8.94, |
|
"learning_rate": 9.260000000000001e-06, |
|
"loss": 1.5222, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 9.360000000000002e-06, |
|
"loss": 1.5358, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 9.13, |
|
"learning_rate": 9.460000000000001e-06, |
|
"loss": 1.4342, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 9.23, |
|
"learning_rate": 9.56e-06, |
|
"loss": 1.4748, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 9.33, |
|
"learning_rate": 9.66e-06, |
|
"loss": 1.414, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 9.42, |
|
"learning_rate": 9.760000000000001e-06, |
|
"loss": 1.4316, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 9.52, |
|
"learning_rate": 9.86e-06, |
|
"loss": 1.4462, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 9.61, |
|
"learning_rate": 9.960000000000001e-06, |
|
"loss": 1.3777, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 9.61, |
|
"eval_loss": 0.8681221008300781, |
|
"eval_runtime": 63.2095, |
|
"eval_samples_per_second": 17.592, |
|
"eval_steps_per_second": 4.398, |
|
"eval_wer": 0.8753363228699551, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 9.71, |
|
"learning_rate": 1.006e-05, |
|
"loss": 1.4522, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 9.8, |
|
"learning_rate": 1.0160000000000001e-05, |
|
"loss": 1.391, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 9.9, |
|
"learning_rate": 1.0260000000000002e-05, |
|
"loss": 1.3864, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 1.036e-05, |
|
"loss": 1.4285, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 10.1, |
|
"learning_rate": 1.0460000000000001e-05, |
|
"loss": 1.3904, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 10.19, |
|
"learning_rate": 1.056e-05, |
|
"loss": 1.4098, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 10.29, |
|
"learning_rate": 1.0660000000000001e-05, |
|
"loss": 1.3707, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 10.38, |
|
"learning_rate": 1.0760000000000002e-05, |
|
"loss": 1.3386, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 10.48, |
|
"learning_rate": 1.0860000000000001e-05, |
|
"loss": 1.3558, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 10.57, |
|
"learning_rate": 1.0960000000000002e-05, |
|
"loss": 1.3114, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 10.67, |
|
"learning_rate": 1.1060000000000003e-05, |
|
"loss": 1.3547, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 10.77, |
|
"learning_rate": 1.1160000000000002e-05, |
|
"loss": 1.331, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 10.86, |
|
"learning_rate": 1.126e-05, |
|
"loss": 1.3099, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 10.96, |
|
"learning_rate": 1.136e-05, |
|
"loss": 1.3869, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 11.06, |
|
"learning_rate": 1.146e-05, |
|
"loss": 1.392, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 11.15, |
|
"learning_rate": 1.156e-05, |
|
"loss": 1.2713, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 11.25, |
|
"learning_rate": 1.166e-05, |
|
"loss": 1.3161, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 11.34, |
|
"learning_rate": 1.1760000000000001e-05, |
|
"loss": 1.2339, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 11.44, |
|
"learning_rate": 1.186e-05, |
|
"loss": 1.3172, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 11.54, |
|
"learning_rate": 1.196e-05, |
|
"loss": 1.3047, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 11.63, |
|
"learning_rate": 1.2060000000000001e-05, |
|
"loss": 1.2366, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 11.73, |
|
"learning_rate": 1.216e-05, |
|
"loss": 1.311, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 11.82, |
|
"learning_rate": 1.2260000000000001e-05, |
|
"loss": 1.2441, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 11.92, |
|
"learning_rate": 1.236e-05, |
|
"loss": 1.2552, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 12.02, |
|
"learning_rate": 1.2460000000000001e-05, |
|
"loss": 1.4014, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 12.11, |
|
"learning_rate": 1.2560000000000002e-05, |
|
"loss": 1.18, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 12.21, |
|
"learning_rate": 1.266e-05, |
|
"loss": 1.2591, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 12.31, |
|
"learning_rate": 1.2760000000000001e-05, |
|
"loss": 1.1673, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 12.4, |
|
"learning_rate": 1.286e-05, |
|
"loss": 1.2424, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 12.5, |
|
"learning_rate": 1.2960000000000001e-05, |
|
"loss": 1.2248, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 12.59, |
|
"learning_rate": 1.3060000000000002e-05, |
|
"loss": 1.152, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 12.69, |
|
"learning_rate": 1.3160000000000001e-05, |
|
"loss": 1.2635, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 12.78, |
|
"learning_rate": 1.3260000000000002e-05, |
|
"loss": 1.2267, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 12.88, |
|
"learning_rate": 1.3360000000000003e-05, |
|
"loss": 1.16, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 12.98, |
|
"learning_rate": 1.3460000000000002e-05, |
|
"loss": 1.2843, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 13.08, |
|
"learning_rate": 1.3560000000000002e-05, |
|
"loss": 1.2263, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 13.17, |
|
"learning_rate": 1.3660000000000001e-05, |
|
"loss": 1.2041, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 13.27, |
|
"learning_rate": 1.376e-05, |
|
"loss": 1.192, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 13.36, |
|
"learning_rate": 1.386e-05, |
|
"loss": 1.15, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 13.46, |
|
"learning_rate": 1.396e-05, |
|
"loss": 1.2492, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 13.56, |
|
"learning_rate": 1.4060000000000001e-05, |
|
"loss": 1.1699, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 13.65, |
|
"learning_rate": 1.416e-05, |
|
"loss": 1.2067, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 13.75, |
|
"learning_rate": 1.426e-05, |
|
"loss": 1.2176, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 13.84, |
|
"learning_rate": 1.4360000000000001e-05, |
|
"loss": 1.139, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 13.94, |
|
"learning_rate": 1.446e-05, |
|
"loss": 1.1917, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 1.4560000000000001e-05, |
|
"loss": 1.2279, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 14.13, |
|
"learning_rate": 1.466e-05, |
|
"loss": 1.1372, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 14.23, |
|
"learning_rate": 1.4760000000000001e-05, |
|
"loss": 1.1766, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 14.33, |
|
"learning_rate": 1.4860000000000002e-05, |
|
"loss": 1.1232, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 14.42, |
|
"learning_rate": 1.496e-05, |
|
"loss": 1.1436, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 14.42, |
|
"eval_loss": 0.625587522983551, |
|
"eval_runtime": 63.9432, |
|
"eval_samples_per_second": 17.39, |
|
"eval_steps_per_second": 4.348, |
|
"eval_wer": 0.7960538116591929, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 14.52, |
|
"learning_rate": 1.5060000000000001e-05, |
|
"loss": 1.169, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 14.61, |
|
"learning_rate": 1.516e-05, |
|
"loss": 1.1012, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 14.71, |
|
"learning_rate": 1.5260000000000003e-05, |
|
"loss": 1.2028, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 14.8, |
|
"learning_rate": 1.5360000000000002e-05, |
|
"loss": 1.1238, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 14.9, |
|
"learning_rate": 1.546e-05, |
|
"loss": 1.1196, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 1.556e-05, |
|
"loss": 1.2336, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 15.1, |
|
"learning_rate": 1.5660000000000003e-05, |
|
"loss": 1.1066, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 15.19, |
|
"learning_rate": 1.576e-05, |
|
"loss": 1.1218, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 15.29, |
|
"learning_rate": 1.586e-05, |
|
"loss": 1.1322, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 15.38, |
|
"learning_rate": 1.5960000000000003e-05, |
|
"loss": 1.1534, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 15.48, |
|
"learning_rate": 1.6060000000000002e-05, |
|
"loss": 1.1725, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 15.57, |
|
"learning_rate": 1.616e-05, |
|
"loss": 1.0691, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 15.67, |
|
"learning_rate": 1.626e-05, |
|
"loss": 1.1454, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 15.77, |
|
"learning_rate": 1.636e-05, |
|
"loss": 1.1552, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 15.86, |
|
"learning_rate": 1.646e-05, |
|
"loss": 1.0275, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 15.96, |
|
"learning_rate": 1.656e-05, |
|
"loss": 1.1397, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 16.06, |
|
"learning_rate": 1.666e-05, |
|
"loss": 1.1147, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 16.15, |
|
"learning_rate": 1.6760000000000002e-05, |
|
"loss": 1.1183, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 16.25, |
|
"learning_rate": 1.686e-05, |
|
"loss": 1.1461, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 16.34, |
|
"learning_rate": 1.696e-05, |
|
"loss": 1.0195, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 16.44, |
|
"learning_rate": 1.7060000000000003e-05, |
|
"loss": 1.1009, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 16.54, |
|
"learning_rate": 1.7160000000000002e-05, |
|
"loss": 1.1064, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 16.63, |
|
"learning_rate": 1.726e-05, |
|
"loss": 1.0747, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 16.73, |
|
"learning_rate": 1.736e-05, |
|
"loss": 1.0235, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 16.82, |
|
"learning_rate": 1.7460000000000002e-05, |
|
"loss": 1.0612, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 16.92, |
|
"learning_rate": 1.756e-05, |
|
"loss": 1.1524, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 17.02, |
|
"learning_rate": 1.766e-05, |
|
"loss": 1.1562, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 17.11, |
|
"learning_rate": 1.7760000000000003e-05, |
|
"loss": 0.9857, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 17.21, |
|
"learning_rate": 1.7860000000000002e-05, |
|
"loss": 1.1104, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 17.31, |
|
"learning_rate": 1.796e-05, |
|
"loss": 1.0533, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 17.4, |
|
"learning_rate": 1.8060000000000003e-05, |
|
"loss": 1.0275, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 17.5, |
|
"learning_rate": 1.8160000000000002e-05, |
|
"loss": 1.1619, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 17.59, |
|
"learning_rate": 1.826e-05, |
|
"loss": 1.0009, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 17.69, |
|
"learning_rate": 1.8360000000000004e-05, |
|
"loss": 1.1132, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 17.78, |
|
"learning_rate": 1.8460000000000003e-05, |
|
"loss": 1.127, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 17.88, |
|
"learning_rate": 1.8560000000000002e-05, |
|
"loss": 1.0414, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 17.98, |
|
"learning_rate": 1.866e-05, |
|
"loss": 1.1011, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 18.08, |
|
"learning_rate": 1.876e-05, |
|
"loss": 1.0154, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 18.17, |
|
"learning_rate": 1.886e-05, |
|
"loss": 1.0592, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 18.27, |
|
"learning_rate": 1.896e-05, |
|
"loss": 1.0987, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 18.36, |
|
"learning_rate": 1.906e-05, |
|
"loss": 0.995, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 18.46, |
|
"learning_rate": 1.916e-05, |
|
"loss": 1.0891, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 18.56, |
|
"learning_rate": 1.9260000000000002e-05, |
|
"loss": 1.0455, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 18.65, |
|
"learning_rate": 1.936e-05, |
|
"loss": 1.0553, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 18.75, |
|
"learning_rate": 1.946e-05, |
|
"loss": 1.0649, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 18.84, |
|
"learning_rate": 1.9560000000000002e-05, |
|
"loss": 0.9852, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 18.94, |
|
"learning_rate": 1.966e-05, |
|
"loss": 1.0887, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 19.04, |
|
"learning_rate": 1.976e-05, |
|
"loss": 1.1287, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 19.13, |
|
"learning_rate": 1.9860000000000003e-05, |
|
"loss": 1.0216, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 19.23, |
|
"learning_rate": 1.9960000000000002e-05, |
|
"loss": 1.0997, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 19.23, |
|
"eval_loss": 0.5243648886680603, |
|
"eval_runtime": 62.7358, |
|
"eval_samples_per_second": 17.725, |
|
"eval_steps_per_second": 4.431, |
|
"eval_wer": 0.6875336322869955, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 19.33, |
|
"learning_rate": 1.9991176470588236e-05, |
|
"loss": 0.9947, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 19.42, |
|
"learning_rate": 1.9976470588235294e-05, |
|
"loss": 1.0263, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 19.52, |
|
"learning_rate": 1.9961764705882354e-05, |
|
"loss": 1.0212, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 19.61, |
|
"learning_rate": 1.9947058823529412e-05, |
|
"loss": 0.9756, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 19.71, |
|
"learning_rate": 1.9932352941176473e-05, |
|
"loss": 1.0833, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 19.8, |
|
"learning_rate": 1.991764705882353e-05, |
|
"loss": 0.9545, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 19.9, |
|
"learning_rate": 1.9902941176470588e-05, |
|
"loss": 0.9908, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"learning_rate": 1.988823529411765e-05, |
|
"loss": 1.1216, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 20.1, |
|
"learning_rate": 1.9873529411764706e-05, |
|
"loss": 0.9908, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 20.19, |
|
"learning_rate": 1.9858823529411767e-05, |
|
"loss": 1.0394, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 20.29, |
|
"learning_rate": 1.9844117647058824e-05, |
|
"loss": 0.9986, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 20.38, |
|
"learning_rate": 1.9829411764705885e-05, |
|
"loss": 0.9719, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 20.48, |
|
"learning_rate": 1.9814705882352943e-05, |
|
"loss": 1.0837, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 20.57, |
|
"learning_rate": 1.98e-05, |
|
"loss": 0.9644, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 20.67, |
|
"learning_rate": 1.978529411764706e-05, |
|
"loss": 1.0182, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 20.77, |
|
"learning_rate": 1.977058823529412e-05, |
|
"loss": 0.9953, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 20.86, |
|
"learning_rate": 1.975588235294118e-05, |
|
"loss": 0.966, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 20.96, |
|
"learning_rate": 1.9741176470588237e-05, |
|
"loss": 1.0316, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 21.06, |
|
"learning_rate": 1.9726470588235294e-05, |
|
"loss": 1.0149, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 21.15, |
|
"learning_rate": 1.9711764705882355e-05, |
|
"loss": 0.9778, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 21.25, |
|
"learning_rate": 1.9697058823529413e-05, |
|
"loss": 1.0777, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 21.34, |
|
"learning_rate": 1.9682352941176473e-05, |
|
"loss": 0.9694, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 21.44, |
|
"learning_rate": 1.966764705882353e-05, |
|
"loss": 1.004, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 21.54, |
|
"learning_rate": 1.9652941176470592e-05, |
|
"loss": 0.9523, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 21.63, |
|
"learning_rate": 1.963823529411765e-05, |
|
"loss": 0.9718, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 21.73, |
|
"learning_rate": 1.9623529411764707e-05, |
|
"loss": 0.9874, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 21.82, |
|
"learning_rate": 1.9608823529411768e-05, |
|
"loss": 0.9471, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 21.92, |
|
"learning_rate": 1.9594117647058825e-05, |
|
"loss": 0.9925, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 22.02, |
|
"learning_rate": 1.9579411764705886e-05, |
|
"loss": 1.0008, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 22.11, |
|
"learning_rate": 1.9564705882352943e-05, |
|
"loss": 0.9036, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 22.21, |
|
"learning_rate": 1.955e-05, |
|
"loss": 0.984, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 22.31, |
|
"learning_rate": 1.9535294117647062e-05, |
|
"loss": 0.9186, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 22.4, |
|
"learning_rate": 1.952058823529412e-05, |
|
"loss": 0.95, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 22.5, |
|
"learning_rate": 1.950588235294118e-05, |
|
"loss": 1.0427, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 22.59, |
|
"learning_rate": 1.9491176470588238e-05, |
|
"loss": 0.9026, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 22.69, |
|
"learning_rate": 1.94764705882353e-05, |
|
"loss": 1.0368, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 22.78, |
|
"learning_rate": 1.9461764705882356e-05, |
|
"loss": 0.9536, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 22.88, |
|
"learning_rate": 1.9447058823529413e-05, |
|
"loss": 0.9426, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 22.98, |
|
"learning_rate": 1.943235294117647e-05, |
|
"loss": 0.9868, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 23.08, |
|
"learning_rate": 1.941764705882353e-05, |
|
"loss": 0.9706, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 23.17, |
|
"learning_rate": 1.940294117647059e-05, |
|
"loss": 0.9642, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 23.27, |
|
"learning_rate": 1.9388235294117647e-05, |
|
"loss": 0.9675, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 23.36, |
|
"learning_rate": 1.9373529411764707e-05, |
|
"loss": 0.9074, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 23.46, |
|
"learning_rate": 1.9358823529411765e-05, |
|
"loss": 0.9888, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 23.56, |
|
"learning_rate": 1.9344117647058826e-05, |
|
"loss": 0.9058, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 23.65, |
|
"learning_rate": 1.9329411764705883e-05, |
|
"loss": 0.9327, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 23.75, |
|
"learning_rate": 1.931470588235294e-05, |
|
"loss": 1.0077, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 23.84, |
|
"learning_rate": 1.93e-05, |
|
"loss": 0.925, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 23.94, |
|
"learning_rate": 1.928529411764706e-05, |
|
"loss": 0.9754, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 24.04, |
|
"learning_rate": 1.927058823529412e-05, |
|
"loss": 1.0363, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 24.04, |
|
"eval_loss": 0.458545058965683, |
|
"eval_runtime": 63.756, |
|
"eval_samples_per_second": 17.441, |
|
"eval_steps_per_second": 4.36, |
|
"eval_wer": 0.6276233183856502, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 24.13, |
|
"learning_rate": 1.9255882352941177e-05, |
|
"loss": 0.906, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 24.23, |
|
"learning_rate": 1.9241176470588235e-05, |
|
"loss": 1.0274, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 24.33, |
|
"learning_rate": 1.9226470588235296e-05, |
|
"loss": 0.8774, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 24.42, |
|
"learning_rate": 1.9211764705882353e-05, |
|
"loss": 0.9585, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 24.52, |
|
"learning_rate": 1.9197058823529414e-05, |
|
"loss": 1.0052, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 24.61, |
|
"learning_rate": 1.918235294117647e-05, |
|
"loss": 0.868, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 24.71, |
|
"learning_rate": 1.9167647058823532e-05, |
|
"loss": 0.9666, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 24.8, |
|
"learning_rate": 1.915294117647059e-05, |
|
"loss": 0.8933, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 24.9, |
|
"learning_rate": 1.9138235294117647e-05, |
|
"loss": 0.967, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"learning_rate": 1.9123529411764708e-05, |
|
"loss": 1.0119, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 25.1, |
|
"learning_rate": 1.9108823529411766e-05, |
|
"loss": 0.9028, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 25.19, |
|
"learning_rate": 1.9094117647058826e-05, |
|
"loss": 0.9376, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 25.29, |
|
"learning_rate": 1.9079411764705884e-05, |
|
"loss": 0.9066, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 25.38, |
|
"learning_rate": 1.906470588235294e-05, |
|
"loss": 0.9027, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 25.48, |
|
"learning_rate": 1.9050000000000002e-05, |
|
"loss": 0.9117, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 25.57, |
|
"learning_rate": 1.903529411764706e-05, |
|
"loss": 0.8336, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 25.67, |
|
"learning_rate": 1.902058823529412e-05, |
|
"loss": 0.9391, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 25.77, |
|
"learning_rate": 1.9005882352941178e-05, |
|
"loss": 0.933, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 25.86, |
|
"learning_rate": 1.899117647058824e-05, |
|
"loss": 0.8725, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 25.96, |
|
"learning_rate": 1.8976470588235296e-05, |
|
"loss": 1.0004, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 26.06, |
|
"learning_rate": 1.8961764705882354e-05, |
|
"loss": 1.0043, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 26.15, |
|
"learning_rate": 1.8947058823529415e-05, |
|
"loss": 0.8783, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 26.25, |
|
"learning_rate": 1.8932352941176472e-05, |
|
"loss": 0.9454, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 26.34, |
|
"learning_rate": 1.8917647058823533e-05, |
|
"loss": 0.8055, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 26.44, |
|
"learning_rate": 1.890294117647059e-05, |
|
"loss": 0.9301, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 26.54, |
|
"learning_rate": 1.8888235294117648e-05, |
|
"loss": 0.9102, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 26.63, |
|
"learning_rate": 1.887352941176471e-05, |
|
"loss": 0.8577, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 26.73, |
|
"learning_rate": 1.8858823529411766e-05, |
|
"loss": 0.9306, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 26.82, |
|
"learning_rate": 1.8844117647058827e-05, |
|
"loss": 0.8715, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 26.92, |
|
"learning_rate": 1.8829411764705885e-05, |
|
"loss": 0.9273, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 27.02, |
|
"learning_rate": 1.8814705882352945e-05, |
|
"loss": 0.9756, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 27.11, |
|
"learning_rate": 1.88e-05, |
|
"loss": 0.8258, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 27.21, |
|
"learning_rate": 1.878529411764706e-05, |
|
"loss": 0.9032, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 27.31, |
|
"learning_rate": 1.8770588235294118e-05, |
|
"loss": 0.8094, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 27.4, |
|
"learning_rate": 1.875588235294118e-05, |
|
"loss": 0.9017, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 27.5, |
|
"learning_rate": 1.8741176470588236e-05, |
|
"loss": 0.8888, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 27.59, |
|
"learning_rate": 1.8726470588235294e-05, |
|
"loss": 0.8205, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 27.69, |
|
"learning_rate": 1.8711764705882355e-05, |
|
"loss": 0.8766, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 27.78, |
|
"learning_rate": 1.8697058823529412e-05, |
|
"loss": 0.8828, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 27.88, |
|
"learning_rate": 1.8682352941176473e-05, |
|
"loss": 0.8596, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 27.98, |
|
"learning_rate": 1.866764705882353e-05, |
|
"loss": 0.9077, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 28.08, |
|
"learning_rate": 1.8652941176470588e-05, |
|
"loss": 0.8044, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 28.17, |
|
"learning_rate": 1.863823529411765e-05, |
|
"loss": 0.89, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 28.27, |
|
"learning_rate": 1.8623529411764706e-05, |
|
"loss": 0.8988, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 28.36, |
|
"learning_rate": 1.8608823529411767e-05, |
|
"loss": 0.8302, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 28.46, |
|
"learning_rate": 1.8594117647058824e-05, |
|
"loss": 0.9261, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 28.56, |
|
"learning_rate": 1.8579411764705882e-05, |
|
"loss": 0.8275, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 28.65, |
|
"learning_rate": 1.8564705882352943e-05, |
|
"loss": 0.8713, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 28.75, |
|
"learning_rate": 1.855e-05, |
|
"loss": 0.9173, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 28.84, |
|
"learning_rate": 1.853529411764706e-05, |
|
"loss": 0.7996, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 28.84, |
|
"eval_loss": 0.40719929337501526, |
|
"eval_runtime": 63.2745, |
|
"eval_samples_per_second": 17.574, |
|
"eval_steps_per_second": 4.394, |
|
"eval_wer": 0.5295067264573992, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 28.94, |
|
"learning_rate": 1.852058823529412e-05, |
|
"loss": 0.9075, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 29.04, |
|
"learning_rate": 1.850588235294118e-05, |
|
"loss": 0.8718, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 29.13, |
|
"learning_rate": 1.8491176470588237e-05, |
|
"loss": 0.7867, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 29.23, |
|
"learning_rate": 1.8476470588235294e-05, |
|
"loss": 0.8962, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 29.33, |
|
"learning_rate": 1.8461764705882355e-05, |
|
"loss": 0.8403, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 29.42, |
|
"learning_rate": 1.8447058823529413e-05, |
|
"loss": 0.845, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 29.52, |
|
"learning_rate": 1.8432352941176474e-05, |
|
"loss": 0.9138, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 29.61, |
|
"learning_rate": 1.841764705882353e-05, |
|
"loss": 0.8396, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 29.71, |
|
"learning_rate": 1.840294117647059e-05, |
|
"loss": 0.912, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 29.8, |
|
"learning_rate": 1.838823529411765e-05, |
|
"loss": 0.8085, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 29.9, |
|
"learning_rate": 1.8373529411764707e-05, |
|
"loss": 0.8836, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"learning_rate": 1.8358823529411768e-05, |
|
"loss": 0.896, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 30.1, |
|
"learning_rate": 1.8344117647058825e-05, |
|
"loss": 0.7952, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 30.19, |
|
"learning_rate": 1.8329411764705886e-05, |
|
"loss": 0.8703, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 30.29, |
|
"learning_rate": 1.8314705882352944e-05, |
|
"loss": 0.8771, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 30.38, |
|
"learning_rate": 1.83e-05, |
|
"loss": 0.8202, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 30.48, |
|
"learning_rate": 1.8285294117647062e-05, |
|
"loss": 0.8612, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 30.57, |
|
"learning_rate": 1.827058823529412e-05, |
|
"loss": 0.8124, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 30.67, |
|
"learning_rate": 1.825588235294118e-05, |
|
"loss": 0.8492, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 30.77, |
|
"learning_rate": 1.8241176470588238e-05, |
|
"loss": 0.8913, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 30.86, |
|
"learning_rate": 1.8226470588235295e-05, |
|
"loss": 0.8128, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 30.96, |
|
"learning_rate": 1.8211764705882356e-05, |
|
"loss": 0.8757, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 31.06, |
|
"learning_rate": 1.8197058823529413e-05, |
|
"loss": 0.8491, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 31.15, |
|
"learning_rate": 1.818235294117647e-05, |
|
"loss": 0.8629, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 31.25, |
|
"learning_rate": 1.816764705882353e-05, |
|
"loss": 0.8527, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 31.34, |
|
"learning_rate": 1.815294117647059e-05, |
|
"loss": 0.7597, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 31.44, |
|
"learning_rate": 1.8138235294117647e-05, |
|
"loss": 0.8674, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 31.54, |
|
"learning_rate": 1.8123529411764708e-05, |
|
"loss": 0.8364, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 31.63, |
|
"learning_rate": 1.8108823529411765e-05, |
|
"loss": 0.7917, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 31.73, |
|
"learning_rate": 1.8094117647058826e-05, |
|
"loss": 0.8861, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 31.82, |
|
"learning_rate": 1.8079411764705883e-05, |
|
"loss": 0.8226, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 31.92, |
|
"learning_rate": 1.806470588235294e-05, |
|
"loss": 0.832, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 32.02, |
|
"learning_rate": 1.805e-05, |
|
"loss": 0.9101, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 32.11, |
|
"learning_rate": 1.803529411764706e-05, |
|
"loss": 0.7659, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 32.21, |
|
"learning_rate": 1.802058823529412e-05, |
|
"loss": 0.8449, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 32.31, |
|
"learning_rate": 1.8005882352941177e-05, |
|
"loss": 0.8284, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 32.4, |
|
"learning_rate": 1.7991176470588235e-05, |
|
"loss": 0.8372, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 32.5, |
|
"learning_rate": 1.7976470588235296e-05, |
|
"loss": 0.8996, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 32.59, |
|
"learning_rate": 1.7961764705882353e-05, |
|
"loss": 0.7571, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 32.69, |
|
"learning_rate": 1.7947058823529414e-05, |
|
"loss": 0.8943, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 32.78, |
|
"learning_rate": 1.793235294117647e-05, |
|
"loss": 0.8313, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 32.88, |
|
"learning_rate": 1.791764705882353e-05, |
|
"loss": 0.7835, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 32.98, |
|
"learning_rate": 1.790294117647059e-05, |
|
"loss": 0.8989, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 33.08, |
|
"learning_rate": 1.7888235294117647e-05, |
|
"loss": 0.8114, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 33.17, |
|
"learning_rate": 1.7873529411764708e-05, |
|
"loss": 0.8343, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 33.27, |
|
"learning_rate": 1.7858823529411766e-05, |
|
"loss": 0.7948, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 33.36, |
|
"learning_rate": 1.7844117647058827e-05, |
|
"loss": 0.7305, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 33.46, |
|
"learning_rate": 1.7829411764705884e-05, |
|
"loss": 0.8569, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 33.56, |
|
"learning_rate": 1.781470588235294e-05, |
|
"loss": 0.7362, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 33.65, |
|
"learning_rate": 1.7800000000000002e-05, |
|
"loss": 0.825, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 33.65, |
|
"eval_loss": 0.35900309681892395, |
|
"eval_runtime": 63.2442, |
|
"eval_samples_per_second": 17.583, |
|
"eval_steps_per_second": 4.396, |
|
"eval_wer": 0.522152466367713, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 33.75, |
|
"learning_rate": 1.778529411764706e-05, |
|
"loss": 0.9049, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 33.84, |
|
"learning_rate": 1.777058823529412e-05, |
|
"loss": 0.7655, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 33.94, |
|
"learning_rate": 1.7755882352941178e-05, |
|
"loss": 0.7926, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 34.04, |
|
"learning_rate": 1.7741176470588236e-05, |
|
"loss": 0.8244, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 34.13, |
|
"learning_rate": 1.7726470588235296e-05, |
|
"loss": 0.7378, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 34.23, |
|
"learning_rate": 1.7711764705882354e-05, |
|
"loss": 0.8426, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 34.33, |
|
"learning_rate": 1.7697058823529415e-05, |
|
"loss": 0.8147, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 34.42, |
|
"learning_rate": 1.7682352941176472e-05, |
|
"loss": 0.7837, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 34.52, |
|
"learning_rate": 1.7667647058823533e-05, |
|
"loss": 0.856, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 34.61, |
|
"learning_rate": 1.765294117647059e-05, |
|
"loss": 0.7904, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 34.71, |
|
"learning_rate": 1.7638235294117648e-05, |
|
"loss": 0.8817, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 34.8, |
|
"learning_rate": 1.762352941176471e-05, |
|
"loss": 0.8031, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 34.9, |
|
"learning_rate": 1.7608823529411766e-05, |
|
"loss": 0.8534, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"learning_rate": 1.7594117647058827e-05, |
|
"loss": 0.8451, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 35.1, |
|
"learning_rate": 1.7579411764705885e-05, |
|
"loss": 0.792, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 35.19, |
|
"learning_rate": 1.7564705882352942e-05, |
|
"loss": 0.7987, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 35.29, |
|
"learning_rate": 1.755e-05, |
|
"loss": 0.8161, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 35.38, |
|
"learning_rate": 1.753529411764706e-05, |
|
"loss": 0.7552, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 35.48, |
|
"learning_rate": 1.7520588235294118e-05, |
|
"loss": 0.8783, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 35.57, |
|
"learning_rate": 1.7505882352941175e-05, |
|
"loss": 0.7565, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 35.67, |
|
"learning_rate": 1.7491176470588236e-05, |
|
"loss": 0.7466, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 35.77, |
|
"learning_rate": 1.7476470588235294e-05, |
|
"loss": 0.8693, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 35.86, |
|
"learning_rate": 1.7461764705882355e-05, |
|
"loss": 0.6972, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 35.96, |
|
"learning_rate": 1.7447058823529412e-05, |
|
"loss": 0.7748, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 36.06, |
|
"learning_rate": 1.7432352941176473e-05, |
|
"loss": 0.7987, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 36.15, |
|
"learning_rate": 1.741764705882353e-05, |
|
"loss": 0.7611, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 36.25, |
|
"learning_rate": 1.7402941176470588e-05, |
|
"loss": 0.8033, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 36.34, |
|
"learning_rate": 1.738823529411765e-05, |
|
"loss": 0.6933, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 36.44, |
|
"learning_rate": 1.7373529411764706e-05, |
|
"loss": 0.8096, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 36.54, |
|
"learning_rate": 1.7358823529411767e-05, |
|
"loss": 0.7609, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 36.63, |
|
"learning_rate": 1.7344117647058825e-05, |
|
"loss": 0.7653, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 36.73, |
|
"learning_rate": 1.7329411764705882e-05, |
|
"loss": 0.7876, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 36.82, |
|
"learning_rate": 1.7314705882352943e-05, |
|
"loss": 0.7517, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 36.92, |
|
"learning_rate": 1.73e-05, |
|
"loss": 0.8098, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 37.02, |
|
"learning_rate": 1.728529411764706e-05, |
|
"loss": 0.8721, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 37.11, |
|
"learning_rate": 1.727058823529412e-05, |
|
"loss": 0.706, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 37.21, |
|
"learning_rate": 1.7255882352941176e-05, |
|
"loss": 0.823, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 37.31, |
|
"learning_rate": 1.7241176470588237e-05, |
|
"loss": 0.7442, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 37.4, |
|
"learning_rate": 1.7226470588235295e-05, |
|
"loss": 0.7708, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 37.5, |
|
"learning_rate": 1.7211764705882355e-05, |
|
"loss": 0.8368, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 37.59, |
|
"learning_rate": 1.7197058823529413e-05, |
|
"loss": 0.7684, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 37.69, |
|
"learning_rate": 1.7182352941176474e-05, |
|
"loss": 0.8044, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 37.78, |
|
"learning_rate": 1.716764705882353e-05, |
|
"loss": 0.7869, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 37.88, |
|
"learning_rate": 1.715294117647059e-05, |
|
"loss": 0.7336, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 37.98, |
|
"learning_rate": 1.713823529411765e-05, |
|
"loss": 0.8777, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 38.08, |
|
"learning_rate": 1.7123529411764707e-05, |
|
"loss": 0.7414, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 38.17, |
|
"learning_rate": 1.7108823529411768e-05, |
|
"loss": 0.7746, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 38.27, |
|
"learning_rate": 1.7094117647058825e-05, |
|
"loss": 0.784, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 38.36, |
|
"learning_rate": 1.7079411764705883e-05, |
|
"loss": 0.7399, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 38.46, |
|
"learning_rate": 1.7064705882352944e-05, |
|
"loss": 0.8018, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 38.46, |
|
"eval_loss": 0.36778753995895386, |
|
"eval_runtime": 63.9779, |
|
"eval_samples_per_second": 17.381, |
|
"eval_steps_per_second": 4.345, |
|
"eval_wer": 0.467085201793722, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 38.56, |
|
"learning_rate": 1.705e-05, |
|
"loss": 0.6784, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 38.65, |
|
"learning_rate": 1.7035294117647062e-05, |
|
"loss": 0.7498, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 38.75, |
|
"learning_rate": 1.702058823529412e-05, |
|
"loss": 0.8099, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 38.84, |
|
"learning_rate": 1.700588235294118e-05, |
|
"loss": 0.6993, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 38.94, |
|
"learning_rate": 1.6991176470588238e-05, |
|
"loss": 0.8063, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 39.04, |
|
"learning_rate": 1.6976470588235295e-05, |
|
"loss": 0.8012, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 39.13, |
|
"learning_rate": 1.6961764705882356e-05, |
|
"loss": 0.7462, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 39.23, |
|
"learning_rate": 1.6947058823529414e-05, |
|
"loss": 0.7782, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 39.33, |
|
"learning_rate": 1.693235294117647e-05, |
|
"loss": 0.7293, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 39.42, |
|
"learning_rate": 1.691764705882353e-05, |
|
"loss": 0.7589, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 39.52, |
|
"learning_rate": 1.690294117647059e-05, |
|
"loss": 0.7356, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 39.61, |
|
"learning_rate": 1.6888235294117647e-05, |
|
"loss": 0.6993, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 39.71, |
|
"learning_rate": 1.6873529411764708e-05, |
|
"loss": 0.8036, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 39.8, |
|
"learning_rate": 1.6858823529411765e-05, |
|
"loss": 0.7366, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 39.9, |
|
"learning_rate": 1.6844117647058823e-05, |
|
"loss": 0.7946, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"learning_rate": 1.6829411764705883e-05, |
|
"loss": 0.8692, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 40.1, |
|
"learning_rate": 1.681470588235294e-05, |
|
"loss": 0.7658, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 40.19, |
|
"learning_rate": 1.6800000000000002e-05, |
|
"loss": 0.841, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 40.29, |
|
"learning_rate": 1.678529411764706e-05, |
|
"loss": 0.7532, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 40.38, |
|
"learning_rate": 1.677058823529412e-05, |
|
"loss": 0.7097, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 40.48, |
|
"learning_rate": 1.6755882352941178e-05, |
|
"loss": 0.7312, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 40.57, |
|
"learning_rate": 1.6741176470588235e-05, |
|
"loss": 0.7157, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 40.67, |
|
"learning_rate": 1.6726470588235296e-05, |
|
"loss": 0.7356, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 40.77, |
|
"learning_rate": 1.6711764705882353e-05, |
|
"loss": 0.7408, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 40.86, |
|
"learning_rate": 1.6697058823529414e-05, |
|
"loss": 0.699, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 40.96, |
|
"learning_rate": 1.6682352941176472e-05, |
|
"loss": 0.7846, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 41.06, |
|
"learning_rate": 1.666764705882353e-05, |
|
"loss": 0.7321, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 41.15, |
|
"learning_rate": 1.665294117647059e-05, |
|
"loss": 0.7253, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 41.25, |
|
"learning_rate": 1.6638235294117647e-05, |
|
"loss": 0.7736, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 41.34, |
|
"learning_rate": 1.662352941176471e-05, |
|
"loss": 0.7096, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 41.44, |
|
"learning_rate": 1.6608823529411766e-05, |
|
"loss": 0.7694, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 41.54, |
|
"learning_rate": 1.6594117647058827e-05, |
|
"loss": 0.7336, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 41.63, |
|
"learning_rate": 1.6579411764705884e-05, |
|
"loss": 0.7723, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 41.73, |
|
"learning_rate": 1.656470588235294e-05, |
|
"loss": 0.821, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 41.82, |
|
"learning_rate": 1.6550000000000002e-05, |
|
"loss": 0.6847, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 41.92, |
|
"learning_rate": 1.653529411764706e-05, |
|
"loss": 0.7324, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 42.02, |
|
"learning_rate": 1.652058823529412e-05, |
|
"loss": 0.7645, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 42.11, |
|
"learning_rate": 1.6505882352941178e-05, |
|
"loss": 0.7115, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 42.21, |
|
"learning_rate": 1.6491176470588236e-05, |
|
"loss": 0.7497, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 42.31, |
|
"learning_rate": 1.6476470588235297e-05, |
|
"loss": 0.7144, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 42.4, |
|
"learning_rate": 1.6461764705882354e-05, |
|
"loss": 0.7473, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 42.5, |
|
"learning_rate": 1.6447058823529415e-05, |
|
"loss": 0.7778, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 42.59, |
|
"learning_rate": 1.6432352941176472e-05, |
|
"loss": 0.6424, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 42.69, |
|
"learning_rate": 1.641764705882353e-05, |
|
"loss": 0.7581, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 42.78, |
|
"learning_rate": 1.640294117647059e-05, |
|
"loss": 0.7509, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 42.88, |
|
"learning_rate": 1.6388235294117648e-05, |
|
"loss": 0.694, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 42.98, |
|
"learning_rate": 1.637352941176471e-05, |
|
"loss": 0.7997, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 43.08, |
|
"learning_rate": 1.6358823529411767e-05, |
|
"loss": 0.6734, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 43.17, |
|
"learning_rate": 1.6344117647058827e-05, |
|
"loss": 0.7276, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 43.27, |
|
"learning_rate": 1.6329411764705885e-05, |
|
"loss": 0.7545, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 43.27, |
|
"eval_loss": 0.3473685681819916, |
|
"eval_runtime": 63.0664, |
|
"eval_samples_per_second": 17.632, |
|
"eval_steps_per_second": 4.408, |
|
"eval_wer": 0.39623318385650225, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 43.36, |
|
"learning_rate": 1.6314705882352942e-05, |
|
"loss": 0.7045, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 43.46, |
|
"learning_rate": 1.63e-05, |
|
"loss": 0.709, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 43.56, |
|
"learning_rate": 1.628529411764706e-05, |
|
"loss": 0.6734, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 43.65, |
|
"learning_rate": 1.6270588235294118e-05, |
|
"loss": 0.7084, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 43.75, |
|
"learning_rate": 1.6255882352941176e-05, |
|
"loss": 0.7622, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 43.84, |
|
"learning_rate": 1.6241176470588236e-05, |
|
"loss": 0.68, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 43.94, |
|
"learning_rate": 1.6226470588235294e-05, |
|
"loss": 0.7237, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 44.04, |
|
"learning_rate": 1.6211764705882355e-05, |
|
"loss": 0.7907, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 44.13, |
|
"learning_rate": 1.6197058823529412e-05, |
|
"loss": 0.6809, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 44.23, |
|
"learning_rate": 1.618235294117647e-05, |
|
"loss": 0.7779, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 44.33, |
|
"learning_rate": 1.616764705882353e-05, |
|
"loss": 0.6766, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 44.42, |
|
"learning_rate": 1.6152941176470588e-05, |
|
"loss": 0.7002, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 44.52, |
|
"learning_rate": 1.613823529411765e-05, |
|
"loss": 0.7409, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 44.61, |
|
"learning_rate": 1.6123529411764706e-05, |
|
"loss": 0.696, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 44.71, |
|
"learning_rate": 1.6108823529411767e-05, |
|
"loss": 0.7951, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 44.8, |
|
"learning_rate": 1.6094117647058825e-05, |
|
"loss": 0.6519, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 44.9, |
|
"learning_rate": 1.6079411764705882e-05, |
|
"loss": 0.7069, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"learning_rate": 1.6064705882352943e-05, |
|
"loss": 0.7394, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 45.1, |
|
"learning_rate": 1.605e-05, |
|
"loss": 0.6931, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 45.19, |
|
"learning_rate": 1.603529411764706e-05, |
|
"loss": 0.7234, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 45.29, |
|
"learning_rate": 1.602058823529412e-05, |
|
"loss": 0.7074, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 45.38, |
|
"learning_rate": 1.6005882352941176e-05, |
|
"loss": 0.6898, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 45.48, |
|
"learning_rate": 1.5991176470588237e-05, |
|
"loss": 0.7865, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 45.57, |
|
"learning_rate": 1.5976470588235295e-05, |
|
"loss": 0.6695, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 45.67, |
|
"learning_rate": 1.5961764705882355e-05, |
|
"loss": 0.7137, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 45.77, |
|
"learning_rate": 1.5947058823529413e-05, |
|
"loss": 0.7248, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 45.86, |
|
"learning_rate": 1.5932352941176474e-05, |
|
"loss": 0.6617, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 45.96, |
|
"learning_rate": 1.591764705882353e-05, |
|
"loss": 0.7628, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 46.06, |
|
"learning_rate": 1.590294117647059e-05, |
|
"loss": 0.7302, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 46.15, |
|
"learning_rate": 1.588823529411765e-05, |
|
"loss": 0.6684, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 46.25, |
|
"learning_rate": 1.5873529411764707e-05, |
|
"loss": 0.7304, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 46.34, |
|
"learning_rate": 1.5858823529411768e-05, |
|
"loss": 0.635, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 46.44, |
|
"learning_rate": 1.5844117647058825e-05, |
|
"loss": 0.7353, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 46.54, |
|
"learning_rate": 1.5829411764705883e-05, |
|
"loss": 0.7193, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 46.63, |
|
"learning_rate": 1.5814705882352944e-05, |
|
"loss": 0.6673, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 46.73, |
|
"learning_rate": 1.58e-05, |
|
"loss": 0.761, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 46.82, |
|
"learning_rate": 1.5785294117647062e-05, |
|
"loss": 0.6406, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 46.92, |
|
"learning_rate": 1.577058823529412e-05, |
|
"loss": 0.686, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 47.02, |
|
"learning_rate": 1.5755882352941177e-05, |
|
"loss": 0.7237, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 47.11, |
|
"learning_rate": 1.5741176470588238e-05, |
|
"loss": 0.6284, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 47.21, |
|
"learning_rate": 1.5726470588235295e-05, |
|
"loss": 0.7495, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 47.31, |
|
"learning_rate": 1.5711764705882356e-05, |
|
"loss": 0.673, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 47.4, |
|
"learning_rate": 1.5697058823529414e-05, |
|
"loss": 0.6968, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 47.5, |
|
"learning_rate": 1.568235294117647e-05, |
|
"loss": 0.7054, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 47.59, |
|
"learning_rate": 1.566764705882353e-05, |
|
"loss": 0.6393, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 47.69, |
|
"learning_rate": 1.565294117647059e-05, |
|
"loss": 0.7054, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 47.78, |
|
"learning_rate": 1.5638235294117647e-05, |
|
"loss": 0.6707, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 47.88, |
|
"learning_rate": 1.5623529411764708e-05, |
|
"loss": 0.683, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 47.98, |
|
"learning_rate": 1.5608823529411765e-05, |
|
"loss": 0.7264, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 48.08, |
|
"learning_rate": 1.5594117647058823e-05, |
|
"loss": 0.7375, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 48.08, |
|
"eval_loss": 0.3224140405654907, |
|
"eval_runtime": 63.7123, |
|
"eval_samples_per_second": 17.453, |
|
"eval_steps_per_second": 4.363, |
|
"eval_wer": 0.38690582959641256, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 48.17, |
|
"learning_rate": 1.5579411764705884e-05, |
|
"loss": 0.6903, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 48.27, |
|
"learning_rate": 1.556470588235294e-05, |
|
"loss": 0.7025, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 48.36, |
|
"learning_rate": 1.5550000000000002e-05, |
|
"loss": 0.6537, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 48.46, |
|
"learning_rate": 1.553529411764706e-05, |
|
"loss": 0.7093, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 48.56, |
|
"learning_rate": 1.5520588235294117e-05, |
|
"loss": 0.6356, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 48.65, |
|
"learning_rate": 1.5505882352941178e-05, |
|
"loss": 0.6667, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 48.75, |
|
"learning_rate": 1.5491176470588235e-05, |
|
"loss": 0.7336, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 48.84, |
|
"learning_rate": 1.5476470588235296e-05, |
|
"loss": 0.6597, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 48.94, |
|
"learning_rate": 1.5461764705882353e-05, |
|
"loss": 0.6578, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 49.04, |
|
"learning_rate": 1.5447058823529414e-05, |
|
"loss": 0.6809, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 49.13, |
|
"learning_rate": 1.5432352941176472e-05, |
|
"loss": 0.7041, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 49.23, |
|
"learning_rate": 1.541764705882353e-05, |
|
"loss": 0.7167, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 49.33, |
|
"learning_rate": 1.540294117647059e-05, |
|
"loss": 0.612, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 49.42, |
|
"learning_rate": 1.5388235294117648e-05, |
|
"loss": 0.7009, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 49.52, |
|
"learning_rate": 1.537352941176471e-05, |
|
"loss": 0.7308, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 49.61, |
|
"learning_rate": 1.5358823529411766e-05, |
|
"loss": 0.6436, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 49.71, |
|
"learning_rate": 1.5344117647058823e-05, |
|
"loss": 0.6969, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 49.8, |
|
"learning_rate": 1.5329411764705884e-05, |
|
"loss": 0.6241, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 49.9, |
|
"learning_rate": 1.5314705882352942e-05, |
|
"loss": 0.6969, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"learning_rate": 1.5300000000000003e-05, |
|
"loss": 0.7205, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 50.1, |
|
"learning_rate": 1.528529411764706e-05, |
|
"loss": 0.6744, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 50.19, |
|
"learning_rate": 1.527058823529412e-05, |
|
"loss": 0.665, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 50.29, |
|
"learning_rate": 1.5255882352941178e-05, |
|
"loss": 0.6083, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 50.38, |
|
"learning_rate": 1.5241176470588238e-05, |
|
"loss": 0.6328, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 50.48, |
|
"learning_rate": 1.5226470588235297e-05, |
|
"loss": 0.6876, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 50.57, |
|
"learning_rate": 1.5211764705882354e-05, |
|
"loss": 0.6444, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 50.67, |
|
"learning_rate": 1.5197058823529413e-05, |
|
"loss": 0.6999, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 50.77, |
|
"learning_rate": 1.5182352941176473e-05, |
|
"loss": 0.7376, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 50.86, |
|
"learning_rate": 1.5167647058823532e-05, |
|
"loss": 0.6226, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 50.96, |
|
"learning_rate": 1.515294117647059e-05, |
|
"loss": 0.6692, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 51.06, |
|
"learning_rate": 1.5138235294117648e-05, |
|
"loss": 0.6899, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 51.15, |
|
"learning_rate": 1.5123529411764707e-05, |
|
"loss": 0.6441, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 51.25, |
|
"learning_rate": 1.5108823529411767e-05, |
|
"loss": 0.6611, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 51.34, |
|
"learning_rate": 1.5094117647058826e-05, |
|
"loss": 0.5911, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 51.44, |
|
"learning_rate": 1.5079411764705885e-05, |
|
"loss": 0.6594, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 51.54, |
|
"learning_rate": 1.5064705882352944e-05, |
|
"loss": 0.6524, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 51.63, |
|
"learning_rate": 1.505e-05, |
|
"loss": 0.6326, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 51.73, |
|
"learning_rate": 1.5035294117647059e-05, |
|
"loss": 0.6649, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 51.82, |
|
"learning_rate": 1.5020588235294118e-05, |
|
"loss": 0.6304, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 51.92, |
|
"learning_rate": 1.5005882352941177e-05, |
|
"loss": 0.6929, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 52.02, |
|
"learning_rate": 1.4991176470588235e-05, |
|
"loss": 0.7116, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 52.11, |
|
"learning_rate": 1.4976470588235294e-05, |
|
"loss": 0.595, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 52.21, |
|
"learning_rate": 1.4961764705882353e-05, |
|
"loss": 0.6809, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 52.31, |
|
"learning_rate": 1.4947058823529412e-05, |
|
"loss": 0.6222, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 52.4, |
|
"learning_rate": 1.4932352941176472e-05, |
|
"loss": 0.6346, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 52.5, |
|
"learning_rate": 1.491764705882353e-05, |
|
"loss": 0.6701, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 52.59, |
|
"learning_rate": 1.4902941176470588e-05, |
|
"loss": 0.6263, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 52.69, |
|
"learning_rate": 1.4888235294117647e-05, |
|
"loss": 0.7159, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 52.78, |
|
"learning_rate": 1.4873529411764706e-05, |
|
"loss": 0.6833, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 52.88, |
|
"learning_rate": 1.4858823529411766e-05, |
|
"loss": 0.6198, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 52.88, |
|
"eval_loss": 0.32332631945610046, |
|
"eval_runtime": 63.48, |
|
"eval_samples_per_second": 17.517, |
|
"eval_steps_per_second": 4.379, |
|
"eval_wer": 0.3630493273542601, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 52.98, |
|
"learning_rate": 1.4844117647058825e-05, |
|
"loss": 0.7255, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 53.08, |
|
"learning_rate": 1.4829411764705884e-05, |
|
"loss": 0.6043, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 53.17, |
|
"learning_rate": 1.4814705882352941e-05, |
|
"loss": 0.5946, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 53.27, |
|
"learning_rate": 1.48e-05, |
|
"loss": 0.6643, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 53.36, |
|
"learning_rate": 1.478529411764706e-05, |
|
"loss": 0.5974, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 53.46, |
|
"learning_rate": 1.4770588235294119e-05, |
|
"loss": 0.655, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 53.56, |
|
"learning_rate": 1.4755882352941178e-05, |
|
"loss": 0.5887, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 53.65, |
|
"learning_rate": 1.4741176470588237e-05, |
|
"loss": 0.6272, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 53.75, |
|
"learning_rate": 1.4726470588235295e-05, |
|
"loss": 0.6723, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 53.84, |
|
"learning_rate": 1.4711764705882354e-05, |
|
"loss": 0.6192, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 53.94, |
|
"learning_rate": 1.4697058823529413e-05, |
|
"loss": 0.6742, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 54.04, |
|
"learning_rate": 1.4682352941176472e-05, |
|
"loss": 0.6684, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 54.13, |
|
"learning_rate": 1.4667647058823531e-05, |
|
"loss": 0.5831, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 54.23, |
|
"learning_rate": 1.465294117647059e-05, |
|
"loss": 0.6729, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 54.33, |
|
"learning_rate": 1.4638235294117648e-05, |
|
"loss": 0.5947, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 54.42, |
|
"learning_rate": 1.4623529411764707e-05, |
|
"loss": 0.6604, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 54.52, |
|
"learning_rate": 1.4608823529411766e-05, |
|
"loss": 0.6306, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 54.61, |
|
"learning_rate": 1.4594117647058826e-05, |
|
"loss": 0.6495, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 54.71, |
|
"learning_rate": 1.4579411764705885e-05, |
|
"loss": 0.7005, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 54.8, |
|
"learning_rate": 1.4564705882352944e-05, |
|
"loss": 0.6115, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 54.9, |
|
"learning_rate": 1.4550000000000001e-05, |
|
"loss": 0.6282, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 55.0, |
|
"learning_rate": 1.453529411764706e-05, |
|
"loss": 0.6748, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 55.1, |
|
"learning_rate": 1.452058823529412e-05, |
|
"loss": 0.6621, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 55.19, |
|
"learning_rate": 1.4505882352941179e-05, |
|
"loss": 0.6849, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 55.29, |
|
"learning_rate": 1.4491176470588238e-05, |
|
"loss": 0.5829, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 55.38, |
|
"learning_rate": 1.4476470588235297e-05, |
|
"loss": 0.6349, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 55.48, |
|
"learning_rate": 1.4461764705882355e-05, |
|
"loss": 0.6789, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 55.57, |
|
"learning_rate": 1.4447058823529414e-05, |
|
"loss": 0.6053, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 55.67, |
|
"learning_rate": 1.4432352941176471e-05, |
|
"loss": 0.6642, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 55.77, |
|
"learning_rate": 1.441764705882353e-05, |
|
"loss": 0.6564, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 55.86, |
|
"learning_rate": 1.4402941176470588e-05, |
|
"loss": 0.609, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 55.96, |
|
"learning_rate": 1.4388235294117647e-05, |
|
"loss": 0.6675, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 56.06, |
|
"learning_rate": 1.4373529411764706e-05, |
|
"loss": 0.6678, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 56.15, |
|
"learning_rate": 1.4358823529411765e-05, |
|
"loss": 0.6351, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 56.25, |
|
"learning_rate": 1.4344117647058825e-05, |
|
"loss": 0.6427, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 56.34, |
|
"learning_rate": 1.4329411764705882e-05, |
|
"loss": 0.5832, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 56.44, |
|
"learning_rate": 1.4314705882352941e-05, |
|
"loss": 0.6484, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 56.54, |
|
"learning_rate": 1.43e-05, |
|
"loss": 0.6515, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 56.63, |
|
"learning_rate": 1.428529411764706e-05, |
|
"loss": 0.6253, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 56.73, |
|
"learning_rate": 1.4270588235294119e-05, |
|
"loss": 0.6931, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 56.82, |
|
"learning_rate": 1.4255882352941178e-05, |
|
"loss": 0.6208, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 56.92, |
|
"learning_rate": 1.4241176470588235e-05, |
|
"loss": 0.6495, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 57.02, |
|
"learning_rate": 1.4226470588235294e-05, |
|
"loss": 0.6669, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 57.11, |
|
"learning_rate": 1.4211764705882354e-05, |
|
"loss": 0.5983, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 57.21, |
|
"learning_rate": 1.4197058823529413e-05, |
|
"loss": 0.6532, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 57.31, |
|
"learning_rate": 1.4182352941176472e-05, |
|
"loss": 0.5822, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 57.4, |
|
"learning_rate": 1.4167647058823531e-05, |
|
"loss": 0.6257, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 57.5, |
|
"learning_rate": 1.4152941176470589e-05, |
|
"loss": 0.658, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 57.59, |
|
"learning_rate": 1.4138235294117648e-05, |
|
"loss": 0.6191, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 57.69, |
|
"learning_rate": 1.4123529411764707e-05, |
|
"loss": 0.6608, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 57.69, |
|
"eval_loss": 0.30293992161750793, |
|
"eval_runtime": 65.7525, |
|
"eval_samples_per_second": 16.912, |
|
"eval_steps_per_second": 4.228, |
|
"eval_wer": 0.330762331838565, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 57.78, |
|
"learning_rate": 1.4108823529411766e-05, |
|
"loss": 0.6108, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 57.88, |
|
"learning_rate": 1.4094117647058825e-05, |
|
"loss": 0.6054, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 57.98, |
|
"learning_rate": 1.4079411764705884e-05, |
|
"loss": 0.6822, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 58.08, |
|
"learning_rate": 1.4064705882352942e-05, |
|
"loss": 0.5838, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 58.17, |
|
"learning_rate": 1.4050000000000001e-05, |
|
"loss": 0.6325, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 58.27, |
|
"learning_rate": 1.403529411764706e-05, |
|
"loss": 0.5843, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 58.36, |
|
"learning_rate": 1.402058823529412e-05, |
|
"loss": 0.562, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 58.46, |
|
"learning_rate": 1.4005882352941178e-05, |
|
"loss": 0.6412, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 58.56, |
|
"learning_rate": 1.3991176470588238e-05, |
|
"loss": 0.6815, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 58.65, |
|
"learning_rate": 1.3976470588235295e-05, |
|
"loss": 0.5983, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 58.75, |
|
"learning_rate": 1.3961764705882354e-05, |
|
"loss": 0.6361, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 58.84, |
|
"learning_rate": 1.3947058823529413e-05, |
|
"loss": 0.5488, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 58.94, |
|
"learning_rate": 1.3932352941176473e-05, |
|
"loss": 0.6384, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 59.04, |
|
"learning_rate": 1.3917647058823532e-05, |
|
"loss": 0.6549, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 59.13, |
|
"learning_rate": 1.3902941176470591e-05, |
|
"loss": 0.5964, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 59.23, |
|
"learning_rate": 1.3888235294117648e-05, |
|
"loss": 0.6409, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 59.33, |
|
"learning_rate": 1.3873529411764708e-05, |
|
"loss": 0.5982, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 59.42, |
|
"learning_rate": 1.3858823529411767e-05, |
|
"loss": 0.608, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 59.52, |
|
"learning_rate": 1.3844117647058826e-05, |
|
"loss": 0.6128, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 59.61, |
|
"learning_rate": 1.3829411764705885e-05, |
|
"loss": 0.583, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 59.71, |
|
"learning_rate": 1.3814705882352944e-05, |
|
"loss": 0.6837, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 59.8, |
|
"learning_rate": 1.38e-05, |
|
"loss": 0.5842, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 59.9, |
|
"learning_rate": 1.378529411764706e-05, |
|
"loss": 0.6115, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 60.0, |
|
"learning_rate": 1.3770588235294118e-05, |
|
"loss": 0.6262, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 60.1, |
|
"learning_rate": 1.3755882352941178e-05, |
|
"loss": 0.5743, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 60.19, |
|
"learning_rate": 1.3741176470588235e-05, |
|
"loss": 0.6657, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 60.29, |
|
"learning_rate": 1.3726470588235294e-05, |
|
"loss": 0.5917, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 60.38, |
|
"learning_rate": 1.3711764705882353e-05, |
|
"loss": 0.6081, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 60.48, |
|
"learning_rate": 1.3697058823529412e-05, |
|
"loss": 0.6253, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 60.57, |
|
"learning_rate": 1.3682352941176472e-05, |
|
"loss": 0.5736, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 60.67, |
|
"learning_rate": 1.3667647058823529e-05, |
|
"loss": 0.5835, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 60.77, |
|
"learning_rate": 1.3652941176470588e-05, |
|
"loss": 0.5963, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 60.86, |
|
"learning_rate": 1.3638235294117647e-05, |
|
"loss": 0.582, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 60.96, |
|
"learning_rate": 1.3623529411764707e-05, |
|
"loss": 0.631, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 61.06, |
|
"learning_rate": 1.3608823529411766e-05, |
|
"loss": 0.6189, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 61.15, |
|
"learning_rate": 1.3594117647058825e-05, |
|
"loss": 0.58, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 61.25, |
|
"learning_rate": 1.3579411764705882e-05, |
|
"loss": 0.6145, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 61.34, |
|
"learning_rate": 1.3564705882352942e-05, |
|
"loss": 0.5785, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 61.44, |
|
"learning_rate": 1.355e-05, |
|
"loss": 0.6151, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 61.54, |
|
"learning_rate": 1.353529411764706e-05, |
|
"loss": 0.5841, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 61.63, |
|
"learning_rate": 1.3520588235294119e-05, |
|
"loss": 0.576, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 61.73, |
|
"learning_rate": 1.3505882352941178e-05, |
|
"loss": 0.6248, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 61.82, |
|
"learning_rate": 1.3491176470588236e-05, |
|
"loss": 0.5985, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 61.92, |
|
"learning_rate": 1.3476470588235295e-05, |
|
"loss": 0.6258, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 62.02, |
|
"learning_rate": 1.3461764705882354e-05, |
|
"loss": 0.628, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 62.11, |
|
"learning_rate": 1.3447058823529413e-05, |
|
"loss": 0.5418, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 62.21, |
|
"learning_rate": 1.3432352941176472e-05, |
|
"loss": 0.5944, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 62.31, |
|
"learning_rate": 1.3417647058823531e-05, |
|
"loss": 0.5303, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 62.4, |
|
"learning_rate": 1.3402941176470589e-05, |
|
"loss": 0.5895, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 62.5, |
|
"learning_rate": 1.3388235294117648e-05, |
|
"loss": 0.645, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 62.5, |
|
"eval_loss": 0.31946861743927, |
|
"eval_runtime": 64.2143, |
|
"eval_samples_per_second": 17.317, |
|
"eval_steps_per_second": 4.329, |
|
"eval_wer": 0.3721973094170404, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 62.59, |
|
"learning_rate": 1.3373529411764707e-05, |
|
"loss": 0.5618, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 62.69, |
|
"learning_rate": 1.3358823529411766e-05, |
|
"loss": 0.5899, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 62.78, |
|
"learning_rate": 1.3344117647058826e-05, |
|
"loss": 0.6129, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 62.88, |
|
"learning_rate": 1.3329411764705885e-05, |
|
"loss": 0.566, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 62.98, |
|
"learning_rate": 1.3314705882352942e-05, |
|
"loss": 0.6321, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 63.08, |
|
"learning_rate": 1.3300000000000001e-05, |
|
"loss": 0.5825, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 63.17, |
|
"learning_rate": 1.328529411764706e-05, |
|
"loss": 0.5806, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 63.27, |
|
"learning_rate": 1.3272058823529412e-05, |
|
"loss": 0.5581, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 63.36, |
|
"learning_rate": 1.3257352941176471e-05, |
|
"loss": 0.5556, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 63.46, |
|
"learning_rate": 1.324264705882353e-05, |
|
"loss": 0.6232, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 63.56, |
|
"learning_rate": 1.322794117647059e-05, |
|
"loss": 0.5705, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 63.65, |
|
"learning_rate": 1.3213235294117647e-05, |
|
"loss": 0.6115, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 63.75, |
|
"learning_rate": 1.3198529411764706e-05, |
|
"loss": 0.6426, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 63.84, |
|
"learning_rate": 1.3183823529411766e-05, |
|
"loss": 0.5589, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 63.94, |
|
"learning_rate": 1.3169117647058825e-05, |
|
"loss": 0.61, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 64.04, |
|
"learning_rate": 1.3154411764705884e-05, |
|
"loss": 0.6278, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 64.13, |
|
"learning_rate": 1.3139705882352941e-05, |
|
"loss": 0.5606, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 64.23, |
|
"learning_rate": 1.3125e-05, |
|
"loss": 0.6356, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 64.33, |
|
"learning_rate": 1.311029411764706e-05, |
|
"loss": 0.5531, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 64.42, |
|
"learning_rate": 1.3095588235294119e-05, |
|
"loss": 0.6166, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 64.52, |
|
"learning_rate": 1.3080882352941178e-05, |
|
"loss": 0.6173, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 64.61, |
|
"learning_rate": 1.3066176470588237e-05, |
|
"loss": 0.5513, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 64.71, |
|
"learning_rate": 1.3051470588235295e-05, |
|
"loss": 0.6087, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 64.8, |
|
"learning_rate": 1.3036764705882354e-05, |
|
"loss": 0.6022, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 64.9, |
|
"learning_rate": 1.3022058823529413e-05, |
|
"loss": 0.5814, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 65.0, |
|
"learning_rate": 1.3007352941176472e-05, |
|
"loss": 0.5967, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 65.1, |
|
"learning_rate": 1.2992647058823531e-05, |
|
"loss": 0.5562, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 65.19, |
|
"learning_rate": 1.297794117647059e-05, |
|
"loss": 0.6277, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 65.29, |
|
"learning_rate": 1.2963235294117648e-05, |
|
"loss": 0.5655, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 65.38, |
|
"learning_rate": 1.2948529411764707e-05, |
|
"loss": 0.5458, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 65.48, |
|
"learning_rate": 1.2933823529411766e-05, |
|
"loss": 0.5877, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 65.57, |
|
"learning_rate": 1.2919117647058825e-05, |
|
"loss": 0.5352, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 65.67, |
|
"learning_rate": 1.2904411764705885e-05, |
|
"loss": 0.588, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 65.77, |
|
"learning_rate": 1.2889705882352944e-05, |
|
"loss": 0.5694, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 65.86, |
|
"learning_rate": 1.2875000000000001e-05, |
|
"loss": 0.564, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 65.96, |
|
"learning_rate": 1.286029411764706e-05, |
|
"loss": 0.5773, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 66.06, |
|
"learning_rate": 1.284558823529412e-05, |
|
"loss": 0.5792, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 66.15, |
|
"learning_rate": 1.2830882352941179e-05, |
|
"loss": 0.5583, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 66.25, |
|
"learning_rate": 1.2816176470588238e-05, |
|
"loss": 0.6152, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 66.34, |
|
"learning_rate": 1.2801470588235294e-05, |
|
"loss": 0.5508, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 66.44, |
|
"learning_rate": 1.2786764705882353e-05, |
|
"loss": 0.5799, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 66.54, |
|
"learning_rate": 1.2772058823529412e-05, |
|
"loss": 0.5303, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 66.63, |
|
"learning_rate": 1.2757352941176471e-05, |
|
"loss": 0.5169, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 66.73, |
|
"learning_rate": 1.274264705882353e-05, |
|
"loss": 0.6059, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 66.82, |
|
"learning_rate": 1.2727941176470588e-05, |
|
"loss": 0.5155, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 66.92, |
|
"learning_rate": 1.2713235294117647e-05, |
|
"loss": 0.5917, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 67.02, |
|
"learning_rate": 1.2698529411764706e-05, |
|
"loss": 0.6176, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 67.11, |
|
"learning_rate": 1.2683823529411765e-05, |
|
"loss": 0.57, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 67.21, |
|
"learning_rate": 1.2669117647058824e-05, |
|
"loss": 0.5843, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 67.31, |
|
"learning_rate": 1.2654411764705884e-05, |
|
"loss": 0.5249, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 67.31, |
|
"eval_loss": 0.3004380464553833, |
|
"eval_runtime": 64.3238, |
|
"eval_samples_per_second": 17.288, |
|
"eval_steps_per_second": 4.322, |
|
"eval_wer": 0.3201793721973094, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 67.4, |
|
"learning_rate": 1.2639705882352941e-05, |
|
"loss": 0.5574, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 67.5, |
|
"learning_rate": 1.2625e-05, |
|
"loss": 0.5894, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 67.59, |
|
"learning_rate": 1.261029411764706e-05, |
|
"loss": 0.5346, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 67.69, |
|
"learning_rate": 1.2595588235294119e-05, |
|
"loss": 0.5894, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 67.78, |
|
"learning_rate": 1.2580882352941178e-05, |
|
"loss": 0.5812, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 67.88, |
|
"learning_rate": 1.2566176470588237e-05, |
|
"loss": 0.5508, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 67.98, |
|
"learning_rate": 1.2551470588235294e-05, |
|
"loss": 0.5751, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 68.08, |
|
"learning_rate": 1.2536764705882354e-05, |
|
"loss": 0.5374, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 68.17, |
|
"learning_rate": 1.2522058823529413e-05, |
|
"loss": 0.5674, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 68.27, |
|
"learning_rate": 1.2507352941176472e-05, |
|
"loss": 0.5831, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 68.36, |
|
"learning_rate": 1.2492647058823531e-05, |
|
"loss": 0.4986, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 68.46, |
|
"learning_rate": 1.2477941176470588e-05, |
|
"loss": 0.6136, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 68.56, |
|
"learning_rate": 1.2463235294117648e-05, |
|
"loss": 0.584, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 68.65, |
|
"learning_rate": 1.2448529411764707e-05, |
|
"loss": 0.5978, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 68.75, |
|
"learning_rate": 1.2433823529411766e-05, |
|
"loss": 0.5997, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 68.84, |
|
"learning_rate": 1.2419117647058825e-05, |
|
"loss": 0.5289, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 68.94, |
|
"learning_rate": 1.2404411764705884e-05, |
|
"loss": 0.5714, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 69.04, |
|
"learning_rate": 1.2389705882352942e-05, |
|
"loss": 0.5923, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 69.13, |
|
"learning_rate": 1.2375000000000001e-05, |
|
"loss": 0.5172, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 69.23, |
|
"learning_rate": 1.236029411764706e-05, |
|
"loss": 0.553, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 69.33, |
|
"learning_rate": 1.234558823529412e-05, |
|
"loss": 0.5201, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 69.42, |
|
"learning_rate": 1.2330882352941178e-05, |
|
"loss": 0.5518, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 69.52, |
|
"learning_rate": 1.2316176470588238e-05, |
|
"loss": 0.5576, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 69.61, |
|
"learning_rate": 1.2301470588235295e-05, |
|
"loss": 0.5296, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 69.71, |
|
"learning_rate": 1.2286764705882354e-05, |
|
"loss": 0.563, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 69.8, |
|
"learning_rate": 1.2272058823529413e-05, |
|
"loss": 0.5248, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 69.9, |
|
"learning_rate": 1.2257352941176473e-05, |
|
"loss": 0.5351, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 70.0, |
|
"learning_rate": 1.2242647058823532e-05, |
|
"loss": 0.5778, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 70.1, |
|
"learning_rate": 1.2227941176470591e-05, |
|
"loss": 0.5017, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 70.19, |
|
"learning_rate": 1.2213235294117648e-05, |
|
"loss": 0.5691, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 70.29, |
|
"learning_rate": 1.2198529411764708e-05, |
|
"loss": 0.5461, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 70.38, |
|
"learning_rate": 1.2183823529411765e-05, |
|
"loss": 0.5056, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 70.48, |
|
"learning_rate": 1.2169117647058824e-05, |
|
"loss": 0.6269, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 70.57, |
|
"learning_rate": 1.2154411764705882e-05, |
|
"loss": 0.5062, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 70.67, |
|
"learning_rate": 1.213970588235294e-05, |
|
"loss": 0.5445, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 70.77, |
|
"learning_rate": 1.2126470588235296e-05, |
|
"loss": 0.5685, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 70.86, |
|
"learning_rate": 1.2111764705882353e-05, |
|
"loss": 0.495, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 70.96, |
|
"learning_rate": 1.2097058823529413e-05, |
|
"loss": 0.5655, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 71.06, |
|
"learning_rate": 1.2082352941176472e-05, |
|
"loss": 0.6005, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 71.15, |
|
"learning_rate": 1.2067647058823531e-05, |
|
"loss": 0.5434, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 71.25, |
|
"learning_rate": 1.205294117647059e-05, |
|
"loss": 0.583, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 71.34, |
|
"learning_rate": 1.203823529411765e-05, |
|
"loss": 0.4755, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 71.44, |
|
"learning_rate": 1.2023529411764707e-05, |
|
"loss": 0.5464, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 71.54, |
|
"learning_rate": 1.2008823529411766e-05, |
|
"loss": 0.5436, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 71.63, |
|
"learning_rate": 1.1994117647058825e-05, |
|
"loss": 0.5371, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 71.73, |
|
"learning_rate": 1.1979411764705884e-05, |
|
"loss": 0.607, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 71.82, |
|
"learning_rate": 1.1964705882352943e-05, |
|
"loss": 0.5427, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 71.92, |
|
"learning_rate": 1.195e-05, |
|
"loss": 0.5756, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 72.02, |
|
"learning_rate": 1.193529411764706e-05, |
|
"loss": 0.5606, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 72.11, |
|
"learning_rate": 1.1920588235294119e-05, |
|
"loss": 0.4875, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 72.11, |
|
"eval_loss": 0.2825987637042999, |
|
"eval_runtime": 65.1196, |
|
"eval_samples_per_second": 17.076, |
|
"eval_steps_per_second": 4.269, |
|
"eval_wer": 0.2991928251121076, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 72.21, |
|
"learning_rate": 1.1905882352941178e-05, |
|
"loss": 0.5382, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 72.31, |
|
"learning_rate": 1.1891176470588237e-05, |
|
"loss": 0.4921, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 72.4, |
|
"learning_rate": 1.1876470588235297e-05, |
|
"loss": 0.5129, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 72.5, |
|
"learning_rate": 1.1861764705882354e-05, |
|
"loss": 0.5687, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 72.59, |
|
"learning_rate": 1.1847058823529413e-05, |
|
"loss": 0.4781, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 72.69, |
|
"learning_rate": 1.1832352941176472e-05, |
|
"loss": 0.5788, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 72.78, |
|
"learning_rate": 1.1817647058823532e-05, |
|
"loss": 0.5007, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 72.88, |
|
"learning_rate": 1.1802941176470587e-05, |
|
"loss": 0.5558, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 72.98, |
|
"learning_rate": 1.1788235294117647e-05, |
|
"loss": 0.5534, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 73.08, |
|
"learning_rate": 1.1773529411764706e-05, |
|
"loss": 0.5321, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 73.17, |
|
"learning_rate": 1.1758823529411765e-05, |
|
"loss": 0.5117, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 73.27, |
|
"learning_rate": 1.1744117647058824e-05, |
|
"loss": 0.5135, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 73.36, |
|
"learning_rate": 1.1729411764705883e-05, |
|
"loss": 0.5021, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 73.46, |
|
"learning_rate": 1.171470588235294e-05, |
|
"loss": 0.5378, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 73.56, |
|
"learning_rate": 1.17e-05, |
|
"loss": 0.5497, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 73.65, |
|
"learning_rate": 1.1685294117647059e-05, |
|
"loss": 0.6035, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 73.75, |
|
"learning_rate": 1.1670588235294118e-05, |
|
"loss": 0.6202, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 73.84, |
|
"learning_rate": 1.1655882352941177e-05, |
|
"loss": 0.5067, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 73.94, |
|
"learning_rate": 1.1641176470588236e-05, |
|
"loss": 0.564, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 74.04, |
|
"learning_rate": 1.1626470588235294e-05, |
|
"loss": 0.603, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 74.13, |
|
"learning_rate": 1.1611764705882353e-05, |
|
"loss": 0.5124, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 74.23, |
|
"learning_rate": 1.1597058823529412e-05, |
|
"loss": 0.6176, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 74.33, |
|
"learning_rate": 1.1582352941176471e-05, |
|
"loss": 0.4979, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 74.42, |
|
"learning_rate": 1.156764705882353e-05, |
|
"loss": 0.5213, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 74.52, |
|
"learning_rate": 1.155294117647059e-05, |
|
"loss": 0.5668, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 74.61, |
|
"learning_rate": 1.1538235294117647e-05, |
|
"loss": 0.5023, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 74.71, |
|
"learning_rate": 1.1523529411764706e-05, |
|
"loss": 0.5328, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 74.8, |
|
"learning_rate": 1.1508823529411766e-05, |
|
"loss": 0.504, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 74.9, |
|
"learning_rate": 1.1494117647058825e-05, |
|
"loss": 0.5274, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 75.0, |
|
"learning_rate": 1.1479411764705884e-05, |
|
"loss": 0.5422, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 75.1, |
|
"learning_rate": 1.1464705882352943e-05, |
|
"loss": 0.4952, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 75.19, |
|
"learning_rate": 1.145e-05, |
|
"loss": 0.5497, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 75.29, |
|
"learning_rate": 1.143529411764706e-05, |
|
"loss": 0.493, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 75.38, |
|
"learning_rate": 1.1420588235294119e-05, |
|
"loss": 0.5036, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 75.48, |
|
"learning_rate": 1.1405882352941178e-05, |
|
"loss": 0.5296, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 75.57, |
|
"learning_rate": 1.1391176470588237e-05, |
|
"loss": 0.4721, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 75.67, |
|
"learning_rate": 1.1376470588235296e-05, |
|
"loss": 0.5258, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 75.77, |
|
"learning_rate": 1.1361764705882354e-05, |
|
"loss": 0.5473, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 75.86, |
|
"learning_rate": 1.1347058823529413e-05, |
|
"loss": 0.4668, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 75.96, |
|
"learning_rate": 1.1332352941176472e-05, |
|
"loss": 0.5451, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 76.06, |
|
"learning_rate": 1.1317647058823531e-05, |
|
"loss": 0.4858, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 76.15, |
|
"learning_rate": 1.130294117647059e-05, |
|
"loss": 0.5053, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 76.25, |
|
"learning_rate": 1.1288235294117648e-05, |
|
"loss": 0.54, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 76.34, |
|
"learning_rate": 1.1273529411764707e-05, |
|
"loss": 0.4802, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 76.44, |
|
"learning_rate": 1.1258823529411766e-05, |
|
"loss": 0.544, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 76.54, |
|
"learning_rate": 1.1244117647058825e-05, |
|
"loss": 0.5039, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 76.63, |
|
"learning_rate": 1.1229411764705885e-05, |
|
"loss": 0.475, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 76.73, |
|
"learning_rate": 1.1214705882352944e-05, |
|
"loss": 0.5507, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 76.82, |
|
"learning_rate": 1.1200000000000001e-05, |
|
"loss": 0.492, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 76.92, |
|
"learning_rate": 1.1185294117647059e-05, |
|
"loss": 0.5171, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 76.92, |
|
"eval_loss": 0.296180784702301, |
|
"eval_runtime": 65.0333, |
|
"eval_samples_per_second": 17.099, |
|
"eval_steps_per_second": 4.275, |
|
"eval_wer": 0.29757847533632287, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 77.02, |
|
"learning_rate": 1.1170588235294118e-05, |
|
"loss": 0.5747, |
|
"step": 8010 |
|
}, |
|
{ |
|
"epoch": 77.11, |
|
"learning_rate": 1.1155882352941177e-05, |
|
"loss": 0.4866, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 77.21, |
|
"learning_rate": 1.1141176470588234e-05, |
|
"loss": 0.5232, |
|
"step": 8030 |
|
}, |
|
{ |
|
"epoch": 77.31, |
|
"learning_rate": 1.1126470588235294e-05, |
|
"loss": 0.4883, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 77.4, |
|
"learning_rate": 1.1111764705882353e-05, |
|
"loss": 0.5459, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 77.5, |
|
"learning_rate": 1.1097058823529412e-05, |
|
"loss": 0.5116, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 77.59, |
|
"learning_rate": 1.1082352941176471e-05, |
|
"loss": 0.4954, |
|
"step": 8070 |
|
}, |
|
{ |
|
"epoch": 77.69, |
|
"learning_rate": 1.106764705882353e-05, |
|
"loss": 0.4944, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 77.78, |
|
"learning_rate": 1.1052941176470588e-05, |
|
"loss": 0.5292, |
|
"step": 8090 |
|
}, |
|
{ |
|
"epoch": 77.88, |
|
"learning_rate": 1.1038235294117647e-05, |
|
"loss": 0.4984, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 77.98, |
|
"learning_rate": 1.1023529411764706e-05, |
|
"loss": 0.5374, |
|
"step": 8110 |
|
}, |
|
{ |
|
"epoch": 78.08, |
|
"learning_rate": 1.1008823529411765e-05, |
|
"loss": 0.5416, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 78.17, |
|
"learning_rate": 1.0994117647058824e-05, |
|
"loss": 0.5029, |
|
"step": 8130 |
|
}, |
|
{ |
|
"epoch": 78.27, |
|
"learning_rate": 1.0979411764705884e-05, |
|
"loss": 0.5227, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 78.36, |
|
"learning_rate": 1.0964705882352941e-05, |
|
"loss": 0.515, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 78.46, |
|
"learning_rate": 1.095e-05, |
|
"loss": 0.5434, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 78.56, |
|
"learning_rate": 1.093529411764706e-05, |
|
"loss": 0.464, |
|
"step": 8170 |
|
}, |
|
{ |
|
"epoch": 78.65, |
|
"learning_rate": 1.0920588235294119e-05, |
|
"loss": 0.4984, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 78.75, |
|
"learning_rate": 1.0905882352941178e-05, |
|
"loss": 0.5363, |
|
"step": 8190 |
|
}, |
|
{ |
|
"epoch": 78.84, |
|
"learning_rate": 1.0891176470588237e-05, |
|
"loss": 0.5057, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 78.94, |
|
"learning_rate": 1.0876470588235294e-05, |
|
"loss": 0.4866, |
|
"step": 8210 |
|
}, |
|
{ |
|
"epoch": 79.04, |
|
"learning_rate": 1.0861764705882353e-05, |
|
"loss": 0.5512, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 79.13, |
|
"learning_rate": 1.0847058823529413e-05, |
|
"loss": 0.4914, |
|
"step": 8230 |
|
}, |
|
{ |
|
"epoch": 79.23, |
|
"learning_rate": 1.0832352941176472e-05, |
|
"loss": 0.5105, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 79.33, |
|
"learning_rate": 1.0817647058823531e-05, |
|
"loss": 0.4772, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 79.42, |
|
"learning_rate": 1.080294117647059e-05, |
|
"loss": 0.5295, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 79.52, |
|
"learning_rate": 1.0788235294117648e-05, |
|
"loss": 0.5131, |
|
"step": 8270 |
|
}, |
|
{ |
|
"epoch": 79.61, |
|
"learning_rate": 1.0773529411764707e-05, |
|
"loss": 0.4812, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 79.71, |
|
"learning_rate": 1.0758823529411766e-05, |
|
"loss": 0.5278, |
|
"step": 8290 |
|
}, |
|
{ |
|
"epoch": 79.8, |
|
"learning_rate": 1.0744117647058825e-05, |
|
"loss": 0.4869, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 79.9, |
|
"learning_rate": 1.0729411764705884e-05, |
|
"loss": 0.548, |
|
"step": 8310 |
|
}, |
|
{ |
|
"epoch": 80.0, |
|
"learning_rate": 1.0714705882352943e-05, |
|
"loss": 0.5351, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 80.1, |
|
"learning_rate": 1.0700000000000001e-05, |
|
"loss": 0.4945, |
|
"step": 8330 |
|
}, |
|
{ |
|
"epoch": 80.19, |
|
"learning_rate": 1.068529411764706e-05, |
|
"loss": 0.4896, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 80.29, |
|
"learning_rate": 1.067058823529412e-05, |
|
"loss": 0.5036, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 80.38, |
|
"learning_rate": 1.0655882352941178e-05, |
|
"loss": 0.5151, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 80.48, |
|
"learning_rate": 1.0641176470588238e-05, |
|
"loss": 0.5666, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 80.57, |
|
"learning_rate": 1.0626470588235297e-05, |
|
"loss": 0.4944, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 80.67, |
|
"learning_rate": 1.0611764705882354e-05, |
|
"loss": 0.4963, |
|
"step": 8390 |
|
}, |
|
{ |
|
"epoch": 80.77, |
|
"learning_rate": 1.0597058823529413e-05, |
|
"loss": 0.5275, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 80.86, |
|
"learning_rate": 1.0582352941176473e-05, |
|
"loss": 0.4922, |
|
"step": 8410 |
|
}, |
|
{ |
|
"epoch": 80.96, |
|
"learning_rate": 1.0567647058823532e-05, |
|
"loss": 0.461, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 81.06, |
|
"learning_rate": 1.0552941176470587e-05, |
|
"loss": 0.5044, |
|
"step": 8430 |
|
}, |
|
{ |
|
"epoch": 81.15, |
|
"learning_rate": 1.0538235294117647e-05, |
|
"loss": 0.5122, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 81.25, |
|
"learning_rate": 1.0523529411764706e-05, |
|
"loss": 0.4778, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 81.34, |
|
"learning_rate": 1.0508823529411765e-05, |
|
"loss": 0.4597, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 81.44, |
|
"learning_rate": 1.0494117647058824e-05, |
|
"loss": 0.483, |
|
"step": 8470 |
|
}, |
|
{ |
|
"epoch": 81.54, |
|
"learning_rate": 1.0479411764705882e-05, |
|
"loss": 0.5168, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 81.63, |
|
"learning_rate": 1.046470588235294e-05, |
|
"loss": 0.4919, |
|
"step": 8490 |
|
}, |
|
{ |
|
"epoch": 81.73, |
|
"learning_rate": 1.045e-05, |
|
"loss": 0.4974, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 81.73, |
|
"eval_loss": 0.2990259826183319, |
|
"eval_runtime": 64.5699, |
|
"eval_samples_per_second": 17.222, |
|
"eval_steps_per_second": 4.305, |
|
"eval_wer": 0.2932735426008969, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 81.82, |
|
"learning_rate": 1.0435294117647059e-05, |
|
"loss": 0.4972, |
|
"step": 8510 |
|
}, |
|
{ |
|
"epoch": 81.92, |
|
"learning_rate": 1.0420588235294118e-05, |
|
"loss": 0.4929, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 82.02, |
|
"learning_rate": 1.0405882352941177e-05, |
|
"loss": 0.5226, |
|
"step": 8530 |
|
}, |
|
{ |
|
"epoch": 82.11, |
|
"learning_rate": 1.0391176470588235e-05, |
|
"loss": 0.447, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 82.21, |
|
"learning_rate": 1.0376470588235294e-05, |
|
"loss": 0.5392, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 82.31, |
|
"learning_rate": 1.0361764705882353e-05, |
|
"loss": 0.471, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 82.4, |
|
"learning_rate": 1.0347058823529412e-05, |
|
"loss": 0.4981, |
|
"step": 8570 |
|
}, |
|
{ |
|
"epoch": 82.5, |
|
"learning_rate": 1.0332352941176472e-05, |
|
"loss": 0.5569, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 82.59, |
|
"learning_rate": 1.031764705882353e-05, |
|
"loss": 0.4502, |
|
"step": 8590 |
|
}, |
|
{ |
|
"epoch": 82.69, |
|
"learning_rate": 1.0302941176470588e-05, |
|
"loss": 0.4962, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 82.78, |
|
"learning_rate": 1.0288235294117647e-05, |
|
"loss": 0.4497, |
|
"step": 8610 |
|
}, |
|
{ |
|
"epoch": 82.88, |
|
"learning_rate": 1.0273529411764706e-05, |
|
"loss": 0.5258, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 82.98, |
|
"learning_rate": 1.0258823529411766e-05, |
|
"loss": 0.4804, |
|
"step": 8630 |
|
}, |
|
{ |
|
"epoch": 83.08, |
|
"learning_rate": 1.0244117647058825e-05, |
|
"loss": 0.5053, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 83.17, |
|
"learning_rate": 1.0229411764705884e-05, |
|
"loss": 0.4822, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 83.27, |
|
"learning_rate": 1.0214705882352941e-05, |
|
"loss": 0.4901, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 83.36, |
|
"learning_rate": 1.02e-05, |
|
"loss": 0.4645, |
|
"step": 8670 |
|
}, |
|
{ |
|
"epoch": 83.46, |
|
"learning_rate": 1.018529411764706e-05, |
|
"loss": 0.4805, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 83.56, |
|
"learning_rate": 1.0170588235294119e-05, |
|
"loss": 0.434, |
|
"step": 8690 |
|
}, |
|
{ |
|
"epoch": 83.65, |
|
"learning_rate": 1.0155882352941178e-05, |
|
"loss": 0.5031, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 83.75, |
|
"learning_rate": 1.0141176470588237e-05, |
|
"loss": 0.5355, |
|
"step": 8710 |
|
}, |
|
{ |
|
"epoch": 83.84, |
|
"learning_rate": 1.0126470588235295e-05, |
|
"loss": 0.4387, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 83.94, |
|
"learning_rate": 1.0111764705882354e-05, |
|
"loss": 0.5048, |
|
"step": 8730 |
|
}, |
|
{ |
|
"epoch": 84.04, |
|
"learning_rate": 1.0097058823529413e-05, |
|
"loss": 0.4859, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 84.13, |
|
"learning_rate": 1.0082352941176472e-05, |
|
"loss": 0.5156, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 84.23, |
|
"learning_rate": 1.0067647058823531e-05, |
|
"loss": 0.4887, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 84.33, |
|
"learning_rate": 1.005294117647059e-05, |
|
"loss": 0.4493, |
|
"step": 8770 |
|
}, |
|
{ |
|
"epoch": 84.42, |
|
"learning_rate": 1.0038235294117648e-05, |
|
"loss": 0.5062, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 84.52, |
|
"learning_rate": 1.0023529411764707e-05, |
|
"loss": 0.5028, |
|
"step": 8790 |
|
}, |
|
{ |
|
"epoch": 84.61, |
|
"learning_rate": 1.0008823529411766e-05, |
|
"loss": 0.4601, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 84.71, |
|
"learning_rate": 9.994117647058824e-06, |
|
"loss": 0.4792, |
|
"step": 8810 |
|
}, |
|
{ |
|
"epoch": 84.8, |
|
"learning_rate": 9.979411764705883e-06, |
|
"loss": 0.4539, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 84.9, |
|
"learning_rate": 9.964705882352942e-06, |
|
"loss": 0.4748, |
|
"step": 8830 |
|
}, |
|
{ |
|
"epoch": 85.0, |
|
"learning_rate": 9.950000000000001e-06, |
|
"loss": 0.494, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 85.1, |
|
"learning_rate": 9.93529411764706e-06, |
|
"loss": 0.4529, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 85.19, |
|
"learning_rate": 9.920588235294118e-06, |
|
"loss": 0.4975, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 85.29, |
|
"learning_rate": 9.905882352941177e-06, |
|
"loss": 0.4953, |
|
"step": 8870 |
|
}, |
|
{ |
|
"epoch": 85.38, |
|
"learning_rate": 9.891176470588236e-06, |
|
"loss": 0.4752, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 85.48, |
|
"learning_rate": 9.876470588235295e-06, |
|
"loss": 0.5153, |
|
"step": 8890 |
|
}, |
|
{ |
|
"epoch": 85.57, |
|
"learning_rate": 9.861764705882355e-06, |
|
"loss": 0.4416, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 85.67, |
|
"learning_rate": 9.847058823529412e-06, |
|
"loss": 0.4874, |
|
"step": 8910 |
|
}, |
|
{ |
|
"epoch": 85.77, |
|
"learning_rate": 9.832352941176471e-06, |
|
"loss": 0.4739, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 85.86, |
|
"learning_rate": 9.81764705882353e-06, |
|
"loss": 0.4526, |
|
"step": 8930 |
|
}, |
|
{ |
|
"epoch": 85.96, |
|
"learning_rate": 9.80294117647059e-06, |
|
"loss": 0.5054, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 86.06, |
|
"learning_rate": 9.788235294117649e-06, |
|
"loss": 0.454, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 86.15, |
|
"learning_rate": 9.773529411764706e-06, |
|
"loss": 0.4802, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 86.25, |
|
"learning_rate": 9.758823529411765e-06, |
|
"loss": 0.4653, |
|
"step": 8970 |
|
}, |
|
{ |
|
"epoch": 86.34, |
|
"learning_rate": 9.744117647058825e-06, |
|
"loss": 0.4155, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 86.44, |
|
"learning_rate": 9.729411764705882e-06, |
|
"loss": 0.4698, |
|
"step": 8990 |
|
}, |
|
{ |
|
"epoch": 86.54, |
|
"learning_rate": 9.714705882352941e-06, |
|
"loss": 0.4387, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 86.54, |
|
"eval_loss": 0.2834252417087555, |
|
"eval_runtime": 64.2806, |
|
"eval_samples_per_second": 17.299, |
|
"eval_steps_per_second": 4.325, |
|
"eval_wer": 0.2755156950672646, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 86.63, |
|
"learning_rate": 9.7e-06, |
|
"loss": 0.458, |
|
"step": 9010 |
|
}, |
|
{ |
|
"epoch": 86.73, |
|
"learning_rate": 9.68529411764706e-06, |
|
"loss": 0.4603, |
|
"step": 9020 |
|
}, |
|
{ |
|
"epoch": 86.82, |
|
"learning_rate": 9.670588235294119e-06, |
|
"loss": 0.4531, |
|
"step": 9030 |
|
}, |
|
{ |
|
"epoch": 86.92, |
|
"learning_rate": 9.655882352941178e-06, |
|
"loss": 0.5019, |
|
"step": 9040 |
|
}, |
|
{ |
|
"epoch": 87.02, |
|
"learning_rate": 9.641176470588235e-06, |
|
"loss": 0.5042, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 87.11, |
|
"learning_rate": 9.626470588235294e-06, |
|
"loss": 0.4676, |
|
"step": 9060 |
|
}, |
|
{ |
|
"epoch": 87.21, |
|
"learning_rate": 9.611764705882354e-06, |
|
"loss": 0.4826, |
|
"step": 9070 |
|
}, |
|
{ |
|
"epoch": 87.31, |
|
"learning_rate": 9.597058823529413e-06, |
|
"loss": 0.4258, |
|
"step": 9080 |
|
}, |
|
{ |
|
"epoch": 87.4, |
|
"learning_rate": 9.582352941176472e-06, |
|
"loss": 0.4513, |
|
"step": 9090 |
|
}, |
|
{ |
|
"epoch": 87.5, |
|
"learning_rate": 9.567647058823531e-06, |
|
"loss": 0.4633, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 87.59, |
|
"learning_rate": 9.552941176470589e-06, |
|
"loss": 0.4382, |
|
"step": 9110 |
|
}, |
|
{ |
|
"epoch": 87.69, |
|
"learning_rate": 9.538235294117648e-06, |
|
"loss": 0.4892, |
|
"step": 9120 |
|
}, |
|
{ |
|
"epoch": 87.78, |
|
"learning_rate": 9.523529411764707e-06, |
|
"loss": 0.488, |
|
"step": 9130 |
|
}, |
|
{ |
|
"epoch": 87.88, |
|
"learning_rate": 9.508823529411766e-06, |
|
"loss": 0.4741, |
|
"step": 9140 |
|
}, |
|
{ |
|
"epoch": 87.98, |
|
"learning_rate": 9.494117647058825e-06, |
|
"loss": 0.5128, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 88.08, |
|
"learning_rate": 9.479411764705884e-06, |
|
"loss": 0.417, |
|
"step": 9160 |
|
}, |
|
{ |
|
"epoch": 88.17, |
|
"learning_rate": 9.464705882352942e-06, |
|
"loss": 0.4903, |
|
"step": 9170 |
|
}, |
|
{ |
|
"epoch": 88.27, |
|
"learning_rate": 9.450000000000001e-06, |
|
"loss": 0.4583, |
|
"step": 9180 |
|
}, |
|
{ |
|
"epoch": 88.36, |
|
"learning_rate": 9.435294117647058e-06, |
|
"loss": 0.4601, |
|
"step": 9190 |
|
}, |
|
{ |
|
"epoch": 88.46, |
|
"learning_rate": 9.420588235294118e-06, |
|
"loss": 0.4945, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 88.56, |
|
"learning_rate": 9.405882352941177e-06, |
|
"loss": 0.4522, |
|
"step": 9210 |
|
}, |
|
{ |
|
"epoch": 88.65, |
|
"learning_rate": 9.391176470588236e-06, |
|
"loss": 0.4722, |
|
"step": 9220 |
|
}, |
|
{ |
|
"epoch": 88.75, |
|
"learning_rate": 9.376470588235295e-06, |
|
"loss": 0.4635, |
|
"step": 9230 |
|
}, |
|
{ |
|
"epoch": 88.84, |
|
"learning_rate": 9.361764705882354e-06, |
|
"loss": 0.4264, |
|
"step": 9240 |
|
}, |
|
{ |
|
"epoch": 88.94, |
|
"learning_rate": 9.347058823529412e-06, |
|
"loss": 0.4896, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 89.04, |
|
"learning_rate": 9.332352941176471e-06, |
|
"loss": 0.5102, |
|
"step": 9260 |
|
}, |
|
{ |
|
"epoch": 89.13, |
|
"learning_rate": 9.31764705882353e-06, |
|
"loss": 0.4935, |
|
"step": 9270 |
|
}, |
|
{ |
|
"epoch": 89.23, |
|
"learning_rate": 9.30294117647059e-06, |
|
"loss": 0.4513, |
|
"step": 9280 |
|
}, |
|
{ |
|
"epoch": 89.33, |
|
"learning_rate": 9.288235294117648e-06, |
|
"loss": 0.4577, |
|
"step": 9290 |
|
}, |
|
{ |
|
"epoch": 89.42, |
|
"learning_rate": 9.273529411764708e-06, |
|
"loss": 0.5036, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 89.52, |
|
"learning_rate": 9.258823529411765e-06, |
|
"loss": 0.4292, |
|
"step": 9310 |
|
}, |
|
{ |
|
"epoch": 89.61, |
|
"learning_rate": 9.244117647058824e-06, |
|
"loss": 0.4526, |
|
"step": 9320 |
|
}, |
|
{ |
|
"epoch": 89.71, |
|
"learning_rate": 9.229411764705883e-06, |
|
"loss": 0.4746, |
|
"step": 9330 |
|
}, |
|
{ |
|
"epoch": 89.8, |
|
"learning_rate": 9.214705882352943e-06, |
|
"loss": 0.4459, |
|
"step": 9340 |
|
}, |
|
{ |
|
"epoch": 89.9, |
|
"learning_rate": 9.200000000000002e-06, |
|
"loss": 0.4879, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 90.0, |
|
"learning_rate": 9.185294117647061e-06, |
|
"loss": 0.4646, |
|
"step": 9360 |
|
}, |
|
{ |
|
"epoch": 90.1, |
|
"learning_rate": 9.170588235294118e-06, |
|
"loss": 0.457, |
|
"step": 9370 |
|
}, |
|
{ |
|
"epoch": 90.19, |
|
"learning_rate": 9.155882352941177e-06, |
|
"loss": 0.4715, |
|
"step": 9380 |
|
}, |
|
{ |
|
"epoch": 90.29, |
|
"learning_rate": 9.141176470588235e-06, |
|
"loss": 0.4499, |
|
"step": 9390 |
|
}, |
|
{ |
|
"epoch": 90.38, |
|
"learning_rate": 9.126470588235294e-06, |
|
"loss": 0.4476, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 90.48, |
|
"learning_rate": 9.111764705882353e-06, |
|
"loss": 0.4863, |
|
"step": 9410 |
|
}, |
|
{ |
|
"epoch": 90.57, |
|
"learning_rate": 9.097058823529412e-06, |
|
"loss": 0.4529, |
|
"step": 9420 |
|
}, |
|
{ |
|
"epoch": 90.67, |
|
"learning_rate": 9.082352941176472e-06, |
|
"loss": 0.442, |
|
"step": 9430 |
|
}, |
|
{ |
|
"epoch": 90.77, |
|
"learning_rate": 9.067647058823529e-06, |
|
"loss": 0.4655, |
|
"step": 9440 |
|
}, |
|
{ |
|
"epoch": 90.86, |
|
"learning_rate": 9.052941176470588e-06, |
|
"loss": 0.4283, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 90.96, |
|
"learning_rate": 9.038235294117647e-06, |
|
"loss": 0.4841, |
|
"step": 9460 |
|
}, |
|
{ |
|
"epoch": 91.06, |
|
"learning_rate": 9.023529411764707e-06, |
|
"loss": 0.4681, |
|
"step": 9470 |
|
}, |
|
{ |
|
"epoch": 91.15, |
|
"learning_rate": 9.008823529411766e-06, |
|
"loss": 0.4466, |
|
"step": 9480 |
|
}, |
|
{ |
|
"epoch": 91.25, |
|
"learning_rate": 8.994117647058825e-06, |
|
"loss": 0.4904, |
|
"step": 9490 |
|
}, |
|
{ |
|
"epoch": 91.34, |
|
"learning_rate": 8.979411764705882e-06, |
|
"loss": 0.4511, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 91.34, |
|
"eval_loss": 0.2885821461677551, |
|
"eval_runtime": 64.801, |
|
"eval_samples_per_second": 17.16, |
|
"eval_steps_per_second": 4.29, |
|
"eval_wer": 0.2787443946188341, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 91.44, |
|
"learning_rate": 8.964705882352942e-06, |
|
"loss": 0.4606, |
|
"step": 9510 |
|
}, |
|
{ |
|
"epoch": 91.54, |
|
"learning_rate": 8.95e-06, |
|
"loss": 0.4366, |
|
"step": 9520 |
|
}, |
|
{ |
|
"epoch": 91.63, |
|
"learning_rate": 8.93529411764706e-06, |
|
"loss": 0.4076, |
|
"step": 9530 |
|
}, |
|
{ |
|
"epoch": 91.73, |
|
"learning_rate": 8.920588235294119e-06, |
|
"loss": 0.4534, |
|
"step": 9540 |
|
}, |
|
{ |
|
"epoch": 91.82, |
|
"learning_rate": 8.905882352941178e-06, |
|
"loss": 0.4321, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 91.92, |
|
"learning_rate": 8.891176470588236e-06, |
|
"loss": 0.473, |
|
"step": 9560 |
|
}, |
|
{ |
|
"epoch": 92.02, |
|
"learning_rate": 8.876470588235295e-06, |
|
"loss": 0.4974, |
|
"step": 9570 |
|
}, |
|
{ |
|
"epoch": 92.11, |
|
"learning_rate": 8.861764705882354e-06, |
|
"loss": 0.4852, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 92.21, |
|
"learning_rate": 8.847058823529413e-06, |
|
"loss": 0.4884, |
|
"step": 9590 |
|
}, |
|
{ |
|
"epoch": 92.31, |
|
"learning_rate": 8.83235294117647e-06, |
|
"loss": 0.4377, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 92.4, |
|
"learning_rate": 8.81764705882353e-06, |
|
"loss": 0.4529, |
|
"step": 9610 |
|
}, |
|
{ |
|
"epoch": 92.5, |
|
"learning_rate": 8.802941176470589e-06, |
|
"loss": 0.4638, |
|
"step": 9620 |
|
}, |
|
{ |
|
"epoch": 92.59, |
|
"learning_rate": 8.788235294117648e-06, |
|
"loss": 0.4472, |
|
"step": 9630 |
|
}, |
|
{ |
|
"epoch": 92.69, |
|
"learning_rate": 8.773529411764706e-06, |
|
"loss": 0.4701, |
|
"step": 9640 |
|
}, |
|
{ |
|
"epoch": 92.78, |
|
"learning_rate": 8.758823529411765e-06, |
|
"loss": 0.4272, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 92.88, |
|
"learning_rate": 8.744117647058824e-06, |
|
"loss": 0.4258, |
|
"step": 9660 |
|
}, |
|
{ |
|
"epoch": 92.98, |
|
"learning_rate": 8.729411764705883e-06, |
|
"loss": 0.4645, |
|
"step": 9670 |
|
}, |
|
{ |
|
"epoch": 93.08, |
|
"learning_rate": 8.714705882352942e-06, |
|
"loss": 0.4258, |
|
"step": 9680 |
|
}, |
|
{ |
|
"epoch": 93.17, |
|
"learning_rate": 8.700000000000001e-06, |
|
"loss": 0.4681, |
|
"step": 9690 |
|
}, |
|
{ |
|
"epoch": 93.27, |
|
"learning_rate": 8.685294117647059e-06, |
|
"loss": 0.4427, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 93.36, |
|
"learning_rate": 8.670588235294118e-06, |
|
"loss": 0.4288, |
|
"step": 9710 |
|
}, |
|
{ |
|
"epoch": 93.46, |
|
"learning_rate": 8.655882352941177e-06, |
|
"loss": 0.448, |
|
"step": 9720 |
|
}, |
|
{ |
|
"epoch": 93.56, |
|
"learning_rate": 8.641176470588236e-06, |
|
"loss": 0.4764, |
|
"step": 9730 |
|
}, |
|
{ |
|
"epoch": 93.65, |
|
"learning_rate": 8.626470588235296e-06, |
|
"loss": 0.4767, |
|
"step": 9740 |
|
}, |
|
{ |
|
"epoch": 93.75, |
|
"learning_rate": 8.611764705882355e-06, |
|
"loss": 0.4457, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 93.84, |
|
"learning_rate": 8.597058823529412e-06, |
|
"loss": 0.4272, |
|
"step": 9760 |
|
}, |
|
{ |
|
"epoch": 93.94, |
|
"learning_rate": 8.582352941176471e-06, |
|
"loss": 0.4269, |
|
"step": 9770 |
|
}, |
|
{ |
|
"epoch": 94.04, |
|
"learning_rate": 8.56764705882353e-06, |
|
"loss": 0.5044, |
|
"step": 9780 |
|
}, |
|
{ |
|
"epoch": 94.13, |
|
"learning_rate": 8.55294117647059e-06, |
|
"loss": 0.4223, |
|
"step": 9790 |
|
}, |
|
{ |
|
"epoch": 94.23, |
|
"learning_rate": 8.538235294117649e-06, |
|
"loss": 0.4618, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 94.33, |
|
"learning_rate": 8.523529411764706e-06, |
|
"loss": 0.4197, |
|
"step": 9810 |
|
}, |
|
{ |
|
"epoch": 94.42, |
|
"learning_rate": 8.508823529411765e-06, |
|
"loss": 0.4606, |
|
"step": 9820 |
|
}, |
|
{ |
|
"epoch": 94.52, |
|
"learning_rate": 8.494117647058825e-06, |
|
"loss": 0.4347, |
|
"step": 9830 |
|
}, |
|
{ |
|
"epoch": 94.61, |
|
"learning_rate": 8.479411764705882e-06, |
|
"loss": 0.4241, |
|
"step": 9840 |
|
}, |
|
{ |
|
"epoch": 94.71, |
|
"learning_rate": 8.464705882352941e-06, |
|
"loss": 0.4623, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 94.8, |
|
"learning_rate": 8.45e-06, |
|
"loss": 0.4137, |
|
"step": 9860 |
|
}, |
|
{ |
|
"epoch": 94.9, |
|
"learning_rate": 8.43529411764706e-06, |
|
"loss": 0.4909, |
|
"step": 9870 |
|
}, |
|
{ |
|
"epoch": 95.0, |
|
"learning_rate": 8.420588235294119e-06, |
|
"loss": 0.4855, |
|
"step": 9880 |
|
}, |
|
{ |
|
"epoch": 95.1, |
|
"learning_rate": 8.405882352941176e-06, |
|
"loss": 0.4239, |
|
"step": 9890 |
|
}, |
|
{ |
|
"epoch": 95.19, |
|
"learning_rate": 8.391176470588235e-06, |
|
"loss": 0.4446, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 95.29, |
|
"learning_rate": 8.376470588235295e-06, |
|
"loss": 0.4171, |
|
"step": 9910 |
|
}, |
|
{ |
|
"epoch": 95.38, |
|
"learning_rate": 8.361764705882354e-06, |
|
"loss": 0.4115, |
|
"step": 9920 |
|
}, |
|
{ |
|
"epoch": 95.48, |
|
"learning_rate": 8.347058823529413e-06, |
|
"loss": 0.4602, |
|
"step": 9930 |
|
}, |
|
{ |
|
"epoch": 95.57, |
|
"learning_rate": 8.332352941176472e-06, |
|
"loss": 0.3836, |
|
"step": 9940 |
|
}, |
|
{ |
|
"epoch": 95.67, |
|
"learning_rate": 8.31764705882353e-06, |
|
"loss": 0.4473, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 95.77, |
|
"learning_rate": 8.302941176470589e-06, |
|
"loss": 0.4429, |
|
"step": 9960 |
|
}, |
|
{ |
|
"epoch": 95.86, |
|
"learning_rate": 8.288235294117648e-06, |
|
"loss": 0.4204, |
|
"step": 9970 |
|
}, |
|
{ |
|
"epoch": 95.96, |
|
"learning_rate": 8.273529411764707e-06, |
|
"loss": 0.4428, |
|
"step": 9980 |
|
}, |
|
{ |
|
"epoch": 96.06, |
|
"learning_rate": 8.258823529411766e-06, |
|
"loss": 0.4324, |
|
"step": 9990 |
|
}, |
|
{ |
|
"epoch": 96.15, |
|
"learning_rate": 8.244117647058825e-06, |
|
"loss": 0.4112, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 96.15, |
|
"eval_loss": 0.3092893362045288, |
|
"eval_runtime": 64.6225, |
|
"eval_samples_per_second": 17.208, |
|
"eval_steps_per_second": 4.302, |
|
"eval_wer": 0.29757847533632287, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 96.25, |
|
"learning_rate": 8.229411764705883e-06, |
|
"loss": 0.4658, |
|
"step": 10010 |
|
}, |
|
{ |
|
"epoch": 96.34, |
|
"learning_rate": 8.214705882352942e-06, |
|
"loss": 0.3744, |
|
"step": 10020 |
|
}, |
|
{ |
|
"epoch": 96.44, |
|
"learning_rate": 8.2e-06, |
|
"loss": 0.4343, |
|
"step": 10030 |
|
}, |
|
{ |
|
"epoch": 96.54, |
|
"learning_rate": 8.185294117647059e-06, |
|
"loss": 0.4076, |
|
"step": 10040 |
|
}, |
|
{ |
|
"epoch": 96.63, |
|
"learning_rate": 8.170588235294118e-06, |
|
"loss": 0.3995, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 96.73, |
|
"learning_rate": 8.155882352941177e-06, |
|
"loss": 0.4331, |
|
"step": 10060 |
|
}, |
|
{ |
|
"epoch": 96.82, |
|
"learning_rate": 8.141176470588236e-06, |
|
"loss": 0.3849, |
|
"step": 10070 |
|
}, |
|
{ |
|
"epoch": 96.92, |
|
"learning_rate": 8.126470588235295e-06, |
|
"loss": 0.4651, |
|
"step": 10080 |
|
}, |
|
{ |
|
"epoch": 97.02, |
|
"learning_rate": 8.111764705882353e-06, |
|
"loss": 0.5139, |
|
"step": 10090 |
|
}, |
|
{ |
|
"epoch": 97.11, |
|
"learning_rate": 8.097058823529412e-06, |
|
"loss": 0.3721, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 97.21, |
|
"learning_rate": 8.082352941176471e-06, |
|
"loss": 0.4509, |
|
"step": 10110 |
|
}, |
|
{ |
|
"epoch": 97.31, |
|
"learning_rate": 8.06764705882353e-06, |
|
"loss": 0.3649, |
|
"step": 10120 |
|
}, |
|
{ |
|
"epoch": 97.4, |
|
"learning_rate": 8.05294117647059e-06, |
|
"loss": 0.4584, |
|
"step": 10130 |
|
}, |
|
{ |
|
"epoch": 97.5, |
|
"learning_rate": 8.038235294117649e-06, |
|
"loss": 0.4277, |
|
"step": 10140 |
|
}, |
|
{ |
|
"epoch": 97.59, |
|
"learning_rate": 8.023529411764706e-06, |
|
"loss": 0.4072, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 97.69, |
|
"learning_rate": 8.008823529411765e-06, |
|
"loss": 0.4345, |
|
"step": 10160 |
|
}, |
|
{ |
|
"epoch": 97.78, |
|
"learning_rate": 7.994117647058824e-06, |
|
"loss": 0.3687, |
|
"step": 10170 |
|
}, |
|
{ |
|
"epoch": 97.88, |
|
"learning_rate": 7.979411764705883e-06, |
|
"loss": 0.386, |
|
"step": 10180 |
|
}, |
|
{ |
|
"epoch": 97.98, |
|
"learning_rate": 7.964705882352943e-06, |
|
"loss": 0.4093, |
|
"step": 10190 |
|
}, |
|
{ |
|
"epoch": 98.08, |
|
"learning_rate": 7.950000000000002e-06, |
|
"loss": 0.396, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 98.17, |
|
"learning_rate": 7.93529411764706e-06, |
|
"loss": 0.4668, |
|
"step": 10210 |
|
}, |
|
{ |
|
"epoch": 98.27, |
|
"learning_rate": 7.920588235294118e-06, |
|
"loss": 0.4236, |
|
"step": 10220 |
|
}, |
|
{ |
|
"epoch": 98.36, |
|
"learning_rate": 7.905882352941176e-06, |
|
"loss": 0.3666, |
|
"step": 10230 |
|
}, |
|
{ |
|
"epoch": 98.46, |
|
"learning_rate": 7.891176470588235e-06, |
|
"loss": 0.4278, |
|
"step": 10240 |
|
}, |
|
{ |
|
"epoch": 98.56, |
|
"learning_rate": 7.876470588235294e-06, |
|
"loss": 0.4228, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 98.65, |
|
"learning_rate": 7.861764705882353e-06, |
|
"loss": 0.425, |
|
"step": 10260 |
|
}, |
|
{ |
|
"epoch": 98.75, |
|
"learning_rate": 7.847058823529413e-06, |
|
"loss": 0.4416, |
|
"step": 10270 |
|
}, |
|
{ |
|
"epoch": 98.84, |
|
"learning_rate": 7.832352941176472e-06, |
|
"loss": 0.4268, |
|
"step": 10280 |
|
}, |
|
{ |
|
"epoch": 98.94, |
|
"learning_rate": 7.81764705882353e-06, |
|
"loss": 0.3955, |
|
"step": 10290 |
|
}, |
|
{ |
|
"epoch": 99.04, |
|
"learning_rate": 7.802941176470588e-06, |
|
"loss": 0.4263, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 99.13, |
|
"learning_rate": 7.788235294117648e-06, |
|
"loss": 0.3934, |
|
"step": 10310 |
|
}, |
|
{ |
|
"epoch": 99.23, |
|
"learning_rate": 7.773529411764707e-06, |
|
"loss": 0.4102, |
|
"step": 10320 |
|
}, |
|
{ |
|
"epoch": 99.33, |
|
"learning_rate": 7.758823529411766e-06, |
|
"loss": 0.3971, |
|
"step": 10330 |
|
}, |
|
{ |
|
"epoch": 99.42, |
|
"learning_rate": 7.744117647058825e-06, |
|
"loss": 0.4102, |
|
"step": 10340 |
|
}, |
|
{ |
|
"epoch": 99.52, |
|
"learning_rate": 7.729411764705882e-06, |
|
"loss": 0.412, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 99.61, |
|
"learning_rate": 7.714705882352942e-06, |
|
"loss": 0.414, |
|
"step": 10360 |
|
}, |
|
{ |
|
"epoch": 99.71, |
|
"learning_rate": 7.7e-06, |
|
"loss": 0.4377, |
|
"step": 10370 |
|
}, |
|
{ |
|
"epoch": 99.8, |
|
"learning_rate": 7.68529411764706e-06, |
|
"loss": 0.4011, |
|
"step": 10380 |
|
}, |
|
{ |
|
"epoch": 99.9, |
|
"learning_rate": 7.670588235294119e-06, |
|
"loss": 0.409, |
|
"step": 10390 |
|
}, |
|
{ |
|
"epoch": 100.0, |
|
"learning_rate": 7.655882352941177e-06, |
|
"loss": 0.4647, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 100.1, |
|
"learning_rate": 7.641176470588236e-06, |
|
"loss": 0.4233, |
|
"step": 10410 |
|
}, |
|
{ |
|
"epoch": 100.19, |
|
"learning_rate": 7.626470588235295e-06, |
|
"loss": 0.4074, |
|
"step": 10420 |
|
}, |
|
{ |
|
"epoch": 100.29, |
|
"learning_rate": 7.611764705882354e-06, |
|
"loss": 0.3793, |
|
"step": 10430 |
|
}, |
|
{ |
|
"epoch": 100.38, |
|
"learning_rate": 7.597058823529413e-06, |
|
"loss": 0.382, |
|
"step": 10440 |
|
}, |
|
{ |
|
"epoch": 100.48, |
|
"learning_rate": 7.582352941176471e-06, |
|
"loss": 0.4269, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 100.57, |
|
"learning_rate": 7.56764705882353e-06, |
|
"loss": 0.3527, |
|
"step": 10460 |
|
}, |
|
{ |
|
"epoch": 100.67, |
|
"learning_rate": 7.552941176470588e-06, |
|
"loss": 0.4196, |
|
"step": 10470 |
|
}, |
|
{ |
|
"epoch": 100.77, |
|
"learning_rate": 7.538235294117647e-06, |
|
"loss": 0.4363, |
|
"step": 10480 |
|
}, |
|
{ |
|
"epoch": 100.86, |
|
"learning_rate": 7.5235294117647065e-06, |
|
"loss": 0.4062, |
|
"step": 10490 |
|
}, |
|
{ |
|
"epoch": 100.96, |
|
"learning_rate": 7.508823529411765e-06, |
|
"loss": 0.4064, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 100.96, |
|
"eval_loss": 0.3123227059841156, |
|
"eval_runtime": 64.2529, |
|
"eval_samples_per_second": 17.307, |
|
"eval_steps_per_second": 4.327, |
|
"eval_wer": 0.2862780269058296, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 101.06, |
|
"learning_rate": 7.494117647058824e-06, |
|
"loss": 0.4024, |
|
"step": 10510 |
|
}, |
|
{ |
|
"epoch": 101.15, |
|
"learning_rate": 7.479411764705883e-06, |
|
"loss": 0.4222, |
|
"step": 10520 |
|
}, |
|
{ |
|
"epoch": 101.25, |
|
"learning_rate": 7.4647058823529415e-06, |
|
"loss": 0.403, |
|
"step": 10530 |
|
}, |
|
{ |
|
"epoch": 101.34, |
|
"learning_rate": 7.450000000000001e-06, |
|
"loss": 0.3858, |
|
"step": 10540 |
|
}, |
|
{ |
|
"epoch": 101.44, |
|
"learning_rate": 7.43529411764706e-06, |
|
"loss": 0.4108, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 101.54, |
|
"learning_rate": 7.420588235294118e-06, |
|
"loss": 0.407, |
|
"step": 10560 |
|
}, |
|
{ |
|
"epoch": 101.63, |
|
"learning_rate": 7.405882352941177e-06, |
|
"loss": 0.4052, |
|
"step": 10570 |
|
}, |
|
{ |
|
"epoch": 101.73, |
|
"learning_rate": 7.3911764705882365e-06, |
|
"loss": 0.4135, |
|
"step": 10580 |
|
}, |
|
{ |
|
"epoch": 101.82, |
|
"learning_rate": 7.376470588235295e-06, |
|
"loss": 0.421, |
|
"step": 10590 |
|
}, |
|
{ |
|
"epoch": 101.92, |
|
"learning_rate": 7.361764705882354e-06, |
|
"loss": 0.4278, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 102.02, |
|
"learning_rate": 7.347058823529412e-06, |
|
"loss": 0.4706, |
|
"step": 10610 |
|
}, |
|
{ |
|
"epoch": 102.11, |
|
"learning_rate": 7.3323529411764714e-06, |
|
"loss": 0.3766, |
|
"step": 10620 |
|
}, |
|
{ |
|
"epoch": 102.21, |
|
"learning_rate": 7.317647058823531e-06, |
|
"loss": 0.4079, |
|
"step": 10630 |
|
}, |
|
{ |
|
"epoch": 102.31, |
|
"learning_rate": 7.302941176470589e-06, |
|
"loss": 0.3916, |
|
"step": 10640 |
|
}, |
|
{ |
|
"epoch": 102.4, |
|
"learning_rate": 7.288235294117648e-06, |
|
"loss": 0.4337, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 102.5, |
|
"learning_rate": 7.2735294117647056e-06, |
|
"loss": 0.4316, |
|
"step": 10660 |
|
}, |
|
{ |
|
"epoch": 102.59, |
|
"learning_rate": 7.258823529411765e-06, |
|
"loss": 0.3651, |
|
"step": 10670 |
|
}, |
|
{ |
|
"epoch": 102.69, |
|
"learning_rate": 7.244117647058824e-06, |
|
"loss": 0.4103, |
|
"step": 10680 |
|
}, |
|
{ |
|
"epoch": 102.78, |
|
"learning_rate": 7.229411764705882e-06, |
|
"loss": 0.4568, |
|
"step": 10690 |
|
}, |
|
{ |
|
"epoch": 102.88, |
|
"learning_rate": 7.214705882352941e-06, |
|
"loss": 0.4147, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 102.98, |
|
"learning_rate": 7.2000000000000005e-06, |
|
"loss": 0.4236, |
|
"step": 10710 |
|
}, |
|
{ |
|
"epoch": 103.08, |
|
"learning_rate": 7.185294117647059e-06, |
|
"loss": 0.413, |
|
"step": 10720 |
|
}, |
|
{ |
|
"epoch": 103.17, |
|
"learning_rate": 7.170588235294118e-06, |
|
"loss": 0.4237, |
|
"step": 10730 |
|
}, |
|
{ |
|
"epoch": 103.27, |
|
"learning_rate": 7.155882352941177e-06, |
|
"loss": 0.4193, |
|
"step": 10740 |
|
}, |
|
{ |
|
"epoch": 103.36, |
|
"learning_rate": 7.1411764705882355e-06, |
|
"loss": 0.3472, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 103.46, |
|
"learning_rate": 7.126470588235295e-06, |
|
"loss": 0.4451, |
|
"step": 10760 |
|
}, |
|
{ |
|
"epoch": 103.56, |
|
"learning_rate": 7.111764705882354e-06, |
|
"loss": 0.3596, |
|
"step": 10770 |
|
}, |
|
{ |
|
"epoch": 103.65, |
|
"learning_rate": 7.097058823529412e-06, |
|
"loss": 0.4148, |
|
"step": 10780 |
|
}, |
|
{ |
|
"epoch": 103.75, |
|
"learning_rate": 7.082352941176471e-06, |
|
"loss": 0.4077, |
|
"step": 10790 |
|
}, |
|
{ |
|
"epoch": 103.84, |
|
"learning_rate": 7.0676470588235305e-06, |
|
"loss": 0.3837, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 103.94, |
|
"learning_rate": 7.052941176470589e-06, |
|
"loss": 0.3791, |
|
"step": 10810 |
|
}, |
|
{ |
|
"epoch": 104.04, |
|
"learning_rate": 7.038235294117648e-06, |
|
"loss": 0.4162, |
|
"step": 10820 |
|
}, |
|
{ |
|
"epoch": 104.13, |
|
"learning_rate": 7.023529411764707e-06, |
|
"loss": 0.4372, |
|
"step": 10830 |
|
}, |
|
{ |
|
"epoch": 104.23, |
|
"learning_rate": 7.008823529411765e-06, |
|
"loss": 0.413, |
|
"step": 10840 |
|
}, |
|
{ |
|
"epoch": 104.33, |
|
"learning_rate": 6.994117647058825e-06, |
|
"loss": 0.3863, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 104.42, |
|
"learning_rate": 6.979411764705884e-06, |
|
"loss": 0.3745, |
|
"step": 10860 |
|
}, |
|
{ |
|
"epoch": 104.52, |
|
"learning_rate": 6.964705882352941e-06, |
|
"loss": 0.4329, |
|
"step": 10870 |
|
}, |
|
{ |
|
"epoch": 104.61, |
|
"learning_rate": 6.95e-06, |
|
"loss": 0.4001, |
|
"step": 10880 |
|
}, |
|
{ |
|
"epoch": 104.71, |
|
"learning_rate": 6.935294117647059e-06, |
|
"loss": 0.4268, |
|
"step": 10890 |
|
}, |
|
{ |
|
"epoch": 104.8, |
|
"learning_rate": 6.920588235294118e-06, |
|
"loss": 0.3947, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 104.9, |
|
"learning_rate": 6.905882352941177e-06, |
|
"loss": 0.4316, |
|
"step": 10910 |
|
}, |
|
{ |
|
"epoch": 105.0, |
|
"learning_rate": 6.891176470588235e-06, |
|
"loss": 0.4296, |
|
"step": 10920 |
|
}, |
|
{ |
|
"epoch": 105.1, |
|
"learning_rate": 6.8764705882352945e-06, |
|
"loss": 0.4397, |
|
"step": 10930 |
|
}, |
|
{ |
|
"epoch": 105.19, |
|
"learning_rate": 6.861764705882354e-06, |
|
"loss": 0.44, |
|
"step": 10940 |
|
}, |
|
{ |
|
"epoch": 105.29, |
|
"learning_rate": 6.847058823529412e-06, |
|
"loss": 0.3805, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 105.38, |
|
"learning_rate": 6.832352941176471e-06, |
|
"loss": 0.3978, |
|
"step": 10960 |
|
}, |
|
{ |
|
"epoch": 105.48, |
|
"learning_rate": 6.81764705882353e-06, |
|
"loss": 0.4091, |
|
"step": 10970 |
|
}, |
|
{ |
|
"epoch": 105.57, |
|
"learning_rate": 6.802941176470589e-06, |
|
"loss": 0.384, |
|
"step": 10980 |
|
}, |
|
{ |
|
"epoch": 105.67, |
|
"learning_rate": 6.788235294117648e-06, |
|
"loss": 0.3767, |
|
"step": 10990 |
|
}, |
|
{ |
|
"epoch": 105.77, |
|
"learning_rate": 6.773529411764707e-06, |
|
"loss": 0.4047, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 105.77, |
|
"eval_loss": 0.29678910970687866, |
|
"eval_runtime": 65.6534, |
|
"eval_samples_per_second": 16.937, |
|
"eval_steps_per_second": 4.234, |
|
"eval_wer": 0.2719282511210762, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 105.86, |
|
"learning_rate": 6.758823529411765e-06, |
|
"loss": 0.4137, |
|
"step": 11010 |
|
}, |
|
{ |
|
"epoch": 105.96, |
|
"learning_rate": 6.7441176470588244e-06, |
|
"loss": 0.4148, |
|
"step": 11020 |
|
}, |
|
{ |
|
"epoch": 106.06, |
|
"learning_rate": 6.729411764705884e-06, |
|
"loss": 0.4109, |
|
"step": 11030 |
|
}, |
|
{ |
|
"epoch": 106.15, |
|
"learning_rate": 6.714705882352942e-06, |
|
"loss": 0.4086, |
|
"step": 11040 |
|
}, |
|
{ |
|
"epoch": 106.25, |
|
"learning_rate": 6.700000000000001e-06, |
|
"loss": 0.3794, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 106.34, |
|
"learning_rate": 6.68529411764706e-06, |
|
"loss": 0.3923, |
|
"step": 11060 |
|
}, |
|
{ |
|
"epoch": 106.44, |
|
"learning_rate": 6.6705882352941186e-06, |
|
"loss": 0.4019, |
|
"step": 11070 |
|
}, |
|
{ |
|
"epoch": 106.54, |
|
"learning_rate": 6.655882352941177e-06, |
|
"loss": 0.3735, |
|
"step": 11080 |
|
}, |
|
{ |
|
"epoch": 106.63, |
|
"learning_rate": 6.641176470588235e-06, |
|
"loss": 0.3908, |
|
"step": 11090 |
|
}, |
|
{ |
|
"epoch": 106.73, |
|
"learning_rate": 6.626470588235294e-06, |
|
"loss": 0.4206, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 106.82, |
|
"learning_rate": 6.611764705882353e-06, |
|
"loss": 0.3573, |
|
"step": 11110 |
|
}, |
|
{ |
|
"epoch": 106.92, |
|
"learning_rate": 6.597058823529412e-06, |
|
"loss": 0.4055, |
|
"step": 11120 |
|
}, |
|
{ |
|
"epoch": 107.02, |
|
"learning_rate": 6.582352941176471e-06, |
|
"loss": 0.4335, |
|
"step": 11130 |
|
}, |
|
{ |
|
"epoch": 107.11, |
|
"learning_rate": 6.567647058823529e-06, |
|
"loss": 0.3998, |
|
"step": 11140 |
|
}, |
|
{ |
|
"epoch": 107.21, |
|
"learning_rate": 6.5529411764705885e-06, |
|
"loss": 0.3945, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 107.31, |
|
"learning_rate": 6.538235294117648e-06, |
|
"loss": 0.4015, |
|
"step": 11160 |
|
}, |
|
{ |
|
"epoch": 107.4, |
|
"learning_rate": 6.523529411764706e-06, |
|
"loss": 0.4011, |
|
"step": 11170 |
|
}, |
|
{ |
|
"epoch": 107.5, |
|
"learning_rate": 6.508823529411765e-06, |
|
"loss": 0.4041, |
|
"step": 11180 |
|
}, |
|
{ |
|
"epoch": 107.59, |
|
"learning_rate": 6.494117647058824e-06, |
|
"loss": 0.3636, |
|
"step": 11190 |
|
}, |
|
{ |
|
"epoch": 107.69, |
|
"learning_rate": 6.479411764705883e-06, |
|
"loss": 0.4269, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 107.78, |
|
"learning_rate": 6.464705882352942e-06, |
|
"loss": 0.3657, |
|
"step": 11210 |
|
}, |
|
{ |
|
"epoch": 107.88, |
|
"learning_rate": 6.450000000000001e-06, |
|
"loss": 0.4016, |
|
"step": 11220 |
|
}, |
|
{ |
|
"epoch": 107.98, |
|
"learning_rate": 6.435294117647059e-06, |
|
"loss": 0.4082, |
|
"step": 11230 |
|
}, |
|
{ |
|
"epoch": 108.08, |
|
"learning_rate": 6.420588235294118e-06, |
|
"loss": 0.4054, |
|
"step": 11240 |
|
}, |
|
{ |
|
"epoch": 108.17, |
|
"learning_rate": 6.4058823529411776e-06, |
|
"loss": 0.4016, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 108.27, |
|
"learning_rate": 6.391176470588236e-06, |
|
"loss": 0.3793, |
|
"step": 11260 |
|
}, |
|
{ |
|
"epoch": 108.36, |
|
"learning_rate": 6.376470588235295e-06, |
|
"loss": 0.4217, |
|
"step": 11270 |
|
}, |
|
{ |
|
"epoch": 108.46, |
|
"learning_rate": 6.361764705882354e-06, |
|
"loss": 0.3891, |
|
"step": 11280 |
|
}, |
|
{ |
|
"epoch": 108.56, |
|
"learning_rate": 6.3470588235294125e-06, |
|
"loss": 0.3814, |
|
"step": 11290 |
|
}, |
|
{ |
|
"epoch": 108.65, |
|
"learning_rate": 6.332352941176471e-06, |
|
"loss": 0.41, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 108.75, |
|
"learning_rate": 6.317647058823529e-06, |
|
"loss": 0.4173, |
|
"step": 11310 |
|
}, |
|
{ |
|
"epoch": 108.84, |
|
"learning_rate": 6.302941176470588e-06, |
|
"loss": 0.3396, |
|
"step": 11320 |
|
}, |
|
{ |
|
"epoch": 108.94, |
|
"learning_rate": 6.2882352941176475e-06, |
|
"loss": 0.4105, |
|
"step": 11330 |
|
}, |
|
{ |
|
"epoch": 109.04, |
|
"learning_rate": 6.273529411764706e-06, |
|
"loss": 0.378, |
|
"step": 11340 |
|
}, |
|
{ |
|
"epoch": 109.13, |
|
"learning_rate": 6.258823529411765e-06, |
|
"loss": 0.3762, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 109.23, |
|
"learning_rate": 6.244117647058824e-06, |
|
"loss": 0.4107, |
|
"step": 11360 |
|
}, |
|
{ |
|
"epoch": 109.33, |
|
"learning_rate": 6.2294117647058825e-06, |
|
"loss": 0.3329, |
|
"step": 11370 |
|
}, |
|
{ |
|
"epoch": 109.42, |
|
"learning_rate": 6.216176470588236e-06, |
|
"loss": 0.4052, |
|
"step": 11380 |
|
}, |
|
{ |
|
"epoch": 109.52, |
|
"learning_rate": 6.201470588235294e-06, |
|
"loss": 0.4056, |
|
"step": 11390 |
|
}, |
|
{ |
|
"epoch": 109.61, |
|
"learning_rate": 6.186764705882353e-06, |
|
"loss": 0.3694, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 109.71, |
|
"learning_rate": 6.1720588235294125e-06, |
|
"loss": 0.415, |
|
"step": 11410 |
|
}, |
|
{ |
|
"epoch": 109.8, |
|
"learning_rate": 6.157352941176471e-06, |
|
"loss": 0.3434, |
|
"step": 11420 |
|
}, |
|
{ |
|
"epoch": 109.9, |
|
"learning_rate": 6.14264705882353e-06, |
|
"loss": 0.3652, |
|
"step": 11430 |
|
}, |
|
{ |
|
"epoch": 110.0, |
|
"learning_rate": 6.127941176470588e-06, |
|
"loss": 0.3531, |
|
"step": 11440 |
|
}, |
|
{ |
|
"epoch": 110.1, |
|
"learning_rate": 6.1132352941176475e-06, |
|
"loss": 0.318, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 110.19, |
|
"learning_rate": 6.098529411764707e-06, |
|
"loss": 0.3946, |
|
"step": 11460 |
|
}, |
|
{ |
|
"epoch": 110.29, |
|
"learning_rate": 6.083823529411765e-06, |
|
"loss": 0.3516, |
|
"step": 11470 |
|
}, |
|
{ |
|
"epoch": 110.38, |
|
"learning_rate": 6.069117647058824e-06, |
|
"loss": 0.4008, |
|
"step": 11480 |
|
}, |
|
{ |
|
"epoch": 110.48, |
|
"learning_rate": 6.054411764705883e-06, |
|
"loss": 0.3657, |
|
"step": 11490 |
|
}, |
|
{ |
|
"epoch": 110.57, |
|
"learning_rate": 6.039705882352942e-06, |
|
"loss": 0.3519, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 110.57, |
|
"eval_loss": 0.31063979864120483, |
|
"eval_runtime": 66.2252, |
|
"eval_samples_per_second": 16.791, |
|
"eval_steps_per_second": 4.198, |
|
"eval_wer": 0.2832286995515695, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 110.67, |
|
"learning_rate": 6.025000000000001e-06, |
|
"loss": 0.417, |
|
"step": 11510 |
|
}, |
|
{ |
|
"epoch": 110.77, |
|
"learning_rate": 6.01029411764706e-06, |
|
"loss": 0.4067, |
|
"step": 11520 |
|
}, |
|
{ |
|
"epoch": 110.86, |
|
"learning_rate": 5.995588235294118e-06, |
|
"loss": 0.4003, |
|
"step": 11530 |
|
}, |
|
{ |
|
"epoch": 110.96, |
|
"learning_rate": 5.980882352941177e-06, |
|
"loss": 0.3703, |
|
"step": 11540 |
|
}, |
|
{ |
|
"epoch": 111.06, |
|
"learning_rate": 5.966176470588235e-06, |
|
"loss": 0.3668, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 111.15, |
|
"learning_rate": 5.951470588235294e-06, |
|
"loss": 0.3864, |
|
"step": 11560 |
|
}, |
|
{ |
|
"epoch": 111.25, |
|
"learning_rate": 5.936764705882353e-06, |
|
"loss": 0.3889, |
|
"step": 11570 |
|
}, |
|
{ |
|
"epoch": 111.34, |
|
"learning_rate": 5.9220588235294115e-06, |
|
"loss": 0.3524, |
|
"step": 11580 |
|
}, |
|
{ |
|
"epoch": 111.44, |
|
"learning_rate": 5.907352941176471e-06, |
|
"loss": 0.3845, |
|
"step": 11590 |
|
}, |
|
{ |
|
"epoch": 111.54, |
|
"learning_rate": 5.89264705882353e-06, |
|
"loss": 0.3644, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 111.63, |
|
"learning_rate": 5.877941176470588e-06, |
|
"loss": 0.4055, |
|
"step": 11610 |
|
}, |
|
{ |
|
"epoch": 111.73, |
|
"learning_rate": 5.863235294117647e-06, |
|
"loss": 0.3602, |
|
"step": 11620 |
|
}, |
|
{ |
|
"epoch": 111.82, |
|
"learning_rate": 5.8485294117647065e-06, |
|
"loss": 0.3532, |
|
"step": 11630 |
|
}, |
|
{ |
|
"epoch": 111.92, |
|
"learning_rate": 5.833823529411765e-06, |
|
"loss": 0.4027, |
|
"step": 11640 |
|
}, |
|
{ |
|
"epoch": 112.02, |
|
"learning_rate": 5.819117647058824e-06, |
|
"loss": 0.412, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 112.11, |
|
"learning_rate": 5.804411764705883e-06, |
|
"loss": 0.3667, |
|
"step": 11660 |
|
}, |
|
{ |
|
"epoch": 112.21, |
|
"learning_rate": 5.7897058823529414e-06, |
|
"loss": 0.4427, |
|
"step": 11670 |
|
}, |
|
{ |
|
"epoch": 112.31, |
|
"learning_rate": 5.775000000000001e-06, |
|
"loss": 0.3766, |
|
"step": 11680 |
|
}, |
|
{ |
|
"epoch": 112.4, |
|
"learning_rate": 5.76029411764706e-06, |
|
"loss": 0.3611, |
|
"step": 11690 |
|
}, |
|
{ |
|
"epoch": 112.5, |
|
"learning_rate": 5.745588235294118e-06, |
|
"loss": 0.3739, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 112.59, |
|
"learning_rate": 5.730882352941177e-06, |
|
"loss": 0.3814, |
|
"step": 11710 |
|
}, |
|
{ |
|
"epoch": 112.69, |
|
"learning_rate": 5.716176470588236e-06, |
|
"loss": 0.3646, |
|
"step": 11720 |
|
}, |
|
{ |
|
"epoch": 112.78, |
|
"learning_rate": 5.701470588235295e-06, |
|
"loss": 0.3779, |
|
"step": 11730 |
|
}, |
|
{ |
|
"epoch": 112.88, |
|
"learning_rate": 5.686764705882354e-06, |
|
"loss": 0.3923, |
|
"step": 11740 |
|
}, |
|
{ |
|
"epoch": 112.98, |
|
"learning_rate": 5.672058823529413e-06, |
|
"loss": 0.3879, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 113.08, |
|
"learning_rate": 5.657352941176471e-06, |
|
"loss": 0.3768, |
|
"step": 11760 |
|
}, |
|
{ |
|
"epoch": 113.17, |
|
"learning_rate": 5.64264705882353e-06, |
|
"loss": 0.3857, |
|
"step": 11770 |
|
}, |
|
{ |
|
"epoch": 113.27, |
|
"learning_rate": 5.627941176470588e-06, |
|
"loss": 0.357, |
|
"step": 11780 |
|
}, |
|
{ |
|
"epoch": 113.36, |
|
"learning_rate": 5.613235294117647e-06, |
|
"loss": 0.3735, |
|
"step": 11790 |
|
}, |
|
{ |
|
"epoch": 113.46, |
|
"learning_rate": 5.598529411764706e-06, |
|
"loss": 0.3877, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 113.56, |
|
"learning_rate": 5.583823529411765e-06, |
|
"loss": 0.3619, |
|
"step": 11810 |
|
}, |
|
{ |
|
"epoch": 113.65, |
|
"learning_rate": 5.569117647058824e-06, |
|
"loss": 0.3786, |
|
"step": 11820 |
|
}, |
|
{ |
|
"epoch": 113.75, |
|
"learning_rate": 5.554411764705883e-06, |
|
"loss": 0.3605, |
|
"step": 11830 |
|
}, |
|
{ |
|
"epoch": 113.84, |
|
"learning_rate": 5.539705882352941e-06, |
|
"loss": 0.3535, |
|
"step": 11840 |
|
}, |
|
{ |
|
"epoch": 113.94, |
|
"learning_rate": 5.5250000000000005e-06, |
|
"loss": 0.3855, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 114.04, |
|
"learning_rate": 5.51029411764706e-06, |
|
"loss": 0.4168, |
|
"step": 11860 |
|
}, |
|
{ |
|
"epoch": 114.13, |
|
"learning_rate": 5.495588235294118e-06, |
|
"loss": 0.3698, |
|
"step": 11870 |
|
}, |
|
{ |
|
"epoch": 114.23, |
|
"learning_rate": 5.480882352941177e-06, |
|
"loss": 0.3924, |
|
"step": 11880 |
|
}, |
|
{ |
|
"epoch": 114.33, |
|
"learning_rate": 5.466176470588236e-06, |
|
"loss": 0.3462, |
|
"step": 11890 |
|
}, |
|
{ |
|
"epoch": 114.42, |
|
"learning_rate": 5.451470588235295e-06, |
|
"loss": 0.3985, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 114.52, |
|
"learning_rate": 5.436764705882354e-06, |
|
"loss": 0.4173, |
|
"step": 11910 |
|
}, |
|
{ |
|
"epoch": 114.61, |
|
"learning_rate": 5.422058823529412e-06, |
|
"loss": 0.3644, |
|
"step": 11920 |
|
}, |
|
{ |
|
"epoch": 114.71, |
|
"learning_rate": 5.407352941176471e-06, |
|
"loss": 0.3644, |
|
"step": 11930 |
|
}, |
|
{ |
|
"epoch": 114.8, |
|
"learning_rate": 5.39264705882353e-06, |
|
"loss": 0.3412, |
|
"step": 11940 |
|
}, |
|
{ |
|
"epoch": 114.9, |
|
"learning_rate": 5.377941176470589e-06, |
|
"loss": 0.385, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 115.0, |
|
"learning_rate": 5.363235294117648e-06, |
|
"loss": 0.3664, |
|
"step": 11960 |
|
}, |
|
{ |
|
"epoch": 115.1, |
|
"learning_rate": 5.348529411764707e-06, |
|
"loss": 0.3467, |
|
"step": 11970 |
|
}, |
|
{ |
|
"epoch": 115.19, |
|
"learning_rate": 5.3338235294117645e-06, |
|
"loss": 0.3733, |
|
"step": 11980 |
|
}, |
|
{ |
|
"epoch": 115.29, |
|
"learning_rate": 5.319117647058824e-06, |
|
"loss": 0.3399, |
|
"step": 11990 |
|
}, |
|
{ |
|
"epoch": 115.38, |
|
"learning_rate": 5.304411764705882e-06, |
|
"loss": 0.3719, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 115.38, |
|
"eval_loss": 0.3029741644859314, |
|
"eval_runtime": 64.0477, |
|
"eval_samples_per_second": 17.362, |
|
"eval_steps_per_second": 4.341, |
|
"eval_wer": 0.2737219730941704, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 115.48, |
|
"learning_rate": 5.289705882352941e-06, |
|
"loss": 0.3697, |
|
"step": 12010 |
|
}, |
|
{ |
|
"epoch": 115.57, |
|
"learning_rate": 5.275e-06, |
|
"loss": 0.3672, |
|
"step": 12020 |
|
}, |
|
{ |
|
"epoch": 115.67, |
|
"learning_rate": 5.260294117647059e-06, |
|
"loss": 0.3549, |
|
"step": 12030 |
|
}, |
|
{ |
|
"epoch": 115.77, |
|
"learning_rate": 5.245588235294118e-06, |
|
"loss": 0.3698, |
|
"step": 12040 |
|
}, |
|
{ |
|
"epoch": 115.86, |
|
"learning_rate": 5.230882352941177e-06, |
|
"loss": 0.3571, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 115.96, |
|
"learning_rate": 5.216176470588235e-06, |
|
"loss": 0.373, |
|
"step": 12060 |
|
}, |
|
{ |
|
"epoch": 116.06, |
|
"learning_rate": 5.2014705882352944e-06, |
|
"loss": 0.3322, |
|
"step": 12070 |
|
}, |
|
{ |
|
"epoch": 116.15, |
|
"learning_rate": 5.186764705882354e-06, |
|
"loss": 0.3121, |
|
"step": 12080 |
|
}, |
|
{ |
|
"epoch": 116.25, |
|
"learning_rate": 5.172058823529412e-06, |
|
"loss": 0.381, |
|
"step": 12090 |
|
}, |
|
{ |
|
"epoch": 116.34, |
|
"learning_rate": 5.157352941176471e-06, |
|
"loss": 0.3475, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 116.44, |
|
"learning_rate": 5.14264705882353e-06, |
|
"loss": 0.3382, |
|
"step": 12110 |
|
}, |
|
{ |
|
"epoch": 116.54, |
|
"learning_rate": 5.1279411764705886e-06, |
|
"loss": 0.3529, |
|
"step": 12120 |
|
}, |
|
{ |
|
"epoch": 116.63, |
|
"learning_rate": 5.113235294117648e-06, |
|
"loss": 0.393, |
|
"step": 12130 |
|
}, |
|
{ |
|
"epoch": 116.73, |
|
"learning_rate": 5.098529411764707e-06, |
|
"loss": 0.4254, |
|
"step": 12140 |
|
}, |
|
{ |
|
"epoch": 116.82, |
|
"learning_rate": 5.083823529411765e-06, |
|
"loss": 0.329, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 116.92, |
|
"learning_rate": 5.069117647058824e-06, |
|
"loss": 0.368, |
|
"step": 12160 |
|
}, |
|
{ |
|
"epoch": 117.02, |
|
"learning_rate": 5.0544117647058835e-06, |
|
"loss": 0.4013, |
|
"step": 12170 |
|
}, |
|
{ |
|
"epoch": 117.11, |
|
"learning_rate": 5.039705882352942e-06, |
|
"loss": 0.3722, |
|
"step": 12180 |
|
}, |
|
{ |
|
"epoch": 117.21, |
|
"learning_rate": 5.025e-06, |
|
"loss": 0.3957, |
|
"step": 12190 |
|
}, |
|
{ |
|
"epoch": 117.31, |
|
"learning_rate": 5.0102941176470585e-06, |
|
"loss": 0.3499, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 117.4, |
|
"learning_rate": 4.9955882352941185e-06, |
|
"loss": 0.3676, |
|
"step": 12210 |
|
}, |
|
{ |
|
"epoch": 117.5, |
|
"learning_rate": 4.980882352941177e-06, |
|
"loss": 0.348, |
|
"step": 12220 |
|
}, |
|
{ |
|
"epoch": 117.59, |
|
"learning_rate": 4.966176470588236e-06, |
|
"loss": 0.3895, |
|
"step": 12230 |
|
}, |
|
{ |
|
"epoch": 117.69, |
|
"learning_rate": 4.951470588235294e-06, |
|
"loss": 0.4196, |
|
"step": 12240 |
|
}, |
|
{ |
|
"epoch": 117.78, |
|
"learning_rate": 4.9367647058823535e-06, |
|
"loss": 0.3694, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 117.88, |
|
"learning_rate": 4.922058823529412e-06, |
|
"loss": 0.3755, |
|
"step": 12260 |
|
}, |
|
{ |
|
"epoch": 117.98, |
|
"learning_rate": 4.907352941176471e-06, |
|
"loss": 0.3805, |
|
"step": 12270 |
|
}, |
|
{ |
|
"epoch": 118.08, |
|
"learning_rate": 4.89264705882353e-06, |
|
"loss": 0.3448, |
|
"step": 12280 |
|
}, |
|
{ |
|
"epoch": 118.17, |
|
"learning_rate": 4.877941176470588e-06, |
|
"loss": 0.3513, |
|
"step": 12290 |
|
}, |
|
{ |
|
"epoch": 118.27, |
|
"learning_rate": 4.863235294117648e-06, |
|
"loss": 0.3578, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 118.36, |
|
"learning_rate": 4.848529411764707e-06, |
|
"loss": 0.3365, |
|
"step": 12310 |
|
}, |
|
{ |
|
"epoch": 118.46, |
|
"learning_rate": 4.833823529411765e-06, |
|
"loss": 0.3456, |
|
"step": 12320 |
|
}, |
|
{ |
|
"epoch": 118.56, |
|
"learning_rate": 4.819117647058824e-06, |
|
"loss": 0.3225, |
|
"step": 12330 |
|
}, |
|
{ |
|
"epoch": 118.65, |
|
"learning_rate": 4.804411764705883e-06, |
|
"loss": 0.3577, |
|
"step": 12340 |
|
}, |
|
{ |
|
"epoch": 118.75, |
|
"learning_rate": 4.789705882352942e-06, |
|
"loss": 0.35, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 118.84, |
|
"learning_rate": 4.775e-06, |
|
"loss": 0.3493, |
|
"step": 12360 |
|
}, |
|
{ |
|
"epoch": 118.94, |
|
"learning_rate": 4.760294117647059e-06, |
|
"loss": 0.4084, |
|
"step": 12370 |
|
}, |
|
{ |
|
"epoch": 119.04, |
|
"learning_rate": 4.745588235294118e-06, |
|
"loss": 0.3678, |
|
"step": 12380 |
|
}, |
|
{ |
|
"epoch": 119.13, |
|
"learning_rate": 4.730882352941177e-06, |
|
"loss": 0.3532, |
|
"step": 12390 |
|
}, |
|
{ |
|
"epoch": 119.23, |
|
"learning_rate": 4.716176470588236e-06, |
|
"loss": 0.4009, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 119.33, |
|
"learning_rate": 4.701470588235294e-06, |
|
"loss": 0.3504, |
|
"step": 12410 |
|
}, |
|
{ |
|
"epoch": 119.42, |
|
"learning_rate": 4.686764705882353e-06, |
|
"loss": 0.4023, |
|
"step": 12420 |
|
}, |
|
{ |
|
"epoch": 119.52, |
|
"learning_rate": 4.6720588235294125e-06, |
|
"loss": 0.3655, |
|
"step": 12430 |
|
}, |
|
{ |
|
"epoch": 119.61, |
|
"learning_rate": 4.657352941176471e-06, |
|
"loss": 0.3455, |
|
"step": 12440 |
|
}, |
|
{ |
|
"epoch": 119.71, |
|
"learning_rate": 4.64264705882353e-06, |
|
"loss": 0.3452, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 119.8, |
|
"learning_rate": 4.627941176470588e-06, |
|
"loss": 0.3224, |
|
"step": 12460 |
|
}, |
|
{ |
|
"epoch": 119.9, |
|
"learning_rate": 4.6132352941176474e-06, |
|
"loss": 0.367, |
|
"step": 12470 |
|
}, |
|
{ |
|
"epoch": 120.0, |
|
"learning_rate": 4.598529411764706e-06, |
|
"loss": 0.3586, |
|
"step": 12480 |
|
}, |
|
{ |
|
"epoch": 120.1, |
|
"learning_rate": 4.583823529411765e-06, |
|
"loss": 0.3808, |
|
"step": 12490 |
|
}, |
|
{ |
|
"epoch": 120.19, |
|
"learning_rate": 4.569117647058824e-06, |
|
"loss": 0.3669, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 120.19, |
|
"eval_loss": 0.2964297831058502, |
|
"eval_runtime": 64.5844, |
|
"eval_samples_per_second": 17.218, |
|
"eval_steps_per_second": 4.304, |
|
"eval_wer": 0.271390134529148, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 120.29, |
|
"learning_rate": 4.554411764705882e-06, |
|
"loss": 0.365, |
|
"step": 12510 |
|
}, |
|
{ |
|
"epoch": 120.38, |
|
"learning_rate": 4.5397058823529416e-06, |
|
"loss": 0.3466, |
|
"step": 12520 |
|
}, |
|
{ |
|
"epoch": 120.48, |
|
"learning_rate": 4.525000000000001e-06, |
|
"loss": 0.3471, |
|
"step": 12530 |
|
}, |
|
{ |
|
"epoch": 120.57, |
|
"learning_rate": 4.510294117647059e-06, |
|
"loss": 0.3426, |
|
"step": 12540 |
|
}, |
|
{ |
|
"epoch": 120.67, |
|
"learning_rate": 4.4970588235294125e-06, |
|
"loss": 0.3496, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 120.77, |
|
"learning_rate": 4.482352941176471e-06, |
|
"loss": 0.3476, |
|
"step": 12560 |
|
}, |
|
{ |
|
"epoch": 120.86, |
|
"learning_rate": 4.46764705882353e-06, |
|
"loss": 0.3531, |
|
"step": 12570 |
|
}, |
|
{ |
|
"epoch": 120.96, |
|
"learning_rate": 4.452941176470589e-06, |
|
"loss": 0.3809, |
|
"step": 12580 |
|
}, |
|
{ |
|
"epoch": 121.06, |
|
"learning_rate": 4.438235294117647e-06, |
|
"loss": 0.3662, |
|
"step": 12590 |
|
}, |
|
{ |
|
"epoch": 121.15, |
|
"learning_rate": 4.423529411764707e-06, |
|
"loss": 0.391, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 121.25, |
|
"learning_rate": 4.408823529411765e-06, |
|
"loss": 0.3381, |
|
"step": 12610 |
|
}, |
|
{ |
|
"epoch": 121.34, |
|
"learning_rate": 4.394117647058824e-06, |
|
"loss": 0.352, |
|
"step": 12620 |
|
}, |
|
{ |
|
"epoch": 121.44, |
|
"learning_rate": 4.379411764705882e-06, |
|
"loss": 0.3506, |
|
"step": 12630 |
|
}, |
|
{ |
|
"epoch": 121.54, |
|
"learning_rate": 4.3647058823529415e-06, |
|
"loss": 0.382, |
|
"step": 12640 |
|
}, |
|
{ |
|
"epoch": 121.63, |
|
"learning_rate": 4.350000000000001e-06, |
|
"loss": 0.341, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 121.73, |
|
"learning_rate": 4.335294117647059e-06, |
|
"loss": 0.319, |
|
"step": 12660 |
|
}, |
|
{ |
|
"epoch": 121.82, |
|
"learning_rate": 4.320588235294118e-06, |
|
"loss": 0.342, |
|
"step": 12670 |
|
}, |
|
{ |
|
"epoch": 121.92, |
|
"learning_rate": 4.305882352941177e-06, |
|
"loss": 0.3313, |
|
"step": 12680 |
|
}, |
|
{ |
|
"epoch": 122.02, |
|
"learning_rate": 4.291176470588236e-06, |
|
"loss": 0.3634, |
|
"step": 12690 |
|
}, |
|
{ |
|
"epoch": 122.11, |
|
"learning_rate": 4.276470588235295e-06, |
|
"loss": 0.3303, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 122.21, |
|
"learning_rate": 4.261764705882353e-06, |
|
"loss": 0.4079, |
|
"step": 12710 |
|
}, |
|
{ |
|
"epoch": 122.31, |
|
"learning_rate": 4.247058823529412e-06, |
|
"loss": 0.3451, |
|
"step": 12720 |
|
}, |
|
{ |
|
"epoch": 122.4, |
|
"learning_rate": 4.232352941176471e-06, |
|
"loss": 0.3871, |
|
"step": 12730 |
|
}, |
|
{ |
|
"epoch": 122.5, |
|
"learning_rate": 4.21764705882353e-06, |
|
"loss": 0.3697, |
|
"step": 12740 |
|
}, |
|
{ |
|
"epoch": 122.59, |
|
"learning_rate": 4.202941176470588e-06, |
|
"loss": 0.3089, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 122.69, |
|
"learning_rate": 4.188235294117647e-06, |
|
"loss": 0.3545, |
|
"step": 12760 |
|
}, |
|
{ |
|
"epoch": 122.78, |
|
"learning_rate": 4.1735294117647064e-06, |
|
"loss": 0.3156, |
|
"step": 12770 |
|
}, |
|
{ |
|
"epoch": 122.88, |
|
"learning_rate": 4.158823529411765e-06, |
|
"loss": 0.3652, |
|
"step": 12780 |
|
}, |
|
{ |
|
"epoch": 122.98, |
|
"learning_rate": 4.144117647058824e-06, |
|
"loss": 0.3508, |
|
"step": 12790 |
|
}, |
|
{ |
|
"epoch": 123.08, |
|
"learning_rate": 4.129411764705883e-06, |
|
"loss": 0.3586, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 123.17, |
|
"learning_rate": 4.114705882352941e-06, |
|
"loss": 0.3716, |
|
"step": 12810 |
|
}, |
|
{ |
|
"epoch": 123.27, |
|
"learning_rate": 4.1e-06, |
|
"loss": 0.3396, |
|
"step": 12820 |
|
}, |
|
{ |
|
"epoch": 123.36, |
|
"learning_rate": 4.085294117647059e-06, |
|
"loss": 0.3549, |
|
"step": 12830 |
|
}, |
|
{ |
|
"epoch": 123.46, |
|
"learning_rate": 4.070588235294118e-06, |
|
"loss": 0.3468, |
|
"step": 12840 |
|
}, |
|
{ |
|
"epoch": 123.56, |
|
"learning_rate": 4.055882352941176e-06, |
|
"loss": 0.3432, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 123.65, |
|
"learning_rate": 4.0411764705882355e-06, |
|
"loss": 0.3367, |
|
"step": 12860 |
|
}, |
|
{ |
|
"epoch": 123.75, |
|
"learning_rate": 4.026470588235295e-06, |
|
"loss": 0.365, |
|
"step": 12870 |
|
}, |
|
{ |
|
"epoch": 123.84, |
|
"learning_rate": 4.011764705882353e-06, |
|
"loss": 0.3475, |
|
"step": 12880 |
|
}, |
|
{ |
|
"epoch": 123.94, |
|
"learning_rate": 3.997058823529412e-06, |
|
"loss": 0.3304, |
|
"step": 12890 |
|
}, |
|
{ |
|
"epoch": 124.04, |
|
"learning_rate": 3.982352941176471e-06, |
|
"loss": 0.3666, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 124.13, |
|
"learning_rate": 3.96764705882353e-06, |
|
"loss": 0.3625, |
|
"step": 12910 |
|
}, |
|
{ |
|
"epoch": 124.23, |
|
"learning_rate": 3.952941176470588e-06, |
|
"loss": 0.3426, |
|
"step": 12920 |
|
}, |
|
{ |
|
"epoch": 124.33, |
|
"learning_rate": 3.938235294117647e-06, |
|
"loss": 0.3524, |
|
"step": 12930 |
|
}, |
|
{ |
|
"epoch": 124.42, |
|
"learning_rate": 3.923529411764706e-06, |
|
"loss": 0.3445, |
|
"step": 12940 |
|
}, |
|
{ |
|
"epoch": 124.52, |
|
"learning_rate": 3.908823529411765e-06, |
|
"loss": 0.3732, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 124.61, |
|
"learning_rate": 3.894117647058824e-06, |
|
"loss": 0.3256, |
|
"step": 12960 |
|
}, |
|
{ |
|
"epoch": 124.71, |
|
"learning_rate": 3.879411764705883e-06, |
|
"loss": 0.352, |
|
"step": 12970 |
|
}, |
|
{ |
|
"epoch": 124.8, |
|
"learning_rate": 3.864705882352941e-06, |
|
"loss": 0.3306, |
|
"step": 12980 |
|
}, |
|
{ |
|
"epoch": 124.9, |
|
"learning_rate": 3.85e-06, |
|
"loss": 0.3564, |
|
"step": 12990 |
|
}, |
|
{ |
|
"epoch": 125.0, |
|
"learning_rate": 3.8352941176470596e-06, |
|
"loss": 0.3386, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 125.0, |
|
"eval_loss": 0.310060054063797, |
|
"eval_runtime": 65.2154, |
|
"eval_samples_per_second": 17.051, |
|
"eval_steps_per_second": 4.263, |
|
"eval_wer": 0.271390134529148, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 125.1, |
|
"learning_rate": 3.820588235294118e-06, |
|
"loss": 0.3322, |
|
"step": 13010 |
|
}, |
|
{ |
|
"epoch": 125.19, |
|
"learning_rate": 3.805882352941177e-06, |
|
"loss": 0.3919, |
|
"step": 13020 |
|
}, |
|
{ |
|
"epoch": 125.29, |
|
"learning_rate": 3.7911764705882354e-06, |
|
"loss": 0.3185, |
|
"step": 13030 |
|
}, |
|
{ |
|
"epoch": 125.38, |
|
"learning_rate": 3.776470588235294e-06, |
|
"loss": 0.3441, |
|
"step": 13040 |
|
}, |
|
{ |
|
"epoch": 125.48, |
|
"learning_rate": 3.7617647058823533e-06, |
|
"loss": 0.3269, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 125.57, |
|
"learning_rate": 3.747058823529412e-06, |
|
"loss": 0.3218, |
|
"step": 13060 |
|
}, |
|
{ |
|
"epoch": 125.67, |
|
"learning_rate": 3.7323529411764708e-06, |
|
"loss": 0.3376, |
|
"step": 13070 |
|
}, |
|
{ |
|
"epoch": 125.77, |
|
"learning_rate": 3.71764705882353e-06, |
|
"loss": 0.3238, |
|
"step": 13080 |
|
}, |
|
{ |
|
"epoch": 125.86, |
|
"learning_rate": 3.7029411764705887e-06, |
|
"loss": 0.3323, |
|
"step": 13090 |
|
}, |
|
{ |
|
"epoch": 125.96, |
|
"learning_rate": 3.6882352941176474e-06, |
|
"loss": 0.3469, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 126.06, |
|
"learning_rate": 3.673529411764706e-06, |
|
"loss": 0.3556, |
|
"step": 13110 |
|
}, |
|
{ |
|
"epoch": 126.15, |
|
"learning_rate": 3.6588235294117653e-06, |
|
"loss": 0.3656, |
|
"step": 13120 |
|
}, |
|
{ |
|
"epoch": 126.25, |
|
"learning_rate": 3.644117647058824e-06, |
|
"loss": 0.3468, |
|
"step": 13130 |
|
}, |
|
{ |
|
"epoch": 126.34, |
|
"learning_rate": 3.6294117647058824e-06, |
|
"loss": 0.2991, |
|
"step": 13140 |
|
}, |
|
{ |
|
"epoch": 126.44, |
|
"learning_rate": 3.614705882352941e-06, |
|
"loss": 0.3644, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 126.54, |
|
"learning_rate": 3.6000000000000003e-06, |
|
"loss": 0.3496, |
|
"step": 13160 |
|
}, |
|
{ |
|
"epoch": 126.63, |
|
"learning_rate": 3.585294117647059e-06, |
|
"loss": 0.3297, |
|
"step": 13170 |
|
}, |
|
{ |
|
"epoch": 126.73, |
|
"learning_rate": 3.5705882352941177e-06, |
|
"loss": 0.3327, |
|
"step": 13180 |
|
}, |
|
{ |
|
"epoch": 126.82, |
|
"learning_rate": 3.555882352941177e-06, |
|
"loss": 0.3279, |
|
"step": 13190 |
|
}, |
|
{ |
|
"epoch": 126.92, |
|
"learning_rate": 3.5411764705882356e-06, |
|
"loss": 0.3287, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 127.02, |
|
"learning_rate": 3.5264705882352944e-06, |
|
"loss": 0.3508, |
|
"step": 13210 |
|
}, |
|
{ |
|
"epoch": 127.11, |
|
"learning_rate": 3.5117647058823535e-06, |
|
"loss": 0.3588, |
|
"step": 13220 |
|
}, |
|
{ |
|
"epoch": 127.21, |
|
"learning_rate": 3.4970588235294123e-06, |
|
"loss": 0.3058, |
|
"step": 13230 |
|
}, |
|
{ |
|
"epoch": 127.31, |
|
"learning_rate": 3.4823529411764706e-06, |
|
"loss": 0.324, |
|
"step": 13240 |
|
}, |
|
{ |
|
"epoch": 127.4, |
|
"learning_rate": 3.4676470588235293e-06, |
|
"loss": 0.3665, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 127.5, |
|
"learning_rate": 3.4529411764705885e-06, |
|
"loss": 0.337, |
|
"step": 13260 |
|
}, |
|
{ |
|
"epoch": 127.59, |
|
"learning_rate": 3.4382352941176472e-06, |
|
"loss": 0.3264, |
|
"step": 13270 |
|
}, |
|
{ |
|
"epoch": 127.69, |
|
"learning_rate": 3.423529411764706e-06, |
|
"loss": 0.3834, |
|
"step": 13280 |
|
}, |
|
{ |
|
"epoch": 127.78, |
|
"learning_rate": 3.408823529411765e-06, |
|
"loss": 0.3465, |
|
"step": 13290 |
|
}, |
|
{ |
|
"epoch": 127.88, |
|
"learning_rate": 3.394117647058824e-06, |
|
"loss": 0.3601, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 127.98, |
|
"learning_rate": 3.3794117647058826e-06, |
|
"loss": 0.3651, |
|
"step": 13310 |
|
}, |
|
{ |
|
"epoch": 128.08, |
|
"learning_rate": 3.364705882352942e-06, |
|
"loss": 0.3504, |
|
"step": 13320 |
|
}, |
|
{ |
|
"epoch": 128.17, |
|
"learning_rate": 3.3500000000000005e-06, |
|
"loss": 0.3526, |
|
"step": 13330 |
|
}, |
|
{ |
|
"epoch": 128.27, |
|
"learning_rate": 3.3352941176470593e-06, |
|
"loss": 0.317, |
|
"step": 13340 |
|
}, |
|
{ |
|
"epoch": 128.36, |
|
"learning_rate": 3.3205882352941176e-06, |
|
"loss": 0.3277, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 128.46, |
|
"learning_rate": 3.3058823529411763e-06, |
|
"loss": 0.3557, |
|
"step": 13360 |
|
}, |
|
{ |
|
"epoch": 128.56, |
|
"learning_rate": 3.2911764705882355e-06, |
|
"loss": 0.3032, |
|
"step": 13370 |
|
}, |
|
{ |
|
"epoch": 128.65, |
|
"learning_rate": 3.2764705882352942e-06, |
|
"loss": 0.3432, |
|
"step": 13380 |
|
}, |
|
{ |
|
"epoch": 128.75, |
|
"learning_rate": 3.261764705882353e-06, |
|
"loss": 0.2866, |
|
"step": 13390 |
|
}, |
|
{ |
|
"epoch": 128.84, |
|
"learning_rate": 3.247058823529412e-06, |
|
"loss": 0.3257, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 128.94, |
|
"learning_rate": 3.232352941176471e-06, |
|
"loss": 0.3889, |
|
"step": 13410 |
|
}, |
|
{ |
|
"epoch": 129.04, |
|
"learning_rate": 3.2176470588235296e-06, |
|
"loss": 0.3334, |
|
"step": 13420 |
|
}, |
|
{ |
|
"epoch": 129.13, |
|
"learning_rate": 3.2029411764705888e-06, |
|
"loss": 0.3346, |
|
"step": 13430 |
|
}, |
|
{ |
|
"epoch": 129.23, |
|
"learning_rate": 3.1882352941176475e-06, |
|
"loss": 0.3365, |
|
"step": 13440 |
|
}, |
|
{ |
|
"epoch": 129.33, |
|
"learning_rate": 3.1735294117647063e-06, |
|
"loss": 0.304, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 129.42, |
|
"learning_rate": 3.1588235294117646e-06, |
|
"loss": 0.3649, |
|
"step": 13460 |
|
}, |
|
{ |
|
"epoch": 129.52, |
|
"learning_rate": 3.1441176470588237e-06, |
|
"loss": 0.3079, |
|
"step": 13470 |
|
}, |
|
{ |
|
"epoch": 129.61, |
|
"learning_rate": 3.1294117647058825e-06, |
|
"loss": 0.3371, |
|
"step": 13480 |
|
}, |
|
{ |
|
"epoch": 129.71, |
|
"learning_rate": 3.1147058823529412e-06, |
|
"loss": 0.3513, |
|
"step": 13490 |
|
}, |
|
{ |
|
"epoch": 129.8, |
|
"learning_rate": 3.1000000000000004e-06, |
|
"loss": 0.3137, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 129.8, |
|
"eval_loss": 0.3063147962093353, |
|
"eval_runtime": 65.5511, |
|
"eval_samples_per_second": 16.964, |
|
"eval_steps_per_second": 4.241, |
|
"eval_wer": 0.27103139013452915, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 129.9, |
|
"learning_rate": 3.085294117647059e-06, |
|
"loss": 0.3268, |
|
"step": 13510 |
|
}, |
|
{ |
|
"epoch": 130.0, |
|
"learning_rate": 3.070588235294118e-06, |
|
"loss": 0.3603, |
|
"step": 13520 |
|
}, |
|
{ |
|
"epoch": 130.1, |
|
"learning_rate": 3.055882352941177e-06, |
|
"loss": 0.3136, |
|
"step": 13530 |
|
}, |
|
{ |
|
"epoch": 130.19, |
|
"learning_rate": 3.0411764705882358e-06, |
|
"loss": 0.3316, |
|
"step": 13540 |
|
}, |
|
{ |
|
"epoch": 130.29, |
|
"learning_rate": 3.0264705882352945e-06, |
|
"loss": 0.3011, |
|
"step": 13550 |
|
}, |
|
{ |
|
"epoch": 130.38, |
|
"learning_rate": 3.011764705882353e-06, |
|
"loss": 0.3355, |
|
"step": 13560 |
|
}, |
|
{ |
|
"epoch": 130.48, |
|
"learning_rate": 2.997058823529412e-06, |
|
"loss": 0.341, |
|
"step": 13570 |
|
}, |
|
{ |
|
"epoch": 130.57, |
|
"learning_rate": 2.9823529411764707e-06, |
|
"loss": 0.3351, |
|
"step": 13580 |
|
}, |
|
{ |
|
"epoch": 130.67, |
|
"learning_rate": 2.9676470588235295e-06, |
|
"loss": 0.3353, |
|
"step": 13590 |
|
}, |
|
{ |
|
"epoch": 130.77, |
|
"learning_rate": 2.9529411764705882e-06, |
|
"loss": 0.3284, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 130.86, |
|
"learning_rate": 2.9382352941176474e-06, |
|
"loss": 0.3309, |
|
"step": 13610 |
|
}, |
|
{ |
|
"epoch": 130.96, |
|
"learning_rate": 2.923529411764706e-06, |
|
"loss": 0.337, |
|
"step": 13620 |
|
}, |
|
{ |
|
"epoch": 131.06, |
|
"learning_rate": 2.908823529411765e-06, |
|
"loss": 0.3489, |
|
"step": 13630 |
|
}, |
|
{ |
|
"epoch": 131.15, |
|
"learning_rate": 2.894117647058824e-06, |
|
"loss": 0.3491, |
|
"step": 13640 |
|
}, |
|
{ |
|
"epoch": 131.25, |
|
"learning_rate": 2.8794117647058828e-06, |
|
"loss": 0.3387, |
|
"step": 13650 |
|
}, |
|
{ |
|
"epoch": 131.34, |
|
"learning_rate": 2.8647058823529415e-06, |
|
"loss": 0.3202, |
|
"step": 13660 |
|
}, |
|
{ |
|
"epoch": 131.44, |
|
"learning_rate": 2.85e-06, |
|
"loss": 0.3506, |
|
"step": 13670 |
|
}, |
|
{ |
|
"epoch": 131.54, |
|
"learning_rate": 2.835294117647059e-06, |
|
"loss": 0.3197, |
|
"step": 13680 |
|
}, |
|
{ |
|
"epoch": 131.63, |
|
"learning_rate": 2.8205882352941177e-06, |
|
"loss": 0.3392, |
|
"step": 13690 |
|
}, |
|
{ |
|
"epoch": 131.73, |
|
"learning_rate": 2.8058823529411765e-06, |
|
"loss": 0.3263, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 131.82, |
|
"learning_rate": 2.7911764705882356e-06, |
|
"loss": 0.3122, |
|
"step": 13710 |
|
}, |
|
{ |
|
"epoch": 131.92, |
|
"learning_rate": 2.7764705882352944e-06, |
|
"loss": 0.3591, |
|
"step": 13720 |
|
}, |
|
{ |
|
"epoch": 132.02, |
|
"learning_rate": 2.761764705882353e-06, |
|
"loss": 0.3493, |
|
"step": 13730 |
|
}, |
|
{ |
|
"epoch": 132.11, |
|
"learning_rate": 2.7470588235294123e-06, |
|
"loss": 0.3399, |
|
"step": 13740 |
|
}, |
|
{ |
|
"epoch": 132.21, |
|
"learning_rate": 2.732352941176471e-06, |
|
"loss": 0.3121, |
|
"step": 13750 |
|
}, |
|
{ |
|
"epoch": 132.31, |
|
"learning_rate": 2.7176470588235297e-06, |
|
"loss": 0.2937, |
|
"step": 13760 |
|
}, |
|
{ |
|
"epoch": 132.4, |
|
"learning_rate": 2.702941176470588e-06, |
|
"loss": 0.3122, |
|
"step": 13770 |
|
}, |
|
{ |
|
"epoch": 132.5, |
|
"learning_rate": 2.6882352941176472e-06, |
|
"loss": 0.319, |
|
"step": 13780 |
|
}, |
|
{ |
|
"epoch": 132.59, |
|
"learning_rate": 2.673529411764706e-06, |
|
"loss": 0.311, |
|
"step": 13790 |
|
}, |
|
{ |
|
"epoch": 132.69, |
|
"learning_rate": 2.6588235294117647e-06, |
|
"loss": 0.3367, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 132.78, |
|
"learning_rate": 2.644117647058824e-06, |
|
"loss": 0.3325, |
|
"step": 13810 |
|
}, |
|
{ |
|
"epoch": 132.88, |
|
"learning_rate": 2.6294117647058826e-06, |
|
"loss": 0.338, |
|
"step": 13820 |
|
}, |
|
{ |
|
"epoch": 132.98, |
|
"learning_rate": 2.6147058823529414e-06, |
|
"loss": 0.3145, |
|
"step": 13830 |
|
}, |
|
{ |
|
"epoch": 133.08, |
|
"learning_rate": 2.6e-06, |
|
"loss": 0.354, |
|
"step": 13840 |
|
}, |
|
{ |
|
"epoch": 133.17, |
|
"learning_rate": 2.5852941176470593e-06, |
|
"loss": 0.3233, |
|
"step": 13850 |
|
}, |
|
{ |
|
"epoch": 133.27, |
|
"learning_rate": 2.570588235294118e-06, |
|
"loss": 0.3667, |
|
"step": 13860 |
|
}, |
|
{ |
|
"epoch": 133.36, |
|
"learning_rate": 2.5558823529411767e-06, |
|
"loss": 0.3157, |
|
"step": 13870 |
|
}, |
|
{ |
|
"epoch": 133.46, |
|
"learning_rate": 2.541176470588235e-06, |
|
"loss": 0.3531, |
|
"step": 13880 |
|
}, |
|
{ |
|
"epoch": 133.56, |
|
"learning_rate": 2.5264705882352942e-06, |
|
"loss": 0.3649, |
|
"step": 13890 |
|
}, |
|
{ |
|
"epoch": 133.65, |
|
"learning_rate": 2.511764705882353e-06, |
|
"loss": 0.3376, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 133.75, |
|
"learning_rate": 2.4970588235294117e-06, |
|
"loss": 0.2959, |
|
"step": 13910 |
|
}, |
|
{ |
|
"epoch": 133.84, |
|
"learning_rate": 2.482352941176471e-06, |
|
"loss": 0.3378, |
|
"step": 13920 |
|
}, |
|
{ |
|
"epoch": 133.94, |
|
"learning_rate": 2.4676470588235296e-06, |
|
"loss": 0.3322, |
|
"step": 13930 |
|
}, |
|
{ |
|
"epoch": 134.04, |
|
"learning_rate": 2.4529411764705883e-06, |
|
"loss": 0.3745, |
|
"step": 13940 |
|
}, |
|
{ |
|
"epoch": 134.13, |
|
"learning_rate": 2.4382352941176475e-06, |
|
"loss": 0.3805, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 134.23, |
|
"learning_rate": 2.423529411764706e-06, |
|
"loss": 0.3204, |
|
"step": 13960 |
|
}, |
|
{ |
|
"epoch": 134.33, |
|
"learning_rate": 2.408823529411765e-06, |
|
"loss": 0.308, |
|
"step": 13970 |
|
}, |
|
{ |
|
"epoch": 134.42, |
|
"learning_rate": 2.3941176470588237e-06, |
|
"loss": 0.3199, |
|
"step": 13980 |
|
}, |
|
{ |
|
"epoch": 134.52, |
|
"learning_rate": 2.3794117647058825e-06, |
|
"loss": 0.3486, |
|
"step": 13990 |
|
}, |
|
{ |
|
"epoch": 134.61, |
|
"learning_rate": 2.3647058823529416e-06, |
|
"loss": 0.3008, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 134.61, |
|
"eval_loss": 0.30818137526512146, |
|
"eval_runtime": 65.7676, |
|
"eval_samples_per_second": 16.908, |
|
"eval_steps_per_second": 4.227, |
|
"eval_wer": 0.26170403587443947, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 134.71, |
|
"learning_rate": 2.35e-06, |
|
"loss": 0.3423, |
|
"step": 14010 |
|
}, |
|
{ |
|
"epoch": 134.8, |
|
"learning_rate": 2.335294117647059e-06, |
|
"loss": 0.3087, |
|
"step": 14020 |
|
}, |
|
{ |
|
"epoch": 134.9, |
|
"learning_rate": 2.320588235294118e-06, |
|
"loss": 0.3474, |
|
"step": 14030 |
|
}, |
|
{ |
|
"epoch": 135.0, |
|
"learning_rate": 2.3058823529411766e-06, |
|
"loss": 0.3414, |
|
"step": 14040 |
|
}, |
|
{ |
|
"epoch": 135.1, |
|
"learning_rate": 2.2911764705882358e-06, |
|
"loss": 0.2999, |
|
"step": 14050 |
|
}, |
|
{ |
|
"epoch": 135.19, |
|
"learning_rate": 2.2764705882352945e-06, |
|
"loss": 0.3283, |
|
"step": 14060 |
|
}, |
|
{ |
|
"epoch": 135.29, |
|
"learning_rate": 2.2617647058823532e-06, |
|
"loss": 0.3161, |
|
"step": 14070 |
|
}, |
|
{ |
|
"epoch": 135.38, |
|
"learning_rate": 2.247058823529412e-06, |
|
"loss": 0.3178, |
|
"step": 14080 |
|
}, |
|
{ |
|
"epoch": 135.48, |
|
"learning_rate": 2.2323529411764707e-06, |
|
"loss": 0.3342, |
|
"step": 14090 |
|
}, |
|
{ |
|
"epoch": 135.57, |
|
"learning_rate": 2.2176470588235295e-06, |
|
"loss": 0.3041, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 135.67, |
|
"learning_rate": 2.2029411764705886e-06, |
|
"loss": 0.3358, |
|
"step": 14110 |
|
}, |
|
{ |
|
"epoch": 135.77, |
|
"learning_rate": 2.188235294117647e-06, |
|
"loss": 0.3114, |
|
"step": 14120 |
|
}, |
|
{ |
|
"epoch": 135.86, |
|
"learning_rate": 2.173529411764706e-06, |
|
"loss": 0.3276, |
|
"step": 14130 |
|
}, |
|
{ |
|
"epoch": 135.96, |
|
"learning_rate": 2.158823529411765e-06, |
|
"loss": 0.3367, |
|
"step": 14140 |
|
}, |
|
{ |
|
"epoch": 136.06, |
|
"learning_rate": 2.1441176470588236e-06, |
|
"loss": 0.3298, |
|
"step": 14150 |
|
}, |
|
{ |
|
"epoch": 136.15, |
|
"learning_rate": 2.1294117647058827e-06, |
|
"loss": 0.359, |
|
"step": 14160 |
|
}, |
|
{ |
|
"epoch": 136.25, |
|
"learning_rate": 2.114705882352941e-06, |
|
"loss": 0.3273, |
|
"step": 14170 |
|
}, |
|
{ |
|
"epoch": 136.34, |
|
"learning_rate": 2.1000000000000002e-06, |
|
"loss": 0.3087, |
|
"step": 14180 |
|
}, |
|
{ |
|
"epoch": 136.44, |
|
"learning_rate": 2.085294117647059e-06, |
|
"loss": 0.3321, |
|
"step": 14190 |
|
}, |
|
{ |
|
"epoch": 136.54, |
|
"learning_rate": 2.0705882352941177e-06, |
|
"loss": 0.3261, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 136.63, |
|
"learning_rate": 2.055882352941177e-06, |
|
"loss": 0.3243, |
|
"step": 14210 |
|
}, |
|
{ |
|
"epoch": 136.73, |
|
"learning_rate": 2.0411764705882356e-06, |
|
"loss": 0.328, |
|
"step": 14220 |
|
}, |
|
{ |
|
"epoch": 136.82, |
|
"learning_rate": 2.0264705882352943e-06, |
|
"loss": 0.285, |
|
"step": 14230 |
|
}, |
|
{ |
|
"epoch": 136.92, |
|
"learning_rate": 2.011764705882353e-06, |
|
"loss": 0.2971, |
|
"step": 14240 |
|
}, |
|
{ |
|
"epoch": 137.02, |
|
"learning_rate": 1.997058823529412e-06, |
|
"loss": 0.3618, |
|
"step": 14250 |
|
}, |
|
{ |
|
"epoch": 137.11, |
|
"learning_rate": 1.982352941176471e-06, |
|
"loss": 0.309, |
|
"step": 14260 |
|
}, |
|
{ |
|
"epoch": 137.21, |
|
"learning_rate": 1.9676470588235297e-06, |
|
"loss": 0.323, |
|
"step": 14270 |
|
}, |
|
{ |
|
"epoch": 137.31, |
|
"learning_rate": 1.9529411764705885e-06, |
|
"loss": 0.336, |
|
"step": 14280 |
|
}, |
|
{ |
|
"epoch": 137.4, |
|
"learning_rate": 1.938235294117647e-06, |
|
"loss": 0.3297, |
|
"step": 14290 |
|
}, |
|
{ |
|
"epoch": 137.5, |
|
"learning_rate": 1.923529411764706e-06, |
|
"loss": 0.3134, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 137.59, |
|
"learning_rate": 1.908823529411765e-06, |
|
"loss": 0.3063, |
|
"step": 14310 |
|
}, |
|
{ |
|
"epoch": 137.69, |
|
"learning_rate": 1.8941176470588239e-06, |
|
"loss": 0.3496, |
|
"step": 14320 |
|
}, |
|
{ |
|
"epoch": 137.78, |
|
"learning_rate": 1.8794117647058824e-06, |
|
"loss": 0.3091, |
|
"step": 14330 |
|
}, |
|
{ |
|
"epoch": 137.88, |
|
"learning_rate": 1.8647058823529413e-06, |
|
"loss": 0.3438, |
|
"step": 14340 |
|
}, |
|
{ |
|
"epoch": 137.98, |
|
"learning_rate": 1.85e-06, |
|
"loss": 0.3314, |
|
"step": 14350 |
|
}, |
|
{ |
|
"epoch": 138.08, |
|
"learning_rate": 1.835294117647059e-06, |
|
"loss": 0.295, |
|
"step": 14360 |
|
}, |
|
{ |
|
"epoch": 138.17, |
|
"learning_rate": 1.820588235294118e-06, |
|
"loss": 0.3285, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 138.27, |
|
"learning_rate": 1.8058823529411767e-06, |
|
"loss": 0.3014, |
|
"step": 14380 |
|
}, |
|
{ |
|
"epoch": 138.36, |
|
"learning_rate": 1.7911764705882352e-06, |
|
"loss": 0.3267, |
|
"step": 14390 |
|
}, |
|
{ |
|
"epoch": 138.46, |
|
"learning_rate": 1.7764705882352942e-06, |
|
"loss": 0.3617, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 138.56, |
|
"learning_rate": 1.7617647058823532e-06, |
|
"loss": 0.3259, |
|
"step": 14410 |
|
}, |
|
{ |
|
"epoch": 138.65, |
|
"learning_rate": 1.7470588235294119e-06, |
|
"loss": 0.3226, |
|
"step": 14420 |
|
}, |
|
{ |
|
"epoch": 138.75, |
|
"learning_rate": 1.7323529411764708e-06, |
|
"loss": 0.2984, |
|
"step": 14430 |
|
}, |
|
{ |
|
"epoch": 138.84, |
|
"learning_rate": 1.7176470588235294e-06, |
|
"loss": 0.3227, |
|
"step": 14440 |
|
}, |
|
{ |
|
"epoch": 138.94, |
|
"learning_rate": 1.7029411764705883e-06, |
|
"loss": 0.356, |
|
"step": 14450 |
|
}, |
|
{ |
|
"epoch": 139.04, |
|
"learning_rate": 1.6882352941176473e-06, |
|
"loss": 0.3408, |
|
"step": 14460 |
|
}, |
|
{ |
|
"epoch": 139.13, |
|
"learning_rate": 1.673529411764706e-06, |
|
"loss": 0.3239, |
|
"step": 14470 |
|
}, |
|
{ |
|
"epoch": 139.23, |
|
"learning_rate": 1.658823529411765e-06, |
|
"loss": 0.3402, |
|
"step": 14480 |
|
}, |
|
{ |
|
"epoch": 139.33, |
|
"learning_rate": 1.6441176470588235e-06, |
|
"loss": 0.3002, |
|
"step": 14490 |
|
}, |
|
{ |
|
"epoch": 139.42, |
|
"learning_rate": 1.6294117647058824e-06, |
|
"loss": 0.301, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 139.42, |
|
"eval_loss": 0.3120834231376648, |
|
"eval_runtime": 64.4568, |
|
"eval_samples_per_second": 17.252, |
|
"eval_steps_per_second": 4.313, |
|
"eval_wer": 0.262780269058296, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 139.52, |
|
"learning_rate": 1.6147058823529412e-06, |
|
"loss": 0.3277, |
|
"step": 14510 |
|
}, |
|
{ |
|
"epoch": 139.61, |
|
"learning_rate": 1.6000000000000001e-06, |
|
"loss": 0.3039, |
|
"step": 14520 |
|
}, |
|
{ |
|
"epoch": 139.71, |
|
"learning_rate": 1.585294117647059e-06, |
|
"loss": 0.31, |
|
"step": 14530 |
|
}, |
|
{ |
|
"epoch": 139.8, |
|
"learning_rate": 1.5705882352941178e-06, |
|
"loss": 0.2857, |
|
"step": 14540 |
|
}, |
|
{ |
|
"epoch": 139.9, |
|
"learning_rate": 1.5558823529411766e-06, |
|
"loss": 0.3276, |
|
"step": 14550 |
|
}, |
|
{ |
|
"epoch": 140.0, |
|
"learning_rate": 1.5411764705882353e-06, |
|
"loss": 0.3451, |
|
"step": 14560 |
|
}, |
|
{ |
|
"epoch": 140.1, |
|
"learning_rate": 1.5264705882352943e-06, |
|
"loss": 0.3222, |
|
"step": 14570 |
|
}, |
|
{ |
|
"epoch": 140.19, |
|
"learning_rate": 1.5117647058823532e-06, |
|
"loss": 0.3316, |
|
"step": 14580 |
|
}, |
|
{ |
|
"epoch": 140.29, |
|
"learning_rate": 1.497058823529412e-06, |
|
"loss": 0.2965, |
|
"step": 14590 |
|
}, |
|
{ |
|
"epoch": 140.38, |
|
"learning_rate": 1.4823529411764707e-06, |
|
"loss": 0.316, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 140.48, |
|
"learning_rate": 1.4676470588235294e-06, |
|
"loss": 0.3711, |
|
"step": 14610 |
|
}, |
|
{ |
|
"epoch": 140.57, |
|
"learning_rate": 1.4529411764705884e-06, |
|
"loss": 0.2935, |
|
"step": 14620 |
|
}, |
|
{ |
|
"epoch": 140.67, |
|
"learning_rate": 1.4382352941176471e-06, |
|
"loss": 0.3388, |
|
"step": 14630 |
|
}, |
|
{ |
|
"epoch": 140.77, |
|
"learning_rate": 1.423529411764706e-06, |
|
"loss": 0.3274, |
|
"step": 14640 |
|
}, |
|
{ |
|
"epoch": 140.86, |
|
"learning_rate": 1.4088235294117646e-06, |
|
"loss": 0.3003, |
|
"step": 14650 |
|
}, |
|
{ |
|
"epoch": 140.96, |
|
"learning_rate": 1.3941176470588236e-06, |
|
"loss": 0.3259, |
|
"step": 14660 |
|
}, |
|
{ |
|
"epoch": 141.06, |
|
"learning_rate": 1.3794117647058825e-06, |
|
"loss": 0.2998, |
|
"step": 14670 |
|
}, |
|
{ |
|
"epoch": 141.15, |
|
"learning_rate": 1.3647058823529413e-06, |
|
"loss": 0.3083, |
|
"step": 14680 |
|
}, |
|
{ |
|
"epoch": 141.25, |
|
"learning_rate": 1.3500000000000002e-06, |
|
"loss": 0.3153, |
|
"step": 14690 |
|
}, |
|
{ |
|
"epoch": 141.34, |
|
"learning_rate": 1.3352941176470587e-06, |
|
"loss": 0.3041, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 141.44, |
|
"learning_rate": 1.3205882352941177e-06, |
|
"loss": 0.3129, |
|
"step": 14710 |
|
}, |
|
{ |
|
"epoch": 141.54, |
|
"learning_rate": 1.3058823529411766e-06, |
|
"loss": 0.3093, |
|
"step": 14720 |
|
}, |
|
{ |
|
"epoch": 141.63, |
|
"learning_rate": 1.2911764705882354e-06, |
|
"loss": 0.3018, |
|
"step": 14730 |
|
}, |
|
{ |
|
"epoch": 141.73, |
|
"learning_rate": 1.2764705882352943e-06, |
|
"loss": 0.3081, |
|
"step": 14740 |
|
}, |
|
{ |
|
"epoch": 141.82, |
|
"learning_rate": 1.261764705882353e-06, |
|
"loss": 0.3192, |
|
"step": 14750 |
|
}, |
|
{ |
|
"epoch": 141.92, |
|
"learning_rate": 1.2470588235294118e-06, |
|
"loss": 0.3146, |
|
"step": 14760 |
|
}, |
|
{ |
|
"epoch": 142.02, |
|
"learning_rate": 1.2323529411764705e-06, |
|
"loss": 0.3772, |
|
"step": 14770 |
|
}, |
|
{ |
|
"epoch": 142.11, |
|
"learning_rate": 1.2176470588235295e-06, |
|
"loss": 0.3015, |
|
"step": 14780 |
|
}, |
|
{ |
|
"epoch": 142.21, |
|
"learning_rate": 1.2029411764705885e-06, |
|
"loss": 0.3446, |
|
"step": 14790 |
|
}, |
|
{ |
|
"epoch": 142.31, |
|
"learning_rate": 1.1882352941176472e-06, |
|
"loss": 0.3154, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 142.4, |
|
"learning_rate": 1.173529411764706e-06, |
|
"loss": 0.3526, |
|
"step": 14810 |
|
}, |
|
{ |
|
"epoch": 142.5, |
|
"learning_rate": 1.1588235294117647e-06, |
|
"loss": 0.2781, |
|
"step": 14820 |
|
}, |
|
{ |
|
"epoch": 142.59, |
|
"learning_rate": 1.1441176470588236e-06, |
|
"loss": 0.2924, |
|
"step": 14830 |
|
}, |
|
{ |
|
"epoch": 142.69, |
|
"learning_rate": 1.1294117647058826e-06, |
|
"loss": 0.3196, |
|
"step": 14840 |
|
}, |
|
{ |
|
"epoch": 142.78, |
|
"learning_rate": 1.1147058823529413e-06, |
|
"loss": 0.3258, |
|
"step": 14850 |
|
}, |
|
{ |
|
"epoch": 142.88, |
|
"learning_rate": 1.1e-06, |
|
"loss": 0.3164, |
|
"step": 14860 |
|
}, |
|
{ |
|
"epoch": 142.98, |
|
"learning_rate": 1.085294117647059e-06, |
|
"loss": 0.3528, |
|
"step": 14870 |
|
}, |
|
{ |
|
"epoch": 143.08, |
|
"learning_rate": 1.0705882352941177e-06, |
|
"loss": 0.3044, |
|
"step": 14880 |
|
}, |
|
{ |
|
"epoch": 143.17, |
|
"learning_rate": 1.0558823529411765e-06, |
|
"loss": 0.3029, |
|
"step": 14890 |
|
}, |
|
{ |
|
"epoch": 143.27, |
|
"learning_rate": 1.0411764705882352e-06, |
|
"loss": 0.355, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 143.36, |
|
"learning_rate": 1.0264705882352942e-06, |
|
"loss": 0.3405, |
|
"step": 14910 |
|
}, |
|
{ |
|
"epoch": 143.46, |
|
"learning_rate": 1.0117647058823531e-06, |
|
"loss": 0.3524, |
|
"step": 14920 |
|
}, |
|
{ |
|
"epoch": 143.56, |
|
"learning_rate": 9.970588235294119e-07, |
|
"loss": 0.3093, |
|
"step": 14930 |
|
}, |
|
{ |
|
"epoch": 143.65, |
|
"learning_rate": 9.823529411764706e-07, |
|
"loss": 0.3122, |
|
"step": 14940 |
|
}, |
|
{ |
|
"epoch": 143.75, |
|
"learning_rate": 9.676470588235296e-07, |
|
"loss": 0.3241, |
|
"step": 14950 |
|
}, |
|
{ |
|
"epoch": 143.84, |
|
"learning_rate": 9.529411764705883e-07, |
|
"loss": 0.3349, |
|
"step": 14960 |
|
}, |
|
{ |
|
"epoch": 143.94, |
|
"learning_rate": 9.382352941176472e-07, |
|
"loss": 0.2934, |
|
"step": 14970 |
|
}, |
|
{ |
|
"epoch": 144.04, |
|
"learning_rate": 9.235294117647059e-07, |
|
"loss": 0.2995, |
|
"step": 14980 |
|
}, |
|
{ |
|
"epoch": 144.13, |
|
"learning_rate": 9.088235294117647e-07, |
|
"loss": 0.2874, |
|
"step": 14990 |
|
}, |
|
{ |
|
"epoch": 144.23, |
|
"learning_rate": 8.941176470588237e-07, |
|
"loss": 0.3291, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 144.23, |
|
"eval_loss": 0.3104746639728546, |
|
"eval_runtime": 64.9185, |
|
"eval_samples_per_second": 17.129, |
|
"eval_steps_per_second": 4.282, |
|
"eval_wer": 0.2611659192825112, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 144.33, |
|
"learning_rate": 8.794117647058824e-07, |
|
"loss": 0.2948, |
|
"step": 15010 |
|
}, |
|
{ |
|
"epoch": 144.42, |
|
"learning_rate": 8.647058823529413e-07, |
|
"loss": 0.3269, |
|
"step": 15020 |
|
}, |
|
{ |
|
"epoch": 144.52, |
|
"learning_rate": 8.500000000000001e-07, |
|
"loss": 0.3024, |
|
"step": 15030 |
|
}, |
|
{ |
|
"epoch": 144.61, |
|
"learning_rate": 8.352941176470589e-07, |
|
"loss": 0.3189, |
|
"step": 15040 |
|
}, |
|
{ |
|
"epoch": 144.71, |
|
"learning_rate": 8.205882352941177e-07, |
|
"loss": 0.3128, |
|
"step": 15050 |
|
}, |
|
{ |
|
"epoch": 144.8, |
|
"learning_rate": 8.058823529411764e-07, |
|
"loss": 0.283, |
|
"step": 15060 |
|
}, |
|
{ |
|
"epoch": 144.9, |
|
"learning_rate": 7.911764705882354e-07, |
|
"loss": 0.3089, |
|
"step": 15070 |
|
}, |
|
{ |
|
"epoch": 145.0, |
|
"learning_rate": 7.764705882352942e-07, |
|
"loss": 0.3062, |
|
"step": 15080 |
|
}, |
|
{ |
|
"epoch": 145.1, |
|
"learning_rate": 7.61764705882353e-07, |
|
"loss": 0.3043, |
|
"step": 15090 |
|
}, |
|
{ |
|
"epoch": 145.19, |
|
"learning_rate": 7.470588235294118e-07, |
|
"loss": 0.3597, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 145.29, |
|
"learning_rate": 7.323529411764707e-07, |
|
"loss": 0.3132, |
|
"step": 15110 |
|
}, |
|
{ |
|
"epoch": 145.38, |
|
"learning_rate": 7.176470588235294e-07, |
|
"loss": 0.3368, |
|
"step": 15120 |
|
}, |
|
{ |
|
"epoch": 145.48, |
|
"learning_rate": 7.029411764705884e-07, |
|
"loss": 0.3084, |
|
"step": 15130 |
|
}, |
|
{ |
|
"epoch": 145.57, |
|
"learning_rate": 6.882352941176471e-07, |
|
"loss": 0.3151, |
|
"step": 15140 |
|
}, |
|
{ |
|
"epoch": 145.67, |
|
"learning_rate": 6.73529411764706e-07, |
|
"loss": 0.3517, |
|
"step": 15150 |
|
}, |
|
{ |
|
"epoch": 145.77, |
|
"learning_rate": 6.588235294117648e-07, |
|
"loss": 0.2851, |
|
"step": 15160 |
|
}, |
|
{ |
|
"epoch": 145.86, |
|
"learning_rate": 6.441176470588235e-07, |
|
"loss": 0.2956, |
|
"step": 15170 |
|
}, |
|
{ |
|
"epoch": 145.96, |
|
"learning_rate": 6.294117647058824e-07, |
|
"loss": 0.3168, |
|
"step": 15180 |
|
}, |
|
{ |
|
"epoch": 146.06, |
|
"learning_rate": 6.147058823529412e-07, |
|
"loss": 0.3001, |
|
"step": 15190 |
|
}, |
|
{ |
|
"epoch": 146.15, |
|
"learning_rate": 6.000000000000001e-07, |
|
"loss": 0.3373, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 146.25, |
|
"learning_rate": 5.852941176470589e-07, |
|
"loss": 0.3195, |
|
"step": 15210 |
|
}, |
|
{ |
|
"epoch": 146.34, |
|
"learning_rate": 5.705882352941177e-07, |
|
"loss": 0.267, |
|
"step": 15220 |
|
}, |
|
{ |
|
"epoch": 146.44, |
|
"learning_rate": 5.558823529411765e-07, |
|
"loss": 0.3197, |
|
"step": 15230 |
|
}, |
|
{ |
|
"epoch": 146.54, |
|
"learning_rate": 5.411764705882354e-07, |
|
"loss": 0.2973, |
|
"step": 15240 |
|
}, |
|
{ |
|
"epoch": 146.63, |
|
"learning_rate": 5.264705882352942e-07, |
|
"loss": 0.3385, |
|
"step": 15250 |
|
}, |
|
{ |
|
"epoch": 146.73, |
|
"learning_rate": 5.11764705882353e-07, |
|
"loss": 0.2951, |
|
"step": 15260 |
|
}, |
|
{ |
|
"epoch": 146.82, |
|
"learning_rate": 4.970588235294118e-07, |
|
"loss": 0.3103, |
|
"step": 15270 |
|
}, |
|
{ |
|
"epoch": 146.92, |
|
"learning_rate": 4.823529411764706e-07, |
|
"loss": 0.3111, |
|
"step": 15280 |
|
}, |
|
{ |
|
"epoch": 147.02, |
|
"learning_rate": 4.676470588235295e-07, |
|
"loss": 0.3462, |
|
"step": 15290 |
|
}, |
|
{ |
|
"epoch": 147.11, |
|
"learning_rate": 4.529411764705883e-07, |
|
"loss": 0.2961, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 147.21, |
|
"learning_rate": 4.3823529411764707e-07, |
|
"loss": 0.2904, |
|
"step": 15310 |
|
}, |
|
{ |
|
"epoch": 147.31, |
|
"learning_rate": 4.235294117647059e-07, |
|
"loss": 0.3077, |
|
"step": 15320 |
|
}, |
|
{ |
|
"epoch": 147.4, |
|
"learning_rate": 4.0882352941176476e-07, |
|
"loss": 0.3472, |
|
"step": 15330 |
|
}, |
|
{ |
|
"epoch": 147.5, |
|
"learning_rate": 3.9411764705882355e-07, |
|
"loss": 0.3002, |
|
"step": 15340 |
|
}, |
|
{ |
|
"epoch": 147.59, |
|
"learning_rate": 3.794117647058824e-07, |
|
"loss": 0.3332, |
|
"step": 15350 |
|
}, |
|
{ |
|
"epoch": 147.69, |
|
"learning_rate": 3.647058823529412e-07, |
|
"loss": 0.2842, |
|
"step": 15360 |
|
}, |
|
{ |
|
"epoch": 147.78, |
|
"learning_rate": 3.5000000000000004e-07, |
|
"loss": 0.2949, |
|
"step": 15370 |
|
}, |
|
{ |
|
"epoch": 147.88, |
|
"learning_rate": 3.36764705882353e-07, |
|
"loss": 0.2753, |
|
"step": 15380 |
|
}, |
|
{ |
|
"epoch": 147.98, |
|
"learning_rate": 3.2205882352941177e-07, |
|
"loss": 0.3131, |
|
"step": 15390 |
|
}, |
|
{ |
|
"epoch": 148.08, |
|
"learning_rate": 3.073529411764706e-07, |
|
"loss": 0.3028, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 148.17, |
|
"learning_rate": 2.9264705882352946e-07, |
|
"loss": 0.328, |
|
"step": 15410 |
|
}, |
|
{ |
|
"epoch": 148.27, |
|
"learning_rate": 2.7794117647058826e-07, |
|
"loss": 0.3404, |
|
"step": 15420 |
|
}, |
|
{ |
|
"epoch": 148.36, |
|
"learning_rate": 2.632352941176471e-07, |
|
"loss": 0.3033, |
|
"step": 15430 |
|
}, |
|
{ |
|
"epoch": 148.46, |
|
"learning_rate": 2.485294117647059e-07, |
|
"loss": 0.2905, |
|
"step": 15440 |
|
}, |
|
{ |
|
"epoch": 148.56, |
|
"learning_rate": 2.3382352941176474e-07, |
|
"loss": 0.2925, |
|
"step": 15450 |
|
}, |
|
{ |
|
"epoch": 148.65, |
|
"learning_rate": 2.1911764705882353e-07, |
|
"loss": 0.3263, |
|
"step": 15460 |
|
}, |
|
{ |
|
"epoch": 148.75, |
|
"learning_rate": 2.0441176470588238e-07, |
|
"loss": 0.3088, |
|
"step": 15470 |
|
}, |
|
{ |
|
"epoch": 148.84, |
|
"learning_rate": 1.897058823529412e-07, |
|
"loss": 0.33, |
|
"step": 15480 |
|
}, |
|
{ |
|
"epoch": 148.94, |
|
"learning_rate": 1.7500000000000002e-07, |
|
"loss": 0.3084, |
|
"step": 15490 |
|
}, |
|
{ |
|
"epoch": 149.04, |
|
"learning_rate": 1.6029411764705884e-07, |
|
"loss": 0.3133, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 149.04, |
|
"eval_loss": 0.31136423349380493, |
|
"eval_runtime": 65.8551, |
|
"eval_samples_per_second": 16.886, |
|
"eval_steps_per_second": 4.221, |
|
"eval_wer": 0.26242152466367713, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 149.13, |
|
"learning_rate": 1.4558823529411766e-07, |
|
"loss": 0.3092, |
|
"step": 15510 |
|
}, |
|
{ |
|
"epoch": 149.23, |
|
"learning_rate": 1.3088235294117648e-07, |
|
"loss": 0.3334, |
|
"step": 15520 |
|
}, |
|
{ |
|
"epoch": 149.33, |
|
"learning_rate": 1.1617647058823531e-07, |
|
"loss": 0.2736, |
|
"step": 15530 |
|
}, |
|
{ |
|
"epoch": 149.42, |
|
"learning_rate": 1.0147058823529413e-07, |
|
"loss": 0.3421, |
|
"step": 15540 |
|
}, |
|
{ |
|
"epoch": 149.52, |
|
"learning_rate": 8.676470588235295e-08, |
|
"loss": 0.3017, |
|
"step": 15550 |
|
}, |
|
{ |
|
"epoch": 149.61, |
|
"learning_rate": 7.205882352941177e-08, |
|
"loss": 0.3209, |
|
"step": 15560 |
|
}, |
|
{ |
|
"epoch": 149.71, |
|
"learning_rate": 5.7352941176470595e-08, |
|
"loss": 0.3358, |
|
"step": 15570 |
|
}, |
|
{ |
|
"epoch": 149.8, |
|
"learning_rate": 4.2647058823529414e-08, |
|
"loss": 0.2969, |
|
"step": 15580 |
|
}, |
|
{ |
|
"epoch": 149.9, |
|
"learning_rate": 2.794117647058824e-08, |
|
"loss": 0.2889, |
|
"step": 15590 |
|
}, |
|
{ |
|
"epoch": 150.0, |
|
"learning_rate": 1.323529411764706e-08, |
|
"loss": 0.2939, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 150.0, |
|
"step": 15600, |
|
"total_flos": 2.1982408826331945e+20, |
|
"train_loss": 0.819076099395752, |
|
"train_runtime": 49902.7564, |
|
"train_samples_per_second": 10.028, |
|
"train_steps_per_second": 0.313 |
|
} |
|
], |
|
"max_steps": 15600, |
|
"num_train_epochs": 150, |
|
"total_flos": 2.1982408826331945e+20, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|