{ "best_metric": null, "best_model_checkpoint": null, "epoch": 100.0, "global_step": 6900, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.14, "learning_rate": 9e-07, "loss": 11.2286, "step": 10 }, { "epoch": 0.29, "learning_rate": 1.8e-06, "loss": 12.1103, "step": 20 }, { "epoch": 0.43, "learning_rate": 2.8000000000000003e-06, "loss": 11.9593, "step": 30 }, { "epoch": 0.58, "learning_rate": 3.8e-06, "loss": 11.6711, "step": 40 }, { "epoch": 0.72, "learning_rate": 4.800000000000001e-06, "loss": 11.9783, "step": 50 }, { "epoch": 0.87, "learning_rate": 5.8e-06, "loss": 10.9667, "step": 60 }, { "epoch": 1.01, "learning_rate": 6.800000000000001e-06, "loss": 12.1446, "step": 70 }, { "epoch": 1.16, "learning_rate": 7.8e-06, "loss": 10.2963, "step": 80 }, { "epoch": 1.3, "learning_rate": 8.7e-06, "loss": 10.8899, "step": 90 }, { "epoch": 1.45, "learning_rate": 9.7e-06, "loss": 9.5968, "step": 100 }, { "epoch": 1.59, "learning_rate": 1.0700000000000001e-05, "loss": 7.6591, "step": 110 }, { "epoch": 1.74, "learning_rate": 1.1700000000000001e-05, "loss": 7.5159, "step": 120 }, { "epoch": 1.88, "learning_rate": 1.27e-05, "loss": 6.2925, "step": 130 }, { "epoch": 2.03, "learning_rate": 1.3700000000000001e-05, "loss": 5.7639, "step": 140 }, { "epoch": 2.17, "learning_rate": 1.47e-05, "loss": 5.6696, "step": 150 }, { "epoch": 2.32, "learning_rate": 1.5700000000000002e-05, "loss": 4.7696, "step": 160 }, { "epoch": 2.46, "learning_rate": 1.6700000000000003e-05, "loss": 5.1383, "step": 170 }, { "epoch": 2.61, "learning_rate": 1.77e-05, "loss": 4.5714, "step": 180 }, { "epoch": 2.75, "learning_rate": 1.87e-05, "loss": 4.5116, "step": 190 }, { "epoch": 2.9, "learning_rate": 1.97e-05, "loss": 4.3733, "step": 200 }, { "epoch": 3.04, "learning_rate": 2.07e-05, "loss": 3.8443, "step": 210 }, { "epoch": 3.19, "learning_rate": 2.1700000000000002e-05, "loss": 3.9388, "step": 220 }, { "epoch": 3.33, "learning_rate": 2.2700000000000003e-05, "loss": 3.8416, "step": 230 }, { "epoch": 3.48, "learning_rate": 2.37e-05, "loss": 3.7135, "step": 240 }, { "epoch": 3.62, "learning_rate": 2.47e-05, "loss": 3.5271, "step": 250 }, { "epoch": 3.77, "learning_rate": 2.57e-05, "loss": 3.5189, "step": 260 }, { "epoch": 3.91, "learning_rate": 2.6700000000000002e-05, "loss": 3.5816, "step": 270 }, { "epoch": 4.06, "learning_rate": 2.7700000000000002e-05, "loss": 3.3417, "step": 280 }, { "epoch": 4.2, "learning_rate": 2.87e-05, "loss": 3.5864, "step": 290 }, { "epoch": 4.35, "learning_rate": 2.97e-05, "loss": 3.327, "step": 300 }, { "epoch": 4.49, "learning_rate": 3.07e-05, "loss": 3.205, "step": 310 }, { "epoch": 4.64, "learning_rate": 3.1700000000000005e-05, "loss": 3.1087, "step": 320 }, { "epoch": 4.78, "learning_rate": 3.27e-05, "loss": 3.1126, "step": 330 }, { "epoch": 4.93, "learning_rate": 3.3700000000000006e-05, "loss": 3.1934, "step": 340 }, { "epoch": 5.07, "learning_rate": 3.4699999999999996e-05, "loss": 3.0881, "step": 350 }, { "epoch": 5.22, "learning_rate": 3.57e-05, "loss": 3.086, "step": 360 }, { "epoch": 5.36, "learning_rate": 3.6700000000000004e-05, "loss": 3.0304, "step": 370 }, { "epoch": 5.51, "learning_rate": 3.77e-05, "loss": 3.0262, "step": 380 }, { "epoch": 5.65, "learning_rate": 3.8700000000000006e-05, "loss": 3.0492, "step": 390 }, { "epoch": 5.8, "learning_rate": 3.97e-05, "loss": 3.0657, "step": 400 }, { "epoch": 5.94, "learning_rate": 4.07e-05, "loss": 2.9912, "step": 410 }, { "epoch": 6.09, "learning_rate": 4.17e-05, "loss": 2.9993, "step": 420 }, { "epoch": 6.23, "learning_rate": 4.27e-05, "loss": 3.0465, "step": 430 }, { "epoch": 6.38, "learning_rate": 4.3700000000000005e-05, "loss": 2.9501, "step": 440 }, { "epoch": 6.52, "learning_rate": 4.47e-05, "loss": 3.0918, "step": 450 }, { "epoch": 6.67, "learning_rate": 4.5700000000000006e-05, "loss": 3.0306, "step": 460 }, { "epoch": 6.81, "learning_rate": 4.6700000000000003e-05, "loss": 2.9349, "step": 470 }, { "epoch": 6.96, "learning_rate": 4.77e-05, "loss": 3.0044, "step": 480 }, { "epoch": 7.1, "learning_rate": 4.87e-05, "loss": 2.9541, "step": 490 }, { "epoch": 7.25, "learning_rate": 4.97e-05, "loss": 3.0132, "step": 500 }, { "epoch": 7.25, "eval_loss": 2.9393208026885986, "eval_runtime": 580.9987, "eval_samples_per_second": 5.842, "eval_steps_per_second": 0.731, "eval_wer": 1.0, "step": 500 }, { "epoch": 7.39, "learning_rate": 5.0700000000000006e-05, "loss": 2.9162, "step": 510 }, { "epoch": 7.54, "learning_rate": 5.17e-05, "loss": 2.9309, "step": 520 }, { "epoch": 7.68, "learning_rate": 5.270000000000001e-05, "loss": 3.0513, "step": 530 }, { "epoch": 7.83, "learning_rate": 5.3700000000000004e-05, "loss": 2.948, "step": 540 }, { "epoch": 7.97, "learning_rate": 5.470000000000001e-05, "loss": 2.9518, "step": 550 }, { "epoch": 8.12, "learning_rate": 5.5700000000000005e-05, "loss": 2.9229, "step": 560 }, { "epoch": 8.26, "learning_rate": 5.6699999999999996e-05, "loss": 2.9511, "step": 570 }, { "epoch": 8.41, "learning_rate": 5.77e-05, "loss": 2.9427, "step": 580 }, { "epoch": 8.55, "learning_rate": 5.87e-05, "loss": 2.9603, "step": 590 }, { "epoch": 8.7, "learning_rate": 5.97e-05, "loss": 2.9397, "step": 600 }, { "epoch": 8.84, "learning_rate": 6.07e-05, "loss": 2.9267, "step": 610 }, { "epoch": 8.99, "learning_rate": 6.170000000000001e-05, "loss": 3.0164, "step": 620 }, { "epoch": 9.13, "learning_rate": 6.27e-05, "loss": 2.913, "step": 630 }, { "epoch": 9.28, "learning_rate": 6.37e-05, "loss": 2.9382, "step": 640 }, { "epoch": 9.42, "learning_rate": 6.47e-05, "loss": 2.9905, "step": 650 }, { "epoch": 9.57, "learning_rate": 6.570000000000001e-05, "loss": 2.8993, "step": 660 }, { "epoch": 9.71, "learning_rate": 6.670000000000001e-05, "loss": 2.9829, "step": 670 }, { "epoch": 9.86, "learning_rate": 6.77e-05, "loss": 2.9161, "step": 680 }, { "epoch": 10.0, "learning_rate": 6.87e-05, "loss": 3.0016, "step": 690 }, { "epoch": 10.14, "learning_rate": 6.97e-05, "loss": 2.9328, "step": 700 }, { "epoch": 10.29, "learning_rate": 7.07e-05, "loss": 3.0272, "step": 710 }, { "epoch": 10.43, "learning_rate": 7.17e-05, "loss": 2.9243, "step": 720 }, { "epoch": 10.58, "learning_rate": 7.27e-05, "loss": 2.9345, "step": 730 }, { "epoch": 10.72, "learning_rate": 7.37e-05, "loss": 2.9191, "step": 740 }, { "epoch": 10.87, "learning_rate": 7.47e-05, "loss": 2.8938, "step": 750 }, { "epoch": 11.01, "learning_rate": 7.570000000000001e-05, "loss": 2.9083, "step": 760 }, { "epoch": 11.16, "learning_rate": 7.670000000000001e-05, "loss": 2.9408, "step": 770 }, { "epoch": 11.3, "learning_rate": 7.77e-05, "loss": 2.9428, "step": 780 }, { "epoch": 11.45, "learning_rate": 7.87e-05, "loss": 2.9443, "step": 790 }, { "epoch": 11.59, "learning_rate": 7.970000000000001e-05, "loss": 2.9159, "step": 800 }, { "epoch": 11.74, "learning_rate": 8.070000000000001e-05, "loss": 2.9294, "step": 810 }, { "epoch": 11.88, "learning_rate": 8.17e-05, "loss": 2.8927, "step": 820 }, { "epoch": 12.03, "learning_rate": 8.27e-05, "loss": 2.9385, "step": 830 }, { "epoch": 12.17, "learning_rate": 8.37e-05, "loss": 2.9407, "step": 840 }, { "epoch": 12.32, "learning_rate": 8.47e-05, "loss": 2.8872, "step": 850 }, { "epoch": 12.46, "learning_rate": 8.57e-05, "loss": 2.9336, "step": 860 }, { "epoch": 12.61, "learning_rate": 8.67e-05, "loss": 2.8768, "step": 870 }, { "epoch": 12.75, "learning_rate": 8.77e-05, "loss": 2.9426, "step": 880 }, { "epoch": 12.9, "learning_rate": 8.87e-05, "loss": 2.8981, "step": 890 }, { "epoch": 13.04, "learning_rate": 8.970000000000001e-05, "loss": 2.8928, "step": 900 }, { "epoch": 13.19, "learning_rate": 9.070000000000001e-05, "loss": 2.9346, "step": 910 }, { "epoch": 13.33, "learning_rate": 9.17e-05, "loss": 2.89, "step": 920 }, { "epoch": 13.48, "learning_rate": 9.27e-05, "loss": 2.9107, "step": 930 }, { "epoch": 13.62, "learning_rate": 9.370000000000001e-05, "loss": 2.8732, "step": 940 }, { "epoch": 13.77, "learning_rate": 9.47e-05, "loss": 2.9441, "step": 950 }, { "epoch": 13.91, "learning_rate": 9.57e-05, "loss": 2.9136, "step": 960 }, { "epoch": 14.06, "learning_rate": 9.67e-05, "loss": 2.8773, "step": 970 }, { "epoch": 14.2, "learning_rate": 9.77e-05, "loss": 2.9015, "step": 980 }, { "epoch": 14.35, "learning_rate": 9.87e-05, "loss": 2.9168, "step": 990 }, { "epoch": 14.49, "learning_rate": 9.970000000000001e-05, "loss": 2.9241, "step": 1000 }, { "epoch": 14.49, "eval_loss": 2.8734302520751953, "eval_runtime": 570.6581, "eval_samples_per_second": 5.948, "eval_steps_per_second": 0.745, "eval_wer": 1.0, "step": 1000 }, { "epoch": 14.64, "learning_rate": 9.988135593220339e-05, "loss": 2.8693, "step": 1010 }, { "epoch": 14.78, "learning_rate": 9.971186440677967e-05, "loss": 2.9199, "step": 1020 }, { "epoch": 14.93, "learning_rate": 9.954237288135594e-05, "loss": 2.8831, "step": 1030 }, { "epoch": 15.07, "learning_rate": 9.937288135593222e-05, "loss": 2.8716, "step": 1040 }, { "epoch": 15.22, "learning_rate": 9.920338983050847e-05, "loss": 2.9457, "step": 1050 }, { "epoch": 15.36, "learning_rate": 9.903389830508475e-05, "loss": 2.8708, "step": 1060 }, { "epoch": 15.51, "learning_rate": 9.886440677966103e-05, "loss": 2.8595, "step": 1070 }, { "epoch": 15.65, "learning_rate": 9.86949152542373e-05, "loss": 2.8718, "step": 1080 }, { "epoch": 15.8, "learning_rate": 9.852542372881356e-05, "loss": 2.8383, "step": 1090 }, { "epoch": 15.94, "learning_rate": 9.835593220338983e-05, "loss": 2.8016, "step": 1100 }, { "epoch": 16.09, "learning_rate": 9.818644067796611e-05, "loss": 2.7608, "step": 1110 }, { "epoch": 16.23, "learning_rate": 9.801694915254239e-05, "loss": 2.7417, "step": 1120 }, { "epoch": 16.38, "learning_rate": 9.784745762711864e-05, "loss": 2.6906, "step": 1130 }, { "epoch": 16.52, "learning_rate": 9.767796610169492e-05, "loss": 2.5662, "step": 1140 }, { "epoch": 16.67, "learning_rate": 9.750847457627119e-05, "loss": 2.4202, "step": 1150 }, { "epoch": 16.81, "learning_rate": 9.733898305084747e-05, "loss": 2.4102, "step": 1160 }, { "epoch": 16.96, "learning_rate": 9.716949152542373e-05, "loss": 2.185, "step": 1170 }, { "epoch": 17.1, "learning_rate": 9.7e-05, "loss": 1.9315, "step": 1180 }, { "epoch": 17.25, "learning_rate": 9.683050847457628e-05, "loss": 2.0025, "step": 1190 }, { "epoch": 17.39, "learning_rate": 9.666101694915255e-05, "loss": 1.8797, "step": 1200 }, { "epoch": 17.54, "learning_rate": 9.649152542372883e-05, "loss": 1.7524, "step": 1210 }, { "epoch": 17.68, "learning_rate": 9.632203389830509e-05, "loss": 1.7243, "step": 1220 }, { "epoch": 17.83, "learning_rate": 9.615254237288136e-05, "loss": 1.5859, "step": 1230 }, { "epoch": 17.97, "learning_rate": 9.598305084745764e-05, "loss": 1.5059, "step": 1240 }, { "epoch": 18.12, "learning_rate": 9.58135593220339e-05, "loss": 1.456, "step": 1250 }, { "epoch": 18.26, "learning_rate": 9.564406779661017e-05, "loss": 1.535, "step": 1260 }, { "epoch": 18.41, "learning_rate": 9.547457627118644e-05, "loss": 1.4808, "step": 1270 }, { "epoch": 18.55, "learning_rate": 9.530508474576272e-05, "loss": 1.4568, "step": 1280 }, { "epoch": 18.7, "learning_rate": 9.5135593220339e-05, "loss": 1.3898, "step": 1290 }, { "epoch": 18.84, "learning_rate": 9.496610169491525e-05, "loss": 1.4865, "step": 1300 }, { "epoch": 18.99, "learning_rate": 9.479661016949153e-05, "loss": 1.391, "step": 1310 }, { "epoch": 19.13, "learning_rate": 9.46271186440678e-05, "loss": 1.4236, "step": 1320 }, { "epoch": 19.28, "learning_rate": 9.445762711864408e-05, "loss": 1.303, "step": 1330 }, { "epoch": 19.42, "learning_rate": 9.428813559322034e-05, "loss": 1.2304, "step": 1340 }, { "epoch": 19.57, "learning_rate": 9.411864406779661e-05, "loss": 1.2014, "step": 1350 }, { "epoch": 19.71, "learning_rate": 9.394915254237289e-05, "loss": 1.2398, "step": 1360 }, { "epoch": 19.86, "learning_rate": 9.377966101694916e-05, "loss": 1.1895, "step": 1370 }, { "epoch": 20.0, "learning_rate": 9.361016949152542e-05, "loss": 1.2296, "step": 1380 }, { "epoch": 20.14, "learning_rate": 9.34406779661017e-05, "loss": 1.3624, "step": 1390 }, { "epoch": 20.29, "learning_rate": 9.327118644067797e-05, "loss": 1.1805, "step": 1400 }, { "epoch": 20.43, "learning_rate": 9.310169491525425e-05, "loss": 1.1705, "step": 1410 }, { "epoch": 20.58, "learning_rate": 9.29322033898305e-05, "loss": 1.1446, "step": 1420 }, { "epoch": 20.72, "learning_rate": 9.276271186440678e-05, "loss": 1.2016, "step": 1430 }, { "epoch": 20.87, "learning_rate": 9.259322033898306e-05, "loss": 1.1053, "step": 1440 }, { "epoch": 21.01, "learning_rate": 9.242372881355933e-05, "loss": 1.1159, "step": 1450 }, { "epoch": 21.16, "learning_rate": 9.225423728813561e-05, "loss": 1.0629, "step": 1460 }, { "epoch": 21.3, "learning_rate": 9.208474576271186e-05, "loss": 1.1606, "step": 1470 }, { "epoch": 21.45, "learning_rate": 9.191525423728814e-05, "loss": 1.051, "step": 1480 }, { "epoch": 21.59, "learning_rate": 9.174576271186442e-05, "loss": 1.0948, "step": 1490 }, { "epoch": 21.74, "learning_rate": 9.157627118644069e-05, "loss": 1.0766, "step": 1500 }, { "epoch": 21.74, "eval_loss": 0.27734482288360596, "eval_runtime": 571.274, "eval_samples_per_second": 5.941, "eval_steps_per_second": 0.744, "eval_wer": 0.2488228188250376, "step": 1500 }, { "epoch": 21.88, "learning_rate": 9.140677966101695e-05, "loss": 0.9907, "step": 1510 }, { "epoch": 22.03, "learning_rate": 9.123728813559322e-05, "loss": 1.0756, "step": 1520 }, { "epoch": 22.17, "learning_rate": 9.10677966101695e-05, "loss": 1.0947, "step": 1530 }, { "epoch": 22.32, "learning_rate": 9.089830508474577e-05, "loss": 1.1119, "step": 1540 }, { "epoch": 22.46, "learning_rate": 9.072881355932203e-05, "loss": 1.039, "step": 1550 }, { "epoch": 22.61, "learning_rate": 9.055932203389831e-05, "loss": 0.9631, "step": 1560 }, { "epoch": 22.75, "learning_rate": 9.038983050847458e-05, "loss": 0.9765, "step": 1570 }, { "epoch": 22.9, "learning_rate": 9.022033898305086e-05, "loss": 0.9932, "step": 1580 }, { "epoch": 23.04, "learning_rate": 9.005084745762712e-05, "loss": 0.9384, "step": 1590 }, { "epoch": 23.19, "learning_rate": 8.988135593220339e-05, "loss": 0.9942, "step": 1600 }, { "epoch": 23.33, "learning_rate": 8.971186440677967e-05, "loss": 0.9546, "step": 1610 }, { "epoch": 23.48, "learning_rate": 8.954237288135594e-05, "loss": 1.0051, "step": 1620 }, { "epoch": 23.62, "learning_rate": 8.93728813559322e-05, "loss": 0.9648, "step": 1630 }, { "epoch": 23.77, "learning_rate": 8.920338983050848e-05, "loss": 1.0198, "step": 1640 }, { "epoch": 23.91, "learning_rate": 8.903389830508475e-05, "loss": 0.9773, "step": 1650 }, { "epoch": 24.06, "learning_rate": 8.886440677966103e-05, "loss": 0.9714, "step": 1660 }, { "epoch": 24.2, "learning_rate": 8.869491525423728e-05, "loss": 0.987, "step": 1670 }, { "epoch": 24.35, "learning_rate": 8.852542372881356e-05, "loss": 1.0105, "step": 1680 }, { "epoch": 24.49, "learning_rate": 8.835593220338983e-05, "loss": 0.9965, "step": 1690 }, { "epoch": 24.64, "learning_rate": 8.818644067796611e-05, "loss": 0.9293, "step": 1700 }, { "epoch": 24.78, "learning_rate": 8.801694915254238e-05, "loss": 0.8939, "step": 1710 }, { "epoch": 24.93, "learning_rate": 8.784745762711864e-05, "loss": 0.9721, "step": 1720 }, { "epoch": 25.07, "learning_rate": 8.767796610169492e-05, "loss": 1.0158, "step": 1730 }, { "epoch": 25.22, "learning_rate": 8.750847457627119e-05, "loss": 1.08, "step": 1740 }, { "epoch": 25.36, "learning_rate": 8.733898305084747e-05, "loss": 1.3222, "step": 1750 }, { "epoch": 25.51, "learning_rate": 8.716949152542373e-05, "loss": 0.9776, "step": 1760 }, { "epoch": 25.65, "learning_rate": 8.7e-05, "loss": 0.9089, "step": 1770 }, { "epoch": 25.8, "learning_rate": 8.683050847457628e-05, "loss": 0.9171, "step": 1780 }, { "epoch": 25.94, "learning_rate": 8.666101694915255e-05, "loss": 0.9366, "step": 1790 }, { "epoch": 26.09, "learning_rate": 8.649152542372881e-05, "loss": 0.8907, "step": 1800 }, { "epoch": 26.23, "learning_rate": 8.632203389830509e-05, "loss": 0.8427, "step": 1810 }, { "epoch": 26.38, "learning_rate": 8.615254237288136e-05, "loss": 0.8692, "step": 1820 }, { "epoch": 26.52, "learning_rate": 8.598305084745764e-05, "loss": 0.8933, "step": 1830 }, { "epoch": 26.67, "learning_rate": 8.581355932203389e-05, "loss": 0.9524, "step": 1840 }, { "epoch": 26.81, "learning_rate": 8.564406779661017e-05, "loss": 0.9623, "step": 1850 }, { "epoch": 26.96, "learning_rate": 8.547457627118645e-05, "loss": 0.9458, "step": 1860 }, { "epoch": 27.1, "learning_rate": 8.530508474576272e-05, "loss": 0.9299, "step": 1870 }, { "epoch": 27.25, "learning_rate": 8.513559322033898e-05, "loss": 0.8612, "step": 1880 }, { "epoch": 27.39, "learning_rate": 8.496610169491525e-05, "loss": 0.9352, "step": 1890 }, { "epoch": 27.54, "learning_rate": 8.479661016949153e-05, "loss": 0.8692, "step": 1900 }, { "epoch": 27.68, "learning_rate": 8.46271186440678e-05, "loss": 0.8834, "step": 1910 }, { "epoch": 27.83, "learning_rate": 8.445762711864406e-05, "loss": 0.8448, "step": 1920 }, { "epoch": 27.97, "learning_rate": 8.428813559322034e-05, "loss": 0.9122, "step": 1930 }, { "epoch": 28.12, "learning_rate": 8.411864406779661e-05, "loss": 0.8632, "step": 1940 }, { "epoch": 28.26, "learning_rate": 8.394915254237289e-05, "loss": 0.8578, "step": 1950 }, { "epoch": 28.41, "learning_rate": 8.377966101694916e-05, "loss": 0.8758, "step": 1960 }, { "epoch": 28.55, "learning_rate": 8.361016949152542e-05, "loss": 0.9024, "step": 1970 }, { "epoch": 28.7, "learning_rate": 8.34406779661017e-05, "loss": 1.1392, "step": 1980 }, { "epoch": 28.84, "learning_rate": 8.327118644067797e-05, "loss": 0.8794, "step": 1990 }, { "epoch": 28.99, "learning_rate": 8.310169491525425e-05, "loss": 0.8416, "step": 2000 }, { "epoch": 28.99, "eval_loss": 0.222377210855484, "eval_runtime": 567.7022, "eval_samples_per_second": 5.978, "eval_steps_per_second": 0.749, "eval_wer": 0.19897443482977098, "step": 2000 }, { "epoch": 29.13, "learning_rate": 8.293220338983052e-05, "loss": 0.8217, "step": 2010 }, { "epoch": 29.28, "learning_rate": 8.276271186440678e-05, "loss": 0.9189, "step": 2020 }, { "epoch": 29.42, "learning_rate": 8.259322033898306e-05, "loss": 0.8251, "step": 2030 }, { "epoch": 29.57, "learning_rate": 8.242372881355933e-05, "loss": 0.9254, "step": 2040 }, { "epoch": 29.71, "learning_rate": 8.22542372881356e-05, "loss": 0.9276, "step": 2050 }, { "epoch": 29.86, "learning_rate": 8.208474576271186e-05, "loss": 0.8893, "step": 2060 }, { "epoch": 30.0, "learning_rate": 8.191525423728814e-05, "loss": 0.7608, "step": 2070 }, { "epoch": 30.14, "learning_rate": 8.174576271186442e-05, "loss": 0.7857, "step": 2080 }, { "epoch": 30.29, "learning_rate": 8.157627118644067e-05, "loss": 0.9132, "step": 2090 }, { "epoch": 30.43, "learning_rate": 8.140677966101695e-05, "loss": 0.8814, "step": 2100 }, { "epoch": 30.58, "learning_rate": 8.123728813559322e-05, "loss": 0.901, "step": 2110 }, { "epoch": 30.72, "learning_rate": 8.10677966101695e-05, "loss": 0.8746, "step": 2120 }, { "epoch": 30.87, "learning_rate": 8.089830508474577e-05, "loss": 0.9291, "step": 2130 }, { "epoch": 31.01, "learning_rate": 8.072881355932203e-05, "loss": 0.8357, "step": 2140 }, { "epoch": 31.16, "learning_rate": 8.055932203389831e-05, "loss": 0.8305, "step": 2150 }, { "epoch": 31.3, "learning_rate": 8.038983050847458e-05, "loss": 0.8275, "step": 2160 }, { "epoch": 31.45, "learning_rate": 8.022033898305085e-05, "loss": 0.8111, "step": 2170 }, { "epoch": 31.59, "learning_rate": 8.005084745762713e-05, "loss": 0.8109, "step": 2180 }, { "epoch": 31.74, "learning_rate": 7.988135593220339e-05, "loss": 0.8761, "step": 2190 }, { "epoch": 31.88, "learning_rate": 7.971186440677967e-05, "loss": 0.8179, "step": 2200 }, { "epoch": 32.03, "learning_rate": 7.954237288135592e-05, "loss": 0.8078, "step": 2210 }, { "epoch": 32.17, "learning_rate": 7.93728813559322e-05, "loss": 0.8174, "step": 2220 }, { "epoch": 32.32, "learning_rate": 7.920338983050848e-05, "loss": 0.8176, "step": 2230 }, { "epoch": 32.46, "learning_rate": 7.903389830508475e-05, "loss": 0.729, "step": 2240 }, { "epoch": 32.61, "learning_rate": 7.886440677966102e-05, "loss": 0.7884, "step": 2250 }, { "epoch": 32.75, "learning_rate": 7.869491525423728e-05, "loss": 0.7896, "step": 2260 }, { "epoch": 32.9, "learning_rate": 7.852542372881356e-05, "loss": 0.8016, "step": 2270 }, { "epoch": 33.04, "learning_rate": 7.835593220338984e-05, "loss": 0.7635, "step": 2280 }, { "epoch": 33.19, "learning_rate": 7.818644067796611e-05, "loss": 0.8308, "step": 2290 }, { "epoch": 33.33, "learning_rate": 7.801694915254238e-05, "loss": 0.7543, "step": 2300 }, { "epoch": 33.48, "learning_rate": 7.784745762711864e-05, "loss": 0.8617, "step": 2310 }, { "epoch": 33.62, "learning_rate": 7.767796610169492e-05, "loss": 0.8434, "step": 2320 }, { "epoch": 33.77, "learning_rate": 7.750847457627119e-05, "loss": 0.948, "step": 2330 }, { "epoch": 33.91, "learning_rate": 7.733898305084746e-05, "loss": 1.0185, "step": 2340 }, { "epoch": 34.06, "learning_rate": 7.716949152542374e-05, "loss": 0.827, "step": 2350 }, { "epoch": 34.2, "learning_rate": 7.7e-05, "loss": 0.7916, "step": 2360 }, { "epoch": 34.35, "learning_rate": 7.683050847457628e-05, "loss": 0.7858, "step": 2370 }, { "epoch": 34.49, "learning_rate": 7.666101694915255e-05, "loss": 0.8583, "step": 2380 }, { "epoch": 34.64, "learning_rate": 7.649152542372881e-05, "loss": 0.7672, "step": 2390 }, { "epoch": 34.78, "learning_rate": 7.63220338983051e-05, "loss": 0.7745, "step": 2400 }, { "epoch": 34.93, "learning_rate": 7.615254237288136e-05, "loss": 0.8104, "step": 2410 }, { "epoch": 35.07, "learning_rate": 7.598305084745763e-05, "loss": 0.7519, "step": 2420 }, { "epoch": 35.22, "learning_rate": 7.58135593220339e-05, "loss": 0.7867, "step": 2430 }, { "epoch": 35.36, "learning_rate": 7.564406779661017e-05, "loss": 0.7614, "step": 2440 }, { "epoch": 35.51, "learning_rate": 7.547457627118645e-05, "loss": 0.8043, "step": 2450 }, { "epoch": 35.65, "learning_rate": 7.53050847457627e-05, "loss": 0.7866, "step": 2460 }, { "epoch": 35.8, "learning_rate": 7.513559322033899e-05, "loss": 0.8356, "step": 2470 }, { "epoch": 35.94, "learning_rate": 7.496610169491525e-05, "loss": 0.761, "step": 2480 }, { "epoch": 36.09, "learning_rate": 7.479661016949153e-05, "loss": 0.9027, "step": 2490 }, { "epoch": 36.23, "learning_rate": 7.46271186440678e-05, "loss": 0.8048, "step": 2500 }, { "epoch": 36.23, "eval_loss": 0.20626655220985413, "eval_runtime": 571.29, "eval_samples_per_second": 5.941, "eval_steps_per_second": 0.744, "eval_wer": 0.17916985101364954, "step": 2500 }, { "epoch": 36.38, "learning_rate": 7.445762711864406e-05, "loss": 0.7866, "step": 2510 }, { "epoch": 36.52, "learning_rate": 7.428813559322034e-05, "loss": 0.7869, "step": 2520 }, { "epoch": 36.67, "learning_rate": 7.411864406779661e-05, "loss": 0.8445, "step": 2530 }, { "epoch": 36.81, "learning_rate": 7.394915254237289e-05, "loss": 0.7745, "step": 2540 }, { "epoch": 36.96, "learning_rate": 7.377966101694916e-05, "loss": 0.7888, "step": 2550 }, { "epoch": 37.1, "learning_rate": 7.361016949152542e-05, "loss": 0.776, "step": 2560 }, { "epoch": 37.25, "learning_rate": 7.34406779661017e-05, "loss": 0.8096, "step": 2570 }, { "epoch": 37.39, "learning_rate": 7.327118644067797e-05, "loss": 0.7561, "step": 2580 }, { "epoch": 37.54, "learning_rate": 7.310169491525424e-05, "loss": 0.7251, "step": 2590 }, { "epoch": 37.68, "learning_rate": 7.293220338983052e-05, "loss": 0.8062, "step": 2600 }, { "epoch": 37.83, "learning_rate": 7.276271186440678e-05, "loss": 0.7388, "step": 2610 }, { "epoch": 37.97, "learning_rate": 7.259322033898306e-05, "loss": 0.7616, "step": 2620 }, { "epoch": 38.12, "learning_rate": 7.242372881355932e-05, "loss": 0.7593, "step": 2630 }, { "epoch": 38.26, "learning_rate": 7.22542372881356e-05, "loss": 0.7399, "step": 2640 }, { "epoch": 38.41, "learning_rate": 7.208474576271188e-05, "loss": 0.7559, "step": 2650 }, { "epoch": 38.55, "learning_rate": 7.191525423728814e-05, "loss": 0.7559, "step": 2660 }, { "epoch": 38.7, "learning_rate": 7.174576271186441e-05, "loss": 0.7218, "step": 2670 }, { "epoch": 38.84, "learning_rate": 7.157627118644067e-05, "loss": 0.7991, "step": 2680 }, { "epoch": 38.99, "learning_rate": 7.140677966101695e-05, "loss": 0.7493, "step": 2690 }, { "epoch": 39.13, "learning_rate": 7.123728813559322e-05, "loss": 0.7502, "step": 2700 }, { "epoch": 39.28, "learning_rate": 7.106779661016949e-05, "loss": 0.7883, "step": 2710 }, { "epoch": 39.42, "learning_rate": 7.089830508474577e-05, "loss": 0.8236, "step": 2720 }, { "epoch": 39.57, "learning_rate": 7.072881355932203e-05, "loss": 0.8141, "step": 2730 }, { "epoch": 39.71, "learning_rate": 7.055932203389831e-05, "loss": 0.7956, "step": 2740 }, { "epoch": 39.86, "learning_rate": 7.038983050847458e-05, "loss": 0.7714, "step": 2750 }, { "epoch": 40.0, "learning_rate": 7.022033898305085e-05, "loss": 0.6689, "step": 2760 }, { "epoch": 40.14, "learning_rate": 7.005084745762713e-05, "loss": 0.7202, "step": 2770 }, { "epoch": 40.29, "learning_rate": 6.988135593220339e-05, "loss": 0.7278, "step": 2780 }, { "epoch": 40.43, "learning_rate": 6.971186440677966e-05, "loss": 0.801, "step": 2790 }, { "epoch": 40.58, "learning_rate": 6.954237288135594e-05, "loss": 0.7865, "step": 2800 }, { "epoch": 40.72, "learning_rate": 6.93728813559322e-05, "loss": 0.6811, "step": 2810 }, { "epoch": 40.87, "learning_rate": 6.920338983050849e-05, "loss": 0.7933, "step": 2820 }, { "epoch": 41.01, "learning_rate": 6.903389830508475e-05, "loss": 0.7826, "step": 2830 }, { "epoch": 41.16, "learning_rate": 6.886440677966102e-05, "loss": 0.7142, "step": 2840 }, { "epoch": 41.3, "learning_rate": 6.869491525423728e-05, "loss": 0.8347, "step": 2850 }, { "epoch": 41.45, "learning_rate": 6.852542372881356e-05, "loss": 0.7712, "step": 2860 }, { "epoch": 41.59, "learning_rate": 6.835593220338984e-05, "loss": 0.7458, "step": 2870 }, { "epoch": 41.74, "learning_rate": 6.81864406779661e-05, "loss": 0.7528, "step": 2880 }, { "epoch": 41.88, "learning_rate": 6.801694915254238e-05, "loss": 0.7699, "step": 2890 }, { "epoch": 42.03, "learning_rate": 6.784745762711864e-05, "loss": 0.8577, "step": 2900 }, { "epoch": 42.17, "learning_rate": 6.767796610169492e-05, "loss": 0.7259, "step": 2910 }, { "epoch": 42.32, "learning_rate": 6.750847457627119e-05, "loss": 0.7039, "step": 2920 }, { "epoch": 42.46, "learning_rate": 6.733898305084746e-05, "loss": 0.8307, "step": 2930 }, { "epoch": 42.61, "learning_rate": 6.716949152542374e-05, "loss": 0.7952, "step": 2940 }, { "epoch": 42.75, "learning_rate": 6.7e-05, "loss": 0.6789, "step": 2950 }, { "epoch": 42.9, "learning_rate": 6.683050847457627e-05, "loss": 0.7708, "step": 2960 }, { "epoch": 43.04, "learning_rate": 6.666101694915255e-05, "loss": 0.6823, "step": 2970 }, { "epoch": 43.19, "learning_rate": 6.649152542372881e-05, "loss": 0.6989, "step": 2980 }, { "epoch": 43.33, "learning_rate": 6.63220338983051e-05, "loss": 0.7944, "step": 2990 }, { "epoch": 43.48, "learning_rate": 6.615254237288135e-05, "loss": 0.7664, "step": 3000 }, { "epoch": 43.48, "eval_loss": 0.2088436335325241, "eval_runtime": 572.961, "eval_samples_per_second": 5.924, "eval_steps_per_second": 0.742, "eval_wer": 0.17475696242059677, "step": 3000 }, { "epoch": 43.62, "learning_rate": 6.598305084745763e-05, "loss": 0.7256, "step": 3010 }, { "epoch": 43.77, "learning_rate": 6.581355932203391e-05, "loss": 0.7146, "step": 3020 }, { "epoch": 43.91, "learning_rate": 6.564406779661017e-05, "loss": 0.7347, "step": 3030 }, { "epoch": 44.06, "learning_rate": 6.547457627118644e-05, "loss": 0.7302, "step": 3040 }, { "epoch": 44.2, "learning_rate": 6.53050847457627e-05, "loss": 0.7772, "step": 3050 }, { "epoch": 44.35, "learning_rate": 6.513559322033899e-05, "loss": 0.7214, "step": 3060 }, { "epoch": 44.49, "learning_rate": 6.496610169491527e-05, "loss": 0.6899, "step": 3070 }, { "epoch": 44.64, "learning_rate": 6.479661016949153e-05, "loss": 0.7488, "step": 3080 }, { "epoch": 44.78, "learning_rate": 6.46271186440678e-05, "loss": 0.831, "step": 3090 }, { "epoch": 44.93, "learning_rate": 6.445762711864407e-05, "loss": 0.799, "step": 3100 }, { "epoch": 45.07, "learning_rate": 6.428813559322035e-05, "loss": 0.7195, "step": 3110 }, { "epoch": 45.22, "learning_rate": 6.411864406779661e-05, "loss": 0.7476, "step": 3120 }, { "epoch": 45.36, "learning_rate": 6.394915254237288e-05, "loss": 0.7475, "step": 3130 }, { "epoch": 45.51, "learning_rate": 6.377966101694916e-05, "loss": 0.767, "step": 3140 }, { "epoch": 45.65, "learning_rate": 6.361016949152542e-05, "loss": 0.7263, "step": 3150 }, { "epoch": 45.8, "learning_rate": 6.34406779661017e-05, "loss": 0.7423, "step": 3160 }, { "epoch": 45.94, "learning_rate": 6.327118644067797e-05, "loss": 0.7693, "step": 3170 }, { "epoch": 46.09, "learning_rate": 6.310169491525424e-05, "loss": 0.8136, "step": 3180 }, { "epoch": 46.23, "learning_rate": 6.293220338983052e-05, "loss": 0.6705, "step": 3190 }, { "epoch": 46.38, "learning_rate": 6.276271186440678e-05, "loss": 0.7041, "step": 3200 }, { "epoch": 46.52, "learning_rate": 6.259322033898305e-05, "loss": 0.7478, "step": 3210 }, { "epoch": 46.67, "learning_rate": 6.242372881355933e-05, "loss": 0.6909, "step": 3220 }, { "epoch": 46.81, "learning_rate": 6.22542372881356e-05, "loss": 0.7702, "step": 3230 }, { "epoch": 46.96, "learning_rate": 6.208474576271188e-05, "loss": 0.7302, "step": 3240 }, { "epoch": 47.1, "learning_rate": 6.191525423728813e-05, "loss": 0.7812, "step": 3250 }, { "epoch": 47.25, "learning_rate": 6.174576271186441e-05, "loss": 0.7043, "step": 3260 }, { "epoch": 47.39, "learning_rate": 6.157627118644068e-05, "loss": 0.7645, "step": 3270 }, { "epoch": 47.54, "learning_rate": 6.140677966101696e-05, "loss": 0.8539, "step": 3280 }, { "epoch": 47.68, "learning_rate": 6.123728813559322e-05, "loss": 0.7747, "step": 3290 }, { "epoch": 47.83, "learning_rate": 6.106779661016949e-05, "loss": 0.6613, "step": 3300 }, { "epoch": 47.97, "learning_rate": 6.089830508474577e-05, "loss": 0.7399, "step": 3310 }, { "epoch": 48.12, "learning_rate": 6.072881355932204e-05, "loss": 0.6675, "step": 3320 }, { "epoch": 48.26, "learning_rate": 6.05593220338983e-05, "loss": 0.6699, "step": 3330 }, { "epoch": 48.41, "learning_rate": 6.0389830508474574e-05, "loss": 0.7534, "step": 3340 }, { "epoch": 48.55, "learning_rate": 6.022033898305085e-05, "loss": 0.7837, "step": 3350 }, { "epoch": 48.7, "learning_rate": 6.005084745762713e-05, "loss": 0.7066, "step": 3360 }, { "epoch": 48.84, "learning_rate": 5.98813559322034e-05, "loss": 0.7931, "step": 3370 }, { "epoch": 48.99, "learning_rate": 5.971186440677966e-05, "loss": 0.7575, "step": 3380 }, { "epoch": 49.13, "learning_rate": 5.954237288135593e-05, "loss": 0.6951, "step": 3390 }, { "epoch": 49.28, "learning_rate": 5.9372881355932206e-05, "loss": 0.788, "step": 3400 }, { "epoch": 49.42, "learning_rate": 5.920338983050848e-05, "loss": 0.7727, "step": 3410 }, { "epoch": 49.57, "learning_rate": 5.9033898305084746e-05, "loss": 0.7102, "step": 3420 }, { "epoch": 49.71, "learning_rate": 5.886440677966102e-05, "loss": 0.6733, "step": 3430 }, { "epoch": 49.86, "learning_rate": 5.869491525423729e-05, "loss": 0.6866, "step": 3440 }, { "epoch": 50.0, "learning_rate": 5.8525423728813565e-05, "loss": 0.7414, "step": 3450 }, { "epoch": 50.14, "learning_rate": 5.835593220338983e-05, "loss": 0.684, "step": 3460 }, { "epoch": 50.29, "learning_rate": 5.8186440677966105e-05, "loss": 0.7363, "step": 3470 }, { "epoch": 50.43, "learning_rate": 5.801694915254238e-05, "loss": 0.7421, "step": 3480 }, { "epoch": 50.58, "learning_rate": 5.784745762711865e-05, "loss": 0.7038, "step": 3490 }, { "epoch": 50.72, "learning_rate": 5.767796610169491e-05, "loss": 0.6571, "step": 3500 }, { "epoch": 50.72, "eval_loss": 0.20422525703907013, "eval_runtime": 570.0136, "eval_samples_per_second": 5.954, "eval_steps_per_second": 0.746, "eval_wer": 0.166761169867449, "step": 3500 }, { "epoch": 50.87, "learning_rate": 5.750847457627119e-05, "loss": 0.7286, "step": 3510 }, { "epoch": 51.01, "learning_rate": 5.7338983050847464e-05, "loss": 0.6604, "step": 3520 }, { "epoch": 51.16, "learning_rate": 5.716949152542374e-05, "loss": 0.7662, "step": 3530 }, { "epoch": 51.3, "learning_rate": 5.6999999999999996e-05, "loss": 0.7688, "step": 3540 }, { "epoch": 51.45, "learning_rate": 5.683050847457627e-05, "loss": 0.6903, "step": 3550 }, { "epoch": 51.59, "learning_rate": 5.666101694915254e-05, "loss": 0.6939, "step": 3560 }, { "epoch": 51.74, "learning_rate": 5.649152542372882e-05, "loss": 0.729, "step": 3570 }, { "epoch": 51.88, "learning_rate": 5.632203389830508e-05, "loss": 0.7307, "step": 3580 }, { "epoch": 52.03, "learning_rate": 5.6152542372881355e-05, "loss": 0.6863, "step": 3590 }, { "epoch": 52.17, "learning_rate": 5.598305084745763e-05, "loss": 0.6843, "step": 3600 }, { "epoch": 52.32, "learning_rate": 5.58135593220339e-05, "loss": 0.7298, "step": 3610 }, { "epoch": 52.46, "learning_rate": 5.5644067796610175e-05, "loss": 0.7545, "step": 3620 }, { "epoch": 52.61, "learning_rate": 5.547457627118644e-05, "loss": 0.6663, "step": 3630 }, { "epoch": 52.75, "learning_rate": 5.5305084745762714e-05, "loss": 0.6822, "step": 3640 }, { "epoch": 52.9, "learning_rate": 5.513559322033899e-05, "loss": 0.806, "step": 3650 }, { "epoch": 53.04, "learning_rate": 5.496610169491526e-05, "loss": 0.6597, "step": 3660 }, { "epoch": 53.19, "learning_rate": 5.479661016949153e-05, "loss": 0.7131, "step": 3670 }, { "epoch": 53.33, "learning_rate": 5.46271186440678e-05, "loss": 0.7221, "step": 3680 }, { "epoch": 53.48, "learning_rate": 5.445762711864407e-05, "loss": 0.6235, "step": 3690 }, { "epoch": 53.62, "learning_rate": 5.4288135593220346e-05, "loss": 0.6547, "step": 3700 }, { "epoch": 53.77, "learning_rate": 5.4118644067796606e-05, "loss": 0.7014, "step": 3710 }, { "epoch": 53.91, "learning_rate": 5.3949152542372886e-05, "loss": 0.6826, "step": 3720 }, { "epoch": 54.06, "learning_rate": 5.377966101694916e-05, "loss": 0.664, "step": 3730 }, { "epoch": 54.2, "learning_rate": 5.361016949152543e-05, "loss": 0.7201, "step": 3740 }, { "epoch": 54.35, "learning_rate": 5.344067796610169e-05, "loss": 0.6765, "step": 3750 }, { "epoch": 54.49, "learning_rate": 5.3271186440677965e-05, "loss": 0.6541, "step": 3760 }, { "epoch": 54.64, "learning_rate": 5.310169491525424e-05, "loss": 0.7299, "step": 3770 }, { "epoch": 54.78, "learning_rate": 5.293220338983051e-05, "loss": 0.7139, "step": 3780 }, { "epoch": 54.93, "learning_rate": 5.276271186440678e-05, "loss": 0.6925, "step": 3790 }, { "epoch": 55.07, "learning_rate": 5.259322033898305e-05, "loss": 0.6712, "step": 3800 }, { "epoch": 55.22, "learning_rate": 5.2423728813559324e-05, "loss": 0.6999, "step": 3810 }, { "epoch": 55.36, "learning_rate": 5.22542372881356e-05, "loss": 0.6577, "step": 3820 }, { "epoch": 55.51, "learning_rate": 5.2084745762711864e-05, "loss": 0.733, "step": 3830 }, { "epoch": 55.65, "learning_rate": 5.191525423728814e-05, "loss": 0.666, "step": 3840 }, { "epoch": 55.8, "learning_rate": 5.174576271186441e-05, "loss": 0.6638, "step": 3850 }, { "epoch": 55.94, "learning_rate": 5.157627118644068e-05, "loss": 0.7019, "step": 3860 }, { "epoch": 56.09, "learning_rate": 5.140677966101695e-05, "loss": 0.7018, "step": 3870 }, { "epoch": 56.23, "learning_rate": 5.123728813559322e-05, "loss": 0.6911, "step": 3880 }, { "epoch": 56.38, "learning_rate": 5.1067796610169496e-05, "loss": 0.7035, "step": 3890 }, { "epoch": 56.52, "learning_rate": 5.089830508474577e-05, "loss": 0.7204, "step": 3900 }, { "epoch": 56.67, "learning_rate": 5.072881355932204e-05, "loss": 0.6888, "step": 3910 }, { "epoch": 56.81, "learning_rate": 5.05593220338983e-05, "loss": 0.7502, "step": 3920 }, { "epoch": 56.96, "learning_rate": 5.0389830508474575e-05, "loss": 0.6246, "step": 3930 }, { "epoch": 57.1, "learning_rate": 5.0220338983050855e-05, "loss": 0.6079, "step": 3940 }, { "epoch": 57.25, "learning_rate": 5.005084745762713e-05, "loss": 0.6618, "step": 3950 }, { "epoch": 57.39, "learning_rate": 4.9881355932203394e-05, "loss": 0.6935, "step": 3960 }, { "epoch": 57.54, "learning_rate": 4.971186440677966e-05, "loss": 0.6872, "step": 3970 }, { "epoch": 57.68, "learning_rate": 4.9542372881355934e-05, "loss": 0.6546, "step": 3980 }, { "epoch": 57.83, "learning_rate": 4.937288135593221e-05, "loss": 0.7091, "step": 3990 }, { "epoch": 57.97, "learning_rate": 4.920338983050848e-05, "loss": 0.7014, "step": 4000 }, { "epoch": 57.97, "eval_loss": 0.21360060572624207, "eval_runtime": 570.979, "eval_samples_per_second": 5.944, "eval_steps_per_second": 0.744, "eval_wer": 0.16485467051253605, "step": 4000 }, { "epoch": 58.12, "learning_rate": 4.9033898305084746e-05, "loss": 0.6837, "step": 4010 }, { "epoch": 58.26, "learning_rate": 4.886440677966102e-05, "loss": 0.6707, "step": 4020 }, { "epoch": 58.41, "learning_rate": 4.8694915254237286e-05, "loss": 0.6715, "step": 4030 }, { "epoch": 58.55, "learning_rate": 4.8525423728813566e-05, "loss": 0.64, "step": 4040 }, { "epoch": 58.7, "learning_rate": 4.835593220338983e-05, "loss": 0.6904, "step": 4050 }, { "epoch": 58.84, "learning_rate": 4.8186440677966105e-05, "loss": 0.6809, "step": 4060 }, { "epoch": 58.99, "learning_rate": 4.801694915254237e-05, "loss": 0.6187, "step": 4070 }, { "epoch": 59.13, "learning_rate": 4.7847457627118645e-05, "loss": 0.7028, "step": 4080 }, { "epoch": 59.28, "learning_rate": 4.767796610169492e-05, "loss": 0.7025, "step": 4090 }, { "epoch": 59.42, "learning_rate": 4.750847457627119e-05, "loss": 0.6471, "step": 4100 }, { "epoch": 59.57, "learning_rate": 4.733898305084746e-05, "loss": 0.6968, "step": 4110 }, { "epoch": 59.71, "learning_rate": 4.716949152542373e-05, "loss": 0.6268, "step": 4120 }, { "epoch": 59.86, "learning_rate": 4.7e-05, "loss": 0.6731, "step": 4130 }, { "epoch": 60.0, "learning_rate": 4.683050847457627e-05, "loss": 0.6614, "step": 4140 }, { "epoch": 60.14, "learning_rate": 4.666101694915255e-05, "loss": 0.6138, "step": 4150 }, { "epoch": 60.29, "learning_rate": 4.649152542372882e-05, "loss": 0.6875, "step": 4160 }, { "epoch": 60.43, "learning_rate": 4.632203389830509e-05, "loss": 0.7305, "step": 4170 }, { "epoch": 60.58, "learning_rate": 4.6152542372881356e-05, "loss": 0.6367, "step": 4180 }, { "epoch": 60.72, "learning_rate": 4.598305084745763e-05, "loss": 0.7106, "step": 4190 }, { "epoch": 60.87, "learning_rate": 4.58135593220339e-05, "loss": 0.7531, "step": 4200 }, { "epoch": 61.01, "learning_rate": 4.5644067796610176e-05, "loss": 0.6545, "step": 4210 }, { "epoch": 61.16, "learning_rate": 4.547457627118644e-05, "loss": 0.7458, "step": 4220 }, { "epoch": 61.3, "learning_rate": 4.5305084745762715e-05, "loss": 0.6348, "step": 4230 }, { "epoch": 61.45, "learning_rate": 4.513559322033898e-05, "loss": 0.6631, "step": 4240 }, { "epoch": 61.59, "learning_rate": 4.4966101694915255e-05, "loss": 0.8233, "step": 4250 }, { "epoch": 61.74, "learning_rate": 4.479661016949153e-05, "loss": 0.6475, "step": 4260 }, { "epoch": 61.88, "learning_rate": 4.46271186440678e-05, "loss": 0.659, "step": 4270 }, { "epoch": 62.03, "learning_rate": 4.445762711864407e-05, "loss": 0.7019, "step": 4280 }, { "epoch": 62.17, "learning_rate": 4.428813559322034e-05, "loss": 0.6716, "step": 4290 }, { "epoch": 62.32, "learning_rate": 4.4118644067796614e-05, "loss": 0.6016, "step": 4300 }, { "epoch": 62.46, "learning_rate": 4.394915254237289e-05, "loss": 0.6532, "step": 4310 }, { "epoch": 62.61, "learning_rate": 4.377966101694915e-05, "loss": 0.6426, "step": 4320 }, { "epoch": 62.75, "learning_rate": 4.3610169491525426e-05, "loss": 0.6543, "step": 4330 }, { "epoch": 62.9, "learning_rate": 4.344067796610169e-05, "loss": 0.7405, "step": 4340 }, { "epoch": 63.04, "learning_rate": 4.3271186440677966e-05, "loss": 0.6795, "step": 4350 }, { "epoch": 63.19, "learning_rate": 4.310169491525424e-05, "loss": 0.726, "step": 4360 }, { "epoch": 63.33, "learning_rate": 4.293220338983051e-05, "loss": 0.8443, "step": 4370 }, { "epoch": 63.48, "learning_rate": 4.276271186440678e-05, "loss": 0.7359, "step": 4380 }, { "epoch": 63.62, "learning_rate": 4.259322033898305e-05, "loss": 0.6781, "step": 4390 }, { "epoch": 63.77, "learning_rate": 4.242372881355932e-05, "loss": 0.6366, "step": 4400 }, { "epoch": 63.91, "learning_rate": 4.22542372881356e-05, "loss": 0.6598, "step": 4410 }, { "epoch": 64.06, "learning_rate": 4.208474576271187e-05, "loss": 0.6668, "step": 4420 }, { "epoch": 64.2, "learning_rate": 4.191525423728814e-05, "loss": 0.7418, "step": 4430 }, { "epoch": 64.35, "learning_rate": 4.174576271186441e-05, "loss": 0.6788, "step": 4440 }, { "epoch": 64.49, "learning_rate": 4.157627118644068e-05, "loss": 0.6549, "step": 4450 }, { "epoch": 64.64, "learning_rate": 4.140677966101695e-05, "loss": 0.6636, "step": 4460 }, { "epoch": 64.78, "learning_rate": 4.1237288135593223e-05, "loss": 0.6163, "step": 4470 }, { "epoch": 64.93, "learning_rate": 4.10677966101695e-05, "loss": 0.7296, "step": 4480 }, { "epoch": 65.07, "learning_rate": 4.089830508474576e-05, "loss": 0.6633, "step": 4490 }, { "epoch": 65.22, "learning_rate": 4.0728813559322036e-05, "loss": 0.6171, "step": 4500 }, { "epoch": 65.22, "eval_loss": 0.21388114988803864, "eval_runtime": 574.1167, "eval_samples_per_second": 5.912, "eval_steps_per_second": 0.74, "eval_wer": 0.16409864490627749, "step": 4500 }, { "epoch": 65.36, "learning_rate": 4.055932203389831e-05, "loss": 0.647, "step": 4510 }, { "epoch": 65.51, "learning_rate": 4.038983050847458e-05, "loss": 0.5946, "step": 4520 }, { "epoch": 65.65, "learning_rate": 4.022033898305085e-05, "loss": 0.6789, "step": 4530 }, { "epoch": 65.8, "learning_rate": 4.005084745762712e-05, "loss": 0.6854, "step": 4540 }, { "epoch": 65.94, "learning_rate": 3.988135593220339e-05, "loss": 0.6547, "step": 4550 }, { "epoch": 66.09, "learning_rate": 3.971186440677966e-05, "loss": 0.7462, "step": 4560 }, { "epoch": 66.23, "learning_rate": 3.9542372881355935e-05, "loss": 0.6724, "step": 4570 }, { "epoch": 66.38, "learning_rate": 3.937288135593221e-05, "loss": 0.6623, "step": 4580 }, { "epoch": 66.52, "learning_rate": 3.9203389830508474e-05, "loss": 0.6438, "step": 4590 }, { "epoch": 66.67, "learning_rate": 3.903389830508475e-05, "loss": 0.6983, "step": 4600 }, { "epoch": 66.81, "learning_rate": 3.8864406779661014e-05, "loss": 0.6905, "step": 4610 }, { "epoch": 66.96, "learning_rate": 3.8694915254237294e-05, "loss": 0.7062, "step": 4620 }, { "epoch": 67.1, "learning_rate": 3.852542372881356e-05, "loss": 0.6685, "step": 4630 }, { "epoch": 67.25, "learning_rate": 3.835593220338983e-05, "loss": 0.6545, "step": 4640 }, { "epoch": 67.39, "learning_rate": 3.81864406779661e-05, "loss": 0.6953, "step": 4650 }, { "epoch": 67.54, "learning_rate": 3.801694915254237e-05, "loss": 0.6793, "step": 4660 }, { "epoch": 67.68, "learning_rate": 3.7847457627118646e-05, "loss": 0.6059, "step": 4670 }, { "epoch": 67.83, "learning_rate": 3.767796610169492e-05, "loss": 0.6555, "step": 4680 }, { "epoch": 67.97, "learning_rate": 3.750847457627119e-05, "loss": 0.6261, "step": 4690 }, { "epoch": 68.12, "learning_rate": 3.733898305084746e-05, "loss": 0.7417, "step": 4700 }, { "epoch": 68.26, "learning_rate": 3.716949152542373e-05, "loss": 0.6732, "step": 4710 }, { "epoch": 68.41, "learning_rate": 3.7e-05, "loss": 0.7073, "step": 4720 }, { "epoch": 68.55, "learning_rate": 3.683050847457628e-05, "loss": 0.6299, "step": 4730 }, { "epoch": 68.7, "learning_rate": 3.6661016949152544e-05, "loss": 0.6789, "step": 4740 }, { "epoch": 68.84, "learning_rate": 3.649152542372882e-05, "loss": 0.6946, "step": 4750 }, { "epoch": 68.99, "learning_rate": 3.6322033898305084e-05, "loss": 0.6954, "step": 4760 }, { "epoch": 69.13, "learning_rate": 3.615254237288136e-05, "loss": 0.6609, "step": 4770 }, { "epoch": 69.28, "learning_rate": 3.598305084745763e-05, "loss": 0.7685, "step": 4780 }, { "epoch": 69.42, "learning_rate": 3.5813559322033903e-05, "loss": 0.6556, "step": 4790 }, { "epoch": 69.57, "learning_rate": 3.564406779661017e-05, "loss": 0.6377, "step": 4800 }, { "epoch": 69.71, "learning_rate": 3.547457627118644e-05, "loss": 0.667, "step": 4810 }, { "epoch": 69.86, "learning_rate": 3.530508474576271e-05, "loss": 0.6439, "step": 4820 }, { "epoch": 70.0, "learning_rate": 3.513559322033899e-05, "loss": 0.6266, "step": 4830 }, { "epoch": 70.14, "learning_rate": 3.4966101694915256e-05, "loss": 0.6563, "step": 4840 }, { "epoch": 70.29, "learning_rate": 3.479661016949153e-05, "loss": 0.606, "step": 4850 }, { "epoch": 70.43, "learning_rate": 3.4627118644067795e-05, "loss": 0.6662, "step": 4860 }, { "epoch": 70.58, "learning_rate": 3.445762711864407e-05, "loss": 0.6453, "step": 4870 }, { "epoch": 70.72, "learning_rate": 3.428813559322034e-05, "loss": 0.6812, "step": 4880 }, { "epoch": 70.87, "learning_rate": 3.4118644067796615e-05, "loss": 0.6655, "step": 4890 }, { "epoch": 71.01, "learning_rate": 3.394915254237288e-05, "loss": 0.6687, "step": 4900 }, { "epoch": 71.16, "learning_rate": 3.3779661016949154e-05, "loss": 0.6349, "step": 4910 }, { "epoch": 71.3, "learning_rate": 3.361016949152542e-05, "loss": 0.5908, "step": 4920 }, { "epoch": 71.45, "learning_rate": 3.3440677966101694e-05, "loss": 0.6648, "step": 4930 }, { "epoch": 71.59, "learning_rate": 3.327118644067797e-05, "loss": 0.6964, "step": 4940 }, { "epoch": 71.74, "learning_rate": 3.310169491525424e-05, "loss": 0.6482, "step": 4950 }, { "epoch": 71.88, "learning_rate": 3.293220338983051e-05, "loss": 0.7086, "step": 4960 }, { "epoch": 72.03, "learning_rate": 3.276271186440678e-05, "loss": 0.6146, "step": 4970 }, { "epoch": 72.17, "learning_rate": 3.259322033898305e-05, "loss": 0.6546, "step": 4980 }, { "epoch": 72.32, "learning_rate": 3.2423728813559326e-05, "loss": 0.608, "step": 4990 }, { "epoch": 72.46, "learning_rate": 3.22542372881356e-05, "loss": 0.6609, "step": 5000 }, { "epoch": 72.46, "eval_loss": 0.2143816202878952, "eval_runtime": 575.4482, "eval_samples_per_second": 5.898, "eval_steps_per_second": 0.739, "eval_wer": 0.16206888050686585, "step": 5000 }, { "epoch": 72.61, "learning_rate": 3.2084745762711865e-05, "loss": 0.6483, "step": 5010 }, { "epoch": 72.75, "learning_rate": 3.191525423728814e-05, "loss": 0.5998, "step": 5020 }, { "epoch": 72.9, "learning_rate": 3.1745762711864405e-05, "loss": 0.6161, "step": 5030 }, { "epoch": 73.04, "learning_rate": 3.157627118644068e-05, "loss": 0.6856, "step": 5040 }, { "epoch": 73.19, "learning_rate": 3.140677966101695e-05, "loss": 0.6095, "step": 5050 }, { "epoch": 73.33, "learning_rate": 3.1237288135593224e-05, "loss": 0.6235, "step": 5060 }, { "epoch": 73.48, "learning_rate": 3.106779661016949e-05, "loss": 0.6827, "step": 5070 }, { "epoch": 73.62, "learning_rate": 3.0898305084745764e-05, "loss": 0.5903, "step": 5080 }, { "epoch": 73.77, "learning_rate": 3.072881355932204e-05, "loss": 0.7406, "step": 5090 }, { "epoch": 73.91, "learning_rate": 3.055932203389831e-05, "loss": 0.6621, "step": 5100 }, { "epoch": 74.06, "learning_rate": 3.0389830508474577e-05, "loss": 0.6619, "step": 5110 }, { "epoch": 74.2, "learning_rate": 3.022033898305085e-05, "loss": 0.6755, "step": 5120 }, { "epoch": 74.35, "learning_rate": 3.005084745762712e-05, "loss": 0.6568, "step": 5130 }, { "epoch": 74.49, "learning_rate": 2.9881355932203393e-05, "loss": 0.6244, "step": 5140 }, { "epoch": 74.64, "learning_rate": 2.971186440677966e-05, "loss": 0.6166, "step": 5150 }, { "epoch": 74.78, "learning_rate": 2.9542372881355936e-05, "loss": 0.6238, "step": 5160 }, { "epoch": 74.93, "learning_rate": 2.9372881355932202e-05, "loss": 0.7148, "step": 5170 }, { "epoch": 75.07, "learning_rate": 2.9203389830508475e-05, "loss": 0.6449, "step": 5180 }, { "epoch": 75.22, "learning_rate": 2.9033898305084745e-05, "loss": 0.6567, "step": 5190 }, { "epoch": 75.36, "learning_rate": 2.8864406779661018e-05, "loss": 0.6167, "step": 5200 }, { "epoch": 75.51, "learning_rate": 2.8694915254237288e-05, "loss": 0.6728, "step": 5210 }, { "epoch": 75.65, "learning_rate": 2.852542372881356e-05, "loss": 0.6267, "step": 5220 }, { "epoch": 75.8, "learning_rate": 2.8355932203389834e-05, "loss": 0.6707, "step": 5230 }, { "epoch": 75.94, "learning_rate": 2.8186440677966104e-05, "loss": 0.6005, "step": 5240 }, { "epoch": 76.09, "learning_rate": 2.8016949152542377e-05, "loss": 0.5838, "step": 5250 }, { "epoch": 76.23, "learning_rate": 2.7847457627118643e-05, "loss": 0.6954, "step": 5260 }, { "epoch": 76.38, "learning_rate": 2.767796610169492e-05, "loss": 0.6658, "step": 5270 }, { "epoch": 76.52, "learning_rate": 2.7508474576271186e-05, "loss": 0.6252, "step": 5280 }, { "epoch": 76.67, "learning_rate": 2.733898305084746e-05, "loss": 0.6967, "step": 5290 }, { "epoch": 76.81, "learning_rate": 2.716949152542373e-05, "loss": 0.6216, "step": 5300 }, { "epoch": 76.96, "learning_rate": 2.7000000000000002e-05, "loss": 0.6392, "step": 5310 }, { "epoch": 77.1, "learning_rate": 2.6830508474576272e-05, "loss": 0.6266, "step": 5320 }, { "epoch": 77.25, "learning_rate": 2.6661016949152545e-05, "loss": 0.5908, "step": 5330 }, { "epoch": 77.39, "learning_rate": 2.6491525423728815e-05, "loss": 0.6788, "step": 5340 }, { "epoch": 77.54, "learning_rate": 2.6322033898305088e-05, "loss": 0.6592, "step": 5350 }, { "epoch": 77.68, "learning_rate": 2.6152542372881355e-05, "loss": 0.6038, "step": 5360 }, { "epoch": 77.83, "learning_rate": 2.598305084745763e-05, "loss": 0.5884, "step": 5370 }, { "epoch": 77.97, "learning_rate": 2.5813559322033898e-05, "loss": 0.6433, "step": 5380 }, { "epoch": 78.12, "learning_rate": 2.564406779661017e-05, "loss": 0.6976, "step": 5390 }, { "epoch": 78.26, "learning_rate": 2.547457627118644e-05, "loss": 0.6922, "step": 5400 }, { "epoch": 78.41, "learning_rate": 2.5305084745762714e-05, "loss": 0.6621, "step": 5410 }, { "epoch": 78.55, "learning_rate": 2.5135593220338983e-05, "loss": 0.6388, "step": 5420 }, { "epoch": 78.7, "learning_rate": 2.4966101694915257e-05, "loss": 0.628, "step": 5430 }, { "epoch": 78.84, "learning_rate": 2.4796610169491526e-05, "loss": 0.5507, "step": 5440 }, { "epoch": 78.99, "learning_rate": 2.46271186440678e-05, "loss": 0.6358, "step": 5450 }, { "epoch": 79.13, "learning_rate": 2.445762711864407e-05, "loss": 0.8279, "step": 5460 }, { "epoch": 79.28, "learning_rate": 2.428813559322034e-05, "loss": 0.6108, "step": 5470 }, { "epoch": 79.42, "learning_rate": 2.4118644067796612e-05, "loss": 0.6535, "step": 5480 }, { "epoch": 79.57, "learning_rate": 2.3949152542372882e-05, "loss": 0.5962, "step": 5490 }, { "epoch": 79.71, "learning_rate": 2.3779661016949155e-05, "loss": 0.6318, "step": 5500 }, { "epoch": 79.71, "eval_loss": 0.21294616162776947, "eval_runtime": 575.7769, "eval_samples_per_second": 5.895, "eval_steps_per_second": 0.738, "eval_wer": 0.16001446309855452, "step": 5500 }, { "epoch": 79.86, "learning_rate": 2.3610169491525425e-05, "loss": 0.5991, "step": 5510 }, { "epoch": 80.0, "learning_rate": 2.3440677966101695e-05, "loss": 0.6389, "step": 5520 }, { "epoch": 80.14, "learning_rate": 2.3271186440677968e-05, "loss": 0.5969, "step": 5530 }, { "epoch": 80.29, "learning_rate": 2.3101694915254237e-05, "loss": 0.626, "step": 5540 }, { "epoch": 80.43, "learning_rate": 2.2932203389830507e-05, "loss": 0.6123, "step": 5550 }, { "epoch": 80.58, "learning_rate": 2.276271186440678e-05, "loss": 0.6056, "step": 5560 }, { "epoch": 80.72, "learning_rate": 2.259322033898305e-05, "loss": 0.6004, "step": 5570 }, { "epoch": 80.87, "learning_rate": 2.2423728813559323e-05, "loss": 0.5924, "step": 5580 }, { "epoch": 81.01, "learning_rate": 2.2254237288135593e-05, "loss": 0.6622, "step": 5590 }, { "epoch": 81.16, "learning_rate": 2.2084745762711863e-05, "loss": 0.6285, "step": 5600 }, { "epoch": 81.3, "learning_rate": 2.1915254237288136e-05, "loss": 0.6199, "step": 5610 }, { "epoch": 81.45, "learning_rate": 2.174576271186441e-05, "loss": 0.6496, "step": 5620 }, { "epoch": 81.59, "learning_rate": 2.157627118644068e-05, "loss": 0.632, "step": 5630 }, { "epoch": 81.74, "learning_rate": 2.1406779661016952e-05, "loss": 0.5825, "step": 5640 }, { "epoch": 81.88, "learning_rate": 2.1237288135593222e-05, "loss": 0.7035, "step": 5650 }, { "epoch": 82.03, "learning_rate": 2.1067796610169495e-05, "loss": 0.5958, "step": 5660 }, { "epoch": 82.17, "learning_rate": 2.0898305084745765e-05, "loss": 0.6213, "step": 5670 }, { "epoch": 82.32, "learning_rate": 2.0728813559322035e-05, "loss": 0.621, "step": 5680 }, { "epoch": 82.46, "learning_rate": 2.0559322033898308e-05, "loss": 0.5801, "step": 5690 }, { "epoch": 82.61, "learning_rate": 2.0389830508474577e-05, "loss": 0.6511, "step": 5700 }, { "epoch": 82.75, "learning_rate": 2.0220338983050847e-05, "loss": 0.6268, "step": 5710 }, { "epoch": 82.9, "learning_rate": 2.005084745762712e-05, "loss": 0.642, "step": 5720 }, { "epoch": 83.04, "learning_rate": 1.988135593220339e-05, "loss": 0.6158, "step": 5730 }, { "epoch": 83.19, "learning_rate": 1.9711864406779663e-05, "loss": 0.6171, "step": 5740 }, { "epoch": 83.33, "learning_rate": 1.9542372881355933e-05, "loss": 0.5943, "step": 5750 }, { "epoch": 83.48, "learning_rate": 1.9372881355932203e-05, "loss": 0.6312, "step": 5760 }, { "epoch": 83.62, "learning_rate": 1.9203389830508476e-05, "loss": 0.6128, "step": 5770 }, { "epoch": 83.77, "learning_rate": 1.9033898305084746e-05, "loss": 0.6094, "step": 5780 }, { "epoch": 83.91, "learning_rate": 1.886440677966102e-05, "loss": 0.6403, "step": 5790 }, { "epoch": 84.06, "learning_rate": 1.869491525423729e-05, "loss": 0.5933, "step": 5800 }, { "epoch": 84.2, "learning_rate": 1.852542372881356e-05, "loss": 0.5684, "step": 5810 }, { "epoch": 84.35, "learning_rate": 1.835593220338983e-05, "loss": 0.6025, "step": 5820 }, { "epoch": 84.49, "learning_rate": 1.81864406779661e-05, "loss": 0.6225, "step": 5830 }, { "epoch": 84.64, "learning_rate": 1.8016949152542374e-05, "loss": 0.6246, "step": 5840 }, { "epoch": 84.78, "learning_rate": 1.7847457627118644e-05, "loss": 0.6186, "step": 5850 }, { "epoch": 84.93, "learning_rate": 1.7677966101694914e-05, "loss": 0.6706, "step": 5860 }, { "epoch": 85.07, "learning_rate": 1.7508474576271187e-05, "loss": 0.5923, "step": 5870 }, { "epoch": 85.22, "learning_rate": 1.7338983050847457e-05, "loss": 0.5892, "step": 5880 }, { "epoch": 85.36, "learning_rate": 1.716949152542373e-05, "loss": 0.6155, "step": 5890 }, { "epoch": 85.51, "learning_rate": 1.7000000000000003e-05, "loss": 0.545, "step": 5900 }, { "epoch": 85.65, "learning_rate": 1.6830508474576273e-05, "loss": 0.6541, "step": 5910 }, { "epoch": 85.8, "learning_rate": 1.6661016949152543e-05, "loss": 0.7079, "step": 5920 }, { "epoch": 85.94, "learning_rate": 1.6491525423728816e-05, "loss": 0.58, "step": 5930 }, { "epoch": 86.09, "learning_rate": 1.6322033898305086e-05, "loss": 0.631, "step": 5940 }, { "epoch": 86.23, "learning_rate": 1.615254237288136e-05, "loss": 0.6232, "step": 5950 }, { "epoch": 86.38, "learning_rate": 1.598305084745763e-05, "loss": 0.5722, "step": 5960 }, { "epoch": 86.52, "learning_rate": 1.58135593220339e-05, "loss": 0.6361, "step": 5970 }, { "epoch": 86.67, "learning_rate": 1.564406779661017e-05, "loss": 0.6603, "step": 5980 }, { "epoch": 86.81, "learning_rate": 1.547457627118644e-05, "loss": 0.6191, "step": 5990 }, { "epoch": 86.96, "learning_rate": 1.5305084745762714e-05, "loss": 0.6222, "step": 6000 }, { "epoch": 86.96, "eval_loss": 0.2124408483505249, "eval_runtime": 572.9918, "eval_samples_per_second": 5.923, "eval_steps_per_second": 0.742, "eval_wer": 0.158223011118507, "step": 6000 }, { "epoch": 87.1, "learning_rate": 1.5135593220338984e-05, "loss": 0.5942, "step": 6010 }, { "epoch": 87.25, "learning_rate": 1.4966101694915256e-05, "loss": 0.6446, "step": 6020 }, { "epoch": 87.39, "learning_rate": 1.4796610169491525e-05, "loss": 0.5991, "step": 6030 }, { "epoch": 87.54, "learning_rate": 1.4627118644067797e-05, "loss": 0.6077, "step": 6040 }, { "epoch": 87.68, "learning_rate": 1.4457627118644068e-05, "loss": 0.5856, "step": 6050 }, { "epoch": 87.83, "learning_rate": 1.428813559322034e-05, "loss": 0.6133, "step": 6060 }, { "epoch": 87.97, "learning_rate": 1.411864406779661e-05, "loss": 0.5992, "step": 6070 }, { "epoch": 88.12, "learning_rate": 1.3949152542372881e-05, "loss": 0.6122, "step": 6080 }, { "epoch": 88.26, "learning_rate": 1.3779661016949153e-05, "loss": 0.6178, "step": 6090 }, { "epoch": 88.41, "learning_rate": 1.3610169491525424e-05, "loss": 0.6505, "step": 6100 }, { "epoch": 88.55, "learning_rate": 1.3440677966101695e-05, "loss": 0.5641, "step": 6110 }, { "epoch": 88.7, "learning_rate": 1.3271186440677965e-05, "loss": 0.6329, "step": 6120 }, { "epoch": 88.84, "learning_rate": 1.3101694915254237e-05, "loss": 0.6119, "step": 6130 }, { "epoch": 88.99, "learning_rate": 1.2932203389830508e-05, "loss": 0.6043, "step": 6140 }, { "epoch": 89.13, "learning_rate": 1.276271186440678e-05, "loss": 0.6661, "step": 6150 }, { "epoch": 89.28, "learning_rate": 1.2593220338983053e-05, "loss": 0.5938, "step": 6160 }, { "epoch": 89.42, "learning_rate": 1.2423728813559323e-05, "loss": 0.6135, "step": 6170 }, { "epoch": 89.57, "learning_rate": 1.2254237288135594e-05, "loss": 0.6354, "step": 6180 }, { "epoch": 89.71, "learning_rate": 1.2084745762711865e-05, "loss": 0.5597, "step": 6190 }, { "epoch": 89.86, "learning_rate": 1.1915254237288135e-05, "loss": 0.6065, "step": 6200 }, { "epoch": 90.0, "learning_rate": 1.1745762711864407e-05, "loss": 0.5579, "step": 6210 }, { "epoch": 90.14, "learning_rate": 1.157627118644068e-05, "loss": 0.5914, "step": 6220 }, { "epoch": 90.29, "learning_rate": 1.140677966101695e-05, "loss": 0.6552, "step": 6230 }, { "epoch": 90.43, "learning_rate": 1.1237288135593221e-05, "loss": 0.5762, "step": 6240 }, { "epoch": 90.58, "learning_rate": 1.1067796610169492e-05, "loss": 0.5902, "step": 6250 }, { "epoch": 90.72, "learning_rate": 1.0898305084745764e-05, "loss": 0.5821, "step": 6260 }, { "epoch": 90.87, "learning_rate": 1.0728813559322035e-05, "loss": 0.6163, "step": 6270 }, { "epoch": 91.01, "learning_rate": 1.0559322033898305e-05, "loss": 0.6451, "step": 6280 }, { "epoch": 91.16, "learning_rate": 1.0389830508474577e-05, "loss": 0.6399, "step": 6290 }, { "epoch": 91.3, "learning_rate": 1.0220338983050848e-05, "loss": 0.6132, "step": 6300 }, { "epoch": 91.45, "learning_rate": 1.005084745762712e-05, "loss": 0.612, "step": 6310 }, { "epoch": 91.59, "learning_rate": 9.88135593220339e-06, "loss": 0.5939, "step": 6320 }, { "epoch": 91.74, "learning_rate": 9.71186440677966e-06, "loss": 0.6379, "step": 6330 }, { "epoch": 91.88, "learning_rate": 9.542372881355932e-06, "loss": 0.6063, "step": 6340 }, { "epoch": 92.03, "learning_rate": 9.372881355932204e-06, "loss": 0.6005, "step": 6350 }, { "epoch": 92.17, "learning_rate": 9.203389830508475e-06, "loss": 0.6076, "step": 6360 }, { "epoch": 92.32, "learning_rate": 9.033898305084747e-06, "loss": 0.6082, "step": 6370 }, { "epoch": 92.46, "learning_rate": 8.864406779661018e-06, "loss": 0.6571, "step": 6380 }, { "epoch": 92.61, "learning_rate": 8.69491525423729e-06, "loss": 0.5666, "step": 6390 }, { "epoch": 92.75, "learning_rate": 8.52542372881356e-06, "loss": 0.5932, "step": 6400 }, { "epoch": 92.9, "learning_rate": 8.35593220338983e-06, "loss": 0.6317, "step": 6410 }, { "epoch": 93.04, "learning_rate": 8.186440677966102e-06, "loss": 0.6665, "step": 6420 }, { "epoch": 93.19, "learning_rate": 8.016949152542374e-06, "loss": 0.5687, "step": 6430 }, { "epoch": 93.33, "learning_rate": 7.847457627118643e-06, "loss": 0.5904, "step": 6440 }, { "epoch": 93.48, "learning_rate": 7.677966101694915e-06, "loss": 0.7073, "step": 6450 }, { "epoch": 93.62, "learning_rate": 7.508474576271186e-06, "loss": 0.619, "step": 6460 }, { "epoch": 93.77, "learning_rate": 7.338983050847458e-06, "loss": 0.6203, "step": 6470 }, { "epoch": 93.91, "learning_rate": 7.1694915254237284e-06, "loss": 0.6576, "step": 6480 }, { "epoch": 94.06, "learning_rate": 7.000000000000001e-06, "loss": 0.6249, "step": 6490 }, { "epoch": 94.2, "learning_rate": 6.830508474576272e-06, "loss": 0.588, "step": 6500 }, { "epoch": 94.2, "eval_loss": 0.2143363356590271, "eval_runtime": 573.5555, "eval_samples_per_second": 5.917, "eval_steps_per_second": 0.741, "eval_wer": 0.15602067565679725, "step": 6500 }, { "epoch": 94.35, "learning_rate": 6.661016949152543e-06, "loss": 0.669, "step": 6510 }, { "epoch": 94.49, "learning_rate": 6.491525423728814e-06, "loss": 0.5974, "step": 6520 }, { "epoch": 94.64, "learning_rate": 6.322033898305085e-06, "loss": 0.6671, "step": 6530 }, { "epoch": 94.78, "learning_rate": 6.152542372881356e-06, "loss": 0.6735, "step": 6540 }, { "epoch": 94.93, "learning_rate": 5.983050847457628e-06, "loss": 0.6356, "step": 6550 }, { "epoch": 95.07, "learning_rate": 5.813559322033898e-06, "loss": 0.6135, "step": 6560 }, { "epoch": 95.22, "learning_rate": 5.64406779661017e-06, "loss": 0.6403, "step": 6570 }, { "epoch": 95.36, "learning_rate": 5.4745762711864405e-06, "loss": 0.5574, "step": 6580 }, { "epoch": 95.51, "learning_rate": 5.305084745762713e-06, "loss": 0.552, "step": 6590 }, { "epoch": 95.65, "learning_rate": 5.135593220338983e-06, "loss": 0.5979, "step": 6600 }, { "epoch": 95.8, "learning_rate": 4.966101694915255e-06, "loss": 0.6342, "step": 6610 }, { "epoch": 95.94, "learning_rate": 4.7966101694915255e-06, "loss": 0.6397, "step": 6620 }, { "epoch": 96.09, "learning_rate": 4.627118644067797e-06, "loss": 0.5847, "step": 6630 }, { "epoch": 96.23, "learning_rate": 4.4576271186440676e-06, "loss": 0.5736, "step": 6640 }, { "epoch": 96.38, "learning_rate": 4.288135593220339e-06, "loss": 0.6148, "step": 6650 }, { "epoch": 96.52, "learning_rate": 4.1186440677966105e-06, "loss": 0.6084, "step": 6660 }, { "epoch": 96.67, "learning_rate": 3.949152542372882e-06, "loss": 0.5835, "step": 6670 }, { "epoch": 96.81, "learning_rate": 3.779661016949153e-06, "loss": 0.5791, "step": 6680 }, { "epoch": 96.96, "learning_rate": 3.610169491525424e-06, "loss": 0.7132, "step": 6690 }, { "epoch": 97.1, "learning_rate": 3.440677966101695e-06, "loss": 0.5572, "step": 6700 }, { "epoch": 97.25, "learning_rate": 3.271186440677966e-06, "loss": 0.6607, "step": 6710 }, { "epoch": 97.39, "learning_rate": 3.1016949152542375e-06, "loss": 0.6343, "step": 6720 }, { "epoch": 97.54, "learning_rate": 2.9322033898305086e-06, "loss": 0.6126, "step": 6730 }, { "epoch": 97.68, "learning_rate": 2.76271186440678e-06, "loss": 0.6536, "step": 6740 }, { "epoch": 97.83, "learning_rate": 2.593220338983051e-06, "loss": 0.5843, "step": 6750 }, { "epoch": 97.97, "learning_rate": 2.423728813559322e-06, "loss": 0.5874, "step": 6760 }, { "epoch": 98.12, "learning_rate": 2.254237288135593e-06, "loss": 0.6067, "step": 6770 }, { "epoch": 98.26, "learning_rate": 2.0847457627118646e-06, "loss": 0.5626, "step": 6780 }, { "epoch": 98.41, "learning_rate": 1.9152542372881356e-06, "loss": 0.6659, "step": 6790 }, { "epoch": 98.55, "learning_rate": 1.7457627118644067e-06, "loss": 0.6462, "step": 6800 }, { "epoch": 98.7, "learning_rate": 1.5762711864406781e-06, "loss": 0.67, "step": 6810 }, { "epoch": 98.84, "learning_rate": 1.4067796610169492e-06, "loss": 0.5734, "step": 6820 }, { "epoch": 98.99, "learning_rate": 1.2372881355932204e-06, "loss": 0.6133, "step": 6830 }, { "epoch": 99.13, "learning_rate": 1.0677966101694917e-06, "loss": 0.596, "step": 6840 }, { "epoch": 99.28, "learning_rate": 9.152542372881356e-07, "loss": 0.6763, "step": 6850 }, { "epoch": 99.42, "learning_rate": 7.457627118644068e-07, "loss": 0.6368, "step": 6860 }, { "epoch": 99.57, "learning_rate": 5.76271186440678e-07, "loss": 0.6463, "step": 6870 }, { "epoch": 99.71, "learning_rate": 4.0677966101694916e-07, "loss": 0.62, "step": 6880 }, { "epoch": 99.86, "learning_rate": 2.3728813559322033e-07, "loss": 0.6346, "step": 6890 }, { "epoch": 100.0, "learning_rate": 6.779661016949153e-08, "loss": 0.5672, "step": 6900 }, { "epoch": 100.0, "step": 6900, "total_flos": 1.0025325448199992e+20, "train_loss": 1.2894020353538402, "train_runtime": 52505.7639, "train_samples_per_second": 4.179, "train_steps_per_second": 0.131 } ], "max_steps": 6900, "num_train_epochs": 100, "total_flos": 1.0025325448199992e+20, "trial_name": null, "trial_params": null }