{ "best_metric": null, "best_model_checkpoint": null, "epoch": 200.0, "global_step": 13800, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.14, "learning_rate": 9e-07, "loss": 11.2286, "step": 10 }, { "epoch": 0.29, "learning_rate": 1.8e-06, "loss": 12.1103, "step": 20 }, { "epoch": 0.43, "learning_rate": 2.8000000000000003e-06, "loss": 11.9593, "step": 30 }, { "epoch": 0.58, "learning_rate": 3.8e-06, "loss": 11.6711, "step": 40 }, { "epoch": 0.72, "learning_rate": 4.800000000000001e-06, "loss": 11.9783, "step": 50 }, { "epoch": 0.87, "learning_rate": 5.8e-06, "loss": 10.9667, "step": 60 }, { "epoch": 1.01, "learning_rate": 6.800000000000001e-06, "loss": 12.1446, "step": 70 }, { "epoch": 1.16, "learning_rate": 7.8e-06, "loss": 10.2963, "step": 80 }, { "epoch": 1.3, "learning_rate": 8.7e-06, "loss": 10.8899, "step": 90 }, { "epoch": 1.45, "learning_rate": 9.7e-06, "loss": 9.5968, "step": 100 }, { "epoch": 1.59, "learning_rate": 1.0700000000000001e-05, "loss": 7.6591, "step": 110 }, { "epoch": 1.74, "learning_rate": 1.1700000000000001e-05, "loss": 7.5159, "step": 120 }, { "epoch": 1.88, "learning_rate": 1.27e-05, "loss": 6.2925, "step": 130 }, { "epoch": 2.03, "learning_rate": 1.3700000000000001e-05, "loss": 5.7639, "step": 140 }, { "epoch": 2.17, "learning_rate": 1.47e-05, "loss": 5.6696, "step": 150 }, { "epoch": 2.32, "learning_rate": 1.5700000000000002e-05, "loss": 4.7696, "step": 160 }, { "epoch": 2.46, "learning_rate": 1.6700000000000003e-05, "loss": 5.1383, "step": 170 }, { "epoch": 2.61, "learning_rate": 1.77e-05, "loss": 4.5714, "step": 180 }, { "epoch": 2.75, "learning_rate": 1.87e-05, "loss": 4.5116, "step": 190 }, { "epoch": 2.9, "learning_rate": 1.97e-05, "loss": 4.3733, "step": 200 }, { "epoch": 3.04, "learning_rate": 2.07e-05, "loss": 3.8443, "step": 210 }, { "epoch": 3.19, "learning_rate": 2.1700000000000002e-05, "loss": 3.9388, "step": 220 }, { "epoch": 3.33, "learning_rate": 2.2700000000000003e-05, "loss": 3.8416, "step": 230 }, { "epoch": 3.48, "learning_rate": 2.37e-05, "loss": 3.7135, "step": 240 }, { "epoch": 3.62, "learning_rate": 2.47e-05, "loss": 3.5271, "step": 250 }, { "epoch": 3.77, "learning_rate": 2.57e-05, "loss": 3.5189, "step": 260 }, { "epoch": 3.91, "learning_rate": 2.6700000000000002e-05, "loss": 3.5816, "step": 270 }, { "epoch": 4.06, "learning_rate": 2.7700000000000002e-05, "loss": 3.3417, "step": 280 }, { "epoch": 4.2, "learning_rate": 2.87e-05, "loss": 3.5864, "step": 290 }, { "epoch": 4.35, "learning_rate": 2.97e-05, "loss": 3.327, "step": 300 }, { "epoch": 4.49, "learning_rate": 3.07e-05, "loss": 3.205, "step": 310 }, { "epoch": 4.64, "learning_rate": 3.1700000000000005e-05, "loss": 3.1087, "step": 320 }, { "epoch": 4.78, "learning_rate": 3.27e-05, "loss": 3.1126, "step": 330 }, { "epoch": 4.93, "learning_rate": 3.3700000000000006e-05, "loss": 3.1934, "step": 340 }, { "epoch": 5.07, "learning_rate": 3.4699999999999996e-05, "loss": 3.0881, "step": 350 }, { "epoch": 5.22, "learning_rate": 3.57e-05, "loss": 3.086, "step": 360 }, { "epoch": 5.36, "learning_rate": 3.6700000000000004e-05, "loss": 3.0304, "step": 370 }, { "epoch": 5.51, "learning_rate": 3.77e-05, "loss": 3.0262, "step": 380 }, { "epoch": 5.65, "learning_rate": 3.8700000000000006e-05, "loss": 3.0492, "step": 390 }, { "epoch": 5.8, "learning_rate": 3.97e-05, "loss": 3.0657, "step": 400 }, { "epoch": 5.94, "learning_rate": 4.07e-05, "loss": 2.9912, "step": 410 }, { "epoch": 6.09, "learning_rate": 4.17e-05, "loss": 2.9993, "step": 420 }, { "epoch": 6.23, "learning_rate": 4.27e-05, "loss": 3.0465, "step": 430 }, { "epoch": 6.38, "learning_rate": 4.3700000000000005e-05, "loss": 2.9501, "step": 440 }, { "epoch": 6.52, "learning_rate": 4.47e-05, "loss": 3.0918, "step": 450 }, { "epoch": 6.67, "learning_rate": 4.5700000000000006e-05, "loss": 3.0306, "step": 460 }, { "epoch": 6.81, "learning_rate": 4.6700000000000003e-05, "loss": 2.9349, "step": 470 }, { "epoch": 6.96, "learning_rate": 4.77e-05, "loss": 3.0044, "step": 480 }, { "epoch": 7.1, "learning_rate": 4.87e-05, "loss": 2.9541, "step": 490 }, { "epoch": 7.25, "learning_rate": 4.97e-05, "loss": 3.0132, "step": 500 }, { "epoch": 7.25, "eval_loss": 2.9393208026885986, "eval_runtime": 580.9987, "eval_samples_per_second": 5.842, "eval_steps_per_second": 0.731, "eval_wer": 1.0, "step": 500 }, { "epoch": 7.39, "learning_rate": 5.0700000000000006e-05, "loss": 2.9162, "step": 510 }, { "epoch": 7.54, "learning_rate": 5.17e-05, "loss": 2.9309, "step": 520 }, { "epoch": 7.68, "learning_rate": 5.270000000000001e-05, "loss": 3.0513, "step": 530 }, { "epoch": 7.83, "learning_rate": 5.3700000000000004e-05, "loss": 2.948, "step": 540 }, { "epoch": 7.97, "learning_rate": 5.470000000000001e-05, "loss": 2.9518, "step": 550 }, { "epoch": 8.12, "learning_rate": 5.5700000000000005e-05, "loss": 2.9229, "step": 560 }, { "epoch": 8.26, "learning_rate": 5.6699999999999996e-05, "loss": 2.9511, "step": 570 }, { "epoch": 8.41, "learning_rate": 5.77e-05, "loss": 2.9427, "step": 580 }, { "epoch": 8.55, "learning_rate": 5.87e-05, "loss": 2.9603, "step": 590 }, { "epoch": 8.7, "learning_rate": 5.97e-05, "loss": 2.9397, "step": 600 }, { "epoch": 8.84, "learning_rate": 6.07e-05, "loss": 2.9267, "step": 610 }, { "epoch": 8.99, "learning_rate": 6.170000000000001e-05, "loss": 3.0164, "step": 620 }, { "epoch": 9.13, "learning_rate": 6.27e-05, "loss": 2.913, "step": 630 }, { "epoch": 9.28, "learning_rate": 6.37e-05, "loss": 2.9382, "step": 640 }, { "epoch": 9.42, "learning_rate": 6.47e-05, "loss": 2.9905, "step": 650 }, { "epoch": 9.57, "learning_rate": 6.570000000000001e-05, "loss": 2.8993, "step": 660 }, { "epoch": 9.71, "learning_rate": 6.670000000000001e-05, "loss": 2.9829, "step": 670 }, { "epoch": 9.86, "learning_rate": 6.77e-05, "loss": 2.9161, "step": 680 }, { "epoch": 10.0, "learning_rate": 6.87e-05, "loss": 3.0016, "step": 690 }, { "epoch": 10.14, "learning_rate": 6.97e-05, "loss": 2.9328, "step": 700 }, { "epoch": 10.29, "learning_rate": 7.07e-05, "loss": 3.0272, "step": 710 }, { "epoch": 10.43, "learning_rate": 7.17e-05, "loss": 2.9243, "step": 720 }, { "epoch": 10.58, "learning_rate": 7.27e-05, "loss": 2.9345, "step": 730 }, { "epoch": 10.72, "learning_rate": 7.37e-05, "loss": 2.9191, "step": 740 }, { "epoch": 10.87, "learning_rate": 7.47e-05, "loss": 2.8938, "step": 750 }, { "epoch": 11.01, "learning_rate": 7.570000000000001e-05, "loss": 2.9083, "step": 760 }, { "epoch": 11.16, "learning_rate": 7.670000000000001e-05, "loss": 2.9408, "step": 770 }, { "epoch": 11.3, "learning_rate": 7.77e-05, "loss": 2.9428, "step": 780 }, { "epoch": 11.45, "learning_rate": 7.87e-05, "loss": 2.9443, "step": 790 }, { "epoch": 11.59, "learning_rate": 7.970000000000001e-05, "loss": 2.9159, "step": 800 }, { "epoch": 11.74, "learning_rate": 8.070000000000001e-05, "loss": 2.9294, "step": 810 }, { "epoch": 11.88, "learning_rate": 8.17e-05, "loss": 2.8927, "step": 820 }, { "epoch": 12.03, "learning_rate": 8.27e-05, "loss": 2.9385, "step": 830 }, { "epoch": 12.17, "learning_rate": 8.37e-05, "loss": 2.9407, "step": 840 }, { "epoch": 12.32, "learning_rate": 8.47e-05, "loss": 2.8872, "step": 850 }, { "epoch": 12.46, "learning_rate": 8.57e-05, "loss": 2.9336, "step": 860 }, { "epoch": 12.61, "learning_rate": 8.67e-05, "loss": 2.8768, "step": 870 }, { "epoch": 12.75, "learning_rate": 8.77e-05, "loss": 2.9426, "step": 880 }, { "epoch": 12.9, "learning_rate": 8.87e-05, "loss": 2.8981, "step": 890 }, { "epoch": 13.04, "learning_rate": 8.970000000000001e-05, "loss": 2.8928, "step": 900 }, { "epoch": 13.19, "learning_rate": 9.070000000000001e-05, "loss": 2.9346, "step": 910 }, { "epoch": 13.33, "learning_rate": 9.17e-05, "loss": 2.89, "step": 920 }, { "epoch": 13.48, "learning_rate": 9.27e-05, "loss": 2.9107, "step": 930 }, { "epoch": 13.62, "learning_rate": 9.370000000000001e-05, "loss": 2.8732, "step": 940 }, { "epoch": 13.77, "learning_rate": 9.47e-05, "loss": 2.9441, "step": 950 }, { "epoch": 13.91, "learning_rate": 9.57e-05, "loss": 2.9136, "step": 960 }, { "epoch": 14.06, "learning_rate": 9.67e-05, "loss": 2.8773, "step": 970 }, { "epoch": 14.2, "learning_rate": 9.77e-05, "loss": 2.9015, "step": 980 }, { "epoch": 14.35, "learning_rate": 9.87e-05, "loss": 2.9168, "step": 990 }, { "epoch": 14.49, "learning_rate": 9.970000000000001e-05, "loss": 2.9241, "step": 1000 }, { "epoch": 14.49, "eval_loss": 2.8734302520751953, "eval_runtime": 570.6581, "eval_samples_per_second": 5.948, "eval_steps_per_second": 0.745, "eval_wer": 1.0, "step": 1000 }, { "epoch": 14.64, "learning_rate": 9.988135593220339e-05, "loss": 2.8693, "step": 1010 }, { "epoch": 14.78, "learning_rate": 9.971186440677967e-05, "loss": 2.9199, "step": 1020 }, { "epoch": 14.93, "learning_rate": 9.954237288135594e-05, "loss": 2.8831, "step": 1030 }, { "epoch": 15.07, "learning_rate": 9.937288135593222e-05, "loss": 2.8716, "step": 1040 }, { "epoch": 15.22, "learning_rate": 9.920338983050847e-05, "loss": 2.9457, "step": 1050 }, { "epoch": 15.36, "learning_rate": 9.903389830508475e-05, "loss": 2.8708, "step": 1060 }, { "epoch": 15.51, "learning_rate": 9.886440677966103e-05, "loss": 2.8595, "step": 1070 }, { "epoch": 15.65, "learning_rate": 9.86949152542373e-05, "loss": 2.8718, "step": 1080 }, { "epoch": 15.8, "learning_rate": 9.852542372881356e-05, "loss": 2.8383, "step": 1090 }, { "epoch": 15.94, "learning_rate": 9.835593220338983e-05, "loss": 2.8016, "step": 1100 }, { "epoch": 16.09, "learning_rate": 9.818644067796611e-05, "loss": 2.7608, "step": 1110 }, { "epoch": 16.23, "learning_rate": 9.801694915254239e-05, "loss": 2.7417, "step": 1120 }, { "epoch": 16.38, "learning_rate": 9.784745762711864e-05, "loss": 2.6906, "step": 1130 }, { "epoch": 16.52, "learning_rate": 9.767796610169492e-05, "loss": 2.5662, "step": 1140 }, { "epoch": 16.67, "learning_rate": 9.750847457627119e-05, "loss": 2.4202, "step": 1150 }, { "epoch": 16.81, "learning_rate": 9.733898305084747e-05, "loss": 2.4102, "step": 1160 }, { "epoch": 16.96, "learning_rate": 9.716949152542373e-05, "loss": 2.185, "step": 1170 }, { "epoch": 17.1, "learning_rate": 9.7e-05, "loss": 1.9315, "step": 1180 }, { "epoch": 17.25, "learning_rate": 9.683050847457628e-05, "loss": 2.0025, "step": 1190 }, { "epoch": 17.39, "learning_rate": 9.666101694915255e-05, "loss": 1.8797, "step": 1200 }, { "epoch": 17.54, "learning_rate": 9.649152542372883e-05, "loss": 1.7524, "step": 1210 }, { "epoch": 17.68, "learning_rate": 9.632203389830509e-05, "loss": 1.7243, "step": 1220 }, { "epoch": 17.83, "learning_rate": 9.615254237288136e-05, "loss": 1.5859, "step": 1230 }, { "epoch": 17.97, "learning_rate": 9.598305084745764e-05, "loss": 1.5059, "step": 1240 }, { "epoch": 18.12, "learning_rate": 9.58135593220339e-05, "loss": 1.456, "step": 1250 }, { "epoch": 18.26, "learning_rate": 9.564406779661017e-05, "loss": 1.535, "step": 1260 }, { "epoch": 18.41, "learning_rate": 9.547457627118644e-05, "loss": 1.4808, "step": 1270 }, { "epoch": 18.55, "learning_rate": 9.530508474576272e-05, "loss": 1.4568, "step": 1280 }, { "epoch": 18.7, "learning_rate": 9.5135593220339e-05, "loss": 1.3898, "step": 1290 }, { "epoch": 18.84, "learning_rate": 9.496610169491525e-05, "loss": 1.4865, "step": 1300 }, { "epoch": 18.99, "learning_rate": 9.479661016949153e-05, "loss": 1.391, "step": 1310 }, { "epoch": 19.13, "learning_rate": 9.46271186440678e-05, "loss": 1.4236, "step": 1320 }, { "epoch": 19.28, "learning_rate": 9.445762711864408e-05, "loss": 1.303, "step": 1330 }, { "epoch": 19.42, "learning_rate": 9.428813559322034e-05, "loss": 1.2304, "step": 1340 }, { "epoch": 19.57, "learning_rate": 9.411864406779661e-05, "loss": 1.2014, "step": 1350 }, { "epoch": 19.71, "learning_rate": 9.394915254237289e-05, "loss": 1.2398, "step": 1360 }, { "epoch": 19.86, "learning_rate": 9.377966101694916e-05, "loss": 1.1895, "step": 1370 }, { "epoch": 20.0, "learning_rate": 9.361016949152542e-05, "loss": 1.2296, "step": 1380 }, { "epoch": 20.14, "learning_rate": 9.34406779661017e-05, "loss": 1.3624, "step": 1390 }, { "epoch": 20.29, "learning_rate": 9.327118644067797e-05, "loss": 1.1805, "step": 1400 }, { "epoch": 20.43, "learning_rate": 9.310169491525425e-05, "loss": 1.1705, "step": 1410 }, { "epoch": 20.58, "learning_rate": 9.29322033898305e-05, "loss": 1.1446, "step": 1420 }, { "epoch": 20.72, "learning_rate": 9.276271186440678e-05, "loss": 1.2016, "step": 1430 }, { "epoch": 20.87, "learning_rate": 9.259322033898306e-05, "loss": 1.1053, "step": 1440 }, { "epoch": 21.01, "learning_rate": 9.242372881355933e-05, "loss": 1.1159, "step": 1450 }, { "epoch": 21.16, "learning_rate": 9.225423728813561e-05, "loss": 1.0629, "step": 1460 }, { "epoch": 21.3, "learning_rate": 9.208474576271186e-05, "loss": 1.1606, "step": 1470 }, { "epoch": 21.45, "learning_rate": 9.191525423728814e-05, "loss": 1.051, "step": 1480 }, { "epoch": 21.59, "learning_rate": 9.174576271186442e-05, "loss": 1.0948, "step": 1490 }, { "epoch": 21.74, "learning_rate": 9.157627118644069e-05, "loss": 1.0766, "step": 1500 }, { "epoch": 21.74, "eval_loss": 0.27734482288360596, "eval_runtime": 571.274, "eval_samples_per_second": 5.941, "eval_steps_per_second": 0.744, "eval_wer": 0.2488228188250376, "step": 1500 }, { "epoch": 21.88, "learning_rate": 9.140677966101695e-05, "loss": 0.9907, "step": 1510 }, { "epoch": 22.03, "learning_rate": 9.123728813559322e-05, "loss": 1.0756, "step": 1520 }, { "epoch": 22.17, "learning_rate": 9.10677966101695e-05, "loss": 1.0947, "step": 1530 }, { "epoch": 22.32, "learning_rate": 9.089830508474577e-05, "loss": 1.1119, "step": 1540 }, { "epoch": 22.46, "learning_rate": 9.072881355932203e-05, "loss": 1.039, "step": 1550 }, { "epoch": 22.61, "learning_rate": 9.055932203389831e-05, "loss": 0.9631, "step": 1560 }, { "epoch": 22.75, "learning_rate": 9.038983050847458e-05, "loss": 0.9765, "step": 1570 }, { "epoch": 22.9, "learning_rate": 9.022033898305086e-05, "loss": 0.9932, "step": 1580 }, { "epoch": 23.04, "learning_rate": 9.005084745762712e-05, "loss": 0.9384, "step": 1590 }, { "epoch": 23.19, "learning_rate": 8.988135593220339e-05, "loss": 0.9942, "step": 1600 }, { "epoch": 23.33, "learning_rate": 8.971186440677967e-05, "loss": 0.9546, "step": 1610 }, { "epoch": 23.48, "learning_rate": 8.954237288135594e-05, "loss": 1.0051, "step": 1620 }, { "epoch": 23.62, "learning_rate": 8.93728813559322e-05, "loss": 0.9648, "step": 1630 }, { "epoch": 23.77, "learning_rate": 8.920338983050848e-05, "loss": 1.0198, "step": 1640 }, { "epoch": 23.91, "learning_rate": 8.903389830508475e-05, "loss": 0.9773, "step": 1650 }, { "epoch": 24.06, "learning_rate": 8.886440677966103e-05, "loss": 0.9714, "step": 1660 }, { "epoch": 24.2, "learning_rate": 8.869491525423728e-05, "loss": 0.987, "step": 1670 }, { "epoch": 24.35, "learning_rate": 8.852542372881356e-05, "loss": 1.0105, "step": 1680 }, { "epoch": 24.49, "learning_rate": 8.835593220338983e-05, "loss": 0.9965, "step": 1690 }, { "epoch": 24.64, "learning_rate": 8.818644067796611e-05, "loss": 0.9293, "step": 1700 }, { "epoch": 24.78, "learning_rate": 8.801694915254238e-05, "loss": 0.8939, "step": 1710 }, { "epoch": 24.93, "learning_rate": 8.784745762711864e-05, "loss": 0.9721, "step": 1720 }, { "epoch": 25.07, "learning_rate": 8.767796610169492e-05, "loss": 1.0158, "step": 1730 }, { "epoch": 25.22, "learning_rate": 8.750847457627119e-05, "loss": 1.08, "step": 1740 }, { "epoch": 25.36, "learning_rate": 8.733898305084747e-05, "loss": 1.3222, "step": 1750 }, { "epoch": 25.51, "learning_rate": 8.716949152542373e-05, "loss": 0.9776, "step": 1760 }, { "epoch": 25.65, "learning_rate": 8.7e-05, "loss": 0.9089, "step": 1770 }, { "epoch": 25.8, "learning_rate": 8.683050847457628e-05, "loss": 0.9171, "step": 1780 }, { "epoch": 25.94, "learning_rate": 8.666101694915255e-05, "loss": 0.9366, "step": 1790 }, { "epoch": 26.09, "learning_rate": 8.649152542372881e-05, "loss": 0.8907, "step": 1800 }, { "epoch": 26.23, "learning_rate": 8.632203389830509e-05, "loss": 0.8427, "step": 1810 }, { "epoch": 26.38, "learning_rate": 8.615254237288136e-05, "loss": 0.8692, "step": 1820 }, { "epoch": 26.52, "learning_rate": 8.598305084745764e-05, "loss": 0.8933, "step": 1830 }, { "epoch": 26.67, "learning_rate": 8.581355932203389e-05, "loss": 0.9524, "step": 1840 }, { "epoch": 26.81, "learning_rate": 8.564406779661017e-05, "loss": 0.9623, "step": 1850 }, { "epoch": 26.96, "learning_rate": 8.547457627118645e-05, "loss": 0.9458, "step": 1860 }, { "epoch": 27.1, "learning_rate": 8.530508474576272e-05, "loss": 0.9299, "step": 1870 }, { "epoch": 27.25, "learning_rate": 8.513559322033898e-05, "loss": 0.8612, "step": 1880 }, { "epoch": 27.39, "learning_rate": 8.496610169491525e-05, "loss": 0.9352, "step": 1890 }, { "epoch": 27.54, "learning_rate": 8.479661016949153e-05, "loss": 0.8692, "step": 1900 }, { "epoch": 27.68, "learning_rate": 8.46271186440678e-05, "loss": 0.8834, "step": 1910 }, { "epoch": 27.83, "learning_rate": 8.445762711864406e-05, "loss": 0.8448, "step": 1920 }, { "epoch": 27.97, "learning_rate": 8.428813559322034e-05, "loss": 0.9122, "step": 1930 }, { "epoch": 28.12, "learning_rate": 8.411864406779661e-05, "loss": 0.8632, "step": 1940 }, { "epoch": 28.26, "learning_rate": 8.394915254237289e-05, "loss": 0.8578, "step": 1950 }, { "epoch": 28.41, "learning_rate": 8.377966101694916e-05, "loss": 0.8758, "step": 1960 }, { "epoch": 28.55, "learning_rate": 8.361016949152542e-05, "loss": 0.9024, "step": 1970 }, { "epoch": 28.7, "learning_rate": 8.34406779661017e-05, "loss": 1.1392, "step": 1980 }, { "epoch": 28.84, "learning_rate": 8.327118644067797e-05, "loss": 0.8794, "step": 1990 }, { "epoch": 28.99, "learning_rate": 8.310169491525425e-05, "loss": 0.8416, "step": 2000 }, { "epoch": 28.99, "eval_loss": 0.222377210855484, "eval_runtime": 567.7022, "eval_samples_per_second": 5.978, "eval_steps_per_second": 0.749, "eval_wer": 0.19897443482977098, "step": 2000 }, { "epoch": 29.13, "learning_rate": 8.293220338983052e-05, "loss": 0.8217, "step": 2010 }, { "epoch": 29.28, "learning_rate": 8.276271186440678e-05, "loss": 0.9189, "step": 2020 }, { "epoch": 29.42, "learning_rate": 8.259322033898306e-05, "loss": 0.8251, "step": 2030 }, { "epoch": 29.57, "learning_rate": 8.242372881355933e-05, "loss": 0.9254, "step": 2040 }, { "epoch": 29.71, "learning_rate": 8.22542372881356e-05, "loss": 0.9276, "step": 2050 }, { "epoch": 29.86, "learning_rate": 8.208474576271186e-05, "loss": 0.8893, "step": 2060 }, { "epoch": 30.0, "learning_rate": 8.191525423728814e-05, "loss": 0.7608, "step": 2070 }, { "epoch": 30.14, "learning_rate": 8.174576271186442e-05, "loss": 0.7857, "step": 2080 }, { "epoch": 30.29, "learning_rate": 8.157627118644067e-05, "loss": 0.9132, "step": 2090 }, { "epoch": 30.43, "learning_rate": 8.140677966101695e-05, "loss": 0.8814, "step": 2100 }, { "epoch": 30.58, "learning_rate": 8.123728813559322e-05, "loss": 0.901, "step": 2110 }, { "epoch": 30.72, "learning_rate": 8.10677966101695e-05, "loss": 0.8746, "step": 2120 }, { "epoch": 30.87, "learning_rate": 8.089830508474577e-05, "loss": 0.9291, "step": 2130 }, { "epoch": 31.01, "learning_rate": 8.072881355932203e-05, "loss": 0.8357, "step": 2140 }, { "epoch": 31.16, "learning_rate": 8.055932203389831e-05, "loss": 0.8305, "step": 2150 }, { "epoch": 31.3, "learning_rate": 8.038983050847458e-05, "loss": 0.8275, "step": 2160 }, { "epoch": 31.45, "learning_rate": 8.022033898305085e-05, "loss": 0.8111, "step": 2170 }, { "epoch": 31.59, "learning_rate": 8.005084745762713e-05, "loss": 0.8109, "step": 2180 }, { "epoch": 31.74, "learning_rate": 7.988135593220339e-05, "loss": 0.8761, "step": 2190 }, { "epoch": 31.88, "learning_rate": 7.971186440677967e-05, "loss": 0.8179, "step": 2200 }, { "epoch": 32.03, "learning_rate": 7.954237288135592e-05, "loss": 0.8078, "step": 2210 }, { "epoch": 32.17, "learning_rate": 7.93728813559322e-05, "loss": 0.8174, "step": 2220 }, { "epoch": 32.32, "learning_rate": 7.920338983050848e-05, "loss": 0.8176, "step": 2230 }, { "epoch": 32.46, "learning_rate": 7.903389830508475e-05, "loss": 0.729, "step": 2240 }, { "epoch": 32.61, "learning_rate": 7.886440677966102e-05, "loss": 0.7884, "step": 2250 }, { "epoch": 32.75, "learning_rate": 7.869491525423728e-05, "loss": 0.7896, "step": 2260 }, { "epoch": 32.9, "learning_rate": 7.852542372881356e-05, "loss": 0.8016, "step": 2270 }, { "epoch": 33.04, "learning_rate": 7.835593220338984e-05, "loss": 0.7635, "step": 2280 }, { "epoch": 33.19, "learning_rate": 7.818644067796611e-05, "loss": 0.8308, "step": 2290 }, { "epoch": 33.33, "learning_rate": 7.801694915254238e-05, "loss": 0.7543, "step": 2300 }, { "epoch": 33.48, "learning_rate": 7.784745762711864e-05, "loss": 0.8617, "step": 2310 }, { "epoch": 33.62, "learning_rate": 7.767796610169492e-05, "loss": 0.8434, "step": 2320 }, { "epoch": 33.77, "learning_rate": 7.750847457627119e-05, "loss": 0.948, "step": 2330 }, { "epoch": 33.91, "learning_rate": 7.733898305084746e-05, "loss": 1.0185, "step": 2340 }, { "epoch": 34.06, "learning_rate": 7.716949152542374e-05, "loss": 0.827, "step": 2350 }, { "epoch": 34.2, "learning_rate": 7.7e-05, "loss": 0.7916, "step": 2360 }, { "epoch": 34.35, "learning_rate": 7.683050847457628e-05, "loss": 0.7858, "step": 2370 }, { "epoch": 34.49, "learning_rate": 7.666101694915255e-05, "loss": 0.8583, "step": 2380 }, { "epoch": 34.64, "learning_rate": 7.649152542372881e-05, "loss": 0.7672, "step": 2390 }, { "epoch": 34.78, "learning_rate": 7.63220338983051e-05, "loss": 0.7745, "step": 2400 }, { "epoch": 34.93, "learning_rate": 7.615254237288136e-05, "loss": 0.8104, "step": 2410 }, { "epoch": 35.07, "learning_rate": 7.598305084745763e-05, "loss": 0.7519, "step": 2420 }, { "epoch": 35.22, "learning_rate": 7.58135593220339e-05, "loss": 0.7867, "step": 2430 }, { "epoch": 35.36, "learning_rate": 7.564406779661017e-05, "loss": 0.7614, "step": 2440 }, { "epoch": 35.51, "learning_rate": 7.547457627118645e-05, "loss": 0.8043, "step": 2450 }, { "epoch": 35.65, "learning_rate": 7.53050847457627e-05, "loss": 0.7866, "step": 2460 }, { "epoch": 35.8, "learning_rate": 7.513559322033899e-05, "loss": 0.8356, "step": 2470 }, { "epoch": 35.94, "learning_rate": 7.496610169491525e-05, "loss": 0.761, "step": 2480 }, { "epoch": 36.09, "learning_rate": 7.479661016949153e-05, "loss": 0.9027, "step": 2490 }, { "epoch": 36.23, "learning_rate": 7.46271186440678e-05, "loss": 0.8048, "step": 2500 }, { "epoch": 36.23, "eval_loss": 0.20626655220985413, "eval_runtime": 571.29, "eval_samples_per_second": 5.941, "eval_steps_per_second": 0.744, "eval_wer": 0.17916985101364954, "step": 2500 }, { "epoch": 36.38, "learning_rate": 7.445762711864406e-05, "loss": 0.7866, "step": 2510 }, { "epoch": 36.52, "learning_rate": 7.428813559322034e-05, "loss": 0.7869, "step": 2520 }, { "epoch": 36.67, "learning_rate": 7.411864406779661e-05, "loss": 0.8445, "step": 2530 }, { "epoch": 36.81, "learning_rate": 7.394915254237289e-05, "loss": 0.7745, "step": 2540 }, { "epoch": 36.96, "learning_rate": 7.377966101694916e-05, "loss": 0.7888, "step": 2550 }, { "epoch": 37.1, "learning_rate": 7.361016949152542e-05, "loss": 0.776, "step": 2560 }, { "epoch": 37.25, "learning_rate": 7.34406779661017e-05, "loss": 0.8096, "step": 2570 }, { "epoch": 37.39, "learning_rate": 7.327118644067797e-05, "loss": 0.7561, "step": 2580 }, { "epoch": 37.54, "learning_rate": 7.310169491525424e-05, "loss": 0.7251, "step": 2590 }, { "epoch": 37.68, "learning_rate": 7.293220338983052e-05, "loss": 0.8062, "step": 2600 }, { "epoch": 37.83, "learning_rate": 7.276271186440678e-05, "loss": 0.7388, "step": 2610 }, { "epoch": 37.97, "learning_rate": 7.259322033898306e-05, "loss": 0.7616, "step": 2620 }, { "epoch": 38.12, "learning_rate": 7.242372881355932e-05, "loss": 0.7593, "step": 2630 }, { "epoch": 38.26, "learning_rate": 7.22542372881356e-05, "loss": 0.7399, "step": 2640 }, { "epoch": 38.41, "learning_rate": 7.208474576271188e-05, "loss": 0.7559, "step": 2650 }, { "epoch": 38.55, "learning_rate": 7.191525423728814e-05, "loss": 0.7559, "step": 2660 }, { "epoch": 38.7, "learning_rate": 7.174576271186441e-05, "loss": 0.7218, "step": 2670 }, { "epoch": 38.84, "learning_rate": 7.157627118644067e-05, "loss": 0.7991, "step": 2680 }, { "epoch": 38.99, "learning_rate": 7.140677966101695e-05, "loss": 0.7493, "step": 2690 }, { "epoch": 39.13, "learning_rate": 7.123728813559322e-05, "loss": 0.7502, "step": 2700 }, { "epoch": 39.28, "learning_rate": 7.106779661016949e-05, "loss": 0.7883, "step": 2710 }, { "epoch": 39.42, "learning_rate": 7.089830508474577e-05, "loss": 0.8236, "step": 2720 }, { "epoch": 39.57, "learning_rate": 7.072881355932203e-05, "loss": 0.8141, "step": 2730 }, { "epoch": 39.71, "learning_rate": 7.055932203389831e-05, "loss": 0.7956, "step": 2740 }, { "epoch": 39.86, "learning_rate": 7.038983050847458e-05, "loss": 0.7714, "step": 2750 }, { "epoch": 40.0, "learning_rate": 7.022033898305085e-05, "loss": 0.6689, "step": 2760 }, { "epoch": 40.14, "learning_rate": 7.005084745762713e-05, "loss": 0.7202, "step": 2770 }, { "epoch": 40.29, "learning_rate": 6.988135593220339e-05, "loss": 0.7278, "step": 2780 }, { "epoch": 40.43, "learning_rate": 6.971186440677966e-05, "loss": 0.801, "step": 2790 }, { "epoch": 40.58, "learning_rate": 6.954237288135594e-05, "loss": 0.7865, "step": 2800 }, { "epoch": 40.72, "learning_rate": 6.93728813559322e-05, "loss": 0.6811, "step": 2810 }, { "epoch": 40.87, "learning_rate": 6.920338983050849e-05, "loss": 0.7933, "step": 2820 }, { "epoch": 41.01, "learning_rate": 6.903389830508475e-05, "loss": 0.7826, "step": 2830 }, { "epoch": 41.16, "learning_rate": 6.886440677966102e-05, "loss": 0.7142, "step": 2840 }, { "epoch": 41.3, "learning_rate": 6.869491525423728e-05, "loss": 0.8347, "step": 2850 }, { "epoch": 41.45, "learning_rate": 6.852542372881356e-05, "loss": 0.7712, "step": 2860 }, { "epoch": 41.59, "learning_rate": 6.835593220338984e-05, "loss": 0.7458, "step": 2870 }, { "epoch": 41.74, "learning_rate": 6.81864406779661e-05, "loss": 0.7528, "step": 2880 }, { "epoch": 41.88, "learning_rate": 6.801694915254238e-05, "loss": 0.7699, "step": 2890 }, { "epoch": 42.03, "learning_rate": 6.784745762711864e-05, "loss": 0.8577, "step": 2900 }, { "epoch": 42.17, "learning_rate": 6.767796610169492e-05, "loss": 0.7259, "step": 2910 }, { "epoch": 42.32, "learning_rate": 6.750847457627119e-05, "loss": 0.7039, "step": 2920 }, { "epoch": 42.46, "learning_rate": 6.733898305084746e-05, "loss": 0.8307, "step": 2930 }, { "epoch": 42.61, "learning_rate": 6.716949152542374e-05, "loss": 0.7952, "step": 2940 }, { "epoch": 42.75, "learning_rate": 6.7e-05, "loss": 0.6789, "step": 2950 }, { "epoch": 42.9, "learning_rate": 6.683050847457627e-05, "loss": 0.7708, "step": 2960 }, { "epoch": 43.04, "learning_rate": 6.666101694915255e-05, "loss": 0.6823, "step": 2970 }, { "epoch": 43.19, "learning_rate": 6.649152542372881e-05, "loss": 0.6989, "step": 2980 }, { "epoch": 43.33, "learning_rate": 6.63220338983051e-05, "loss": 0.7944, "step": 2990 }, { "epoch": 43.48, "learning_rate": 6.615254237288135e-05, "loss": 0.7664, "step": 3000 }, { "epoch": 43.48, "eval_loss": 0.2088436335325241, "eval_runtime": 572.961, "eval_samples_per_second": 5.924, "eval_steps_per_second": 0.742, "eval_wer": 0.17475696242059677, "step": 3000 }, { "epoch": 43.62, "learning_rate": 6.598305084745763e-05, "loss": 0.7256, "step": 3010 }, { "epoch": 43.77, "learning_rate": 6.581355932203391e-05, "loss": 0.7146, "step": 3020 }, { "epoch": 43.91, "learning_rate": 6.564406779661017e-05, "loss": 0.7347, "step": 3030 }, { "epoch": 44.06, "learning_rate": 6.547457627118644e-05, "loss": 0.7302, "step": 3040 }, { "epoch": 44.2, "learning_rate": 6.53050847457627e-05, "loss": 0.7772, "step": 3050 }, { "epoch": 44.35, "learning_rate": 6.513559322033899e-05, "loss": 0.7214, "step": 3060 }, { "epoch": 44.49, "learning_rate": 6.496610169491527e-05, "loss": 0.6899, "step": 3070 }, { "epoch": 44.64, "learning_rate": 6.479661016949153e-05, "loss": 0.7488, "step": 3080 }, { "epoch": 44.78, "learning_rate": 6.46271186440678e-05, "loss": 0.831, "step": 3090 }, { "epoch": 44.93, "learning_rate": 6.445762711864407e-05, "loss": 0.799, "step": 3100 }, { "epoch": 45.07, "learning_rate": 6.428813559322035e-05, "loss": 0.7195, "step": 3110 }, { "epoch": 45.22, "learning_rate": 6.411864406779661e-05, "loss": 0.7476, "step": 3120 }, { "epoch": 45.36, "learning_rate": 6.394915254237288e-05, "loss": 0.7475, "step": 3130 }, { "epoch": 45.51, "learning_rate": 6.377966101694916e-05, "loss": 0.767, "step": 3140 }, { "epoch": 45.65, "learning_rate": 6.361016949152542e-05, "loss": 0.7263, "step": 3150 }, { "epoch": 45.8, "learning_rate": 6.34406779661017e-05, "loss": 0.7423, "step": 3160 }, { "epoch": 45.94, "learning_rate": 6.327118644067797e-05, "loss": 0.7693, "step": 3170 }, { "epoch": 46.09, "learning_rate": 6.310169491525424e-05, "loss": 0.8136, "step": 3180 }, { "epoch": 46.23, "learning_rate": 6.293220338983052e-05, "loss": 0.6705, "step": 3190 }, { "epoch": 46.38, "learning_rate": 6.276271186440678e-05, "loss": 0.7041, "step": 3200 }, { "epoch": 46.52, "learning_rate": 6.259322033898305e-05, "loss": 0.7478, "step": 3210 }, { "epoch": 46.67, "learning_rate": 6.242372881355933e-05, "loss": 0.6909, "step": 3220 }, { "epoch": 46.81, "learning_rate": 6.22542372881356e-05, "loss": 0.7702, "step": 3230 }, { "epoch": 46.96, "learning_rate": 6.208474576271188e-05, "loss": 0.7302, "step": 3240 }, { "epoch": 47.1, "learning_rate": 6.191525423728813e-05, "loss": 0.7812, "step": 3250 }, { "epoch": 47.25, "learning_rate": 6.174576271186441e-05, "loss": 0.7043, "step": 3260 }, { "epoch": 47.39, "learning_rate": 6.157627118644068e-05, "loss": 0.7645, "step": 3270 }, { "epoch": 47.54, "learning_rate": 6.140677966101696e-05, "loss": 0.8539, "step": 3280 }, { "epoch": 47.68, "learning_rate": 6.123728813559322e-05, "loss": 0.7747, "step": 3290 }, { "epoch": 47.83, "learning_rate": 6.106779661016949e-05, "loss": 0.6613, "step": 3300 }, { "epoch": 47.97, "learning_rate": 6.089830508474577e-05, "loss": 0.7399, "step": 3310 }, { "epoch": 48.12, "learning_rate": 6.072881355932204e-05, "loss": 0.6675, "step": 3320 }, { "epoch": 48.26, "learning_rate": 6.05593220338983e-05, "loss": 0.6699, "step": 3330 }, { "epoch": 48.41, "learning_rate": 6.0389830508474574e-05, "loss": 0.7534, "step": 3340 }, { "epoch": 48.55, "learning_rate": 6.022033898305085e-05, "loss": 0.7837, "step": 3350 }, { "epoch": 48.7, "learning_rate": 6.005084745762713e-05, "loss": 0.7066, "step": 3360 }, { "epoch": 48.84, "learning_rate": 5.98813559322034e-05, "loss": 0.7931, "step": 3370 }, { "epoch": 48.99, "learning_rate": 5.971186440677966e-05, "loss": 0.7575, "step": 3380 }, { "epoch": 49.13, "learning_rate": 5.954237288135593e-05, "loss": 0.6951, "step": 3390 }, { "epoch": 49.28, "learning_rate": 5.9372881355932206e-05, "loss": 0.788, "step": 3400 }, { "epoch": 49.42, "learning_rate": 5.920338983050848e-05, "loss": 0.7727, "step": 3410 }, { "epoch": 49.57, "learning_rate": 5.9033898305084746e-05, "loss": 0.7102, "step": 3420 }, { "epoch": 49.71, "learning_rate": 5.886440677966102e-05, "loss": 0.6733, "step": 3430 }, { "epoch": 49.86, "learning_rate": 5.869491525423729e-05, "loss": 0.6866, "step": 3440 }, { "epoch": 50.0, "learning_rate": 5.8525423728813565e-05, "loss": 0.7414, "step": 3450 }, { "epoch": 50.14, "learning_rate": 5.835593220338983e-05, "loss": 0.684, "step": 3460 }, { "epoch": 50.29, "learning_rate": 5.8186440677966105e-05, "loss": 0.7363, "step": 3470 }, { "epoch": 50.43, "learning_rate": 5.801694915254238e-05, "loss": 0.7421, "step": 3480 }, { "epoch": 50.58, "learning_rate": 5.784745762711865e-05, "loss": 0.7038, "step": 3490 }, { "epoch": 50.72, "learning_rate": 5.767796610169491e-05, "loss": 0.6571, "step": 3500 }, { "epoch": 50.72, "eval_loss": 0.20422525703907013, "eval_runtime": 570.0136, "eval_samples_per_second": 5.954, "eval_steps_per_second": 0.746, "eval_wer": 0.166761169867449, "step": 3500 }, { "epoch": 50.87, "learning_rate": 5.750847457627119e-05, "loss": 0.7286, "step": 3510 }, { "epoch": 51.01, "learning_rate": 5.7338983050847464e-05, "loss": 0.6604, "step": 3520 }, { "epoch": 51.16, "learning_rate": 5.716949152542374e-05, "loss": 0.7662, "step": 3530 }, { "epoch": 51.3, "learning_rate": 5.6999999999999996e-05, "loss": 0.7688, "step": 3540 }, { "epoch": 51.45, "learning_rate": 5.683050847457627e-05, "loss": 0.6903, "step": 3550 }, { "epoch": 51.59, "learning_rate": 5.666101694915254e-05, "loss": 0.6939, "step": 3560 }, { "epoch": 51.74, "learning_rate": 5.649152542372882e-05, "loss": 0.729, "step": 3570 }, { "epoch": 51.88, "learning_rate": 5.632203389830508e-05, "loss": 0.7307, "step": 3580 }, { "epoch": 52.03, "learning_rate": 5.6152542372881355e-05, "loss": 0.6863, "step": 3590 }, { "epoch": 52.17, "learning_rate": 5.598305084745763e-05, "loss": 0.6843, "step": 3600 }, { "epoch": 52.32, "learning_rate": 5.58135593220339e-05, "loss": 0.7298, "step": 3610 }, { "epoch": 52.46, "learning_rate": 5.5644067796610175e-05, "loss": 0.7545, "step": 3620 }, { "epoch": 52.61, "learning_rate": 5.547457627118644e-05, "loss": 0.6663, "step": 3630 }, { "epoch": 52.75, "learning_rate": 5.5305084745762714e-05, "loss": 0.6822, "step": 3640 }, { "epoch": 52.9, "learning_rate": 5.513559322033899e-05, "loss": 0.806, "step": 3650 }, { "epoch": 53.04, "learning_rate": 5.496610169491526e-05, "loss": 0.6597, "step": 3660 }, { "epoch": 53.19, "learning_rate": 5.479661016949153e-05, "loss": 0.7131, "step": 3670 }, { "epoch": 53.33, "learning_rate": 5.46271186440678e-05, "loss": 0.7221, "step": 3680 }, { "epoch": 53.48, "learning_rate": 5.445762711864407e-05, "loss": 0.6235, "step": 3690 }, { "epoch": 53.62, "learning_rate": 5.4288135593220346e-05, "loss": 0.6547, "step": 3700 }, { "epoch": 53.77, "learning_rate": 5.4118644067796606e-05, "loss": 0.7014, "step": 3710 }, { "epoch": 53.91, "learning_rate": 5.3949152542372886e-05, "loss": 0.6826, "step": 3720 }, { "epoch": 54.06, "learning_rate": 5.377966101694916e-05, "loss": 0.664, "step": 3730 }, { "epoch": 54.2, "learning_rate": 5.361016949152543e-05, "loss": 0.7201, "step": 3740 }, { "epoch": 54.35, "learning_rate": 5.344067796610169e-05, "loss": 0.6765, "step": 3750 }, { "epoch": 54.49, "learning_rate": 5.3271186440677965e-05, "loss": 0.6541, "step": 3760 }, { "epoch": 54.64, "learning_rate": 5.310169491525424e-05, "loss": 0.7299, "step": 3770 }, { "epoch": 54.78, "learning_rate": 5.293220338983051e-05, "loss": 0.7139, "step": 3780 }, { "epoch": 54.93, "learning_rate": 5.276271186440678e-05, "loss": 0.6925, "step": 3790 }, { "epoch": 55.07, "learning_rate": 5.259322033898305e-05, "loss": 0.6712, "step": 3800 }, { "epoch": 55.22, "learning_rate": 5.2423728813559324e-05, "loss": 0.6999, "step": 3810 }, { "epoch": 55.36, "learning_rate": 5.22542372881356e-05, "loss": 0.6577, "step": 3820 }, { "epoch": 55.51, "learning_rate": 5.2084745762711864e-05, "loss": 0.733, "step": 3830 }, { "epoch": 55.65, "learning_rate": 5.191525423728814e-05, "loss": 0.666, "step": 3840 }, { "epoch": 55.8, "learning_rate": 5.174576271186441e-05, "loss": 0.6638, "step": 3850 }, { "epoch": 55.94, "learning_rate": 5.157627118644068e-05, "loss": 0.7019, "step": 3860 }, { "epoch": 56.09, "learning_rate": 5.140677966101695e-05, "loss": 0.7018, "step": 3870 }, { "epoch": 56.23, "learning_rate": 5.123728813559322e-05, "loss": 0.6911, "step": 3880 }, { "epoch": 56.38, "learning_rate": 5.1067796610169496e-05, "loss": 0.7035, "step": 3890 }, { "epoch": 56.52, "learning_rate": 5.089830508474577e-05, "loss": 0.7204, "step": 3900 }, { "epoch": 56.67, "learning_rate": 5.072881355932204e-05, "loss": 0.6888, "step": 3910 }, { "epoch": 56.81, "learning_rate": 5.05593220338983e-05, "loss": 0.7502, "step": 3920 }, { "epoch": 56.96, "learning_rate": 5.0389830508474575e-05, "loss": 0.6246, "step": 3930 }, { "epoch": 57.1, "learning_rate": 5.0220338983050855e-05, "loss": 0.6079, "step": 3940 }, { "epoch": 57.25, "learning_rate": 5.005084745762713e-05, "loss": 0.6618, "step": 3950 }, { "epoch": 57.39, "learning_rate": 4.9881355932203394e-05, "loss": 0.6935, "step": 3960 }, { "epoch": 57.54, "learning_rate": 4.971186440677966e-05, "loss": 0.6872, "step": 3970 }, { "epoch": 57.68, "learning_rate": 4.9542372881355934e-05, "loss": 0.6546, "step": 3980 }, { "epoch": 57.83, "learning_rate": 4.937288135593221e-05, "loss": 0.7091, "step": 3990 }, { "epoch": 57.97, "learning_rate": 4.920338983050848e-05, "loss": 0.7014, "step": 4000 }, { "epoch": 57.97, "eval_loss": 0.21360060572624207, "eval_runtime": 570.979, "eval_samples_per_second": 5.944, "eval_steps_per_second": 0.744, "eval_wer": 0.16485467051253605, "step": 4000 }, { "epoch": 58.12, "learning_rate": 4.9033898305084746e-05, "loss": 0.6837, "step": 4010 }, { "epoch": 58.26, "learning_rate": 4.886440677966102e-05, "loss": 0.6707, "step": 4020 }, { "epoch": 58.41, "learning_rate": 4.8694915254237286e-05, "loss": 0.6715, "step": 4030 }, { "epoch": 58.55, "learning_rate": 4.8525423728813566e-05, "loss": 0.64, "step": 4040 }, { "epoch": 58.7, "learning_rate": 4.835593220338983e-05, "loss": 0.6904, "step": 4050 }, { "epoch": 58.84, "learning_rate": 4.8186440677966105e-05, "loss": 0.6809, "step": 4060 }, { "epoch": 58.99, "learning_rate": 4.801694915254237e-05, "loss": 0.6187, "step": 4070 }, { "epoch": 59.13, "learning_rate": 4.7847457627118645e-05, "loss": 0.7028, "step": 4080 }, { "epoch": 59.28, "learning_rate": 4.767796610169492e-05, "loss": 0.7025, "step": 4090 }, { "epoch": 59.42, "learning_rate": 4.750847457627119e-05, "loss": 0.6471, "step": 4100 }, { "epoch": 59.57, "learning_rate": 4.733898305084746e-05, "loss": 0.6968, "step": 4110 }, { "epoch": 59.71, "learning_rate": 4.716949152542373e-05, "loss": 0.6268, "step": 4120 }, { "epoch": 59.86, "learning_rate": 4.7e-05, "loss": 0.6731, "step": 4130 }, { "epoch": 60.0, "learning_rate": 4.683050847457627e-05, "loss": 0.6614, "step": 4140 }, { "epoch": 60.14, "learning_rate": 4.666101694915255e-05, "loss": 0.6138, "step": 4150 }, { "epoch": 60.29, "learning_rate": 4.649152542372882e-05, "loss": 0.6875, "step": 4160 }, { "epoch": 60.43, "learning_rate": 4.632203389830509e-05, "loss": 0.7305, "step": 4170 }, { "epoch": 60.58, "learning_rate": 4.6152542372881356e-05, "loss": 0.6367, "step": 4180 }, { "epoch": 60.72, "learning_rate": 4.598305084745763e-05, "loss": 0.7106, "step": 4190 }, { "epoch": 60.87, "learning_rate": 4.58135593220339e-05, "loss": 0.7531, "step": 4200 }, { "epoch": 61.01, "learning_rate": 4.5644067796610176e-05, "loss": 0.6545, "step": 4210 }, { "epoch": 61.16, "learning_rate": 4.547457627118644e-05, "loss": 0.7458, "step": 4220 }, { "epoch": 61.3, "learning_rate": 4.5305084745762715e-05, "loss": 0.6348, "step": 4230 }, { "epoch": 61.45, "learning_rate": 4.513559322033898e-05, "loss": 0.6631, "step": 4240 }, { "epoch": 61.59, "learning_rate": 4.4966101694915255e-05, "loss": 0.8233, "step": 4250 }, { "epoch": 61.74, "learning_rate": 4.479661016949153e-05, "loss": 0.6475, "step": 4260 }, { "epoch": 61.88, "learning_rate": 4.46271186440678e-05, "loss": 0.659, "step": 4270 }, { "epoch": 62.03, "learning_rate": 4.445762711864407e-05, "loss": 0.7019, "step": 4280 }, { "epoch": 62.17, "learning_rate": 4.428813559322034e-05, "loss": 0.6716, "step": 4290 }, { "epoch": 62.32, "learning_rate": 4.4118644067796614e-05, "loss": 0.6016, "step": 4300 }, { "epoch": 62.46, "learning_rate": 4.394915254237289e-05, "loss": 0.6532, "step": 4310 }, { "epoch": 62.61, "learning_rate": 4.377966101694915e-05, "loss": 0.6426, "step": 4320 }, { "epoch": 62.75, "learning_rate": 4.3610169491525426e-05, "loss": 0.6543, "step": 4330 }, { "epoch": 62.9, "learning_rate": 4.344067796610169e-05, "loss": 0.7405, "step": 4340 }, { "epoch": 63.04, "learning_rate": 4.3271186440677966e-05, "loss": 0.6795, "step": 4350 }, { "epoch": 63.19, "learning_rate": 4.310169491525424e-05, "loss": 0.726, "step": 4360 }, { "epoch": 63.33, "learning_rate": 4.293220338983051e-05, "loss": 0.8443, "step": 4370 }, { "epoch": 63.48, "learning_rate": 4.276271186440678e-05, "loss": 0.7359, "step": 4380 }, { "epoch": 63.62, "learning_rate": 4.259322033898305e-05, "loss": 0.6781, "step": 4390 }, { "epoch": 63.77, "learning_rate": 4.242372881355932e-05, "loss": 0.6366, "step": 4400 }, { "epoch": 63.91, "learning_rate": 4.22542372881356e-05, "loss": 0.6598, "step": 4410 }, { "epoch": 64.06, "learning_rate": 4.208474576271187e-05, "loss": 0.6668, "step": 4420 }, { "epoch": 64.2, "learning_rate": 4.191525423728814e-05, "loss": 0.7418, "step": 4430 }, { "epoch": 64.35, "learning_rate": 4.174576271186441e-05, "loss": 0.6788, "step": 4440 }, { "epoch": 64.49, "learning_rate": 4.157627118644068e-05, "loss": 0.6549, "step": 4450 }, { "epoch": 64.64, "learning_rate": 4.140677966101695e-05, "loss": 0.6636, "step": 4460 }, { "epoch": 64.78, "learning_rate": 4.1237288135593223e-05, "loss": 0.6163, "step": 4470 }, { "epoch": 64.93, "learning_rate": 4.10677966101695e-05, "loss": 0.7296, "step": 4480 }, { "epoch": 65.07, "learning_rate": 4.089830508474576e-05, "loss": 0.6633, "step": 4490 }, { "epoch": 65.22, "learning_rate": 4.0728813559322036e-05, "loss": 0.6171, "step": 4500 }, { "epoch": 65.22, "eval_loss": 0.21388114988803864, "eval_runtime": 574.1167, "eval_samples_per_second": 5.912, "eval_steps_per_second": 0.74, "eval_wer": 0.16409864490627749, "step": 4500 }, { "epoch": 65.36, "learning_rate": 4.055932203389831e-05, "loss": 0.647, "step": 4510 }, { "epoch": 65.51, "learning_rate": 4.038983050847458e-05, "loss": 0.5946, "step": 4520 }, { "epoch": 65.65, "learning_rate": 4.022033898305085e-05, "loss": 0.6789, "step": 4530 }, { "epoch": 65.8, "learning_rate": 4.005084745762712e-05, "loss": 0.6854, "step": 4540 }, { "epoch": 65.94, "learning_rate": 3.988135593220339e-05, "loss": 0.6547, "step": 4550 }, { "epoch": 66.09, "learning_rate": 3.971186440677966e-05, "loss": 0.7462, "step": 4560 }, { "epoch": 66.23, "learning_rate": 3.9542372881355935e-05, "loss": 0.6724, "step": 4570 }, { "epoch": 66.38, "learning_rate": 3.937288135593221e-05, "loss": 0.6623, "step": 4580 }, { "epoch": 66.52, "learning_rate": 3.9203389830508474e-05, "loss": 0.6438, "step": 4590 }, { "epoch": 66.67, "learning_rate": 3.903389830508475e-05, "loss": 0.6983, "step": 4600 }, { "epoch": 66.81, "learning_rate": 3.8864406779661014e-05, "loss": 0.6905, "step": 4610 }, { "epoch": 66.96, "learning_rate": 3.8694915254237294e-05, "loss": 0.7062, "step": 4620 }, { "epoch": 67.1, "learning_rate": 3.852542372881356e-05, "loss": 0.6685, "step": 4630 }, { "epoch": 67.25, "learning_rate": 3.835593220338983e-05, "loss": 0.6545, "step": 4640 }, { "epoch": 67.39, "learning_rate": 3.81864406779661e-05, "loss": 0.6953, "step": 4650 }, { "epoch": 67.54, "learning_rate": 3.801694915254237e-05, "loss": 0.6793, "step": 4660 }, { "epoch": 67.68, "learning_rate": 3.7847457627118646e-05, "loss": 0.6059, "step": 4670 }, { "epoch": 67.83, "learning_rate": 3.767796610169492e-05, "loss": 0.6555, "step": 4680 }, { "epoch": 67.97, "learning_rate": 3.750847457627119e-05, "loss": 0.6261, "step": 4690 }, { "epoch": 68.12, "learning_rate": 3.733898305084746e-05, "loss": 0.7417, "step": 4700 }, { "epoch": 68.26, "learning_rate": 3.716949152542373e-05, "loss": 0.6732, "step": 4710 }, { "epoch": 68.41, "learning_rate": 3.7e-05, "loss": 0.7073, "step": 4720 }, { "epoch": 68.55, "learning_rate": 3.683050847457628e-05, "loss": 0.6299, "step": 4730 }, { "epoch": 68.7, "learning_rate": 3.6661016949152544e-05, "loss": 0.6789, "step": 4740 }, { "epoch": 68.84, "learning_rate": 3.649152542372882e-05, "loss": 0.6946, "step": 4750 }, { "epoch": 68.99, "learning_rate": 3.6322033898305084e-05, "loss": 0.6954, "step": 4760 }, { "epoch": 69.13, "learning_rate": 3.615254237288136e-05, "loss": 0.6609, "step": 4770 }, { "epoch": 69.28, "learning_rate": 3.598305084745763e-05, "loss": 0.7685, "step": 4780 }, { "epoch": 69.42, "learning_rate": 3.5813559322033903e-05, "loss": 0.6556, "step": 4790 }, { "epoch": 69.57, "learning_rate": 3.564406779661017e-05, "loss": 0.6377, "step": 4800 }, { "epoch": 69.71, "learning_rate": 3.547457627118644e-05, "loss": 0.667, "step": 4810 }, { "epoch": 69.86, "learning_rate": 3.530508474576271e-05, "loss": 0.6439, "step": 4820 }, { "epoch": 70.0, "learning_rate": 3.513559322033899e-05, "loss": 0.6266, "step": 4830 }, { "epoch": 70.14, "learning_rate": 3.4966101694915256e-05, "loss": 0.6563, "step": 4840 }, { "epoch": 70.29, "learning_rate": 3.479661016949153e-05, "loss": 0.606, "step": 4850 }, { "epoch": 70.43, "learning_rate": 3.4627118644067795e-05, "loss": 0.6662, "step": 4860 }, { "epoch": 70.58, "learning_rate": 3.445762711864407e-05, "loss": 0.6453, "step": 4870 }, { "epoch": 70.72, "learning_rate": 3.428813559322034e-05, "loss": 0.6812, "step": 4880 }, { "epoch": 70.87, "learning_rate": 3.4118644067796615e-05, "loss": 0.6655, "step": 4890 }, { "epoch": 71.01, "learning_rate": 3.394915254237288e-05, "loss": 0.6687, "step": 4900 }, { "epoch": 71.16, "learning_rate": 3.3779661016949154e-05, "loss": 0.6349, "step": 4910 }, { "epoch": 71.3, "learning_rate": 3.361016949152542e-05, "loss": 0.5908, "step": 4920 }, { "epoch": 71.45, "learning_rate": 3.3440677966101694e-05, "loss": 0.6648, "step": 4930 }, { "epoch": 71.59, "learning_rate": 3.327118644067797e-05, "loss": 0.6964, "step": 4940 }, { "epoch": 71.74, "learning_rate": 3.310169491525424e-05, "loss": 0.6482, "step": 4950 }, { "epoch": 71.88, "learning_rate": 3.293220338983051e-05, "loss": 0.7086, "step": 4960 }, { "epoch": 72.03, "learning_rate": 3.276271186440678e-05, "loss": 0.6146, "step": 4970 }, { "epoch": 72.17, "learning_rate": 3.259322033898305e-05, "loss": 0.6546, "step": 4980 }, { "epoch": 72.32, "learning_rate": 3.2423728813559326e-05, "loss": 0.608, "step": 4990 }, { "epoch": 72.46, "learning_rate": 3.22542372881356e-05, "loss": 0.6609, "step": 5000 }, { "epoch": 72.46, "eval_loss": 0.2143816202878952, "eval_runtime": 575.4482, "eval_samples_per_second": 5.898, "eval_steps_per_second": 0.739, "eval_wer": 0.16206888050686585, "step": 5000 }, { "epoch": 72.61, "learning_rate": 3.2084745762711865e-05, "loss": 0.6483, "step": 5010 }, { "epoch": 72.75, "learning_rate": 3.191525423728814e-05, "loss": 0.5998, "step": 5020 }, { "epoch": 72.9, "learning_rate": 3.1745762711864405e-05, "loss": 0.6161, "step": 5030 }, { "epoch": 73.04, "learning_rate": 3.157627118644068e-05, "loss": 0.6856, "step": 5040 }, { "epoch": 73.19, "learning_rate": 3.140677966101695e-05, "loss": 0.6095, "step": 5050 }, { "epoch": 73.33, "learning_rate": 3.1237288135593224e-05, "loss": 0.6235, "step": 5060 }, { "epoch": 73.48, "learning_rate": 3.106779661016949e-05, "loss": 0.6827, "step": 5070 }, { "epoch": 73.62, "learning_rate": 3.0898305084745764e-05, "loss": 0.5903, "step": 5080 }, { "epoch": 73.77, "learning_rate": 3.072881355932204e-05, "loss": 0.7406, "step": 5090 }, { "epoch": 73.91, "learning_rate": 3.055932203389831e-05, "loss": 0.6621, "step": 5100 }, { "epoch": 74.06, "learning_rate": 3.0389830508474577e-05, "loss": 0.6619, "step": 5110 }, { "epoch": 74.2, "learning_rate": 3.022033898305085e-05, "loss": 0.6755, "step": 5120 }, { "epoch": 74.35, "learning_rate": 3.005084745762712e-05, "loss": 0.6568, "step": 5130 }, { "epoch": 74.49, "learning_rate": 2.9881355932203393e-05, "loss": 0.6244, "step": 5140 }, { "epoch": 74.64, "learning_rate": 2.971186440677966e-05, "loss": 0.6166, "step": 5150 }, { "epoch": 74.78, "learning_rate": 2.9542372881355936e-05, "loss": 0.6238, "step": 5160 }, { "epoch": 74.93, "learning_rate": 2.9372881355932202e-05, "loss": 0.7148, "step": 5170 }, { "epoch": 75.07, "learning_rate": 2.9203389830508475e-05, "loss": 0.6449, "step": 5180 }, { "epoch": 75.22, "learning_rate": 2.9033898305084745e-05, "loss": 0.6567, "step": 5190 }, { "epoch": 75.36, "learning_rate": 2.8864406779661018e-05, "loss": 0.6167, "step": 5200 }, { "epoch": 75.51, "learning_rate": 2.8694915254237288e-05, "loss": 0.6728, "step": 5210 }, { "epoch": 75.65, "learning_rate": 2.852542372881356e-05, "loss": 0.6267, "step": 5220 }, { "epoch": 75.8, "learning_rate": 2.8355932203389834e-05, "loss": 0.6707, "step": 5230 }, { "epoch": 75.94, "learning_rate": 2.8186440677966104e-05, "loss": 0.6005, "step": 5240 }, { "epoch": 76.09, "learning_rate": 2.8016949152542377e-05, "loss": 0.5838, "step": 5250 }, { "epoch": 76.23, "learning_rate": 2.7847457627118643e-05, "loss": 0.6954, "step": 5260 }, { "epoch": 76.38, "learning_rate": 2.767796610169492e-05, "loss": 0.6658, "step": 5270 }, { "epoch": 76.52, "learning_rate": 2.7508474576271186e-05, "loss": 0.6252, "step": 5280 }, { "epoch": 76.67, "learning_rate": 2.733898305084746e-05, "loss": 0.6967, "step": 5290 }, { "epoch": 76.81, "learning_rate": 2.716949152542373e-05, "loss": 0.6216, "step": 5300 }, { "epoch": 76.96, "learning_rate": 2.7000000000000002e-05, "loss": 0.6392, "step": 5310 }, { "epoch": 77.1, "learning_rate": 2.6830508474576272e-05, "loss": 0.6266, "step": 5320 }, { "epoch": 77.25, "learning_rate": 2.6661016949152545e-05, "loss": 0.5908, "step": 5330 }, { "epoch": 77.39, "learning_rate": 2.6491525423728815e-05, "loss": 0.6788, "step": 5340 }, { "epoch": 77.54, "learning_rate": 2.6322033898305088e-05, "loss": 0.6592, "step": 5350 }, { "epoch": 77.68, "learning_rate": 2.6152542372881355e-05, "loss": 0.6038, "step": 5360 }, { "epoch": 77.83, "learning_rate": 2.598305084745763e-05, "loss": 0.5884, "step": 5370 }, { "epoch": 77.97, "learning_rate": 2.5813559322033898e-05, "loss": 0.6433, "step": 5380 }, { "epoch": 78.12, "learning_rate": 2.564406779661017e-05, "loss": 0.6976, "step": 5390 }, { "epoch": 78.26, "learning_rate": 2.547457627118644e-05, "loss": 0.6922, "step": 5400 }, { "epoch": 78.41, "learning_rate": 2.5305084745762714e-05, "loss": 0.6621, "step": 5410 }, { "epoch": 78.55, "learning_rate": 2.5135593220338983e-05, "loss": 0.6388, "step": 5420 }, { "epoch": 78.7, "learning_rate": 2.4966101694915257e-05, "loss": 0.628, "step": 5430 }, { "epoch": 78.84, "learning_rate": 2.4796610169491526e-05, "loss": 0.5507, "step": 5440 }, { "epoch": 78.99, "learning_rate": 2.46271186440678e-05, "loss": 0.6358, "step": 5450 }, { "epoch": 79.13, "learning_rate": 2.445762711864407e-05, "loss": 0.8279, "step": 5460 }, { "epoch": 79.28, "learning_rate": 2.428813559322034e-05, "loss": 0.6108, "step": 5470 }, { "epoch": 79.42, "learning_rate": 2.4118644067796612e-05, "loss": 0.6535, "step": 5480 }, { "epoch": 79.57, "learning_rate": 2.3949152542372882e-05, "loss": 0.5962, "step": 5490 }, { "epoch": 79.71, "learning_rate": 2.3779661016949155e-05, "loss": 0.6318, "step": 5500 }, { "epoch": 79.71, "eval_loss": 0.21294616162776947, "eval_runtime": 575.7769, "eval_samples_per_second": 5.895, "eval_steps_per_second": 0.738, "eval_wer": 0.16001446309855452, "step": 5500 }, { "epoch": 79.86, "learning_rate": 2.3610169491525425e-05, "loss": 0.5991, "step": 5510 }, { "epoch": 80.0, "learning_rate": 2.3440677966101695e-05, "loss": 0.6389, "step": 5520 }, { "epoch": 80.14, "learning_rate": 2.3271186440677968e-05, "loss": 0.5969, "step": 5530 }, { "epoch": 80.29, "learning_rate": 2.3101694915254237e-05, "loss": 0.626, "step": 5540 }, { "epoch": 80.43, "learning_rate": 2.2932203389830507e-05, "loss": 0.6123, "step": 5550 }, { "epoch": 80.58, "learning_rate": 2.276271186440678e-05, "loss": 0.6056, "step": 5560 }, { "epoch": 80.72, "learning_rate": 2.259322033898305e-05, "loss": 0.6004, "step": 5570 }, { "epoch": 80.87, "learning_rate": 2.2423728813559323e-05, "loss": 0.5924, "step": 5580 }, { "epoch": 81.01, "learning_rate": 2.2254237288135593e-05, "loss": 0.6622, "step": 5590 }, { "epoch": 81.16, "learning_rate": 2.2084745762711863e-05, "loss": 0.6285, "step": 5600 }, { "epoch": 81.3, "learning_rate": 2.1915254237288136e-05, "loss": 0.6199, "step": 5610 }, { "epoch": 81.45, "learning_rate": 2.174576271186441e-05, "loss": 0.6496, "step": 5620 }, { "epoch": 81.59, "learning_rate": 2.157627118644068e-05, "loss": 0.632, "step": 5630 }, { "epoch": 81.74, "learning_rate": 2.1406779661016952e-05, "loss": 0.5825, "step": 5640 }, { "epoch": 81.88, "learning_rate": 2.1237288135593222e-05, "loss": 0.7035, "step": 5650 }, { "epoch": 82.03, "learning_rate": 2.1067796610169495e-05, "loss": 0.5958, "step": 5660 }, { "epoch": 82.17, "learning_rate": 2.0898305084745765e-05, "loss": 0.6213, "step": 5670 }, { "epoch": 82.32, "learning_rate": 2.0728813559322035e-05, "loss": 0.621, "step": 5680 }, { "epoch": 82.46, "learning_rate": 2.0559322033898308e-05, "loss": 0.5801, "step": 5690 }, { "epoch": 82.61, "learning_rate": 2.0389830508474577e-05, "loss": 0.6511, "step": 5700 }, { "epoch": 82.75, "learning_rate": 2.0220338983050847e-05, "loss": 0.6268, "step": 5710 }, { "epoch": 82.9, "learning_rate": 2.005084745762712e-05, "loss": 0.642, "step": 5720 }, { "epoch": 83.04, "learning_rate": 1.988135593220339e-05, "loss": 0.6158, "step": 5730 }, { "epoch": 83.19, "learning_rate": 1.9711864406779663e-05, "loss": 0.6171, "step": 5740 }, { "epoch": 83.33, "learning_rate": 1.9542372881355933e-05, "loss": 0.5943, "step": 5750 }, { "epoch": 83.48, "learning_rate": 1.9372881355932203e-05, "loss": 0.6312, "step": 5760 }, { "epoch": 83.62, "learning_rate": 1.9203389830508476e-05, "loss": 0.6128, "step": 5770 }, { "epoch": 83.77, "learning_rate": 1.9033898305084746e-05, "loss": 0.6094, "step": 5780 }, { "epoch": 83.91, "learning_rate": 1.886440677966102e-05, "loss": 0.6403, "step": 5790 }, { "epoch": 84.06, "learning_rate": 1.869491525423729e-05, "loss": 0.5933, "step": 5800 }, { "epoch": 84.2, "learning_rate": 1.852542372881356e-05, "loss": 0.5684, "step": 5810 }, { "epoch": 84.35, "learning_rate": 1.835593220338983e-05, "loss": 0.6025, "step": 5820 }, { "epoch": 84.49, "learning_rate": 1.81864406779661e-05, "loss": 0.6225, "step": 5830 }, { "epoch": 84.64, "learning_rate": 1.8016949152542374e-05, "loss": 0.6246, "step": 5840 }, { "epoch": 84.78, "learning_rate": 1.7847457627118644e-05, "loss": 0.6186, "step": 5850 }, { "epoch": 84.93, "learning_rate": 1.7677966101694914e-05, "loss": 0.6706, "step": 5860 }, { "epoch": 85.07, "learning_rate": 1.7508474576271187e-05, "loss": 0.5923, "step": 5870 }, { "epoch": 85.22, "learning_rate": 1.7338983050847457e-05, "loss": 0.5892, "step": 5880 }, { "epoch": 85.36, "learning_rate": 1.716949152542373e-05, "loss": 0.6155, "step": 5890 }, { "epoch": 85.51, "learning_rate": 1.7000000000000003e-05, "loss": 0.545, "step": 5900 }, { "epoch": 85.65, "learning_rate": 1.6830508474576273e-05, "loss": 0.6541, "step": 5910 }, { "epoch": 85.8, "learning_rate": 1.6661016949152543e-05, "loss": 0.7079, "step": 5920 }, { "epoch": 85.94, "learning_rate": 1.6491525423728816e-05, "loss": 0.58, "step": 5930 }, { "epoch": 86.09, "learning_rate": 1.6322033898305086e-05, "loss": 0.631, "step": 5940 }, { "epoch": 86.23, "learning_rate": 1.615254237288136e-05, "loss": 0.6232, "step": 5950 }, { "epoch": 86.38, "learning_rate": 1.598305084745763e-05, "loss": 0.5722, "step": 5960 }, { "epoch": 86.52, "learning_rate": 1.58135593220339e-05, "loss": 0.6361, "step": 5970 }, { "epoch": 86.67, "learning_rate": 1.564406779661017e-05, "loss": 0.6603, "step": 5980 }, { "epoch": 86.81, "learning_rate": 1.547457627118644e-05, "loss": 0.6191, "step": 5990 }, { "epoch": 86.96, "learning_rate": 1.5305084745762714e-05, "loss": 0.6222, "step": 6000 }, { "epoch": 86.96, "eval_loss": 0.2124408483505249, "eval_runtime": 572.9918, "eval_samples_per_second": 5.923, "eval_steps_per_second": 0.742, "eval_wer": 0.158223011118507, "step": 6000 }, { "epoch": 87.1, "learning_rate": 6.088281250000001e-05, "loss": 0.595, "step": 6010 }, { "epoch": 87.25, "learning_rate": 6.08046875e-05, "loss": 0.6517, "step": 6020 }, { "epoch": 87.39, "learning_rate": 6.07265625e-05, "loss": 0.6079, "step": 6030 }, { "epoch": 87.54, "learning_rate": 6.06484375e-05, "loss": 0.6203, "step": 6040 }, { "epoch": 87.68, "learning_rate": 6.05703125e-05, "loss": 0.6005, "step": 6050 }, { "epoch": 87.83, "learning_rate": 6.0492187500000006e-05, "loss": 0.6321, "step": 6060 }, { "epoch": 87.97, "learning_rate": 6.04140625e-05, "loss": 0.6156, "step": 6070 }, { "epoch": 88.12, "learning_rate": 6.0335937500000005e-05, "loss": 0.6329, "step": 6080 }, { "epoch": 88.26, "learning_rate": 6.02578125e-05, "loss": 0.6311, "step": 6090 }, { "epoch": 88.41, "learning_rate": 6.0179687500000005e-05, "loss": 0.6689, "step": 6100 }, { "epoch": 88.55, "learning_rate": 6.010156250000001e-05, "loss": 0.5824, "step": 6110 }, { "epoch": 88.7, "learning_rate": 6.0023437500000005e-05, "loss": 0.6491, "step": 6120 }, { "epoch": 88.84, "learning_rate": 5.994531250000001e-05, "loss": 0.6326, "step": 6130 }, { "epoch": 88.99, "learning_rate": 5.98671875e-05, "loss": 0.6272, "step": 6140 }, { "epoch": 89.13, "learning_rate": 5.97890625e-05, "loss": 0.6865, "step": 6150 }, { "epoch": 89.28, "learning_rate": 5.971093750000001e-05, "loss": 0.6149, "step": 6160 }, { "epoch": 89.42, "learning_rate": 5.96328125e-05, "loss": 0.6385, "step": 6170 }, { "epoch": 89.57, "learning_rate": 5.9554687500000003e-05, "loss": 0.6621, "step": 6180 }, { "epoch": 89.71, "learning_rate": 5.94765625e-05, "loss": 0.5817, "step": 6190 }, { "epoch": 89.86, "learning_rate": 5.93984375e-05, "loss": 0.6333, "step": 6200 }, { "epoch": 90.0, "learning_rate": 5.9320312500000006e-05, "loss": 0.5794, "step": 6210 }, { "epoch": 90.14, "learning_rate": 5.92421875e-05, "loss": 0.6153, "step": 6220 }, { "epoch": 90.29, "learning_rate": 5.9164062500000006e-05, "loss": 0.6794, "step": 6230 }, { "epoch": 90.43, "learning_rate": 5.90859375e-05, "loss": 0.601, "step": 6240 }, { "epoch": 90.58, "learning_rate": 5.9007812500000005e-05, "loss": 0.6176, "step": 6250 }, { "epoch": 90.72, "learning_rate": 5.892968750000001e-05, "loss": 0.6055, "step": 6260 }, { "epoch": 90.87, "learning_rate": 5.88515625e-05, "loss": 0.6389, "step": 6270 }, { "epoch": 91.01, "learning_rate": 5.877343750000001e-05, "loss": 0.6727, "step": 6280 }, { "epoch": 91.16, "learning_rate": 5.86953125e-05, "loss": 0.6592, "step": 6290 }, { "epoch": 91.3, "learning_rate": 5.86171875e-05, "loss": 0.6367, "step": 6300 }, { "epoch": 91.45, "learning_rate": 5.853906250000001e-05, "loss": 0.6366, "step": 6310 }, { "epoch": 91.59, "learning_rate": 5.84609375e-05, "loss": 0.6171, "step": 6320 }, { "epoch": 91.74, "learning_rate": 5.8382812500000004e-05, "loss": 0.6642, "step": 6330 }, { "epoch": 91.88, "learning_rate": 5.83046875e-05, "loss": 0.6345, "step": 6340 }, { "epoch": 92.03, "learning_rate": 5.82265625e-05, "loss": 0.6265, "step": 6350 }, { "epoch": 92.17, "learning_rate": 5.8148437500000006e-05, "loss": 0.6302, "step": 6360 }, { "epoch": 92.32, "learning_rate": 5.80703125e-05, "loss": 0.6347, "step": 6370 }, { "epoch": 92.46, "learning_rate": 5.7992187500000006e-05, "loss": 0.6858, "step": 6380 }, { "epoch": 92.61, "learning_rate": 5.79140625e-05, "loss": 0.5876, "step": 6390 }, { "epoch": 92.75, "learning_rate": 5.7835937500000006e-05, "loss": 0.6195, "step": 6400 }, { "epoch": 92.9, "learning_rate": 5.775781250000001e-05, "loss": 0.6559, "step": 6410 }, { "epoch": 93.04, "learning_rate": 5.76796875e-05, "loss": 0.6908, "step": 6420 }, { "epoch": 93.19, "learning_rate": 5.760156250000001e-05, "loss": 0.5887, "step": 6430 }, { "epoch": 93.33, "learning_rate": 5.75234375e-05, "loss": 0.6118, "step": 6440 }, { "epoch": 93.48, "learning_rate": 5.74453125e-05, "loss": 0.7352, "step": 6450 }, { "epoch": 93.62, "learning_rate": 5.736718750000001e-05, "loss": 0.6462, "step": 6460 }, { "epoch": 93.77, "learning_rate": 5.72890625e-05, "loss": 0.6448, "step": 6470 }, { "epoch": 93.91, "learning_rate": 5.7210937500000004e-05, "loss": 0.6806, "step": 6480 }, { "epoch": 94.06, "learning_rate": 5.71328125e-05, "loss": 0.6465, "step": 6490 }, { "epoch": 94.2, "learning_rate": 5.7062500000000005e-05, "loss": 0.608, "step": 6500 }, { "epoch": 94.2, "eval_loss": 0.22552849352359772, "eval_runtime": 596.3798, "eval_samples_per_second": 5.691, "eval_steps_per_second": 0.713, "eval_wer": 0.1638849854958131, "step": 6500 }, { "epoch": 94.35, "learning_rate": 5.6984375e-05, "loss": 0.6931, "step": 6510 }, { "epoch": 94.49, "learning_rate": 5.6906250000000004e-05, "loss": 0.6172, "step": 6520 }, { "epoch": 94.64, "learning_rate": 5.6828125e-05, "loss": 0.686, "step": 6530 }, { "epoch": 94.78, "learning_rate": 5.6750000000000004e-05, "loss": 0.6993, "step": 6540 }, { "epoch": 94.93, "learning_rate": 5.667187500000001e-05, "loss": 0.6582, "step": 6550 }, { "epoch": 95.07, "learning_rate": 5.6593750000000003e-05, "loss": 0.6366, "step": 6560 }, { "epoch": 95.22, "learning_rate": 5.6515625000000007e-05, "loss": 0.6609, "step": 6570 }, { "epoch": 95.36, "learning_rate": 5.6437499999999996e-05, "loss": 0.5802, "step": 6580 }, { "epoch": 95.51, "learning_rate": 5.6359375000000006e-05, "loss": 0.5748, "step": 6590 }, { "epoch": 95.65, "learning_rate": 5.628125000000001e-05, "loss": 0.6195, "step": 6600 }, { "epoch": 95.8, "learning_rate": 5.6203125e-05, "loss": 0.6524, "step": 6610 }, { "epoch": 95.94, "learning_rate": 5.6125e-05, "loss": 0.6621, "step": 6620 }, { "epoch": 96.09, "learning_rate": 5.6046875e-05, "loss": 0.6074, "step": 6630 }, { "epoch": 96.23, "learning_rate": 5.596875e-05, "loss": 0.598, "step": 6640 }, { "epoch": 96.38, "learning_rate": 5.5890625000000005e-05, "loss": 0.6296, "step": 6650 }, { "epoch": 96.52, "learning_rate": 5.58125e-05, "loss": 0.6353, "step": 6660 }, { "epoch": 96.67, "learning_rate": 5.5734375000000005e-05, "loss": 0.6013, "step": 6670 }, { "epoch": 96.81, "learning_rate": 5.565625e-05, "loss": 0.606, "step": 6680 }, { "epoch": 96.96, "learning_rate": 5.5578125000000004e-05, "loss": 0.742, "step": 6690 }, { "epoch": 97.1, "learning_rate": 5.550000000000001e-05, "loss": 0.5763, "step": 6700 }, { "epoch": 97.25, "learning_rate": 5.5421875000000004e-05, "loss": 0.6839, "step": 6710 }, { "epoch": 97.39, "learning_rate": 5.534375000000001e-05, "loss": 0.6565, "step": 6720 }, { "epoch": 97.54, "learning_rate": 5.5265624999999997e-05, "loss": 0.6341, "step": 6730 }, { "epoch": 97.68, "learning_rate": 5.51875e-05, "loss": 0.6721, "step": 6740 }, { "epoch": 97.83, "learning_rate": 5.510937500000001e-05, "loss": 0.6134, "step": 6750 }, { "epoch": 97.97, "learning_rate": 5.503125e-05, "loss": 0.6161, "step": 6760 }, { "epoch": 98.12, "learning_rate": 5.4953125e-05, "loss": 0.6273, "step": 6770 }, { "epoch": 98.26, "learning_rate": 5.4875e-05, "loss": 0.5778, "step": 6780 }, { "epoch": 98.41, "learning_rate": 5.4796875e-05, "loss": 0.6879, "step": 6790 }, { "epoch": 98.55, "learning_rate": 5.4718750000000005e-05, "loss": 0.6652, "step": 6800 }, { "epoch": 98.7, "learning_rate": 5.4640625e-05, "loss": 0.6894, "step": 6810 }, { "epoch": 98.84, "learning_rate": 5.4562500000000005e-05, "loss": 0.5951, "step": 6820 }, { "epoch": 98.99, "learning_rate": 5.4484375e-05, "loss": 0.6397, "step": 6830 }, { "epoch": 99.13, "learning_rate": 5.4406250000000004e-05, "loss": 0.6221, "step": 6840 }, { "epoch": 99.28, "learning_rate": 5.432812500000001e-05, "loss": 0.7, "step": 6850 }, { "epoch": 99.42, "learning_rate": 5.4250000000000004e-05, "loss": 0.6638, "step": 6860 }, { "epoch": 99.57, "learning_rate": 5.417187500000001e-05, "loss": 0.6702, "step": 6870 }, { "epoch": 99.71, "learning_rate": 5.409375e-05, "loss": 0.637, "step": 6880 }, { "epoch": 99.86, "learning_rate": 5.4015625e-05, "loss": 0.6518, "step": 6890 }, { "epoch": 100.0, "learning_rate": 5.393750000000001e-05, "loss": 0.5912, "step": 6900 }, { "epoch": 100.14, "learning_rate": 5.3859375e-05, "loss": 0.6999, "step": 6910 }, { "epoch": 100.29, "learning_rate": 5.378125e-05, "loss": 0.5682, "step": 6920 }, { "epoch": 100.43, "learning_rate": 5.3703125e-05, "loss": 0.6023, "step": 6930 }, { "epoch": 100.58, "learning_rate": 5.3625e-05, "loss": 0.6437, "step": 6940 }, { "epoch": 100.72, "learning_rate": 5.3546875000000006e-05, "loss": 0.7168, "step": 6950 }, { "epoch": 100.87, "learning_rate": 5.346875e-05, "loss": 0.5913, "step": 6960 }, { "epoch": 101.01, "learning_rate": 5.3390625000000005e-05, "loss": 0.6047, "step": 6970 }, { "epoch": 101.16, "learning_rate": 5.33125e-05, "loss": 0.6355, "step": 6980 }, { "epoch": 101.3, "learning_rate": 5.3234375000000005e-05, "loss": 0.7046, "step": 6990 }, { "epoch": 101.45, "learning_rate": 5.315625000000001e-05, "loss": 0.6099, "step": 7000 }, { "epoch": 101.45, "eval_loss": 0.22652995586395264, "eval_runtime": 583.6707, "eval_samples_per_second": 5.815, "eval_steps_per_second": 0.728, "eval_wer": 0.1621839278817313, "step": 7000 }, { "epoch": 101.59, "learning_rate": 5.3078125e-05, "loss": 0.5706, "step": 7010 }, { "epoch": 101.74, "learning_rate": 5.300000000000001e-05, "loss": 0.5656, "step": 7020 }, { "epoch": 101.88, "learning_rate": 5.2921875e-05, "loss": 0.6755, "step": 7030 }, { "epoch": 102.03, "learning_rate": 5.284375e-05, "loss": 0.6343, "step": 7040 }, { "epoch": 102.17, "learning_rate": 5.276562500000001e-05, "loss": 0.6129, "step": 7050 }, { "epoch": 102.32, "learning_rate": 5.26875e-05, "loss": 0.6321, "step": 7060 }, { "epoch": 102.46, "learning_rate": 5.2609375e-05, "loss": 0.6189, "step": 7070 }, { "epoch": 102.61, "learning_rate": 5.253125e-05, "loss": 0.6267, "step": 7080 }, { "epoch": 102.75, "learning_rate": 5.2453125e-05, "loss": 0.601, "step": 7090 }, { "epoch": 102.9, "learning_rate": 5.2375000000000006e-05, "loss": 0.6051, "step": 7100 }, { "epoch": 103.04, "learning_rate": 5.2296875e-05, "loss": 0.538, "step": 7110 }, { "epoch": 103.19, "learning_rate": 5.2218750000000006e-05, "loss": 0.6136, "step": 7120 }, { "epoch": 103.33, "learning_rate": 5.2140624999999995e-05, "loss": 0.625, "step": 7130 }, { "epoch": 103.48, "learning_rate": 5.2062500000000005e-05, "loss": 0.5999, "step": 7140 }, { "epoch": 103.62, "learning_rate": 5.198437500000001e-05, "loss": 0.694, "step": 7150 }, { "epoch": 103.77, "learning_rate": 5.190625e-05, "loss": 0.6446, "step": 7160 }, { "epoch": 103.91, "learning_rate": 5.182812500000001e-05, "loss": 0.5879, "step": 7170 }, { "epoch": 104.06, "learning_rate": 5.175e-05, "loss": 0.6115, "step": 7180 }, { "epoch": 104.2, "learning_rate": 5.1671875e-05, "loss": 0.5663, "step": 7190 }, { "epoch": 104.35, "learning_rate": 5.159375000000001e-05, "loss": 0.5825, "step": 7200 }, { "epoch": 104.49, "learning_rate": 5.1515625e-05, "loss": 0.5692, "step": 7210 }, { "epoch": 104.64, "learning_rate": 5.1437500000000003e-05, "loss": 0.614, "step": 7220 }, { "epoch": 104.78, "learning_rate": 5.1359375e-05, "loss": 0.6005, "step": 7230 }, { "epoch": 104.93, "learning_rate": 5.128125e-05, "loss": 0.6169, "step": 7240 }, { "epoch": 105.07, "learning_rate": 5.1203125000000006e-05, "loss": 0.6383, "step": 7250 }, { "epoch": 105.22, "learning_rate": 5.1125e-05, "loss": 0.5705, "step": 7260 }, { "epoch": 105.36, "learning_rate": 5.1046875000000006e-05, "loss": 0.6212, "step": 7270 }, { "epoch": 105.51, "learning_rate": 5.0968749999999995e-05, "loss": 0.5747, "step": 7280 }, { "epoch": 105.65, "learning_rate": 5.0890625000000005e-05, "loss": 0.6407, "step": 7290 }, { "epoch": 105.8, "learning_rate": 5.081250000000001e-05, "loss": 0.5993, "step": 7300 }, { "epoch": 105.94, "learning_rate": 5.0734375e-05, "loss": 0.6226, "step": 7310 }, { "epoch": 106.09, "learning_rate": 5.065625000000001e-05, "loss": 0.6316, "step": 7320 }, { "epoch": 106.23, "learning_rate": 5.0578125e-05, "loss": 0.6922, "step": 7330 }, { "epoch": 106.38, "learning_rate": 5.05e-05, "loss": 0.6252, "step": 7340 }, { "epoch": 106.52, "learning_rate": 5.0421875000000004e-05, "loss": 0.5841, "step": 7350 }, { "epoch": 106.67, "learning_rate": 5.034375e-05, "loss": 0.6174, "step": 7360 }, { "epoch": 106.81, "learning_rate": 5.0265625000000004e-05, "loss": 0.6482, "step": 7370 }, { "epoch": 106.96, "learning_rate": 5.01875e-05, "loss": 0.5965, "step": 7380 }, { "epoch": 107.1, "learning_rate": 5.0109375e-05, "loss": 0.6417, "step": 7390 }, { "epoch": 107.25, "learning_rate": 5.0031250000000007e-05, "loss": 0.6373, "step": 7400 }, { "epoch": 107.39, "learning_rate": 4.9953125e-05, "loss": 0.6205, "step": 7410 }, { "epoch": 107.54, "learning_rate": 4.9875000000000006e-05, "loss": 0.5891, "step": 7420 }, { "epoch": 107.68, "learning_rate": 4.9796875e-05, "loss": 0.6333, "step": 7430 }, { "epoch": 107.83, "learning_rate": 4.9718750000000006e-05, "loss": 0.6479, "step": 7440 }, { "epoch": 107.97, "learning_rate": 4.9640625e-05, "loss": 0.5854, "step": 7450 }, { "epoch": 108.12, "learning_rate": 4.95625e-05, "loss": 0.602, "step": 7460 }, { "epoch": 108.26, "learning_rate": 4.9484375e-05, "loss": 0.6362, "step": 7470 }, { "epoch": 108.41, "learning_rate": 4.9406250000000005e-05, "loss": 0.7472, "step": 7480 }, { "epoch": 108.55, "learning_rate": 4.9328125e-05, "loss": 0.6158, "step": 7490 }, { "epoch": 108.7, "learning_rate": 4.9250000000000004e-05, "loss": 0.6069, "step": 7500 }, { "epoch": 108.7, "eval_loss": 0.22459882497787476, "eval_runtime": 574.1645, "eval_samples_per_second": 5.911, "eval_steps_per_second": 0.74, "eval_wer": 0.1592502198226627, "step": 7500 }, { "epoch": 108.84, "learning_rate": 4.9171875e-05, "loss": 0.7501, "step": 7510 }, { "epoch": 108.99, "learning_rate": 4.9093750000000004e-05, "loss": 0.585, "step": 7520 }, { "epoch": 109.13, "learning_rate": 4.901562500000001e-05, "loss": 0.6811, "step": 7530 }, { "epoch": 109.28, "learning_rate": 4.8937500000000004e-05, "loss": 0.7343, "step": 7540 }, { "epoch": 109.42, "learning_rate": 4.8859375e-05, "loss": 0.6275, "step": 7550 }, { "epoch": 109.57, "learning_rate": 4.878125e-05, "loss": 0.6326, "step": 7560 }, { "epoch": 109.71, "learning_rate": 4.8703125000000006e-05, "loss": 0.6374, "step": 7570 }, { "epoch": 109.86, "learning_rate": 4.8625e-05, "loss": 0.6079, "step": 7580 }, { "epoch": 110.0, "learning_rate": 4.8546875000000006e-05, "loss": 0.5572, "step": 7590 }, { "epoch": 110.14, "learning_rate": 4.846875e-05, "loss": 0.6443, "step": 7600 }, { "epoch": 110.29, "learning_rate": 4.8390625e-05, "loss": 0.5898, "step": 7610 }, { "epoch": 110.43, "learning_rate": 4.83125e-05, "loss": 0.6258, "step": 7620 }, { "epoch": 110.58, "learning_rate": 4.8234375000000005e-05, "loss": 0.6048, "step": 7630 }, { "epoch": 110.72, "learning_rate": 4.815625e-05, "loss": 0.6092, "step": 7640 }, { "epoch": 110.87, "learning_rate": 4.8078125000000005e-05, "loss": 0.5939, "step": 7650 }, { "epoch": 111.01, "learning_rate": 4.8e-05, "loss": 0.6584, "step": 7660 }, { "epoch": 111.16, "learning_rate": 4.7921875000000004e-05, "loss": 0.6788, "step": 7670 }, { "epoch": 111.3, "learning_rate": 4.784375e-05, "loss": 0.6859, "step": 7680 }, { "epoch": 111.45, "learning_rate": 4.7765625000000004e-05, "loss": 0.6178, "step": 7690 }, { "epoch": 111.59, "learning_rate": 4.76875e-05, "loss": 0.5804, "step": 7700 }, { "epoch": 111.74, "learning_rate": 4.7609375000000004e-05, "loss": 0.693, "step": 7710 }, { "epoch": 111.88, "learning_rate": 4.753125000000001e-05, "loss": 0.6894, "step": 7720 }, { "epoch": 112.03, "learning_rate": 4.7453125e-05, "loss": 0.5552, "step": 7730 }, { "epoch": 112.17, "learning_rate": 4.7375e-05, "loss": 0.5774, "step": 7740 }, { "epoch": 112.32, "learning_rate": 4.7296875e-05, "loss": 0.5482, "step": 7750 }, { "epoch": 112.46, "learning_rate": 4.721875e-05, "loss": 0.6571, "step": 7760 }, { "epoch": 112.61, "learning_rate": 4.7140625e-05, "loss": 0.6455, "step": 7770 }, { "epoch": 112.75, "learning_rate": 4.7062500000000006e-05, "loss": 0.6115, "step": 7780 }, { "epoch": 112.9, "learning_rate": 4.6984375e-05, "loss": 0.6006, "step": 7790 }, { "epoch": 113.04, "learning_rate": 4.690625e-05, "loss": 0.5906, "step": 7800 }, { "epoch": 113.19, "learning_rate": 4.6828125e-05, "loss": 0.5477, "step": 7810 }, { "epoch": 113.33, "learning_rate": 4.6750000000000005e-05, "loss": 0.6492, "step": 7820 }, { "epoch": 113.48, "learning_rate": 4.6671875e-05, "loss": 0.5919, "step": 7830 }, { "epoch": 113.62, "learning_rate": 4.6593750000000004e-05, "loss": 0.5931, "step": 7840 }, { "epoch": 113.77, "learning_rate": 4.6515625e-05, "loss": 0.6467, "step": 7850 }, { "epoch": 113.91, "learning_rate": 4.64375e-05, "loss": 0.5646, "step": 7860 }, { "epoch": 114.06, "learning_rate": 4.635937500000001e-05, "loss": 0.6137, "step": 7870 }, { "epoch": 114.2, "learning_rate": 4.6281250000000003e-05, "loss": 0.5523, "step": 7880 }, { "epoch": 114.35, "learning_rate": 4.6203125e-05, "loss": 0.6965, "step": 7890 }, { "epoch": 114.49, "learning_rate": 4.6125e-05, "loss": 0.542, "step": 7900 }, { "epoch": 114.64, "learning_rate": 4.6046875e-05, "loss": 0.5662, "step": 7910 }, { "epoch": 114.78, "learning_rate": 4.596875e-05, "loss": 0.5677, "step": 7920 }, { "epoch": 114.93, "learning_rate": 4.5890625000000006e-05, "loss": 0.5547, "step": 7930 }, { "epoch": 115.07, "learning_rate": 4.58125e-05, "loss": 0.6085, "step": 7940 }, { "epoch": 115.22, "learning_rate": 4.5734375e-05, "loss": 0.5735, "step": 7950 }, { "epoch": 115.36, "learning_rate": 4.565625e-05, "loss": 0.5977, "step": 7960 }, { "epoch": 115.51, "learning_rate": 4.5578125000000005e-05, "loss": 0.6654, "step": 7970 }, { "epoch": 115.65, "learning_rate": 4.55e-05, "loss": 0.6036, "step": 7980 }, { "epoch": 115.8, "learning_rate": 4.5421875000000005e-05, "loss": 0.6112, "step": 7990 }, { "epoch": 115.94, "learning_rate": 4.534375e-05, "loss": 0.5929, "step": 8000 }, { "epoch": 115.94, "eval_loss": 0.2322680950164795, "eval_runtime": 572.4419, "eval_samples_per_second": 5.929, "eval_steps_per_second": 0.742, "eval_wer": 0.16172373838226956, "step": 8000 }, { "epoch": 116.09, "learning_rate": 4.5265625e-05, "loss": 0.5724, "step": 8010 }, { "epoch": 116.23, "learning_rate": 4.518750000000001e-05, "loss": 0.589, "step": 8020 }, { "epoch": 116.38, "learning_rate": 4.5109375000000004e-05, "loss": 0.607, "step": 8030 }, { "epoch": 116.52, "learning_rate": 4.503125e-05, "loss": 0.6171, "step": 8040 }, { "epoch": 116.67, "learning_rate": 4.4953125000000003e-05, "loss": 0.6011, "step": 8050 }, { "epoch": 116.81, "learning_rate": 4.4875e-05, "loss": 0.6247, "step": 8060 }, { "epoch": 116.96, "learning_rate": 4.4796875e-05, "loss": 0.566, "step": 8070 }, { "epoch": 117.1, "learning_rate": 4.4718750000000006e-05, "loss": 0.6967, "step": 8080 }, { "epoch": 117.25, "learning_rate": 4.4640625e-05, "loss": 0.6296, "step": 8090 }, { "epoch": 117.39, "learning_rate": 4.45625e-05, "loss": 0.6186, "step": 8100 }, { "epoch": 117.54, "learning_rate": 4.4484375e-05, "loss": 0.5811, "step": 8110 }, { "epoch": 117.68, "learning_rate": 4.4406250000000005e-05, "loss": 0.5486, "step": 8120 }, { "epoch": 117.83, "learning_rate": 4.4328125e-05, "loss": 0.5805, "step": 8130 }, { "epoch": 117.97, "learning_rate": 4.4250000000000005e-05, "loss": 0.572, "step": 8140 }, { "epoch": 118.12, "learning_rate": 4.4171875e-05, "loss": 0.6108, "step": 8150 }, { "epoch": 118.26, "learning_rate": 4.409375e-05, "loss": 0.666, "step": 8160 }, { "epoch": 118.41, "learning_rate": 4.401562500000001e-05, "loss": 0.5758, "step": 8170 }, { "epoch": 118.55, "learning_rate": 4.3937500000000004e-05, "loss": 0.6212, "step": 8180 }, { "epoch": 118.7, "learning_rate": 4.3859375e-05, "loss": 0.6679, "step": 8190 }, { "epoch": 118.84, "learning_rate": 4.3781250000000004e-05, "loss": 0.6261, "step": 8200 }, { "epoch": 118.99, "learning_rate": 4.3703125e-05, "loss": 0.6025, "step": 8210 }, { "epoch": 119.13, "learning_rate": 4.3625e-05, "loss": 0.6722, "step": 8220 }, { "epoch": 119.28, "learning_rate": 4.3546875000000006e-05, "loss": 0.5637, "step": 8230 }, { "epoch": 119.42, "learning_rate": 4.346875e-05, "loss": 0.6138, "step": 8240 }, { "epoch": 119.57, "learning_rate": 4.3390625e-05, "loss": 0.6397, "step": 8250 }, { "epoch": 119.71, "learning_rate": 4.33125e-05, "loss": 0.5852, "step": 8260 }, { "epoch": 119.86, "learning_rate": 4.3234375000000006e-05, "loss": 0.616, "step": 8270 }, { "epoch": 120.0, "learning_rate": 4.315625e-05, "loss": 0.5824, "step": 8280 }, { "epoch": 120.14, "learning_rate": 4.3078125000000005e-05, "loss": 0.5909, "step": 8290 }, { "epoch": 120.29, "learning_rate": 4.3e-05, "loss": 0.6523, "step": 8300 }, { "epoch": 120.43, "learning_rate": 4.2921875e-05, "loss": 0.6134, "step": 8310 }, { "epoch": 120.58, "learning_rate": 4.284375000000001e-05, "loss": 0.5903, "step": 8320 }, { "epoch": 120.72, "learning_rate": 4.2765625000000004e-05, "loss": 0.5471, "step": 8330 }, { "epoch": 120.87, "learning_rate": 4.26875e-05, "loss": 0.62, "step": 8340 }, { "epoch": 121.01, "learning_rate": 4.2609375000000004e-05, "loss": 0.6299, "step": 8350 }, { "epoch": 121.16, "learning_rate": 4.253125e-05, "loss": 0.5991, "step": 8360 }, { "epoch": 121.3, "learning_rate": 4.2453125000000004e-05, "loss": 0.597, "step": 8370 }, { "epoch": 121.45, "learning_rate": 4.237500000000001e-05, "loss": 0.5925, "step": 8380 }, { "epoch": 121.59, "learning_rate": 4.2296875e-05, "loss": 0.6336, "step": 8390 }, { "epoch": 121.74, "learning_rate": 4.221875e-05, "loss": 0.5826, "step": 8400 }, { "epoch": 121.88, "learning_rate": 4.2140625e-05, "loss": 0.5964, "step": 8410 }, { "epoch": 122.03, "learning_rate": 4.2062500000000006e-05, "loss": 0.5978, "step": 8420 }, { "epoch": 122.17, "learning_rate": 4.1984375e-05, "loss": 0.6456, "step": 8430 }, { "epoch": 122.32, "learning_rate": 4.1906250000000006e-05, "loss": 0.6088, "step": 8440 }, { "epoch": 122.46, "learning_rate": 4.1828125e-05, "loss": 0.6204, "step": 8450 }, { "epoch": 122.61, "learning_rate": 4.175e-05, "loss": 0.5545, "step": 8460 }, { "epoch": 122.75, "learning_rate": 4.1671875e-05, "loss": 0.5823, "step": 8470 }, { "epoch": 122.9, "learning_rate": 4.1593750000000005e-05, "loss": 0.5386, "step": 8480 }, { "epoch": 123.04, "learning_rate": 4.1515625e-05, "loss": 0.5619, "step": 8490 }, { "epoch": 123.19, "learning_rate": 4.1437500000000004e-05, "loss": 0.6218, "step": 8500 }, { "epoch": 123.19, "eval_loss": 0.22871814668178558, "eval_runtime": 573.0196, "eval_samples_per_second": 5.923, "eval_steps_per_second": 0.742, "eval_wer": 0.1565712595222247, "step": 8500 }, { "epoch": 123.33, "learning_rate": 4.1359375e-05, "loss": 0.6351, "step": 8510 }, { "epoch": 123.48, "learning_rate": 4.1281250000000004e-05, "loss": 0.5678, "step": 8520 }, { "epoch": 123.62, "learning_rate": 4.1203125e-05, "loss": 0.7051, "step": 8530 }, { "epoch": 123.77, "learning_rate": 4.1125000000000004e-05, "loss": 0.5785, "step": 8540 }, { "epoch": 123.91, "learning_rate": 4.1046875e-05, "loss": 0.6527, "step": 8550 }, { "epoch": 124.06, "learning_rate": 4.096875e-05, "loss": 0.5509, "step": 8560 }, { "epoch": 124.2, "learning_rate": 4.0890625000000006e-05, "loss": 0.6065, "step": 8570 }, { "epoch": 124.35, "learning_rate": 4.08125e-05, "loss": 0.6249, "step": 8580 }, { "epoch": 124.49, "learning_rate": 4.0734375e-05, "loss": 0.5477, "step": 8590 }, { "epoch": 124.64, "learning_rate": 4.065625e-05, "loss": 0.6066, "step": 8600 }, { "epoch": 124.78, "learning_rate": 4.0585937500000007e-05, "loss": 0.5868, "step": 8610 }, { "epoch": 124.93, "learning_rate": 4.05078125e-05, "loss": 0.5902, "step": 8620 }, { "epoch": 125.07, "learning_rate": 4.04296875e-05, "loss": 0.6138, "step": 8630 }, { "epoch": 125.22, "learning_rate": 4.03515625e-05, "loss": 0.6382, "step": 8640 }, { "epoch": 125.36, "learning_rate": 4.02734375e-05, "loss": 0.6279, "step": 8650 }, { "epoch": 125.51, "learning_rate": 4.01953125e-05, "loss": 0.6699, "step": 8660 }, { "epoch": 125.65, "learning_rate": 4.0117187500000005e-05, "loss": 0.6396, "step": 8670 }, { "epoch": 125.8, "learning_rate": 4.00390625e-05, "loss": 0.564, "step": 8680 }, { "epoch": 125.94, "learning_rate": 3.99609375e-05, "loss": 0.556, "step": 8690 }, { "epoch": 126.09, "learning_rate": 3.98828125e-05, "loss": 0.6425, "step": 8700 }, { "epoch": 126.23, "learning_rate": 3.9804687500000004e-05, "loss": 0.5753, "step": 8710 }, { "epoch": 126.38, "learning_rate": 3.97265625e-05, "loss": 0.5892, "step": 8720 }, { "epoch": 126.52, "learning_rate": 3.9648437500000004e-05, "loss": 0.5827, "step": 8730 }, { "epoch": 126.67, "learning_rate": 3.95703125e-05, "loss": 0.6517, "step": 8740 }, { "epoch": 126.81, "learning_rate": 3.94921875e-05, "loss": 0.5346, "step": 8750 }, { "epoch": 126.96, "learning_rate": 3.941406250000001e-05, "loss": 0.5868, "step": 8760 }, { "epoch": 127.1, "learning_rate": 3.93359375e-05, "loss": 0.5673, "step": 8770 }, { "epoch": 127.25, "learning_rate": 3.92578125e-05, "loss": 0.59, "step": 8780 }, { "epoch": 127.39, "learning_rate": 3.91796875e-05, "loss": 0.5705, "step": 8790 }, { "epoch": 127.54, "learning_rate": 3.91015625e-05, "loss": 0.5958, "step": 8800 }, { "epoch": 127.68, "learning_rate": 3.90234375e-05, "loss": 0.5522, "step": 8810 }, { "epoch": 127.83, "learning_rate": 3.8945312500000006e-05, "loss": 0.581, "step": 8820 }, { "epoch": 127.97, "learning_rate": 3.88671875e-05, "loss": 0.5469, "step": 8830 }, { "epoch": 128.12, "learning_rate": 3.87890625e-05, "loss": 0.6312, "step": 8840 }, { "epoch": 128.26, "learning_rate": 3.87109375e-05, "loss": 0.6486, "step": 8850 }, { "epoch": 128.41, "learning_rate": 3.8632812500000005e-05, "loss": 0.6362, "step": 8860 }, { "epoch": 128.55, "learning_rate": 3.85546875e-05, "loss": 0.5602, "step": 8870 }, { "epoch": 128.7, "learning_rate": 3.8476562500000004e-05, "loss": 0.5554, "step": 8880 }, { "epoch": 128.84, "learning_rate": 3.83984375e-05, "loss": 0.5491, "step": 8890 }, { "epoch": 128.99, "learning_rate": 3.83203125e-05, "loss": 0.5681, "step": 8900 }, { "epoch": 129.13, "learning_rate": 3.824218750000001e-05, "loss": 0.6239, "step": 8910 }, { "epoch": 129.28, "learning_rate": 3.8164062500000004e-05, "loss": 0.5759, "step": 8920 }, { "epoch": 129.42, "learning_rate": 3.80859375e-05, "loss": 0.5868, "step": 8930 }, { "epoch": 129.57, "learning_rate": 3.8015625e-05, "loss": 0.5796, "step": 8940 }, { "epoch": 129.71, "learning_rate": 3.79375e-05, "loss": 0.5501, "step": 8950 }, { "epoch": 129.86, "learning_rate": 3.7859375000000004e-05, "loss": 0.5976, "step": 8960 }, { "epoch": 130.0, "learning_rate": 3.778125e-05, "loss": 0.5488, "step": 8970 }, { "epoch": 130.14, "learning_rate": 3.7703125e-05, "loss": 0.5688, "step": 8980 }, { "epoch": 130.29, "learning_rate": 3.7625e-05, "loss": 0.5448, "step": 8990 }, { "epoch": 130.43, "learning_rate": 3.7546875e-05, "loss": 0.5751, "step": 9000 }, { "epoch": 130.43, "eval_loss": 0.22747375071048737, "eval_runtime": 574.7485, "eval_samples_per_second": 5.905, "eval_steps_per_second": 0.739, "eval_wer": 0.1562672057457946, "step": 9000 }, { "epoch": 130.58, "learning_rate": 3.746875e-05, "loss": 0.6, "step": 9010 }, { "epoch": 130.72, "learning_rate": 3.7390625e-05, "loss": 0.5781, "step": 9020 }, { "epoch": 130.87, "learning_rate": 3.73125e-05, "loss": 0.5545, "step": 9030 }, { "epoch": 131.01, "learning_rate": 3.7234375e-05, "loss": 0.6116, "step": 9040 }, { "epoch": 131.16, "learning_rate": 3.7156250000000005e-05, "loss": 0.5634, "step": 9050 }, { "epoch": 131.3, "learning_rate": 3.7078125e-05, "loss": 0.6436, "step": 9060 }, { "epoch": 131.45, "learning_rate": 3.7e-05, "loss": 0.5882, "step": 9070 }, { "epoch": 131.59, "learning_rate": 3.6921875e-05, "loss": 0.5391, "step": 9080 }, { "epoch": 131.74, "learning_rate": 3.684375e-05, "loss": 0.5211, "step": 9090 }, { "epoch": 131.88, "learning_rate": 3.6765625e-05, "loss": 0.5971, "step": 9100 }, { "epoch": 132.03, "learning_rate": 3.6687500000000004e-05, "loss": 0.5314, "step": 9110 }, { "epoch": 132.17, "learning_rate": 3.6609375e-05, "loss": 0.573, "step": 9120 }, { "epoch": 132.32, "learning_rate": 3.653125e-05, "loss": 0.5678, "step": 9130 }, { "epoch": 132.46, "learning_rate": 3.6453125e-05, "loss": 0.5369, "step": 9140 }, { "epoch": 132.61, "learning_rate": 3.6375e-05, "loss": 0.5976, "step": 9150 }, { "epoch": 132.75, "learning_rate": 3.6296875e-05, "loss": 0.609, "step": 9160 }, { "epoch": 132.9, "learning_rate": 3.621875e-05, "loss": 0.5859, "step": 9170 }, { "epoch": 133.04, "learning_rate": 3.6140625e-05, "loss": 0.6005, "step": 9180 }, { "epoch": 133.19, "learning_rate": 3.60625e-05, "loss": 0.5796, "step": 9190 }, { "epoch": 133.33, "learning_rate": 3.5984375000000006e-05, "loss": 0.5125, "step": 9200 }, { "epoch": 133.48, "learning_rate": 3.590625e-05, "loss": 0.5465, "step": 9210 }, { "epoch": 133.62, "learning_rate": 3.5828125e-05, "loss": 0.5985, "step": 9220 }, { "epoch": 133.77, "learning_rate": 3.575e-05, "loss": 0.5687, "step": 9230 }, { "epoch": 133.91, "learning_rate": 3.5671875e-05, "loss": 0.6586, "step": 9240 }, { "epoch": 134.06, "learning_rate": 3.559375e-05, "loss": 0.5656, "step": 9250 }, { "epoch": 134.2, "learning_rate": 3.5515625000000004e-05, "loss": 0.6136, "step": 9260 }, { "epoch": 134.35, "learning_rate": 3.54375e-05, "loss": 0.6063, "step": 9270 }, { "epoch": 134.49, "learning_rate": 3.5359375e-05, "loss": 0.5242, "step": 9280 }, { "epoch": 134.64, "learning_rate": 3.528125e-05, "loss": 0.5605, "step": 9290 }, { "epoch": 134.78, "learning_rate": 3.5203125000000004e-05, "loss": 0.5774, "step": 9300 }, { "epoch": 134.93, "learning_rate": 3.5125e-05, "loss": 0.5937, "step": 9310 }, { "epoch": 135.07, "learning_rate": 3.5046875e-05, "loss": 0.5875, "step": 9320 }, { "epoch": 135.22, "learning_rate": 3.496875e-05, "loss": 0.5483, "step": 9330 }, { "epoch": 135.36, "learning_rate": 3.4890624999999996e-05, "loss": 0.5705, "step": 9340 }, { "epoch": 135.51, "learning_rate": 3.4812500000000006e-05, "loss": 0.6026, "step": 9350 }, { "epoch": 135.65, "learning_rate": 3.4734375e-05, "loss": 0.5888, "step": 9360 }, { "epoch": 135.8, "learning_rate": 3.465625e-05, "loss": 0.5913, "step": 9370 }, { "epoch": 135.94, "learning_rate": 3.4578125e-05, "loss": 0.5792, "step": 9380 }, { "epoch": 136.09, "learning_rate": 3.45e-05, "loss": 0.615, "step": 9390 }, { "epoch": 136.23, "learning_rate": 3.4421875e-05, "loss": 0.5827, "step": 9400 }, { "epoch": 136.38, "learning_rate": 3.4343750000000005e-05, "loss": 0.5956, "step": 9410 }, { "epoch": 136.52, "learning_rate": 3.4265625e-05, "loss": 0.4951, "step": 9420 }, { "epoch": 136.67, "learning_rate": 3.41875e-05, "loss": 0.5786, "step": 9430 }, { "epoch": 136.81, "learning_rate": 3.4109375e-05, "loss": 0.5307, "step": 9440 }, { "epoch": 136.96, "learning_rate": 3.4031250000000004e-05, "loss": 0.6652, "step": 9450 }, { "epoch": 137.1, "learning_rate": 3.3953125e-05, "loss": 0.5879, "step": 9460 }, { "epoch": 137.25, "learning_rate": 3.3875000000000003e-05, "loss": 0.6949, "step": 9470 }, { "epoch": 137.39, "learning_rate": 3.3796875e-05, "loss": 0.5537, "step": 9480 }, { "epoch": 137.54, "learning_rate": 3.3718749999999996e-05, "loss": 0.5923, "step": 9490 }, { "epoch": 137.68, "learning_rate": 3.3640625000000006e-05, "loss": 0.5181, "step": 9500 }, { "epoch": 137.68, "eval_loss": 0.23162005841732025, "eval_runtime": 574.0095, "eval_samples_per_second": 5.913, "eval_steps_per_second": 0.74, "eval_wer": 0.15791895734207692, "step": 9500 }, { "epoch": 137.83, "learning_rate": 3.35625e-05, "loss": 0.5543, "step": 9510 }, { "epoch": 137.97, "learning_rate": 3.3484375e-05, "loss": 0.5875, "step": 9520 }, { "epoch": 138.12, "learning_rate": 3.340625e-05, "loss": 0.5757, "step": 9530 }, { "epoch": 138.26, "learning_rate": 3.3328125e-05, "loss": 0.5366, "step": 9540 }, { "epoch": 138.41, "learning_rate": 3.325e-05, "loss": 0.5328, "step": 9550 }, { "epoch": 138.55, "learning_rate": 3.3171875000000005e-05, "loss": 0.5193, "step": 9560 }, { "epoch": 138.7, "learning_rate": 3.309375e-05, "loss": 0.5377, "step": 9570 }, { "epoch": 138.84, "learning_rate": 3.3015625e-05, "loss": 0.5687, "step": 9580 }, { "epoch": 138.99, "learning_rate": 3.29375e-05, "loss": 0.5416, "step": 9590 }, { "epoch": 139.13, "learning_rate": 3.2859375000000004e-05, "loss": 0.5427, "step": 9600 }, { "epoch": 139.28, "learning_rate": 3.278125e-05, "loss": 0.577, "step": 9610 }, { "epoch": 139.42, "learning_rate": 3.2703125000000004e-05, "loss": 0.5943, "step": 9620 }, { "epoch": 139.57, "learning_rate": 3.2625e-05, "loss": 0.5481, "step": 9630 }, { "epoch": 139.71, "learning_rate": 3.2546874999999997e-05, "loss": 0.5904, "step": 9640 }, { "epoch": 139.86, "learning_rate": 3.2468750000000007e-05, "loss": 0.6591, "step": 9650 }, { "epoch": 140.0, "learning_rate": 3.2390625e-05, "loss": 0.6146, "step": 9660 }, { "epoch": 140.14, "learning_rate": 3.23125e-05, "loss": 0.5974, "step": 9670 }, { "epoch": 140.29, "learning_rate": 3.2234375e-05, "loss": 0.546, "step": 9680 }, { "epoch": 140.43, "learning_rate": 3.215625e-05, "loss": 0.5433, "step": 9690 }, { "epoch": 140.58, "learning_rate": 3.2078125e-05, "loss": 0.5357, "step": 9700 }, { "epoch": 140.72, "learning_rate": 3.2000000000000005e-05, "loss": 0.6074, "step": 9710 }, { "epoch": 140.87, "learning_rate": 3.1921875e-05, "loss": 0.5543, "step": 9720 }, { "epoch": 141.01, "learning_rate": 3.184375e-05, "loss": 0.5415, "step": 9730 }, { "epoch": 141.16, "learning_rate": 3.1765625e-05, "loss": 0.5853, "step": 9740 }, { "epoch": 141.3, "learning_rate": 3.1687500000000005e-05, "loss": 0.5143, "step": 9750 }, { "epoch": 141.45, "learning_rate": 3.1609375e-05, "loss": 0.503, "step": 9760 }, { "epoch": 141.59, "learning_rate": 3.1531250000000004e-05, "loss": 0.5154, "step": 9770 }, { "epoch": 141.74, "learning_rate": 3.1453125e-05, "loss": 0.562, "step": 9780 }, { "epoch": 141.88, "learning_rate": 3.1375e-05, "loss": 0.5454, "step": 9790 }, { "epoch": 142.03, "learning_rate": 3.1296875e-05, "loss": 0.5722, "step": 9800 }, { "epoch": 142.17, "learning_rate": 3.121875e-05, "loss": 0.6911, "step": 9810 }, { "epoch": 142.32, "learning_rate": 3.1140625e-05, "loss": 0.5725, "step": 9820 }, { "epoch": 142.46, "learning_rate": 3.10625e-05, "loss": 0.5842, "step": 9830 }, { "epoch": 142.61, "learning_rate": 3.0984375e-05, "loss": 0.5597, "step": 9840 }, { "epoch": 142.75, "learning_rate": 3.090625e-05, "loss": 0.592, "step": 9850 }, { "epoch": 142.9, "learning_rate": 3.0828125e-05, "loss": 0.5506, "step": 9860 }, { "epoch": 143.04, "learning_rate": 3.075e-05, "loss": 0.6292, "step": 9870 }, { "epoch": 143.19, "learning_rate": 3.0671875e-05, "loss": 0.5672, "step": 9880 }, { "epoch": 143.33, "learning_rate": 3.059375e-05, "loss": 0.5315, "step": 9890 }, { "epoch": 143.48, "learning_rate": 3.0515625000000005e-05, "loss": 0.5563, "step": 9900 }, { "epoch": 143.62, "learning_rate": 3.04375e-05, "loss": 0.5877, "step": 9910 }, { "epoch": 143.77, "learning_rate": 3.0359375e-05, "loss": 0.546, "step": 9920 }, { "epoch": 143.91, "learning_rate": 3.028125e-05, "loss": 0.5425, "step": 9930 }, { "epoch": 144.06, "learning_rate": 3.0203124999999997e-05, "loss": 0.5368, "step": 9940 }, { "epoch": 144.2, "learning_rate": 3.0125000000000004e-05, "loss": 0.538, "step": 9950 }, { "epoch": 144.35, "learning_rate": 3.0046875000000004e-05, "loss": 0.5367, "step": 9960 }, { "epoch": 144.49, "learning_rate": 2.996875e-05, "loss": 0.5591, "step": 9970 }, { "epoch": 144.64, "learning_rate": 2.9890625e-05, "loss": 0.5554, "step": 9980 }, { "epoch": 144.78, "learning_rate": 2.98125e-05, "loss": 0.4985, "step": 9990 }, { "epoch": 144.93, "learning_rate": 2.9734375000000003e-05, "loss": 0.6306, "step": 10000 }, { "epoch": 144.93, "eval_loss": 0.23715920746326447, "eval_runtime": 577.4706, "eval_samples_per_second": 5.877, "eval_steps_per_second": 0.736, "eval_wer": 0.15560157450550172, "step": 10000 }, { "epoch": 145.07, "learning_rate": 2.9656250000000003e-05, "loss": 0.5794, "step": 10010 }, { "epoch": 145.22, "learning_rate": 2.9578125000000002e-05, "loss": 0.5792, "step": 10020 }, { "epoch": 145.36, "learning_rate": 2.95e-05, "loss": 0.5664, "step": 10030 }, { "epoch": 145.51, "learning_rate": 2.9421875e-05, "loss": 0.5737, "step": 10040 }, { "epoch": 145.65, "learning_rate": 2.9343750000000002e-05, "loss": 0.5735, "step": 10050 }, { "epoch": 145.8, "learning_rate": 2.9265625e-05, "loss": 0.5972, "step": 10060 }, { "epoch": 145.94, "learning_rate": 2.91875e-05, "loss": 0.5998, "step": 10070 }, { "epoch": 146.09, "learning_rate": 2.9109375e-05, "loss": 0.5357, "step": 10080 }, { "epoch": 146.23, "learning_rate": 2.9031249999999998e-05, "loss": 0.522, "step": 10090 }, { "epoch": 146.38, "learning_rate": 2.8953125000000004e-05, "loss": 0.5231, "step": 10100 }, { "epoch": 146.52, "learning_rate": 2.8875e-05, "loss": 0.5343, "step": 10110 }, { "epoch": 146.67, "learning_rate": 2.8796875e-05, "loss": 0.5996, "step": 10120 }, { "epoch": 146.81, "learning_rate": 2.871875e-05, "loss": 0.5595, "step": 10130 }, { "epoch": 146.96, "learning_rate": 2.8640625e-05, "loss": 0.5386, "step": 10140 }, { "epoch": 147.1, "learning_rate": 2.8562500000000003e-05, "loss": 0.5338, "step": 10150 }, { "epoch": 147.25, "learning_rate": 2.8484375000000003e-05, "loss": 0.5273, "step": 10160 }, { "epoch": 147.39, "learning_rate": 2.840625e-05, "loss": 0.628, "step": 10170 }, { "epoch": 147.54, "learning_rate": 2.8328125e-05, "loss": 0.5602, "step": 10180 }, { "epoch": 147.68, "learning_rate": 2.825e-05, "loss": 0.621, "step": 10190 }, { "epoch": 147.83, "learning_rate": 2.8171875000000002e-05, "loss": 0.6112, "step": 10200 }, { "epoch": 147.97, "learning_rate": 2.8093750000000002e-05, "loss": 0.4686, "step": 10210 }, { "epoch": 148.12, "learning_rate": 2.8015625e-05, "loss": 0.5582, "step": 10220 }, { "epoch": 148.26, "learning_rate": 2.79375e-05, "loss": 0.5353, "step": 10230 }, { "epoch": 148.41, "learning_rate": 2.7859374999999998e-05, "loss": 0.5545, "step": 10240 }, { "epoch": 148.55, "learning_rate": 2.7781250000000004e-05, "loss": 0.497, "step": 10250 }, { "epoch": 148.7, "learning_rate": 2.7703125e-05, "loss": 0.6022, "step": 10260 }, { "epoch": 148.84, "learning_rate": 2.7625e-05, "loss": 0.5438, "step": 10270 }, { "epoch": 148.99, "learning_rate": 2.7546875e-05, "loss": 0.6067, "step": 10280 }, { "epoch": 149.13, "learning_rate": 2.746875e-05, "loss": 0.5966, "step": 10290 }, { "epoch": 149.28, "learning_rate": 2.7390625000000003e-05, "loss": 0.6345, "step": 10300 }, { "epoch": 149.42, "learning_rate": 2.7312500000000003e-05, "loss": 0.5105, "step": 10310 }, { "epoch": 149.57, "learning_rate": 2.7234375e-05, "loss": 0.573, "step": 10320 }, { "epoch": 149.71, "learning_rate": 2.715625e-05, "loss": 0.5542, "step": 10330 }, { "epoch": 149.86, "learning_rate": 2.7078125e-05, "loss": 0.6154, "step": 10340 }, { "epoch": 150.0, "learning_rate": 2.7000000000000002e-05, "loss": 0.5475, "step": 10350 }, { "epoch": 150.14, "learning_rate": 2.6921875000000002e-05, "loss": 0.5476, "step": 10360 }, { "epoch": 150.29, "learning_rate": 2.6843750000000002e-05, "loss": 0.5667, "step": 10370 }, { "epoch": 150.43, "learning_rate": 2.6765625e-05, "loss": 0.5752, "step": 10380 }, { "epoch": 150.58, "learning_rate": 2.6687499999999998e-05, "loss": 0.5797, "step": 10390 }, { "epoch": 150.72, "learning_rate": 2.6609375000000005e-05, "loss": 0.5585, "step": 10400 }, { "epoch": 150.87, "learning_rate": 2.653125e-05, "loss": 0.6719, "step": 10410 }, { "epoch": 151.01, "learning_rate": 2.6453125e-05, "loss": 0.535, "step": 10420 }, { "epoch": 151.16, "learning_rate": 2.6375e-05, "loss": 0.5853, "step": 10430 }, { "epoch": 151.3, "learning_rate": 2.6296874999999997e-05, "loss": 0.5341, "step": 10440 }, { "epoch": 151.45, "learning_rate": 2.6218750000000004e-05, "loss": 0.4932, "step": 10450 }, { "epoch": 151.59, "learning_rate": 2.6140625000000004e-05, "loss": 0.5924, "step": 10460 }, { "epoch": 151.74, "learning_rate": 2.60625e-05, "loss": 0.5665, "step": 10470 }, { "epoch": 151.88, "learning_rate": 2.5984375e-05, "loss": 0.5202, "step": 10480 }, { "epoch": 152.03, "learning_rate": 2.590625e-05, "loss": 0.5209, "step": 10490 }, { "epoch": 152.17, "learning_rate": 2.5828125000000003e-05, "loss": 0.5874, "step": 10500 }, { "epoch": 152.17, "eval_loss": 0.23621943593025208, "eval_runtime": 575.3543, "eval_samples_per_second": 5.899, "eval_steps_per_second": 0.739, "eval_wer": 0.15330884467782627, "step": 10500 }, { "epoch": 152.32, "learning_rate": 2.5750000000000002e-05, "loss": 0.57, "step": 10510 }, { "epoch": 152.46, "learning_rate": 2.5671875000000002e-05, "loss": 0.58, "step": 10520 }, { "epoch": 152.61, "learning_rate": 2.559375e-05, "loss": 0.5269, "step": 10530 }, { "epoch": 152.75, "learning_rate": 2.5515625e-05, "loss": 0.5519, "step": 10540 }, { "epoch": 152.9, "learning_rate": 2.54375e-05, "loss": 0.497, "step": 10550 }, { "epoch": 153.04, "learning_rate": 2.5359375e-05, "loss": 0.5686, "step": 10560 }, { "epoch": 153.19, "learning_rate": 2.528125e-05, "loss": 0.5219, "step": 10570 }, { "epoch": 153.33, "learning_rate": 2.5203125e-05, "loss": 0.5695, "step": 10580 }, { "epoch": 153.48, "learning_rate": 2.5124999999999997e-05, "loss": 0.5747, "step": 10590 }, { "epoch": 153.62, "learning_rate": 2.5046875000000004e-05, "loss": 0.5297, "step": 10600 }, { "epoch": 153.77, "learning_rate": 2.496875e-05, "loss": 0.5956, "step": 10610 }, { "epoch": 153.91, "learning_rate": 2.4890625e-05, "loss": 0.5752, "step": 10620 }, { "epoch": 154.06, "learning_rate": 2.4812500000000003e-05, "loss": 0.5765, "step": 10630 }, { "epoch": 154.2, "learning_rate": 2.4734375e-05, "loss": 0.5627, "step": 10640 }, { "epoch": 154.35, "learning_rate": 2.465625e-05, "loss": 0.5591, "step": 10650 }, { "epoch": 154.49, "learning_rate": 2.4578125000000003e-05, "loss": 0.5595, "step": 10660 }, { "epoch": 154.64, "learning_rate": 2.45e-05, "loss": 0.5978, "step": 10670 }, { "epoch": 154.78, "learning_rate": 2.4421875000000002e-05, "loss": 0.5249, "step": 10680 }, { "epoch": 154.93, "learning_rate": 2.4343750000000002e-05, "loss": 0.5522, "step": 10690 }, { "epoch": 155.07, "learning_rate": 2.4265625e-05, "loss": 0.5152, "step": 10700 }, { "epoch": 155.22, "learning_rate": 2.4187500000000002e-05, "loss": 0.6, "step": 10710 }, { "epoch": 155.36, "learning_rate": 2.4109375e-05, "loss": 0.588, "step": 10720 }, { "epoch": 155.51, "learning_rate": 2.403125e-05, "loss": 0.52, "step": 10730 }, { "epoch": 155.65, "learning_rate": 2.3953125e-05, "loss": 0.5191, "step": 10740 }, { "epoch": 155.8, "learning_rate": 2.3875e-05, "loss": 0.5456, "step": 10750 }, { "epoch": 155.94, "learning_rate": 2.3796875e-05, "loss": 0.6365, "step": 10760 }, { "epoch": 156.09, "learning_rate": 2.371875e-05, "loss": 0.5344, "step": 10770 }, { "epoch": 156.23, "learning_rate": 2.3640625000000004e-05, "loss": 0.5185, "step": 10780 }, { "epoch": 156.38, "learning_rate": 2.35625e-05, "loss": 0.5619, "step": 10790 }, { "epoch": 156.52, "learning_rate": 2.3484375e-05, "loss": 0.5248, "step": 10800 }, { "epoch": 156.67, "learning_rate": 2.3406250000000003e-05, "loss": 0.499, "step": 10810 }, { "epoch": 156.81, "learning_rate": 2.3328125e-05, "loss": 0.5657, "step": 10820 }, { "epoch": 156.96, "learning_rate": 2.3250000000000003e-05, "loss": 0.52, "step": 10830 }, { "epoch": 157.1, "learning_rate": 2.3171875000000003e-05, "loss": 0.5119, "step": 10840 }, { "epoch": 157.25, "learning_rate": 2.309375e-05, "loss": 0.5749, "step": 10850 }, { "epoch": 157.39, "learning_rate": 2.3015625000000002e-05, "loss": 0.57, "step": 10860 }, { "epoch": 157.54, "learning_rate": 2.2937500000000002e-05, "loss": 0.5096, "step": 10870 }, { "epoch": 157.68, "learning_rate": 2.2859375e-05, "loss": 0.5025, "step": 10880 }, { "epoch": 157.83, "learning_rate": 2.278125e-05, "loss": 0.5884, "step": 10890 }, { "epoch": 157.97, "learning_rate": 2.2703125e-05, "loss": 0.4872, "step": 10900 }, { "epoch": 158.12, "learning_rate": 2.2625e-05, "loss": 0.5518, "step": 10910 }, { "epoch": 158.26, "learning_rate": 2.2546875e-05, "loss": 0.5348, "step": 10920 }, { "epoch": 158.41, "learning_rate": 2.246875e-05, "loss": 0.5997, "step": 10930 }, { "epoch": 158.55, "learning_rate": 2.2390625e-05, "loss": 0.5602, "step": 10940 }, { "epoch": 158.7, "learning_rate": 2.23125e-05, "loss": 0.5996, "step": 10950 }, { "epoch": 158.84, "learning_rate": 2.2234375e-05, "loss": 0.6152, "step": 10960 }, { "epoch": 158.99, "learning_rate": 2.215625e-05, "loss": 0.5291, "step": 10970 }, { "epoch": 159.13, "learning_rate": 2.2078125000000003e-05, "loss": 0.55, "step": 10980 }, { "epoch": 159.28, "learning_rate": 2.2000000000000003e-05, "loss": 0.6242, "step": 10990 }, { "epoch": 159.42, "learning_rate": 2.1921875e-05, "loss": 0.5546, "step": 11000 }, { "epoch": 159.42, "eval_loss": 0.23421239852905273, "eval_runtime": 576.3026, "eval_samples_per_second": 5.889, "eval_steps_per_second": 0.737, "eval_wer": 0.15427852969454922, "step": 11000 }, { "epoch": 159.57, "learning_rate": 2.1843750000000002e-05, "loss": 0.5893, "step": 11010 }, { "epoch": 159.71, "learning_rate": 2.1765625000000002e-05, "loss": 0.5654, "step": 11020 }, { "epoch": 159.86, "learning_rate": 2.1687500000000002e-05, "loss": 0.5748, "step": 11030 }, { "epoch": 160.0, "learning_rate": 2.1609375000000002e-05, "loss": 0.5317, "step": 11040 }, { "epoch": 160.14, "learning_rate": 2.153125e-05, "loss": 0.4892, "step": 11050 }, { "epoch": 160.29, "learning_rate": 2.1453125e-05, "loss": 0.544, "step": 11060 }, { "epoch": 160.43, "learning_rate": 2.1375e-05, "loss": 0.5197, "step": 11070 }, { "epoch": 160.58, "learning_rate": 2.1296875e-05, "loss": 0.5127, "step": 11080 }, { "epoch": 160.72, "learning_rate": 2.121875e-05, "loss": 0.5438, "step": 11090 }, { "epoch": 160.87, "learning_rate": 2.1140625e-05, "loss": 0.5184, "step": 11100 }, { "epoch": 161.01, "learning_rate": 2.10625e-05, "loss": 0.6122, "step": 11110 }, { "epoch": 161.16, "learning_rate": 2.0984375e-05, "loss": 0.5086, "step": 11120 }, { "epoch": 161.3, "learning_rate": 2.0906250000000003e-05, "loss": 0.6104, "step": 11130 }, { "epoch": 161.45, "learning_rate": 2.0828125e-05, "loss": 0.5168, "step": 11140 }, { "epoch": 161.59, "learning_rate": 2.075e-05, "loss": 0.5145, "step": 11150 }, { "epoch": 161.74, "learning_rate": 2.0671875000000003e-05, "loss": 0.5389, "step": 11160 }, { "epoch": 161.88, "learning_rate": 2.059375e-05, "loss": 0.671, "step": 11170 }, { "epoch": 162.03, "learning_rate": 2.05234375e-05, "loss": 0.5659, "step": 11180 }, { "epoch": 162.17, "learning_rate": 2.0453125e-05, "loss": 0.5035, "step": 11190 }, { "epoch": 162.32, "learning_rate": 2.0375e-05, "loss": 0.5038, "step": 11200 }, { "epoch": 162.46, "learning_rate": 2.0296875e-05, "loss": 0.6035, "step": 11210 }, { "epoch": 162.61, "learning_rate": 2.021875e-05, "loss": 0.5472, "step": 11220 }, { "epoch": 162.75, "learning_rate": 2.0140625000000003e-05, "loss": 0.5149, "step": 11230 }, { "epoch": 162.9, "learning_rate": 2.00625e-05, "loss": 0.6043, "step": 11240 }, { "epoch": 163.04, "learning_rate": 1.9984375e-05, "loss": 0.5377, "step": 11250 }, { "epoch": 163.19, "learning_rate": 1.9906250000000003e-05, "loss": 0.5328, "step": 11260 }, { "epoch": 163.33, "learning_rate": 1.9828125e-05, "loss": 0.4993, "step": 11270 }, { "epoch": 163.48, "learning_rate": 1.9750000000000002e-05, "loss": 0.5234, "step": 11280 }, { "epoch": 163.62, "learning_rate": 1.9671875000000002e-05, "loss": 0.6392, "step": 11290 }, { "epoch": 163.77, "learning_rate": 1.959375e-05, "loss": 0.5308, "step": 11300 }, { "epoch": 163.91, "learning_rate": 1.9515625000000002e-05, "loss": 0.5907, "step": 11310 }, { "epoch": 164.06, "learning_rate": 1.94375e-05, "loss": 0.5679, "step": 11320 }, { "epoch": 164.2, "learning_rate": 1.9359375e-05, "loss": 0.5094, "step": 11330 }, { "epoch": 164.35, "learning_rate": 1.928125e-05, "loss": 0.5877, "step": 11340 }, { "epoch": 164.49, "learning_rate": 1.9203125e-05, "loss": 0.5177, "step": 11350 }, { "epoch": 164.64, "learning_rate": 1.9125e-05, "loss": 0.5452, "step": 11360 }, { "epoch": 164.78, "learning_rate": 1.9046875e-05, "loss": 0.568, "step": 11370 }, { "epoch": 164.93, "learning_rate": 1.896875e-05, "loss": 0.5717, "step": 11380 }, { "epoch": 165.07, "learning_rate": 1.8890625e-05, "loss": 0.5861, "step": 11390 }, { "epoch": 165.22, "learning_rate": 1.88125e-05, "loss": 0.5153, "step": 11400 }, { "epoch": 165.36, "learning_rate": 1.8734375e-05, "loss": 0.5797, "step": 11410 }, { "epoch": 165.51, "learning_rate": 1.865625e-05, "loss": 0.5237, "step": 11420 }, { "epoch": 165.65, "learning_rate": 1.8578125000000003e-05, "loss": 0.5606, "step": 11430 }, { "epoch": 165.8, "learning_rate": 1.85e-05, "loss": 0.5389, "step": 11440 }, { "epoch": 165.94, "learning_rate": 1.8421875e-05, "loss": 0.572, "step": 11450 }, { "epoch": 166.09, "learning_rate": 1.8343750000000002e-05, "loss": 0.5169, "step": 11460 }, { "epoch": 166.23, "learning_rate": 1.8265625e-05, "loss": 0.506, "step": 11470 }, { "epoch": 166.38, "learning_rate": 1.81875e-05, "loss": 0.5406, "step": 11480 }, { "epoch": 166.52, "learning_rate": 1.8109375e-05, "loss": 0.5913, "step": 11490 }, { "epoch": 166.67, "learning_rate": 1.803125e-05, "loss": 0.6294, "step": 11500 }, { "epoch": 166.67, "eval_loss": 0.23806829750537872, "eval_runtime": 576.0097, "eval_samples_per_second": 5.892, "eval_steps_per_second": 0.738, "eval_wer": 0.15362933379352284, "step": 11500 }, { "epoch": 166.81, "learning_rate": 1.7953125e-05, "loss": 0.51, "step": 11510 }, { "epoch": 166.96, "learning_rate": 1.7875e-05, "loss": 0.5439, "step": 11520 }, { "epoch": 167.1, "learning_rate": 1.7796875e-05, "loss": 0.4877, "step": 11530 }, { "epoch": 167.25, "learning_rate": 1.771875e-05, "loss": 0.5299, "step": 11540 }, { "epoch": 167.39, "learning_rate": 1.7640625e-05, "loss": 0.5475, "step": 11550 }, { "epoch": 167.54, "learning_rate": 1.75625e-05, "loss": 0.5661, "step": 11560 }, { "epoch": 167.68, "learning_rate": 1.7484375e-05, "loss": 0.5416, "step": 11570 }, { "epoch": 167.83, "learning_rate": 1.7406250000000003e-05, "loss": 0.4929, "step": 11580 }, { "epoch": 167.97, "learning_rate": 1.7328125e-05, "loss": 0.5195, "step": 11590 }, { "epoch": 168.12, "learning_rate": 1.725e-05, "loss": 0.5883, "step": 11600 }, { "epoch": 168.26, "learning_rate": 1.7171875000000002e-05, "loss": 0.5723, "step": 11610 }, { "epoch": 168.41, "learning_rate": 1.709375e-05, "loss": 0.5621, "step": 11620 }, { "epoch": 168.55, "learning_rate": 1.7015625000000002e-05, "loss": 0.5567, "step": 11630 }, { "epoch": 168.7, "learning_rate": 1.6937500000000002e-05, "loss": 0.4995, "step": 11640 }, { "epoch": 168.84, "learning_rate": 1.6859374999999998e-05, "loss": 0.5601, "step": 11650 }, { "epoch": 168.99, "learning_rate": 1.678125e-05, "loss": 0.5122, "step": 11660 }, { "epoch": 169.13, "learning_rate": 1.6703125e-05, "loss": 0.5406, "step": 11670 }, { "epoch": 169.28, "learning_rate": 1.6625e-05, "loss": 0.4914, "step": 11680 }, { "epoch": 169.42, "learning_rate": 1.6546875e-05, "loss": 0.6525, "step": 11690 }, { "epoch": 169.57, "learning_rate": 1.646875e-05, "loss": 0.5181, "step": 11700 }, { "epoch": 169.71, "learning_rate": 1.6390625e-05, "loss": 0.5328, "step": 11710 }, { "epoch": 169.86, "learning_rate": 1.63125e-05, "loss": 0.5583, "step": 11720 }, { "epoch": 170.0, "learning_rate": 1.6234375000000003e-05, "loss": 0.5376, "step": 11730 }, { "epoch": 170.14, "learning_rate": 1.615625e-05, "loss": 0.5182, "step": 11740 }, { "epoch": 170.29, "learning_rate": 1.6078125e-05, "loss": 0.4756, "step": 11750 }, { "epoch": 170.43, "learning_rate": 1.6000000000000003e-05, "loss": 0.5295, "step": 11760 }, { "epoch": 170.58, "learning_rate": 1.5921875e-05, "loss": 0.5482, "step": 11770 }, { "epoch": 170.72, "learning_rate": 1.5843750000000002e-05, "loss": 0.5513, "step": 11780 }, { "epoch": 170.87, "learning_rate": 1.5765625000000002e-05, "loss": 0.5341, "step": 11790 }, { "epoch": 171.01, "learning_rate": 1.56875e-05, "loss": 0.4755, "step": 11800 }, { "epoch": 171.16, "learning_rate": 1.5609375e-05, "loss": 0.5352, "step": 11810 }, { "epoch": 171.3, "learning_rate": 1.553125e-05, "loss": 0.5318, "step": 11820 }, { "epoch": 171.45, "learning_rate": 1.5453125e-05, "loss": 0.4488, "step": 11830 }, { "epoch": 171.59, "learning_rate": 1.5375e-05, "loss": 0.6219, "step": 11840 }, { "epoch": 171.74, "learning_rate": 1.5296875e-05, "loss": 0.5449, "step": 11850 }, { "epoch": 171.88, "learning_rate": 1.521875e-05, "loss": 0.5278, "step": 11860 }, { "epoch": 172.03, "learning_rate": 1.5140625e-05, "loss": 0.539, "step": 11870 }, { "epoch": 172.17, "learning_rate": 1.5062500000000002e-05, "loss": 0.5436, "step": 11880 }, { "epoch": 172.32, "learning_rate": 1.4984375e-05, "loss": 0.5874, "step": 11890 }, { "epoch": 172.46, "learning_rate": 1.490625e-05, "loss": 0.5529, "step": 11900 }, { "epoch": 172.61, "learning_rate": 1.4828125000000001e-05, "loss": 0.5403, "step": 11910 }, { "epoch": 172.75, "learning_rate": 1.475e-05, "loss": 0.4999, "step": 11920 }, { "epoch": 172.9, "learning_rate": 1.4671875000000001e-05, "loss": 0.6045, "step": 11930 }, { "epoch": 173.04, "learning_rate": 1.459375e-05, "loss": 0.5237, "step": 11940 }, { "epoch": 173.19, "learning_rate": 1.4515624999999999e-05, "loss": 0.5556, "step": 11950 }, { "epoch": 173.33, "learning_rate": 1.44375e-05, "loss": 0.5837, "step": 11960 }, { "epoch": 173.48, "learning_rate": 1.4359375e-05, "loss": 0.5021, "step": 11970 }, { "epoch": 173.62, "learning_rate": 1.4281250000000002e-05, "loss": 0.5294, "step": 11980 }, { "epoch": 173.77, "learning_rate": 1.4203125e-05, "loss": 0.5379, "step": 11990 }, { "epoch": 173.91, "learning_rate": 1.4125e-05, "loss": 0.5989, "step": 12000 }, { "epoch": 173.91, "eval_loss": 0.23597407341003418, "eval_runtime": 575.4715, "eval_samples_per_second": 5.898, "eval_steps_per_second": 0.739, "eval_wer": 0.1526843017856996, "step": 12000 }, { "epoch": 174.06, "learning_rate": 1.4046875000000001e-05, "loss": 0.5167, "step": 12010 }, { "epoch": 174.2, "learning_rate": 1.396875e-05, "loss": 0.5127, "step": 12020 }, { "epoch": 174.35, "learning_rate": 1.3890625000000002e-05, "loss": 0.613, "step": 12030 }, { "epoch": 174.49, "learning_rate": 1.38125e-05, "loss": 0.5287, "step": 12040 }, { "epoch": 174.64, "learning_rate": 1.3734375e-05, "loss": 0.5609, "step": 12050 }, { "epoch": 174.78, "learning_rate": 1.3656250000000002e-05, "loss": 0.5278, "step": 12060 }, { "epoch": 174.93, "learning_rate": 1.3578125e-05, "loss": 0.5928, "step": 12070 }, { "epoch": 175.07, "learning_rate": 1.3500000000000001e-05, "loss": 0.56, "step": 12080 }, { "epoch": 175.22, "learning_rate": 1.3421875000000001e-05, "loss": 0.5716, "step": 12090 }, { "epoch": 175.36, "learning_rate": 1.3343749999999999e-05, "loss": 0.5499, "step": 12100 }, { "epoch": 175.51, "learning_rate": 1.3265625e-05, "loss": 0.476, "step": 12110 }, { "epoch": 175.65, "learning_rate": 1.31875e-05, "loss": 0.5134, "step": 12120 }, { "epoch": 175.8, "learning_rate": 1.3109375000000002e-05, "loss": 0.5093, "step": 12130 }, { "epoch": 175.94, "learning_rate": 1.303125e-05, "loss": 0.579, "step": 12140 }, { "epoch": 176.09, "learning_rate": 1.2953125e-05, "loss": 0.4885, "step": 12150 }, { "epoch": 176.23, "learning_rate": 1.2875000000000001e-05, "loss": 0.5433, "step": 12160 }, { "epoch": 176.38, "learning_rate": 1.2796875e-05, "loss": 0.5137, "step": 12170 }, { "epoch": 176.52, "learning_rate": 1.271875e-05, "loss": 0.5202, "step": 12180 }, { "epoch": 176.67, "learning_rate": 1.2640625e-05, "loss": 0.5336, "step": 12190 }, { "epoch": 176.81, "learning_rate": 1.2562499999999999e-05, "loss": 0.5573, "step": 12200 }, { "epoch": 176.96, "learning_rate": 1.2484375e-05, "loss": 0.4605, "step": 12210 }, { "epoch": 177.1, "learning_rate": 1.2406250000000002e-05, "loss": 0.4848, "step": 12220 }, { "epoch": 177.25, "learning_rate": 1.2328125e-05, "loss": 0.5071, "step": 12230 }, { "epoch": 177.39, "learning_rate": 1.225e-05, "loss": 0.4797, "step": 12240 }, { "epoch": 177.54, "learning_rate": 1.2171875000000001e-05, "loss": 0.5308, "step": 12250 }, { "epoch": 177.68, "learning_rate": 1.2093750000000001e-05, "loss": 0.6085, "step": 12260 }, { "epoch": 177.83, "learning_rate": 1.2015625e-05, "loss": 0.5489, "step": 12270 }, { "epoch": 177.97, "learning_rate": 1.19375e-05, "loss": 0.4995, "step": 12280 }, { "epoch": 178.12, "learning_rate": 1.1859375e-05, "loss": 0.5686, "step": 12290 }, { "epoch": 178.26, "learning_rate": 1.178125e-05, "loss": 0.5471, "step": 12300 }, { "epoch": 178.41, "learning_rate": 1.1703125000000002e-05, "loss": 0.5439, "step": 12310 }, { "epoch": 178.55, "learning_rate": 1.1625000000000001e-05, "loss": 0.53, "step": 12320 }, { "epoch": 178.7, "learning_rate": 1.1546875e-05, "loss": 0.5401, "step": 12330 }, { "epoch": 178.84, "learning_rate": 1.1468750000000001e-05, "loss": 0.5493, "step": 12340 }, { "epoch": 178.99, "learning_rate": 1.1390625e-05, "loss": 0.5021, "step": 12350 }, { "epoch": 179.13, "learning_rate": 1.13125e-05, "loss": 0.5518, "step": 12360 }, { "epoch": 179.28, "learning_rate": 1.1234375e-05, "loss": 0.512, "step": 12370 }, { "epoch": 179.42, "learning_rate": 1.115625e-05, "loss": 0.503, "step": 12380 }, { "epoch": 179.57, "learning_rate": 1.1078125e-05, "loss": 0.5771, "step": 12390 }, { "epoch": 179.71, "learning_rate": 1.1000000000000001e-05, "loss": 0.5227, "step": 12400 }, { "epoch": 179.86, "learning_rate": 1.0921875000000001e-05, "loss": 0.5993, "step": 12410 }, { "epoch": 180.0, "learning_rate": 1.0843750000000001e-05, "loss": 0.4581, "step": 12420 }, { "epoch": 180.14, "learning_rate": 1.0765625e-05, "loss": 0.523, "step": 12430 }, { "epoch": 180.29, "learning_rate": 1.06875e-05, "loss": 0.5472, "step": 12440 }, { "epoch": 180.43, "learning_rate": 1.0609375e-05, "loss": 0.595, "step": 12450 }, { "epoch": 180.58, "learning_rate": 1.053125e-05, "loss": 0.5503, "step": 12460 }, { "epoch": 180.72, "learning_rate": 1.0453125000000002e-05, "loss": 0.5245, "step": 12470 }, { "epoch": 180.87, "learning_rate": 1.0375e-05, "loss": 0.5567, "step": 12480 }, { "epoch": 181.01, "learning_rate": 1.0296875e-05, "loss": 0.4907, "step": 12490 }, { "epoch": 181.16, "learning_rate": 1.0218750000000001e-05, "loss": 0.5697, "step": 12500 }, { "epoch": 181.16, "eval_loss": 0.23992973566055298, "eval_runtime": 578.3824, "eval_samples_per_second": 5.868, "eval_steps_per_second": 0.735, "eval_wer": 0.15259390741973392, "step": 12500 }, { "epoch": 181.3, "learning_rate": 1.0140625000000001e-05, "loss": 0.5403, "step": 12510 }, { "epoch": 181.45, "learning_rate": 1.00625e-05, "loss": 0.5109, "step": 12520 }, { "epoch": 181.59, "learning_rate": 9.984375e-06, "loss": 0.4829, "step": 12530 }, { "epoch": 181.74, "learning_rate": 9.90625e-06, "loss": 0.5025, "step": 12540 }, { "epoch": 181.88, "learning_rate": 9.828125e-06, "loss": 0.5442, "step": 12550 }, { "epoch": 182.03, "learning_rate": 9.750000000000002e-06, "loss": 0.5034, "step": 12560 }, { "epoch": 182.17, "learning_rate": 9.671875000000001e-06, "loss": 0.4793, "step": 12570 }, { "epoch": 182.32, "learning_rate": 9.59375e-06, "loss": 0.5074, "step": 12580 }, { "epoch": 182.46, "learning_rate": 9.515625000000001e-06, "loss": 0.5132, "step": 12590 }, { "epoch": 182.61, "learning_rate": 9.4375e-06, "loss": 0.5744, "step": 12600 }, { "epoch": 182.75, "learning_rate": 9.359375e-06, "loss": 0.4714, "step": 12610 }, { "epoch": 182.9, "learning_rate": 9.28125e-06, "loss": 0.566, "step": 12620 }, { "epoch": 183.04, "learning_rate": 9.203125e-06, "loss": 0.5804, "step": 12630 }, { "epoch": 183.19, "learning_rate": 9.125e-06, "loss": 0.4781, "step": 12640 }, { "epoch": 183.33, "learning_rate": 9.046875e-06, "loss": 0.4948, "step": 12650 }, { "epoch": 183.48, "learning_rate": 8.968750000000001e-06, "loss": 0.4981, "step": 12660 }, { "epoch": 183.62, "learning_rate": 8.890625000000001e-06, "loss": 0.5249, "step": 12670 }, { "epoch": 183.77, "learning_rate": 8.8125e-06, "loss": 0.5197, "step": 12680 }, { "epoch": 183.91, "learning_rate": 8.734375e-06, "loss": 0.5002, "step": 12690 }, { "epoch": 184.06, "learning_rate": 8.65625e-06, "loss": 0.5451, "step": 12700 }, { "epoch": 184.2, "learning_rate": 8.578125e-06, "loss": 0.5132, "step": 12710 }, { "epoch": 184.35, "learning_rate": 8.500000000000002e-06, "loss": 0.4806, "step": 12720 }, { "epoch": 184.49, "learning_rate": 8.421875e-06, "loss": 0.5207, "step": 12730 }, { "epoch": 184.64, "learning_rate": 8.34375e-06, "loss": 0.5618, "step": 12740 }, { "epoch": 184.78, "learning_rate": 8.265625000000001e-06, "loss": 0.5574, "step": 12750 }, { "epoch": 184.93, "learning_rate": 8.1875e-06, "loss": 0.5181, "step": 12760 }, { "epoch": 185.07, "learning_rate": 8.109375e-06, "loss": 0.5471, "step": 12770 }, { "epoch": 185.22, "learning_rate": 8.03125e-06, "loss": 0.4863, "step": 12780 }, { "epoch": 185.36, "learning_rate": 7.953125e-06, "loss": 0.547, "step": 12790 }, { "epoch": 185.51, "learning_rate": 7.875e-06, "loss": 0.5103, "step": 12800 }, { "epoch": 185.65, "learning_rate": 7.796875000000001e-06, "loss": 0.5406, "step": 12810 }, { "epoch": 185.8, "learning_rate": 7.718750000000001e-06, "loss": 0.5136, "step": 12820 }, { "epoch": 185.94, "learning_rate": 7.640625e-06, "loss": 0.4991, "step": 12830 }, { "epoch": 186.09, "learning_rate": 7.5625e-06, "loss": 0.5515, "step": 12840 }, { "epoch": 186.23, "learning_rate": 7.484375000000001e-06, "loss": 0.4893, "step": 12850 }, { "epoch": 186.38, "learning_rate": 7.4062500000000005e-06, "loss": 0.6101, "step": 12860 }, { "epoch": 186.52, "learning_rate": 7.328125000000001e-06, "loss": 0.5348, "step": 12870 }, { "epoch": 186.67, "learning_rate": 7.25e-06, "loss": 0.554, "step": 12880 }, { "epoch": 186.81, "learning_rate": 7.171875e-06, "loss": 0.5119, "step": 12890 }, { "epoch": 186.96, "learning_rate": 7.0937500000000005e-06, "loss": 0.5348, "step": 12900 }, { "epoch": 187.1, "learning_rate": 7.015625e-06, "loss": 0.5884, "step": 12910 }, { "epoch": 187.25, "learning_rate": 6.937500000000001e-06, "loss": 0.496, "step": 12920 }, { "epoch": 187.39, "learning_rate": 6.859375e-06, "loss": 0.5119, "step": 12930 }, { "epoch": 187.54, "learning_rate": 6.7812500000000005e-06, "loss": 0.5284, "step": 12940 }, { "epoch": 187.68, "learning_rate": 6.703125e-06, "loss": 0.5597, "step": 12950 }, { "epoch": 187.83, "learning_rate": 6.625000000000001e-06, "loss": 0.5816, "step": 12960 }, { "epoch": 187.97, "learning_rate": 6.546875000000001e-06, "loss": 0.5311, "step": 12970 }, { "epoch": 188.12, "learning_rate": 6.46875e-06, "loss": 0.5023, "step": 12980 }, { "epoch": 188.26, "learning_rate": 6.390625e-06, "loss": 0.5199, "step": 12990 }, { "epoch": 188.41, "learning_rate": 6.3125e-06, "loss": 0.5379, "step": 13000 }, { "epoch": 188.41, "eval_loss": 0.2375340759754181, "eval_runtime": 574.5803, "eval_samples_per_second": 5.907, "eval_steps_per_second": 0.74, "eval_wer": 0.15229807131293707, "step": 13000 }, { "epoch": 188.55, "learning_rate": 6.234375000000001e-06, "loss": 0.5227, "step": 13010 }, { "epoch": 188.7, "learning_rate": 6.1562500000000006e-06, "loss": 0.4946, "step": 13020 }, { "epoch": 188.84, "learning_rate": 6.078125e-06, "loss": 0.5618, "step": 13030 }, { "epoch": 188.99, "learning_rate": 6e-06, "loss": 0.5023, "step": 13040 }, { "epoch": 189.13, "learning_rate": 5.921875e-06, "loss": 0.4884, "step": 13050 }, { "epoch": 189.28, "learning_rate": 5.843750000000001e-06, "loss": 0.4912, "step": 13060 }, { "epoch": 189.42, "learning_rate": 5.765625e-06, "loss": 0.5511, "step": 13070 }, { "epoch": 189.57, "learning_rate": 5.6875e-06, "loss": 0.542, "step": 13080 }, { "epoch": 189.71, "learning_rate": 5.609375e-06, "loss": 0.5275, "step": 13090 }, { "epoch": 189.86, "learning_rate": 5.531250000000001e-06, "loss": 0.5719, "step": 13100 }, { "epoch": 190.0, "learning_rate": 5.453125e-06, "loss": 0.5686, "step": 13110 }, { "epoch": 190.14, "learning_rate": 5.375e-06, "loss": 0.5425, "step": 13120 }, { "epoch": 190.29, "learning_rate": 5.296875000000001e-06, "loss": 0.5322, "step": 13130 }, { "epoch": 190.43, "learning_rate": 5.21875e-06, "loss": 0.5146, "step": 13140 }, { "epoch": 190.58, "learning_rate": 5.1406250000000004e-06, "loss": 0.5364, "step": 13150 }, { "epoch": 190.72, "learning_rate": 5.0625e-06, "loss": 0.496, "step": 13160 }, { "epoch": 190.87, "learning_rate": 4.984375e-06, "loss": 0.6167, "step": 13170 }, { "epoch": 191.01, "learning_rate": 4.906250000000001e-06, "loss": 0.4898, "step": 13180 }, { "epoch": 191.16, "learning_rate": 4.828125e-06, "loss": 0.5216, "step": 13190 }, { "epoch": 191.3, "learning_rate": 4.75e-06, "loss": 0.5122, "step": 13200 }, { "epoch": 191.45, "learning_rate": 4.671875000000001e-06, "loss": 0.5293, "step": 13210 }, { "epoch": 191.59, "learning_rate": 4.59375e-06, "loss": 0.5072, "step": 13220 }, { "epoch": 191.74, "learning_rate": 4.5156250000000005e-06, "loss": 0.4934, "step": 13230 }, { "epoch": 191.88, "learning_rate": 4.4375e-06, "loss": 0.5224, "step": 13240 }, { "epoch": 192.03, "learning_rate": 4.359375e-06, "loss": 0.4804, "step": 13250 }, { "epoch": 192.17, "learning_rate": 4.281250000000001e-06, "loss": 0.6039, "step": 13260 }, { "epoch": 192.32, "learning_rate": 4.2031250000000005e-06, "loss": 0.4833, "step": 13270 }, { "epoch": 192.46, "learning_rate": 4.125e-06, "loss": 0.5236, "step": 13280 }, { "epoch": 192.61, "learning_rate": 4.046875e-06, "loss": 0.5312, "step": 13290 }, { "epoch": 192.75, "learning_rate": 3.96875e-06, "loss": 0.5886, "step": 13300 }, { "epoch": 192.9, "learning_rate": 3.8906250000000005e-06, "loss": 0.5098, "step": 13310 }, { "epoch": 193.04, "learning_rate": 3.8125e-06, "loss": 0.5223, "step": 13320 }, { "epoch": 193.19, "learning_rate": 3.7343750000000006e-06, "loss": 0.581, "step": 13330 }, { "epoch": 193.33, "learning_rate": 3.65625e-06, "loss": 0.4837, "step": 13340 }, { "epoch": 193.48, "learning_rate": 3.578125e-06, "loss": 0.5184, "step": 13350 }, { "epoch": 193.62, "learning_rate": 3.5000000000000004e-06, "loss": 0.5145, "step": 13360 }, { "epoch": 193.77, "learning_rate": 3.421875e-06, "loss": 0.4956, "step": 13370 }, { "epoch": 193.91, "learning_rate": 3.3437500000000004e-06, "loss": 0.5211, "step": 13380 }, { "epoch": 194.06, "learning_rate": 3.2656249999999998e-06, "loss": 0.5218, "step": 13390 }, { "epoch": 194.2, "learning_rate": 3.1875000000000004e-06, "loss": 0.5457, "step": 13400 }, { "epoch": 194.35, "learning_rate": 3.109375e-06, "loss": 0.4737, "step": 13410 }, { "epoch": 194.49, "learning_rate": 3.03125e-06, "loss": 0.5828, "step": 13420 }, { "epoch": 194.64, "learning_rate": 2.9531249999999998e-06, "loss": 0.5597, "step": 13430 }, { "epoch": 194.78, "learning_rate": 2.8750000000000004e-06, "loss": 0.5, "step": 13440 }, { "epoch": 194.93, "learning_rate": 2.7968750000000002e-06, "loss": 0.5694, "step": 13450 }, { "epoch": 195.07, "learning_rate": 2.71875e-06, "loss": 0.5176, "step": 13460 }, { "epoch": 195.22, "learning_rate": 2.6406250000000002e-06, "loss": 0.5216, "step": 13470 }, { "epoch": 195.36, "learning_rate": 2.5625e-06, "loss": 0.489, "step": 13480 }, { "epoch": 195.51, "learning_rate": 2.4843750000000002e-06, "loss": 0.5186, "step": 13490 }, { "epoch": 195.65, "learning_rate": 2.40625e-06, "loss": 0.5022, "step": 13500 }, { "epoch": 195.65, "eval_loss": 0.23950409889221191, "eval_runtime": 592.332, "eval_samples_per_second": 5.73, "eval_steps_per_second": 0.718, "eval_wer": 0.15194471151870753, "step": 13500 }, { "epoch": 195.8, "learning_rate": 2.3281250000000003e-06, "loss": 0.5106, "step": 13510 }, { "epoch": 195.94, "learning_rate": 2.25e-06, "loss": 0.4966, "step": 13520 }, { "epoch": 196.09, "learning_rate": 2.171875e-06, "loss": 0.5053, "step": 13530 }, { "epoch": 196.23, "learning_rate": 2.09375e-06, "loss": 0.5594, "step": 13540 }, { "epoch": 196.38, "learning_rate": 2.0156250000000003e-06, "loss": 0.5047, "step": 13550 }, { "epoch": 196.52, "learning_rate": 1.9375e-06, "loss": 0.5078, "step": 13560 }, { "epoch": 196.67, "learning_rate": 1.859375e-06, "loss": 0.5466, "step": 13570 }, { "epoch": 196.81, "learning_rate": 1.7812499999999999e-06, "loss": 0.5371, "step": 13580 }, { "epoch": 196.96, "learning_rate": 1.703125e-06, "loss": 0.5551, "step": 13590 }, { "epoch": 197.1, "learning_rate": 1.6250000000000001e-06, "loss": 0.5615, "step": 13600 }, { "epoch": 197.25, "learning_rate": 1.5468750000000001e-06, "loss": 0.5464, "step": 13610 }, { "epoch": 197.39, "learning_rate": 1.46875e-06, "loss": 0.5196, "step": 13620 }, { "epoch": 197.54, "learning_rate": 1.3906250000000001e-06, "loss": 0.4493, "step": 13630 }, { "epoch": 197.68, "learning_rate": 1.3125e-06, "loss": 0.5603, "step": 13640 }, { "epoch": 197.83, "learning_rate": 1.2343750000000001e-06, "loss": 0.5227, "step": 13650 }, { "epoch": 197.97, "learning_rate": 1.15625e-06, "loss": 0.4889, "step": 13660 }, { "epoch": 198.12, "learning_rate": 1.078125e-06, "loss": 0.5893, "step": 13670 }, { "epoch": 198.26, "learning_rate": 1.0000000000000002e-06, "loss": 0.5509, "step": 13680 }, { "epoch": 198.41, "learning_rate": 9.21875e-07, "loss": 0.513, "step": 13690 }, { "epoch": 198.55, "learning_rate": 8.437500000000001e-07, "loss": 0.5744, "step": 13700 }, { "epoch": 198.7, "learning_rate": 7.65625e-07, "loss": 0.5457, "step": 13710 }, { "epoch": 198.84, "learning_rate": 6.875000000000001e-07, "loss": 0.5007, "step": 13720 }, { "epoch": 198.99, "learning_rate": 6.093750000000001e-07, "loss": 0.5829, "step": 13730 }, { "epoch": 199.13, "learning_rate": 5.312500000000001e-07, "loss": 0.5047, "step": 13740 }, { "epoch": 199.28, "learning_rate": 4.53125e-07, "loss": 0.5173, "step": 13750 }, { "epoch": 199.42, "learning_rate": 3.75e-07, "loss": 0.533, "step": 13760 }, { "epoch": 199.57, "learning_rate": 2.96875e-07, "loss": 0.5328, "step": 13770 }, { "epoch": 199.71, "learning_rate": 2.1875000000000002e-07, "loss": 0.5111, "step": 13780 }, { "epoch": 199.86, "learning_rate": 1.40625e-07, "loss": 0.5071, "step": 13790 }, { "epoch": 200.0, "learning_rate": 6.250000000000001e-08, "loss": 0.5151, "step": 13800 }, { "epoch": 200.0, "step": 13800, "total_flos": 2.0051176394243018e+20, "train_loss": 0.32463424516760786, "train_runtime": 60486.5281, "train_samples_per_second": 7.255, "train_steps_per_second": 0.228 } ], "max_steps": 13800, "num_train_epochs": 200, "total_flos": 2.0051176394243018e+20, "trial_name": null, "trial_params": null }