|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 100.0, |
|
"global_step": 15600, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.3949999999999997e-06, |
|
"loss": 18.8158, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 6.895e-06, |
|
"loss": 9.3721, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.0394999999999998e-05, |
|
"loss": 5.1453, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 1.3895e-05, |
|
"loss": 4.2355, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 1.7395e-05, |
|
"loss": 3.7107, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 2.0894999999999996e-05, |
|
"loss": 3.3647, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"learning_rate": 2.4394999999999996e-05, |
|
"loss": 3.2191, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 5.13, |
|
"learning_rate": 2.7895e-05, |
|
"loss": 3.1339, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 5.77, |
|
"learning_rate": 3.1395e-05, |
|
"loss": 3.0484, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 6.41, |
|
"learning_rate": 3.4895e-05, |
|
"loss": 3.0086, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 7.05, |
|
"learning_rate": 3.8394999999999994e-05, |
|
"loss": 2.9794, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 7.69, |
|
"learning_rate": 4.1895e-05, |
|
"loss": 2.9544, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 8.33, |
|
"learning_rate": 4.5394999999999995e-05, |
|
"loss": 2.901, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 8.97, |
|
"learning_rate": 4.8895e-05, |
|
"loss": 2.6256, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 9.62, |
|
"learning_rate": 5.2395e-05, |
|
"loss": 2.0673, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 10.26, |
|
"learning_rate": 5.589499999999999e-05, |
|
"loss": 1.763, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 10.9, |
|
"learning_rate": 5.9394999999999996e-05, |
|
"loss": 1.6053, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 11.54, |
|
"learning_rate": 6.289499999999999e-05, |
|
"loss": 1.5242, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 12.18, |
|
"learning_rate": 6.639499999999999e-05, |
|
"loss": 1.4635, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 12.82, |
|
"learning_rate": 6.9895e-05, |
|
"loss": 1.4235, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 12.82, |
|
"eval_loss": 0.4475284516811371, |
|
"eval_runtime": 99.8092, |
|
"eval_samples_per_second": 20.88, |
|
"eval_steps_per_second": 20.88, |
|
"eval_wer": 0.455098317453882, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 13.46, |
|
"learning_rate": 6.950073529411765e-05, |
|
"loss": 1.3615, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 14.1, |
|
"learning_rate": 6.89860294117647e-05, |
|
"loss": 1.2973, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 14.74, |
|
"learning_rate": 6.847132352941176e-05, |
|
"loss": 1.261, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 15.38, |
|
"learning_rate": 6.795661764705882e-05, |
|
"loss": 1.2356, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 16.03, |
|
"learning_rate": 6.744191176470588e-05, |
|
"loss": 1.1895, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 16.67, |
|
"learning_rate": 6.692720588235293e-05, |
|
"loss": 1.144, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 17.31, |
|
"learning_rate": 6.64125e-05, |
|
"loss": 1.1081, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 17.95, |
|
"learning_rate": 6.589779411764706e-05, |
|
"loss": 1.1043, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 18.59, |
|
"learning_rate": 6.538308823529412e-05, |
|
"loss": 1.074, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 19.23, |
|
"learning_rate": 6.486838235294117e-05, |
|
"loss": 1.0529, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 19.87, |
|
"learning_rate": 6.435367647058823e-05, |
|
"loss": 1.0452, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 20.51, |
|
"learning_rate": 6.383897058823529e-05, |
|
"loss": 1.0405, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 21.15, |
|
"learning_rate": 6.332426470588235e-05, |
|
"loss": 1.0131, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 21.79, |
|
"learning_rate": 6.28095588235294e-05, |
|
"loss": 1.0007, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 22.44, |
|
"learning_rate": 6.229485294117647e-05, |
|
"loss": 0.9744, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 23.08, |
|
"learning_rate": 6.178014705882353e-05, |
|
"loss": 0.9773, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 23.72, |
|
"learning_rate": 6.126544117647059e-05, |
|
"loss": 0.9767, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 24.36, |
|
"learning_rate": 6.0750735294117646e-05, |
|
"loss": 0.9495, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"learning_rate": 6.024117647058823e-05, |
|
"loss": 0.9667, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 25.64, |
|
"learning_rate": 5.9726470588235294e-05, |
|
"loss": 0.9383, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 25.64, |
|
"eval_loss": 0.223493292927742, |
|
"eval_runtime": 99.8226, |
|
"eval_samples_per_second": 20.877, |
|
"eval_steps_per_second": 20.877, |
|
"eval_wer": 0.23281978512061627, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 26.28, |
|
"learning_rate": 5.921176470588234e-05, |
|
"loss": 0.9534, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 26.92, |
|
"learning_rate": 5.8697058823529405e-05, |
|
"loss": 0.9294, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 27.56, |
|
"learning_rate": 5.818235294117647e-05, |
|
"loss": 0.9348, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 28.21, |
|
"learning_rate": 5.766764705882352e-05, |
|
"loss": 0.9209, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 28.85, |
|
"learning_rate": 5.7152941176470585e-05, |
|
"loss": 0.9104, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 29.49, |
|
"learning_rate": 5.663823529411764e-05, |
|
"loss": 0.9009, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 30.13, |
|
"learning_rate": 5.61235294117647e-05, |
|
"loss": 0.9109, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 30.77, |
|
"learning_rate": 5.560882352941176e-05, |
|
"loss": 0.8973, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 31.41, |
|
"learning_rate": 5.509411764705882e-05, |
|
"loss": 0.864, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 32.05, |
|
"learning_rate": 5.4579411764705876e-05, |
|
"loss": 0.8654, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 32.69, |
|
"learning_rate": 5.406470588235294e-05, |
|
"loss": 0.8523, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 33.33, |
|
"learning_rate": 5.3549999999999994e-05, |
|
"loss": 0.8794, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 33.97, |
|
"learning_rate": 5.3035294117647056e-05, |
|
"loss": 0.8577, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 34.62, |
|
"learning_rate": 5.252058823529411e-05, |
|
"loss": 0.8585, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 35.26, |
|
"learning_rate": 5.2005882352941174e-05, |
|
"loss": 0.8473, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 35.9, |
|
"learning_rate": 5.149117647058823e-05, |
|
"loss": 0.8465, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 36.54, |
|
"learning_rate": 5.097647058823529e-05, |
|
"loss": 0.838, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 37.18, |
|
"learning_rate": 5.046176470588235e-05, |
|
"loss": 0.8402, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 37.82, |
|
"learning_rate": 4.99470588235294e-05, |
|
"loss": 0.8432, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 38.46, |
|
"learning_rate": 4.9432352941176465e-05, |
|
"loss": 0.8359, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 38.46, |
|
"eval_loss": 0.20040307939052582, |
|
"eval_runtime": 101.1606, |
|
"eval_samples_per_second": 20.601, |
|
"eval_steps_per_second": 20.601, |
|
"eval_wer": 0.20981147374822623, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 39.1, |
|
"learning_rate": 4.891764705882353e-05, |
|
"loss": 0.8456, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 39.74, |
|
"learning_rate": 4.840294117647059e-05, |
|
"loss": 0.8285, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 40.38, |
|
"learning_rate": 4.788823529411764e-05, |
|
"loss": 0.8054, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 41.03, |
|
"learning_rate": 4.73735294117647e-05, |
|
"loss": 0.82, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 41.67, |
|
"learning_rate": 4.685882352941176e-05, |
|
"loss": 0.8082, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 42.31, |
|
"learning_rate": 4.6344117647058825e-05, |
|
"loss": 0.8049, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 42.95, |
|
"learning_rate": 4.5829411764705873e-05, |
|
"loss": 0.8088, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 43.59, |
|
"learning_rate": 4.5314705882352936e-05, |
|
"loss": 0.8092, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 44.23, |
|
"learning_rate": 4.48e-05, |
|
"loss": 0.8212, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 44.87, |
|
"learning_rate": 4.428529411764706e-05, |
|
"loss": 0.7847, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 45.51, |
|
"learning_rate": 4.3775735294117646e-05, |
|
"loss": 0.7846, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 46.15, |
|
"learning_rate": 4.32610294117647e-05, |
|
"loss": 0.7966, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 46.79, |
|
"learning_rate": 4.2746323529411764e-05, |
|
"loss": 0.7811, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 47.44, |
|
"learning_rate": 4.223161764705882e-05, |
|
"loss": 0.7706, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 48.08, |
|
"learning_rate": 4.1716911764705875e-05, |
|
"loss": 0.7958, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 48.72, |
|
"learning_rate": 4.120220588235294e-05, |
|
"loss": 0.7617, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 49.36, |
|
"learning_rate": 4.06875e-05, |
|
"loss": 0.7662, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"learning_rate": 4.0172794117647055e-05, |
|
"loss": 0.7545, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 50.64, |
|
"learning_rate": 3.965808823529411e-05, |
|
"loss": 0.7482, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 51.28, |
|
"learning_rate": 3.914338235294117e-05, |
|
"loss": 0.7633, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 51.28, |
|
"eval_loss": 0.19603605568408966, |
|
"eval_runtime": 98.7836, |
|
"eval_samples_per_second": 21.097, |
|
"eval_steps_per_second": 21.097, |
|
"eval_wer": 0.1882221771741334, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 51.92, |
|
"learning_rate": 3.8628676470588235e-05, |
|
"loss": 0.7522, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 52.56, |
|
"learning_rate": 3.811397058823529e-05, |
|
"loss": 0.7411, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 53.21, |
|
"learning_rate": 3.7599264705882346e-05, |
|
"loss": 0.7443, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 53.85, |
|
"learning_rate": 3.708455882352941e-05, |
|
"loss": 0.7583, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 54.49, |
|
"learning_rate": 3.656985294117647e-05, |
|
"loss": 0.7254, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 55.13, |
|
"learning_rate": 3.605514705882352e-05, |
|
"loss": 0.7329, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 55.77, |
|
"learning_rate": 3.554044117647058e-05, |
|
"loss": 0.7216, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 56.41, |
|
"learning_rate": 3.5025735294117644e-05, |
|
"loss": 0.7322, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 57.05, |
|
"learning_rate": 3.4511029411764706e-05, |
|
"loss": 0.7405, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 57.69, |
|
"learning_rate": 3.399632352941176e-05, |
|
"loss": 0.7194, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 58.33, |
|
"learning_rate": 3.3481617647058824e-05, |
|
"loss": 0.7146, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 58.97, |
|
"learning_rate": 3.297205882352941e-05, |
|
"loss": 0.7128, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 59.62, |
|
"learning_rate": 3.2457352941176466e-05, |
|
"loss": 0.7186, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 60.26, |
|
"learning_rate": 3.194264705882353e-05, |
|
"loss": 0.7038, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 60.9, |
|
"learning_rate": 3.142794117647058e-05, |
|
"loss": 0.696, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 61.54, |
|
"learning_rate": 3.0913235294117646e-05, |
|
"loss": 0.707, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 62.18, |
|
"learning_rate": 3.0398529411764704e-05, |
|
"loss": 0.7113, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 62.82, |
|
"learning_rate": 2.9883823529411763e-05, |
|
"loss": 0.6997, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 63.46, |
|
"learning_rate": 2.9369117647058822e-05, |
|
"loss": 0.6815, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 64.1, |
|
"learning_rate": 2.885441176470588e-05, |
|
"loss": 0.7001, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 64.1, |
|
"eval_loss": 0.19019167125225067, |
|
"eval_runtime": 99.1402, |
|
"eval_samples_per_second": 21.021, |
|
"eval_steps_per_second": 21.021, |
|
"eval_wer": 0.18092438678289074, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 64.74, |
|
"learning_rate": 2.833970588235294e-05, |
|
"loss": 0.7032, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 65.38, |
|
"learning_rate": 2.7825e-05, |
|
"loss": 0.7102, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 66.03, |
|
"learning_rate": 2.7310294117647054e-05, |
|
"loss": 0.6916, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 66.67, |
|
"learning_rate": 2.6795588235294116e-05, |
|
"loss": 0.6907, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 67.31, |
|
"learning_rate": 2.6280882352941172e-05, |
|
"loss": 0.6711, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 67.95, |
|
"learning_rate": 2.5766176470588234e-05, |
|
"loss": 0.6843, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 68.59, |
|
"learning_rate": 2.525147058823529e-05, |
|
"loss": 0.6847, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 69.23, |
|
"learning_rate": 2.4736764705882352e-05, |
|
"loss": 0.6533, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 69.87, |
|
"learning_rate": 2.4222058823529407e-05, |
|
"loss": 0.6883, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 70.51, |
|
"learning_rate": 2.3712499999999997e-05, |
|
"loss": 0.6779, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 71.15, |
|
"learning_rate": 2.319779411764706e-05, |
|
"loss": 0.6739, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 71.79, |
|
"learning_rate": 2.2683088235294115e-05, |
|
"loss": 0.6662, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 72.44, |
|
"learning_rate": 2.2168382352941177e-05, |
|
"loss": 0.6539, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 73.08, |
|
"learning_rate": 2.1653676470588232e-05, |
|
"loss": 0.6589, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 73.72, |
|
"learning_rate": 2.1138970588235295e-05, |
|
"loss": 0.6545, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 74.36, |
|
"learning_rate": 2.062426470588235e-05, |
|
"loss": 0.6585, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 75.0, |
|
"learning_rate": 2.0109558823529412e-05, |
|
"loss": 0.6472, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 75.64, |
|
"learning_rate": 1.9594852941176468e-05, |
|
"loss": 0.6657, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 76.28, |
|
"learning_rate": 1.9080147058823527e-05, |
|
"loss": 0.65, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 76.92, |
|
"learning_rate": 1.8565441176470586e-05, |
|
"loss": 0.652, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 76.92, |
|
"eval_loss": 0.197893887758255, |
|
"eval_runtime": 98.6246, |
|
"eval_samples_per_second": 21.131, |
|
"eval_steps_per_second": 21.131, |
|
"eval_wer": 0.17747820798702615, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 77.56, |
|
"learning_rate": 1.8050735294117645e-05, |
|
"loss": 0.6559, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 78.21, |
|
"learning_rate": 1.7536029411764703e-05, |
|
"loss": 0.6282, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 78.85, |
|
"learning_rate": 1.7021323529411762e-05, |
|
"loss": 0.6251, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 79.49, |
|
"learning_rate": 1.650661764705882e-05, |
|
"loss": 0.6147, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 80.13, |
|
"learning_rate": 1.599191176470588e-05, |
|
"loss": 0.6412, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 80.77, |
|
"learning_rate": 1.547720588235294e-05, |
|
"loss": 0.6309, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 81.41, |
|
"learning_rate": 1.4962499999999998e-05, |
|
"loss": 0.6197, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 82.05, |
|
"learning_rate": 1.4447794117647057e-05, |
|
"loss": 0.6238, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 82.69, |
|
"learning_rate": 1.3933088235294116e-05, |
|
"loss": 0.635, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 83.33, |
|
"learning_rate": 1.3418382352941174e-05, |
|
"loss": 0.6294, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 83.97, |
|
"learning_rate": 1.2903676470588233e-05, |
|
"loss": 0.6171, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 84.62, |
|
"learning_rate": 1.2388970588235292e-05, |
|
"loss": 0.6255, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 85.26, |
|
"learning_rate": 1.1874264705882353e-05, |
|
"loss": 0.6238, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 85.9, |
|
"learning_rate": 1.1359558823529412e-05, |
|
"loss": 0.5923, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 86.54, |
|
"learning_rate": 1.084485294117647e-05, |
|
"loss": 0.6066, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 87.18, |
|
"learning_rate": 1.033014705882353e-05, |
|
"loss": 0.6143, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 87.82, |
|
"learning_rate": 9.815441176470588e-06, |
|
"loss": 0.6146, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 88.46, |
|
"learning_rate": 9.300735294117647e-06, |
|
"loss": 0.6102, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 89.1, |
|
"learning_rate": 8.786029411764706e-06, |
|
"loss": 0.6112, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 89.74, |
|
"learning_rate": 8.271323529411765e-06, |
|
"loss": 0.6025, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 89.74, |
|
"eval_loss": 0.18655425310134888, |
|
"eval_runtime": 99.4533, |
|
"eval_samples_per_second": 20.955, |
|
"eval_steps_per_second": 20.955, |
|
"eval_wer": 0.16957226839651326, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 90.38, |
|
"learning_rate": 7.756617647058824e-06, |
|
"loss": 0.6147, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 91.03, |
|
"learning_rate": 7.241911764705882e-06, |
|
"loss": 0.5954, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 91.67, |
|
"learning_rate": 6.7272058823529405e-06, |
|
"loss": 0.5998, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 92.31, |
|
"learning_rate": 6.212499999999999e-06, |
|
"loss": 0.6078, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 92.95, |
|
"learning_rate": 5.697794117647058e-06, |
|
"loss": 0.6068, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 93.59, |
|
"learning_rate": 5.183088235294118e-06, |
|
"loss": 0.6015, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 94.23, |
|
"learning_rate": 4.668382352941176e-06, |
|
"loss": 0.6013, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 94.87, |
|
"learning_rate": 4.153676470588235e-06, |
|
"loss": 0.5948, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 95.51, |
|
"learning_rate": 3.6389705882352937e-06, |
|
"loss": 0.5909, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 96.15, |
|
"learning_rate": 3.1242647058823526e-06, |
|
"loss": 0.6129, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 96.79, |
|
"learning_rate": 2.614705882352941e-06, |
|
"loss": 0.5795, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 97.44, |
|
"learning_rate": 2.1e-06, |
|
"loss": 0.5967, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 98.08, |
|
"learning_rate": 1.5852941176470587e-06, |
|
"loss": 0.6013, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 98.72, |
|
"learning_rate": 1.0705882352941175e-06, |
|
"loss": 0.5981, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 99.36, |
|
"learning_rate": 5.558823529411764e-07, |
|
"loss": 0.6133, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 100.0, |
|
"learning_rate": 4.117647058823529e-08, |
|
"loss": 0.5932, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 100.0, |
|
"step": 15600, |
|
"total_flos": 5.043365508466233e+19, |
|
"train_loss": 1.1756210368718856, |
|
"train_runtime": 18507.9964, |
|
"train_samples_per_second": 26.815, |
|
"train_steps_per_second": 0.843 |
|
} |
|
], |
|
"max_steps": 15600, |
|
"num_train_epochs": 100, |
|
"total_flos": 5.043365508466233e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|