|
{ |
|
"best_metric": 20.446244243290458, |
|
"best_model_checkpoint": "./whisper-medium-lt/checkpoint-2000", |
|
"epoch": 18.685979142526072, |
|
"global_step": 2000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.6000000000000004e-07, |
|
"loss": 3.6989, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.600000000000001e-07, |
|
"loss": 2.5535, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.46e-06, |
|
"loss": 1.7181, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.9600000000000003e-06, |
|
"loss": 1.2891, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 2.46e-06, |
|
"loss": 0.7856, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 2.96e-06, |
|
"loss": 0.3772, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 3.46e-06, |
|
"loss": 0.3373, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 3.96e-06, |
|
"loss": 0.3153, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 4.4600000000000005e-06, |
|
"loss": 0.2667, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 4.960000000000001e-06, |
|
"loss": 0.2029, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 5.460000000000001e-06, |
|
"loss": 0.1957, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 5.9600000000000005e-06, |
|
"loss": 0.1887, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 6.460000000000001e-06, |
|
"loss": 0.1782, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 6.96e-06, |
|
"loss": 0.1022, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 7.4600000000000006e-06, |
|
"loss": 0.099, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 7.960000000000002e-06, |
|
"loss": 0.1057, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 8.46e-06, |
|
"loss": 0.1052, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"learning_rate": 8.96e-06, |
|
"loss": 0.0607, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 9.460000000000001e-06, |
|
"loss": 0.0531, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"learning_rate": 9.960000000000001e-06, |
|
"loss": 0.0532, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"learning_rate": 9.94888888888889e-06, |
|
"loss": 0.0565, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 9.893333333333334e-06, |
|
"loss": 0.0432, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 5.37, |
|
"learning_rate": 9.837777777777778e-06, |
|
"loss": 0.0302, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 5.6, |
|
"learning_rate": 9.782222222222222e-06, |
|
"loss": 0.0306, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 5.83, |
|
"learning_rate": 9.726666666666668e-06, |
|
"loss": 0.0305, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"learning_rate": 9.671111111111112e-06, |
|
"loss": 0.0267, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 6.31, |
|
"learning_rate": 9.615555555555558e-06, |
|
"loss": 0.0163, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 6.54, |
|
"learning_rate": 9.56e-06, |
|
"loss": 0.0189, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"learning_rate": 9.504444444444446e-06, |
|
"loss": 0.0194, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 9.44888888888889e-06, |
|
"loss": 0.0209, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 7.24, |
|
"learning_rate": 9.393333333333334e-06, |
|
"loss": 0.0119, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 7.47, |
|
"learning_rate": 9.33777777777778e-06, |
|
"loss": 0.0124, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 7.7, |
|
"learning_rate": 9.282222222222222e-06, |
|
"loss": 0.0129, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 7.94, |
|
"learning_rate": 9.226666666666668e-06, |
|
"loss": 0.0118, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 8.18, |
|
"learning_rate": 9.171111111111112e-06, |
|
"loss": 0.009, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 8.41, |
|
"learning_rate": 9.115555555555556e-06, |
|
"loss": 0.0069, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 8.64, |
|
"learning_rate": 9.060000000000001e-06, |
|
"loss": 0.0076, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 8.87, |
|
"learning_rate": 9.004444444444445e-06, |
|
"loss": 0.0076, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 9.11, |
|
"learning_rate": 8.94888888888889e-06, |
|
"loss": 0.0071, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 9.34, |
|
"learning_rate": 8.893333333333333e-06, |
|
"loss": 0.0056, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 9.34, |
|
"eval_loss": 0.3252415060997009, |
|
"eval_runtime": 2671.1564, |
|
"eval_samples_per_second": 1.404, |
|
"eval_steps_per_second": 0.176, |
|
"eval_wer": 20.553438145148483, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"learning_rate": 8.83777777777778e-06, |
|
"loss": 0.0058, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 9.81, |
|
"learning_rate": 8.782222222222223e-06, |
|
"loss": 0.0058, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 10.05, |
|
"learning_rate": 8.726666666666667e-06, |
|
"loss": 0.0057, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 10.28, |
|
"learning_rate": 8.671111111111113e-06, |
|
"loss": 0.0054, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 10.51, |
|
"learning_rate": 8.615555555555555e-06, |
|
"loss": 0.0042, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 10.74, |
|
"learning_rate": 8.560000000000001e-06, |
|
"loss": 0.0035, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 10.97, |
|
"learning_rate": 8.504444444444445e-06, |
|
"loss": 0.0048, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 11.21, |
|
"learning_rate": 8.448888888888889e-06, |
|
"loss": 0.0033, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 11.44, |
|
"learning_rate": 8.393333333333335e-06, |
|
"loss": 0.0033, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 11.68, |
|
"learning_rate": 8.337777777777777e-06, |
|
"loss": 0.003, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 11.91, |
|
"learning_rate": 8.282222222222223e-06, |
|
"loss": 0.0033, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 12.15, |
|
"learning_rate": 8.226666666666667e-06, |
|
"loss": 0.0031, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 12.38, |
|
"learning_rate": 8.171111111111113e-06, |
|
"loss": 0.0027, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 12.61, |
|
"learning_rate": 8.115555555555557e-06, |
|
"loss": 0.0026, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 12.84, |
|
"learning_rate": 8.06e-06, |
|
"loss": 0.0026, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 13.08, |
|
"learning_rate": 8.004444444444445e-06, |
|
"loss": 0.0021, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 13.32, |
|
"learning_rate": 7.948888888888889e-06, |
|
"loss": 0.0014, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 13.55, |
|
"learning_rate": 7.893333333333335e-06, |
|
"loss": 0.0013, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 13.78, |
|
"learning_rate": 7.837777777777779e-06, |
|
"loss": 0.0018, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 14.02, |
|
"learning_rate": 7.782222222222223e-06, |
|
"loss": 0.0019, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 14.25, |
|
"learning_rate": 7.726666666666667e-06, |
|
"loss": 0.0014, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 14.48, |
|
"learning_rate": 7.67111111111111e-06, |
|
"loss": 0.0013, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 14.71, |
|
"learning_rate": 7.6155555555555564e-06, |
|
"loss": 0.0012, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 14.95, |
|
"learning_rate": 7.5600000000000005e-06, |
|
"loss": 0.0017, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 15.19, |
|
"learning_rate": 7.504444444444445e-06, |
|
"loss": 0.0014, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 15.42, |
|
"learning_rate": 7.44888888888889e-06, |
|
"loss": 0.0014, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 15.65, |
|
"learning_rate": 7.393333333333333e-06, |
|
"loss": 0.0014, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 15.88, |
|
"learning_rate": 7.337777777777778e-06, |
|
"loss": 0.0012, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 16.12, |
|
"learning_rate": 7.282222222222222e-06, |
|
"loss": 0.0012, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 16.35, |
|
"learning_rate": 7.226666666666667e-06, |
|
"loss": 0.001, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 16.58, |
|
"learning_rate": 7.171111111111112e-06, |
|
"loss": 0.0011, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 16.82, |
|
"learning_rate": 7.115555555555557e-06, |
|
"loss": 0.0012, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 17.06, |
|
"learning_rate": 7.06e-06, |
|
"loss": 0.0011, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 17.29, |
|
"learning_rate": 7.004444444444445e-06, |
|
"loss": 0.001, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 17.52, |
|
"learning_rate": 6.948888888888889e-06, |
|
"loss": 0.001, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 17.75, |
|
"learning_rate": 6.893333333333334e-06, |
|
"loss": 0.0014, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 17.98, |
|
"learning_rate": 6.837777777777779e-06, |
|
"loss": 0.0015, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 18.22, |
|
"learning_rate": 6.782222222222222e-06, |
|
"loss": 0.0016, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 18.45, |
|
"learning_rate": 6.726666666666667e-06, |
|
"loss": 0.0018, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 18.69, |
|
"learning_rate": 6.671111111111112e-06, |
|
"loss": 0.0023, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 18.69, |
|
"eval_loss": 0.3549511432647705, |
|
"eval_runtime": 2667.5472, |
|
"eval_samples_per_second": 1.405, |
|
"eval_steps_per_second": 0.176, |
|
"eval_wer": 20.446244243290458, |
|
"step": 2000 |
|
} |
|
], |
|
"max_steps": 5000, |
|
"num_train_epochs": 47, |
|
"total_flos": 2.6324054137700352e+20, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|