|
{ |
|
"best_metric": 1.0062872171401978, |
|
"best_model_checkpoint": "ai-light-dance_stepmania_ft_wav2vec2-large-xlsr-53-v6/checkpoint-2256", |
|
"epoch": 10.0, |
|
"global_step": 3760, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.9023, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.8546, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.2e-05, |
|
"loss": 0.8402, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 0.8535, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 2e-05, |
|
"loss": 0.8511, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 2.4e-05, |
|
"loss": 0.8989, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 2.8e-05, |
|
"loss": 0.8645, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 3.2000000000000005e-05, |
|
"loss": 0.8518, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 3.6e-05, |
|
"loss": 0.8493, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4e-05, |
|
"loss": 0.8645, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 3.989071038251366e-05, |
|
"loss": 0.9279, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 3.9781420765027323e-05, |
|
"loss": 0.8863, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 3.9672131147540983e-05, |
|
"loss": 0.8678, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 3.956284153005465e-05, |
|
"loss": 0.8731, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 3.945355191256831e-05, |
|
"loss": 0.8731, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 3.934426229508197e-05, |
|
"loss": 0.9165, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 3.923497267759563e-05, |
|
"loss": 0.8585, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 3.912568306010929e-05, |
|
"loss": 0.869, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 3.901639344262295e-05, |
|
"loss": 0.8579, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 3.890710382513661e-05, |
|
"loss": 0.8642, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 3.879781420765028e-05, |
|
"loss": 0.916, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.868852459016394e-05, |
|
"loss": 0.8699, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.85792349726776e-05, |
|
"loss": 0.8871, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.846994535519126e-05, |
|
"loss": 0.8526, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.836065573770492e-05, |
|
"loss": 0.863, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.825136612021858e-05, |
|
"loss": 0.914, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.8142076502732245e-05, |
|
"loss": 0.8703, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.8032786885245905e-05, |
|
"loss": 0.8514, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.7923497267759565e-05, |
|
"loss": 0.8656, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.781420765027323e-05, |
|
"loss": 0.8687, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.770491803278689e-05, |
|
"loss": 0.9105, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.759562841530055e-05, |
|
"loss": 0.8647, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.748633879781421e-05, |
|
"loss": 0.868, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 3.737704918032787e-05, |
|
"loss": 0.8642, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 3.726775956284153e-05, |
|
"loss": 0.8642, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 3.71584699453552e-05, |
|
"loss": 0.8784, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.704918032786886e-05, |
|
"loss": 0.8572, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 1.0507858991622925, |
|
"eval_runtime": 360.6121, |
|
"eval_samples_per_second": 7.415, |
|
"eval_steps_per_second": 0.929, |
|
"eval_wer": 0.6600650741855132, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 3.693989071038252e-05, |
|
"loss": 0.8909, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 3.683060109289618e-05, |
|
"loss": 0.8889, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 3.672131147540984e-05, |
|
"loss": 0.8695, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 3.66120218579235e-05, |
|
"loss": 0.8653, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 3.650273224043716e-05, |
|
"loss": 0.8808, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 3.6393442622950826e-05, |
|
"loss": 0.9, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 3.6284153005464486e-05, |
|
"loss": 0.9073, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 3.6174863387978146e-05, |
|
"loss": 0.8674, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 3.6065573770491806e-05, |
|
"loss": 0.8733, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 3.5956284153005466e-05, |
|
"loss": 0.8699, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 3.5846994535519126e-05, |
|
"loss": 0.8951, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 3.5737704918032786e-05, |
|
"loss": 0.8887, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 3.562841530054645e-05, |
|
"loss": 0.8633, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 3.551912568306011e-05, |
|
"loss": 0.8592, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 3.5409836065573773e-05, |
|
"loss": 0.863, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 3.5300546448087433e-05, |
|
"loss": 0.8857, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 3.5191256830601094e-05, |
|
"loss": 0.893, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 3.5081967213114754e-05, |
|
"loss": 0.8761, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 3.4972677595628414e-05, |
|
"loss": 0.8663, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 3.486338797814208e-05, |
|
"loss": 0.8711, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 3.475409836065574e-05, |
|
"loss": 0.8966, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 3.46448087431694e-05, |
|
"loss": 0.8819, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 3.453551912568306e-05, |
|
"loss": 0.8561, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 3.442622950819672e-05, |
|
"loss": 0.8643, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 3.431693989071038e-05, |
|
"loss": 0.8713, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 3.420765027322405e-05, |
|
"loss": 0.8968, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 3.409836065573771e-05, |
|
"loss": 0.9026, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 3.398907103825137e-05, |
|
"loss": 0.8785, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.3879781420765035e-05, |
|
"loss": 0.8609, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.3770491803278695e-05, |
|
"loss": 0.8678, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 3.3661202185792355e-05, |
|
"loss": 0.898, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 3.3551912568306015e-05, |
|
"loss": 0.8957, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 3.3442622950819675e-05, |
|
"loss": 0.8644, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 3.3333333333333335e-05, |
|
"loss": 0.8691, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 3.3224043715847e-05, |
|
"loss": 0.8695, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 3.311475409836066e-05, |
|
"loss": 0.8985, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 3.300546448087432e-05, |
|
"loss": 0.8724, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 3.289617486338798e-05, |
|
"loss": 0.8671, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 1.0755400657653809, |
|
"eval_runtime": 359.2021, |
|
"eval_samples_per_second": 7.444, |
|
"eval_steps_per_second": 0.933, |
|
"eval_wer": 0.6580638781766044, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 3.278688524590164e-05, |
|
"loss": 0.9214, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 3.26775956284153e-05, |
|
"loss": 0.8734, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 3.256830601092896e-05, |
|
"loss": 0.8595, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 3.245901639344263e-05, |
|
"loss": 0.8591, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 3.234972677595629e-05, |
|
"loss": 0.8733, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 3.224043715846995e-05, |
|
"loss": 0.9151, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 3.213114754098361e-05, |
|
"loss": 0.8615, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 3.203278688524591e-05, |
|
"loss": 0.8529, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 3.192349726775957e-05, |
|
"loss": 0.879, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 3.181420765027323e-05, |
|
"loss": 0.8572, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 3.170491803278689e-05, |
|
"loss": 0.9139, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 3.159562841530055e-05, |
|
"loss": 0.8654, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 3.148633879781421e-05, |
|
"loss": 0.8701, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 3.1377049180327875e-05, |
|
"loss": 0.8842, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 3.1267759562841535e-05, |
|
"loss": 0.8644, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 3.1158469945355195e-05, |
|
"loss": 0.9083, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 3.1049180327868855e-05, |
|
"loss": 0.857, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 3.0939890710382515e-05, |
|
"loss": 0.8563, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 3.0830601092896175e-05, |
|
"loss": 0.8621, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 3.0721311475409836e-05, |
|
"loss": 0.87, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 3.06120218579235e-05, |
|
"loss": 0.9134, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 3.050273224043716e-05, |
|
"loss": 0.8582, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 3.0393442622950822e-05, |
|
"loss": 0.8574, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 3.0284153005464483e-05, |
|
"loss": 0.8864, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 3.0174863387978143e-05, |
|
"loss": 0.8538, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 3.0065573770491806e-05, |
|
"loss": 0.9179, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.9956284153005466e-05, |
|
"loss": 0.8733, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 2.9846994535519126e-05, |
|
"loss": 0.8676, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 2.973770491803279e-05, |
|
"loss": 0.8642, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 2.962841530054645e-05, |
|
"loss": 0.8728, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 2.951912568306011e-05, |
|
"loss": 0.9145, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 2.940983606557377e-05, |
|
"loss": 0.856, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 2.9300546448087433e-05, |
|
"loss": 0.8666, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 2.9191256830601093e-05, |
|
"loss": 0.8574, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 2.9081967213114753e-05, |
|
"loss": 0.8595, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 2.8972677595628417e-05, |
|
"loss": 0.8913, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 2.8863387978142077e-05, |
|
"loss": 0.8578, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 1.0152311325073242, |
|
"eval_runtime": 365.393, |
|
"eval_samples_per_second": 7.318, |
|
"eval_steps_per_second": 0.917, |
|
"eval_wer": 0.6786973862025776, |
|
"step": 1128 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 2.8754098360655744e-05, |
|
"loss": 0.8817, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 2.8644808743169404e-05, |
|
"loss": 0.8914, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 2.8535519125683064e-05, |
|
"loss": 0.8723, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 2.8426229508196724e-05, |
|
"loss": 0.8644, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 2.8316939890710387e-05, |
|
"loss": 0.862, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 2.8207650273224047e-05, |
|
"loss": 0.8653, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 2.8098360655737708e-05, |
|
"loss": 0.8903, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 2.798907103825137e-05, |
|
"loss": 0.8701, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 2.787978142076503e-05, |
|
"loss": 0.8636, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 2.777049180327869e-05, |
|
"loss": 0.8575, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 2.7661202185792355e-05, |
|
"loss": 0.8736, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 2.7551912568306015e-05, |
|
"loss": 0.8842, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 2.7442622950819675e-05, |
|
"loss": 0.8454, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 2.7333333333333335e-05, |
|
"loss": 0.8648, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 2.7224043715846998e-05, |
|
"loss": 0.8473, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 2.7114754098360658e-05, |
|
"loss": 0.8692, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 2.700546448087432e-05, |
|
"loss": 0.8916, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 2.6896174863387982e-05, |
|
"loss": 0.8669, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 2.6786885245901642e-05, |
|
"loss": 0.8551, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 2.6677595628415302e-05, |
|
"loss": 0.8571, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 2.6568306010928962e-05, |
|
"loss": 0.8733, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 2.6459016393442625e-05, |
|
"loss": 0.9035, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 2.6349726775956286e-05, |
|
"loss": 0.8585, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 2.6240437158469946e-05, |
|
"loss": 0.8579, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 2.613114754098361e-05, |
|
"loss": 0.8658, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 2.602185792349727e-05, |
|
"loss": 0.8735, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 2.591256830601093e-05, |
|
"loss": 0.908, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"learning_rate": 2.5803278688524593e-05, |
|
"loss": 0.8503, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 2.5693989071038253e-05, |
|
"loss": 0.8539, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 2.5584699453551913e-05, |
|
"loss": 0.8513, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 2.5475409836065573e-05, |
|
"loss": 0.853, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 2.5366120218579236e-05, |
|
"loss": 0.892, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 2.5256830601092896e-05, |
|
"loss": 0.857, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 2.5147540983606556e-05, |
|
"loss": 0.8481, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 2.503825136612022e-05, |
|
"loss": 0.8503, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 2.492896174863388e-05, |
|
"loss": 0.8751, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 2.481967213114754e-05, |
|
"loss": 0.8701, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 2.4710382513661207e-05, |
|
"loss": 0.8552, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 1.0537359714508057, |
|
"eval_runtime": 358.2144, |
|
"eval_samples_per_second": 7.465, |
|
"eval_steps_per_second": 0.935, |
|
"eval_wer": 0.6557001125378461, |
|
"step": 1504 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 2.4601092896174867e-05, |
|
"loss": 0.8939, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 2.4491803278688527e-05, |
|
"loss": 0.8571, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 2.438251366120219e-05, |
|
"loss": 0.862, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 2.427322404371585e-05, |
|
"loss": 0.858, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 4.12, |
|
"learning_rate": 2.416393442622951e-05, |
|
"loss": 0.8522, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"learning_rate": 2.4054644808743174e-05, |
|
"loss": 0.8906, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 4.18, |
|
"learning_rate": 2.3945355191256834e-05, |
|
"loss": 0.8618, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"learning_rate": 2.3836065573770494e-05, |
|
"loss": 0.8576, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 4.23, |
|
"learning_rate": 2.3726775956284158e-05, |
|
"loss": 0.8459, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 4.26, |
|
"learning_rate": 2.3617486338797818e-05, |
|
"loss": 0.8544, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 2.3508196721311478e-05, |
|
"loss": 0.9056, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 2.3398907103825138e-05, |
|
"loss": 0.8703, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 2.32896174863388e-05, |
|
"loss": 0.8526, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 2.318032786885246e-05, |
|
"loss": 0.8652, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"learning_rate": 2.307103825136612e-05, |
|
"loss": 0.8505, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 2.2961748633879785e-05, |
|
"loss": 0.8932, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 2.2852459016393445e-05, |
|
"loss": 0.8703, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 2.2743169398907105e-05, |
|
"loss": 0.859, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"learning_rate": 2.263387978142077e-05, |
|
"loss": 0.8521, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 2.252459016393443e-05, |
|
"loss": 0.8544, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"learning_rate": 2.241530054644809e-05, |
|
"loss": 0.8936, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 2.230601092896175e-05, |
|
"loss": 0.8668, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 2.2196721311475412e-05, |
|
"loss": 0.8601, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 4.63, |
|
"learning_rate": 2.2087431693989072e-05, |
|
"loss": 0.86, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"learning_rate": 2.1978142076502732e-05, |
|
"loss": 0.8507, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 2.1868852459016396e-05, |
|
"loss": 0.9023, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 4.71, |
|
"learning_rate": 2.1759562841530056e-05, |
|
"loss": 0.8565, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 2.1650273224043716e-05, |
|
"loss": 0.8655, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 4.76, |
|
"learning_rate": 2.1540983606557376e-05, |
|
"loss": 0.8488, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 2.143169398907104e-05, |
|
"loss": 0.8517, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 2.13224043715847e-05, |
|
"loss": 0.8877, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"learning_rate": 2.121311475409836e-05, |
|
"loss": 0.8726, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 2.1103825136612023e-05, |
|
"loss": 0.8537, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 2.0994535519125683e-05, |
|
"loss": 0.8486, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 4.92, |
|
"learning_rate": 2.0885245901639343e-05, |
|
"loss": 0.8477, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 2.077595628415301e-05, |
|
"loss": 0.8784, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 4.97, |
|
"learning_rate": 2.066666666666667e-05, |
|
"loss": 0.8456, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 2.055737704918033e-05, |
|
"loss": 0.8354, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 1.038648009300232, |
|
"eval_runtime": 360.6739, |
|
"eval_samples_per_second": 7.414, |
|
"eval_steps_per_second": 0.929, |
|
"eval_wer": 0.6605500699241429, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 2.0448087431693993e-05, |
|
"loss": 0.9179, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"learning_rate": 2.0338797814207653e-05, |
|
"loss": 0.8421, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 5.08, |
|
"learning_rate": 2.0229508196721313e-05, |
|
"loss": 0.8479, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 5.11, |
|
"learning_rate": 2.0120218579234977e-05, |
|
"loss": 0.8555, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 5.13, |
|
"learning_rate": 2.0010928961748637e-05, |
|
"loss": 0.848, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 5.16, |
|
"learning_rate": 1.9901639344262297e-05, |
|
"loss": 0.9042, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 5.19, |
|
"learning_rate": 1.979234972677596e-05, |
|
"loss": 0.8496, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"learning_rate": 1.968306010928962e-05, |
|
"loss": 0.8603, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 5.24, |
|
"learning_rate": 1.957377049180328e-05, |
|
"loss": 0.8457, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 5.27, |
|
"learning_rate": 1.946448087431694e-05, |
|
"loss": 0.8492, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 5.29, |
|
"learning_rate": 1.9355191256830604e-05, |
|
"loss": 0.8977, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 5.32, |
|
"learning_rate": 1.9245901639344264e-05, |
|
"loss": 0.8486, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 5.35, |
|
"learning_rate": 1.9136612021857924e-05, |
|
"loss": 0.837, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 5.37, |
|
"learning_rate": 1.9027322404371588e-05, |
|
"loss": 0.8396, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 5.4, |
|
"learning_rate": 1.8918032786885248e-05, |
|
"loss": 0.8493, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 5.43, |
|
"learning_rate": 1.8808743169398908e-05, |
|
"loss": 0.8894, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 5.45, |
|
"learning_rate": 1.869945355191257e-05, |
|
"loss": 0.842, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 5.48, |
|
"learning_rate": 1.859016393442623e-05, |
|
"loss": 0.8328, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 5.51, |
|
"learning_rate": 1.848087431693989e-05, |
|
"loss": 0.8287, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"learning_rate": 1.837158469945355e-05, |
|
"loss": 0.837, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 5.56, |
|
"learning_rate": 1.8262295081967215e-05, |
|
"loss": 0.8962, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 5.59, |
|
"learning_rate": 1.8153005464480875e-05, |
|
"loss": 0.8476, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 5.61, |
|
"learning_rate": 1.8043715846994535e-05, |
|
"loss": 0.84, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 5.64, |
|
"learning_rate": 1.79344262295082e-05, |
|
"loss": 0.846, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 5.66, |
|
"learning_rate": 1.782513661202186e-05, |
|
"loss": 0.8291, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 1.7715846994535522e-05, |
|
"loss": 0.9029, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 5.72, |
|
"learning_rate": 1.7606557377049182e-05, |
|
"loss": 0.8283, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 5.74, |
|
"learning_rate": 1.7497267759562842e-05, |
|
"loss": 0.8471, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 5.77, |
|
"learning_rate": 1.7387978142076506e-05, |
|
"loss": 0.8461, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"learning_rate": 1.7278688524590166e-05, |
|
"loss": 0.846, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 1.7169398907103826e-05, |
|
"loss": 0.8967, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 5.85, |
|
"learning_rate": 1.706010928961749e-05, |
|
"loss": 0.8457, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 5.88, |
|
"learning_rate": 1.695081967213115e-05, |
|
"loss": 0.8474, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 5.9, |
|
"learning_rate": 1.684153005464481e-05, |
|
"loss": 0.8363, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 5.93, |
|
"learning_rate": 1.6732240437158473e-05, |
|
"loss": 0.8591, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 5.96, |
|
"learning_rate": 1.6622950819672133e-05, |
|
"loss": 0.8584, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 5.98, |
|
"learning_rate": 1.6513661202185793e-05, |
|
"loss": 0.8543, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 1.0062872171401978, |
|
"eval_runtime": 363.0682, |
|
"eval_samples_per_second": 7.365, |
|
"eval_steps_per_second": 0.923, |
|
"eval_wer": 0.6580026651222142, |
|
"step": 2256 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 1.6404371584699453e-05, |
|
"loss": 0.8769, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 1.6295081967213116e-05, |
|
"loss": 0.8632, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 6.06, |
|
"learning_rate": 1.6185792349726776e-05, |
|
"loss": 0.8381, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"learning_rate": 1.6076502732240437e-05, |
|
"loss": 0.8396, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 6.12, |
|
"learning_rate": 1.59672131147541e-05, |
|
"loss": 0.847, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 6.14, |
|
"learning_rate": 1.585792349726776e-05, |
|
"loss": 0.8782, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 6.17, |
|
"learning_rate": 1.5748633879781424e-05, |
|
"loss": 0.8695, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 6.2, |
|
"learning_rate": 1.5639344262295084e-05, |
|
"loss": 0.8467, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 6.22, |
|
"learning_rate": 1.5530054644808744e-05, |
|
"loss": 0.8342, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 1.5420765027322407e-05, |
|
"loss": 0.8255, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 6.28, |
|
"learning_rate": 1.5311475409836067e-05, |
|
"loss": 0.8762, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 6.3, |
|
"learning_rate": 1.5202185792349729e-05, |
|
"loss": 0.8643, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 6.33, |
|
"learning_rate": 1.5092896174863389e-05, |
|
"loss": 0.8425, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 6.36, |
|
"learning_rate": 1.498360655737705e-05, |
|
"loss": 0.8416, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"learning_rate": 1.4874316939890713e-05, |
|
"loss": 0.8412, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 6.41, |
|
"learning_rate": 1.4765027322404373e-05, |
|
"loss": 0.8716, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 6.44, |
|
"learning_rate": 1.4655737704918034e-05, |
|
"loss": 0.8713, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 6.46, |
|
"learning_rate": 1.4546448087431694e-05, |
|
"loss": 0.8283, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 6.49, |
|
"learning_rate": 1.4437158469945356e-05, |
|
"loss": 0.8338, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 6.52, |
|
"learning_rate": 1.4327868852459016e-05, |
|
"loss": 0.8402, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 6.54, |
|
"learning_rate": 1.4218579234972678e-05, |
|
"loss": 0.8705, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 6.57, |
|
"learning_rate": 1.410928961748634e-05, |
|
"loss": 0.8612, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 1.4e-05, |
|
"loss": 0.841, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 6.62, |
|
"learning_rate": 1.3890710382513662e-05, |
|
"loss": 0.8485, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"learning_rate": 1.3781420765027325e-05, |
|
"loss": 0.8808, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 6.68, |
|
"learning_rate": 1.3672131147540985e-05, |
|
"loss": 0.8681, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 6.7, |
|
"learning_rate": 1.3562841530054647e-05, |
|
"loss": 0.8503, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 6.73, |
|
"learning_rate": 1.3453551912568309e-05, |
|
"loss": 0.8398, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 6.76, |
|
"learning_rate": 1.3344262295081969e-05, |
|
"loss": 0.8275, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 6.78, |
|
"learning_rate": 1.323497267759563e-05, |
|
"loss": 0.8246, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 6.81, |
|
"learning_rate": 1.312568306010929e-05, |
|
"loss": 0.8519, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 6.84, |
|
"learning_rate": 1.3016393442622952e-05, |
|
"loss": 0.8676, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 6.86, |
|
"learning_rate": 1.2907103825136614e-05, |
|
"loss": 0.8386, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 6.89, |
|
"learning_rate": 1.2797814207650274e-05, |
|
"loss": 0.8457, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 6.91, |
|
"learning_rate": 1.2688524590163936e-05, |
|
"loss": 0.8426, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 6.94, |
|
"learning_rate": 1.2579234972677596e-05, |
|
"loss": 0.8799, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 6.97, |
|
"learning_rate": 1.2469945355191258e-05, |
|
"loss": 0.8464, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 6.99, |
|
"learning_rate": 1.236065573770492e-05, |
|
"loss": 0.8556, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 1.048650860786438, |
|
"eval_runtime": 362.39, |
|
"eval_samples_per_second": 7.379, |
|
"eval_steps_per_second": 0.924, |
|
"eval_wer": 0.6498754549777985, |
|
"step": 2632 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 1.225136612021858e-05, |
|
"loss": 0.8925, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 7.05, |
|
"learning_rate": 1.2142076502732241e-05, |
|
"loss": 0.8356, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 7.07, |
|
"learning_rate": 1.2032786885245901e-05, |
|
"loss": 0.8401, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 7.1, |
|
"learning_rate": 1.1923497267759563e-05, |
|
"loss": 0.8348, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 7.13, |
|
"learning_rate": 1.1814207650273223e-05, |
|
"loss": 0.834, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 7.15, |
|
"learning_rate": 1.1704918032786887e-05, |
|
"loss": 0.8962, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 7.18, |
|
"learning_rate": 1.1595628415300548e-05, |
|
"loss": 0.8275, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 7.21, |
|
"learning_rate": 1.148633879781421e-05, |
|
"loss": 0.8286, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 7.23, |
|
"learning_rate": 1.137704918032787e-05, |
|
"loss": 0.8419, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 7.26, |
|
"learning_rate": 1.1267759562841532e-05, |
|
"loss": 0.8249, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 7.29, |
|
"learning_rate": 1.1158469945355192e-05, |
|
"loss": 0.8883, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 7.31, |
|
"learning_rate": 1.1049180327868854e-05, |
|
"loss": 0.8398, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 7.34, |
|
"learning_rate": 1.0939890710382515e-05, |
|
"loss": 0.832, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 7.37, |
|
"learning_rate": 1.0830601092896176e-05, |
|
"loss": 0.835, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 7.39, |
|
"learning_rate": 1.0721311475409837e-05, |
|
"loss": 0.8384, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 7.42, |
|
"learning_rate": 1.0612021857923497e-05, |
|
"loss": 0.9041, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 7.45, |
|
"learning_rate": 1.0502732240437159e-05, |
|
"loss": 0.8281, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 7.47, |
|
"learning_rate": 1.0393442622950821e-05, |
|
"loss": 0.8426, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 1.0284153005464481e-05, |
|
"loss": 0.8341, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 7.53, |
|
"learning_rate": 1.0174863387978143e-05, |
|
"loss": 0.8367, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 7.55, |
|
"learning_rate": 1.0065573770491803e-05, |
|
"loss": 0.8891, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 7.58, |
|
"learning_rate": 9.956284153005466e-06, |
|
"loss": 0.8495, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 7.61, |
|
"learning_rate": 9.846994535519126e-06, |
|
"loss": 0.8458, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 7.63, |
|
"learning_rate": 9.737704918032788e-06, |
|
"loss": 0.8416, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 7.66, |
|
"learning_rate": 9.628415300546448e-06, |
|
"loss": 0.8304, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 7.69, |
|
"learning_rate": 9.51912568306011e-06, |
|
"loss": 0.8864, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 7.71, |
|
"learning_rate": 9.409836065573772e-06, |
|
"loss": 0.8444, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 7.74, |
|
"learning_rate": 9.300546448087432e-06, |
|
"loss": 0.8378, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 7.77, |
|
"learning_rate": 9.191256830601093e-06, |
|
"loss": 0.8369, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 7.79, |
|
"learning_rate": 9.081967213114755e-06, |
|
"loss": 0.8266, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 7.82, |
|
"learning_rate": 8.972677595628417e-06, |
|
"loss": 0.8802, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 7.85, |
|
"learning_rate": 8.863387978142077e-06, |
|
"loss": 0.8445, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 7.87, |
|
"learning_rate": 8.765027322404372e-06, |
|
"loss": 0.8247, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 7.9, |
|
"learning_rate": 8.655737704918034e-06, |
|
"loss": 0.8262, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 7.93, |
|
"learning_rate": 8.546448087431694e-06, |
|
"loss": 0.8228, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 7.95, |
|
"learning_rate": 8.437158469945356e-06, |
|
"loss": 0.8594, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 7.98, |
|
"learning_rate": 8.327868852459016e-06, |
|
"loss": 0.8356, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 1.0407345294952393, |
|
"eval_runtime": 360.9048, |
|
"eval_samples_per_second": 7.409, |
|
"eval_steps_per_second": 0.928, |
|
"eval_wer": 0.6548855080448079, |
|
"step": 3008 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 8.218579234972678e-06, |
|
"loss": 0.8511, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 8.10928961748634e-06, |
|
"loss": 0.8626, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"learning_rate": 8.010928961748633e-06, |
|
"loss": 0.8381, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 8.09, |
|
"learning_rate": 7.901639344262295e-06, |
|
"loss": 0.8257, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 8.11, |
|
"learning_rate": 7.792349726775957e-06, |
|
"loss": 0.8267, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 8.14, |
|
"learning_rate": 7.683060109289619e-06, |
|
"loss": 0.8519, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 8.16, |
|
"learning_rate": 7.5737704918032795e-06, |
|
"loss": 0.8679, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 8.19, |
|
"learning_rate": 7.46448087431694e-06, |
|
"loss": 0.8369, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 8.22, |
|
"learning_rate": 7.355191256830601e-06, |
|
"loss": 0.8326, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 8.24, |
|
"learning_rate": 7.245901639344263e-06, |
|
"loss": 0.8234, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 8.27, |
|
"learning_rate": 7.136612021857924e-06, |
|
"loss": 0.8373, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 8.3, |
|
"learning_rate": 7.027322404371585e-06, |
|
"loss": 0.8728, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 8.32, |
|
"learning_rate": 6.918032786885246e-06, |
|
"loss": 0.8309, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 8.35, |
|
"learning_rate": 6.808743169398908e-06, |
|
"loss": 0.8274, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 8.38, |
|
"learning_rate": 6.699453551912569e-06, |
|
"loss": 0.8257, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 8.4, |
|
"learning_rate": 6.59016393442623e-06, |
|
"loss": 0.8457, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 8.43, |
|
"learning_rate": 6.480874316939891e-06, |
|
"loss": 0.8646, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 8.46, |
|
"learning_rate": 6.371584699453552e-06, |
|
"loss": 0.8135, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 8.48, |
|
"learning_rate": 6.262295081967214e-06, |
|
"loss": 0.8343, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 8.51, |
|
"learning_rate": 6.153005464480875e-06, |
|
"loss": 0.8413, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 8.54, |
|
"learning_rate": 6.043715846994536e-06, |
|
"loss": 0.8499, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 8.56, |
|
"learning_rate": 5.9344262295081965e-06, |
|
"loss": 0.8735, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 8.59, |
|
"learning_rate": 5.825136612021859e-06, |
|
"loss": 0.8273, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 8.62, |
|
"learning_rate": 5.71584699453552e-06, |
|
"loss": 0.8266, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 8.64, |
|
"learning_rate": 5.606557377049181e-06, |
|
"loss": 0.8421, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 8.67, |
|
"learning_rate": 5.497267759562842e-06, |
|
"loss": 0.8458, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 8.7, |
|
"learning_rate": 5.387978142076503e-06, |
|
"loss": 0.8547, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 8.72, |
|
"learning_rate": 5.2786885245901645e-06, |
|
"loss": 0.8247, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"learning_rate": 5.1693989071038254e-06, |
|
"loss": 0.8175, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 8.78, |
|
"learning_rate": 5.060109289617486e-06, |
|
"loss": 0.8322, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 8.8, |
|
"learning_rate": 4.950819672131148e-06, |
|
"loss": 0.8364, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 8.83, |
|
"learning_rate": 4.841530054644809e-06, |
|
"loss": 0.867, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 8.86, |
|
"learning_rate": 4.73224043715847e-06, |
|
"loss": 0.8345, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 8.88, |
|
"learning_rate": 4.622950819672132e-06, |
|
"loss": 0.8281, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 8.91, |
|
"learning_rate": 4.513661202185793e-06, |
|
"loss": 0.8189, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 8.94, |
|
"learning_rate": 4.4043715846994535e-06, |
|
"loss": 0.8394, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 8.96, |
|
"learning_rate": 4.295081967213115e-06, |
|
"loss": 0.8385, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 8.99, |
|
"learning_rate": 4.185792349726776e-06, |
|
"loss": 0.8227, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 1.038238525390625, |
|
"eval_runtime": 357.9219, |
|
"eval_samples_per_second": 7.471, |
|
"eval_steps_per_second": 0.936, |
|
"eval_wer": 0.6506194290234635, |
|
"step": 3384 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 4.076502732240438e-06, |
|
"loss": 0.8718, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 3.967213114754099e-06, |
|
"loss": 0.8406, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 9.07, |
|
"learning_rate": 3.85792349726776e-06, |
|
"loss": 0.8334, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 9.1, |
|
"learning_rate": 3.748633879781421e-06, |
|
"loss": 0.8303, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 9.12, |
|
"learning_rate": 3.6393442622950824e-06, |
|
"loss": 0.8278, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 9.15, |
|
"learning_rate": 3.5300546448087438e-06, |
|
"loss": 0.8616, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 9.18, |
|
"learning_rate": 3.4207650273224047e-06, |
|
"loss": 0.829, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 9.2, |
|
"learning_rate": 3.3114754098360656e-06, |
|
"loss": 0.8231, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 9.23, |
|
"learning_rate": 3.202185792349727e-06, |
|
"loss": 0.8128, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 9.26, |
|
"learning_rate": 3.0928961748633883e-06, |
|
"loss": 0.8224, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 9.28, |
|
"learning_rate": 2.9836065573770496e-06, |
|
"loss": 0.8604, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 9.31, |
|
"learning_rate": 2.8743169398907105e-06, |
|
"loss": 0.8285, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 9.34, |
|
"learning_rate": 2.765027322404372e-06, |
|
"loss": 0.8291, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 9.36, |
|
"learning_rate": 2.6557377049180328e-06, |
|
"loss": 0.8288, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 9.39, |
|
"learning_rate": 2.5464480874316945e-06, |
|
"loss": 0.8224, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 9.41, |
|
"learning_rate": 2.4371584699453554e-06, |
|
"loss": 0.8664, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 9.44, |
|
"learning_rate": 2.3278688524590163e-06, |
|
"loss": 0.8186, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 9.47, |
|
"learning_rate": 2.2185792349726777e-06, |
|
"loss": 0.8271, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 9.49, |
|
"learning_rate": 2.109289617486339e-06, |
|
"loss": 0.8177, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 9.52, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 0.8203, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 9.55, |
|
"learning_rate": 1.8907103825136612e-06, |
|
"loss": 0.8688, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"learning_rate": 1.7814207650273224e-06, |
|
"loss": 0.8323, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 9.6, |
|
"learning_rate": 1.6721311475409837e-06, |
|
"loss": 0.8193, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 9.63, |
|
"learning_rate": 1.5628415300546448e-06, |
|
"loss": 0.8203, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 9.65, |
|
"learning_rate": 1.4535519125683062e-06, |
|
"loss": 0.8243, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 9.68, |
|
"learning_rate": 1.3442622950819673e-06, |
|
"loss": 0.8603, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 9.71, |
|
"learning_rate": 1.2349726775956284e-06, |
|
"loss": 0.8327, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 9.73, |
|
"learning_rate": 1.1256830601092897e-06, |
|
"loss": 0.8274, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 9.76, |
|
"learning_rate": 1.0163934426229509e-06, |
|
"loss": 0.8206, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 9.79, |
|
"learning_rate": 9.071038251366121e-07, |
|
"loss": 0.8213, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 9.81, |
|
"learning_rate": 7.978142076502733e-07, |
|
"loss": 0.8608, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 9.84, |
|
"learning_rate": 6.885245901639345e-07, |
|
"loss": 0.8235, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 9.87, |
|
"learning_rate": 5.792349726775957e-07, |
|
"loss": 0.8329, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 9.89, |
|
"learning_rate": 4.699453551912569e-07, |
|
"loss": 0.8279, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 9.92, |
|
"learning_rate": 3.6065573770491807e-07, |
|
"loss": 0.8218, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 9.95, |
|
"learning_rate": 2.5136612021857925e-07, |
|
"loss": 0.8451, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 9.97, |
|
"learning_rate": 1.4207650273224045e-07, |
|
"loss": 0.8189, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 3.278688524590164e-08, |
|
"loss": 0.8148, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 1.044033408164978, |
|
"eval_runtime": 359.0191, |
|
"eval_samples_per_second": 7.448, |
|
"eval_steps_per_second": 0.933, |
|
"eval_wer": 0.6499837549971041, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"step": 3760, |
|
"total_flos": 7.2906157544976376e+19, |
|
"train_loss": 0.8583133852228205, |
|
"train_runtime": 77803.3749, |
|
"train_samples_per_second": 3.093, |
|
"train_steps_per_second": 0.048 |
|
} |
|
], |
|
"max_steps": 3760, |
|
"num_train_epochs": 10, |
|
"total_flos": 7.2906157544976376e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|