|
{ |
|
"best_metric": 0.6776940822601318, |
|
"best_model_checkpoint": "ai-light-dance_singing3_ft_wav2vec2-large-xlsr-53-v1/checkpoint-720", |
|
"epoch": 9.996539792387543, |
|
"global_step": 720, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 0.634, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.6419, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6e-06, |
|
"loss": 0.6223, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.6284, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1e-05, |
|
"loss": 0.6029, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.850746268656717e-06, |
|
"loss": 0.6024, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.701492537313434e-06, |
|
"loss": 0.6227, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.7283244729042053, |
|
"eval_runtime": 128.3245, |
|
"eval_samples_per_second": 4.146, |
|
"eval_steps_per_second": 0.522, |
|
"eval_wer": 0.46589959503760364, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 9.552238805970149e-06, |
|
"loss": 0.6452, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 9.402985074626867e-06, |
|
"loss": 0.6194, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.253731343283582e-06, |
|
"loss": 0.6242, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 9.104477611940299e-06, |
|
"loss": 0.6169, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 8.955223880597016e-06, |
|
"loss": 0.6081, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 8.805970149253732e-06, |
|
"loss": 0.6511, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 8.656716417910447e-06, |
|
"loss": 0.6027, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.7168184518814087, |
|
"eval_runtime": 129.6874, |
|
"eval_samples_per_second": 4.102, |
|
"eval_steps_per_second": 0.517, |
|
"eval_wer": 0.4427588866748088, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 8.507462686567165e-06, |
|
"loss": 0.6228, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 8.35820895522388e-06, |
|
"loss": 0.6092, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 8.208955223880599e-06, |
|
"loss": 0.625, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 8.059701492537314e-06, |
|
"loss": 0.5997, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 7.91044776119403e-06, |
|
"loss": 0.5925, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 7.761194029850747e-06, |
|
"loss": 0.5989, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 7.611940298507463e-06, |
|
"loss": 0.5973, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.706081211566925, |
|
"eval_runtime": 130.3715, |
|
"eval_samples_per_second": 4.081, |
|
"eval_steps_per_second": 0.514, |
|
"eval_wer": 0.4312528122388635, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 7.46268656716418e-06, |
|
"loss": 0.6123, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 7.313432835820896e-06, |
|
"loss": 0.6008, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 7.164179104477612e-06, |
|
"loss": 0.6303, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 7.014925373134329e-06, |
|
"loss": 0.5887, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 6.865671641791045e-06, |
|
"loss": 0.5873, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 6.7164179104477625e-06, |
|
"loss": 0.589, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 6.567164179104478e-06, |
|
"loss": 0.5773, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 0.7067840695381165, |
|
"eval_runtime": 128.8032, |
|
"eval_samples_per_second": 4.13, |
|
"eval_steps_per_second": 0.52, |
|
"eval_wer": 0.4336311628205952, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 6.417910447761194e-06, |
|
"loss": 0.6105, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 6.2686567164179116e-06, |
|
"loss": 0.5964, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 6.119402985074627e-06, |
|
"loss": 0.5959, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 5.970149253731343e-06, |
|
"loss": 0.5843, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"learning_rate": 5.820895522388061e-06, |
|
"loss": 0.6068, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 5.671641791044776e-06, |
|
"loss": 0.6045, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 5.522388059701493e-06, |
|
"loss": 0.5906, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 5.37313432835821e-06, |
|
"loss": 0.5965, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 0.7037749886512756, |
|
"eval_runtime": 129.081, |
|
"eval_samples_per_second": 4.121, |
|
"eval_steps_per_second": 0.519, |
|
"eval_wer": 0.4339525615478563, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 5.2238805970149255e-06, |
|
"loss": 0.5885, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 5.28, |
|
"learning_rate": 5.074626865671642e-06, |
|
"loss": 0.5907, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 5.42, |
|
"learning_rate": 4.925373134328359e-06, |
|
"loss": 0.5862, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 5.55, |
|
"learning_rate": 4.7761194029850745e-06, |
|
"loss": 0.6002, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 4.626865671641791e-06, |
|
"loss": 0.6035, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 5.83, |
|
"learning_rate": 4.477611940298508e-06, |
|
"loss": 0.6163, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 5.97, |
|
"learning_rate": 4.3283582089552236e-06, |
|
"loss": 0.5854, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 0.705437958240509, |
|
"eval_runtime": 128.9432, |
|
"eval_samples_per_second": 4.126, |
|
"eval_steps_per_second": 0.52, |
|
"eval_wer": 0.43086713376615027, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 6.11, |
|
"learning_rate": 4.17910447761194e-06, |
|
"loss": 0.6481, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 4.029850746268657e-06, |
|
"loss": 0.5806, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 6.39, |
|
"learning_rate": 3.8805970149253735e-06, |
|
"loss": 0.5838, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 3.73134328358209e-06, |
|
"loss": 0.6078, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 6.66, |
|
"learning_rate": 3.582089552238806e-06, |
|
"loss": 0.6097, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 6.8, |
|
"learning_rate": 3.447761194029851e-06, |
|
"loss": 0.5909, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 6.94, |
|
"learning_rate": 3.2985074626865677e-06, |
|
"loss": 0.5832, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 0.699042022228241, |
|
"eval_runtime": 127.7248, |
|
"eval_samples_per_second": 4.165, |
|
"eval_steps_per_second": 0.525, |
|
"eval_wer": 0.42527479591180817, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 7.08, |
|
"learning_rate": 3.1492537313432834e-06, |
|
"loss": 0.6283, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 7.22, |
|
"learning_rate": 3e-06, |
|
"loss": 0.5905, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 7.36, |
|
"learning_rate": 2.8507462686567167e-06, |
|
"loss": 0.6143, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 2.7014925373134333e-06, |
|
"loss": 0.6072, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 7.64, |
|
"learning_rate": 2.5522388059701496e-06, |
|
"loss": 0.5894, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 7.78, |
|
"learning_rate": 2.402985074626866e-06, |
|
"loss": 0.604, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 7.91, |
|
"learning_rate": 2.2537313432835824e-06, |
|
"loss": 0.5968, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 0.698808491230011, |
|
"eval_runtime": 129.2631, |
|
"eval_samples_per_second": 4.116, |
|
"eval_steps_per_second": 0.518, |
|
"eval_wer": 0.42315356431188533, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"learning_rate": 2.1044776119402986e-06, |
|
"loss": 0.6225, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 8.19, |
|
"learning_rate": 1.955223880597015e-06, |
|
"loss": 0.5931, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 8.33, |
|
"learning_rate": 1.8059701492537314e-06, |
|
"loss": 0.6011, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 8.47, |
|
"learning_rate": 1.6567164179104479e-06, |
|
"loss": 0.5992, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 8.61, |
|
"learning_rate": 1.5074626865671643e-06, |
|
"loss": 0.6182, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"learning_rate": 1.3582089552238805e-06, |
|
"loss": 0.6231, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 8.89, |
|
"learning_rate": 1.2089552238805971e-06, |
|
"loss": 0.5895, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 0.6866586208343506, |
|
"eval_runtime": 126.5198, |
|
"eval_samples_per_second": 4.205, |
|
"eval_steps_per_second": 0.53, |
|
"eval_wer": 0.4221893681301022, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 1.0597014925373136e-06, |
|
"loss": 0.6387, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 9.17, |
|
"learning_rate": 9.104477611940299e-07, |
|
"loss": 0.6074, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 9.3, |
|
"learning_rate": 7.611940298507463e-07, |
|
"loss": 0.6425, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 9.44, |
|
"learning_rate": 6.119402985074627e-07, |
|
"loss": 0.6131, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 9.58, |
|
"learning_rate": 4.6268656716417914e-07, |
|
"loss": 0.6328, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 9.72, |
|
"learning_rate": 3.134328358208955e-07, |
|
"loss": 0.61, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"learning_rate": 1.6417910447761195e-07, |
|
"loss": 0.6417, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 1.492537313432836e-08, |
|
"loss": 0.6378, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 0.6776940822601318, |
|
"eval_runtime": 127.2273, |
|
"eval_samples_per_second": 4.181, |
|
"eval_steps_per_second": 0.527, |
|
"eval_wer": 0.4203895352574404, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"step": 720, |
|
"total_flos": 2.7197958559499194e+19, |
|
"train_loss": 0.6085318591859605, |
|
"train_runtime": 19393.8759, |
|
"train_samples_per_second": 2.381, |
|
"train_steps_per_second": 0.037 |
|
} |
|
], |
|
"max_steps": 720, |
|
"num_train_epochs": 10, |
|
"total_flos": 2.7197958559499194e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|