|
{ |
|
"best_metric": 0.5577101707458496, |
|
"best_model_checkpoint": "ai-light-dance_singing3_ft_wav2vec2-large-xlsr-53-v1/checkpoint-13536", |
|
"epoch": 49.9974025974026, |
|
"global_step": 14400, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.6e-08, |
|
"loss": 0.5118, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 3.2e-08, |
|
"loss": 0.5028, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.639999999999999e-08, |
|
"loss": 0.4975, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 6.239999999999999e-08, |
|
"loss": 0.5018, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 7.839999999999999e-08, |
|
"loss": 0.5586, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.44e-08, |
|
"loss": 0.4638, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.104e-07, |
|
"loss": 0.4929, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.264e-07, |
|
"loss": 0.4867, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.424e-07, |
|
"loss": 0.5007, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.5840000000000002e-07, |
|
"loss": 0.5239, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.744e-07, |
|
"loss": 0.4796, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.904e-07, |
|
"loss": 0.4702, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 2.064e-07, |
|
"loss": 0.5091, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 2.2239999999999998e-07, |
|
"loss": 0.4932, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 2.384e-07, |
|
"loss": 0.4892, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 2.544e-07, |
|
"loss": 0.4815, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 2.704e-07, |
|
"loss": 0.5209, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 2.8639999999999997e-07, |
|
"loss": 0.4332, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.024e-07, |
|
"loss": 0.4571, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.184e-07, |
|
"loss": 0.4908, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.344e-07, |
|
"loss": 0.4846, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.488e-07, |
|
"loss": 0.4669, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 3.648e-07, |
|
"loss": 0.4448, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.808e-07, |
|
"loss": 0.4909, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.9679999999999995e-07, |
|
"loss": 0.5204, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 4.128e-07, |
|
"loss": 0.4829, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 4.288e-07, |
|
"loss": 0.4745, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.4479999999999996e-07, |
|
"loss": 0.4853, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.5759792327880859, |
|
"eval_runtime": 131.0362, |
|
"eval_samples_per_second": 4.06, |
|
"eval_steps_per_second": 1.015, |
|
"eval_wer": 0.3097640933341904, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 4.6079999999999994e-07, |
|
"loss": 0.5331, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 4.768e-07, |
|
"loss": 0.4663, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 4.928e-07, |
|
"loss": 0.4768, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 5.088e-07, |
|
"loss": 0.4554, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 5.248e-07, |
|
"loss": 0.5003, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 5.408e-07, |
|
"loss": 0.4539, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 5.567999999999999e-07, |
|
"loss": 0.4418, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 5.727999999999999e-07, |
|
"loss": 0.4517, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 5.888e-07, |
|
"loss": 0.445, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.048e-07, |
|
"loss": 0.4677, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 6.208e-07, |
|
"loss": 0.5141, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 6.368e-07, |
|
"loss": 0.4952, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 6.528e-07, |
|
"loss": 0.4539, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 6.688e-07, |
|
"loss": 0.4497, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 6.847999999999999e-07, |
|
"loss": 0.4799, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 7.007999999999999e-07, |
|
"loss": 0.478, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 7.168e-07, |
|
"loss": 0.4606, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 7.328e-07, |
|
"loss": 0.4826, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 7.488e-07, |
|
"loss": 0.4652, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 7.648e-07, |
|
"loss": 0.4844, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 7.808e-07, |
|
"loss": 0.5098, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 7.967999999999999e-07, |
|
"loss": 0.4814, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 8.128e-07, |
|
"loss": 0.4684, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 8.287999999999999e-07, |
|
"loss": 0.4636, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 8.447999999999999e-07, |
|
"loss": 0.486, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 8.608e-07, |
|
"loss": 0.4856, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 8.768e-07, |
|
"loss": 0.4739, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 8.928e-07, |
|
"loss": 0.4527, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 9.088e-07, |
|
"loss": 0.48, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.5787187218666077, |
|
"eval_runtime": 132.722, |
|
"eval_samples_per_second": 4.008, |
|
"eval_steps_per_second": 1.002, |
|
"eval_wer": 0.30847849842514624, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.247999999999999e-07, |
|
"loss": 0.511, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 9.408e-07, |
|
"loss": 0.4765, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 9.567999999999999e-07, |
|
"loss": 0.4452, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 9.728e-07, |
|
"loss": 0.4748, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 9.888e-07, |
|
"loss": 0.4958, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.0047999999999999e-06, |
|
"loss": 0.4684, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.0207999999999998e-06, |
|
"loss": 0.483, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.0368e-06, |
|
"loss": 0.4703, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 1.0528e-06, |
|
"loss": 0.4534, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 1.0687999999999998e-06, |
|
"loss": 0.4886, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 1.0848e-06, |
|
"loss": 0.5094, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 1.1008e-06, |
|
"loss": 0.4495, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 1.1168e-06, |
|
"loss": 0.4246, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 1.1328e-06, |
|
"loss": 0.4479, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 1.1488e-06, |
|
"loss": 0.4625, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 1.1648e-06, |
|
"loss": 0.5005, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 1.1808e-06, |
|
"loss": 0.463, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 1.1968000000000001e-06, |
|
"loss": 0.4572, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 1.2128e-06, |
|
"loss": 0.4968, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 1.2287999999999998e-06, |
|
"loss": 0.5079, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 1.2448e-06, |
|
"loss": 0.5009, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.2607999999999999e-06, |
|
"loss": 0.4464, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.2767999999999998e-06, |
|
"loss": 0.4434, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.2928e-06, |
|
"loss": 0.4415, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.3087999999999999e-06, |
|
"loss": 0.4516, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 1.3248e-06, |
|
"loss": 0.4565, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.3408e-06, |
|
"loss": 0.4381, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.3567999999999999e-06, |
|
"loss": 0.4579, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.3728e-06, |
|
"loss": 0.4625, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.5924614071846008, |
|
"eval_runtime": 132.5598, |
|
"eval_samples_per_second": 4.013, |
|
"eval_steps_per_second": 1.003, |
|
"eval_wer": 0.311178247734139, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 1.3888e-06, |
|
"loss": 0.5048, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 1.4048e-06, |
|
"loss": 0.4607, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 1.4208e-06, |
|
"loss": 0.4824, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 1.4368e-06, |
|
"loss": 0.4697, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 1.4528e-06, |
|
"loss": 0.5011, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 1.4688e-06, |
|
"loss": 0.4597, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 1.4847999999999997e-06, |
|
"loss": 0.4611, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 1.5007999999999999e-06, |
|
"loss": 0.4511, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 1.5167999999999998e-06, |
|
"loss": 0.4736, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 1.5328e-06, |
|
"loss": 0.4919, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 1.5487999999999999e-06, |
|
"loss": 0.4886, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 1.5647999999999998e-06, |
|
"loss": 0.4864, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 1.5808e-06, |
|
"loss": 0.4288, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 1.5968e-06, |
|
"loss": 0.4678, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 1.6128e-06, |
|
"loss": 0.4673, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 1.6288e-06, |
|
"loss": 0.4817, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 1.6448e-06, |
|
"loss": 0.4684, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 1.6608e-06, |
|
"loss": 0.4431, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 1.6768e-06, |
|
"loss": 0.4504, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 1.6928e-06, |
|
"loss": 0.472, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 1.7088e-06, |
|
"loss": 0.411, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 1.7248e-06, |
|
"loss": 0.4365, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 1.7408e-06, |
|
"loss": 0.4423, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 1.7567999999999998e-06, |
|
"loss": 0.4762, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 1.7727999999999998e-06, |
|
"loss": 0.4218, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 1.7888e-06, |
|
"loss": 0.4694, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"learning_rate": 1.8047999999999999e-06, |
|
"loss": 0.4443, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 1.8207999999999998e-06, |
|
"loss": 0.4547, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 1.8368e-06, |
|
"loss": 0.4704, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 0.6065183877944946, |
|
"eval_runtime": 130.2735, |
|
"eval_samples_per_second": 4.084, |
|
"eval_steps_per_second": 1.021, |
|
"eval_wer": 0.31079256926142573, |
|
"step": 1152 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 1.8527999999999999e-06, |
|
"loss": 0.4844, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"learning_rate": 1.8688e-06, |
|
"loss": 0.4618, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 1.8848e-06, |
|
"loss": 0.4071, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"learning_rate": 1.9007999999999999e-06, |
|
"loss": 0.4612, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 1.9168e-06, |
|
"loss": 0.452, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"learning_rate": 1.9328e-06, |
|
"loss": 0.4601, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"learning_rate": 1.9488e-06, |
|
"loss": 0.4423, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"learning_rate": 1.9648e-06, |
|
"loss": 0.4362, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 1.9807999999999997e-06, |
|
"loss": 0.4682, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 1.9967999999999997e-06, |
|
"loss": 0.4773, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 2.0128e-06, |
|
"loss": 0.4693, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 2.0288e-06, |
|
"loss": 0.4618, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 2.0448e-06, |
|
"loss": 0.4496, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 4.48, |
|
"learning_rate": 2.0608e-06, |
|
"loss": 0.4575, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"learning_rate": 2.0767999999999998e-06, |
|
"loss": 0.4846, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"learning_rate": 2.0928e-06, |
|
"loss": 0.4672, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"learning_rate": 2.1088e-06, |
|
"loss": 0.468, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"learning_rate": 2.1248e-06, |
|
"loss": 0.4403, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"learning_rate": 2.1408e-06, |
|
"loss": 0.4849, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 4.69, |
|
"learning_rate": 2.1568e-06, |
|
"loss": 0.4618, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 2.1727999999999998e-06, |
|
"loss": 0.443, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 2.1888e-06, |
|
"loss": 0.4592, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 2.2048e-06, |
|
"loss": 0.443, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"learning_rate": 2.2208e-06, |
|
"loss": 0.4589, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 2.2368e-06, |
|
"loss": 0.4802, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 2.2528e-06, |
|
"loss": 0.4471, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 2.2688e-06, |
|
"loss": 0.444, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 2.2848e-06, |
|
"loss": 0.4299, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 2.3008e-06, |
|
"loss": 0.4854, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 0.6036490797996521, |
|
"eval_runtime": 131.0346, |
|
"eval_samples_per_second": 4.06, |
|
"eval_steps_per_second": 1.015, |
|
"eval_wer": 0.3112425274795912, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 5.03, |
|
"learning_rate": 2.3168e-06, |
|
"loss": 0.4727, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"learning_rate": 2.3328e-06, |
|
"loss": 0.4395, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 5.1, |
|
"learning_rate": 2.3488000000000003e-06, |
|
"loss": 0.428, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 2.3647999999999998e-06, |
|
"loss": 0.4543, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 5.17, |
|
"learning_rate": 2.3807999999999997e-06, |
|
"loss": 0.4907, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"learning_rate": 2.3967999999999997e-06, |
|
"loss": 0.4558, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 5.24, |
|
"learning_rate": 2.4127999999999996e-06, |
|
"loss": 0.4319, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 5.28, |
|
"learning_rate": 2.4288e-06, |
|
"loss": 0.4274, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 5.31, |
|
"learning_rate": 2.4448e-06, |
|
"loss": 0.4645, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 5.35, |
|
"learning_rate": 2.4608e-06, |
|
"loss": 0.4675, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 5.38, |
|
"learning_rate": 2.4767999999999997e-06, |
|
"loss": 0.4447, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 5.42, |
|
"learning_rate": 2.4927999999999997e-06, |
|
"loss": 0.4348, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 5.45, |
|
"learning_rate": 2.5087999999999996e-06, |
|
"loss": 0.4521, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 5.48, |
|
"learning_rate": 2.5248e-06, |
|
"loss": 0.4734, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 5.52, |
|
"learning_rate": 2.5408e-06, |
|
"loss": 0.4693, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 5.55, |
|
"learning_rate": 2.5568e-06, |
|
"loss": 0.4504, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 5.59, |
|
"learning_rate": 2.5727999999999997e-06, |
|
"loss": 0.4334, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 5.62, |
|
"learning_rate": 2.5887999999999997e-06, |
|
"loss": 0.4632, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 5.66, |
|
"learning_rate": 2.6048e-06, |
|
"loss": 0.4563, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 2.6208e-06, |
|
"loss": 0.4762, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 5.73, |
|
"learning_rate": 2.6368e-06, |
|
"loss": 0.464, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 5.76, |
|
"learning_rate": 2.6528e-06, |
|
"loss": 0.4757, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"learning_rate": 2.6687999999999998e-06, |
|
"loss": 0.4655, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 5.83, |
|
"learning_rate": 2.6848e-06, |
|
"loss": 0.472, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 5.87, |
|
"learning_rate": 2.7008e-06, |
|
"loss": 0.4678, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 5.9, |
|
"learning_rate": 2.7168e-06, |
|
"loss": 0.4486, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 5.94, |
|
"learning_rate": 2.7328e-06, |
|
"loss": 0.4481, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 5.97, |
|
"learning_rate": 2.7488e-06, |
|
"loss": 0.4918, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 0.6006949543952942, |
|
"eval_runtime": 131.3568, |
|
"eval_samples_per_second": 4.05, |
|
"eval_steps_per_second": 1.013, |
|
"eval_wer": 0.31484219322491486, |
|
"step": 1728 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 2.7648e-06, |
|
"loss": 0.5089, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 2.7808e-06, |
|
"loss": 0.4866, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 6.08, |
|
"learning_rate": 2.7968e-06, |
|
"loss": 0.4609, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 6.11, |
|
"learning_rate": 2.8128e-06, |
|
"loss": 0.4424, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 6.15, |
|
"learning_rate": 2.8288e-06, |
|
"loss": 0.4567, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 6.18, |
|
"learning_rate": 2.8448000000000003e-06, |
|
"loss": 0.4713, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 6.21, |
|
"learning_rate": 2.8607999999999998e-06, |
|
"loss": 0.4763, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 2.8767999999999997e-06, |
|
"loss": 0.412, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 6.28, |
|
"learning_rate": 2.8927999999999996e-06, |
|
"loss": 0.4248, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 6.32, |
|
"learning_rate": 2.9087999999999996e-06, |
|
"loss": 0.4873, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 6.35, |
|
"learning_rate": 2.9247999999999995e-06, |
|
"loss": 0.4969, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 6.39, |
|
"learning_rate": 2.9408e-06, |
|
"loss": 0.4437, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 6.42, |
|
"learning_rate": 2.9568e-06, |
|
"loss": 0.4533, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 6.46, |
|
"learning_rate": 2.9727999999999997e-06, |
|
"loss": 0.4637, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 6.49, |
|
"learning_rate": 2.9887999999999997e-06, |
|
"loss": 0.4406, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 3.0047999999999996e-06, |
|
"loss": 0.4835, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 6.56, |
|
"learning_rate": 3.0208e-06, |
|
"loss": 0.4592, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 3.0368e-06, |
|
"loss": 0.4705, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 6.63, |
|
"learning_rate": 3.0528e-06, |
|
"loss": 0.4672, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 6.66, |
|
"learning_rate": 3.0687999999999997e-06, |
|
"loss": 0.4663, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 6.7, |
|
"learning_rate": 3.0847999999999997e-06, |
|
"loss": 0.4574, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 6.73, |
|
"learning_rate": 3.1008e-06, |
|
"loss": 0.4277, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"learning_rate": 3.1168e-06, |
|
"loss": 0.463, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 6.8, |
|
"learning_rate": 3.1328e-06, |
|
"loss": 0.4274, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 6.84, |
|
"learning_rate": 3.1488e-06, |
|
"loss": 0.4742, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 6.87, |
|
"learning_rate": 3.1647999999999998e-06, |
|
"loss": 0.4731, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 6.91, |
|
"learning_rate": 3.1808e-06, |
|
"loss": 0.4362, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 6.94, |
|
"learning_rate": 3.1968e-06, |
|
"loss": 0.4304, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 6.98, |
|
"learning_rate": 3.2128e-06, |
|
"loss": 0.4549, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"eval_loss": 0.6039400696754456, |
|
"eval_runtime": 132.6852, |
|
"eval_samples_per_second": 4.009, |
|
"eval_steps_per_second": 1.002, |
|
"eval_wer": 0.30725718326155427, |
|
"step": 2016 |
|
}, |
|
{ |
|
"epoch": 7.01, |
|
"learning_rate": 3.2288e-06, |
|
"loss": 0.4989, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 7.05, |
|
"learning_rate": 3.2448e-06, |
|
"loss": 0.4714, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 7.08, |
|
"learning_rate": 3.2608e-06, |
|
"loss": 0.4323, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 7.12, |
|
"learning_rate": 3.2768e-06, |
|
"loss": 0.4592, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 7.15, |
|
"learning_rate": 3.2928e-06, |
|
"loss": 0.4588, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 7.19, |
|
"learning_rate": 3.3088e-06, |
|
"loss": 0.4579, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 7.22, |
|
"learning_rate": 3.3248e-06, |
|
"loss": 0.4626, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 7.26, |
|
"learning_rate": 3.3408000000000003e-06, |
|
"loss": 0.444, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 7.29, |
|
"learning_rate": 3.3567999999999998e-06, |
|
"loss": 0.4645, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 7.33, |
|
"learning_rate": 3.3727999999999997e-06, |
|
"loss": 0.4798, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 7.36, |
|
"learning_rate": 3.3887999999999996e-06, |
|
"loss": 0.4486, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 7.39, |
|
"learning_rate": 3.4047999999999996e-06, |
|
"loss": 0.4526, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 7.43, |
|
"learning_rate": 3.4207999999999995e-06, |
|
"loss": 0.4478, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 7.46, |
|
"learning_rate": 3.4368e-06, |
|
"loss": 0.4503, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 3.4528e-06, |
|
"loss": 0.4482, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 7.53, |
|
"learning_rate": 3.4687999999999997e-06, |
|
"loss": 0.4819, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 7.57, |
|
"learning_rate": 3.4847999999999997e-06, |
|
"loss": 0.4618, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 7.6, |
|
"learning_rate": 3.5007999999999996e-06, |
|
"loss": 0.4388, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 7.64, |
|
"learning_rate": 3.5168e-06, |
|
"loss": 0.461, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 7.67, |
|
"learning_rate": 3.5328e-06, |
|
"loss": 0.4777, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 7.71, |
|
"learning_rate": 3.5488e-06, |
|
"loss": 0.4705, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 7.74, |
|
"learning_rate": 3.5647999999999997e-06, |
|
"loss": 0.4377, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 7.78, |
|
"learning_rate": 3.5807999999999997e-06, |
|
"loss": 0.4403, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 7.81, |
|
"learning_rate": 3.5968e-06, |
|
"loss": 0.4308, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 7.85, |
|
"learning_rate": 3.6128e-06, |
|
"loss": 0.4447, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 7.88, |
|
"learning_rate": 3.6288e-06, |
|
"loss": 0.5049, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 7.91, |
|
"learning_rate": 3.6448e-06, |
|
"loss": 0.4638, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 7.95, |
|
"learning_rate": 3.6607999999999997e-06, |
|
"loss": 0.4377, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 7.98, |
|
"learning_rate": 3.6768e-06, |
|
"loss": 0.4546, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 0.612916111946106, |
|
"eval_runtime": 131.5991, |
|
"eval_samples_per_second": 4.043, |
|
"eval_steps_per_second": 1.011, |
|
"eval_wer": 0.30725718326155427, |
|
"step": 2304 |
|
}, |
|
{ |
|
"epoch": 8.02, |
|
"learning_rate": 3.6928e-06, |
|
"loss": 0.4812, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"learning_rate": 3.7088e-06, |
|
"loss": 0.4806, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 8.09, |
|
"learning_rate": 3.7248e-06, |
|
"loss": 0.4407, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 8.12, |
|
"learning_rate": 3.7408e-06, |
|
"loss": 0.4493, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 8.16, |
|
"learning_rate": 3.7567999999999998e-06, |
|
"loss": 0.4571, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 8.19, |
|
"learning_rate": 3.7728e-06, |
|
"loss": 0.4674, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 8.23, |
|
"learning_rate": 3.7888e-06, |
|
"loss": 0.4685, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 8.26, |
|
"learning_rate": 3.8048e-06, |
|
"loss": 0.4256, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 8.3, |
|
"learning_rate": 3.8208e-06, |
|
"loss": 0.4657, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 8.33, |
|
"learning_rate": 3.8368e-06, |
|
"loss": 0.4747, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 8.37, |
|
"learning_rate": 3.8528e-06, |
|
"loss": 0.4538, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 8.4, |
|
"learning_rate": 3.868799999999999e-06, |
|
"loss": 0.4635, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 8.44, |
|
"learning_rate": 3.8848e-06, |
|
"loss": 0.4257, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 8.47, |
|
"learning_rate": 3.9008e-06, |
|
"loss": 0.4537, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 8.51, |
|
"learning_rate": 3.9167999999999995e-06, |
|
"loss": 0.4674, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 8.54, |
|
"learning_rate": 3.9328e-06, |
|
"loss": 0.4263, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 8.57, |
|
"learning_rate": 3.948799999999999e-06, |
|
"loss": 0.4476, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 8.61, |
|
"learning_rate": 3.9648e-06, |
|
"loss": 0.4592, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 8.64, |
|
"learning_rate": 3.9808e-06, |
|
"loss": 0.4608, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 8.68, |
|
"learning_rate": 3.9968e-06, |
|
"loss": 0.4702, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 8.71, |
|
"learning_rate": 4.0128e-06, |
|
"loss": 0.4665, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"learning_rate": 4.0288e-06, |
|
"loss": 0.4524, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 8.78, |
|
"learning_rate": 4.0448e-06, |
|
"loss": 0.4461, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 8.82, |
|
"learning_rate": 4.0608e-06, |
|
"loss": 0.448, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 8.85, |
|
"learning_rate": 4.0768000000000005e-06, |
|
"loss": 0.4737, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 8.89, |
|
"learning_rate": 4.0928e-06, |
|
"loss": 0.4799, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 8.92, |
|
"learning_rate": 4.1087999999999995e-06, |
|
"loss": 0.4435, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 8.96, |
|
"learning_rate": 4.124799999999999e-06, |
|
"loss": 0.4626, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 8.99, |
|
"learning_rate": 4.140799999999999e-06, |
|
"loss": 0.4404, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"eval_loss": 0.6061944961547852, |
|
"eval_runtime": 132.6658, |
|
"eval_samples_per_second": 4.01, |
|
"eval_steps_per_second": 1.003, |
|
"eval_wer": 0.30539307064344023, |
|
"step": 2592 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 4.1568e-06, |
|
"loss": 0.4639, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"learning_rate": 4.172799999999999e-06, |
|
"loss": 0.4538, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 9.1, |
|
"learning_rate": 4.1888e-06, |
|
"loss": 0.4543, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 9.13, |
|
"learning_rate": 4.204799999999999e-06, |
|
"loss": 0.4562, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 9.17, |
|
"learning_rate": 4.2207999999999995e-06, |
|
"loss": 0.4686, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 9.2, |
|
"learning_rate": 4.2368e-06, |
|
"loss": 0.4693, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 9.24, |
|
"learning_rate": 4.252799999999999e-06, |
|
"loss": 0.4708, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 9.27, |
|
"learning_rate": 4.2688e-06, |
|
"loss": 0.4286, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 9.3, |
|
"learning_rate": 4.284799999999999e-06, |
|
"loss": 0.4662, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 9.34, |
|
"learning_rate": 4.3007999999999995e-06, |
|
"loss": 0.4749, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 9.37, |
|
"learning_rate": 4.3152e-06, |
|
"loss": 0.4793, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 9.41, |
|
"learning_rate": 4.3312e-06, |
|
"loss": 0.4502, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 9.44, |
|
"learning_rate": 4.3472e-06, |
|
"loss": 0.4208, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 9.48, |
|
"learning_rate": 4.3632e-06, |
|
"loss": 0.482, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 9.51, |
|
"learning_rate": 4.3791999999999995e-06, |
|
"loss": 0.4361, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 9.55, |
|
"learning_rate": 4.3952e-06, |
|
"loss": 0.4828, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 9.58, |
|
"learning_rate": 4.4112e-06, |
|
"loss": 0.4436, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 9.62, |
|
"learning_rate": 4.4272e-06, |
|
"loss": 0.4264, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 9.65, |
|
"learning_rate": 4.4432e-06, |
|
"loss": 0.4762, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 9.69, |
|
"learning_rate": 4.4592e-06, |
|
"loss": 0.4899, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 9.72, |
|
"learning_rate": 4.4752e-06, |
|
"loss": 0.4511, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 9.75, |
|
"learning_rate": 4.4912e-06, |
|
"loss": 0.4373, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 9.79, |
|
"learning_rate": 4.5072e-06, |
|
"loss": 0.432, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 9.82, |
|
"learning_rate": 4.5232e-06, |
|
"loss": 0.4703, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"learning_rate": 4.5392e-06, |
|
"loss": 0.4761, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 9.89, |
|
"learning_rate": 4.5552e-06, |
|
"loss": 0.4312, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 9.93, |
|
"learning_rate": 4.5711999999999995e-06, |
|
"loss": 0.4858, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 9.96, |
|
"learning_rate": 4.5872e-06, |
|
"loss": 0.4656, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 4.6032e-06, |
|
"loss": 0.4681, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 0.6062560081481934, |
|
"eval_runtime": 132.3245, |
|
"eval_samples_per_second": 4.02, |
|
"eval_steps_per_second": 1.005, |
|
"eval_wer": 0.3074500224979109, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 4.6192e-06, |
|
"loss": 0.4374, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 10.07, |
|
"learning_rate": 4.6352e-06, |
|
"loss": 0.4345, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 10.1, |
|
"learning_rate": 4.6512e-06, |
|
"loss": 0.4305, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 10.14, |
|
"learning_rate": 4.6672e-06, |
|
"loss": 0.4713, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 10.17, |
|
"learning_rate": 4.6832e-06, |
|
"loss": 0.4838, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 10.21, |
|
"learning_rate": 4.6992e-06, |
|
"loss": 0.4283, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 10.24, |
|
"learning_rate": 4.7152e-06, |
|
"loss": 0.4301, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 10.28, |
|
"learning_rate": 4.7312e-06, |
|
"loss": 0.4475, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 10.31, |
|
"learning_rate": 4.7472e-06, |
|
"loss": 0.4428, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 10.35, |
|
"learning_rate": 4.7632e-06, |
|
"loss": 0.4786, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 10.38, |
|
"learning_rate": 4.7792e-06, |
|
"loss": 0.4545, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 10.42, |
|
"learning_rate": 4.7952e-06, |
|
"loss": 0.4996, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 10.45, |
|
"learning_rate": 4.8112e-06, |
|
"loss": 0.4714, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 10.48, |
|
"learning_rate": 4.8272e-06, |
|
"loss": 0.4784, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 10.52, |
|
"learning_rate": 4.8432000000000005e-06, |
|
"loss": 0.5106, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 10.55, |
|
"learning_rate": 4.8592e-06, |
|
"loss": 0.4746, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 10.59, |
|
"learning_rate": 4.8752e-06, |
|
"loss": 0.4775, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 10.62, |
|
"learning_rate": 4.8912e-06, |
|
"loss": 0.4351, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 10.66, |
|
"learning_rate": 4.907199999999999e-06, |
|
"loss": 0.4612, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 10.69, |
|
"learning_rate": 4.9232e-06, |
|
"loss": 0.4428, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 10.73, |
|
"learning_rate": 4.939199999999999e-06, |
|
"loss": 0.4598, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 10.76, |
|
"learning_rate": 4.9552e-06, |
|
"loss": 0.4646, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 10.8, |
|
"learning_rate": 4.971199999999999e-06, |
|
"loss": 0.4778, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 10.83, |
|
"learning_rate": 4.9871999999999994e-06, |
|
"loss": 0.4717, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 10.87, |
|
"learning_rate": 5.0032e-06, |
|
"loss": 0.4726, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 10.9, |
|
"learning_rate": 5.019199999999999e-06, |
|
"loss": 0.4781, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 10.94, |
|
"learning_rate": 5.0352e-06, |
|
"loss": 0.4336, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 10.97, |
|
"learning_rate": 5.051199999999999e-06, |
|
"loss": 0.469, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"eval_loss": 0.5881296992301941, |
|
"eval_runtime": 132.5378, |
|
"eval_samples_per_second": 4.014, |
|
"eval_steps_per_second": 1.003, |
|
"eval_wer": 0.30307899980716074, |
|
"step": 3168 |
|
}, |
|
{ |
|
"epoch": 11.01, |
|
"learning_rate": 5.0671999999999995e-06, |
|
"loss": 0.5434, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 5.0832e-06, |
|
"loss": 0.4576, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 11.08, |
|
"learning_rate": 5.099199999999999e-06, |
|
"loss": 0.4659, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 11.11, |
|
"learning_rate": 5.1152e-06, |
|
"loss": 0.4789, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 11.15, |
|
"learning_rate": 5.131199999999999e-06, |
|
"loss": 0.488, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 11.18, |
|
"learning_rate": 5.1472e-06, |
|
"loss": 0.5179, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 11.21, |
|
"learning_rate": 5.1632e-06, |
|
"loss": 0.5029, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 11.25, |
|
"learning_rate": 5.1791999999999995e-06, |
|
"loss": 0.5185, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 11.28, |
|
"learning_rate": 5.1952e-06, |
|
"loss": 0.4456, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 11.32, |
|
"learning_rate": 5.211199999999999e-06, |
|
"loss": 0.4869, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 11.35, |
|
"learning_rate": 5.2272e-06, |
|
"loss": 0.5245, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 11.39, |
|
"learning_rate": 5.2432e-06, |
|
"loss": 0.5119, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 11.42, |
|
"learning_rate": 5.2591999999999996e-06, |
|
"loss": 0.4462, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 11.46, |
|
"learning_rate": 5.2752e-06, |
|
"loss": 0.4932, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 11.49, |
|
"learning_rate": 5.291199999999999e-06, |
|
"loss": 0.5009, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 11.53, |
|
"learning_rate": 5.3072e-06, |
|
"loss": 0.5098, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 11.56, |
|
"learning_rate": 5.323199999999999e-06, |
|
"loss": 0.4766, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 11.6, |
|
"learning_rate": 5.3392e-06, |
|
"loss": 0.4616, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 11.63, |
|
"learning_rate": 5.3552e-06, |
|
"loss": 0.489, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 11.66, |
|
"learning_rate": 5.3711999999999995e-06, |
|
"loss": 0.4912, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 11.7, |
|
"learning_rate": 5.3872e-06, |
|
"loss": 0.5002, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 11.73, |
|
"learning_rate": 5.403199999999999e-06, |
|
"loss": 0.4864, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 11.77, |
|
"learning_rate": 5.4192e-06, |
|
"loss": 0.4808, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 11.8, |
|
"learning_rate": 5.4352e-06, |
|
"loss": 0.5101, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 11.84, |
|
"learning_rate": 5.4512e-06, |
|
"loss": 0.4796, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 11.87, |
|
"learning_rate": 5.4672e-06, |
|
"loss": 0.4847, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 11.91, |
|
"learning_rate": 5.4831999999999994e-06, |
|
"loss": 0.4861, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 11.94, |
|
"learning_rate": 5.4992e-06, |
|
"loss": 0.483, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 11.98, |
|
"learning_rate": 5.5152e-06, |
|
"loss": 0.4903, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_loss": 0.5912908911705017, |
|
"eval_runtime": 131.3368, |
|
"eval_samples_per_second": 4.051, |
|
"eval_steps_per_second": 1.013, |
|
"eval_wer": 0.304685993443466, |
|
"step": 3456 |
|
}, |
|
{ |
|
"epoch": 12.01, |
|
"learning_rate": 5.5312e-06, |
|
"loss": 0.5251, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 12.05, |
|
"learning_rate": 5.5472e-06, |
|
"loss": 0.505, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 12.08, |
|
"learning_rate": 5.5631999999999995e-06, |
|
"loss": 0.4575, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 12.12, |
|
"learning_rate": 5.5792e-06, |
|
"loss": 0.4521, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 12.15, |
|
"learning_rate": 5.5952e-06, |
|
"loss": 0.4952, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 12.19, |
|
"learning_rate": 5.6112e-06, |
|
"loss": 0.4898, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 12.22, |
|
"learning_rate": 5.6272e-06, |
|
"loss": 0.484, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 12.26, |
|
"learning_rate": 5.6432e-06, |
|
"loss": 0.4785, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 12.29, |
|
"learning_rate": 5.6592e-06, |
|
"loss": 0.487, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 12.33, |
|
"learning_rate": 5.6752e-06, |
|
"loss": 0.4844, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 12.36, |
|
"learning_rate": 5.6912e-06, |
|
"loss": 0.5375, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 12.39, |
|
"learning_rate": 5.7072e-06, |
|
"loss": 0.4902, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 12.43, |
|
"learning_rate": 5.7232e-06, |
|
"loss": 0.4738, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 12.46, |
|
"learning_rate": 5.7392e-06, |
|
"loss": 0.4767, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 12.5, |
|
"learning_rate": 5.7552e-06, |
|
"loss": 0.4886, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 12.53, |
|
"learning_rate": 5.7712e-06, |
|
"loss": 0.4927, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 12.57, |
|
"learning_rate": 5.7872e-06, |
|
"loss": 0.4633, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 12.6, |
|
"learning_rate": 5.8032e-06, |
|
"loss": 0.4881, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 12.64, |
|
"learning_rate": 5.8192e-06, |
|
"loss": 0.4908, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 12.67, |
|
"learning_rate": 5.8352000000000005e-06, |
|
"loss": 0.4823, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 12.71, |
|
"learning_rate": 5.8512e-06, |
|
"loss": 0.4796, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 12.74, |
|
"learning_rate": 5.8672e-06, |
|
"loss": 0.5192, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 12.78, |
|
"learning_rate": 5.8832e-06, |
|
"loss": 0.4772, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 12.81, |
|
"learning_rate": 5.8992e-06, |
|
"loss": 0.4726, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 12.85, |
|
"learning_rate": 5.9152e-06, |
|
"loss": 0.5245, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 12.88, |
|
"learning_rate": 5.931199999999999e-06, |
|
"loss": 0.5215, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 12.91, |
|
"learning_rate": 5.9471999999999996e-06, |
|
"loss": 0.4807, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 12.95, |
|
"learning_rate": 5.963199999999999e-06, |
|
"loss": 0.4689, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 12.98, |
|
"learning_rate": 5.9791999999999994e-06, |
|
"loss": 0.4677, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"eval_loss": 0.5920921564102173, |
|
"eval_runtime": 131.4646, |
|
"eval_samples_per_second": 4.047, |
|
"eval_steps_per_second": 1.012, |
|
"eval_wer": 0.30545735038889243, |
|
"step": 3744 |
|
}, |
|
{ |
|
"epoch": 13.02, |
|
"learning_rate": 5.9952e-06, |
|
"loss": 0.523, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 13.06, |
|
"learning_rate": 6.011199999999999e-06, |
|
"loss": 0.4964, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 13.09, |
|
"learning_rate": 6.0272e-06, |
|
"loss": 0.4605, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 13.12, |
|
"learning_rate": 6.043199999999999e-06, |
|
"loss": 0.4585, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 13.16, |
|
"learning_rate": 6.0591999999999995e-06, |
|
"loss": 0.4852, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 13.19, |
|
"learning_rate": 6.075199999999999e-06, |
|
"loss": 0.5152, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 13.23, |
|
"learning_rate": 6.091199999999999e-06, |
|
"loss": 0.4635, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 13.26, |
|
"learning_rate": 6.1072e-06, |
|
"loss": 0.4971, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 13.3, |
|
"learning_rate": 6.123199999999999e-06, |
|
"loss": 0.4915, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 13.33, |
|
"learning_rate": 6.1392e-06, |
|
"loss": 0.4639, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 13.37, |
|
"learning_rate": 6.155199999999999e-06, |
|
"loss": 0.4812, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 13.4, |
|
"learning_rate": 6.1711999999999995e-06, |
|
"loss": 0.5318, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 13.44, |
|
"learning_rate": 6.1872e-06, |
|
"loss": 0.4435, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 13.47, |
|
"learning_rate": 6.203199999999999e-06, |
|
"loss": 0.485, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 13.51, |
|
"learning_rate": 6.2192e-06, |
|
"loss": 0.522, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 13.54, |
|
"learning_rate": 6.235199999999999e-06, |
|
"loss": 0.5032, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 13.57, |
|
"learning_rate": 6.2511999999999995e-06, |
|
"loss": 0.4688, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 13.61, |
|
"learning_rate": 6.2672e-06, |
|
"loss": 0.477, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 13.64, |
|
"learning_rate": 6.283199999999999e-06, |
|
"loss": 0.4973, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 13.68, |
|
"learning_rate": 6.2992e-06, |
|
"loss": 0.4963, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 13.71, |
|
"learning_rate": 6.315199999999999e-06, |
|
"loss": 0.482, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 13.75, |
|
"learning_rate": 6.3312e-06, |
|
"loss": 0.4992, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 13.78, |
|
"learning_rate": 6.3472e-06, |
|
"loss": 0.4664, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 13.82, |
|
"learning_rate": 6.3631999999999995e-06, |
|
"loss": 0.5029, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 13.85, |
|
"learning_rate": 6.3792e-06, |
|
"loss": 0.4791, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 13.89, |
|
"learning_rate": 6.395199999999999e-06, |
|
"loss": 0.4984, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 13.92, |
|
"learning_rate": 6.4112e-06, |
|
"loss": 0.4782, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 13.96, |
|
"learning_rate": 6.4272e-06, |
|
"loss": 0.4688, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 13.99, |
|
"learning_rate": 6.4431999999999996e-06, |
|
"loss": 0.502, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_loss": 0.5905393362045288, |
|
"eval_runtime": 132.5982, |
|
"eval_samples_per_second": 4.012, |
|
"eval_steps_per_second": 1.003, |
|
"eval_wer": 0.3042360352253005, |
|
"step": 4032 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 6.4592e-06, |
|
"loss": 0.5202, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 14.06, |
|
"learning_rate": 6.4751999999999994e-06, |
|
"loss": 0.495, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 14.1, |
|
"learning_rate": 6.4912e-06, |
|
"loss": 0.4539, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 14.13, |
|
"learning_rate": 6.5072e-06, |
|
"loss": 0.4991, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 14.17, |
|
"learning_rate": 6.5232e-06, |
|
"loss": 0.4867, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 14.2, |
|
"learning_rate": 6.5392e-06, |
|
"loss": 0.4813, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 14.24, |
|
"learning_rate": 6.5551999999999995e-06, |
|
"loss": 0.4829, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 14.27, |
|
"learning_rate": 6.5712e-06, |
|
"loss": 0.4749, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 14.3, |
|
"learning_rate": 6.5872e-06, |
|
"loss": 0.4711, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 14.34, |
|
"learning_rate": 6.6032e-06, |
|
"loss": 0.4772, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 14.37, |
|
"learning_rate": 6.6192e-06, |
|
"loss": 0.4815, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 14.41, |
|
"learning_rate": 6.6352e-06, |
|
"loss": 0.4601, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 14.44, |
|
"learning_rate": 6.6512e-06, |
|
"loss": 0.4428, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 14.48, |
|
"learning_rate": 6.6672e-06, |
|
"loss": 0.4912, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 14.51, |
|
"learning_rate": 6.6832e-06, |
|
"loss": 0.49, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 14.55, |
|
"learning_rate": 6.6992e-06, |
|
"loss": 0.4895, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 14.58, |
|
"learning_rate": 6.7152e-06, |
|
"loss": 0.5205, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 14.62, |
|
"learning_rate": 6.7312e-06, |
|
"loss": 0.4868, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 14.65, |
|
"learning_rate": 6.7472e-06, |
|
"loss": 0.4892, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 14.69, |
|
"learning_rate": 6.7632e-06, |
|
"loss": 0.5081, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 14.72, |
|
"learning_rate": 6.7792e-06, |
|
"loss": 0.4815, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 14.75, |
|
"learning_rate": 6.7952e-06, |
|
"loss": 0.4592, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 14.79, |
|
"learning_rate": 6.8112e-06, |
|
"loss": 0.4593, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 14.82, |
|
"learning_rate": 6.8272000000000005e-06, |
|
"loss": 0.4962, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 14.86, |
|
"learning_rate": 6.8432e-06, |
|
"loss": 0.5118, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 14.89, |
|
"learning_rate": 6.8592e-06, |
|
"loss": 0.4986, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 14.93, |
|
"learning_rate": 6.8752e-06, |
|
"loss": 0.4582, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 14.96, |
|
"learning_rate": 6.8912e-06, |
|
"loss": 0.4658, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 6.907199999999999e-06, |
|
"loss": 0.5028, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_loss": 0.5988514423370361, |
|
"eval_runtime": 132.6201, |
|
"eval_samples_per_second": 4.011, |
|
"eval_steps_per_second": 1.003, |
|
"eval_wer": 0.3087998971524073, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 15.03, |
|
"learning_rate": 6.923199999999999e-06, |
|
"loss": 0.5183, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 15.07, |
|
"learning_rate": 6.9391999999999996e-06, |
|
"loss": 0.4793, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 15.1, |
|
"learning_rate": 6.955199999999999e-06, |
|
"loss": 0.4721, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 15.14, |
|
"learning_rate": 6.9711999999999994e-06, |
|
"loss": 0.5019, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 15.17, |
|
"learning_rate": 6.987199999999999e-06, |
|
"loss": 0.5078, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 15.21, |
|
"learning_rate": 7.003199999999999e-06, |
|
"loss": 0.4837, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 15.24, |
|
"learning_rate": 7.0192e-06, |
|
"loss": 0.4749, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 15.28, |
|
"learning_rate": 7.035199999999999e-06, |
|
"loss": 0.4345, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 15.31, |
|
"learning_rate": 7.0511999999999995e-06, |
|
"loss": 0.4767, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 15.35, |
|
"learning_rate": 7.067199999999999e-06, |
|
"loss": 0.4765, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 15.38, |
|
"learning_rate": 7.083199999999999e-06, |
|
"loss": 0.4711, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 15.42, |
|
"learning_rate": 7.0992e-06, |
|
"loss": 0.4657, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 15.45, |
|
"learning_rate": 7.115199999999999e-06, |
|
"loss": 0.4828, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 15.48, |
|
"learning_rate": 7.1312e-06, |
|
"loss": 0.4796, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 15.52, |
|
"learning_rate": 7.147199999999999e-06, |
|
"loss": 0.5007, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 15.55, |
|
"learning_rate": 7.1631999999999995e-06, |
|
"loss": 0.4637, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 15.59, |
|
"learning_rate": 7.1792e-06, |
|
"loss": 0.4844, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 15.62, |
|
"learning_rate": 7.195199999999999e-06, |
|
"loss": 0.4453, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 15.66, |
|
"learning_rate": 7.2112e-06, |
|
"loss": 0.4787, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 15.69, |
|
"learning_rate": 7.227199999999999e-06, |
|
"loss": 0.4795, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 15.73, |
|
"learning_rate": 7.2431999999999995e-06, |
|
"loss": 0.4906, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 15.76, |
|
"learning_rate": 7.2592e-06, |
|
"loss": 0.4923, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 15.8, |
|
"learning_rate": 7.275199999999999e-06, |
|
"loss": 0.4904, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 15.83, |
|
"learning_rate": 7.2912e-06, |
|
"loss": 0.4902, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 15.87, |
|
"learning_rate": 7.307199999999999e-06, |
|
"loss": 0.5582, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 15.9, |
|
"learning_rate": 7.3232e-06, |
|
"loss": 0.495, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 15.94, |
|
"learning_rate": 7.3392e-06, |
|
"loss": 0.4784, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 15.97, |
|
"learning_rate": 7.3551999999999995e-06, |
|
"loss": 0.4706, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_loss": 0.5664656758308411, |
|
"eval_runtime": 134.177, |
|
"eval_samples_per_second": 3.965, |
|
"eval_steps_per_second": 0.991, |
|
"eval_wer": 0.30655010606158, |
|
"step": 4608 |
|
}, |
|
{ |
|
"epoch": 16.01, |
|
"learning_rate": 7.3712e-06, |
|
"loss": 0.5272, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 16.04, |
|
"learning_rate": 7.387199999999999e-06, |
|
"loss": 0.4749, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 16.08, |
|
"learning_rate": 7.4032e-06, |
|
"loss": 0.4663, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 16.11, |
|
"learning_rate": 7.4192e-06, |
|
"loss": 0.4622, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 16.15, |
|
"learning_rate": 7.4351999999999996e-06, |
|
"loss": 0.487, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 16.18, |
|
"learning_rate": 7.4512e-06, |
|
"loss": 0.4853, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 16.21, |
|
"learning_rate": 7.467199999999999e-06, |
|
"loss": 0.4826, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 16.25, |
|
"learning_rate": 7.4832e-06, |
|
"loss": 0.4536, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 16.28, |
|
"learning_rate": 7.4992e-06, |
|
"loss": 0.4747, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 16.32, |
|
"learning_rate": 7.5152e-06, |
|
"loss": 0.4925, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 16.35, |
|
"learning_rate": 7.5312e-06, |
|
"loss": 0.5197, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 16.39, |
|
"learning_rate": 7.5471999999999995e-06, |
|
"loss": 0.481, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 16.42, |
|
"learning_rate": 7.5632e-06, |
|
"loss": 0.4855, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 16.46, |
|
"learning_rate": 7.5792e-06, |
|
"loss": 0.4839, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 16.49, |
|
"learning_rate": 7.5952e-06, |
|
"loss": 0.4653, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 16.53, |
|
"learning_rate": 7.6112e-06, |
|
"loss": 0.5252, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 16.56, |
|
"learning_rate": 7.6272e-06, |
|
"loss": 0.4747, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 16.6, |
|
"learning_rate": 7.6432e-06, |
|
"loss": 0.4679, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 16.63, |
|
"learning_rate": 7.6592e-06, |
|
"loss": 0.4705, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 16.66, |
|
"learning_rate": 7.6752e-06, |
|
"loss": 0.4689, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 16.7, |
|
"learning_rate": 7.6912e-06, |
|
"loss": 0.4694, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 16.73, |
|
"learning_rate": 7.7072e-06, |
|
"loss": 0.5155, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 16.77, |
|
"learning_rate": 7.7232e-06, |
|
"loss": 0.5148, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 16.8, |
|
"learning_rate": 7.7392e-06, |
|
"loss": 0.493, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 16.84, |
|
"learning_rate": 7.7552e-06, |
|
"loss": 0.5057, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 16.87, |
|
"learning_rate": 7.7712e-06, |
|
"loss": 0.4833, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 16.91, |
|
"learning_rate": 7.7872e-06, |
|
"loss": 0.4429, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 16.94, |
|
"learning_rate": 7.8032e-06, |
|
"loss": 0.4557, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 16.98, |
|
"learning_rate": 7.8192e-06, |
|
"loss": 0.4839, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_loss": 0.6003334522247314, |
|
"eval_runtime": 132.3698, |
|
"eval_samples_per_second": 4.019, |
|
"eval_steps_per_second": 1.005, |
|
"eval_wer": 0.3111139679886868, |
|
"step": 4896 |
|
}, |
|
{ |
|
"epoch": 17.01, |
|
"learning_rate": 7.8352e-06, |
|
"loss": 0.5301, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 17.05, |
|
"learning_rate": 7.8512e-06, |
|
"loss": 0.4963, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 17.08, |
|
"learning_rate": 7.8672e-06, |
|
"loss": 0.4488, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 17.12, |
|
"learning_rate": 7.8832e-06, |
|
"loss": 0.4498, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 17.15, |
|
"learning_rate": 7.8992e-06, |
|
"loss": 0.4564, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 17.19, |
|
"learning_rate": 7.9152e-06, |
|
"loss": 0.47, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 17.22, |
|
"learning_rate": 7.9312e-06, |
|
"loss": 0.4789, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 17.26, |
|
"learning_rate": 7.947199999999999e-06, |
|
"loss": 0.4687, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 17.29, |
|
"learning_rate": 7.963199999999999e-06, |
|
"loss": 0.4753, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 17.33, |
|
"learning_rate": 7.9792e-06, |
|
"loss": 0.4691, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 17.36, |
|
"learning_rate": 7.9952e-06, |
|
"loss": 0.4769, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 17.39, |
|
"learning_rate": 7.994042553191488e-06, |
|
"loss": 0.4754, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 17.43, |
|
"learning_rate": 7.985531914893617e-06, |
|
"loss": 0.4758, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 17.46, |
|
"learning_rate": 7.977021276595745e-06, |
|
"loss": 0.4789, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 17.5, |
|
"learning_rate": 7.968510638297872e-06, |
|
"loss": 0.5191, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 17.53, |
|
"learning_rate": 7.96e-06, |
|
"loss": 0.4745, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 17.57, |
|
"learning_rate": 7.951489361702127e-06, |
|
"loss": 0.4778, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 17.6, |
|
"learning_rate": 7.942978723404255e-06, |
|
"loss": 0.474, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 17.64, |
|
"learning_rate": 7.934468085106383e-06, |
|
"loss": 0.4794, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 17.67, |
|
"learning_rate": 7.92595744680851e-06, |
|
"loss": 0.4932, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 17.71, |
|
"learning_rate": 7.917446808510638e-06, |
|
"loss": 0.5037, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 17.74, |
|
"learning_rate": 7.908936170212767e-06, |
|
"loss": 0.4708, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 17.78, |
|
"learning_rate": 7.900425531914893e-06, |
|
"loss": 0.4446, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 17.81, |
|
"learning_rate": 7.891914893617021e-06, |
|
"loss": 0.4956, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 17.85, |
|
"learning_rate": 7.883404255319148e-06, |
|
"loss": 0.4956, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 17.88, |
|
"learning_rate": 7.874893617021276e-06, |
|
"loss": 0.4907, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 17.91, |
|
"learning_rate": 7.866382978723405e-06, |
|
"loss": 0.4608, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 17.95, |
|
"learning_rate": 7.857872340425531e-06, |
|
"loss": 0.4997, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 17.98, |
|
"learning_rate": 7.849361702127658e-06, |
|
"loss": 0.4733, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"eval_loss": 0.5936566591262817, |
|
"eval_runtime": 133.7823, |
|
"eval_samples_per_second": 3.977, |
|
"eval_steps_per_second": 0.994, |
|
"eval_wer": 0.30385035675258726, |
|
"step": 5184 |
|
}, |
|
{ |
|
"epoch": 18.02, |
|
"learning_rate": 7.840851063829786e-06, |
|
"loss": 0.5199, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 18.06, |
|
"learning_rate": 7.832340425531915e-06, |
|
"loss": 0.4623, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 18.09, |
|
"learning_rate": 7.823829787234041e-06, |
|
"loss": 0.4811, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 18.12, |
|
"learning_rate": 7.81531914893617e-06, |
|
"loss": 0.4668, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 18.16, |
|
"learning_rate": 7.806808510638298e-06, |
|
"loss": 0.4912, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 18.19, |
|
"learning_rate": 7.798297872340425e-06, |
|
"loss": 0.4839, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 18.23, |
|
"learning_rate": 7.789787234042553e-06, |
|
"loss": 0.4634, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 18.26, |
|
"learning_rate": 7.781276595744681e-06, |
|
"loss": 0.4374, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 18.3, |
|
"learning_rate": 7.772765957446808e-06, |
|
"loss": 0.4947, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 18.33, |
|
"learning_rate": 7.764255319148936e-06, |
|
"loss": 0.5125, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 18.37, |
|
"learning_rate": 7.755744680851063e-06, |
|
"loss": 0.5111, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 18.4, |
|
"learning_rate": 7.747234042553191e-06, |
|
"loss": 0.4681, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 18.44, |
|
"learning_rate": 7.73872340425532e-06, |
|
"loss": 0.4393, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 18.47, |
|
"learning_rate": 7.730212765957446e-06, |
|
"loss": 0.4822, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 18.51, |
|
"learning_rate": 7.721702127659574e-06, |
|
"loss": 0.4626, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 18.54, |
|
"learning_rate": 7.713191489361703e-06, |
|
"loss": 0.5142, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 18.57, |
|
"learning_rate": 7.70468085106383e-06, |
|
"loss": 0.4912, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 18.61, |
|
"learning_rate": 7.696170212765958e-06, |
|
"loss": 0.4883, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 18.64, |
|
"learning_rate": 7.687659574468084e-06, |
|
"loss": 0.4442, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 18.68, |
|
"learning_rate": 7.679148936170212e-06, |
|
"loss": 0.4706, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 18.71, |
|
"learning_rate": 7.67063829787234e-06, |
|
"loss": 0.4691, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 18.75, |
|
"learning_rate": 7.662127659574467e-06, |
|
"loss": 0.4764, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 18.78, |
|
"learning_rate": 7.653617021276596e-06, |
|
"loss": 0.4469, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 18.82, |
|
"learning_rate": 7.645106382978724e-06, |
|
"loss": 0.4961, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 18.85, |
|
"learning_rate": 7.63659574468085e-06, |
|
"loss": 0.5096, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 18.89, |
|
"learning_rate": 7.628085106382978e-06, |
|
"loss": 0.5096, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 18.92, |
|
"learning_rate": 7.6195744680851065e-06, |
|
"loss": 0.474, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 18.96, |
|
"learning_rate": 7.611063829787234e-06, |
|
"loss": 0.4563, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 18.99, |
|
"learning_rate": 7.602553191489361e-06, |
|
"loss": 0.4544, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"eval_loss": 0.5903141498565674, |
|
"eval_runtime": 132.6544, |
|
"eval_samples_per_second": 4.01, |
|
"eval_steps_per_second": 1.003, |
|
"eval_wer": 0.3025004820980909, |
|
"step": 5472 |
|
}, |
|
{ |
|
"epoch": 19.03, |
|
"learning_rate": 7.594042553191489e-06, |
|
"loss": 0.5091, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 19.06, |
|
"learning_rate": 7.585531914893617e-06, |
|
"loss": 0.4623, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 19.1, |
|
"learning_rate": 7.577021276595745e-06, |
|
"loss": 0.4279, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 19.13, |
|
"learning_rate": 7.568510638297872e-06, |
|
"loss": 0.4936, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 19.17, |
|
"learning_rate": 7.56e-06, |
|
"loss": 0.4928, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 19.2, |
|
"learning_rate": 7.551489361702128e-06, |
|
"loss": 0.4678, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 19.24, |
|
"learning_rate": 7.542978723404255e-06, |
|
"loss": 0.5013, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 19.27, |
|
"learning_rate": 7.534468085106382e-06, |
|
"loss": 0.467, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 19.3, |
|
"learning_rate": 7.5259574468085095e-06, |
|
"loss": 0.4751, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 19.34, |
|
"learning_rate": 7.517446808510638e-06, |
|
"loss": 0.4792, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 19.37, |
|
"learning_rate": 7.508936170212765e-06, |
|
"loss": 0.4714, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 19.41, |
|
"learning_rate": 7.500425531914893e-06, |
|
"loss": 0.4475, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 19.44, |
|
"learning_rate": 7.491914893617021e-06, |
|
"loss": 0.4478, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 19.48, |
|
"learning_rate": 7.4834042553191485e-06, |
|
"loss": 0.4516, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 19.51, |
|
"learning_rate": 7.474893617021276e-06, |
|
"loss": 0.4917, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 19.55, |
|
"learning_rate": 7.4663829787234035e-06, |
|
"loss": 0.4799, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 19.58, |
|
"learning_rate": 7.457872340425532e-06, |
|
"loss": 0.4615, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 19.62, |
|
"learning_rate": 7.449361702127659e-06, |
|
"loss": 0.4605, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 19.65, |
|
"learning_rate": 7.440851063829787e-06, |
|
"loss": 0.4824, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 19.69, |
|
"learning_rate": 7.432340425531914e-06, |
|
"loss": 0.4775, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 19.72, |
|
"learning_rate": 7.424680851063829e-06, |
|
"loss": 0.4762, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 19.75, |
|
"learning_rate": 7.416170212765957e-06, |
|
"loss": 0.4654, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 19.79, |
|
"learning_rate": 7.407659574468085e-06, |
|
"loss": 0.4472, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 19.82, |
|
"learning_rate": 7.399148936170213e-06, |
|
"loss": 0.4971, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 19.86, |
|
"learning_rate": 7.39063829787234e-06, |
|
"loss": 0.5079, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 19.89, |
|
"learning_rate": 7.382127659574468e-06, |
|
"loss": 0.5002, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 19.93, |
|
"learning_rate": 7.373617021276596e-06, |
|
"loss": 0.4706, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 19.96, |
|
"learning_rate": 7.365106382978723e-06, |
|
"loss": 0.458, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"learning_rate": 7.356595744680851e-06, |
|
"loss": 0.4616, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_loss": 0.6064006686210632, |
|
"eval_runtime": 132.8156, |
|
"eval_samples_per_second": 4.006, |
|
"eval_steps_per_second": 1.001, |
|
"eval_wer": 0.2967795847528444, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 20.03, |
|
"learning_rate": 7.348085106382979e-06, |
|
"loss": 0.4895, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 20.07, |
|
"learning_rate": 7.339574468085106e-06, |
|
"loss": 0.4836, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 20.1, |
|
"learning_rate": 7.331063829787233e-06, |
|
"loss": 0.4503, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 20.14, |
|
"learning_rate": 7.322553191489361e-06, |
|
"loss": 0.4281, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 20.17, |
|
"learning_rate": 7.314042553191489e-06, |
|
"loss": 0.4656, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 20.21, |
|
"learning_rate": 7.3055319148936165e-06, |
|
"loss": 0.4534, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 20.24, |
|
"learning_rate": 7.297021276595744e-06, |
|
"loss": 0.4536, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 20.28, |
|
"learning_rate": 7.2885106382978714e-06, |
|
"loss": 0.4409, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 20.31, |
|
"learning_rate": 7.28e-06, |
|
"loss": 0.4836, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 20.35, |
|
"learning_rate": 7.271489361702127e-06, |
|
"loss": 0.4729, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 20.38, |
|
"learning_rate": 7.262978723404255e-06, |
|
"loss": 0.4831, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 20.42, |
|
"learning_rate": 7.254468085106383e-06, |
|
"loss": 0.4458, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 20.45, |
|
"learning_rate": 7.2459574468085105e-06, |
|
"loss": 0.4617, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 20.48, |
|
"learning_rate": 7.237446808510638e-06, |
|
"loss": 0.4805, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 20.52, |
|
"learning_rate": 7.2289361702127654e-06, |
|
"loss": 0.4856, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 20.55, |
|
"learning_rate": 7.220425531914894e-06, |
|
"loss": 0.4696, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 20.59, |
|
"learning_rate": 7.211914893617021e-06, |
|
"loss": 0.4622, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 20.62, |
|
"learning_rate": 7.203404255319149e-06, |
|
"loss": 0.4636, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 20.66, |
|
"learning_rate": 7.194893617021276e-06, |
|
"loss": 0.4748, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 20.69, |
|
"learning_rate": 7.1863829787234045e-06, |
|
"loss": 0.4811, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 20.73, |
|
"learning_rate": 7.177872340425532e-06, |
|
"loss": 0.4762, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 20.76, |
|
"learning_rate": 7.169361702127659e-06, |
|
"loss": 0.4812, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 20.8, |
|
"learning_rate": 7.160851063829787e-06, |
|
"loss": 0.4609, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 20.83, |
|
"learning_rate": 7.152340425531915e-06, |
|
"loss": 0.4819, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 20.87, |
|
"learning_rate": 7.143829787234043e-06, |
|
"loss": 0.5133, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 20.9, |
|
"learning_rate": 7.135319148936169e-06, |
|
"loss": 0.4576, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 20.94, |
|
"learning_rate": 7.126808510638297e-06, |
|
"loss": 0.4457, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 20.97, |
|
"learning_rate": 7.118297872340425e-06, |
|
"loss": 0.475, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 21.0, |
|
"eval_loss": 0.588329553604126, |
|
"eval_runtime": 133.391, |
|
"eval_samples_per_second": 3.988, |
|
"eval_steps_per_second": 0.997, |
|
"eval_wer": 0.29600822780741787, |
|
"step": 6048 |
|
}, |
|
{ |
|
"epoch": 21.01, |
|
"learning_rate": 7.1097872340425526e-06, |
|
"loss": 0.5207, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 21.04, |
|
"learning_rate": 7.10127659574468e-06, |
|
"loss": 0.46, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 21.08, |
|
"learning_rate": 7.092765957446808e-06, |
|
"loss": 0.4391, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 21.11, |
|
"learning_rate": 7.084255319148936e-06, |
|
"loss": 0.4726, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 21.15, |
|
"learning_rate": 7.075744680851063e-06, |
|
"loss": 0.4933, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 21.18, |
|
"learning_rate": 7.067234042553191e-06, |
|
"loss": 0.5056, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 21.21, |
|
"learning_rate": 7.058723404255319e-06, |
|
"loss": 0.4792, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 21.25, |
|
"learning_rate": 7.0502127659574465e-06, |
|
"loss": 0.4418, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 21.28, |
|
"learning_rate": 7.041702127659574e-06, |
|
"loss": 0.4512, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 21.32, |
|
"learning_rate": 7.0331914893617015e-06, |
|
"loss": 0.4536, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 21.35, |
|
"learning_rate": 7.02468085106383e-06, |
|
"loss": 0.4961, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 21.39, |
|
"learning_rate": 7.016170212765957e-06, |
|
"loss": 0.4739, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 21.42, |
|
"learning_rate": 7.007659574468085e-06, |
|
"loss": 0.4557, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 21.46, |
|
"learning_rate": 6.999148936170212e-06, |
|
"loss": 0.452, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 21.49, |
|
"learning_rate": 6.9906382978723405e-06, |
|
"loss": 0.4847, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 21.53, |
|
"learning_rate": 6.982127659574468e-06, |
|
"loss": 0.4795, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 21.56, |
|
"learning_rate": 6.9736170212765955e-06, |
|
"loss": 0.4672, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 21.6, |
|
"learning_rate": 6.965106382978723e-06, |
|
"loss": 0.4572, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 21.63, |
|
"learning_rate": 6.956595744680851e-06, |
|
"loss": 0.4136, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 21.66, |
|
"learning_rate": 6.948085106382979e-06, |
|
"loss": 0.4891, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 21.7, |
|
"learning_rate": 6.939574468085106e-06, |
|
"loss": 0.4632, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 21.73, |
|
"learning_rate": 6.9310638297872345e-06, |
|
"loss": 0.4404, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 21.77, |
|
"learning_rate": 6.922553191489361e-06, |
|
"loss": 0.4809, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 21.8, |
|
"learning_rate": 6.914042553191489e-06, |
|
"loss": 0.4557, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 21.84, |
|
"learning_rate": 6.905531914893616e-06, |
|
"loss": 0.4788, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 21.87, |
|
"learning_rate": 6.897021276595744e-06, |
|
"loss": 0.4687, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 21.91, |
|
"learning_rate": 6.888510638297872e-06, |
|
"loss": 0.4798, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 21.94, |
|
"learning_rate": 6.879999999999999e-06, |
|
"loss": 0.4697, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 21.98, |
|
"learning_rate": 6.871489361702127e-06, |
|
"loss": 0.4707, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 22.0, |
|
"eval_loss": 0.5899641513824463, |
|
"eval_runtime": 132.5359, |
|
"eval_samples_per_second": 4.014, |
|
"eval_steps_per_second": 1.004, |
|
"eval_wer": 0.2888088963167706, |
|
"step": 6336 |
|
}, |
|
{ |
|
"epoch": 22.01, |
|
"learning_rate": 6.862978723404255e-06, |
|
"loss": 0.562, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 22.05, |
|
"learning_rate": 6.854468085106383e-06, |
|
"loss": 0.443, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 22.08, |
|
"learning_rate": 6.84595744680851e-06, |
|
"loss": 0.4499, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 22.12, |
|
"learning_rate": 6.8374468085106376e-06, |
|
"loss": 0.458, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 22.15, |
|
"learning_rate": 6.828936170212766e-06, |
|
"loss": 0.4721, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 22.19, |
|
"learning_rate": 6.820425531914893e-06, |
|
"loss": 0.4804, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 22.22, |
|
"learning_rate": 6.811914893617021e-06, |
|
"loss": 0.4745, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 22.26, |
|
"learning_rate": 6.803404255319148e-06, |
|
"loss": 0.4592, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 22.29, |
|
"learning_rate": 6.794893617021277e-06, |
|
"loss": 0.4853, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 22.33, |
|
"learning_rate": 6.786382978723404e-06, |
|
"loss": 0.4459, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 22.36, |
|
"learning_rate": 6.7778723404255315e-06, |
|
"loss": 0.4619, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 22.39, |
|
"learning_rate": 6.76936170212766e-06, |
|
"loss": 0.4411, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 22.43, |
|
"learning_rate": 6.760851063829787e-06, |
|
"loss": 0.4136, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 22.46, |
|
"learning_rate": 6.752340425531915e-06, |
|
"loss": 0.4543, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 22.5, |
|
"learning_rate": 6.743829787234042e-06, |
|
"loss": 0.5094, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 22.53, |
|
"learning_rate": 6.735319148936171e-06, |
|
"loss": 0.4592, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 22.57, |
|
"learning_rate": 6.726808510638298e-06, |
|
"loss": 0.4606, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 22.6, |
|
"learning_rate": 6.7182978723404255e-06, |
|
"loss": 0.4549, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 22.64, |
|
"learning_rate": 6.709787234042552e-06, |
|
"loss": 0.4621, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 22.67, |
|
"learning_rate": 6.7012765957446805e-06, |
|
"loss": 0.4713, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 22.71, |
|
"learning_rate": 6.692765957446808e-06, |
|
"loss": 0.4998, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 22.74, |
|
"learning_rate": 6.684255319148935e-06, |
|
"loss": 0.4612, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 22.78, |
|
"learning_rate": 6.675744680851063e-06, |
|
"loss": 0.4323, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 22.81, |
|
"learning_rate": 6.667234042553191e-06, |
|
"loss": 0.4485, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 22.85, |
|
"learning_rate": 6.658723404255319e-06, |
|
"loss": 0.4559, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 22.88, |
|
"learning_rate": 6.650212765957446e-06, |
|
"loss": 0.4806, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 22.91, |
|
"learning_rate": 6.6417021276595745e-06, |
|
"loss": 0.4695, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 22.95, |
|
"learning_rate": 6.633191489361702e-06, |
|
"loss": 0.4478, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 22.98, |
|
"learning_rate": 6.624680851063829e-06, |
|
"loss": 0.4562, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 23.0, |
|
"eval_loss": 0.5642163157463074, |
|
"eval_runtime": 132.8576, |
|
"eval_samples_per_second": 4.004, |
|
"eval_steps_per_second": 1.001, |
|
"eval_wer": 0.2955582695892524, |
|
"step": 6624 |
|
}, |
|
{ |
|
"epoch": 23.02, |
|
"learning_rate": 6.616170212765957e-06, |
|
"loss": 0.4739, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 23.06, |
|
"learning_rate": 6.607659574468085e-06, |
|
"loss": 0.4726, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 23.09, |
|
"learning_rate": 6.599148936170213e-06, |
|
"loss": 0.4628, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 23.12, |
|
"learning_rate": 6.59063829787234e-06, |
|
"loss": 0.4497, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 23.16, |
|
"learning_rate": 6.582127659574468e-06, |
|
"loss": 0.4641, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 23.19, |
|
"learning_rate": 6.573617021276596e-06, |
|
"loss": 0.477, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 23.23, |
|
"learning_rate": 6.565106382978723e-06, |
|
"loss": 0.4764, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 23.26, |
|
"learning_rate": 6.556595744680851e-06, |
|
"loss": 0.4382, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 23.3, |
|
"learning_rate": 6.548085106382978e-06, |
|
"loss": 0.4743, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 23.33, |
|
"learning_rate": 6.539574468085107e-06, |
|
"loss": 0.4374, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 23.37, |
|
"learning_rate": 6.531063829787234e-06, |
|
"loss": 0.4587, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 23.4, |
|
"learning_rate": 6.522553191489362e-06, |
|
"loss": 0.4717, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 23.44, |
|
"learning_rate": 6.514042553191489e-06, |
|
"loss": 0.4758, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 23.47, |
|
"learning_rate": 6.505531914893617e-06, |
|
"loss": 0.4621, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 23.51, |
|
"learning_rate": 6.497021276595744e-06, |
|
"loss": 0.4925, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 23.54, |
|
"learning_rate": 6.4885106382978715e-06, |
|
"loss": 0.5269, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 23.57, |
|
"learning_rate": 6.48e-06, |
|
"loss": 0.4422, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 23.61, |
|
"learning_rate": 6.471489361702127e-06, |
|
"loss": 0.4422, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 23.64, |
|
"learning_rate": 6.462978723404255e-06, |
|
"loss": 0.4644, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 23.68, |
|
"learning_rate": 6.454468085106382e-06, |
|
"loss": 0.4728, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 23.71, |
|
"learning_rate": 6.4459574468085105e-06, |
|
"loss": 0.4857, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 23.75, |
|
"learning_rate": 6.437446808510638e-06, |
|
"loss": 0.4416, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 23.78, |
|
"learning_rate": 6.4289361702127655e-06, |
|
"loss": 0.4135, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 23.82, |
|
"learning_rate": 6.420425531914893e-06, |
|
"loss": 0.4769, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 23.85, |
|
"learning_rate": 6.411914893617021e-06, |
|
"loss": 0.4428, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 23.89, |
|
"learning_rate": 6.403404255319149e-06, |
|
"loss": 0.4495, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 23.92, |
|
"learning_rate": 6.394893617021276e-06, |
|
"loss": 0.4582, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 23.96, |
|
"learning_rate": 6.386382978723404e-06, |
|
"loss": 0.4369, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 23.99, |
|
"learning_rate": 6.377872340425532e-06, |
|
"loss": 0.455, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_loss": 0.5732349753379822, |
|
"eval_runtime": 133.5797, |
|
"eval_samples_per_second": 3.983, |
|
"eval_steps_per_second": 0.996, |
|
"eval_wer": 0.28932313428038825, |
|
"step": 6912 |
|
}, |
|
{ |
|
"epoch": 24.03, |
|
"learning_rate": 6.3693617021276595e-06, |
|
"loss": 0.5, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 24.06, |
|
"learning_rate": 6.360851063829787e-06, |
|
"loss": 0.4288, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 24.1, |
|
"learning_rate": 6.352340425531914e-06, |
|
"loss": 0.4418, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 24.13, |
|
"learning_rate": 6.343829787234043e-06, |
|
"loss": 0.4671, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 24.17, |
|
"learning_rate": 6.33531914893617e-06, |
|
"loss": 0.4597, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 24.2, |
|
"learning_rate": 6.326808510638298e-06, |
|
"loss": 0.455, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 24.24, |
|
"learning_rate": 6.318297872340426e-06, |
|
"loss": 0.4744, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 24.27, |
|
"learning_rate": 6.3097872340425534e-06, |
|
"loss": 0.4441, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 24.3, |
|
"learning_rate": 6.301276595744681e-06, |
|
"loss": 0.4755, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 24.34, |
|
"learning_rate": 6.2927659574468075e-06, |
|
"loss": 0.4446, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 24.37, |
|
"learning_rate": 6.284255319148936e-06, |
|
"loss": 0.4696, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 24.41, |
|
"learning_rate": 6.275744680851063e-06, |
|
"loss": 0.4845, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 24.44, |
|
"learning_rate": 6.267234042553191e-06, |
|
"loss": 0.42, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 24.48, |
|
"learning_rate": 6.258723404255318e-06, |
|
"loss": 0.4748, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 24.51, |
|
"learning_rate": 6.250212765957447e-06, |
|
"loss": 0.4793, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 24.55, |
|
"learning_rate": 6.241702127659574e-06, |
|
"loss": 0.4424, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 24.58, |
|
"learning_rate": 6.2331914893617015e-06, |
|
"loss": 0.4361, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 24.62, |
|
"learning_rate": 6.224680851063829e-06, |
|
"loss": 0.4495, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 24.65, |
|
"learning_rate": 6.216170212765957e-06, |
|
"loss": 0.4462, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 24.69, |
|
"learning_rate": 6.207659574468085e-06, |
|
"loss": 0.4664, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 24.72, |
|
"learning_rate": 6.199148936170212e-06, |
|
"loss": 0.4676, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 24.75, |
|
"learning_rate": 6.19063829787234e-06, |
|
"loss": 0.4344, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 24.79, |
|
"learning_rate": 6.182127659574468e-06, |
|
"loss": 0.4299, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 24.82, |
|
"learning_rate": 6.1736170212765955e-06, |
|
"loss": 0.4401, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 24.86, |
|
"learning_rate": 6.165106382978723e-06, |
|
"loss": 0.4615, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 24.89, |
|
"learning_rate": 6.156595744680851e-06, |
|
"loss": 0.4462, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 24.93, |
|
"learning_rate": 6.148085106382979e-06, |
|
"loss": 0.4036, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 24.96, |
|
"learning_rate": 6.139574468085106e-06, |
|
"loss": 0.4456, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"learning_rate": 6.131063829787234e-06, |
|
"loss": 0.5011, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"eval_loss": 0.5611743330955505, |
|
"eval_runtime": 133.0769, |
|
"eval_samples_per_second": 3.998, |
|
"eval_steps_per_second": 0.999, |
|
"eval_wer": 0.2875875811531786, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 25.03, |
|
"learning_rate": 6.122553191489362e-06, |
|
"loss": 0.5137, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 25.07, |
|
"learning_rate": 6.1140425531914895e-06, |
|
"loss": 0.458, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 25.1, |
|
"learning_rate": 6.105531914893617e-06, |
|
"loss": 0.439, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 25.14, |
|
"learning_rate": 6.0970212765957444e-06, |
|
"loss": 0.4453, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 25.17, |
|
"learning_rate": 6.088510638297873e-06, |
|
"loss": 0.4853, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 25.21, |
|
"learning_rate": 6.079999999999999e-06, |
|
"loss": 0.4421, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 25.24, |
|
"learning_rate": 6.071489361702127e-06, |
|
"loss": 0.4661, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 25.28, |
|
"learning_rate": 6.062978723404254e-06, |
|
"loss": 0.4201, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 25.31, |
|
"learning_rate": 6.054468085106383e-06, |
|
"loss": 0.4676, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 25.35, |
|
"learning_rate": 6.04595744680851e-06, |
|
"loss": 0.4829, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 25.38, |
|
"learning_rate": 6.037446808510638e-06, |
|
"loss": 0.4608, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 25.42, |
|
"learning_rate": 6.028936170212766e-06, |
|
"loss": 0.4362, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 25.45, |
|
"learning_rate": 6.020425531914893e-06, |
|
"loss": 0.4492, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 25.48, |
|
"learning_rate": 6.011914893617021e-06, |
|
"loss": 0.4829, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 25.52, |
|
"learning_rate": 6.003404255319148e-06, |
|
"loss": 0.4489, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 25.55, |
|
"learning_rate": 5.994893617021277e-06, |
|
"loss": 0.4291, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 25.59, |
|
"learning_rate": 5.986382978723404e-06, |
|
"loss": 0.4341, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 25.62, |
|
"learning_rate": 5.9778723404255316e-06, |
|
"loss": 0.4451, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 25.66, |
|
"learning_rate": 5.969361702127659e-06, |
|
"loss": 0.44, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 25.69, |
|
"learning_rate": 5.960851063829787e-06, |
|
"loss": 0.4811, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 25.73, |
|
"learning_rate": 5.952340425531915e-06, |
|
"loss": 0.4657, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 25.76, |
|
"learning_rate": 5.943829787234042e-06, |
|
"loss": 0.4709, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 25.8, |
|
"learning_rate": 5.93531914893617e-06, |
|
"loss": 0.4489, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 25.83, |
|
"learning_rate": 5.926808510638298e-06, |
|
"loss": 0.426, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 25.87, |
|
"learning_rate": 5.9182978723404256e-06, |
|
"loss": 0.4735, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 25.9, |
|
"learning_rate": 5.909787234042553e-06, |
|
"loss": 0.4428, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 25.94, |
|
"learning_rate": 5.9012765957446805e-06, |
|
"loss": 0.4407, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 25.97, |
|
"learning_rate": 5.892765957446809e-06, |
|
"loss": 0.4658, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 26.0, |
|
"eval_loss": 0.563091516494751, |
|
"eval_runtime": 133.7955, |
|
"eval_samples_per_second": 3.976, |
|
"eval_steps_per_second": 0.994, |
|
"eval_wer": 0.29150864562576334, |
|
"step": 7488 |
|
}, |
|
{ |
|
"epoch": 26.01, |
|
"learning_rate": 5.884255319148936e-06, |
|
"loss": 0.5104, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 26.04, |
|
"learning_rate": 5.875744680851063e-06, |
|
"loss": 0.4606, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 26.08, |
|
"learning_rate": 5.867234042553191e-06, |
|
"loss": 0.4494, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 26.11, |
|
"learning_rate": 5.858723404255319e-06, |
|
"loss": 0.4411, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 26.15, |
|
"learning_rate": 5.850212765957446e-06, |
|
"loss": 0.4716, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 26.18, |
|
"learning_rate": 5.841702127659574e-06, |
|
"loss": 0.4661, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 26.21, |
|
"learning_rate": 5.833191489361702e-06, |
|
"loss": 0.4434, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 26.25, |
|
"learning_rate": 5.8246808510638294e-06, |
|
"loss": 0.448, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 26.28, |
|
"learning_rate": 5.816170212765957e-06, |
|
"loss": 0.4399, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 26.32, |
|
"learning_rate": 5.807659574468084e-06, |
|
"loss": 0.4914, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 26.35, |
|
"learning_rate": 5.799148936170213e-06, |
|
"loss": 0.4773, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 26.39, |
|
"learning_rate": 5.79063829787234e-06, |
|
"loss": 0.4539, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 26.42, |
|
"learning_rate": 5.782127659574468e-06, |
|
"loss": 0.4469, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 26.46, |
|
"learning_rate": 5.773617021276595e-06, |
|
"loss": 0.438, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 26.49, |
|
"learning_rate": 5.7651063829787234e-06, |
|
"loss": 0.449, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 26.53, |
|
"learning_rate": 5.756595744680851e-06, |
|
"loss": 0.4478, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 26.56, |
|
"learning_rate": 5.748085106382978e-06, |
|
"loss": 0.4542, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 26.6, |
|
"learning_rate": 5.739574468085106e-06, |
|
"loss": 0.4279, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 26.63, |
|
"learning_rate": 5.731063829787234e-06, |
|
"loss": 0.4255, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 26.66, |
|
"learning_rate": 5.722553191489362e-06, |
|
"loss": 0.4399, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 26.7, |
|
"learning_rate": 5.714042553191489e-06, |
|
"loss": 0.4897, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 26.73, |
|
"learning_rate": 5.705531914893617e-06, |
|
"loss": 0.4693, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 26.77, |
|
"learning_rate": 5.697021276595745e-06, |
|
"loss": 0.4241, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 26.8, |
|
"learning_rate": 5.688510638297872e-06, |
|
"loss": 0.4551, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 26.84, |
|
"learning_rate": 5.68e-06, |
|
"loss": 0.4498, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 26.87, |
|
"learning_rate": 5.671489361702128e-06, |
|
"loss": 0.435, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 26.91, |
|
"learning_rate": 5.662978723404256e-06, |
|
"loss": 0.4649, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 26.94, |
|
"learning_rate": 5.654468085106382e-06, |
|
"loss": 0.4215, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 26.98, |
|
"learning_rate": 5.64595744680851e-06, |
|
"loss": 0.4423, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 27.0, |
|
"eval_loss": 0.5668238997459412, |
|
"eval_runtime": 134.2481, |
|
"eval_samples_per_second": 3.963, |
|
"eval_steps_per_second": 0.991, |
|
"eval_wer": 0.2852735103168991, |
|
"step": 7776 |
|
}, |
|
{ |
|
"epoch": 27.01, |
|
"learning_rate": 5.637446808510638e-06, |
|
"loss": 0.4848, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 27.05, |
|
"learning_rate": 5.6289361702127655e-06, |
|
"loss": 0.4462, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 27.08, |
|
"learning_rate": 5.620425531914893e-06, |
|
"loss": 0.4358, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 27.12, |
|
"learning_rate": 5.6119148936170204e-06, |
|
"loss": 0.4406, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 27.15, |
|
"learning_rate": 5.603404255319149e-06, |
|
"loss": 0.4588, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 27.19, |
|
"learning_rate": 5.594893617021276e-06, |
|
"loss": 0.4681, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 27.22, |
|
"learning_rate": 5.586382978723404e-06, |
|
"loss": 0.457, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 27.26, |
|
"learning_rate": 5.577872340425531e-06, |
|
"loss": 0.3994, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 27.29, |
|
"learning_rate": 5.5693617021276595e-06, |
|
"loss": 0.4579, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 27.33, |
|
"learning_rate": 5.561702127659575e-06, |
|
"loss": 0.4507, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 27.36, |
|
"learning_rate": 5.553191489361702e-06, |
|
"loss": 0.4456, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 27.39, |
|
"learning_rate": 5.54468085106383e-06, |
|
"loss": 0.4618, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 27.43, |
|
"learning_rate": 5.536170212765957e-06, |
|
"loss": 0.4263, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 27.46, |
|
"learning_rate": 5.527659574468085e-06, |
|
"loss": 0.432, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 27.5, |
|
"learning_rate": 5.519148936170213e-06, |
|
"loss": 0.4393, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 27.53, |
|
"learning_rate": 5.51063829787234e-06, |
|
"loss": 0.4445, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 27.57, |
|
"learning_rate": 5.502127659574468e-06, |
|
"loss": 0.4639, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 27.6, |
|
"learning_rate": 5.493617021276596e-06, |
|
"loss": 0.4471, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 27.64, |
|
"learning_rate": 5.4851063829787236e-06, |
|
"loss": 0.4213, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 27.67, |
|
"learning_rate": 5.476595744680851e-06, |
|
"loss": 0.4747, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 27.71, |
|
"learning_rate": 5.468085106382978e-06, |
|
"loss": 0.4627, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 27.74, |
|
"learning_rate": 5.459574468085106e-06, |
|
"loss": 0.4492, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 27.78, |
|
"learning_rate": 5.4510638297872334e-06, |
|
"loss": 0.4158, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 27.81, |
|
"learning_rate": 5.442553191489361e-06, |
|
"loss": 0.4476, |
|
"step": 8010 |
|
}, |
|
{ |
|
"epoch": 27.85, |
|
"learning_rate": 5.434042553191489e-06, |
|
"loss": 0.4516, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 27.88, |
|
"learning_rate": 5.425531914893617e-06, |
|
"loss": 0.4435, |
|
"step": 8030 |
|
}, |
|
{ |
|
"epoch": 27.91, |
|
"learning_rate": 5.417021276595744e-06, |
|
"loss": 0.4603, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 27.95, |
|
"learning_rate": 5.408510638297872e-06, |
|
"loss": 0.4339, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 27.98, |
|
"learning_rate": 5.4e-06, |
|
"loss": 0.4287, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_loss": 0.5664469599723816, |
|
"eval_runtime": 133.0789, |
|
"eval_samples_per_second": 3.998, |
|
"eval_steps_per_second": 0.999, |
|
"eval_wer": 0.28469499260782927, |
|
"step": 8064 |
|
}, |
|
{ |
|
"epoch": 28.02, |
|
"learning_rate": 5.3914893617021274e-06, |
|
"loss": 0.4908, |
|
"step": 8070 |
|
}, |
|
{ |
|
"epoch": 28.06, |
|
"learning_rate": 5.382978723404255e-06, |
|
"loss": 0.4646, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 28.09, |
|
"learning_rate": 5.374468085106382e-06, |
|
"loss": 0.4049, |
|
"step": 8090 |
|
}, |
|
{ |
|
"epoch": 28.12, |
|
"learning_rate": 5.365957446808511e-06, |
|
"loss": 0.4424, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 28.16, |
|
"learning_rate": 5.357446808510638e-06, |
|
"loss": 0.4506, |
|
"step": 8110 |
|
}, |
|
{ |
|
"epoch": 28.19, |
|
"learning_rate": 5.348936170212766e-06, |
|
"loss": 0.4401, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 28.23, |
|
"learning_rate": 5.340425531914893e-06, |
|
"loss": 0.4478, |
|
"step": 8130 |
|
}, |
|
{ |
|
"epoch": 28.26, |
|
"learning_rate": 5.3319148936170214e-06, |
|
"loss": 0.4495, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 28.3, |
|
"learning_rate": 5.323404255319149e-06, |
|
"loss": 0.4561, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 28.33, |
|
"learning_rate": 5.314893617021276e-06, |
|
"loss": 0.4784, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 28.37, |
|
"learning_rate": 5.306382978723405e-06, |
|
"loss": 0.4633, |
|
"step": 8170 |
|
}, |
|
{ |
|
"epoch": 28.4, |
|
"learning_rate": 5.297872340425532e-06, |
|
"loss": 0.4623, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 28.44, |
|
"learning_rate": 5.28936170212766e-06, |
|
"loss": 0.4255, |
|
"step": 8190 |
|
}, |
|
{ |
|
"epoch": 28.47, |
|
"learning_rate": 5.280851063829787e-06, |
|
"loss": 0.4213, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 28.51, |
|
"learning_rate": 5.272340425531915e-06, |
|
"loss": 0.4664, |
|
"step": 8210 |
|
}, |
|
{ |
|
"epoch": 28.54, |
|
"learning_rate": 5.263829787234043e-06, |
|
"loss": 0.4707, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 28.57, |
|
"learning_rate": 5.2553191489361695e-06, |
|
"loss": 0.4314, |
|
"step": 8230 |
|
}, |
|
{ |
|
"epoch": 28.61, |
|
"learning_rate": 5.246808510638297e-06, |
|
"loss": 0.4262, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 28.64, |
|
"learning_rate": 5.238297872340425e-06, |
|
"loss": 0.4501, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 28.68, |
|
"learning_rate": 5.229787234042553e-06, |
|
"loss": 0.4495, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 28.71, |
|
"learning_rate": 5.22127659574468e-06, |
|
"loss": 0.4672, |
|
"step": 8270 |
|
}, |
|
{ |
|
"epoch": 28.75, |
|
"learning_rate": 5.212765957446808e-06, |
|
"loss": 0.444, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 28.78, |
|
"learning_rate": 5.204255319148936e-06, |
|
"loss": 0.4379, |
|
"step": 8290 |
|
}, |
|
{ |
|
"epoch": 28.82, |
|
"learning_rate": 5.1957446808510635e-06, |
|
"loss": 0.453, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 28.85, |
|
"learning_rate": 5.187234042553191e-06, |
|
"loss": 0.4379, |
|
"step": 8310 |
|
}, |
|
{ |
|
"epoch": 28.89, |
|
"learning_rate": 5.1787234042553184e-06, |
|
"loss": 0.4383, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 28.92, |
|
"learning_rate": 5.170212765957447e-06, |
|
"loss": 0.4279, |
|
"step": 8330 |
|
}, |
|
{ |
|
"epoch": 28.96, |
|
"learning_rate": 5.161702127659574e-06, |
|
"loss": 0.426, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 28.99, |
|
"learning_rate": 5.153191489361702e-06, |
|
"loss": 0.4634, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 29.0, |
|
"eval_loss": 0.5687233209609985, |
|
"eval_runtime": 134.6333, |
|
"eval_samples_per_second": 3.951, |
|
"eval_steps_per_second": 0.988, |
|
"eval_wer": 0.2874590216622742, |
|
"step": 8352 |
|
}, |
|
{ |
|
"epoch": 29.03, |
|
"learning_rate": 5.14468085106383e-06, |
|
"loss": 0.4244, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 29.06, |
|
"learning_rate": 5.1361702127659575e-06, |
|
"loss": 0.4715, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 29.1, |
|
"learning_rate": 5.127659574468085e-06, |
|
"loss": 0.4182, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 29.13, |
|
"learning_rate": 5.1191489361702124e-06, |
|
"loss": 0.459, |
|
"step": 8390 |
|
}, |
|
{ |
|
"epoch": 29.17, |
|
"learning_rate": 5.110638297872341e-06, |
|
"loss": 0.4341, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 29.2, |
|
"learning_rate": 5.102127659574468e-06, |
|
"loss": 0.4355, |
|
"step": 8410 |
|
}, |
|
{ |
|
"epoch": 29.24, |
|
"learning_rate": 5.093617021276596e-06, |
|
"loss": 0.4569, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 29.27, |
|
"learning_rate": 5.085106382978723e-06, |
|
"loss": 0.3944, |
|
"step": 8430 |
|
}, |
|
{ |
|
"epoch": 29.3, |
|
"learning_rate": 5.0765957446808515e-06, |
|
"loss": 0.4266, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 29.34, |
|
"learning_rate": 5.068085106382979e-06, |
|
"loss": 0.4668, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 29.37, |
|
"learning_rate": 5.059574468085106e-06, |
|
"loss": 0.4222, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 29.41, |
|
"learning_rate": 5.051063829787233e-06, |
|
"loss": 0.4542, |
|
"step": 8470 |
|
}, |
|
{ |
|
"epoch": 29.44, |
|
"learning_rate": 5.042553191489361e-06, |
|
"loss": 0.42, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 29.48, |
|
"learning_rate": 5.034042553191489e-06, |
|
"loss": 0.4644, |
|
"step": 8490 |
|
}, |
|
{ |
|
"epoch": 29.51, |
|
"learning_rate": 5.025531914893616e-06, |
|
"loss": 0.4568, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 29.55, |
|
"learning_rate": 5.017021276595744e-06, |
|
"loss": 0.4369, |
|
"step": 8510 |
|
}, |
|
{ |
|
"epoch": 29.58, |
|
"learning_rate": 5.008510638297872e-06, |
|
"loss": 0.4198, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 29.62, |
|
"learning_rate": 4.9999999999999996e-06, |
|
"loss": 0.4286, |
|
"step": 8530 |
|
}, |
|
{ |
|
"epoch": 29.65, |
|
"learning_rate": 4.991489361702127e-06, |
|
"loss": 0.4463, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 29.69, |
|
"learning_rate": 4.982978723404255e-06, |
|
"loss": 0.4491, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 29.72, |
|
"learning_rate": 4.974468085106383e-06, |
|
"loss": 0.4651, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 29.75, |
|
"learning_rate": 4.96595744680851e-06, |
|
"loss": 0.4399, |
|
"step": 8570 |
|
}, |
|
{ |
|
"epoch": 29.79, |
|
"learning_rate": 4.957446808510638e-06, |
|
"loss": 0.4016, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 29.82, |
|
"learning_rate": 4.948936170212766e-06, |
|
"loss": 0.4548, |
|
"step": 8590 |
|
}, |
|
{ |
|
"epoch": 29.86, |
|
"learning_rate": 4.9404255319148936e-06, |
|
"loss": 0.4949, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 29.89, |
|
"learning_rate": 4.931914893617021e-06, |
|
"loss": 0.4648, |
|
"step": 8610 |
|
}, |
|
{ |
|
"epoch": 29.93, |
|
"learning_rate": 4.9234042553191485e-06, |
|
"loss": 0.4253, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 29.96, |
|
"learning_rate": 4.914893617021277e-06, |
|
"loss": 0.4524, |
|
"step": 8630 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"learning_rate": 4.906382978723404e-06, |
|
"loss": 0.4413, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"eval_loss": 0.5683608055114746, |
|
"eval_runtime": 131.8613, |
|
"eval_samples_per_second": 4.035, |
|
"eval_steps_per_second": 1.009, |
|
"eval_wer": 0.295429710098348, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 30.03, |
|
"learning_rate": 4.897872340425532e-06, |
|
"loss": 0.4603, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 30.07, |
|
"learning_rate": 4.889361702127659e-06, |
|
"loss": 0.4516, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 30.1, |
|
"learning_rate": 4.8808510638297875e-06, |
|
"loss": 0.4476, |
|
"step": 8670 |
|
}, |
|
{ |
|
"epoch": 30.14, |
|
"learning_rate": 4.872340425531915e-06, |
|
"loss": 0.4528, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 30.17, |
|
"learning_rate": 4.8638297872340425e-06, |
|
"loss": 0.4464, |
|
"step": 8690 |
|
}, |
|
{ |
|
"epoch": 30.21, |
|
"learning_rate": 4.85531914893617e-06, |
|
"loss": 0.4193, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 30.24, |
|
"learning_rate": 4.846808510638298e-06, |
|
"loss": 0.4178, |
|
"step": 8710 |
|
}, |
|
{ |
|
"epoch": 30.28, |
|
"learning_rate": 4.838297872340425e-06, |
|
"loss": 0.399, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 30.31, |
|
"learning_rate": 4.829787234042552e-06, |
|
"loss": 0.4386, |
|
"step": 8730 |
|
}, |
|
{ |
|
"epoch": 30.35, |
|
"learning_rate": 4.821276595744681e-06, |
|
"loss": 0.465, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 30.38, |
|
"learning_rate": 4.812765957446808e-06, |
|
"loss": 0.4304, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 30.42, |
|
"learning_rate": 4.804255319148936e-06, |
|
"loss": 0.4718, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 30.45, |
|
"learning_rate": 4.795744680851063e-06, |
|
"loss": 0.4136, |
|
"step": 8770 |
|
}, |
|
{ |
|
"epoch": 30.48, |
|
"learning_rate": 4.787234042553191e-06, |
|
"loss": 0.4442, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 30.52, |
|
"learning_rate": 4.778723404255319e-06, |
|
"loss": 0.4509, |
|
"step": 8790 |
|
}, |
|
{ |
|
"epoch": 30.55, |
|
"learning_rate": 4.770212765957446e-06, |
|
"loss": 0.4563, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 30.59, |
|
"learning_rate": 4.761702127659574e-06, |
|
"loss": 0.4422, |
|
"step": 8810 |
|
}, |
|
{ |
|
"epoch": 30.62, |
|
"learning_rate": 4.753191489361702e-06, |
|
"loss": 0.4259, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 30.66, |
|
"learning_rate": 4.74468085106383e-06, |
|
"loss": 0.4331, |
|
"step": 8830 |
|
}, |
|
{ |
|
"epoch": 30.69, |
|
"learning_rate": 4.736170212765957e-06, |
|
"loss": 0.4595, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 30.73, |
|
"learning_rate": 4.7276595744680846e-06, |
|
"loss": 0.4269, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 30.76, |
|
"learning_rate": 4.719148936170213e-06, |
|
"loss": 0.4405, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 30.8, |
|
"learning_rate": 4.71063829787234e-06, |
|
"loss": 0.4133, |
|
"step": 8870 |
|
}, |
|
{ |
|
"epoch": 30.83, |
|
"learning_rate": 4.702127659574468e-06, |
|
"loss": 0.445, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 30.87, |
|
"learning_rate": 4.693617021276595e-06, |
|
"loss": 0.4755, |
|
"step": 8890 |
|
}, |
|
{ |
|
"epoch": 30.9, |
|
"learning_rate": 4.685106382978724e-06, |
|
"loss": 0.4147, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 30.94, |
|
"learning_rate": 4.676595744680851e-06, |
|
"loss": 0.408, |
|
"step": 8910 |
|
}, |
|
{ |
|
"epoch": 30.97, |
|
"learning_rate": 4.6680851063829785e-06, |
|
"loss": 0.4385, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 31.0, |
|
"eval_loss": 0.5601801872253418, |
|
"eval_runtime": 132.6008, |
|
"eval_samples_per_second": 4.012, |
|
"eval_steps_per_second": 1.003, |
|
"eval_wer": 0.2800668509352703, |
|
"step": 8928 |
|
}, |
|
{ |
|
"epoch": 31.01, |
|
"learning_rate": 4.659574468085107e-06, |
|
"loss": 0.4796, |
|
"step": 8930 |
|
}, |
|
{ |
|
"epoch": 31.04, |
|
"learning_rate": 4.651063829787234e-06, |
|
"loss": 0.4404, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 31.08, |
|
"learning_rate": 4.642553191489362e-06, |
|
"loss": 0.4353, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 31.11, |
|
"learning_rate": 4.6340425531914884e-06, |
|
"loss": 0.4253, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 31.15, |
|
"learning_rate": 4.625531914893617e-06, |
|
"loss": 0.4624, |
|
"step": 8970 |
|
}, |
|
{ |
|
"epoch": 31.18, |
|
"learning_rate": 4.617021276595744e-06, |
|
"loss": 0.4239, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 31.21, |
|
"learning_rate": 4.608510638297872e-06, |
|
"loss": 0.4264, |
|
"step": 8990 |
|
}, |
|
{ |
|
"epoch": 31.25, |
|
"learning_rate": 4.599999999999999e-06, |
|
"loss": 0.4259, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 31.28, |
|
"learning_rate": 4.5914893617021275e-06, |
|
"loss": 0.4292, |
|
"step": 9010 |
|
}, |
|
{ |
|
"epoch": 31.32, |
|
"learning_rate": 4.582978723404255e-06, |
|
"loss": 0.4115, |
|
"step": 9020 |
|
}, |
|
{ |
|
"epoch": 31.35, |
|
"learning_rate": 4.574468085106382e-06, |
|
"loss": 0.4872, |
|
"step": 9030 |
|
}, |
|
{ |
|
"epoch": 31.39, |
|
"learning_rate": 4.56595744680851e-06, |
|
"loss": 0.4342, |
|
"step": 9040 |
|
}, |
|
{ |
|
"epoch": 31.42, |
|
"learning_rate": 4.557446808510638e-06, |
|
"loss": 0.4272, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 31.46, |
|
"learning_rate": 4.548936170212766e-06, |
|
"loss": 0.4144, |
|
"step": 9060 |
|
}, |
|
{ |
|
"epoch": 31.49, |
|
"learning_rate": 4.540425531914893e-06, |
|
"loss": 0.4412, |
|
"step": 9070 |
|
}, |
|
{ |
|
"epoch": 31.53, |
|
"learning_rate": 4.5319148936170215e-06, |
|
"loss": 0.4591, |
|
"step": 9080 |
|
}, |
|
{ |
|
"epoch": 31.56, |
|
"learning_rate": 4.523404255319149e-06, |
|
"loss": 0.4421, |
|
"step": 9090 |
|
}, |
|
{ |
|
"epoch": 31.6, |
|
"learning_rate": 4.514893617021276e-06, |
|
"loss": 0.4473, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 31.63, |
|
"learning_rate": 4.506382978723404e-06, |
|
"loss": 0.404, |
|
"step": 9110 |
|
}, |
|
{ |
|
"epoch": 31.66, |
|
"learning_rate": 4.497872340425532e-06, |
|
"loss": 0.4556, |
|
"step": 9120 |
|
}, |
|
{ |
|
"epoch": 31.7, |
|
"learning_rate": 4.48936170212766e-06, |
|
"loss": 0.4476, |
|
"step": 9130 |
|
}, |
|
{ |
|
"epoch": 31.73, |
|
"learning_rate": 4.480851063829787e-06, |
|
"loss": 0.4237, |
|
"step": 9140 |
|
}, |
|
{ |
|
"epoch": 31.77, |
|
"learning_rate": 4.472340425531915e-06, |
|
"loss": 0.4422, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 31.8, |
|
"learning_rate": 4.463829787234043e-06, |
|
"loss": 0.4305, |
|
"step": 9160 |
|
}, |
|
{ |
|
"epoch": 31.84, |
|
"learning_rate": 4.45531914893617e-06, |
|
"loss": 0.4295, |
|
"step": 9170 |
|
}, |
|
{ |
|
"epoch": 31.87, |
|
"learning_rate": 4.446808510638298e-06, |
|
"loss": 0.4797, |
|
"step": 9180 |
|
}, |
|
{ |
|
"epoch": 31.91, |
|
"learning_rate": 4.438297872340425e-06, |
|
"loss": 0.4231, |
|
"step": 9190 |
|
}, |
|
{ |
|
"epoch": 31.94, |
|
"learning_rate": 4.429787234042554e-06, |
|
"loss": 0.4371, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 31.98, |
|
"learning_rate": 4.421276595744681e-06, |
|
"loss": 0.4557, |
|
"step": 9210 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_loss": 0.5636705756187439, |
|
"eval_runtime": 133.3263, |
|
"eval_samples_per_second": 3.99, |
|
"eval_steps_per_second": 0.998, |
|
"eval_wer": 0.27466735231728484, |
|
"step": 9216 |
|
}, |
|
{ |
|
"epoch": 32.01, |
|
"learning_rate": 4.412765957446808e-06, |
|
"loss": 0.4727, |
|
"step": 9220 |
|
}, |
|
{ |
|
"epoch": 32.05, |
|
"learning_rate": 4.404255319148935e-06, |
|
"loss": 0.4599, |
|
"step": 9230 |
|
}, |
|
{ |
|
"epoch": 32.08, |
|
"learning_rate": 4.3957446808510635e-06, |
|
"loss": 0.4019, |
|
"step": 9240 |
|
}, |
|
{ |
|
"epoch": 32.12, |
|
"learning_rate": 4.387234042553191e-06, |
|
"loss": 0.4084, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 32.15, |
|
"learning_rate": 4.3787234042553185e-06, |
|
"loss": 0.4386, |
|
"step": 9260 |
|
}, |
|
{ |
|
"epoch": 32.19, |
|
"learning_rate": 4.370212765957447e-06, |
|
"loss": 0.4857, |
|
"step": 9270 |
|
}, |
|
{ |
|
"epoch": 32.22, |
|
"learning_rate": 4.361702127659574e-06, |
|
"loss": 0.4431, |
|
"step": 9280 |
|
}, |
|
{ |
|
"epoch": 32.26, |
|
"learning_rate": 4.353191489361702e-06, |
|
"loss": 0.4415, |
|
"step": 9290 |
|
}, |
|
{ |
|
"epoch": 32.29, |
|
"learning_rate": 4.344680851063829e-06, |
|
"loss": 0.4383, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 32.33, |
|
"learning_rate": 4.3361702127659575e-06, |
|
"loss": 0.4288, |
|
"step": 9310 |
|
}, |
|
{ |
|
"epoch": 32.36, |
|
"learning_rate": 4.327659574468085e-06, |
|
"loss": 0.4504, |
|
"step": 9320 |
|
}, |
|
{ |
|
"epoch": 32.39, |
|
"learning_rate": 4.3191489361702125e-06, |
|
"loss": 0.4293, |
|
"step": 9330 |
|
}, |
|
{ |
|
"epoch": 32.43, |
|
"learning_rate": 4.31063829787234e-06, |
|
"loss": 0.4084, |
|
"step": 9340 |
|
}, |
|
{ |
|
"epoch": 32.46, |
|
"learning_rate": 4.302127659574468e-06, |
|
"loss": 0.4499, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 32.5, |
|
"learning_rate": 4.293617021276596e-06, |
|
"loss": 0.4591, |
|
"step": 9360 |
|
}, |
|
{ |
|
"epoch": 32.53, |
|
"learning_rate": 4.285106382978723e-06, |
|
"loss": 0.4292, |
|
"step": 9370 |
|
}, |
|
{ |
|
"epoch": 32.57, |
|
"learning_rate": 4.276595744680851e-06, |
|
"loss": 0.4, |
|
"step": 9380 |
|
}, |
|
{ |
|
"epoch": 32.6, |
|
"learning_rate": 4.268085106382979e-06, |
|
"loss": 0.3918, |
|
"step": 9390 |
|
}, |
|
{ |
|
"epoch": 32.64, |
|
"learning_rate": 4.2595744680851065e-06, |
|
"loss": 0.4227, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 32.67, |
|
"learning_rate": 4.251063829787234e-06, |
|
"loss": 0.4472, |
|
"step": 9410 |
|
}, |
|
{ |
|
"epoch": 32.71, |
|
"learning_rate": 4.242553191489361e-06, |
|
"loss": 0.4557, |
|
"step": 9420 |
|
}, |
|
{ |
|
"epoch": 32.74, |
|
"learning_rate": 4.23404255319149e-06, |
|
"loss": 0.4267, |
|
"step": 9430 |
|
}, |
|
{ |
|
"epoch": 32.78, |
|
"learning_rate": 4.225531914893617e-06, |
|
"loss": 0.4154, |
|
"step": 9440 |
|
}, |
|
{ |
|
"epoch": 32.81, |
|
"learning_rate": 4.217021276595745e-06, |
|
"loss": 0.4383, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 32.85, |
|
"learning_rate": 4.208510638297873e-06, |
|
"loss": 0.4675, |
|
"step": 9460 |
|
}, |
|
{ |
|
"epoch": 32.88, |
|
"learning_rate": 4.2e-06, |
|
"loss": 0.4231, |
|
"step": 9470 |
|
}, |
|
{ |
|
"epoch": 32.91, |
|
"learning_rate": 4.191489361702127e-06, |
|
"loss": 0.4229, |
|
"step": 9480 |
|
}, |
|
{ |
|
"epoch": 32.95, |
|
"learning_rate": 4.1829787234042545e-06, |
|
"loss": 0.4006, |
|
"step": 9490 |
|
}, |
|
{ |
|
"epoch": 32.98, |
|
"learning_rate": 4.174468085106383e-06, |
|
"loss": 0.4344, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 33.0, |
|
"eval_loss": 0.5689535737037659, |
|
"eval_runtime": 134.4411, |
|
"eval_samples_per_second": 3.957, |
|
"eval_steps_per_second": 0.989, |
|
"eval_wer": 0.2853377900623513, |
|
"step": 9504 |
|
}, |
|
{ |
|
"epoch": 33.02, |
|
"learning_rate": 4.16595744680851e-06, |
|
"loss": 0.4484, |
|
"step": 9510 |
|
}, |
|
{ |
|
"epoch": 33.06, |
|
"learning_rate": 4.157446808510638e-06, |
|
"loss": 0.4377, |
|
"step": 9520 |
|
}, |
|
{ |
|
"epoch": 33.09, |
|
"learning_rate": 4.148936170212765e-06, |
|
"loss": 0.427, |
|
"step": 9530 |
|
}, |
|
{ |
|
"epoch": 33.12, |
|
"learning_rate": 4.140425531914894e-06, |
|
"loss": 0.4336, |
|
"step": 9540 |
|
}, |
|
{ |
|
"epoch": 33.16, |
|
"learning_rate": 4.131914893617021e-06, |
|
"loss": 0.4454, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 33.19, |
|
"learning_rate": 4.1234042553191485e-06, |
|
"loss": 0.4398, |
|
"step": 9560 |
|
}, |
|
{ |
|
"epoch": 33.23, |
|
"learning_rate": 4.114893617021276e-06, |
|
"loss": 0.4501, |
|
"step": 9570 |
|
}, |
|
{ |
|
"epoch": 33.26, |
|
"learning_rate": 4.106382978723404e-06, |
|
"loss": 0.4232, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 33.3, |
|
"learning_rate": 4.097872340425532e-06, |
|
"loss": 0.4278, |
|
"step": 9590 |
|
}, |
|
{ |
|
"epoch": 33.33, |
|
"learning_rate": 4.089361702127659e-06, |
|
"loss": 0.4328, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 33.37, |
|
"learning_rate": 4.080851063829787e-06, |
|
"loss": 0.4647, |
|
"step": 9610 |
|
}, |
|
{ |
|
"epoch": 33.4, |
|
"learning_rate": 4.072340425531915e-06, |
|
"loss": 0.4371, |
|
"step": 9620 |
|
}, |
|
{ |
|
"epoch": 33.44, |
|
"learning_rate": 4.0638297872340425e-06, |
|
"loss": 0.424, |
|
"step": 9630 |
|
}, |
|
{ |
|
"epoch": 33.47, |
|
"learning_rate": 4.05531914893617e-06, |
|
"loss": 0.4555, |
|
"step": 9640 |
|
}, |
|
{ |
|
"epoch": 33.51, |
|
"learning_rate": 4.046808510638298e-06, |
|
"loss": 0.4805, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 33.54, |
|
"learning_rate": 4.038297872340426e-06, |
|
"loss": 0.4349, |
|
"step": 9660 |
|
}, |
|
{ |
|
"epoch": 33.57, |
|
"learning_rate": 4.029787234042553e-06, |
|
"loss": 0.4339, |
|
"step": 9670 |
|
}, |
|
{ |
|
"epoch": 33.61, |
|
"learning_rate": 4.021276595744681e-06, |
|
"loss": 0.4088, |
|
"step": 9680 |
|
}, |
|
{ |
|
"epoch": 33.64, |
|
"learning_rate": 4.012765957446809e-06, |
|
"loss": 0.4114, |
|
"step": 9690 |
|
}, |
|
{ |
|
"epoch": 33.68, |
|
"learning_rate": 4.0042553191489365e-06, |
|
"loss": 0.4259, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 33.71, |
|
"learning_rate": 3.995744680851063e-06, |
|
"loss": 0.4222, |
|
"step": 9710 |
|
}, |
|
{ |
|
"epoch": 33.75, |
|
"learning_rate": 3.9872340425531914e-06, |
|
"loss": 0.4105, |
|
"step": 9720 |
|
}, |
|
{ |
|
"epoch": 33.78, |
|
"learning_rate": 3.978723404255319e-06, |
|
"loss": 0.4159, |
|
"step": 9730 |
|
}, |
|
{ |
|
"epoch": 33.82, |
|
"learning_rate": 3.970212765957446e-06, |
|
"loss": 0.453, |
|
"step": 9740 |
|
}, |
|
{ |
|
"epoch": 33.85, |
|
"learning_rate": 3.961702127659574e-06, |
|
"loss": 0.4618, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 33.89, |
|
"learning_rate": 3.953191489361702e-06, |
|
"loss": 0.4249, |
|
"step": 9760 |
|
}, |
|
{ |
|
"epoch": 33.92, |
|
"learning_rate": 3.94468085106383e-06, |
|
"loss": 0.4381, |
|
"step": 9770 |
|
}, |
|
{ |
|
"epoch": 33.96, |
|
"learning_rate": 3.936170212765957e-06, |
|
"loss": 0.4322, |
|
"step": 9780 |
|
}, |
|
{ |
|
"epoch": 33.99, |
|
"learning_rate": 3.927659574468085e-06, |
|
"loss": 0.4264, |
|
"step": 9790 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_loss": 0.5652894973754883, |
|
"eval_runtime": 133.9685, |
|
"eval_samples_per_second": 3.971, |
|
"eval_steps_per_second": 0.993, |
|
"eval_wer": 0.2865591052259433, |
|
"step": 9792 |
|
}, |
|
{ |
|
"epoch": 34.03, |
|
"learning_rate": 3.919148936170213e-06, |
|
"loss": 0.4907, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 34.06, |
|
"learning_rate": 3.91063829787234e-06, |
|
"loss": 0.4436, |
|
"step": 9810 |
|
}, |
|
{ |
|
"epoch": 34.1, |
|
"learning_rate": 3.902127659574468e-06, |
|
"loss": 0.4247, |
|
"step": 9820 |
|
}, |
|
{ |
|
"epoch": 34.13, |
|
"learning_rate": 3.893617021276596e-06, |
|
"loss": 0.4515, |
|
"step": 9830 |
|
}, |
|
{ |
|
"epoch": 34.17, |
|
"learning_rate": 3.885106382978723e-06, |
|
"loss": 0.4615, |
|
"step": 9840 |
|
}, |
|
{ |
|
"epoch": 34.2, |
|
"learning_rate": 3.876595744680851e-06, |
|
"loss": 0.4238, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 34.24, |
|
"learning_rate": 3.868085106382979e-06, |
|
"loss": 0.4262, |
|
"step": 9860 |
|
}, |
|
{ |
|
"epoch": 34.27, |
|
"learning_rate": 3.859574468085106e-06, |
|
"loss": 0.4094, |
|
"step": 9870 |
|
}, |
|
{ |
|
"epoch": 34.3, |
|
"learning_rate": 3.8510638297872335e-06, |
|
"loss": 0.4608, |
|
"step": 9880 |
|
}, |
|
{ |
|
"epoch": 34.34, |
|
"learning_rate": 3.842553191489362e-06, |
|
"loss": 0.4506, |
|
"step": 9890 |
|
}, |
|
{ |
|
"epoch": 34.37, |
|
"learning_rate": 3.834042553191489e-06, |
|
"loss": 0.4427, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 34.41, |
|
"learning_rate": 3.825531914893617e-06, |
|
"loss": 0.4177, |
|
"step": 9910 |
|
}, |
|
{ |
|
"epoch": 34.44, |
|
"learning_rate": 3.817021276595744e-06, |
|
"loss": 0.4178, |
|
"step": 9920 |
|
}, |
|
{ |
|
"epoch": 34.48, |
|
"learning_rate": 3.808510638297872e-06, |
|
"loss": 0.4193, |
|
"step": 9930 |
|
}, |
|
{ |
|
"epoch": 34.51, |
|
"learning_rate": 3.7999999999999996e-06, |
|
"loss": 0.4132, |
|
"step": 9940 |
|
}, |
|
{ |
|
"epoch": 34.55, |
|
"learning_rate": 3.7914893617021275e-06, |
|
"loss": 0.432, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 34.58, |
|
"learning_rate": 3.782978723404255e-06, |
|
"loss": 0.4413, |
|
"step": 9960 |
|
}, |
|
{ |
|
"epoch": 34.62, |
|
"learning_rate": 3.774468085106383e-06, |
|
"loss": 0.3887, |
|
"step": 9970 |
|
}, |
|
{ |
|
"epoch": 34.65, |
|
"learning_rate": 3.7659574468085108e-06, |
|
"loss": 0.4263, |
|
"step": 9980 |
|
}, |
|
{ |
|
"epoch": 34.69, |
|
"learning_rate": 3.7574468085106382e-06, |
|
"loss": 0.467, |
|
"step": 9990 |
|
}, |
|
{ |
|
"epoch": 34.72, |
|
"learning_rate": 3.748936170212766e-06, |
|
"loss": 0.4154, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 34.75, |
|
"learning_rate": 3.7404255319148936e-06, |
|
"loss": 0.4348, |
|
"step": 10010 |
|
}, |
|
{ |
|
"epoch": 34.79, |
|
"learning_rate": 3.731914893617021e-06, |
|
"loss": 0.3973, |
|
"step": 10020 |
|
}, |
|
{ |
|
"epoch": 34.82, |
|
"learning_rate": 3.7234042553191485e-06, |
|
"loss": 0.4246, |
|
"step": 10030 |
|
}, |
|
{ |
|
"epoch": 34.86, |
|
"learning_rate": 3.7148936170212764e-06, |
|
"loss": 0.4527, |
|
"step": 10040 |
|
}, |
|
{ |
|
"epoch": 34.89, |
|
"learning_rate": 3.706382978723404e-06, |
|
"loss": 0.424, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 34.93, |
|
"learning_rate": 3.697872340425532e-06, |
|
"loss": 0.3972, |
|
"step": 10060 |
|
}, |
|
{ |
|
"epoch": 34.96, |
|
"learning_rate": 3.6893617021276593e-06, |
|
"loss": 0.4456, |
|
"step": 10070 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"learning_rate": 3.680851063829787e-06, |
|
"loss": 0.4395, |
|
"step": 10080 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"eval_loss": 0.5764341354370117, |
|
"eval_runtime": 132.9016, |
|
"eval_samples_per_second": 4.003, |
|
"eval_steps_per_second": 1.001, |
|
"eval_wer": 0.2807739281352446, |
|
"step": 10080 |
|
}, |
|
{ |
|
"epoch": 35.03, |
|
"learning_rate": 3.6723404255319146e-06, |
|
"loss": 0.4322, |
|
"step": 10090 |
|
}, |
|
{ |
|
"epoch": 35.07, |
|
"learning_rate": 3.6638297872340425e-06, |
|
"loss": 0.425, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 35.1, |
|
"learning_rate": 3.65531914893617e-06, |
|
"loss": 0.3907, |
|
"step": 10110 |
|
}, |
|
{ |
|
"epoch": 35.14, |
|
"learning_rate": 3.646808510638298e-06, |
|
"loss": 0.4557, |
|
"step": 10120 |
|
}, |
|
{ |
|
"epoch": 35.17, |
|
"learning_rate": 3.6382978723404254e-06, |
|
"loss": 0.451, |
|
"step": 10130 |
|
}, |
|
{ |
|
"epoch": 35.21, |
|
"learning_rate": 3.629787234042553e-06, |
|
"loss": 0.436, |
|
"step": 10140 |
|
}, |
|
{ |
|
"epoch": 35.24, |
|
"learning_rate": 3.6212765957446803e-06, |
|
"loss": 0.4283, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 35.28, |
|
"learning_rate": 3.612765957446808e-06, |
|
"loss": 0.4228, |
|
"step": 10160 |
|
}, |
|
{ |
|
"epoch": 35.31, |
|
"learning_rate": 3.604255319148936e-06, |
|
"loss": 0.4419, |
|
"step": 10170 |
|
}, |
|
{ |
|
"epoch": 35.35, |
|
"learning_rate": 3.5957446808510636e-06, |
|
"loss": 0.4605, |
|
"step": 10180 |
|
}, |
|
{ |
|
"epoch": 35.38, |
|
"learning_rate": 3.5872340425531915e-06, |
|
"loss": 0.4174, |
|
"step": 10190 |
|
}, |
|
{ |
|
"epoch": 35.42, |
|
"learning_rate": 3.578723404255319e-06, |
|
"loss": 0.4318, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 35.45, |
|
"learning_rate": 3.570212765957447e-06, |
|
"loss": 0.4059, |
|
"step": 10210 |
|
}, |
|
{ |
|
"epoch": 35.48, |
|
"learning_rate": 3.5617021276595743e-06, |
|
"loss": 0.444, |
|
"step": 10220 |
|
}, |
|
{ |
|
"epoch": 35.52, |
|
"learning_rate": 3.553191489361702e-06, |
|
"loss": 0.4261, |
|
"step": 10230 |
|
}, |
|
{ |
|
"epoch": 35.55, |
|
"learning_rate": 3.5446808510638297e-06, |
|
"loss": 0.4416, |
|
"step": 10240 |
|
}, |
|
{ |
|
"epoch": 35.59, |
|
"learning_rate": 3.5361702127659576e-06, |
|
"loss": 0.4183, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 35.62, |
|
"learning_rate": 3.5276595744680846e-06, |
|
"loss": 0.4159, |
|
"step": 10260 |
|
}, |
|
{ |
|
"epoch": 35.66, |
|
"learning_rate": 3.5191489361702125e-06, |
|
"loss": 0.4209, |
|
"step": 10270 |
|
}, |
|
{ |
|
"epoch": 35.69, |
|
"learning_rate": 3.51063829787234e-06, |
|
"loss": 0.4594, |
|
"step": 10280 |
|
}, |
|
{ |
|
"epoch": 35.73, |
|
"learning_rate": 3.502127659574468e-06, |
|
"loss": 0.4146, |
|
"step": 10290 |
|
}, |
|
{ |
|
"epoch": 35.76, |
|
"learning_rate": 3.4936170212765953e-06, |
|
"loss": 0.429, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 35.8, |
|
"learning_rate": 3.4851063829787232e-06, |
|
"loss": 0.4081, |
|
"step": 10310 |
|
}, |
|
{ |
|
"epoch": 35.83, |
|
"learning_rate": 3.4765957446808507e-06, |
|
"loss": 0.4198, |
|
"step": 10320 |
|
}, |
|
{ |
|
"epoch": 35.87, |
|
"learning_rate": 3.4680851063829786e-06, |
|
"loss": 0.444, |
|
"step": 10330 |
|
}, |
|
{ |
|
"epoch": 35.9, |
|
"learning_rate": 3.4595744680851065e-06, |
|
"loss": 0.4591, |
|
"step": 10340 |
|
}, |
|
{ |
|
"epoch": 35.94, |
|
"learning_rate": 3.451063829787234e-06, |
|
"loss": 0.3977, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 35.97, |
|
"learning_rate": 3.442553191489362e-06, |
|
"loss": 0.4278, |
|
"step": 10360 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"eval_loss": 0.5757846832275391, |
|
"eval_runtime": 134.8971, |
|
"eval_samples_per_second": 3.944, |
|
"eval_steps_per_second": 0.986, |
|
"eval_wer": 0.2760815067172334, |
|
"step": 10368 |
|
}, |
|
{ |
|
"epoch": 36.01, |
|
"learning_rate": 3.4340425531914893e-06, |
|
"loss": 0.4533, |
|
"step": 10370 |
|
}, |
|
{ |
|
"epoch": 36.04, |
|
"learning_rate": 3.4255319148936172e-06, |
|
"loss": 0.402, |
|
"step": 10380 |
|
}, |
|
{ |
|
"epoch": 36.08, |
|
"learning_rate": 3.4170212765957443e-06, |
|
"loss": 0.4124, |
|
"step": 10390 |
|
}, |
|
{ |
|
"epoch": 36.11, |
|
"learning_rate": 3.408510638297872e-06, |
|
"loss": 0.4056, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 36.15, |
|
"learning_rate": 3.3999999999999996e-06, |
|
"loss": 0.401, |
|
"step": 10410 |
|
}, |
|
{ |
|
"epoch": 36.18, |
|
"learning_rate": 3.3914893617021275e-06, |
|
"loss": 0.4451, |
|
"step": 10420 |
|
}, |
|
{ |
|
"epoch": 36.21, |
|
"learning_rate": 3.382978723404255e-06, |
|
"loss": 0.4369, |
|
"step": 10430 |
|
}, |
|
{ |
|
"epoch": 36.25, |
|
"learning_rate": 3.374468085106383e-06, |
|
"loss": 0.4122, |
|
"step": 10440 |
|
}, |
|
{ |
|
"epoch": 36.28, |
|
"learning_rate": 3.3659574468085104e-06, |
|
"loss": 0.399, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 36.32, |
|
"learning_rate": 3.3574468085106383e-06, |
|
"loss": 0.4201, |
|
"step": 10460 |
|
}, |
|
{ |
|
"epoch": 36.35, |
|
"learning_rate": 3.3489361702127657e-06, |
|
"loss": 0.4234, |
|
"step": 10470 |
|
}, |
|
{ |
|
"epoch": 36.39, |
|
"learning_rate": 3.3404255319148936e-06, |
|
"loss": 0.4362, |
|
"step": 10480 |
|
}, |
|
{ |
|
"epoch": 36.42, |
|
"learning_rate": 3.331914893617021e-06, |
|
"loss": 0.4357, |
|
"step": 10490 |
|
}, |
|
{ |
|
"epoch": 36.46, |
|
"learning_rate": 3.323404255319149e-06, |
|
"loss": 0.424, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 36.49, |
|
"learning_rate": 3.314893617021276e-06, |
|
"loss": 0.4489, |
|
"step": 10510 |
|
}, |
|
{ |
|
"epoch": 36.53, |
|
"learning_rate": 3.306382978723404e-06, |
|
"loss": 0.4841, |
|
"step": 10520 |
|
}, |
|
{ |
|
"epoch": 36.56, |
|
"learning_rate": 3.297872340425532e-06, |
|
"loss": 0.4156, |
|
"step": 10530 |
|
}, |
|
{ |
|
"epoch": 36.6, |
|
"learning_rate": 3.2893617021276593e-06, |
|
"loss": 0.4043, |
|
"step": 10540 |
|
}, |
|
{ |
|
"epoch": 36.63, |
|
"learning_rate": 3.280851063829787e-06, |
|
"loss": 0.419, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 36.66, |
|
"learning_rate": 3.2723404255319147e-06, |
|
"loss": 0.4406, |
|
"step": 10560 |
|
}, |
|
{ |
|
"epoch": 36.7, |
|
"learning_rate": 3.2638297872340426e-06, |
|
"loss": 0.4529, |
|
"step": 10570 |
|
}, |
|
{ |
|
"epoch": 36.73, |
|
"learning_rate": 3.25531914893617e-06, |
|
"loss": 0.4346, |
|
"step": 10580 |
|
}, |
|
{ |
|
"epoch": 36.77, |
|
"learning_rate": 3.246808510638298e-06, |
|
"loss": 0.428, |
|
"step": 10590 |
|
}, |
|
{ |
|
"epoch": 36.8, |
|
"learning_rate": 3.2382978723404254e-06, |
|
"loss": 0.4344, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 36.84, |
|
"learning_rate": 3.2297872340425533e-06, |
|
"loss": 0.4354, |
|
"step": 10610 |
|
}, |
|
{ |
|
"epoch": 36.87, |
|
"learning_rate": 3.2212765957446808e-06, |
|
"loss": 0.4207, |
|
"step": 10620 |
|
}, |
|
{ |
|
"epoch": 36.91, |
|
"learning_rate": 3.2127659574468086e-06, |
|
"loss": 0.4382, |
|
"step": 10630 |
|
}, |
|
{ |
|
"epoch": 36.94, |
|
"learning_rate": 3.2042553191489357e-06, |
|
"loss": 0.4119, |
|
"step": 10640 |
|
}, |
|
{ |
|
"epoch": 36.98, |
|
"learning_rate": 3.1957446808510636e-06, |
|
"loss": 0.44, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 37.0, |
|
"eval_loss": 0.5815668106079102, |
|
"eval_runtime": 133.3811, |
|
"eval_samples_per_second": 3.989, |
|
"eval_steps_per_second": 0.997, |
|
"eval_wer": 0.27704570289901653, |
|
"step": 10656 |
|
}, |
|
{ |
|
"epoch": 37.01, |
|
"learning_rate": 3.187234042553191e-06, |
|
"loss": 0.4719, |
|
"step": 10660 |
|
}, |
|
{ |
|
"epoch": 37.05, |
|
"learning_rate": 3.178723404255319e-06, |
|
"loss": 0.4245, |
|
"step": 10670 |
|
}, |
|
{ |
|
"epoch": 37.08, |
|
"learning_rate": 3.1702127659574464e-06, |
|
"loss": 0.4144, |
|
"step": 10680 |
|
}, |
|
{ |
|
"epoch": 37.12, |
|
"learning_rate": 3.1617021276595743e-06, |
|
"loss": 0.4384, |
|
"step": 10690 |
|
}, |
|
{ |
|
"epoch": 37.15, |
|
"learning_rate": 3.1531914893617022e-06, |
|
"loss": 0.4373, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 37.19, |
|
"learning_rate": 3.1446808510638297e-06, |
|
"loss": 0.4426, |
|
"step": 10710 |
|
}, |
|
{ |
|
"epoch": 37.22, |
|
"learning_rate": 3.1361702127659576e-06, |
|
"loss": 0.4305, |
|
"step": 10720 |
|
}, |
|
{ |
|
"epoch": 37.26, |
|
"learning_rate": 3.127659574468085e-06, |
|
"loss": 0.433, |
|
"step": 10730 |
|
}, |
|
{ |
|
"epoch": 37.29, |
|
"learning_rate": 3.119148936170213e-06, |
|
"loss": 0.4056, |
|
"step": 10740 |
|
}, |
|
{ |
|
"epoch": 37.33, |
|
"learning_rate": 3.1106382978723404e-06, |
|
"loss": 0.4173, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 37.36, |
|
"learning_rate": 3.102127659574468e-06, |
|
"loss": 0.4224, |
|
"step": 10760 |
|
}, |
|
{ |
|
"epoch": 37.39, |
|
"learning_rate": 3.0936170212765954e-06, |
|
"loss": 0.4335, |
|
"step": 10770 |
|
}, |
|
{ |
|
"epoch": 37.43, |
|
"learning_rate": 3.0851063829787233e-06, |
|
"loss": 0.4112, |
|
"step": 10780 |
|
}, |
|
{ |
|
"epoch": 37.46, |
|
"learning_rate": 3.0765957446808507e-06, |
|
"loss": 0.4268, |
|
"step": 10790 |
|
}, |
|
{ |
|
"epoch": 37.5, |
|
"learning_rate": 3.0680851063829786e-06, |
|
"loss": 0.4459, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 37.53, |
|
"learning_rate": 3.059574468085106e-06, |
|
"loss": 0.4, |
|
"step": 10810 |
|
}, |
|
{ |
|
"epoch": 37.57, |
|
"learning_rate": 3.051063829787234e-06, |
|
"loss": 0.4398, |
|
"step": 10820 |
|
}, |
|
{ |
|
"epoch": 37.6, |
|
"learning_rate": 3.0425531914893615e-06, |
|
"loss": 0.3947, |
|
"step": 10830 |
|
}, |
|
{ |
|
"epoch": 37.64, |
|
"learning_rate": 3.0340425531914893e-06, |
|
"loss": 0.4136, |
|
"step": 10840 |
|
}, |
|
{ |
|
"epoch": 37.67, |
|
"learning_rate": 3.025531914893617e-06, |
|
"loss": 0.4447, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 37.71, |
|
"learning_rate": 3.0170212765957447e-06, |
|
"loss": 0.4635, |
|
"step": 10860 |
|
}, |
|
{ |
|
"epoch": 37.74, |
|
"learning_rate": 3.008510638297872e-06, |
|
"loss": 0.4247, |
|
"step": 10870 |
|
}, |
|
{ |
|
"epoch": 37.78, |
|
"learning_rate": 3e-06, |
|
"loss": 0.425, |
|
"step": 10880 |
|
}, |
|
{ |
|
"epoch": 37.81, |
|
"learning_rate": 2.9914893617021275e-06, |
|
"loss": 0.4365, |
|
"step": 10890 |
|
}, |
|
{ |
|
"epoch": 37.85, |
|
"learning_rate": 2.982978723404255e-06, |
|
"loss": 0.4025, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 37.88, |
|
"learning_rate": 2.974468085106383e-06, |
|
"loss": 0.4368, |
|
"step": 10910 |
|
}, |
|
{ |
|
"epoch": 37.91, |
|
"learning_rate": 2.9659574468085104e-06, |
|
"loss": 0.4028, |
|
"step": 10920 |
|
}, |
|
{ |
|
"epoch": 37.95, |
|
"learning_rate": 2.9574468085106383e-06, |
|
"loss": 0.3819, |
|
"step": 10930 |
|
}, |
|
{ |
|
"epoch": 37.98, |
|
"learning_rate": 2.9489361702127657e-06, |
|
"loss": 0.4356, |
|
"step": 10940 |
|
}, |
|
{ |
|
"epoch": 38.0, |
|
"eval_loss": 0.5813676118850708, |
|
"eval_runtime": 134.6483, |
|
"eval_samples_per_second": 3.951, |
|
"eval_steps_per_second": 0.988, |
|
"eval_wer": 0.2783955775535129, |
|
"step": 10944 |
|
}, |
|
{ |
|
"epoch": 38.02, |
|
"learning_rate": 2.9404255319148936e-06, |
|
"loss": 0.4929, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 38.06, |
|
"learning_rate": 2.931914893617021e-06, |
|
"loss": 0.4202, |
|
"step": 10960 |
|
}, |
|
{ |
|
"epoch": 38.09, |
|
"learning_rate": 2.923404255319149e-06, |
|
"loss": 0.4092, |
|
"step": 10970 |
|
}, |
|
{ |
|
"epoch": 38.12, |
|
"learning_rate": 2.9148936170212765e-06, |
|
"loss": 0.4292, |
|
"step": 10980 |
|
}, |
|
{ |
|
"epoch": 38.16, |
|
"learning_rate": 2.9063829787234044e-06, |
|
"loss": 0.41, |
|
"step": 10990 |
|
}, |
|
{ |
|
"epoch": 38.19, |
|
"learning_rate": 2.897872340425532e-06, |
|
"loss": 0.3993, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 38.23, |
|
"learning_rate": 2.8893617021276593e-06, |
|
"loss": 0.4267, |
|
"step": 11010 |
|
}, |
|
{ |
|
"epoch": 38.26, |
|
"learning_rate": 2.8808510638297868e-06, |
|
"loss": 0.3903, |
|
"step": 11020 |
|
}, |
|
{ |
|
"epoch": 38.3, |
|
"learning_rate": 2.8723404255319147e-06, |
|
"loss": 0.403, |
|
"step": 11030 |
|
}, |
|
{ |
|
"epoch": 38.33, |
|
"learning_rate": 2.863829787234042e-06, |
|
"loss": 0.4547, |
|
"step": 11040 |
|
}, |
|
{ |
|
"epoch": 38.37, |
|
"learning_rate": 2.85531914893617e-06, |
|
"loss": 0.415, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 38.4, |
|
"learning_rate": 2.8468085106382975e-06, |
|
"loss": 0.3951, |
|
"step": 11060 |
|
}, |
|
{ |
|
"epoch": 38.44, |
|
"learning_rate": 2.8382978723404254e-06, |
|
"loss": 0.4191, |
|
"step": 11070 |
|
}, |
|
{ |
|
"epoch": 38.47, |
|
"learning_rate": 2.8297872340425533e-06, |
|
"loss": 0.4491, |
|
"step": 11080 |
|
}, |
|
{ |
|
"epoch": 38.51, |
|
"learning_rate": 2.8212765957446808e-06, |
|
"loss": 0.4279, |
|
"step": 11090 |
|
}, |
|
{ |
|
"epoch": 38.54, |
|
"learning_rate": 2.8127659574468087e-06, |
|
"loss": 0.4283, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 38.57, |
|
"learning_rate": 2.804255319148936e-06, |
|
"loss": 0.438, |
|
"step": 11110 |
|
}, |
|
{ |
|
"epoch": 38.61, |
|
"learning_rate": 2.795744680851064e-06, |
|
"loss": 0.4026, |
|
"step": 11120 |
|
}, |
|
{ |
|
"epoch": 38.64, |
|
"learning_rate": 2.787234042553191e-06, |
|
"loss": 0.4396, |
|
"step": 11130 |
|
}, |
|
{ |
|
"epoch": 38.68, |
|
"learning_rate": 2.778723404255319e-06, |
|
"loss": 0.4181, |
|
"step": 11140 |
|
}, |
|
{ |
|
"epoch": 38.71, |
|
"learning_rate": 2.7702127659574464e-06, |
|
"loss": 0.4366, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 38.75, |
|
"learning_rate": 2.7617021276595743e-06, |
|
"loss": 0.4305, |
|
"step": 11160 |
|
}, |
|
{ |
|
"epoch": 38.78, |
|
"learning_rate": 2.753191489361702e-06, |
|
"loss": 0.372, |
|
"step": 11170 |
|
}, |
|
{ |
|
"epoch": 38.82, |
|
"learning_rate": 2.7446808510638297e-06, |
|
"loss": 0.4088, |
|
"step": 11180 |
|
}, |
|
{ |
|
"epoch": 38.85, |
|
"learning_rate": 2.736170212765957e-06, |
|
"loss": 0.4269, |
|
"step": 11190 |
|
}, |
|
{ |
|
"epoch": 38.89, |
|
"learning_rate": 2.727659574468085e-06, |
|
"loss": 0.4222, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 38.92, |
|
"learning_rate": 2.7191489361702125e-06, |
|
"loss": 0.4304, |
|
"step": 11210 |
|
}, |
|
{ |
|
"epoch": 38.96, |
|
"learning_rate": 2.7106382978723404e-06, |
|
"loss": 0.4269, |
|
"step": 11220 |
|
}, |
|
{ |
|
"epoch": 38.99, |
|
"learning_rate": 2.702127659574468e-06, |
|
"loss": 0.487, |
|
"step": 11230 |
|
}, |
|
{ |
|
"epoch": 39.0, |
|
"eval_loss": 0.5693573951721191, |
|
"eval_runtime": 131.675, |
|
"eval_samples_per_second": 4.04, |
|
"eval_steps_per_second": 1.01, |
|
"eval_wer": 0.2834093976987851, |
|
"step": 11232 |
|
}, |
|
{ |
|
"epoch": 39.03, |
|
"learning_rate": 2.693617021276596e-06, |
|
"loss": 0.4475, |
|
"step": 11240 |
|
}, |
|
{ |
|
"epoch": 39.06, |
|
"learning_rate": 2.6851063829787237e-06, |
|
"loss": 0.4228, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 39.1, |
|
"learning_rate": 2.6765957446808507e-06, |
|
"loss": 0.3902, |
|
"step": 11260 |
|
}, |
|
{ |
|
"epoch": 39.13, |
|
"learning_rate": 2.6680851063829786e-06, |
|
"loss": 0.4326, |
|
"step": 11270 |
|
}, |
|
{ |
|
"epoch": 39.17, |
|
"learning_rate": 2.659574468085106e-06, |
|
"loss": 0.4123, |
|
"step": 11280 |
|
}, |
|
{ |
|
"epoch": 39.2, |
|
"learning_rate": 2.651063829787234e-06, |
|
"loss": 0.417, |
|
"step": 11290 |
|
}, |
|
{ |
|
"epoch": 39.24, |
|
"learning_rate": 2.6425531914893615e-06, |
|
"loss": 0.4347, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 39.27, |
|
"learning_rate": 2.6340425531914894e-06, |
|
"loss": 0.3865, |
|
"step": 11310 |
|
}, |
|
{ |
|
"epoch": 39.3, |
|
"learning_rate": 2.625531914893617e-06, |
|
"loss": 0.4238, |
|
"step": 11320 |
|
}, |
|
{ |
|
"epoch": 39.34, |
|
"learning_rate": 2.6170212765957447e-06, |
|
"loss": 0.4533, |
|
"step": 11330 |
|
}, |
|
{ |
|
"epoch": 39.37, |
|
"learning_rate": 2.608510638297872e-06, |
|
"loss": 0.4112, |
|
"step": 11340 |
|
}, |
|
{ |
|
"epoch": 39.41, |
|
"learning_rate": 2.6e-06, |
|
"loss": 0.4321, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 39.44, |
|
"learning_rate": 2.5914893617021276e-06, |
|
"loss": 0.3737, |
|
"step": 11360 |
|
}, |
|
{ |
|
"epoch": 39.48, |
|
"learning_rate": 2.5829787234042555e-06, |
|
"loss": 0.4157, |
|
"step": 11370 |
|
}, |
|
{ |
|
"epoch": 39.51, |
|
"learning_rate": 2.5744680851063825e-06, |
|
"loss": 0.4174, |
|
"step": 11380 |
|
}, |
|
{ |
|
"epoch": 39.55, |
|
"learning_rate": 2.5659574468085104e-06, |
|
"loss": 0.4229, |
|
"step": 11390 |
|
}, |
|
{ |
|
"epoch": 39.58, |
|
"learning_rate": 2.557446808510638e-06, |
|
"loss": 0.4275, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 39.62, |
|
"learning_rate": 2.5489361702127658e-06, |
|
"loss": 0.4251, |
|
"step": 11410 |
|
}, |
|
{ |
|
"epoch": 39.65, |
|
"learning_rate": 2.5404255319148932e-06, |
|
"loss": 0.4355, |
|
"step": 11420 |
|
}, |
|
{ |
|
"epoch": 39.69, |
|
"learning_rate": 2.531914893617021e-06, |
|
"loss": 0.4321, |
|
"step": 11430 |
|
}, |
|
{ |
|
"epoch": 39.72, |
|
"learning_rate": 2.523404255319149e-06, |
|
"loss": 0.4074, |
|
"step": 11440 |
|
}, |
|
{ |
|
"epoch": 39.75, |
|
"learning_rate": 2.5148936170212765e-06, |
|
"loss": 0.389, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 39.79, |
|
"learning_rate": 2.5063829787234044e-06, |
|
"loss": 0.4206, |
|
"step": 11460 |
|
}, |
|
{ |
|
"epoch": 39.82, |
|
"learning_rate": 2.497872340425532e-06, |
|
"loss": 0.4245, |
|
"step": 11470 |
|
}, |
|
{ |
|
"epoch": 39.86, |
|
"learning_rate": 2.4893617021276598e-06, |
|
"loss": 0.4461, |
|
"step": 11480 |
|
}, |
|
{ |
|
"epoch": 39.89, |
|
"learning_rate": 2.4808510638297872e-06, |
|
"loss": 0.4278, |
|
"step": 11490 |
|
}, |
|
{ |
|
"epoch": 39.93, |
|
"learning_rate": 2.472340425531915e-06, |
|
"loss": 0.408, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 39.96, |
|
"learning_rate": 2.463829787234042e-06, |
|
"loss": 0.4275, |
|
"step": 11510 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"learning_rate": 2.45531914893617e-06, |
|
"loss": 0.44, |
|
"step": 11520 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_loss": 0.5636621117591858, |
|
"eval_runtime": 132.4068, |
|
"eval_samples_per_second": 4.018, |
|
"eval_steps_per_second": 1.004, |
|
"eval_wer": 0.27473163206273704, |
|
"step": 11520 |
|
}, |
|
{ |
|
"epoch": 40.03, |
|
"learning_rate": 2.4468085106382975e-06, |
|
"loss": 0.4403, |
|
"step": 11530 |
|
}, |
|
{ |
|
"epoch": 40.07, |
|
"learning_rate": 2.4382978723404254e-06, |
|
"loss": 0.4279, |
|
"step": 11540 |
|
}, |
|
{ |
|
"epoch": 40.1, |
|
"learning_rate": 2.429787234042553e-06, |
|
"loss": 0.416, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 40.14, |
|
"learning_rate": 2.421276595744681e-06, |
|
"loss": 0.4315, |
|
"step": 11560 |
|
}, |
|
{ |
|
"epoch": 40.17, |
|
"learning_rate": 2.4127659574468083e-06, |
|
"loss": 0.4413, |
|
"step": 11570 |
|
}, |
|
{ |
|
"epoch": 40.21, |
|
"learning_rate": 2.404255319148936e-06, |
|
"loss": 0.416, |
|
"step": 11580 |
|
}, |
|
{ |
|
"epoch": 40.24, |
|
"learning_rate": 2.3957446808510636e-06, |
|
"loss": 0.4333, |
|
"step": 11590 |
|
}, |
|
{ |
|
"epoch": 40.28, |
|
"learning_rate": 2.3872340425531915e-06, |
|
"loss": 0.4119, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 40.31, |
|
"learning_rate": 2.3787234042553194e-06, |
|
"loss": 0.4066, |
|
"step": 11610 |
|
}, |
|
{ |
|
"epoch": 40.35, |
|
"learning_rate": 2.370212765957447e-06, |
|
"loss": 0.4227, |
|
"step": 11620 |
|
}, |
|
{ |
|
"epoch": 40.38, |
|
"learning_rate": 2.3617021276595744e-06, |
|
"loss": 0.4336, |
|
"step": 11630 |
|
}, |
|
{ |
|
"epoch": 40.42, |
|
"learning_rate": 2.353191489361702e-06, |
|
"loss": 0.3754, |
|
"step": 11640 |
|
}, |
|
{ |
|
"epoch": 40.45, |
|
"learning_rate": 2.3446808510638297e-06, |
|
"loss": 0.3989, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 40.48, |
|
"learning_rate": 2.336170212765957e-06, |
|
"loss": 0.4487, |
|
"step": 11660 |
|
}, |
|
{ |
|
"epoch": 40.52, |
|
"learning_rate": 2.327659574468085e-06, |
|
"loss": 0.4478, |
|
"step": 11670 |
|
}, |
|
{ |
|
"epoch": 40.55, |
|
"learning_rate": 2.3191489361702126e-06, |
|
"loss": 0.4249, |
|
"step": 11680 |
|
}, |
|
{ |
|
"epoch": 40.59, |
|
"learning_rate": 2.3106382978723405e-06, |
|
"loss": 0.4058, |
|
"step": 11690 |
|
}, |
|
{ |
|
"epoch": 40.62, |
|
"learning_rate": 2.302127659574468e-06, |
|
"loss": 0.3722, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 40.66, |
|
"learning_rate": 2.293617021276596e-06, |
|
"loss": 0.4134, |
|
"step": 11710 |
|
}, |
|
{ |
|
"epoch": 40.69, |
|
"learning_rate": 2.2851063829787233e-06, |
|
"loss": 0.4582, |
|
"step": 11720 |
|
}, |
|
{ |
|
"epoch": 40.73, |
|
"learning_rate": 2.276595744680851e-06, |
|
"loss": 0.4146, |
|
"step": 11730 |
|
}, |
|
{ |
|
"epoch": 40.76, |
|
"learning_rate": 2.2680851063829787e-06, |
|
"loss": 0.4251, |
|
"step": 11740 |
|
}, |
|
{ |
|
"epoch": 40.8, |
|
"learning_rate": 2.259574468085106e-06, |
|
"loss": 0.4115, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 40.83, |
|
"learning_rate": 2.2510638297872336e-06, |
|
"loss": 0.4261, |
|
"step": 11760 |
|
}, |
|
{ |
|
"epoch": 40.87, |
|
"learning_rate": 2.2425531914893615e-06, |
|
"loss": 0.4373, |
|
"step": 11770 |
|
}, |
|
{ |
|
"epoch": 40.9, |
|
"learning_rate": 2.234042553191489e-06, |
|
"loss": 0.4125, |
|
"step": 11780 |
|
}, |
|
{ |
|
"epoch": 40.94, |
|
"learning_rate": 2.225531914893617e-06, |
|
"loss": 0.4014, |
|
"step": 11790 |
|
}, |
|
{ |
|
"epoch": 40.97, |
|
"learning_rate": 2.2170212765957447e-06, |
|
"loss": 0.4151, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 41.0, |
|
"eval_loss": 0.5682585835456848, |
|
"eval_runtime": 134.2217, |
|
"eval_samples_per_second": 3.964, |
|
"eval_steps_per_second": 0.991, |
|
"eval_wer": 0.2763386256990422, |
|
"step": 11808 |
|
}, |
|
{ |
|
"epoch": 41.01, |
|
"learning_rate": 2.2085106382978722e-06, |
|
"loss": 0.4544, |
|
"step": 11810 |
|
}, |
|
{ |
|
"epoch": 41.04, |
|
"learning_rate": 2.2e-06, |
|
"loss": 0.409, |
|
"step": 11820 |
|
}, |
|
{ |
|
"epoch": 41.08, |
|
"learning_rate": 2.1914893617021276e-06, |
|
"loss": 0.4006, |
|
"step": 11830 |
|
}, |
|
{ |
|
"epoch": 41.11, |
|
"learning_rate": 2.1829787234042555e-06, |
|
"loss": 0.4066, |
|
"step": 11840 |
|
}, |
|
{ |
|
"epoch": 41.15, |
|
"learning_rate": 2.174468085106383e-06, |
|
"loss": 0.4313, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 41.18, |
|
"learning_rate": 2.165957446808511e-06, |
|
"loss": 0.4554, |
|
"step": 11860 |
|
}, |
|
{ |
|
"epoch": 41.21, |
|
"learning_rate": 2.157446808510638e-06, |
|
"loss": 0.4092, |
|
"step": 11870 |
|
}, |
|
{ |
|
"epoch": 41.25, |
|
"learning_rate": 2.1489361702127658e-06, |
|
"loss": 0.4217, |
|
"step": 11880 |
|
}, |
|
{ |
|
"epoch": 41.28, |
|
"learning_rate": 2.1404255319148933e-06, |
|
"loss": 0.4233, |
|
"step": 11890 |
|
}, |
|
{ |
|
"epoch": 41.32, |
|
"learning_rate": 2.131914893617021e-06, |
|
"loss": 0.4132, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 41.35, |
|
"learning_rate": 2.1234042553191486e-06, |
|
"loss": 0.4182, |
|
"step": 11910 |
|
}, |
|
{ |
|
"epoch": 41.39, |
|
"learning_rate": 2.1157446808510638e-06, |
|
"loss": 0.4362, |
|
"step": 11920 |
|
}, |
|
{ |
|
"epoch": 41.42, |
|
"learning_rate": 2.1072340425531917e-06, |
|
"loss": 0.4363, |
|
"step": 11930 |
|
}, |
|
{ |
|
"epoch": 41.46, |
|
"learning_rate": 2.098723404255319e-06, |
|
"loss": 0.3856, |
|
"step": 11940 |
|
}, |
|
{ |
|
"epoch": 41.49, |
|
"learning_rate": 2.090212765957447e-06, |
|
"loss": 0.4307, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 41.53, |
|
"learning_rate": 2.0817021276595745e-06, |
|
"loss": 0.4284, |
|
"step": 11960 |
|
}, |
|
{ |
|
"epoch": 41.56, |
|
"learning_rate": 2.0731914893617024e-06, |
|
"loss": 0.4275, |
|
"step": 11970 |
|
}, |
|
{ |
|
"epoch": 41.6, |
|
"learning_rate": 2.0646808510638294e-06, |
|
"loss": 0.4241, |
|
"step": 11980 |
|
}, |
|
{ |
|
"epoch": 41.63, |
|
"learning_rate": 2.0561702127659573e-06, |
|
"loss": 0.4151, |
|
"step": 11990 |
|
}, |
|
{ |
|
"epoch": 41.66, |
|
"learning_rate": 2.047659574468085e-06, |
|
"loss": 0.4277, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 41.7, |
|
"learning_rate": 2.0391489361702127e-06, |
|
"loss": 0.4429, |
|
"step": 12010 |
|
}, |
|
{ |
|
"epoch": 41.73, |
|
"learning_rate": 2.03063829787234e-06, |
|
"loss": 0.403, |
|
"step": 12020 |
|
}, |
|
{ |
|
"epoch": 41.77, |
|
"learning_rate": 2.022127659574468e-06, |
|
"loss": 0.397, |
|
"step": 12030 |
|
}, |
|
{ |
|
"epoch": 41.8, |
|
"learning_rate": 2.0136170212765955e-06, |
|
"loss": 0.3967, |
|
"step": 12040 |
|
}, |
|
{ |
|
"epoch": 41.84, |
|
"learning_rate": 2.0051063829787234e-06, |
|
"loss": 0.4115, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 41.87, |
|
"learning_rate": 1.996595744680851e-06, |
|
"loss": 0.4237, |
|
"step": 12060 |
|
}, |
|
{ |
|
"epoch": 41.91, |
|
"learning_rate": 1.988085106382979e-06, |
|
"loss": 0.4216, |
|
"step": 12070 |
|
}, |
|
{ |
|
"epoch": 41.94, |
|
"learning_rate": 1.9795744680851063e-06, |
|
"loss": 0.4019, |
|
"step": 12080 |
|
}, |
|
{ |
|
"epoch": 41.98, |
|
"learning_rate": 1.971063829787234e-06, |
|
"loss": 0.4208, |
|
"step": 12090 |
|
}, |
|
{ |
|
"epoch": 42.0, |
|
"eval_loss": 0.571990966796875, |
|
"eval_runtime": 134.5195, |
|
"eval_samples_per_second": 3.955, |
|
"eval_steps_per_second": 0.989, |
|
"eval_wer": 0.27318891817188407, |
|
"step": 12096 |
|
}, |
|
{ |
|
"epoch": 42.01, |
|
"learning_rate": 1.9625531914893616e-06, |
|
"loss": 0.468, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 42.05, |
|
"learning_rate": 1.954042553191489e-06, |
|
"loss": 0.4384, |
|
"step": 12110 |
|
}, |
|
{ |
|
"epoch": 42.08, |
|
"learning_rate": 1.945531914893617e-06, |
|
"loss": 0.4088, |
|
"step": 12120 |
|
}, |
|
{ |
|
"epoch": 42.12, |
|
"learning_rate": 1.9370212765957445e-06, |
|
"loss": 0.4038, |
|
"step": 12130 |
|
}, |
|
{ |
|
"epoch": 42.15, |
|
"learning_rate": 1.9285106382978724e-06, |
|
"loss": 0.4097, |
|
"step": 12140 |
|
}, |
|
{ |
|
"epoch": 42.19, |
|
"learning_rate": 1.92e-06, |
|
"loss": 0.4487, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 42.22, |
|
"learning_rate": 1.9114893617021273e-06, |
|
"loss": 0.4229, |
|
"step": 12160 |
|
}, |
|
{ |
|
"epoch": 42.26, |
|
"learning_rate": 1.9029787234042552e-06, |
|
"loss": 0.4127, |
|
"step": 12170 |
|
}, |
|
{ |
|
"epoch": 42.29, |
|
"learning_rate": 1.8944680851063829e-06, |
|
"loss": 0.4359, |
|
"step": 12180 |
|
}, |
|
{ |
|
"epoch": 42.33, |
|
"learning_rate": 1.8859574468085106e-06, |
|
"loss": 0.4067, |
|
"step": 12190 |
|
}, |
|
{ |
|
"epoch": 42.36, |
|
"learning_rate": 1.8774468085106382e-06, |
|
"loss": 0.4246, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 42.39, |
|
"learning_rate": 1.868936170212766e-06, |
|
"loss": 0.4416, |
|
"step": 12210 |
|
}, |
|
{ |
|
"epoch": 42.43, |
|
"learning_rate": 1.8604255319148936e-06, |
|
"loss": 0.4094, |
|
"step": 12220 |
|
}, |
|
{ |
|
"epoch": 42.46, |
|
"learning_rate": 1.8519148936170213e-06, |
|
"loss": 0.4193, |
|
"step": 12230 |
|
}, |
|
{ |
|
"epoch": 42.5, |
|
"learning_rate": 1.843404255319149e-06, |
|
"loss": 0.4137, |
|
"step": 12240 |
|
}, |
|
{ |
|
"epoch": 42.53, |
|
"learning_rate": 1.8348936170212764e-06, |
|
"loss": 0.4068, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 42.57, |
|
"learning_rate": 1.8263829787234041e-06, |
|
"loss": 0.4028, |
|
"step": 12260 |
|
}, |
|
{ |
|
"epoch": 42.6, |
|
"learning_rate": 1.8178723404255318e-06, |
|
"loss": 0.3841, |
|
"step": 12270 |
|
}, |
|
{ |
|
"epoch": 42.64, |
|
"learning_rate": 1.8093617021276595e-06, |
|
"loss": 0.4044, |
|
"step": 12280 |
|
}, |
|
{ |
|
"epoch": 42.67, |
|
"learning_rate": 1.8008510638297872e-06, |
|
"loss": 0.4422, |
|
"step": 12290 |
|
}, |
|
{ |
|
"epoch": 42.71, |
|
"learning_rate": 1.7923404255319149e-06, |
|
"loss": 0.4219, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 42.74, |
|
"learning_rate": 1.7838297872340423e-06, |
|
"loss": 0.4225, |
|
"step": 12310 |
|
}, |
|
{ |
|
"epoch": 42.78, |
|
"learning_rate": 1.77531914893617e-06, |
|
"loss": 0.401, |
|
"step": 12320 |
|
}, |
|
{ |
|
"epoch": 42.81, |
|
"learning_rate": 1.7668085106382977e-06, |
|
"loss": 0.4085, |
|
"step": 12330 |
|
}, |
|
{ |
|
"epoch": 42.85, |
|
"learning_rate": 1.7582978723404254e-06, |
|
"loss": 0.4278, |
|
"step": 12340 |
|
}, |
|
{ |
|
"epoch": 42.88, |
|
"learning_rate": 1.749787234042553e-06, |
|
"loss": 0.442, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 42.91, |
|
"learning_rate": 1.7412765957446807e-06, |
|
"loss": 0.4061, |
|
"step": 12360 |
|
}, |
|
{ |
|
"epoch": 42.95, |
|
"learning_rate": 1.7327659574468086e-06, |
|
"loss": 0.4076, |
|
"step": 12370 |
|
}, |
|
{ |
|
"epoch": 42.98, |
|
"learning_rate": 1.724255319148936e-06, |
|
"loss": 0.4354, |
|
"step": 12380 |
|
}, |
|
{ |
|
"epoch": 43.0, |
|
"eval_loss": 0.5656567215919495, |
|
"eval_runtime": 135.0364, |
|
"eval_samples_per_second": 3.94, |
|
"eval_steps_per_second": 0.985, |
|
"eval_wer": 0.27710998264446873, |
|
"step": 12384 |
|
}, |
|
{ |
|
"epoch": 43.02, |
|
"learning_rate": 1.7157446808510638e-06, |
|
"loss": 0.4398, |
|
"step": 12390 |
|
}, |
|
{ |
|
"epoch": 43.06, |
|
"learning_rate": 1.7072340425531915e-06, |
|
"loss": 0.4191, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 43.09, |
|
"learning_rate": 1.6987234042553191e-06, |
|
"loss": 0.4114, |
|
"step": 12410 |
|
}, |
|
{ |
|
"epoch": 43.12, |
|
"learning_rate": 1.6902127659574468e-06, |
|
"loss": 0.4304, |
|
"step": 12420 |
|
}, |
|
{ |
|
"epoch": 43.16, |
|
"learning_rate": 1.6817021276595745e-06, |
|
"loss": 0.4, |
|
"step": 12430 |
|
}, |
|
{ |
|
"epoch": 43.19, |
|
"learning_rate": 1.673191489361702e-06, |
|
"loss": 0.4245, |
|
"step": 12440 |
|
}, |
|
{ |
|
"epoch": 43.23, |
|
"learning_rate": 1.6646808510638297e-06, |
|
"loss": 0.3886, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 43.26, |
|
"learning_rate": 1.6561702127659574e-06, |
|
"loss": 0.3894, |
|
"step": 12460 |
|
}, |
|
{ |
|
"epoch": 43.3, |
|
"learning_rate": 1.647659574468085e-06, |
|
"loss": 0.3997, |
|
"step": 12470 |
|
}, |
|
{ |
|
"epoch": 43.33, |
|
"learning_rate": 1.6391489361702127e-06, |
|
"loss": 0.4366, |
|
"step": 12480 |
|
}, |
|
{ |
|
"epoch": 43.37, |
|
"learning_rate": 1.6306382978723404e-06, |
|
"loss": 0.4191, |
|
"step": 12490 |
|
}, |
|
{ |
|
"epoch": 43.4, |
|
"learning_rate": 1.6221276595744679e-06, |
|
"loss": 0.4086, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 43.44, |
|
"learning_rate": 1.6136170212765956e-06, |
|
"loss": 0.4006, |
|
"step": 12510 |
|
}, |
|
{ |
|
"epoch": 43.47, |
|
"learning_rate": 1.6051063829787232e-06, |
|
"loss": 0.4089, |
|
"step": 12520 |
|
}, |
|
{ |
|
"epoch": 43.51, |
|
"learning_rate": 1.596595744680851e-06, |
|
"loss": 0.4287, |
|
"step": 12530 |
|
}, |
|
{ |
|
"epoch": 43.54, |
|
"learning_rate": 1.5880851063829786e-06, |
|
"loss": 0.4002, |
|
"step": 12540 |
|
}, |
|
{ |
|
"epoch": 43.57, |
|
"learning_rate": 1.5795744680851065e-06, |
|
"loss": 0.4306, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 43.61, |
|
"learning_rate": 1.571063829787234e-06, |
|
"loss": 0.3923, |
|
"step": 12560 |
|
}, |
|
{ |
|
"epoch": 43.64, |
|
"learning_rate": 1.5625531914893616e-06, |
|
"loss": 0.4052, |
|
"step": 12570 |
|
}, |
|
{ |
|
"epoch": 43.68, |
|
"learning_rate": 1.5540425531914893e-06, |
|
"loss": 0.4205, |
|
"step": 12580 |
|
}, |
|
{ |
|
"epoch": 43.71, |
|
"learning_rate": 1.545531914893617e-06, |
|
"loss": 0.4204, |
|
"step": 12590 |
|
}, |
|
{ |
|
"epoch": 43.75, |
|
"learning_rate": 1.5370212765957447e-06, |
|
"loss": 0.411, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 43.78, |
|
"learning_rate": 1.5285106382978724e-06, |
|
"loss": 0.4026, |
|
"step": 12610 |
|
}, |
|
{ |
|
"epoch": 43.82, |
|
"learning_rate": 1.5199999999999998e-06, |
|
"loss": 0.4251, |
|
"step": 12620 |
|
}, |
|
{ |
|
"epoch": 43.85, |
|
"learning_rate": 1.5114893617021275e-06, |
|
"loss": 0.403, |
|
"step": 12630 |
|
}, |
|
{ |
|
"epoch": 43.89, |
|
"learning_rate": 1.5029787234042552e-06, |
|
"loss": 0.4309, |
|
"step": 12640 |
|
}, |
|
{ |
|
"epoch": 43.92, |
|
"learning_rate": 1.4944680851063829e-06, |
|
"loss": 0.4174, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 43.96, |
|
"learning_rate": 1.4859574468085106e-06, |
|
"loss": 0.4075, |
|
"step": 12660 |
|
}, |
|
{ |
|
"epoch": 43.99, |
|
"learning_rate": 1.4774468085106383e-06, |
|
"loss": 0.4304, |
|
"step": 12670 |
|
}, |
|
{ |
|
"epoch": 44.0, |
|
"eval_loss": 0.5735185742378235, |
|
"eval_runtime": 133.4317, |
|
"eval_samples_per_second": 3.987, |
|
"eval_steps_per_second": 0.997, |
|
"eval_wer": 0.27241756122645755, |
|
"step": 12672 |
|
}, |
|
{ |
|
"epoch": 44.03, |
|
"learning_rate": 1.4689361702127657e-06, |
|
"loss": 0.4386, |
|
"step": 12680 |
|
}, |
|
{ |
|
"epoch": 44.06, |
|
"learning_rate": 1.4604255319148934e-06, |
|
"loss": 0.4332, |
|
"step": 12690 |
|
}, |
|
{ |
|
"epoch": 44.1, |
|
"learning_rate": 1.451914893617021e-06, |
|
"loss": 0.4041, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 44.13, |
|
"learning_rate": 1.4434042553191488e-06, |
|
"loss": 0.4219, |
|
"step": 12710 |
|
}, |
|
{ |
|
"epoch": 44.17, |
|
"learning_rate": 1.4348936170212765e-06, |
|
"loss": 0.4385, |
|
"step": 12720 |
|
}, |
|
{ |
|
"epoch": 44.2, |
|
"learning_rate": 1.4263829787234044e-06, |
|
"loss": 0.4258, |
|
"step": 12730 |
|
}, |
|
{ |
|
"epoch": 44.24, |
|
"learning_rate": 1.417872340425532e-06, |
|
"loss": 0.4333, |
|
"step": 12740 |
|
}, |
|
{ |
|
"epoch": 44.27, |
|
"learning_rate": 1.4093617021276595e-06, |
|
"loss": 0.4241, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 44.3, |
|
"learning_rate": 1.4008510638297872e-06, |
|
"loss": 0.4153, |
|
"step": 12760 |
|
}, |
|
{ |
|
"epoch": 44.34, |
|
"learning_rate": 1.3923404255319149e-06, |
|
"loss": 0.429, |
|
"step": 12770 |
|
}, |
|
{ |
|
"epoch": 44.37, |
|
"learning_rate": 1.3838297872340426e-06, |
|
"loss": 0.4148, |
|
"step": 12780 |
|
}, |
|
{ |
|
"epoch": 44.41, |
|
"learning_rate": 1.3753191489361702e-06, |
|
"loss": 0.4239, |
|
"step": 12790 |
|
}, |
|
{ |
|
"epoch": 44.44, |
|
"learning_rate": 1.366808510638298e-06, |
|
"loss": 0.3846, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 44.48, |
|
"learning_rate": 1.3582978723404254e-06, |
|
"loss": 0.426, |
|
"step": 12810 |
|
}, |
|
{ |
|
"epoch": 44.51, |
|
"learning_rate": 1.349787234042553e-06, |
|
"loss": 0.4258, |
|
"step": 12820 |
|
}, |
|
{ |
|
"epoch": 44.55, |
|
"learning_rate": 1.3412765957446808e-06, |
|
"loss": 0.4203, |
|
"step": 12830 |
|
}, |
|
{ |
|
"epoch": 44.58, |
|
"learning_rate": 1.3327659574468084e-06, |
|
"loss": 0.4069, |
|
"step": 12840 |
|
}, |
|
{ |
|
"epoch": 44.62, |
|
"learning_rate": 1.3242553191489361e-06, |
|
"loss": 0.3836, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 44.65, |
|
"learning_rate": 1.3157446808510638e-06, |
|
"loss": 0.4022, |
|
"step": 12860 |
|
}, |
|
{ |
|
"epoch": 44.69, |
|
"learning_rate": 1.3072340425531913e-06, |
|
"loss": 0.4059, |
|
"step": 12870 |
|
}, |
|
{ |
|
"epoch": 44.72, |
|
"learning_rate": 1.298723404255319e-06, |
|
"loss": 0.4013, |
|
"step": 12880 |
|
}, |
|
{ |
|
"epoch": 44.75, |
|
"learning_rate": 1.2902127659574466e-06, |
|
"loss": 0.3931, |
|
"step": 12890 |
|
}, |
|
{ |
|
"epoch": 44.79, |
|
"learning_rate": 1.2817021276595743e-06, |
|
"loss": 0.4007, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 44.82, |
|
"learning_rate": 1.2731914893617022e-06, |
|
"loss": 0.4356, |
|
"step": 12910 |
|
}, |
|
{ |
|
"epoch": 44.86, |
|
"learning_rate": 1.26468085106383e-06, |
|
"loss": 0.4132, |
|
"step": 12920 |
|
}, |
|
{ |
|
"epoch": 44.89, |
|
"learning_rate": 1.2561702127659574e-06, |
|
"loss": 0.4018, |
|
"step": 12930 |
|
}, |
|
{ |
|
"epoch": 44.93, |
|
"learning_rate": 1.247659574468085e-06, |
|
"loss": 0.4033, |
|
"step": 12940 |
|
}, |
|
{ |
|
"epoch": 44.96, |
|
"learning_rate": 1.2391489361702127e-06, |
|
"loss": 0.4141, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"learning_rate": 1.2306382978723404e-06, |
|
"loss": 0.3991, |
|
"step": 12960 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"eval_loss": 0.5637593865394592, |
|
"eval_runtime": 134.7734, |
|
"eval_samples_per_second": 3.947, |
|
"eval_steps_per_second": 0.987, |
|
"eval_wer": 0.2687536157356817, |
|
"step": 12960 |
|
}, |
|
{ |
|
"epoch": 45.03, |
|
"learning_rate": 1.222127659574468e-06, |
|
"loss": 0.4509, |
|
"step": 12970 |
|
}, |
|
{ |
|
"epoch": 45.07, |
|
"learning_rate": 1.2136170212765958e-06, |
|
"loss": 0.4135, |
|
"step": 12980 |
|
}, |
|
{ |
|
"epoch": 45.1, |
|
"learning_rate": 1.2051063829787233e-06, |
|
"loss": 0.3954, |
|
"step": 12990 |
|
}, |
|
{ |
|
"epoch": 45.14, |
|
"learning_rate": 1.196595744680851e-06, |
|
"loss": 0.4192, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 45.17, |
|
"learning_rate": 1.1880851063829786e-06, |
|
"loss": 0.4081, |
|
"step": 13010 |
|
}, |
|
{ |
|
"epoch": 45.21, |
|
"learning_rate": 1.1795744680851063e-06, |
|
"loss": 0.3951, |
|
"step": 13020 |
|
}, |
|
{ |
|
"epoch": 45.24, |
|
"learning_rate": 1.171063829787234e-06, |
|
"loss": 0.4076, |
|
"step": 13030 |
|
}, |
|
{ |
|
"epoch": 45.28, |
|
"learning_rate": 1.1625531914893617e-06, |
|
"loss": 0.3835, |
|
"step": 13040 |
|
}, |
|
{ |
|
"epoch": 45.31, |
|
"learning_rate": 1.1540425531914893e-06, |
|
"loss": 0.4094, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 45.35, |
|
"learning_rate": 1.1455319148936168e-06, |
|
"loss": 0.4492, |
|
"step": 13060 |
|
}, |
|
{ |
|
"epoch": 45.38, |
|
"learning_rate": 1.1370212765957445e-06, |
|
"loss": 0.3998, |
|
"step": 13070 |
|
}, |
|
{ |
|
"epoch": 45.42, |
|
"learning_rate": 1.1285106382978722e-06, |
|
"loss": 0.399, |
|
"step": 13080 |
|
}, |
|
{ |
|
"epoch": 45.45, |
|
"learning_rate": 1.12e-06, |
|
"loss": 0.4092, |
|
"step": 13090 |
|
}, |
|
{ |
|
"epoch": 45.48, |
|
"learning_rate": 1.1114893617021278e-06, |
|
"loss": 0.4328, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 45.52, |
|
"learning_rate": 1.1029787234042554e-06, |
|
"loss": 0.4434, |
|
"step": 13110 |
|
}, |
|
{ |
|
"epoch": 45.55, |
|
"learning_rate": 1.094468085106383e-06, |
|
"loss": 0.4291, |
|
"step": 13120 |
|
}, |
|
{ |
|
"epoch": 45.59, |
|
"learning_rate": 1.0859574468085106e-06, |
|
"loss": 0.4066, |
|
"step": 13130 |
|
}, |
|
{ |
|
"epoch": 45.62, |
|
"learning_rate": 1.0774468085106383e-06, |
|
"loss": 0.4124, |
|
"step": 13140 |
|
}, |
|
{ |
|
"epoch": 45.66, |
|
"learning_rate": 1.068936170212766e-06, |
|
"loss": 0.4146, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 45.69, |
|
"learning_rate": 1.0604255319148936e-06, |
|
"loss": 0.4516, |
|
"step": 13160 |
|
}, |
|
{ |
|
"epoch": 45.73, |
|
"learning_rate": 1.0519148936170213e-06, |
|
"loss": 0.3924, |
|
"step": 13170 |
|
}, |
|
{ |
|
"epoch": 45.76, |
|
"learning_rate": 1.0434042553191488e-06, |
|
"loss": 0.4024, |
|
"step": 13180 |
|
}, |
|
{ |
|
"epoch": 45.8, |
|
"learning_rate": 1.0348936170212765e-06, |
|
"loss": 0.4006, |
|
"step": 13190 |
|
}, |
|
{ |
|
"epoch": 45.83, |
|
"learning_rate": 1.0263829787234042e-06, |
|
"loss": 0.4287, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 45.87, |
|
"learning_rate": 1.0178723404255318e-06, |
|
"loss": 0.4035, |
|
"step": 13210 |
|
}, |
|
{ |
|
"epoch": 45.9, |
|
"learning_rate": 1.0093617021276595e-06, |
|
"loss": 0.4037, |
|
"step": 13220 |
|
}, |
|
{ |
|
"epoch": 45.94, |
|
"learning_rate": 1.0008510638297872e-06, |
|
"loss": 0.4273, |
|
"step": 13230 |
|
}, |
|
{ |
|
"epoch": 45.97, |
|
"learning_rate": 9.923404255319149e-07, |
|
"loss": 0.4348, |
|
"step": 13240 |
|
}, |
|
{ |
|
"epoch": 46.0, |
|
"eval_loss": 0.5639462471008301, |
|
"eval_runtime": 133.2564, |
|
"eval_samples_per_second": 3.992, |
|
"eval_steps_per_second": 0.998, |
|
"eval_wer": 0.26991065115382146, |
|
"step": 13248 |
|
}, |
|
{ |
|
"epoch": 46.01, |
|
"learning_rate": 9.838297872340426e-07, |
|
"loss": 0.4588, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 46.04, |
|
"learning_rate": 9.753191489361703e-07, |
|
"loss": 0.408, |
|
"step": 13260 |
|
}, |
|
{ |
|
"epoch": 46.08, |
|
"learning_rate": 9.668085106382977e-07, |
|
"loss": 0.4361, |
|
"step": 13270 |
|
}, |
|
{ |
|
"epoch": 46.11, |
|
"learning_rate": 9.582978723404254e-07, |
|
"loss": 0.4064, |
|
"step": 13280 |
|
}, |
|
{ |
|
"epoch": 46.15, |
|
"learning_rate": 9.497872340425531e-07, |
|
"loss": 0.4334, |
|
"step": 13290 |
|
}, |
|
{ |
|
"epoch": 46.18, |
|
"learning_rate": 9.412765957446808e-07, |
|
"loss": 0.4312, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 46.21, |
|
"learning_rate": 9.327659574468085e-07, |
|
"loss": 0.4124, |
|
"step": 13310 |
|
}, |
|
{ |
|
"epoch": 46.25, |
|
"learning_rate": 9.242553191489361e-07, |
|
"loss": 0.4274, |
|
"step": 13320 |
|
}, |
|
{ |
|
"epoch": 46.28, |
|
"learning_rate": 9.157446808510638e-07, |
|
"loss": 0.417, |
|
"step": 13330 |
|
}, |
|
{ |
|
"epoch": 46.32, |
|
"learning_rate": 9.072340425531915e-07, |
|
"loss": 0.4202, |
|
"step": 13340 |
|
}, |
|
{ |
|
"epoch": 46.35, |
|
"learning_rate": 8.987234042553191e-07, |
|
"loss": 0.4243, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 46.39, |
|
"learning_rate": 8.902127659574468e-07, |
|
"loss": 0.4138, |
|
"step": 13360 |
|
}, |
|
{ |
|
"epoch": 46.42, |
|
"learning_rate": 8.817021276595744e-07, |
|
"loss": 0.3794, |
|
"step": 13370 |
|
}, |
|
{ |
|
"epoch": 46.46, |
|
"learning_rate": 8.73191489361702e-07, |
|
"loss": 0.3794, |
|
"step": 13380 |
|
}, |
|
{ |
|
"epoch": 46.49, |
|
"learning_rate": 8.646808510638297e-07, |
|
"loss": 0.413, |
|
"step": 13390 |
|
}, |
|
{ |
|
"epoch": 46.53, |
|
"learning_rate": 8.561702127659574e-07, |
|
"loss": 0.4112, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 46.56, |
|
"learning_rate": 8.476595744680851e-07, |
|
"loss": 0.4195, |
|
"step": 13410 |
|
}, |
|
{ |
|
"epoch": 46.6, |
|
"learning_rate": 8.391489361702128e-07, |
|
"loss": 0.3927, |
|
"step": 13420 |
|
}, |
|
{ |
|
"epoch": 46.63, |
|
"learning_rate": 8.306382978723404e-07, |
|
"loss": 0.3942, |
|
"step": 13430 |
|
}, |
|
{ |
|
"epoch": 46.66, |
|
"learning_rate": 8.22127659574468e-07, |
|
"loss": 0.4077, |
|
"step": 13440 |
|
}, |
|
{ |
|
"epoch": 46.7, |
|
"learning_rate": 8.136170212765957e-07, |
|
"loss": 0.416, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 46.73, |
|
"learning_rate": 8.051063829787234e-07, |
|
"loss": 0.4035, |
|
"step": 13460 |
|
}, |
|
{ |
|
"epoch": 46.77, |
|
"learning_rate": 7.96595744680851e-07, |
|
"loss": 0.4068, |
|
"step": 13470 |
|
}, |
|
{ |
|
"epoch": 46.8, |
|
"learning_rate": 7.880851063829786e-07, |
|
"loss": 0.4026, |
|
"step": 13480 |
|
}, |
|
{ |
|
"epoch": 46.84, |
|
"learning_rate": 7.795744680851063e-07, |
|
"loss": 0.4216, |
|
"step": 13490 |
|
}, |
|
{ |
|
"epoch": 46.87, |
|
"learning_rate": 7.710638297872341e-07, |
|
"loss": 0.4101, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 46.91, |
|
"learning_rate": 7.625531914893617e-07, |
|
"loss": 0.4174, |
|
"step": 13510 |
|
}, |
|
{ |
|
"epoch": 46.94, |
|
"learning_rate": 7.540425531914894e-07, |
|
"loss": 0.3828, |
|
"step": 13520 |
|
}, |
|
{ |
|
"epoch": 46.98, |
|
"learning_rate": 7.45531914893617e-07, |
|
"loss": 0.4291, |
|
"step": 13530 |
|
}, |
|
{ |
|
"epoch": 47.0, |
|
"eval_loss": 0.5577101707458496, |
|
"eval_runtime": 135.0346, |
|
"eval_samples_per_second": 3.94, |
|
"eval_steps_per_second": 0.985, |
|
"eval_wer": 0.2681750980266118, |
|
"step": 13536 |
|
}, |
|
{ |
|
"epoch": 47.01, |
|
"learning_rate": 7.370212765957446e-07, |
|
"loss": 0.4639, |
|
"step": 13540 |
|
}, |
|
{ |
|
"epoch": 47.05, |
|
"learning_rate": 7.285106382978723e-07, |
|
"loss": 0.4172, |
|
"step": 13550 |
|
}, |
|
{ |
|
"epoch": 47.08, |
|
"learning_rate": 7.2e-07, |
|
"loss": 0.3836, |
|
"step": 13560 |
|
}, |
|
{ |
|
"epoch": 47.12, |
|
"learning_rate": 7.114893617021276e-07, |
|
"loss": 0.4125, |
|
"step": 13570 |
|
}, |
|
{ |
|
"epoch": 47.15, |
|
"learning_rate": 7.029787234042552e-07, |
|
"loss": 0.4288, |
|
"step": 13580 |
|
}, |
|
{ |
|
"epoch": 47.19, |
|
"learning_rate": 6.94468085106383e-07, |
|
"loss": 0.4026, |
|
"step": 13590 |
|
}, |
|
{ |
|
"epoch": 47.22, |
|
"learning_rate": 6.859574468085106e-07, |
|
"loss": 0.4049, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 47.26, |
|
"learning_rate": 6.774468085106383e-07, |
|
"loss": 0.432, |
|
"step": 13610 |
|
}, |
|
{ |
|
"epoch": 47.29, |
|
"learning_rate": 6.68936170212766e-07, |
|
"loss": 0.4045, |
|
"step": 13620 |
|
}, |
|
{ |
|
"epoch": 47.33, |
|
"learning_rate": 6.604255319148936e-07, |
|
"loss": 0.4128, |
|
"step": 13630 |
|
}, |
|
{ |
|
"epoch": 47.36, |
|
"learning_rate": 6.519148936170212e-07, |
|
"loss": 0.4183, |
|
"step": 13640 |
|
}, |
|
{ |
|
"epoch": 47.39, |
|
"learning_rate": 6.434042553191489e-07, |
|
"loss": 0.3997, |
|
"step": 13650 |
|
}, |
|
{ |
|
"epoch": 47.43, |
|
"learning_rate": 6.348936170212765e-07, |
|
"loss": 0.3937, |
|
"step": 13660 |
|
}, |
|
{ |
|
"epoch": 47.46, |
|
"learning_rate": 6.263829787234042e-07, |
|
"loss": 0.4155, |
|
"step": 13670 |
|
}, |
|
{ |
|
"epoch": 47.5, |
|
"learning_rate": 6.17872340425532e-07, |
|
"loss": 0.4143, |
|
"step": 13680 |
|
}, |
|
{ |
|
"epoch": 47.53, |
|
"learning_rate": 6.093617021276595e-07, |
|
"loss": 0.414, |
|
"step": 13690 |
|
}, |
|
{ |
|
"epoch": 47.57, |
|
"learning_rate": 6.008510638297872e-07, |
|
"loss": 0.3861, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 47.6, |
|
"learning_rate": 5.923404255319149e-07, |
|
"loss": 0.4069, |
|
"step": 13710 |
|
}, |
|
{ |
|
"epoch": 47.64, |
|
"learning_rate": 5.838297872340425e-07, |
|
"loss": 0.4269, |
|
"step": 13720 |
|
}, |
|
{ |
|
"epoch": 47.67, |
|
"learning_rate": 5.753191489361702e-07, |
|
"loss": 0.414, |
|
"step": 13730 |
|
}, |
|
{ |
|
"epoch": 47.71, |
|
"learning_rate": 5.668085106382978e-07, |
|
"loss": 0.4392, |
|
"step": 13740 |
|
}, |
|
{ |
|
"epoch": 47.74, |
|
"learning_rate": 5.582978723404254e-07, |
|
"loss": 0.3977, |
|
"step": 13750 |
|
}, |
|
{ |
|
"epoch": 47.78, |
|
"learning_rate": 5.497872340425531e-07, |
|
"loss": 0.367, |
|
"step": 13760 |
|
}, |
|
{ |
|
"epoch": 47.81, |
|
"learning_rate": 5.412765957446809e-07, |
|
"loss": 0.4072, |
|
"step": 13770 |
|
}, |
|
{ |
|
"epoch": 47.85, |
|
"learning_rate": 5.327659574468085e-07, |
|
"loss": 0.4229, |
|
"step": 13780 |
|
}, |
|
{ |
|
"epoch": 47.88, |
|
"learning_rate": 5.242553191489362e-07, |
|
"loss": 0.428, |
|
"step": 13790 |
|
}, |
|
{ |
|
"epoch": 47.91, |
|
"learning_rate": 5.157446808510638e-07, |
|
"loss": 0.4134, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 47.95, |
|
"learning_rate": 5.072340425531914e-07, |
|
"loss": 0.4104, |
|
"step": 13810 |
|
}, |
|
{ |
|
"epoch": 47.98, |
|
"learning_rate": 4.987234042553191e-07, |
|
"loss": 0.4252, |
|
"step": 13820 |
|
}, |
|
{ |
|
"epoch": 48.0, |
|
"eval_loss": 0.5610881447792053, |
|
"eval_runtime": 132.1309, |
|
"eval_samples_per_second": 4.026, |
|
"eval_steps_per_second": 1.007, |
|
"eval_wer": 0.2680465385357074, |
|
"step": 13824 |
|
}, |
|
{ |
|
"epoch": 48.02, |
|
"learning_rate": 4.91063829787234e-07, |
|
"loss": 0.4403, |
|
"step": 13830 |
|
}, |
|
{ |
|
"epoch": 48.06, |
|
"learning_rate": 4.825531914893617e-07, |
|
"loss": 0.4265, |
|
"step": 13840 |
|
}, |
|
{ |
|
"epoch": 48.09, |
|
"learning_rate": 4.7404255319148935e-07, |
|
"loss": 0.4026, |
|
"step": 13850 |
|
}, |
|
{ |
|
"epoch": 48.12, |
|
"learning_rate": 4.65531914893617e-07, |
|
"loss": 0.4031, |
|
"step": 13860 |
|
}, |
|
{ |
|
"epoch": 48.16, |
|
"learning_rate": 4.570212765957447e-07, |
|
"loss": 0.4244, |
|
"step": 13870 |
|
}, |
|
{ |
|
"epoch": 48.19, |
|
"learning_rate": 4.4851063829787234e-07, |
|
"loss": 0.4124, |
|
"step": 13880 |
|
}, |
|
{ |
|
"epoch": 48.23, |
|
"learning_rate": 4.3999999999999997e-07, |
|
"loss": 0.4199, |
|
"step": 13890 |
|
}, |
|
{ |
|
"epoch": 48.26, |
|
"learning_rate": 4.3148936170212765e-07, |
|
"loss": 0.3883, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 48.3, |
|
"learning_rate": 4.229787234042553e-07, |
|
"loss": 0.4297, |
|
"step": 13910 |
|
}, |
|
{ |
|
"epoch": 48.33, |
|
"learning_rate": 4.1446808510638296e-07, |
|
"loss": 0.4189, |
|
"step": 13920 |
|
}, |
|
{ |
|
"epoch": 48.37, |
|
"learning_rate": 4.0595744680851065e-07, |
|
"loss": 0.4313, |
|
"step": 13930 |
|
}, |
|
{ |
|
"epoch": 48.4, |
|
"learning_rate": 3.974468085106383e-07, |
|
"loss": 0.4304, |
|
"step": 13940 |
|
}, |
|
{ |
|
"epoch": 48.44, |
|
"learning_rate": 3.889361702127659e-07, |
|
"loss": 0.4021, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 48.47, |
|
"learning_rate": 3.8042553191489364e-07, |
|
"loss": 0.3968, |
|
"step": 13960 |
|
}, |
|
{ |
|
"epoch": 48.51, |
|
"learning_rate": 3.7191489361702127e-07, |
|
"loss": 0.4168, |
|
"step": 13970 |
|
}, |
|
{ |
|
"epoch": 48.54, |
|
"learning_rate": 3.634042553191489e-07, |
|
"loss": 0.425, |
|
"step": 13980 |
|
}, |
|
{ |
|
"epoch": 48.57, |
|
"learning_rate": 3.548936170212766e-07, |
|
"loss": 0.4012, |
|
"step": 13990 |
|
}, |
|
{ |
|
"epoch": 48.61, |
|
"learning_rate": 3.463829787234042e-07, |
|
"loss": 0.388, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 48.64, |
|
"learning_rate": 3.378723404255319e-07, |
|
"loss": 0.4405, |
|
"step": 14010 |
|
}, |
|
{ |
|
"epoch": 48.68, |
|
"learning_rate": 3.293617021276596e-07, |
|
"loss": 0.4384, |
|
"step": 14020 |
|
}, |
|
{ |
|
"epoch": 48.71, |
|
"learning_rate": 3.208510638297872e-07, |
|
"loss": 0.4048, |
|
"step": 14030 |
|
}, |
|
{ |
|
"epoch": 48.75, |
|
"learning_rate": 3.123404255319149e-07, |
|
"loss": 0.4088, |
|
"step": 14040 |
|
}, |
|
{ |
|
"epoch": 48.78, |
|
"learning_rate": 3.0382978723404257e-07, |
|
"loss": 0.3933, |
|
"step": 14050 |
|
}, |
|
{ |
|
"epoch": 48.82, |
|
"learning_rate": 2.953191489361702e-07, |
|
"loss": 0.4251, |
|
"step": 14060 |
|
}, |
|
{ |
|
"epoch": 48.85, |
|
"learning_rate": 2.868085106382979e-07, |
|
"loss": 0.3987, |
|
"step": 14070 |
|
}, |
|
{ |
|
"epoch": 48.89, |
|
"learning_rate": 2.782978723404255e-07, |
|
"loss": 0.3836, |
|
"step": 14080 |
|
}, |
|
{ |
|
"epoch": 48.92, |
|
"learning_rate": 2.6978723404255314e-07, |
|
"loss": 0.4018, |
|
"step": 14090 |
|
}, |
|
{ |
|
"epoch": 48.96, |
|
"learning_rate": 2.612765957446809e-07, |
|
"loss": 0.395, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 48.99, |
|
"learning_rate": 2.527659574468085e-07, |
|
"loss": 0.4253, |
|
"step": 14110 |
|
}, |
|
{ |
|
"epoch": 49.0, |
|
"eval_loss": 0.5621383786201477, |
|
"eval_runtime": 132.5942, |
|
"eval_samples_per_second": 4.012, |
|
"eval_steps_per_second": 1.003, |
|
"eval_wer": 0.268239377772064, |
|
"step": 14112 |
|
}, |
|
{ |
|
"epoch": 49.03, |
|
"learning_rate": 2.4425531914893614e-07, |
|
"loss": 0.4524, |
|
"step": 14120 |
|
}, |
|
{ |
|
"epoch": 49.06, |
|
"learning_rate": 2.3574468085106382e-07, |
|
"loss": 0.3945, |
|
"step": 14130 |
|
}, |
|
{ |
|
"epoch": 49.1, |
|
"learning_rate": 2.2723404255319148e-07, |
|
"loss": 0.4015, |
|
"step": 14140 |
|
}, |
|
{ |
|
"epoch": 49.13, |
|
"learning_rate": 2.1872340425531913e-07, |
|
"loss": 0.4088, |
|
"step": 14150 |
|
}, |
|
{ |
|
"epoch": 49.17, |
|
"learning_rate": 2.1021276595744681e-07, |
|
"loss": 0.4228, |
|
"step": 14160 |
|
}, |
|
{ |
|
"epoch": 49.2, |
|
"learning_rate": 2.0170212765957444e-07, |
|
"loss": 0.4073, |
|
"step": 14170 |
|
}, |
|
{ |
|
"epoch": 49.24, |
|
"learning_rate": 1.9319148936170213e-07, |
|
"loss": 0.4412, |
|
"step": 14180 |
|
}, |
|
{ |
|
"epoch": 49.27, |
|
"learning_rate": 1.8468085106382978e-07, |
|
"loss": 0.398, |
|
"step": 14190 |
|
}, |
|
{ |
|
"epoch": 49.3, |
|
"learning_rate": 1.7617021276595744e-07, |
|
"loss": 0.3931, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 49.34, |
|
"learning_rate": 1.676595744680851e-07, |
|
"loss": 0.4079, |
|
"step": 14210 |
|
}, |
|
{ |
|
"epoch": 49.37, |
|
"learning_rate": 1.5914893617021278e-07, |
|
"loss": 0.4437, |
|
"step": 14220 |
|
}, |
|
{ |
|
"epoch": 49.41, |
|
"learning_rate": 1.506382978723404e-07, |
|
"loss": 0.4097, |
|
"step": 14230 |
|
}, |
|
{ |
|
"epoch": 49.44, |
|
"learning_rate": 1.4212765957446806e-07, |
|
"loss": 0.3792, |
|
"step": 14240 |
|
}, |
|
{ |
|
"epoch": 49.48, |
|
"learning_rate": 1.3361702127659574e-07, |
|
"loss": 0.4091, |
|
"step": 14250 |
|
}, |
|
{ |
|
"epoch": 49.51, |
|
"learning_rate": 1.251063829787234e-07, |
|
"loss": 0.4166, |
|
"step": 14260 |
|
}, |
|
{ |
|
"epoch": 49.55, |
|
"learning_rate": 1.1659574468085106e-07, |
|
"loss": 0.395, |
|
"step": 14270 |
|
}, |
|
{ |
|
"epoch": 49.58, |
|
"learning_rate": 1.0808510638297871e-07, |
|
"loss": 0.3977, |
|
"step": 14280 |
|
}, |
|
{ |
|
"epoch": 49.62, |
|
"learning_rate": 9.957446808510637e-08, |
|
"loss": 0.401, |
|
"step": 14290 |
|
}, |
|
{ |
|
"epoch": 49.65, |
|
"learning_rate": 9.106382978723404e-08, |
|
"loss": 0.4239, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 49.69, |
|
"learning_rate": 8.25531914893617e-08, |
|
"loss": 0.418, |
|
"step": 14310 |
|
}, |
|
{ |
|
"epoch": 49.72, |
|
"learning_rate": 7.404255319148936e-08, |
|
"loss": 0.4267, |
|
"step": 14320 |
|
}, |
|
{ |
|
"epoch": 49.75, |
|
"learning_rate": 6.553191489361702e-08, |
|
"loss": 0.3983, |
|
"step": 14330 |
|
}, |
|
{ |
|
"epoch": 49.79, |
|
"learning_rate": 5.7021276595744676e-08, |
|
"loss": 0.3946, |
|
"step": 14340 |
|
}, |
|
{ |
|
"epoch": 49.82, |
|
"learning_rate": 4.851063829787234e-08, |
|
"loss": 0.408, |
|
"step": 14350 |
|
}, |
|
{ |
|
"epoch": 49.86, |
|
"learning_rate": 4e-08, |
|
"loss": 0.4219, |
|
"step": 14360 |
|
}, |
|
{ |
|
"epoch": 49.89, |
|
"learning_rate": 3.1489361702127663e-08, |
|
"loss": 0.3991, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 49.93, |
|
"learning_rate": 2.297872340425532e-08, |
|
"loss": 0.3836, |
|
"step": 14380 |
|
}, |
|
{ |
|
"epoch": 49.96, |
|
"learning_rate": 1.4468085106382979e-08, |
|
"loss": 0.4113, |
|
"step": 14390 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"learning_rate": 5.957446808510638e-09, |
|
"loss": 0.4298, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"eval_loss": 0.5591704845428467, |
|
"eval_runtime": 132.9336, |
|
"eval_samples_per_second": 4.002, |
|
"eval_steps_per_second": 1.0, |
|
"eval_wer": 0.2671466220993765, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"step": 14400, |
|
"total_flos": 1.3591623959806149e+20, |
|
"train_loss": 0.4495800813370281, |
|
"train_runtime": 106829.4814, |
|
"train_samples_per_second": 2.161, |
|
"train_steps_per_second": 0.135 |
|
} |
|
], |
|
"max_steps": 14400, |
|
"num_train_epochs": 50, |
|
"total_flos": 1.3591623959806149e+20, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|