|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9353078721745908, |
|
"eval_steps": 500, |
|
"global_step": 9000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 5e-06, |
|
"loss": 11.5375, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1e-05, |
|
"loss": 11.2877, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.5e-05, |
|
"loss": 10.631, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2e-05, |
|
"loss": 9.8892, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.5e-05, |
|
"loss": 9.0342, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 3e-05, |
|
"loss": 8.2813, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 3.5e-05, |
|
"loss": 7.8183, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 4e-05, |
|
"loss": 7.4274, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 4.5e-05, |
|
"loss": 7.3671, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 5e-05, |
|
"loss": 7.2528, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 5.500000000000001e-05, |
|
"loss": 7.1583, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 6e-05, |
|
"loss": 7.0445, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 6.500000000000001e-05, |
|
"loss": 6.9933, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 7e-05, |
|
"loss": 6.826, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 6.7937, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8e-05, |
|
"loss": 6.7451, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.5e-05, |
|
"loss": 6.4212, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9e-05, |
|
"loss": 6.4033, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.5e-05, |
|
"loss": 6.3906, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0001, |
|
"loss": 6.1864, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.99999319664223e-05, |
|
"loss": 6.0284, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.999972786587433e-05, |
|
"loss": 5.7979, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.99993876989115e-05, |
|
"loss": 5.9609, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.999891146645955e-05, |
|
"loss": 5.7689, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.999829916981445e-05, |
|
"loss": 5.6893, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.999755081064248e-05, |
|
"loss": 5.6071, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.999666639098018e-05, |
|
"loss": 5.5475, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.999564591323436e-05, |
|
"loss": 5.5319, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.99944893801821e-05, |
|
"loss": 5.3992, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.99931967949707e-05, |
|
"loss": 5.3752, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.999176816111775e-05, |
|
"loss": 5.3229, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.999020348251104e-05, |
|
"loss": 5.1756, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.99885027634086e-05, |
|
"loss": 5.1552, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.998666600843867e-05, |
|
"loss": 5.1344, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.99846932225997e-05, |
|
"loss": 5.1621, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.99825844112603e-05, |
|
"loss": 5.0381, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.998033958015928e-05, |
|
"loss": 5.1088, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.997795873540561e-05, |
|
"loss": 5.0639, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.997544188347834e-05, |
|
"loss": 4.9108, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.997278903122673e-05, |
|
"loss": 4.9241, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.997000018587011e-05, |
|
"loss": 4.9309, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.996707535499784e-05, |
|
"loss": 4.8356, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.99640145465694e-05, |
|
"loss": 4.9571, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.996081776891434e-05, |
|
"loss": 4.8575, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.995748503073215e-05, |
|
"loss": 4.8101, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.995401634109236e-05, |
|
"loss": 4.7393, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.995041170943447e-05, |
|
"loss": 4.6531, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.994667114556792e-05, |
|
"loss": 4.7505, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.994279465967206e-05, |
|
"loss": 4.7601, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.993878226229616e-05, |
|
"loss": 4.7105, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.993463396435929e-05, |
|
"loss": 4.5959, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.993034977715044e-05, |
|
"loss": 4.7773, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.992592971232833e-05, |
|
"loss": 4.6829, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.992137378192145e-05, |
|
"loss": 4.7622, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.991668199832809e-05, |
|
"loss": 4.5865, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.991185437431618e-05, |
|
"loss": 4.5472, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.990689092302336e-05, |
|
"loss": 4.5461, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.990179165795686e-05, |
|
"loss": 4.5179, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.989655659299353e-05, |
|
"loss": 4.625, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.98911857423798e-05, |
|
"loss": 4.5834, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.988567912073158e-05, |
|
"loss": 4.6043, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.98800367430343e-05, |
|
"loss": 4.5443, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.987425862464277e-05, |
|
"loss": 4.5195, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.986834478128127e-05, |
|
"loss": 4.497, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.986229522904336e-05, |
|
"loss": 4.4615, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.985610998439197e-05, |
|
"loss": 4.4588, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.984978906415927e-05, |
|
"loss": 4.4382, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.984333248554665e-05, |
|
"loss": 4.4523, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.983674026612467e-05, |
|
"loss": 4.4456, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.983001242383302e-05, |
|
"loss": 4.386, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.98231489769805e-05, |
|
"loss": 4.4249, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.981614994424487e-05, |
|
"loss": 4.3047, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.980901534467291e-05, |
|
"loss": 4.3131, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.980174519768031e-05, |
|
"loss": 4.3479, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.979433952305163e-05, |
|
"loss": 4.3042, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.978679834094025e-05, |
|
"loss": 4.2703, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.977912167186834e-05, |
|
"loss": 4.1616, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.977130953672673e-05, |
|
"loss": 4.413, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.976336195677491e-05, |
|
"loss": 4.2935, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.975527895364098e-05, |
|
"loss": 4.3209, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.974706054932158e-05, |
|
"loss": 4.2606, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.973870676618179e-05, |
|
"loss": 4.4179, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.973021762695514e-05, |
|
"loss": 4.1093, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.972159315474348e-05, |
|
"loss": 4.2495, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.971283337301694e-05, |
|
"loss": 4.2522, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.97039383056139e-05, |
|
"loss": 4.2719, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.969490797674091e-05, |
|
"loss": 4.2665, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.968574241097259e-05, |
|
"loss": 4.2985, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.967644163325156e-05, |
|
"loss": 4.2518, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.966700566888847e-05, |
|
"loss": 4.211, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.965743454356178e-05, |
|
"loss": 4.3097, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.964772828331781e-05, |
|
"loss": 4.2268, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.963788691457066e-05, |
|
"loss": 4.1852, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.962791046410202e-05, |
|
"loss": 4.2971, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.961779895906127e-05, |
|
"loss": 4.2119, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.960755242696528e-05, |
|
"loss": 4.2781, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.959717089569837e-05, |
|
"loss": 4.2193, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.958665439351224e-05, |
|
"loss": 4.1444, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.957600294902593e-05, |
|
"loss": 4.2276, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.956521659122567e-05, |
|
"loss": 4.1808, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.95542953494648e-05, |
|
"loss": 4.2218, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.954323925346381e-05, |
|
"loss": 4.1615, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.953204833331011e-05, |
|
"loss": 4.1061, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.952072261945804e-05, |
|
"loss": 4.0275, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.950926214272876e-05, |
|
"loss": 3.9907, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.949766693431015e-05, |
|
"loss": 4.0716, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.948593702575675e-05, |
|
"loss": 4.1418, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.947407244898967e-05, |
|
"loss": 4.0215, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.946207323629648e-05, |
|
"loss": 4.0912, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.944993942033119e-05, |
|
"loss": 4.1922, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.943767103411403e-05, |
|
"loss": 4.1028, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.942526811103152e-05, |
|
"loss": 4.129, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.941273068483625e-05, |
|
"loss": 4.0818, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.940005878964689e-05, |
|
"loss": 4.1768, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.938725245994798e-05, |
|
"loss": 4.0915, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.937431173058996e-05, |
|
"loss": 4.0007, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.9361236636789e-05, |
|
"loss": 3.9288, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.934802721412686e-05, |
|
"loss": 3.9529, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.933468349855099e-05, |
|
"loss": 4.0354, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.932120552637418e-05, |
|
"loss": 4.0737, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.930759333427459e-05, |
|
"loss": 4.0666, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.929384695929572e-05, |
|
"loss": 4.0967, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.927996643884614e-05, |
|
"loss": 4.0246, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.92659518106995e-05, |
|
"loss": 3.8967, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.925180311299446e-05, |
|
"loss": 4.1447, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.923752038423443e-05, |
|
"loss": 4.089, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.922310366328763e-05, |
|
"loss": 3.9846, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.920855298938691e-05, |
|
"loss": 4.0517, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.919386840212964e-05, |
|
"loss": 4.138, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.917904994147764e-05, |
|
"loss": 4.069, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.916409764775699e-05, |
|
"loss": 4.006, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.914901156165804e-05, |
|
"loss": 3.932, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.91337917242352e-05, |
|
"loss": 4.0524, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.911843817690686e-05, |
|
"loss": 3.9053, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.91029509614553e-05, |
|
"loss": 3.9072, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.908733012002655e-05, |
|
"loss": 3.9842, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.907157569513027e-05, |
|
"loss": 3.95, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.905568772963967e-05, |
|
"loss": 3.9991, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.903966626679133e-05, |
|
"loss": 3.9474, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.902351135018517e-05, |
|
"loss": 3.9754, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.900722302378425e-05, |
|
"loss": 3.989, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.89908013319147e-05, |
|
"loss": 3.9484, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.897424631926557e-05, |
|
"loss": 4.0325, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.895755803088874e-05, |
|
"loss": 4.0261, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.894073651219877e-05, |
|
"loss": 3.8771, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.892378180897277e-05, |
|
"loss": 3.8138, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.890669396735032e-05, |
|
"loss": 3.9609, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.888947303383329e-05, |
|
"loss": 3.9424, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.887211905528576e-05, |
|
"loss": 3.9631, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.885463207893383e-05, |
|
"loss": 3.8561, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.88370121523656e-05, |
|
"loss": 3.8925, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.881925932353091e-05, |
|
"loss": 3.8925, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.880137364074132e-05, |
|
"loss": 3.8758, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.878335515266989e-05, |
|
"loss": 3.806, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.876520390835112e-05, |
|
"loss": 3.9016, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.874691995718077e-05, |
|
"loss": 3.7877, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.872850334891575e-05, |
|
"loss": 3.866, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.870995413367397e-05, |
|
"loss": 3.8179, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.869127236193419e-05, |
|
"loss": 3.9106, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.867245808453594e-05, |
|
"loss": 3.9662, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.865351135267931e-05, |
|
"loss": 3.8928, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.863443221792487e-05, |
|
"loss": 3.8844, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.861522073219349e-05, |
|
"loss": 3.7855, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.859587694776621e-05, |
|
"loss": 3.8677, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.857640091728411e-05, |
|
"loss": 3.8445, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.855679269374816e-05, |
|
"loss": 3.8411, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.853705233051904e-05, |
|
"loss": 3.8302, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.851717988131705e-05, |
|
"loss": 3.8467, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.849717540022199e-05, |
|
"loss": 3.8176, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.847703894167286e-05, |
|
"loss": 3.7066, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.84567705604679e-05, |
|
"loss": 3.7558, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.843637031176433e-05, |
|
"loss": 3.7346, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.841583825107821e-05, |
|
"loss": 3.7371, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.839517443428434e-05, |
|
"loss": 3.8364, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.837437891761606e-05, |
|
"loss": 3.7334, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.835345175766509e-05, |
|
"loss": 3.865, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.833239301138142e-05, |
|
"loss": 3.7821, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.83112027360731e-05, |
|
"loss": 3.6758, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.828988098940619e-05, |
|
"loss": 3.7046, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.826842782940445e-05, |
|
"loss": 3.9129, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.824684331444927e-05, |
|
"loss": 3.8495, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.822512750327953e-05, |
|
"loss": 3.8108, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.820328045499143e-05, |
|
"loss": 3.753, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.818130222903828e-05, |
|
"loss": 3.8771, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.815919288523035e-05, |
|
"loss": 3.7107, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.813695248373474e-05, |
|
"loss": 3.7011, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.811458108507527e-05, |
|
"loss": 3.7602, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.809207875013213e-05, |
|
"loss": 3.8076, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.806944554014193e-05, |
|
"loss": 3.7961, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.80466815166974e-05, |
|
"loss": 3.7413, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.802378674174724e-05, |
|
"loss": 3.7162, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.800076127759599e-05, |
|
"loss": 3.7689, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.797760518690385e-05, |
|
"loss": 3.629, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.795431853268648e-05, |
|
"loss": 3.7742, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.793090137831486e-05, |
|
"loss": 3.6976, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.79073537875151e-05, |
|
"loss": 3.7106, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.788367582436828e-05, |
|
"loss": 3.7185, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.785986755331025e-05, |
|
"loss": 3.7897, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.78359290391315e-05, |
|
"loss": 3.6348, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.781186034697692e-05, |
|
"loss": 3.5979, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.778766154234569e-05, |
|
"loss": 3.7523, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.776333269109107e-05, |
|
"loss": 3.7874, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.77388738594202e-05, |
|
"loss": 3.7577, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.771428511389395e-05, |
|
"loss": 3.7654, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.768956652142674e-05, |
|
"loss": 3.8066, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.766471814928635e-05, |
|
"loss": 3.7506, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.76397400650937e-05, |
|
"loss": 3.8497, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.761463233682276e-05, |
|
"loss": 3.6532, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.758939503280024e-05, |
|
"loss": 3.7796, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.756402822170553e-05, |
|
"loss": 3.6971, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.753853197257042e-05, |
|
"loss": 3.8117, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.751290635477895e-05, |
|
"loss": 3.7334, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.748715143806719e-05, |
|
"loss": 3.8371, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.746126729252316e-05, |
|
"loss": 3.6015, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.743525398858646e-05, |
|
"loss": 3.6297, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.740911159704824e-05, |
|
"loss": 3.6828, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.738284018905091e-05, |
|
"loss": 3.7113, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.735643983608797e-05, |
|
"loss": 3.7889, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.732991061000385e-05, |
|
"loss": 3.5754, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.730325258299366e-05, |
|
"loss": 3.6713, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.727646582760308e-05, |
|
"loss": 3.5681, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.724955041672803e-05, |
|
"loss": 3.7093, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.722250642361456e-05, |
|
"loss": 3.514, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.719533392185872e-05, |
|
"loss": 3.6651, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.716803298540613e-05, |
|
"loss": 3.707, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.714060368855206e-05, |
|
"loss": 3.6944, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.711304610594104e-05, |
|
"loss": 3.7269, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.708536031256666e-05, |
|
"loss": 3.6314, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.705754638377151e-05, |
|
"loss": 3.5697, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.70296043952468e-05, |
|
"loss": 3.7098, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.70015344230323e-05, |
|
"loss": 3.696, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.697333654351602e-05, |
|
"loss": 3.7505, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.694501083343405e-05, |
|
"loss": 3.5885, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.691655736987038e-05, |
|
"loss": 3.7345, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.688797623025666e-05, |
|
"loss": 3.5453, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.685926749237195e-05, |
|
"loss": 3.6337, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.683043123434261e-05, |
|
"loss": 3.5826, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.680146753464197e-05, |
|
"loss": 3.7254, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.677237647209017e-05, |
|
"loss": 3.6509, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.674315812585402e-05, |
|
"loss": 3.7221, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.671381257544665e-05, |
|
"loss": 3.7415, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.668433990072737e-05, |
|
"loss": 3.6499, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.665474018190142e-05, |
|
"loss": 3.5825, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.662501349951981e-05, |
|
"loss": 3.5002, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.659515993447904e-05, |
|
"loss": 3.5725, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.656517956802092e-05, |
|
"loss": 3.6267, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.653507248173227e-05, |
|
"loss": 3.6351, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.650483875754483e-05, |
|
"loss": 3.4773, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.647447847773497e-05, |
|
"loss": 3.7241, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.644399172492336e-05, |
|
"loss": 3.5076, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.641337858207495e-05, |
|
"loss": 3.7403, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.638263913249861e-05, |
|
"loss": 3.5662, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.635177345984692e-05, |
|
"loss": 3.6316, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.632078164811598e-05, |
|
"loss": 3.5475, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.628966378164513e-05, |
|
"loss": 3.5504, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.625841994511675e-05, |
|
"loss": 3.6074, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.622705022355607e-05, |
|
"loss": 3.5816, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.619555470233083e-05, |
|
"loss": 3.5766, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.616393346715119e-05, |
|
"loss": 3.6748, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.613218660406933e-05, |
|
"loss": 3.5986, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.61003141994794e-05, |
|
"loss": 3.731, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.606831634011712e-05, |
|
"loss": 3.6098, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.603619311305968e-05, |
|
"loss": 3.5826, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.600394460572537e-05, |
|
"loss": 3.635, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.597157090587346e-05, |
|
"loss": 3.5129, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.593907210160387e-05, |
|
"loss": 3.5163, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.590644828135702e-05, |
|
"loss": 3.4268, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.587369953391352e-05, |
|
"loss": 3.5015, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.584082594839394e-05, |
|
"loss": 3.6242, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.580782761425859e-05, |
|
"loss": 3.5733, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.577470462130726e-05, |
|
"loss": 3.5834, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.574145705967898e-05, |
|
"loss": 3.4741, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.570808501985175e-05, |
|
"loss": 3.5284, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.567458859264238e-05, |
|
"loss": 3.6601, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.56409678692061e-05, |
|
"loss": 3.7271, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.560722294103646e-05, |
|
"loss": 3.56, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.557335389996499e-05, |
|
"loss": 3.7228, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.553936083816095e-05, |
|
"loss": 3.6206, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.550524384813114e-05, |
|
"loss": 3.5725, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.54710030227196e-05, |
|
"loss": 3.63, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.543663845510736e-05, |
|
"loss": 3.5948, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.540215023881219e-05, |
|
"loss": 3.5457, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.536753846768835e-05, |
|
"loss": 3.7054, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.53328032359264e-05, |
|
"loss": 3.6052, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.529794463805275e-05, |
|
"loss": 3.4509, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.526296276892965e-05, |
|
"loss": 3.4801, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.522785772375475e-05, |
|
"loss": 3.5778, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.519262959806095e-05, |
|
"loss": 3.5404, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.515727848771605e-05, |
|
"loss": 3.6708, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.512180448892254e-05, |
|
"loss": 3.4611, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.508620769821734e-05, |
|
"loss": 3.6322, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.505048821247156e-05, |
|
"loss": 3.5811, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.501464612889015e-05, |
|
"loss": 3.6099, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.497868154501172e-05, |
|
"loss": 3.6795, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.494259455870824e-05, |
|
"loss": 3.4475, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.49063852681848e-05, |
|
"loss": 3.5502, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.48700537719793e-05, |
|
"loss": 3.4725, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.48336001689622e-05, |
|
"loss": 3.536, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.479702455833626e-05, |
|
"loss": 3.4377, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.476032703963625e-05, |
|
"loss": 3.4536, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.472350771272875e-05, |
|
"loss": 3.5803, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.468656667781174e-05, |
|
"loss": 3.5074, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.464950403541447e-05, |
|
"loss": 3.5433, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.461231988639711e-05, |
|
"loss": 3.4499, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.457501433195048e-05, |
|
"loss": 3.5926, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.45375874735958e-05, |
|
"loss": 3.5056, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.450003941318438e-05, |
|
"loss": 3.5455, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.44623702528974e-05, |
|
"loss": 3.5595, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.442458009524554e-05, |
|
"loss": 3.3979, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.43866690430688e-05, |
|
"loss": 3.5457, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.434863719953618e-05, |
|
"loss": 3.4334, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.431048466814534e-05, |
|
"loss": 3.5655, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.427221155272241e-05, |
|
"loss": 3.4996, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.42338179574217e-05, |
|
"loss": 3.5764, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.419530398672533e-05, |
|
"loss": 3.5163, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.415666974544305e-05, |
|
"loss": 3.4246, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.411791533871186e-05, |
|
"loss": 3.4725, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.407904087199582e-05, |
|
"loss": 3.3925, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.404004645108568e-05, |
|
"loss": 3.4705, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.400093218209865e-05, |
|
"loss": 3.4844, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.396169817147805e-05, |
|
"loss": 3.4981, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.392234452599312e-05, |
|
"loss": 3.4456, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.388287135273861e-05, |
|
"loss": 3.5698, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.384327875913457e-05, |
|
"loss": 3.4827, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.380356685292603e-05, |
|
"loss": 3.5414, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.376373574218272e-05, |
|
"loss": 3.5588, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.372378553529875e-05, |
|
"loss": 3.4879, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.368371634099236e-05, |
|
"loss": 3.5712, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.364352826830555e-05, |
|
"loss": 3.5887, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.360322142660388e-05, |
|
"loss": 3.5668, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.356279592557607e-05, |
|
"loss": 3.4418, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.352225187523378e-05, |
|
"loss": 3.3832, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.34815893859113e-05, |
|
"loss": 3.5223, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.344080856826521e-05, |
|
"loss": 3.5331, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.339990953327411e-05, |
|
"loss": 3.484, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.33588923922383e-05, |
|
"loss": 3.4791, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.331775725677949e-05, |
|
"loss": 3.4218, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.327650423884051e-05, |
|
"loss": 3.42, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.323513345068496e-05, |
|
"loss": 3.5308, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.319364500489696e-05, |
|
"loss": 3.3971, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.315203901438082e-05, |
|
"loss": 3.4279, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.311031559236067e-05, |
|
"loss": 3.4812, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.30684748523803e-05, |
|
"loss": 3.3762, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.302651690830271e-05, |
|
"loss": 3.4395, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.298444187430986e-05, |
|
"loss": 3.497, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.294224986490236e-05, |
|
"loss": 3.4741, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.289994099489913e-05, |
|
"loss": 3.415, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.285751537943712e-05, |
|
"loss": 3.5428, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.281497313397101e-05, |
|
"loss": 3.3536, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.27723143742728e-05, |
|
"loss": 3.4309, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.272953921643168e-05, |
|
"loss": 3.4613, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.268664777685345e-05, |
|
"loss": 3.5508, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.264364017226051e-05, |
|
"loss": 3.536, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.260051651969128e-05, |
|
"loss": 3.4405, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.255727693649998e-05, |
|
"loss": 3.4646, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.251392154035641e-05, |
|
"loss": 3.4193, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.247045044924542e-05, |
|
"loss": 3.4704, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.242686378146682e-05, |
|
"loss": 3.3396, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.238316165563485e-05, |
|
"loss": 3.5639, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.233934419067802e-05, |
|
"loss": 3.41, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.229541150583865e-05, |
|
"loss": 3.47, |
|
"step": 1805 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.225136372067266e-05, |
|
"loss": 3.4233, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.22072009550492e-05, |
|
"loss": 3.3309, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.216292332915031e-05, |
|
"loss": 3.5253, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.211853096347058e-05, |
|
"loss": 3.4263, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.20740239788169e-05, |
|
"loss": 3.5337, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.202940249630801e-05, |
|
"loss": 3.3848, |
|
"step": 1835 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.198466663737431e-05, |
|
"loss": 3.3998, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.193981652375738e-05, |
|
"loss": 3.4307, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.189485227750981e-05, |
|
"loss": 3.4196, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.18497740209947e-05, |
|
"loss": 3.4977, |
|
"step": 1855 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.180458187688549e-05, |
|
"loss": 3.3734, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.175927596816549e-05, |
|
"loss": 3.3519, |
|
"step": 1865 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.171385641812763e-05, |
|
"loss": 3.3348, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.166832335037408e-05, |
|
"loss": 3.3296, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.162267688881596e-05, |
|
"loss": 3.4177, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.157691715767293e-05, |
|
"loss": 3.5309, |
|
"step": 1885 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.153104428147294e-05, |
|
"loss": 3.3949, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.148505838505181e-05, |
|
"loss": 3.4182, |
|
"step": 1895 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.143895959355293e-05, |
|
"loss": 3.4144, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.139274803242697e-05, |
|
"loss": 3.4485, |
|
"step": 1905 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.13464238274314e-05, |
|
"loss": 3.3505, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.12999871046303e-05, |
|
"loss": 3.4214, |
|
"step": 1915 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.125343799039394e-05, |
|
"loss": 3.6238, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.120677661139839e-05, |
|
"loss": 3.5022, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.11600030946253e-05, |
|
"loss": 3.2656, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.111311756736145e-05, |
|
"loss": 3.4656, |
|
"step": 1935 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.106612015719845e-05, |
|
"loss": 3.4179, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.101901099203239e-05, |
|
"loss": 3.4982, |
|
"step": 1945 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.097179020006343e-05, |
|
"loss": 3.4987, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.09244579097956e-05, |
|
"loss": 3.4155, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.087701425003627e-05, |
|
"loss": 3.3359, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.082945934989591e-05, |
|
"loss": 3.3327, |
|
"step": 1965 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.078179333878774e-05, |
|
"loss": 3.4443, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.073401634642733e-05, |
|
"loss": 3.4336, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.068612850283227e-05, |
|
"loss": 3.3127, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.063812993832179e-05, |
|
"loss": 3.402, |
|
"step": 1985 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.059002078351648e-05, |
|
"loss": 3.3454, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.054180116933783e-05, |
|
"loss": 3.5579, |
|
"step": 1995 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.049347122700799e-05, |
|
"loss": 3.313, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.04450310880493e-05, |
|
"loss": 3.3515, |
|
"step": 2005 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.039648088428397e-05, |
|
"loss": 3.4245, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.034782074783379e-05, |
|
"loss": 3.3838, |
|
"step": 2015 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.02990508111197e-05, |
|
"loss": 3.423, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.025017120686142e-05, |
|
"loss": 3.4128, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.02011820680771e-05, |
|
"loss": 3.4144, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.015208352808301e-05, |
|
"loss": 3.4192, |
|
"step": 2035 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.010287572049314e-05, |
|
"loss": 3.4221, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.00535587792188e-05, |
|
"loss": 3.4177, |
|
"step": 2045 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.000413283846831e-05, |
|
"loss": 3.3476, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.995459803274664e-05, |
|
"loss": 3.3584, |
|
"step": 2055 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.990495449685494e-05, |
|
"loss": 3.338, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 8.985520236589036e-05, |
|
"loss": 3.4804, |
|
"step": 2065 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.980534177524547e-05, |
|
"loss": 3.3623, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.97553728606081e-05, |
|
"loss": 3.3682, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.970529575796073e-05, |
|
"loss": 3.3774, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.965511060358043e-05, |
|
"loss": 3.3451, |
|
"step": 2085 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.960481753403815e-05, |
|
"loss": 3.3615, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.955441668619864e-05, |
|
"loss": 3.3598, |
|
"step": 2095 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.950390819721986e-05, |
|
"loss": 3.4915, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.945329220455279e-05, |
|
"loss": 3.2871, |
|
"step": 2105 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.940256884594085e-05, |
|
"loss": 3.3905, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.935173825941973e-05, |
|
"loss": 3.3348, |
|
"step": 2115 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.930080058331691e-05, |
|
"loss": 3.4252, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.924975595625128e-05, |
|
"loss": 3.316, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.919860451713277e-05, |
|
"loss": 3.3096, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.914734640516201e-05, |
|
"loss": 3.3754, |
|
"step": 2135 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.909598175982989e-05, |
|
"loss": 3.3597, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.904451072091724e-05, |
|
"loss": 3.3051, |
|
"step": 2145 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.899293342849444e-05, |
|
"loss": 3.1761, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.894125002292097e-05, |
|
"loss": 3.4439, |
|
"step": 2155 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.888946064484514e-05, |
|
"loss": 3.4385, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 8.883756543520359e-05, |
|
"loss": 3.3402, |
|
"step": 2165 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.8785564535221e-05, |
|
"loss": 3.3721, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.873345808640964e-05, |
|
"loss": 3.3845, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.86812462305691e-05, |
|
"loss": 3.4275, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.862892910978568e-05, |
|
"loss": 3.4157, |
|
"step": 2185 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.857650686643226e-05, |
|
"loss": 3.3359, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.852397964316773e-05, |
|
"loss": 3.3565, |
|
"step": 2195 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.847134758293669e-05, |
|
"loss": 3.3251, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.841861082896905e-05, |
|
"loss": 3.4929, |
|
"step": 2205 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.836576952477959e-05, |
|
"loss": 3.4097, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.831282381416764e-05, |
|
"loss": 3.2644, |
|
"step": 2215 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.825977384121666e-05, |
|
"loss": 3.3957, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.820661975029381e-05, |
|
"loss": 3.3589, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.81533616860496e-05, |
|
"loss": 3.2689, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.80999997934175e-05, |
|
"loss": 3.4637, |
|
"step": 2235 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.804653421761355e-05, |
|
"loss": 3.3386, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.799296510413592e-05, |
|
"loss": 3.4938, |
|
"step": 2245 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.793929259876453e-05, |
|
"loss": 3.3582, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.78855168475607e-05, |
|
"loss": 3.3692, |
|
"step": 2255 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 8.783163799686669e-05, |
|
"loss": 3.2405, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.777765619330534e-05, |
|
"loss": 3.226, |
|
"step": 2265 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.772357158377966e-05, |
|
"loss": 3.2608, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.766938431547242e-05, |
|
"loss": 3.354, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.761509453584579e-05, |
|
"loss": 3.3514, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.75607023926409e-05, |
|
"loss": 3.4127, |
|
"step": 2285 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.75062080338774e-05, |
|
"loss": 3.4036, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.745161160785312e-05, |
|
"loss": 3.3129, |
|
"step": 2295 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.739691326314373e-05, |
|
"loss": 3.2615, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.734211314860215e-05, |
|
"loss": 3.3197, |
|
"step": 2305 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.728721141335829e-05, |
|
"loss": 3.3757, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.723220820681863e-05, |
|
"loss": 3.3568, |
|
"step": 2315 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.717710367866579e-05, |
|
"loss": 3.2747, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.712189797885804e-05, |
|
"loss": 3.3347, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.706659125762905e-05, |
|
"loss": 3.2547, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.701118366548741e-05, |
|
"loss": 3.2634, |
|
"step": 2335 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.695567535321616e-05, |
|
"loss": 3.3791, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.690006647187249e-05, |
|
"loss": 3.3769, |
|
"step": 2345 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.684435717278723e-05, |
|
"loss": 3.3526, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 8.67885476075645e-05, |
|
"loss": 3.4036, |
|
"step": 2355 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.673263792808126e-05, |
|
"loss": 3.3095, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.667662828648695e-05, |
|
"loss": 3.1342, |
|
"step": 2365 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.662051883520304e-05, |
|
"loss": 3.4006, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.656430972692254e-05, |
|
"loss": 3.2912, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.650800111460978e-05, |
|
"loss": 3.315, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.645159315149978e-05, |
|
"loss": 3.2575, |
|
"step": 2385 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.639508599109798e-05, |
|
"loss": 3.4536, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.633847978717972e-05, |
|
"loss": 3.2418, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.628177469378995e-05, |
|
"loss": 3.3163, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.622497086524266e-05, |
|
"loss": 3.3075, |
|
"step": 2405 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.616806845612055e-05, |
|
"loss": 3.3349, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.61110676212746e-05, |
|
"loss": 3.4049, |
|
"step": 2415 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.605396851582366e-05, |
|
"loss": 3.3102, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.599677129515397e-05, |
|
"loss": 3.3275, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.593947611491879e-05, |
|
"loss": 3.3388, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.588208313103798e-05, |
|
"loss": 3.271, |
|
"step": 2435 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.582459249969752e-05, |
|
"loss": 3.3382, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.576700437734916e-05, |
|
"loss": 3.3221, |
|
"step": 2445 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 8.570931892070992e-05, |
|
"loss": 3.2905, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.565153628676175e-05, |
|
"loss": 3.3314, |
|
"step": 2455 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.559365663275101e-05, |
|
"loss": 3.2394, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.553568011618809e-05, |
|
"loss": 3.3375, |
|
"step": 2465 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.547760689484698e-05, |
|
"loss": 3.2946, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.541943712676484e-05, |
|
"loss": 3.3408, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.536117097024158e-05, |
|
"loss": 3.3747, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.530280858383942e-05, |
|
"loss": 3.3417, |
|
"step": 2485 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.524435012638241e-05, |
|
"loss": 3.3327, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.518579575695606e-05, |
|
"loss": 3.2889, |
|
"step": 2495 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.512714563490694e-05, |
|
"loss": 3.3986, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.506839991984212e-05, |
|
"loss": 3.2444, |
|
"step": 2505 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.500955877162886e-05, |
|
"loss": 3.267, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.495062235039411e-05, |
|
"loss": 3.2563, |
|
"step": 2515 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.48915908165241e-05, |
|
"loss": 3.2339, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.483246433066389e-05, |
|
"loss": 3.284, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.477324305371692e-05, |
|
"loss": 3.3038, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.471392714684462e-05, |
|
"loss": 3.3061, |
|
"step": 2535 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.465451677146592e-05, |
|
"loss": 3.4221, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 8.459501208925681e-05, |
|
"loss": 3.3609, |
|
"step": 2545 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.453541326215e-05, |
|
"loss": 3.214, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.44757204523343e-05, |
|
"loss": 3.3066, |
|
"step": 2555 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.441593382225434e-05, |
|
"loss": 3.3019, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.435605353461007e-05, |
|
"loss": 3.1965, |
|
"step": 2565 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.42960797523563e-05, |
|
"loss": 3.3415, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.423601263870222e-05, |
|
"loss": 3.258, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.417585235711112e-05, |
|
"loss": 3.3111, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.41155990712997e-05, |
|
"loss": 3.3403, |
|
"step": 2585 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.405525294523791e-05, |
|
"loss": 3.3113, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.39948141431482e-05, |
|
"loss": 3.2371, |
|
"step": 2595 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.393428282950529e-05, |
|
"loss": 3.224, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.38736591690357e-05, |
|
"loss": 3.2535, |
|
"step": 2605 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.381294332671714e-05, |
|
"loss": 3.2127, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.375213546777829e-05, |
|
"loss": 3.3192, |
|
"step": 2615 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.369123575769822e-05, |
|
"loss": 3.263, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.36302443622059e-05, |
|
"loss": 3.2654, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.356916144727985e-05, |
|
"loss": 3.1785, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.350798717914766e-05, |
|
"loss": 3.2524, |
|
"step": 2635 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.344672172428547e-05, |
|
"loss": 3.2547, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 8.338536524941763e-05, |
|
"loss": 3.2824, |
|
"step": 2645 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.332391792151613e-05, |
|
"loss": 3.2834, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.326237990780028e-05, |
|
"loss": 3.3773, |
|
"step": 2655 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.320075137573609e-05, |
|
"loss": 3.2653, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.313903249303595e-05, |
|
"loss": 3.1791, |
|
"step": 2665 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.307722342765813e-05, |
|
"loss": 3.2213, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.30153243478063e-05, |
|
"loss": 3.2951, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.295333542192907e-05, |
|
"loss": 3.1415, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.289125681871961e-05, |
|
"loss": 3.2968, |
|
"step": 2685 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.282908870711507e-05, |
|
"loss": 3.299, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.276683125629625e-05, |
|
"loss": 3.2173, |
|
"step": 2695 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.270448463568697e-05, |
|
"loss": 3.3282, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.264204901495385e-05, |
|
"loss": 3.3958, |
|
"step": 2705 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.257952456400559e-05, |
|
"loss": 3.3657, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.251691145299269e-05, |
|
"loss": 3.2049, |
|
"step": 2715 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.245420985230691e-05, |
|
"loss": 3.3265, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.239141993258082e-05, |
|
"loss": 3.2992, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.23285418646873e-05, |
|
"loss": 3.3062, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.226557581973919e-05, |
|
"loss": 3.2234, |
|
"step": 2735 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 8.22025219690887e-05, |
|
"loss": 3.2928, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.213938048432697e-05, |
|
"loss": 3.2126, |
|
"step": 2745 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.207615153728365e-05, |
|
"loss": 3.1843, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.201283530002641e-05, |
|
"loss": 3.1663, |
|
"step": 2755 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.194943194486045e-05, |
|
"loss": 3.1322, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.188594164432806e-05, |
|
"loss": 3.2378, |
|
"step": 2765 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.18223645712081e-05, |
|
"loss": 3.2049, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.175870089851564e-05, |
|
"loss": 3.3502, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.169495079950138e-05, |
|
"loss": 3.2646, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.163111444765119e-05, |
|
"loss": 3.3117, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.156719201668567e-05, |
|
"loss": 3.264, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.150318368055973e-05, |
|
"loss": 3.1868, |
|
"step": 2795 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.143908961346197e-05, |
|
"loss": 3.218, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.137490998981437e-05, |
|
"loss": 3.1588, |
|
"step": 2805 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.13106449842717e-05, |
|
"loss": 3.2438, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.124629477172107e-05, |
|
"loss": 3.2773, |
|
"step": 2815 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.118185952728152e-05, |
|
"loss": 3.1023, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.111733942630345e-05, |
|
"loss": 3.2641, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.105273464436816e-05, |
|
"loss": 3.185, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 8.098804535728746e-05, |
|
"loss": 3.2098, |
|
"step": 2835 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.092327174110309e-05, |
|
"loss": 3.2594, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.085841397208628e-05, |
|
"loss": 3.2184, |
|
"step": 2845 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.079347222673727e-05, |
|
"loss": 3.2367, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.072844668178485e-05, |
|
"loss": 3.298, |
|
"step": 2855 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.066333751418583e-05, |
|
"loss": 3.2417, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.059814490112456e-05, |
|
"loss": 3.1765, |
|
"step": 2865 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.053286902001256e-05, |
|
"loss": 3.2471, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.046751004848787e-05, |
|
"loss": 3.1743, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.040206816441469e-05, |
|
"loss": 3.2923, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.033654354588282e-05, |
|
"loss": 3.2636, |
|
"step": 2885 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.027093637120725e-05, |
|
"loss": 3.3305, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.020524681892761e-05, |
|
"loss": 3.2531, |
|
"step": 2895 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.013947506780768e-05, |
|
"loss": 3.15, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.007362129683501e-05, |
|
"loss": 3.1532, |
|
"step": 2905 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 8.000768568522028e-05, |
|
"loss": 3.1407, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.994166841239692e-05, |
|
"loss": 3.2168, |
|
"step": 2915 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.987556965802056e-05, |
|
"loss": 3.1482, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.980938960196861e-05, |
|
"loss": 3.1502, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 7.974312842433971e-05, |
|
"loss": 3.1375, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.967678630545326e-05, |
|
"loss": 3.1295, |
|
"step": 2935 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.961036342584891e-05, |
|
"loss": 3.2147, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.954385996628611e-05, |
|
"loss": 3.2256, |
|
"step": 2945 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.947727610774361e-05, |
|
"loss": 3.272, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.941061203141893e-05, |
|
"loss": 3.2253, |
|
"step": 2955 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.934386791872787e-05, |
|
"loss": 3.2288, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.927704395130408e-05, |
|
"loss": 3.2354, |
|
"step": 2965 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.921014031099852e-05, |
|
"loss": 3.2107, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.914315717987892e-05, |
|
"loss": 3.2168, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.907609474022939e-05, |
|
"loss": 3.2445, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.900895317454981e-05, |
|
"loss": 3.1948, |
|
"step": 2985 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.894173266555545e-05, |
|
"loss": 3.1794, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.887443339617633e-05, |
|
"loss": 3.1999, |
|
"step": 2995 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.880705554955689e-05, |
|
"loss": 3.2405, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.873959930905536e-05, |
|
"loss": 3.2159, |
|
"step": 3005 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.867206485824332e-05, |
|
"loss": 3.2134, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.860445238090518e-05, |
|
"loss": 3.2412, |
|
"step": 3015 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.853676206103768e-05, |
|
"loss": 3.058, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.846899408284942e-05, |
|
"loss": 3.1502, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 7.840114863076031e-05, |
|
"loss": 3.2644, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.833322588940111e-05, |
|
"loss": 3.1445, |
|
"step": 3035 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.826522604361289e-05, |
|
"loss": 3.2321, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.819714927844658e-05, |
|
"loss": 3.2214, |
|
"step": 3045 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.81289957791624e-05, |
|
"loss": 3.1481, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.806076573122942e-05, |
|
"loss": 3.1775, |
|
"step": 3055 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.7992459320325e-05, |
|
"loss": 3.1945, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.792407673233433e-05, |
|
"loss": 3.1555, |
|
"step": 3065 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.78556181533499e-05, |
|
"loss": 3.1334, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.778708376967096e-05, |
|
"loss": 3.1462, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.77184737678031e-05, |
|
"loss": 3.2397, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.76497883344577e-05, |
|
"loss": 3.0973, |
|
"step": 3085 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.758102765655137e-05, |
|
"loss": 3.1067, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.751219192120549e-05, |
|
"loss": 3.2301, |
|
"step": 3095 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.744328131574575e-05, |
|
"loss": 3.1452, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.737429602770152e-05, |
|
"loss": 3.173, |
|
"step": 3105 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.730523624480546e-05, |
|
"loss": 3.2283, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.723610215499291e-05, |
|
"loss": 3.2139, |
|
"step": 3115 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.716689394640148e-05, |
|
"loss": 3.2393, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 7.709761180737043e-05, |
|
"loss": 3.1475, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.702825592644025e-05, |
|
"loss": 3.1161, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.695882649235206e-05, |
|
"loss": 3.2175, |
|
"step": 3135 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.68893236940472e-05, |
|
"loss": 3.0892, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.68197477206666e-05, |
|
"loss": 3.2444, |
|
"step": 3145 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.675009876155038e-05, |
|
"loss": 3.2191, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.668037700623724e-05, |
|
"loss": 3.2492, |
|
"step": 3155 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.661058264446403e-05, |
|
"loss": 3.2339, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.654071586616513e-05, |
|
"loss": 3.1808, |
|
"step": 3165 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.647077686147202e-05, |
|
"loss": 3.1602, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.640076582071271e-05, |
|
"loss": 3.1421, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.63306829344113e-05, |
|
"loss": 3.1525, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.626052839328734e-05, |
|
"loss": 3.1926, |
|
"step": 3185 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.619030238825541e-05, |
|
"loss": 3.1067, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.612000511042457e-05, |
|
"loss": 3.148, |
|
"step": 3195 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.604963675109786e-05, |
|
"loss": 3.1369, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.597919750177168e-05, |
|
"loss": 3.0461, |
|
"step": 3205 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.590868755413543e-05, |
|
"loss": 3.127, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.583810710007087e-05, |
|
"loss": 3.2704, |
|
"step": 3215 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 7.576745633165159e-05, |
|
"loss": 3.2469, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.569673544114264e-05, |
|
"loss": 3.1398, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.562594462099977e-05, |
|
"loss": 3.1535, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.555508406386912e-05, |
|
"loss": 3.0938, |
|
"step": 3235 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.548415396258657e-05, |
|
"loss": 3.1336, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.541315451017725e-05, |
|
"loss": 3.2197, |
|
"step": 3245 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.534208589985507e-05, |
|
"loss": 3.2466, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.527094832502206e-05, |
|
"loss": 3.1807, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.519974197926801e-05, |
|
"loss": 3.1906, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.512846705636975e-05, |
|
"loss": 3.113, |
|
"step": 3265 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.505712375029085e-05, |
|
"loss": 3.0678, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.498571225518096e-05, |
|
"loss": 3.1234, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.491423276537519e-05, |
|
"loss": 3.2494, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.48426854753938e-05, |
|
"loss": 3.1463, |
|
"step": 3285 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.477107057994149e-05, |
|
"loss": 3.1431, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.469938827390697e-05, |
|
"loss": 3.1355, |
|
"step": 3295 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.46276387523624e-05, |
|
"loss": 3.1921, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.455582221056282e-05, |
|
"loss": 3.1494, |
|
"step": 3305 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.448393884394573e-05, |
|
"loss": 3.2458, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 7.441198884813039e-05, |
|
"loss": 3.0447, |
|
"step": 3315 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.433997241891742e-05, |
|
"loss": 3.1718, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.426788975228828e-05, |
|
"loss": 3.1271, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.41957410444046e-05, |
|
"loss": 3.0893, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.412352649160779e-05, |
|
"loss": 3.1669, |
|
"step": 3335 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.405124629041838e-05, |
|
"loss": 3.0632, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.397890063753565e-05, |
|
"loss": 3.1882, |
|
"step": 3345 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.390648972983693e-05, |
|
"loss": 3.1121, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.383401376437716e-05, |
|
"loss": 3.1172, |
|
"step": 3355 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.376147293838827e-05, |
|
"loss": 3.0802, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.368886744927875e-05, |
|
"loss": 3.013, |
|
"step": 3365 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.361619749463306e-05, |
|
"loss": 3.0762, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.354346327221106e-05, |
|
"loss": 3.2118, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.347066497994756e-05, |
|
"loss": 3.2867, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.339780281595166e-05, |
|
"loss": 3.1968, |
|
"step": 3385 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.332487697850633e-05, |
|
"loss": 3.0677, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.325188766606776e-05, |
|
"loss": 3.2039, |
|
"step": 3395 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.317883507726496e-05, |
|
"loss": 3.2133, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.310571941089906e-05, |
|
"loss": 3.054, |
|
"step": 3405 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.303254086594287e-05, |
|
"loss": 3.2114, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 7.295929964154033e-05, |
|
"loss": 3.074, |
|
"step": 3415 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.288599593700592e-05, |
|
"loss": 3.1789, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.281262995182422e-05, |
|
"loss": 3.07, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.273920188564921e-05, |
|
"loss": 3.166, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.266571193830387e-05, |
|
"loss": 3.1676, |
|
"step": 3435 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.259216030977954e-05, |
|
"loss": 3.158, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.251854720023545e-05, |
|
"loss": 3.1018, |
|
"step": 3445 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.244487280999813e-05, |
|
"loss": 3.1207, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.237113733956087e-05, |
|
"loss": 3.1879, |
|
"step": 3455 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.22973409895832e-05, |
|
"loss": 3.0686, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.222348396089029e-05, |
|
"loss": 3.1069, |
|
"step": 3465 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.214956645447244e-05, |
|
"loss": 3.1542, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.207558867148458e-05, |
|
"loss": 3.0487, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.200155081324565e-05, |
|
"loss": 3.1688, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.192745308123802e-05, |
|
"loss": 3.1539, |
|
"step": 3485 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.185329567710706e-05, |
|
"loss": 3.1241, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.177907880266051e-05, |
|
"loss": 3.164, |
|
"step": 3495 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.170480265986798e-05, |
|
"loss": 3.1179, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.16304674508603e-05, |
|
"loss": 3.1395, |
|
"step": 3505 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 7.155607337792912e-05, |
|
"loss": 3.1721, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.148162064352619e-05, |
|
"loss": 3.2053, |
|
"step": 3515 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.140710945026296e-05, |
|
"loss": 3.1553, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.133254000090997e-05, |
|
"loss": 3.129, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.125791249839626e-05, |
|
"loss": 3.2526, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.118322714580888e-05, |
|
"loss": 3.1089, |
|
"step": 3535 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.110848414639229e-05, |
|
"loss": 3.2093, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.103368370354783e-05, |
|
"loss": 3.2533, |
|
"step": 3545 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.095882602083322e-05, |
|
"loss": 3.1246, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.088391130196184e-05, |
|
"loss": 3.0657, |
|
"step": 3555 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.080893975080237e-05, |
|
"loss": 3.2354, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.07339115713781e-05, |
|
"loss": 3.1765, |
|
"step": 3565 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.065882696786646e-05, |
|
"loss": 3.1773, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.058368614459845e-05, |
|
"loss": 3.1519, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.0508489306058e-05, |
|
"loss": 3.0885, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.043323665688153e-05, |
|
"loss": 3.1815, |
|
"step": 3585 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.035792840185728e-05, |
|
"loss": 3.0248, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.02825647459249e-05, |
|
"loss": 3.1001, |
|
"step": 3595 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.020714589417471e-05, |
|
"loss": 3.1015, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 7.013167205184732e-05, |
|
"loss": 3.085, |
|
"step": 3605 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 7.005614342433291e-05, |
|
"loss": 3.1053, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.998056021717083e-05, |
|
"loss": 3.1589, |
|
"step": 3615 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.990492263604889e-05, |
|
"loss": 3.1577, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.982923088680293e-05, |
|
"loss": 3.2004, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.975348517541614e-05, |
|
"loss": 3.1613, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.967768570801862e-05, |
|
"loss": 3.1487, |
|
"step": 3635 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.960183269088669e-05, |
|
"loss": 3.1031, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.952592633044247e-05, |
|
"loss": 3.1244, |
|
"step": 3645 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.944996683325319e-05, |
|
"loss": 3.2323, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.937395440603073e-05, |
|
"loss": 3.1502, |
|
"step": 3655 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.929788925563095e-05, |
|
"loss": 3.1587, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.922177158905325e-05, |
|
"loss": 3.0308, |
|
"step": 3665 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.914560161343991e-05, |
|
"loss": 3.0768, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.906937953607555e-05, |
|
"loss": 3.0908, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.899310556438663e-05, |
|
"loss": 3.1149, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.891677990594075e-05, |
|
"loss": 3.0646, |
|
"step": 3685 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.884040276844627e-05, |
|
"loss": 3.1447, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.876397435975152e-05, |
|
"loss": 3.1473, |
|
"step": 3695 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 6.86874948878445e-05, |
|
"loss": 3.1373, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.861096456085206e-05, |
|
"loss": 2.9776, |
|
"step": 3705 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.853438358703946e-05, |
|
"loss": 3.1365, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.845775217480981e-05, |
|
"loss": 3.164, |
|
"step": 3715 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.838107053270348e-05, |
|
"loss": 3.0909, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.830433886939754e-05, |
|
"loss": 3.0687, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.822755739370518e-05, |
|
"loss": 3.146, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.815072631457512e-05, |
|
"loss": 3.0139, |
|
"step": 3735 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.80738458410911e-05, |
|
"loss": 3.1475, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.799691618247124e-05, |
|
"loss": 3.1542, |
|
"step": 3745 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.791993754806757e-05, |
|
"loss": 3.084, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.784291014736534e-05, |
|
"loss": 3.0876, |
|
"step": 3755 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.776583418998255e-05, |
|
"loss": 3.1758, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.768870988566934e-05, |
|
"loss": 3.1975, |
|
"step": 3765 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.761153744430739e-05, |
|
"loss": 3.0681, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.753431707590939e-05, |
|
"loss": 3.0481, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.745704899061843e-05, |
|
"loss": 3.1188, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.737973339870753e-05, |
|
"loss": 3.105, |
|
"step": 3785 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.730237051057892e-05, |
|
"loss": 2.9708, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.722496053676356e-05, |
|
"loss": 3.0351, |
|
"step": 3795 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.714750368792055e-05, |
|
"loss": 3.1342, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.707000017483655e-05, |
|
"loss": 3.2341, |
|
"step": 3805 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.699245020842524e-05, |
|
"loss": 3.0404, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.691485399972665e-05, |
|
"loss": 3.1263, |
|
"step": 3815 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.68372117599067e-05, |
|
"loss": 3.1401, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.675952370025656e-05, |
|
"loss": 3.0714, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.66817900321921e-05, |
|
"loss": 3.095, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.660401096725331e-05, |
|
"loss": 2.995, |
|
"step": 3835 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.652618671710369e-05, |
|
"loss": 3.0723, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.644831749352974e-05, |
|
"loss": 3.1983, |
|
"step": 3845 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.637040350844035e-05, |
|
"loss": 3.2112, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.629244497386619e-05, |
|
"loss": 3.0518, |
|
"step": 3855 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.62144421019592e-05, |
|
"loss": 2.9698, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.613639510499193e-05, |
|
"loss": 3.0943, |
|
"step": 3865 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.605830419535702e-05, |
|
"loss": 3.0889, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.598016958556667e-05, |
|
"loss": 3.1251, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.590199148825197e-05, |
|
"loss": 3.1085, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.582377011616231e-05, |
|
"loss": 3.0247, |
|
"step": 3885 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.574550568216491e-05, |
|
"loss": 3.1328, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 6.566719839924412e-05, |
|
"loss": 3.1151, |
|
"step": 3895 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.558884848050094e-05, |
|
"loss": 3.04, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.55104561391524e-05, |
|
"loss": 3.1058, |
|
"step": 3905 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.543202158853091e-05, |
|
"loss": 3.0598, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.535354504208385e-05, |
|
"loss": 3.0513, |
|
"step": 3915 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.527502671337278e-05, |
|
"loss": 3.0771, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.519646681607305e-05, |
|
"loss": 3.1076, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.511786556397308e-05, |
|
"loss": 3.1108, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.503922317097385e-05, |
|
"loss": 3.1839, |
|
"step": 3935 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.496053985108829e-05, |
|
"loss": 3.1131, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.48818158184407e-05, |
|
"loss": 3.0072, |
|
"step": 3945 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.480305128726618e-05, |
|
"loss": 3.0962, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.472424647191007e-05, |
|
"loss": 3.1292, |
|
"step": 3955 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.464540158682731e-05, |
|
"loss": 3.0864, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.456651684658188e-05, |
|
"loss": 3.0322, |
|
"step": 3965 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.44875924658462e-05, |
|
"loss": 3.1003, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.44086286594006e-05, |
|
"loss": 3.1056, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.432962564213272e-05, |
|
"loss": 3.1058, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.425058362903684e-05, |
|
"loss": 2.9969, |
|
"step": 3985 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.417150283521341e-05, |
|
"loss": 3.07, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.40923834758684e-05, |
|
"loss": 3.038, |
|
"step": 3995 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.401322576631276e-05, |
|
"loss": 3.01, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.393402992196176e-05, |
|
"loss": 3.0808, |
|
"step": 4005 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.385479615833445e-05, |
|
"loss": 3.0561, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.37755246910531e-05, |
|
"loss": 2.95, |
|
"step": 4015 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.369621573584256e-05, |
|
"loss": 3.0461, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.361686950852971e-05, |
|
"loss": 3.0327, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.353748622504288e-05, |
|
"loss": 2.9993, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.345806610141121e-05, |
|
"loss": 3.0402, |
|
"step": 4035 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.33786093537641e-05, |
|
"loss": 3.0468, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.329911619833063e-05, |
|
"loss": 3.1014, |
|
"step": 4045 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.321958685143894e-05, |
|
"loss": 3.0086, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.314002152951569e-05, |
|
"loss": 3.1249, |
|
"step": 4055 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.30604204490854e-05, |
|
"loss": 3.0183, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.298078382676993e-05, |
|
"loss": 3.0314, |
|
"step": 4065 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.290111187928784e-05, |
|
"loss": 3.0055, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.282140482345387e-05, |
|
"loss": 2.9902, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.274166287617824e-05, |
|
"loss": 3.0762, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 6.266188625446615e-05, |
|
"loss": 3.1289, |
|
"step": 4085 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.258207517541717e-05, |
|
"loss": 3.0808, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.250222985622461e-05, |
|
"loss": 3.0976, |
|
"step": 4095 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.2422350514175e-05, |
|
"loss": 3.1466, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.234243736664742e-05, |
|
"loss": 3.0128, |
|
"step": 4105 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.226249063111299e-05, |
|
"loss": 3.0281, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.218251052513418e-05, |
|
"loss": 2.986, |
|
"step": 4115 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.210249726636431e-05, |
|
"loss": 3.1206, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.202245107254693e-05, |
|
"loss": 3.0731, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.194237216151516e-05, |
|
"loss": 3.036, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.186226075119123e-05, |
|
"loss": 3.1311, |
|
"step": 4135 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.178211705958572e-05, |
|
"loss": 3.1126, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.170194130479717e-05, |
|
"loss": 2.972, |
|
"step": 4145 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.16217337050113e-05, |
|
"loss": 3.0654, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.15414944785005e-05, |
|
"loss": 3.0943, |
|
"step": 4155 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.146122384362326e-05, |
|
"loss": 3.0756, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.138092201882349e-05, |
|
"loss": 3.0197, |
|
"step": 4165 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.130058922263e-05, |
|
"loss": 3.1162, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.122022567365592e-05, |
|
"loss": 3.087, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.113983159059803e-05, |
|
"loss": 3.0941, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 6.105940719223621e-05, |
|
"loss": 2.9955, |
|
"step": 4185 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.097895269743287e-05, |
|
"loss": 3.0133, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.089846832513224e-05, |
|
"loss": 3.0481, |
|
"step": 4195 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.0817954294359955e-05, |
|
"loss": 2.8821, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.0737410824222316e-05, |
|
"loss": 3.035, |
|
"step": 4205 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.0656838133905727e-05, |
|
"loss": 2.9994, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.0576236442676104e-05, |
|
"loss": 3.1524, |
|
"step": 4215 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.049560596987833e-05, |
|
"loss": 3.0584, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.04149469349356e-05, |
|
"loss": 2.9864, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.03342595573488e-05, |
|
"loss": 3.0731, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.0253544056695976e-05, |
|
"loss": 3.0851, |
|
"step": 4235 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.01728006526317e-05, |
|
"loss": 3.0273, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.009202956488647e-05, |
|
"loss": 3.09, |
|
"step": 4245 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 6.001123101326613e-05, |
|
"loss": 3.0294, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.9930405217651274e-05, |
|
"loss": 3.0173, |
|
"step": 4255 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.984955239799661e-05, |
|
"loss": 3.1223, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.976867277433043e-05, |
|
"loss": 3.0614, |
|
"step": 4265 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.968776656675389e-05, |
|
"loss": 2.9893, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.9606833995440584e-05, |
|
"loss": 3.0115, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 5.952587528063579e-05, |
|
"loss": 3.1049, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.944489064265595e-05, |
|
"loss": 3.1301, |
|
"step": 4285 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.936388030188804e-05, |
|
"loss": 3.06, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.9282844478789014e-05, |
|
"loss": 2.9623, |
|
"step": 4295 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.920178339388513e-05, |
|
"loss": 3.0338, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.912069726777144e-05, |
|
"loss": 3.0864, |
|
"step": 4305 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.9039586321111096e-05, |
|
"loss": 3.0815, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.895845077463482e-05, |
|
"loss": 3.1435, |
|
"step": 4315 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.887729084914024e-05, |
|
"loss": 2.9858, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.8796106765491424e-05, |
|
"loss": 3.0735, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.8714898744618055e-05, |
|
"loss": 3.0907, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.8633667007515056e-05, |
|
"loss": 3.0273, |
|
"step": 4335 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.8552411775241845e-05, |
|
"loss": 3.0662, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.8471133268921794e-05, |
|
"loss": 3.0199, |
|
"step": 4345 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.8389831709741605e-05, |
|
"loss": 3.0126, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.830850731895071e-05, |
|
"loss": 3.1145, |
|
"step": 4355 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.822716031786069e-05, |
|
"loss": 2.9883, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.8145790927844614e-05, |
|
"loss": 3.0954, |
|
"step": 4365 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.806439937033654e-05, |
|
"loss": 3.0301, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 5.798298586683084e-05, |
|
"loss": 3.0994, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.790155063888155e-05, |
|
"loss": 3.0692, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.78200939081019e-05, |
|
"loss": 3.0307, |
|
"step": 4385 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.7738615896163586e-05, |
|
"loss": 3.057, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.765711682479621e-05, |
|
"loss": 3.0394, |
|
"step": 4395 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.7575596915786754e-05, |
|
"loss": 3.0503, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.749405639097882e-05, |
|
"loss": 3.004, |
|
"step": 4405 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.74124954722722e-05, |
|
"loss": 3.0699, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.7330914381622104e-05, |
|
"loss": 3.1242, |
|
"step": 4415 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.724931334103867e-05, |
|
"loss": 3.0656, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.716769257258634e-05, |
|
"loss": 3.0341, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.708605229838323e-05, |
|
"loss": 3.0247, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.700439274060053e-05, |
|
"loss": 3.0948, |
|
"step": 4435 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.692271412146192e-05, |
|
"loss": 2.9225, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.6841016663242916e-05, |
|
"loss": 2.972, |
|
"step": 4445 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.675930058827039e-05, |
|
"loss": 3.0104, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.66775661189218e-05, |
|
"loss": 3.049, |
|
"step": 4455 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.659581347762466e-05, |
|
"loss": 2.9567, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.651404288685598e-05, |
|
"loss": 3.1455, |
|
"step": 4465 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 5.643225456914156e-05, |
|
"loss": 3.0998, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.635044874705551e-05, |
|
"loss": 3.0398, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.626862564321952e-05, |
|
"loss": 3.0931, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.618678548030235e-05, |
|
"loss": 3.0014, |
|
"step": 4485 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.610492848101915e-05, |
|
"loss": 2.9121, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.602305486813089e-05, |
|
"loss": 3.0461, |
|
"step": 4495 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.5941164864443785e-05, |
|
"loss": 3.1352, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.585925869280861e-05, |
|
"loss": 2.9522, |
|
"step": 4505 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.577733657612019e-05, |
|
"loss": 3.0116, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.5695398737316685e-05, |
|
"loss": 3.0293, |
|
"step": 4515 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.561344539937907e-05, |
|
"loss": 2.9686, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.55314767853305e-05, |
|
"loss": 3.0624, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.5449493118235715e-05, |
|
"loss": 3.0677, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.536749462120039e-05, |
|
"loss": 3.0206, |
|
"step": 4535 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.528548151737056e-05, |
|
"loss": 3.0829, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.520345402993203e-05, |
|
"loss": 3.1213, |
|
"step": 4545 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.512141238210974e-05, |
|
"loss": 3.0561, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.5039356797167155e-05, |
|
"loss": 3.1096, |
|
"step": 4555 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.4957287498405686e-05, |
|
"loss": 2.9436, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.487520470916404e-05, |
|
"loss": 3.0278, |
|
"step": 4565 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 5.479310865281766e-05, |
|
"loss": 3.0438, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.471099955277809e-05, |
|
"loss": 3.0986, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.462887763249235e-05, |
|
"loss": 2.9591, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.454674311544235e-05, |
|
"loss": 2.9862, |
|
"step": 4585 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.446459622514433e-05, |
|
"loss": 3.0033, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.4382437185148125e-05, |
|
"loss": 2.9845, |
|
"step": 4595 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.430026621903669e-05, |
|
"loss": 2.9696, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.4218083550425416e-05, |
|
"loss": 2.9929, |
|
"step": 4605 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.4135889402961547e-05, |
|
"loss": 3.1634, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.405368400032357e-05, |
|
"loss": 3.0643, |
|
"step": 4615 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.3971467566220546e-05, |
|
"loss": 3.0154, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.3889240324391645e-05, |
|
"loss": 3.0445, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.380700249860538e-05, |
|
"loss": 3.0543, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.372475431265912e-05, |
|
"loss": 2.9665, |
|
"step": 4635 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.3642495990378386e-05, |
|
"loss": 3.0134, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.356022775561627e-05, |
|
"loss": 3.0128, |
|
"step": 4645 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.347794983225289e-05, |
|
"loss": 3.0196, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.339566244419471e-05, |
|
"loss": 3.1539, |
|
"step": 4655 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.331336581537395e-05, |
|
"loss": 2.9392, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 5.323106016974795e-05, |
|
"loss": 3.0064, |
|
"step": 4665 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.314874573129862e-05, |
|
"loss": 2.903, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.3066422724031794e-05, |
|
"loss": 3.0323, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.298409137197663e-05, |
|
"loss": 3.0275, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.2901751899184984e-05, |
|
"loss": 2.9456, |
|
"step": 4685 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.2819404529730796e-05, |
|
"loss": 3.0856, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.273704948770951e-05, |
|
"loss": 3.0298, |
|
"step": 4695 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.265468699723748e-05, |
|
"loss": 2.8966, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.2572317282451286e-05, |
|
"loss": 2.9719, |
|
"step": 4705 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.2489940567507165e-05, |
|
"loss": 2.9296, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.240755707658046e-05, |
|
"loss": 2.9683, |
|
"step": 4715 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.232516703386489e-05, |
|
"loss": 2.9552, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.2242770663572027e-05, |
|
"loss": 3.03, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.2160368189930694e-05, |
|
"loss": 2.9519, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.2077959837186263e-05, |
|
"loss": 3.116, |
|
"step": 4735 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.1995545829600166e-05, |
|
"loss": 3.0781, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.191312639144915e-05, |
|
"loss": 3.0392, |
|
"step": 4745 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.1830701747024825e-05, |
|
"loss": 3.0529, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.1748272120632925e-05, |
|
"loss": 2.9831, |
|
"step": 4755 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 5.1665837736592736e-05, |
|
"loss": 3.0534, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.15833988192365e-05, |
|
"loss": 2.9701, |
|
"step": 4765 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.15009555929088e-05, |
|
"loss": 2.9424, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.141850828196595e-05, |
|
"loss": 3.005, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.133605711077536e-05, |
|
"loss": 3.0597, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.1253602303714945e-05, |
|
"loss": 3.0784, |
|
"step": 4785 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.1171144085172553e-05, |
|
"loss": 3.0135, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.108868267954526e-05, |
|
"loss": 3.0173, |
|
"step": 4795 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.1006218311238884e-05, |
|
"loss": 2.9668, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.092375120466724e-05, |
|
"loss": 2.9921, |
|
"step": 4805 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.084128158425161e-05, |
|
"loss": 2.9828, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.075880967442014e-05, |
|
"loss": 2.9582, |
|
"step": 4815 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.067633569960718e-05, |
|
"loss": 3.0719, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.059385988425274e-05, |
|
"loss": 2.9479, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.0511382452801794e-05, |
|
"loss": 3.0166, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.0428903629703737e-05, |
|
"loss": 3.0477, |
|
"step": 4835 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.034642363941174e-05, |
|
"loss": 3.0076, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.026394270638215e-05, |
|
"loss": 3.0304, |
|
"step": 4845 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.0181461055073907e-05, |
|
"loss": 3.0762, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 5.009897890994788e-05, |
|
"loss": 2.9893, |
|
"step": 4855 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 5.001649649546626e-05, |
|
"loss": 2.9915, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.993401403609203e-05, |
|
"loss": 3.0271, |
|
"step": 4865 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.9851531756288234e-05, |
|
"loss": 3.0551, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.976904988051748e-05, |
|
"loss": 2.9895, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.968656863324124e-05, |
|
"loss": 3.0064, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.96040882389193e-05, |
|
"loss": 3.074, |
|
"step": 4885 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.952160892200909e-05, |
|
"loss": 3.1097, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.9439130906965144e-05, |
|
"loss": 2.9972, |
|
"step": 4895 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.9356654418238444e-05, |
|
"loss": 2.9557, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.927417968027581e-05, |
|
"loss": 2.9784, |
|
"step": 4905 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.91917069175193e-05, |
|
"loss": 2.9535, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.910923635440558e-05, |
|
"loss": 2.94, |
|
"step": 4915 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.9026768215365386e-05, |
|
"loss": 3.0684, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.8944302724822793e-05, |
|
"loss": 3.0555, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.886184010719471e-05, |
|
"loss": 2.9557, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.877938058689022e-05, |
|
"loss": 2.9694, |
|
"step": 4935 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.8696924388309946e-05, |
|
"loss": 2.9223, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.86144717358455e-05, |
|
"loss": 2.9208, |
|
"step": 4945 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.853202285387886e-05, |
|
"loss": 3.019, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.84495779667817e-05, |
|
"loss": 3.043, |
|
"step": 4955 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.836713729891487e-05, |
|
"loss": 2.9888, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.8284701074627694e-05, |
|
"loss": 2.9286, |
|
"step": 4965 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.820226951825741e-05, |
|
"loss": 2.9686, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.81198428541286e-05, |
|
"loss": 3.1102, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.8037421306552485e-05, |
|
"loss": 3.0098, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.795500509982636e-05, |
|
"loss": 2.9617, |
|
"step": 4985 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.7872594458233e-05, |
|
"loss": 3.0371, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.779018960604007e-05, |
|
"loss": 3.0059, |
|
"step": 4995 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.770779076749941e-05, |
|
"loss": 3.022, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.7625398166846565e-05, |
|
"loss": 2.9216, |
|
"step": 5005 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.754301202830006e-05, |
|
"loss": 3.0249, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.746063257606082e-05, |
|
"loss": 2.8526, |
|
"step": 5015 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.737826003431162e-05, |
|
"loss": 3.0094, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.729589462721642e-05, |
|
"loss": 3.1334, |
|
"step": 5025 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.7213536578919734e-05, |
|
"loss": 3.0408, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.7131186113546076e-05, |
|
"loss": 2.9893, |
|
"step": 5035 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.704884345519932e-05, |
|
"loss": 3.0477, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.6966508827962075e-05, |
|
"loss": 2.9936, |
|
"step": 5045 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 4.688418245589514e-05, |
|
"loss": 3.034, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.6801864563036805e-05, |
|
"loss": 2.982, |
|
"step": 5055 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.6719555373402294e-05, |
|
"loss": 3.0017, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.6637255110983154e-05, |
|
"loss": 3.0727, |
|
"step": 5065 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.655496399974664e-05, |
|
"loss": 2.9249, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.6472682263635106e-05, |
|
"loss": 2.974, |
|
"step": 5075 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.63904101265654e-05, |
|
"loss": 3.0252, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.630814781242821e-05, |
|
"loss": 2.8929, |
|
"step": 5085 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.6225895545087535e-05, |
|
"loss": 2.9902, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.614365354838e-05, |
|
"loss": 3.0188, |
|
"step": 5095 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.60614220461143e-05, |
|
"loss": 2.9119, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.597920126207058e-05, |
|
"loss": 2.952, |
|
"step": 5105 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.5896991419999796e-05, |
|
"loss": 2.893, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.581479274362314e-05, |
|
"loss": 2.9398, |
|
"step": 5115 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.5732605456631386e-05, |
|
"loss": 2.93, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.565042978268438e-05, |
|
"loss": 2.9845, |
|
"step": 5125 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.55682659454103e-05, |
|
"loss": 2.9244, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.5486114168405156e-05, |
|
"loss": 3.0629, |
|
"step": 5135 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.5403974675232104e-05, |
|
"loss": 2.9369, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.532184768942089e-05, |
|
"loss": 2.9826, |
|
"step": 5145 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.523973343446723e-05, |
|
"loss": 3.097, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.5157632133832185e-05, |
|
"loss": 3.043, |
|
"step": 5155 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.5075544010941564e-05, |
|
"loss": 3.0134, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.499346928918531e-05, |
|
"loss": 2.988, |
|
"step": 5165 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.491140819191691e-05, |
|
"loss": 3.057, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.482936094245274e-05, |
|
"loss": 2.9222, |
|
"step": 5175 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.4747327764071543e-05, |
|
"loss": 3.0549, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.466530888001374e-05, |
|
"loss": 3.0851, |
|
"step": 5185 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.458330451348085e-05, |
|
"loss": 3.0138, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.450131488763488e-05, |
|
"loss": 3.032, |
|
"step": 5195 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.441934022559776e-05, |
|
"loss": 3.0443, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.433738075045065e-05, |
|
"loss": 3.0077, |
|
"step": 5205 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.425543668523341e-05, |
|
"loss": 3.1165, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.417350825294395e-05, |
|
"loss": 2.9383, |
|
"step": 5215 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.409159567653765e-05, |
|
"loss": 2.9451, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.400969917892674e-05, |
|
"loss": 2.9542, |
|
"step": 5225 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.392781898297969e-05, |
|
"loss": 3.0694, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.384595531152061e-05, |
|
"loss": 2.9181, |
|
"step": 5235 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 4.376410838732863e-05, |
|
"loss": 2.9679, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.368227843313731e-05, |
|
"loss": 2.9548, |
|
"step": 5245 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.360046567163403e-05, |
|
"loss": 3.0431, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.351867032545939e-05, |
|
"loss": 2.9602, |
|
"step": 5255 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.3436892617206605e-05, |
|
"loss": 2.908, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.335513276942086e-05, |
|
"loss": 2.9248, |
|
"step": 5265 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.3273391004598745e-05, |
|
"loss": 2.9548, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.319166754518768e-05, |
|
"loss": 2.9168, |
|
"step": 5275 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.3109962613585214e-05, |
|
"loss": 2.8516, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.302827643213851e-05, |
|
"loss": 2.9677, |
|
"step": 5285 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.294660922314367e-05, |
|
"loss": 3.1188, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.286496120884522e-05, |
|
"loss": 2.9361, |
|
"step": 5295 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.278333261143541e-05, |
|
"loss": 2.9715, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.270172365305367e-05, |
|
"loss": 2.9817, |
|
"step": 5305 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.2620134555785976e-05, |
|
"loss": 2.878, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.2538565541664245e-05, |
|
"loss": 3.0183, |
|
"step": 5315 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.245701683266575e-05, |
|
"loss": 2.9741, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.237548865071251e-05, |
|
"loss": 3.0651, |
|
"step": 5325 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.229398121767069e-05, |
|
"loss": 3.0065, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.221249475534997e-05, |
|
"loss": 2.9313, |
|
"step": 5335 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.2131029485502995e-05, |
|
"loss": 3.0811, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.204958562982468e-05, |
|
"loss": 2.9344, |
|
"step": 5345 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.196816340995172e-05, |
|
"loss": 2.9898, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.1886763047461925e-05, |
|
"loss": 3.1129, |
|
"step": 5355 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.180538476387359e-05, |
|
"loss": 3.0285, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.172402878064495e-05, |
|
"loss": 3.0211, |
|
"step": 5365 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.164269531917354e-05, |
|
"loss": 3.0428, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.156138460079563e-05, |
|
"loss": 2.9048, |
|
"step": 5375 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.14800968467856e-05, |
|
"loss": 2.9882, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.1398832278355296e-05, |
|
"loss": 3.0682, |
|
"step": 5385 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.131759111665349e-05, |
|
"loss": 3.0709, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.123637358276526e-05, |
|
"loss": 2.898, |
|
"step": 5395 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.1155179897711374e-05, |
|
"loss": 2.9611, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.107401028244773e-05, |
|
"loss": 3.0413, |
|
"step": 5405 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.0992864957864694e-05, |
|
"loss": 2.9746, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.0911744144786523e-05, |
|
"loss": 3.0358, |
|
"step": 5415 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.083064806397078e-05, |
|
"loss": 2.9667, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.074957693610775e-05, |
|
"loss": 3.0153, |
|
"step": 5425 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.0668530981819776e-05, |
|
"loss": 2.8481, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 4.0587510421660704e-05, |
|
"loss": 2.9695, |
|
"step": 5435 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.050651547611526e-05, |
|
"loss": 3.0039, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.04255463655985e-05, |
|
"loss": 2.9089, |
|
"step": 5445 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.0344603310455154e-05, |
|
"loss": 3.101, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.026368653095906e-05, |
|
"loss": 2.9567, |
|
"step": 5455 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.0182796247312526e-05, |
|
"loss": 3.0554, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.010193267964576e-05, |
|
"loss": 2.981, |
|
"step": 5465 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 4.0021096048016297e-05, |
|
"loss": 3.0242, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.99402865724083e-05, |
|
"loss": 2.975, |
|
"step": 5475 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.985950447273214e-05, |
|
"loss": 2.9443, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.9778749968823586e-05, |
|
"loss": 2.9989, |
|
"step": 5485 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.969802328044337e-05, |
|
"loss": 2.9051, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.9617324627276487e-05, |
|
"loss": 2.8675, |
|
"step": 5495 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.9536654228931685e-05, |
|
"loss": 2.8715, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.945601230494079e-05, |
|
"loss": 3.0438, |
|
"step": 5505 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.937539907475814e-05, |
|
"loss": 3.0196, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.9294814757759985e-05, |
|
"loss": 2.9482, |
|
"step": 5515 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.921425957324392e-05, |
|
"loss": 3.0152, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.913373374042822e-05, |
|
"loss": 3.037, |
|
"step": 5525 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.905323747845133e-05, |
|
"loss": 2.8559, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.8972771006371193e-05, |
|
"loss": 2.9082, |
|
"step": 5535 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.8892334543164675e-05, |
|
"loss": 2.9559, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.8811928307727005e-05, |
|
"loss": 2.977, |
|
"step": 5545 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.8731552518871115e-05, |
|
"loss": 2.9108, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.8651207395327135e-05, |
|
"loss": 2.9273, |
|
"step": 5555 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.8570893155741686e-05, |
|
"loss": 3.0549, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.849061001867739e-05, |
|
"loss": 3.0003, |
|
"step": 5565 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.8410358202612184e-05, |
|
"loss": 2.7616, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.8330137925938825e-05, |
|
"loss": 3.0067, |
|
"step": 5575 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.82499494069642e-05, |
|
"loss": 2.9336, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.816979286390877e-05, |
|
"loss": 3.0131, |
|
"step": 5585 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.808966851490599e-05, |
|
"loss": 2.9417, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.800957657800173e-05, |
|
"loss": 2.9149, |
|
"step": 5595 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.792951727115359e-05, |
|
"loss": 2.8807, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.784949081223046e-05, |
|
"loss": 2.9032, |
|
"step": 5605 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.7769497419011766e-05, |
|
"loss": 2.9831, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.768953730918699e-05, |
|
"loss": 2.976, |
|
"step": 5615 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.760961070035501e-05, |
|
"loss": 2.9412, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.752971781002355e-05, |
|
"loss": 2.9446, |
|
"step": 5625 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.74498588556086e-05, |
|
"loss": 2.9199, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.737003405443376e-05, |
|
"loss": 2.9459, |
|
"step": 5635 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.72902436237297e-05, |
|
"loss": 2.9315, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.7210487780633564e-05, |
|
"loss": 2.9288, |
|
"step": 5645 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.713076674218836e-05, |
|
"loss": 2.9751, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.70510807253424e-05, |
|
"loss": 2.9068, |
|
"step": 5655 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.6971429946948665e-05, |
|
"loss": 2.9021, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.6891814623764254e-05, |
|
"loss": 3.0042, |
|
"step": 5665 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.6812234972449774e-05, |
|
"loss": 3.0421, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.673269120956875e-05, |
|
"loss": 3.0051, |
|
"step": 5675 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.6653183551587086e-05, |
|
"loss": 2.9744, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.657371221487239e-05, |
|
"loss": 2.9316, |
|
"step": 5685 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.6494277415693424e-05, |
|
"loss": 3.0179, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.6414879370219526e-05, |
|
"loss": 2.9394, |
|
"step": 5695 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.633551829452003e-05, |
|
"loss": 2.8471, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.625619440456366e-05, |
|
"loss": 2.9791, |
|
"step": 5705 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.617690791621793e-05, |
|
"loss": 2.9971, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.609765904524857e-05, |
|
"loss": 2.9016, |
|
"step": 5715 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.601844800731895e-05, |
|
"loss": 2.9531, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.59392750179895e-05, |
|
"loss": 2.9115, |
|
"step": 5725 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.586014029271708e-05, |
|
"loss": 2.9262, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.5781044046854416e-05, |
|
"loss": 2.9121, |
|
"step": 5735 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.5701986495649555e-05, |
|
"loss": 2.9112, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.562296785424518e-05, |
|
"loss": 3.0054, |
|
"step": 5745 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.554398833767816e-05, |
|
"loss": 2.9217, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.546504816087886e-05, |
|
"loss": 3.0008, |
|
"step": 5755 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.538614753867058e-05, |
|
"loss": 2.8544, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.530728668576898e-05, |
|
"loss": 3.0141, |
|
"step": 5765 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.52284658167815e-05, |
|
"loss": 2.8511, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.5149685146206756e-05, |
|
"loss": 3.0225, |
|
"step": 5775 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.507094488843401e-05, |
|
"loss": 2.8771, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.499224525774251e-05, |
|
"loss": 3.0127, |
|
"step": 5785 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.4913586468300934e-05, |
|
"loss": 2.8895, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.483496873416685e-05, |
|
"loss": 2.92, |
|
"step": 5795 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.4756392269286095e-05, |
|
"loss": 2.907, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.467785728749219e-05, |
|
"loss": 2.9166, |
|
"step": 5805 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.4599364002505755e-05, |
|
"loss": 3.049, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.452091262793395e-05, |
|
"loss": 2.9727, |
|
"step": 5815 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.444250337726989e-05, |
|
"loss": 2.8995, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.4364136463892024e-05, |
|
"loss": 2.9885, |
|
"step": 5825 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.428581210106366e-05, |
|
"loss": 3.0731, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.420753050193223e-05, |
|
"loss": 2.8772, |
|
"step": 5835 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.412929187952884e-05, |
|
"loss": 2.9897, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.4051096446767615e-05, |
|
"loss": 3.023, |
|
"step": 5845 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.397294441644515e-05, |
|
"loss": 2.9753, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.3894836001239946e-05, |
|
"loss": 2.97, |
|
"step": 5855 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.3816771413711815e-05, |
|
"loss": 2.8528, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.373875086630125e-05, |
|
"loss": 2.8336, |
|
"step": 5865 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.3660774571328944e-05, |
|
"loss": 2.8747, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.358284274099516e-05, |
|
"loss": 2.9985, |
|
"step": 5875 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.3504955587379147e-05, |
|
"loss": 3.0302, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.342711332243856e-05, |
|
"loss": 2.877, |
|
"step": 5885 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.334931615800892e-05, |
|
"loss": 3.0142, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.3271564305802994e-05, |
|
"loss": 2.949, |
|
"step": 5895 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.319385797741025e-05, |
|
"loss": 3.0545, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.311619738429628e-05, |
|
"loss": 2.9261, |
|
"step": 5905 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.303858273780221e-05, |
|
"loss": 2.8461, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.296101424914411e-05, |
|
"loss": 2.9121, |
|
"step": 5915 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.288349212941246e-05, |
|
"loss": 2.9994, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.280601658957153e-05, |
|
"loss": 2.9453, |
|
"step": 5925 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.272858784045887e-05, |
|
"loss": 2.8655, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.265120609278466e-05, |
|
"loss": 2.9476, |
|
"step": 5935 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.257387155713119e-05, |
|
"loss": 2.9384, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.249658444395224e-05, |
|
"loss": 2.8973, |
|
"step": 5945 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.24193449635726e-05, |
|
"loss": 2.8922, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.234215332618739e-05, |
|
"loss": 2.961, |
|
"step": 5955 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.2265009741861516e-05, |
|
"loss": 3.077, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.2187914420529174e-05, |
|
"loss": 2.8913, |
|
"step": 5965 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.211086757199315e-05, |
|
"loss": 2.949, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.2033869405924386e-05, |
|
"loss": 2.9077, |
|
"step": 5975 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.195692013186127e-05, |
|
"loss": 2.8589, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.1880019959209225e-05, |
|
"loss": 2.9046, |
|
"step": 5985 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.180316909724e-05, |
|
"loss": 2.9405, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.172636775509114e-05, |
|
"loss": 2.7514, |
|
"step": 5995 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.164961614176543e-05, |
|
"loss": 2.9554, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.1572914466130385e-05, |
|
"loss": 2.9045, |
|
"step": 6005 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.1496262936917564e-05, |
|
"loss": 2.9913, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.141966176272208e-05, |
|
"loss": 2.9642, |
|
"step": 6015 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.134311115200199e-05, |
|
"loss": 2.8793, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.126661131307781e-05, |
|
"loss": 2.9875, |
|
"step": 6025 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.119016245413182e-05, |
|
"loss": 2.9724, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.11137647832076e-05, |
|
"loss": 2.9, |
|
"step": 6035 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.103741850820944e-05, |
|
"loss": 2.9434, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.0961123836901715e-05, |
|
"loss": 3.0319, |
|
"step": 6045 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.088488097690844e-05, |
|
"loss": 2.8912, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.080869013571257e-05, |
|
"loss": 2.93, |
|
"step": 6055 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.073255152065555e-05, |
|
"loss": 2.9719, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.065646533893667e-05, |
|
"loss": 2.9573, |
|
"step": 6065 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.058043179761252e-05, |
|
"loss": 2.9199, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.050445110359646e-05, |
|
"loss": 2.9656, |
|
"step": 6075 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.0428523463658044e-05, |
|
"loss": 2.9889, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.0352649084422414e-05, |
|
"loss": 2.8312, |
|
"step": 6085 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.0276828172369796e-05, |
|
"loss": 2.9099, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.020106093383489e-05, |
|
"loss": 2.8909, |
|
"step": 6095 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.012534757500638e-05, |
|
"loss": 3.1077, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.004968830192627e-05, |
|
"loss": 2.9411, |
|
"step": 6105 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 2.9974083320489415e-05, |
|
"loss": 3.0338, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.98985328364429e-05, |
|
"loss": 2.9307, |
|
"step": 6115 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.9823037055385506e-05, |
|
"loss": 2.9608, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.9747596182767178e-05, |
|
"loss": 2.9714, |
|
"step": 6125 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.967221042388838e-05, |
|
"loss": 2.8675, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.9596879983899672e-05, |
|
"loss": 2.9623, |
|
"step": 6135 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.9521605067801016e-05, |
|
"loss": 2.8146, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.9446385880441274e-05, |
|
"loss": 2.8223, |
|
"step": 6145 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.9371222626517648e-05, |
|
"loss": 2.9929, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.9296115510575177e-05, |
|
"loss": 2.816, |
|
"step": 6155 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.9221064737006077e-05, |
|
"loss": 2.9419, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.914607051004924e-05, |
|
"loss": 2.8513, |
|
"step": 6165 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.907113303378972e-05, |
|
"loss": 3.023, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.8996252512158062e-05, |
|
"loss": 2.8467, |
|
"step": 6175 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.8921429148929903e-05, |
|
"loss": 2.8152, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.8846663147725238e-05, |
|
"loss": 2.889, |
|
"step": 6185 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.8771954712008038e-05, |
|
"loss": 3.0059, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.86973040450856e-05, |
|
"loss": 2.9976, |
|
"step": 6195 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.8622711350107966e-05, |
|
"loss": 2.8969, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 2.8548176830067474e-05, |
|
"loss": 2.8465, |
|
"step": 6205 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.8473700687798134e-05, |
|
"loss": 2.8471, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.839928312597506e-05, |
|
"loss": 2.9779, |
|
"step": 6215 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.8324924347114013e-05, |
|
"loss": 2.7989, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.8250624553570694e-05, |
|
"loss": 2.8406, |
|
"step": 6225 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.8176383947540376e-05, |
|
"loss": 2.9092, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.810220273105718e-05, |
|
"loss": 2.9164, |
|
"step": 6235 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.802808110599367e-05, |
|
"loss": 2.834, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.795401927406024e-05, |
|
"loss": 2.9236, |
|
"step": 6245 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.7880017436804497e-05, |
|
"loss": 2.9497, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.780607579561088e-05, |
|
"loss": 2.9739, |
|
"step": 6255 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.7732194551699923e-05, |
|
"loss": 2.9375, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.7658373906127867e-05, |
|
"loss": 2.7406, |
|
"step": 6265 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.7584614059786016e-05, |
|
"loss": 2.9097, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.7510915213400202e-05, |
|
"loss": 2.8839, |
|
"step": 6275 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.743727756753028e-05, |
|
"loss": 2.9065, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.7363701322569557e-05, |
|
"loss": 2.8454, |
|
"step": 6285 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.7290186678744235e-05, |
|
"loss": 2.8946, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.72167338361129e-05, |
|
"loss": 2.8756, |
|
"step": 6295 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 2.7143342994565913e-05, |
|
"loss": 2.9266, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.7070014353824956e-05, |
|
"loss": 3.0255, |
|
"step": 6305 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.6996748113442394e-05, |
|
"loss": 3.0523, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.6923544472800826e-05, |
|
"loss": 3.0364, |
|
"step": 6315 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.6850403631112487e-05, |
|
"loss": 2.9404, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.6777325787418674e-05, |
|
"loss": 2.9068, |
|
"step": 6325 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.6704311140589293e-05, |
|
"loss": 2.9351, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.6631359889322228e-05, |
|
"loss": 2.8579, |
|
"step": 6335 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.6558472232142877e-05, |
|
"loss": 2.9638, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.6485648367403587e-05, |
|
"loss": 2.8894, |
|
"step": 6345 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.641288849328304e-05, |
|
"loss": 2.9169, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.634019280778583e-05, |
|
"loss": 2.9032, |
|
"step": 6355 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.6267561508741868e-05, |
|
"loss": 2.942, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.6194994793805837e-05, |
|
"loss": 2.8697, |
|
"step": 6365 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.612249286045669e-05, |
|
"loss": 3.0204, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.6050055905997018e-05, |
|
"loss": 2.9513, |
|
"step": 6375 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.5977684127552665e-05, |
|
"loss": 2.8773, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.5905377722072045e-05, |
|
"loss": 2.9156, |
|
"step": 6385 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.5833136886325705e-05, |
|
"loss": 3.0443, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 2.576096181690576e-05, |
|
"loss": 2.9093, |
|
"step": 6395 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.5688852710225315e-05, |
|
"loss": 2.9503, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.561680976251802e-05, |
|
"loss": 3.0527, |
|
"step": 6405 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.554483316983741e-05, |
|
"loss": 2.9008, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.5472923128056526e-05, |
|
"loss": 2.8578, |
|
"step": 6415 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.5401079832867265e-05, |
|
"loss": 3.1319, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.5329303479779854e-05, |
|
"loss": 2.8457, |
|
"step": 6425 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.5257594264122397e-05, |
|
"loss": 2.9139, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.5185952381040266e-05, |
|
"loss": 2.985, |
|
"step": 6435 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.5114378025495622e-05, |
|
"loss": 2.8521, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.504287139226684e-05, |
|
"loss": 2.9344, |
|
"step": 6445 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.4971432675947986e-05, |
|
"loss": 3.0079, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.490006207094835e-05, |
|
"loss": 2.9127, |
|
"step": 6455 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.4828759771491794e-05, |
|
"loss": 2.9941, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.4757525971616362e-05, |
|
"loss": 2.9152, |
|
"step": 6465 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.468636086517368e-05, |
|
"loss": 2.9293, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.4615264645828385e-05, |
|
"loss": 2.9036, |
|
"step": 6475 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.4544237507057723e-05, |
|
"loss": 2.9462, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.4473279642150864e-05, |
|
"loss": 2.9845, |
|
"step": 6485 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.4402391244208527e-05, |
|
"loss": 3.0078, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 2.4331572506142392e-05, |
|
"loss": 2.9214, |
|
"step": 6495 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.4260823620674494e-05, |
|
"loss": 2.9839, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.419014478033685e-05, |
|
"loss": 2.9875, |
|
"step": 6505 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.411953617747083e-05, |
|
"loss": 2.9337, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.4048998004226675e-05, |
|
"loss": 2.9176, |
|
"step": 6515 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.3978530452562974e-05, |
|
"loss": 2.9536, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.3908133714246078e-05, |
|
"loss": 2.8946, |
|
"step": 6525 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.3837807980849696e-05, |
|
"loss": 2.9506, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.3767553443754243e-05, |
|
"loss": 2.91, |
|
"step": 6535 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.3697370294146444e-05, |
|
"loss": 2.9307, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.3627258723018737e-05, |
|
"loss": 2.8291, |
|
"step": 6545 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.3557218921168735e-05, |
|
"loss": 2.8649, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.34872510791988e-05, |
|
"loss": 2.9299, |
|
"step": 6555 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.341735538751541e-05, |
|
"loss": 2.7958, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.3347532036328735e-05, |
|
"loss": 2.8647, |
|
"step": 6565 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.327778121565209e-05, |
|
"loss": 2.8582, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.3208103115301354e-05, |
|
"loss": 2.79, |
|
"step": 6575 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.3138497924894558e-05, |
|
"loss": 2.8581, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.306896583385132e-05, |
|
"loss": 2.9129, |
|
"step": 6585 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 2.2999507031392303e-05, |
|
"loss": 2.8519, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.2930121706538764e-05, |
|
"loss": 2.6874, |
|
"step": 6595 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.2860810048111946e-05, |
|
"loss": 2.9205, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.279157224473269e-05, |
|
"loss": 3.0293, |
|
"step": 6605 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.2722408484820767e-05, |
|
"loss": 2.8354, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.2653318956594526e-05, |
|
"loss": 2.9465, |
|
"step": 6615 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.25843038480703e-05, |
|
"loss": 2.976, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.2515363347061837e-05, |
|
"loss": 2.9244, |
|
"step": 6625 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.2446497641179927e-05, |
|
"loss": 2.8918, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.2377706917831753e-05, |
|
"loss": 3.0385, |
|
"step": 6635 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.2308991364220493e-05, |
|
"loss": 2.9077, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.2240351167344763e-05, |
|
"loss": 2.8777, |
|
"step": 6645 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.2171786513998056e-05, |
|
"loss": 3.0116, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.2103297590768334e-05, |
|
"loss": 2.9378, |
|
"step": 6655 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.2034884584037458e-05, |
|
"loss": 2.8471, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.1966547679980692e-05, |
|
"loss": 2.8558, |
|
"step": 6665 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.1898287064566216e-05, |
|
"loss": 2.8937, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.1830102923554546e-05, |
|
"loss": 2.8693, |
|
"step": 6675 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.176199544249817e-05, |
|
"loss": 2.912, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 2.1693964806740868e-05, |
|
"loss": 2.8784, |
|
"step": 6685 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.1626011201417363e-05, |
|
"loss": 2.8303, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.1558134811452745e-05, |
|
"loss": 3.0053, |
|
"step": 6695 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.1490335821561923e-05, |
|
"loss": 2.8993, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.142261441624926e-05, |
|
"loss": 2.9888, |
|
"step": 6705 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.1354970779807882e-05, |
|
"loss": 2.9881, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.128740509631936e-05, |
|
"loss": 2.9555, |
|
"step": 6715 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.1219917549653123e-05, |
|
"loss": 2.9375, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.1152508323465897e-05, |
|
"loss": 2.9425, |
|
"step": 6725 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.108517760120134e-05, |
|
"loss": 2.9474, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.1017925566089435e-05, |
|
"loss": 2.8324, |
|
"step": 6735 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.095075240114606e-05, |
|
"loss": 2.8604, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.0883658289172464e-05, |
|
"loss": 2.863, |
|
"step": 6745 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.0816643412754694e-05, |
|
"loss": 2.9563, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.074970795426327e-05, |
|
"loss": 2.8494, |
|
"step": 6755 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.0682852095852505e-05, |
|
"loss": 2.9259, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.061607601946014e-05, |
|
"loss": 2.9161, |
|
"step": 6765 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.0549379906806815e-05, |
|
"loss": 2.9611, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.0482763939395495e-05, |
|
"loss": 2.706, |
|
"step": 6775 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 2.041622829851112e-05, |
|
"loss": 2.8842, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.0349773165219965e-05, |
|
"loss": 2.7763, |
|
"step": 6785 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.0283398720369267e-05, |
|
"loss": 2.8833, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.0217105144586678e-05, |
|
"loss": 2.9402, |
|
"step": 6795 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.015089261827973e-05, |
|
"loss": 2.8516, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.0084761321635454e-05, |
|
"loss": 2.9836, |
|
"step": 6805 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 2.001871143461978e-05, |
|
"loss": 2.8904, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.995274313697712e-05, |
|
"loss": 2.8632, |
|
"step": 6815 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.9886856608229864e-05, |
|
"loss": 2.8973, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.9821052027677833e-05, |
|
"loss": 2.9326, |
|
"step": 6825 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.9755329574397897e-05, |
|
"loss": 2.9903, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.968968942724337e-05, |
|
"loss": 2.8984, |
|
"step": 6835 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.962413176484362e-05, |
|
"loss": 2.9656, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.9558656765603583e-05, |
|
"loss": 2.9501, |
|
"step": 6845 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.9493264607703138e-05, |
|
"loss": 2.9534, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.9427955469096825e-05, |
|
"loss": 2.8833, |
|
"step": 6855 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.936272952751319e-05, |
|
"loss": 2.859, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.9297586960454404e-05, |
|
"loss": 2.92, |
|
"step": 6865 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.9232527945195778e-05, |
|
"loss": 2.9643, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.9167552658785165e-05, |
|
"loss": 2.927, |
|
"step": 6875 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.910266127804264e-05, |
|
"loss": 2.9116, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.9037853979559923e-05, |
|
"loss": 2.8302, |
|
"step": 6885 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8973130939699858e-05, |
|
"loss": 2.9509, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8908492334596122e-05, |
|
"loss": 2.8555, |
|
"step": 6895 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.884393834015248e-05, |
|
"loss": 3.0024, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8779469132042534e-05, |
|
"loss": 2.9214, |
|
"step": 6905 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8715084885709084e-05, |
|
"loss": 2.8809, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8650785776363772e-05, |
|
"loss": 2.8551, |
|
"step": 6915 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8586571978986562e-05, |
|
"loss": 2.8375, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.852244366832518e-05, |
|
"loss": 2.8239, |
|
"step": 6925 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8458401018894807e-05, |
|
"loss": 2.9718, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8394444204977428e-05, |
|
"loss": 2.9168, |
|
"step": 6935 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.833057340062149e-05, |
|
"loss": 2.8779, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8266788779641396e-05, |
|
"loss": 2.931, |
|
"step": 6945 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8203090515616946e-05, |
|
"loss": 2.928, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8139478781892977e-05, |
|
"loss": 2.9042, |
|
"step": 6955 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8075953751578872e-05, |
|
"loss": 2.8956, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.801251559754798e-05, |
|
"loss": 2.9258, |
|
"step": 6965 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.7949164492437348e-05, |
|
"loss": 2.8822, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.7885900608647028e-05, |
|
"loss": 2.9457, |
|
"step": 6975 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.782272411833978e-05, |
|
"loss": 2.8477, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7759635193440476e-05, |
|
"loss": 2.9183, |
|
"step": 6985 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7696634005635753e-05, |
|
"loss": 2.9364, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7633720726373465e-05, |
|
"loss": 2.8664, |
|
"step": 6995 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7570895526862202e-05, |
|
"loss": 2.8978, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7508158578070928e-05, |
|
"loss": 2.8647, |
|
"step": 7005 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7445510050728358e-05, |
|
"loss": 2.8948, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.738295011532266e-05, |
|
"loss": 2.9046, |
|
"step": 7015 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7320478942100894e-05, |
|
"loss": 2.9088, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7258096701068522e-05, |
|
"loss": 2.9672, |
|
"step": 7025 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7195803561989048e-05, |
|
"loss": 2.8572, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.713359969438349e-05, |
|
"loss": 2.92, |
|
"step": 7035 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7071485267529864e-05, |
|
"loss": 2.8882, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.70094604504629e-05, |
|
"loss": 2.9217, |
|
"step": 7045 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.6947525411973364e-05, |
|
"loss": 2.8347, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.6885680320607776e-05, |
|
"loss": 2.9347, |
|
"step": 7055 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.6823925344667818e-05, |
|
"loss": 2.8641, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.6762260652209983e-05, |
|
"loss": 2.8052, |
|
"step": 7065 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.6700686411045075e-05, |
|
"loss": 2.8709, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.6639202788737694e-05, |
|
"loss": 2.9251, |
|
"step": 7075 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.6577809952605904e-05, |
|
"loss": 2.7836, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.6516508069720653e-05, |
|
"loss": 2.8996, |
|
"step": 7085 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.6455297306905404e-05, |
|
"loss": 2.9278, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.6394177830735668e-05, |
|
"loss": 2.98, |
|
"step": 7095 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.6333149807538473e-05, |
|
"loss": 2.7282, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.6272213403392023e-05, |
|
"loss": 2.856, |
|
"step": 7105 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.6211368784125204e-05, |
|
"loss": 2.9134, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.6150616115317052e-05, |
|
"loss": 2.873, |
|
"step": 7115 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.6089955562296483e-05, |
|
"loss": 3.0006, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.6029387290141633e-05, |
|
"loss": 2.8852, |
|
"step": 7125 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.596891146367958e-05, |
|
"loss": 2.8632, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.5908528247485767e-05, |
|
"loss": 2.8915, |
|
"step": 7135 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.584823780588366e-05, |
|
"loss": 2.9205, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.5788040302944245e-05, |
|
"loss": 2.9478, |
|
"step": 7145 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.5727935902485563e-05, |
|
"loss": 2.9553, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.5667924768072335e-05, |
|
"loss": 2.8743, |
|
"step": 7155 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.560800706301542e-05, |
|
"loss": 2.8947, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.554818295037146e-05, |
|
"loss": 2.8892, |
|
"step": 7165 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.548845259294241e-05, |
|
"loss": 2.927, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.5428816153275038e-05, |
|
"loss": 2.9317, |
|
"step": 7175 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.5369273793660573e-05, |
|
"loss": 2.9011, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.5309825676134217e-05, |
|
"loss": 2.9101, |
|
"step": 7185 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.5250471962474656e-05, |
|
"loss": 2.9748, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.5191212814203765e-05, |
|
"loss": 2.8822, |
|
"step": 7195 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.5132048392585974e-05, |
|
"loss": 2.8518, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.5072978858628006e-05, |
|
"loss": 3.0075, |
|
"step": 7205 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.5014004373078294e-05, |
|
"loss": 2.85, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.4955125096426675e-05, |
|
"loss": 2.7951, |
|
"step": 7215 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.4896341188903867e-05, |
|
"loss": 2.88, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.4837652810481023e-05, |
|
"loss": 3.0368, |
|
"step": 7225 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.4779060120869392e-05, |
|
"loss": 2.7971, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.4720563279519756e-05, |
|
"loss": 2.7829, |
|
"step": 7235 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.4662162445622097e-05, |
|
"loss": 2.844, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.4603857778105152e-05, |
|
"loss": 2.9245, |
|
"step": 7245 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.4545649435635878e-05, |
|
"loss": 2.9121, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.4487537576619176e-05, |
|
"loss": 2.9732, |
|
"step": 7255 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.4429522359197362e-05, |
|
"loss": 2.8346, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.43716039412497e-05, |
|
"loss": 2.824, |
|
"step": 7265 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.4313782480392147e-05, |
|
"loss": 2.9197, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.4256058133976686e-05, |
|
"loss": 2.8023, |
|
"step": 7275 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.41984310590911e-05, |
|
"loss": 2.8808, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.41409014125584e-05, |
|
"loss": 2.887, |
|
"step": 7285 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.4083469350936513e-05, |
|
"loss": 2.9202, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.4026135030517796e-05, |
|
"loss": 2.8866, |
|
"step": 7295 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.3968898607328573e-05, |
|
"loss": 2.8599, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.3911760237128829e-05, |
|
"loss": 2.9011, |
|
"step": 7305 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.3854720075411637e-05, |
|
"loss": 2.9945, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.3797778277402862e-05, |
|
"loss": 2.9268, |
|
"step": 7315 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.3740934998060694e-05, |
|
"loss": 2.7898, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.3684190392075164e-05, |
|
"loss": 2.9648, |
|
"step": 7325 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.3627544613867853e-05, |
|
"loss": 2.812, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.3570997817591318e-05, |
|
"loss": 2.9451, |
|
"step": 7335 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.3514550157128814e-05, |
|
"loss": 2.8403, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.3458201786093794e-05, |
|
"loss": 2.6865, |
|
"step": 7345 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.3401952857829509e-05, |
|
"loss": 2.7784, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.3345803525408606e-05, |
|
"loss": 2.8942, |
|
"step": 7355 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.3289753941632649e-05, |
|
"loss": 2.9349, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.3233804259031813e-05, |
|
"loss": 2.9317, |
|
"step": 7365 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.3177954629864386e-05, |
|
"loss": 2.9265, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.3122205206116345e-05, |
|
"loss": 2.9964, |
|
"step": 7375 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.306655613950103e-05, |
|
"loss": 2.9694, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.3011007581458611e-05, |
|
"loss": 2.7936, |
|
"step": 7385 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2955559683155799e-05, |
|
"loss": 2.879, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2900212595485356e-05, |
|
"loss": 2.8039, |
|
"step": 7395 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2844966469065672e-05, |
|
"loss": 2.8918, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2789821454240435e-05, |
|
"loss": 2.9659, |
|
"step": 7405 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2734777701078133e-05, |
|
"loss": 2.9603, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2679835359371706e-05, |
|
"loss": 2.9423, |
|
"step": 7415 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2624994578638127e-05, |
|
"loss": 2.8943, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2570255508117978e-05, |
|
"loss": 2.8678, |
|
"step": 7425 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2515618296775056e-05, |
|
"loss": 2.8788, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.246108309329594e-05, |
|
"loss": 2.8355, |
|
"step": 7435 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2406650046089635e-05, |
|
"loss": 2.8009, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2352319303287163e-05, |
|
"loss": 2.7742, |
|
"step": 7445 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.229809101274108e-05, |
|
"loss": 2.8486, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2243965322025202e-05, |
|
"loss": 2.8947, |
|
"step": 7455 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.2189942378434083e-05, |
|
"loss": 2.8822, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.2136022328982693e-05, |
|
"loss": 2.8753, |
|
"step": 7465 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.2082205320406004e-05, |
|
"loss": 2.881, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.2028491499158534e-05, |
|
"loss": 2.868, |
|
"step": 7475 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1974881011414046e-05, |
|
"loss": 2.8985, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1921374003065044e-05, |
|
"loss": 2.8125, |
|
"step": 7485 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.186797061972248e-05, |
|
"loss": 2.9751, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1814671006715267e-05, |
|
"loss": 2.9307, |
|
"step": 7495 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1761475309089948e-05, |
|
"loss": 3.0034, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1708383671610284e-05, |
|
"loss": 2.9065, |
|
"step": 7505 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1655396238756805e-05, |
|
"loss": 2.9129, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1602513154726512e-05, |
|
"loss": 2.8419, |
|
"step": 7515 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1549734563432436e-05, |
|
"loss": 2.907, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1497060608503202e-05, |
|
"loss": 2.8648, |
|
"step": 7525 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1444491433282739e-05, |
|
"loss": 2.9375, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1392027180829795e-05, |
|
"loss": 2.8749, |
|
"step": 7535 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1339667993917602e-05, |
|
"loss": 2.8733, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1287414015033493e-05, |
|
"loss": 2.8616, |
|
"step": 7545 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.1235265386378446e-05, |
|
"loss": 2.8642, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.1183222249866793e-05, |
|
"loss": 2.8825, |
|
"step": 7555 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.1131284747125747e-05, |
|
"loss": 2.9617, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.107945301949508e-05, |
|
"loss": 2.9455, |
|
"step": 7565 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.1027727208026711e-05, |
|
"loss": 2.876, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.0976107453484314e-05, |
|
"loss": 2.8741, |
|
"step": 7575 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.0924593896342977e-05, |
|
"loss": 2.8908, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.0873186676788726e-05, |
|
"loss": 2.8954, |
|
"step": 7585 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.0821885934718262e-05, |
|
"loss": 2.872, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.0770691809738514e-05, |
|
"loss": 2.896, |
|
"step": 7595 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.0719604441166247e-05, |
|
"loss": 2.7755, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.066862396802773e-05, |
|
"loss": 2.8576, |
|
"step": 7605 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.0617750529058312e-05, |
|
"loss": 2.8723, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.0566984262702073e-05, |
|
"loss": 2.9218, |
|
"step": 7615 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.0516325307111468e-05, |
|
"loss": 2.8971, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.0465773800146862e-05, |
|
"loss": 2.9214, |
|
"step": 7625 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.0415329879376278e-05, |
|
"loss": 2.8813, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.0364993682074915e-05, |
|
"loss": 2.8865, |
|
"step": 7635 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.0314765345224831e-05, |
|
"loss": 2.9054, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.0264645005514584e-05, |
|
"loss": 2.8731, |
|
"step": 7645 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.0214632799338802e-05, |
|
"loss": 2.8814, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.0164728862797878e-05, |
|
"loss": 2.9242, |
|
"step": 7655 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.0114933331697513e-05, |
|
"loss": 2.8808, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.006524634154845e-05, |
|
"loss": 2.871, |
|
"step": 7665 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.0015668027566056e-05, |
|
"loss": 2.8632, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.966198524669896e-06, |
|
"loss": 2.9597, |
|
"step": 7675 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.91683796748349e-06, |
|
"loss": 2.8929, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.86758649033383e-06, |
|
"loss": 2.8624, |
|
"step": 7685 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.818444227251089e-06, |
|
"loss": 2.8156, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.769411311968247e-06, |
|
"loss": 2.8784, |
|
"step": 7695 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.720487877920658e-06, |
|
"loss": 2.865, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.671674058245805e-06, |
|
"loss": 2.891, |
|
"step": 7705 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.622969985782803e-06, |
|
"loss": 2.8756, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.57437579307217e-06, |
|
"loss": 2.8465, |
|
"step": 7715 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.525891612355364e-06, |
|
"loss": 2.9065, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.47751757557449e-06, |
|
"loss": 2.904, |
|
"step": 7725 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.429253814371897e-06, |
|
"loss": 2.9035, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.381100460089826e-06, |
|
"loss": 2.8806, |
|
"step": 7735 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.333057643770083e-06, |
|
"loss": 2.8941, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.285125496153667e-06, |
|
"loss": 2.8788, |
|
"step": 7745 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.237304147680375e-06, |
|
"loss": 2.9316, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.189593728488533e-06, |
|
"loss": 2.8426, |
|
"step": 7755 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.141994368414524e-06, |
|
"loss": 2.8847, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.094506196992564e-06, |
|
"loss": 2.8059, |
|
"step": 7765 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.047129343454268e-06, |
|
"loss": 2.9604, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.999863936728287e-06, |
|
"loss": 2.7716, |
|
"step": 7775 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.952710105440027e-06, |
|
"loss": 2.8111, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.905667977911225e-06, |
|
"loss": 2.8593, |
|
"step": 7785 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.858737682159663e-06, |
|
"loss": 2.7311, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.811919345898772e-06, |
|
"loss": 2.9675, |
|
"step": 7795 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.765213096537317e-06, |
|
"loss": 2.9184, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.718619061179028e-06, |
|
"loss": 2.8842, |
|
"step": 7805 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.672137366622252e-06, |
|
"loss": 2.8258, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.625768139359631e-06, |
|
"loss": 2.8813, |
|
"step": 7815 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.579511505577759e-06, |
|
"loss": 2.7782, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.533367591156778e-06, |
|
"loss": 2.8901, |
|
"step": 7825 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.487336521670141e-06, |
|
"loss": 2.8051, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.441418422384151e-06, |
|
"loss": 2.9374, |
|
"step": 7835 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.39561341825773e-06, |
|
"loss": 2.8758, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.349921633942021e-06, |
|
"loss": 2.8147, |
|
"step": 7845 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.304343193780017e-06, |
|
"loss": 2.8288, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.258878221806326e-06, |
|
"loss": 2.9349, |
|
"step": 7855 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.213526841746699e-06, |
|
"loss": 2.7969, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.168289177017824e-06, |
|
"loss": 2.8592, |
|
"step": 7865 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.123165350726919e-06, |
|
"loss": 2.7655, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.078155485671357e-06, |
|
"loss": 2.9649, |
|
"step": 7875 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.033259704338475e-06, |
|
"loss": 2.8298, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.988478128905063e-06, |
|
"loss": 2.7694, |
|
"step": 7885 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.943810881237163e-06, |
|
"loss": 2.8258, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.899258082889688e-06, |
|
"loss": 2.8975, |
|
"step": 7895 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.854819855106078e-06, |
|
"loss": 2.8528, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.81049631881801e-06, |
|
"loss": 2.9486, |
|
"step": 7905 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.766287594645017e-06, |
|
"loss": 2.9073, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.722193802894217e-06, |
|
"loss": 2.9155, |
|
"step": 7915 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.678215063559956e-06, |
|
"loss": 3.0423, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.63435149632345e-06, |
|
"loss": 2.7638, |
|
"step": 7925 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.590603220552539e-06, |
|
"loss": 2.94, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 7.546970355301264e-06, |
|
"loss": 2.8597, |
|
"step": 7935 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.5034530193096475e-06, |
|
"loss": 3.0005, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.460051331003287e-06, |
|
"loss": 2.9032, |
|
"step": 7945 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.416765408493042e-06, |
|
"loss": 2.8776, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.373595369574804e-06, |
|
"loss": 2.7591, |
|
"step": 7955 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.33054133172903e-06, |
|
"loss": 2.8881, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.287603412120531e-06, |
|
"loss": 2.9873, |
|
"step": 7965 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.244781727598137e-06, |
|
"loss": 2.8301, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.202076394694324e-06, |
|
"loss": 2.8317, |
|
"step": 7975 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.159487529624975e-06, |
|
"loss": 2.8781, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.117015248288977e-06, |
|
"loss": 2.8637, |
|
"step": 7985 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.074659666268002e-06, |
|
"loss": 2.887, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 7.03242089882612e-06, |
|
"loss": 2.8215, |
|
"step": 7995 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.990299060909494e-06, |
|
"loss": 2.8774, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.948294267146121e-06, |
|
"loss": 2.9101, |
|
"step": 8005 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.906406631845436e-06, |
|
"loss": 2.7999, |
|
"step": 8010 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.86463626899807e-06, |
|
"loss": 2.8326, |
|
"step": 8015 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.822983292275525e-06, |
|
"loss": 2.7986, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.781447815029818e-06, |
|
"loss": 2.8634, |
|
"step": 8025 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 6.740029950293269e-06, |
|
"loss": 2.8693, |
|
"step": 8030 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.698729810778065e-06, |
|
"loss": 2.8484, |
|
"step": 8035 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.657547508876078e-06, |
|
"loss": 2.8413, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.616483156658476e-06, |
|
"loss": 2.8747, |
|
"step": 8045 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.575536865875443e-06, |
|
"loss": 2.8632, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.534708747955903e-06, |
|
"loss": 2.8095, |
|
"step": 8055 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.493998914007149e-06, |
|
"loss": 2.8387, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.453407474814616e-06, |
|
"loss": 2.8881, |
|
"step": 8065 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.412934540841553e-06, |
|
"loss": 2.8793, |
|
"step": 8070 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.3725802222286805e-06, |
|
"loss": 2.7694, |
|
"step": 8075 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.332344628793957e-06, |
|
"loss": 2.9183, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.292227870032225e-06, |
|
"loss": 2.8543, |
|
"step": 8085 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.2522300551149605e-06, |
|
"loss": 2.8417, |
|
"step": 8090 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.212351292889945e-06, |
|
"loss": 2.9838, |
|
"step": 8095 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.172591691880947e-06, |
|
"loss": 2.8515, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.1329513602875195e-06, |
|
"loss": 2.913, |
|
"step": 8105 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.093430405984585e-06, |
|
"loss": 2.8196, |
|
"step": 8110 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.054028936522216e-06, |
|
"loss": 2.8937, |
|
"step": 8115 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 6.014747059125353e-06, |
|
"loss": 2.8529, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 5.975584880693436e-06, |
|
"loss": 2.9127, |
|
"step": 8125 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 5.936542507800208e-06, |
|
"loss": 2.9065, |
|
"step": 8130 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.8976200466933475e-06, |
|
"loss": 2.8673, |
|
"step": 8135 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.8588176032942255e-06, |
|
"loss": 2.8579, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.820135283197625e-06, |
|
"loss": 2.9074, |
|
"step": 8145 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.781573191671386e-06, |
|
"loss": 2.7958, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.743131433656207e-06, |
|
"loss": 2.9314, |
|
"step": 8155 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.704810113765291e-06, |
|
"loss": 2.9389, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.666609336284096e-06, |
|
"loss": 2.9184, |
|
"step": 8165 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.628529205170063e-06, |
|
"loss": 2.8807, |
|
"step": 8170 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.590569824052266e-06, |
|
"loss": 2.8066, |
|
"step": 8175 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.552731296231239e-06, |
|
"loss": 2.8473, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.515013724678569e-06, |
|
"loss": 2.8549, |
|
"step": 8185 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.477417212036723e-06, |
|
"loss": 2.8706, |
|
"step": 8190 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.4399418606187195e-06, |
|
"loss": 2.7804, |
|
"step": 8195 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.402587772407824e-06, |
|
"loss": 2.8332, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.365355049057358e-06, |
|
"loss": 2.8729, |
|
"step": 8205 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.328243791890303e-06, |
|
"loss": 2.9654, |
|
"step": 8210 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.291254101899146e-06, |
|
"loss": 2.7877, |
|
"step": 8215 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.254386079745527e-06, |
|
"loss": 2.8073, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 5.217639825759962e-06, |
|
"loss": 2.8717, |
|
"step": 8225 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 5.181015439941645e-06, |
|
"loss": 2.8088, |
|
"step": 8230 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 5.14451302195807e-06, |
|
"loss": 2.8124, |
|
"step": 8235 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 5.108132671144849e-06, |
|
"loss": 2.913, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 5.071874486505412e-06, |
|
"loss": 2.8362, |
|
"step": 8245 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 5.0357385667106906e-06, |
|
"loss": 2.9117, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.99972501009896e-06, |
|
"loss": 2.8891, |
|
"step": 8255 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.963833914675431e-06, |
|
"loss": 2.9069, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.928065378112106e-06, |
|
"loss": 2.7637, |
|
"step": 8265 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.892419497747447e-06, |
|
"loss": 2.8374, |
|
"step": 8270 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.8568963705861114e-06, |
|
"loss": 2.8136, |
|
"step": 8275 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.821496093298733e-06, |
|
"loss": 2.795, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.786218762221595e-06, |
|
"loss": 2.9271, |
|
"step": 8285 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.751064473356431e-06, |
|
"loss": 2.9181, |
|
"step": 8290 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.716033322370123e-06, |
|
"loss": 2.8116, |
|
"step": 8295 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.6811254045944445e-06, |
|
"loss": 2.8575, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.646340815025824e-06, |
|
"loss": 2.8371, |
|
"step": 8305 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.611679648325057e-06, |
|
"loss": 2.8703, |
|
"step": 8310 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.577141998817075e-06, |
|
"loss": 2.8521, |
|
"step": 8315 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.542727960490684e-06, |
|
"loss": 2.8273, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.508437626998269e-06, |
|
"loss": 2.8532, |
|
"step": 8325 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.474271091655602e-06, |
|
"loss": 2.9128, |
|
"step": 8330 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.440228447441552e-06, |
|
"loss": 2.8762, |
|
"step": 8335 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.406309786997825e-06, |
|
"loss": 2.8786, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.372515202628758e-06, |
|
"loss": 2.8667, |
|
"step": 8345 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.338844786300977e-06, |
|
"loss": 2.7771, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.3052986296432615e-06, |
|
"loss": 2.8549, |
|
"step": 8355 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.271876823946203e-06, |
|
"loss": 2.8453, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.238579460161996e-06, |
|
"loss": 2.8916, |
|
"step": 8365 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.205406628904213e-06, |
|
"loss": 2.7955, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.172358420447492e-06, |
|
"loss": 2.7991, |
|
"step": 8375 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.139434924727359e-06, |
|
"loss": 2.8524, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.106636231339922e-06, |
|
"loss": 2.9343, |
|
"step": 8385 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.0739624295417e-06, |
|
"loss": 2.881, |
|
"step": 8390 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.041413608249312e-06, |
|
"loss": 2.8243, |
|
"step": 8395 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.008989856039264e-06, |
|
"loss": 2.9094, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.976691261147714e-06, |
|
"loss": 2.9381, |
|
"step": 8405 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.944517911470219e-06, |
|
"loss": 2.8515, |
|
"step": 8410 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.9124698945615046e-06, |
|
"loss": 2.8254, |
|
"step": 8415 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.880547297635234e-06, |
|
"loss": 2.7336, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.848750207563717e-06, |
|
"loss": 2.8598, |
|
"step": 8425 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.817078710877775e-06, |
|
"loss": 2.8624, |
|
"step": 8430 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.7855328937663914e-06, |
|
"loss": 2.8908, |
|
"step": 8435 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.7541128420765734e-06, |
|
"loss": 2.8764, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.7228186413130637e-06, |
|
"loss": 2.9249, |
|
"step": 8445 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.6916503766381116e-06, |
|
"loss": 2.9115, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.6606081328712683e-06, |
|
"loss": 2.8248, |
|
"step": 8455 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.629691994489115e-06, |
|
"loss": 2.8825, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.5989020456250854e-06, |
|
"loss": 2.7977, |
|
"step": 8465 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.5682383700691933e-06, |
|
"loss": 2.75, |
|
"step": 8470 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.5377010512678167e-06, |
|
"loss": 2.7819, |
|
"step": 8475 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.507290172323474e-06, |
|
"loss": 2.8601, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.4770058159946106e-06, |
|
"loss": 2.9211, |
|
"step": 8485 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.4468480646953404e-06, |
|
"loss": 2.9062, |
|
"step": 8490 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.4168170004952706e-06, |
|
"loss": 2.8822, |
|
"step": 8495 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.3869127051192105e-06, |
|
"loss": 2.7757, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.3571352599470253e-06, |
|
"loss": 2.9033, |
|
"step": 8505 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.3274847460133372e-06, |
|
"loss": 2.8923, |
|
"step": 8510 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 3.2979612440073857e-06, |
|
"loss": 2.8077, |
|
"step": 8515 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.2685648342727527e-06, |
|
"loss": 2.9168, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.2392955968071404e-06, |
|
"loss": 2.8472, |
|
"step": 8525 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.2101536112622064e-06, |
|
"loss": 2.8074, |
|
"step": 8530 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.1811389569432692e-06, |
|
"loss": 2.8736, |
|
"step": 8535 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.1522517128091677e-06, |
|
"loss": 2.8609, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.1234919574720135e-06, |
|
"loss": 2.8778, |
|
"step": 8545 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.094859769196956e-06, |
|
"loss": 2.9457, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.0663552259020067e-06, |
|
"loss": 2.9051, |
|
"step": 8555 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.0379784051578097e-06, |
|
"loss": 2.8775, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 3.009729384187432e-06, |
|
"loss": 2.7186, |
|
"step": 8565 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.981608239866157e-06, |
|
"loss": 2.8407, |
|
"step": 8570 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.953615048721253e-06, |
|
"loss": 2.8583, |
|
"step": 8575 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.925749886931817e-06, |
|
"loss": 2.9134, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.898012830328495e-06, |
|
"loss": 2.8832, |
|
"step": 8585 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.8704039543933416e-06, |
|
"loss": 2.8545, |
|
"step": 8590 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.8429233342595894e-06, |
|
"loss": 2.8149, |
|
"step": 8595 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.8155710447114236e-06, |
|
"loss": 2.8752, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.788347160183824e-06, |
|
"loss": 2.8725, |
|
"step": 8605 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.761251754762295e-06, |
|
"loss": 2.841, |
|
"step": 8610 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.7342849021827544e-06, |
|
"loss": 2.8485, |
|
"step": 8615 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.70744667583126e-06, |
|
"loss": 2.7597, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.680737148743817e-06, |
|
"loss": 2.8591, |
|
"step": 8625 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.654156393606222e-06, |
|
"loss": 2.8849, |
|
"step": 8630 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.6277044827538346e-06, |
|
"loss": 2.7434, |
|
"step": 8635 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.6013814881713784e-06, |
|
"loss": 2.7657, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.5751874814927522e-06, |
|
"loss": 2.9131, |
|
"step": 8645 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.5491225340008306e-06, |
|
"loss": 2.7557, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.5231867166272905e-06, |
|
"loss": 2.9101, |
|
"step": 8655 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.4973800999523788e-06, |
|
"loss": 2.8779, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.471702754204752e-06, |
|
"loss": 2.8746, |
|
"step": 8665 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.4461547492612923e-06, |
|
"loss": 2.8075, |
|
"step": 8670 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.4207361546468744e-06, |
|
"loss": 2.9346, |
|
"step": 8675 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.395447039534221e-06, |
|
"loss": 2.8424, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.3702874727436876e-06, |
|
"loss": 2.8498, |
|
"step": 8685 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.3452575227430885e-06, |
|
"loss": 2.9311, |
|
"step": 8690 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.32035725764751e-06, |
|
"loss": 2.9577, |
|
"step": 8695 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.295586745219108e-06, |
|
"loss": 2.9036, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.2709460528669513e-06, |
|
"loss": 2.8055, |
|
"step": 8705 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.246435247646822e-06, |
|
"loss": 2.8272, |
|
"step": 8710 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.222054396261025e-06, |
|
"loss": 2.9431, |
|
"step": 8715 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.197803565058232e-06, |
|
"loss": 2.7475, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.1736828200332625e-06, |
|
"loss": 2.9193, |
|
"step": 8725 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.1496922268269493e-06, |
|
"loss": 2.8035, |
|
"step": 8730 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.125831850725918e-06, |
|
"loss": 2.8738, |
|
"step": 8735 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.1021017566624447e-06, |
|
"loss": 2.893, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.0785020092142672e-06, |
|
"loss": 2.9523, |
|
"step": 8745 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.0550326726043734e-06, |
|
"loss": 2.8293, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.031693810700902e-06, |
|
"loss": 2.8159, |
|
"step": 8755 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.0084854870168868e-06, |
|
"loss": 2.8093, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.985407764710151e-06, |
|
"loss": 2.8171, |
|
"step": 8765 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.962460706583097e-06, |
|
"loss": 2.9431, |
|
"step": 8770 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.939644375082528e-06, |
|
"loss": 2.8466, |
|
"step": 8775 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.9169588322995314e-06, |
|
"loss": 2.8888, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.8944041399692248e-06, |
|
"loss": 2.8591, |
|
"step": 8785 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.871980359470693e-06, |
|
"loss": 2.835, |
|
"step": 8790 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.8496875518267286e-06, |
|
"loss": 2.9509, |
|
"step": 8795 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.8275257777036981e-06, |
|
"loss": 2.9184, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.8054950974114137e-06, |
|
"loss": 2.8473, |
|
"step": 8805 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.7835955709028907e-06, |
|
"loss": 2.8985, |
|
"step": 8810 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.7618272577742733e-06, |
|
"loss": 2.9569, |
|
"step": 8815 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.7401902172646034e-06, |
|
"loss": 2.8695, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.7186845082556914e-06, |
|
"loss": 2.8379, |
|
"step": 8825 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.6973101892719556e-06, |
|
"loss": 2.9668, |
|
"step": 8830 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.6760673184802401e-06, |
|
"loss": 2.8445, |
|
"step": 8835 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.6549559536896964e-06, |
|
"loss": 2.835, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.6339761523515907e-06, |
|
"loss": 2.9824, |
|
"step": 8845 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.6131279715591584e-06, |
|
"loss": 2.9564, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.59241146804745e-06, |
|
"loss": 2.7621, |
|
"step": 8855 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.571826698193174e-06, |
|
"loss": 2.8287, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.5513737180145705e-06, |
|
"loss": 2.8275, |
|
"step": 8865 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.5310525831711996e-06, |
|
"loss": 2.9048, |
|
"step": 8870 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.510863348963848e-06, |
|
"loss": 2.8928, |
|
"step": 8875 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.4908060703343495e-06, |
|
"loss": 2.9125, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.4708808018654374e-06, |
|
"loss": 2.8363, |
|
"step": 8885 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.4510875977806092e-06, |
|
"loss": 2.9201, |
|
"step": 8890 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.4314265119439618e-06, |
|
"loss": 2.8847, |
|
"step": 8895 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.4118975978600513e-06, |
|
"loss": 2.9236, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.3925009086737661e-06, |
|
"loss": 2.8667, |
|
"step": 8905 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.373236497170133e-06, |
|
"loss": 2.875, |
|
"step": 8910 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.3541044157742378e-06, |
|
"loss": 3.0322, |
|
"step": 8915 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.3351047165510444e-06, |
|
"loss": 2.835, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.3162374512052433e-06, |
|
"loss": 2.8812, |
|
"step": 8925 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.297502671081141e-06, |
|
"loss": 2.9387, |
|
"step": 8930 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.2789004271624884e-06, |
|
"loss": 2.8301, |
|
"step": 8935 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.2604307700724017e-06, |
|
"loss": 2.8831, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.242093750073131e-06, |
|
"loss": 2.935, |
|
"step": 8945 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.2238894170660098e-06, |
|
"loss": 2.823, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.2058178205912762e-06, |
|
"loss": 2.9259, |
|
"step": 8955 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.1878790098279358e-06, |
|
"loss": 2.8865, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.170073033593655e-06, |
|
"loss": 2.8884, |
|
"step": 8965 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.1523999403446007e-06, |
|
"loss": 2.8522, |
|
"step": 8970 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.134859778175329e-06, |
|
"loss": 2.7615, |
|
"step": 8975 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.1174525948186354e-06, |
|
"loss": 2.8425, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.1001784376454326e-06, |
|
"loss": 2.8441, |
|
"step": 8985 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.0830373536646343e-06, |
|
"loss": 2.9573, |
|
"step": 8990 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.0660293895230156e-06, |
|
"loss": 2.8495, |
|
"step": 8995 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.0491545915050804e-06, |
|
"loss": 2.8729, |
|
"step": 9000 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 9622, |
|
"num_train_epochs": 1, |
|
"save_steps": 1000, |
|
"total_flos": 1.348118192848896e+18, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|