|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.1695906432748537, |
|
"eval_steps": 100, |
|
"global_step": 5000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.005847953216374269, |
|
"grad_norm": 19.25, |
|
"learning_rate": 2.5e-05, |
|
"loss": 1.3094, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.011695906432748537, |
|
"grad_norm": 20.515625, |
|
"learning_rate": 2.4874371859296484e-05, |
|
"loss": 0.6781, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.017543859649122806, |
|
"grad_norm": 9.015625, |
|
"learning_rate": 2.4748743718592964e-05, |
|
"loss": 0.6127, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.023391812865497075, |
|
"grad_norm": 9.4765625, |
|
"learning_rate": 2.462311557788945e-05, |
|
"loss": 0.6338, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.023391812865497075, |
|
"eval_loss": 0.6474161744117737, |
|
"eval_runtime": 125.9492, |
|
"eval_samples_per_second": 3.573, |
|
"eval_steps_per_second": 0.453, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.029239766081871343, |
|
"grad_norm": 16.71875, |
|
"learning_rate": 2.449748743718593e-05, |
|
"loss": 0.506, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.03508771929824561, |
|
"grad_norm": 21.125, |
|
"learning_rate": 2.4371859296482413e-05, |
|
"loss": 0.501, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.04093567251461988, |
|
"grad_norm": 17.953125, |
|
"learning_rate": 2.4246231155778896e-05, |
|
"loss": 0.6208, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.04678362573099415, |
|
"grad_norm": 7.9609375, |
|
"learning_rate": 2.4120603015075376e-05, |
|
"loss": 0.5748, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.04678362573099415, |
|
"eval_loss": 0.5898649096488953, |
|
"eval_runtime": 129.2682, |
|
"eval_samples_per_second": 3.481, |
|
"eval_steps_per_second": 0.441, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.05263157894736842, |
|
"grad_norm": 7.4921875, |
|
"learning_rate": 2.3994974874371863e-05, |
|
"loss": 0.5821, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.05847953216374269, |
|
"grad_norm": 8.703125, |
|
"learning_rate": 2.3869346733668342e-05, |
|
"loss": 0.5764, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.06432748538011696, |
|
"grad_norm": 10.3828125, |
|
"learning_rate": 2.3743718592964825e-05, |
|
"loss": 0.5921, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.07017543859649122, |
|
"grad_norm": 14.90625, |
|
"learning_rate": 2.361809045226131e-05, |
|
"loss": 0.4678, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.07017543859649122, |
|
"eval_loss": 0.5602560639381409, |
|
"eval_runtime": 129.0871, |
|
"eval_samples_per_second": 3.486, |
|
"eval_steps_per_second": 0.442, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.07602339181286549, |
|
"grad_norm": 16.6875, |
|
"learning_rate": 2.3492462311557788e-05, |
|
"loss": 0.5118, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.08187134502923976, |
|
"grad_norm": 17.921875, |
|
"learning_rate": 2.3366834170854275e-05, |
|
"loss": 0.5199, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.08771929824561403, |
|
"grad_norm": 16.5625, |
|
"learning_rate": 2.3241206030150754e-05, |
|
"loss": 0.4611, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.0935672514619883, |
|
"grad_norm": 10.5, |
|
"learning_rate": 2.3115577889447238e-05, |
|
"loss": 0.428, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.0935672514619883, |
|
"eval_loss": 0.524933397769928, |
|
"eval_runtime": 127.1828, |
|
"eval_samples_per_second": 3.538, |
|
"eval_steps_per_second": 0.448, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.09941520467836257, |
|
"grad_norm": 11.6328125, |
|
"learning_rate": 2.298994974874372e-05, |
|
"loss": 0.4953, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.10526315789473684, |
|
"grad_norm": 5.4296875, |
|
"learning_rate": 2.28643216080402e-05, |
|
"loss": 0.4173, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.1111111111111111, |
|
"grad_norm": 8.8125, |
|
"learning_rate": 2.2738693467336687e-05, |
|
"loss": 0.4225, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.11695906432748537, |
|
"grad_norm": 8.84375, |
|
"learning_rate": 2.2613065326633167e-05, |
|
"loss": 0.3798, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.11695906432748537, |
|
"eval_loss": 0.49835124611854553, |
|
"eval_runtime": 128.3365, |
|
"eval_samples_per_second": 3.506, |
|
"eval_steps_per_second": 0.444, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.12280701754385964, |
|
"grad_norm": 11.4609375, |
|
"learning_rate": 2.248743718592965e-05, |
|
"loss": 0.4386, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.1286549707602339, |
|
"grad_norm": 16.4375, |
|
"learning_rate": 2.2361809045226133e-05, |
|
"loss": 0.4537, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.13450292397660818, |
|
"grad_norm": 10.140625, |
|
"learning_rate": 2.2236180904522613e-05, |
|
"loss": 0.3514, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.14035087719298245, |
|
"grad_norm": 3.265625, |
|
"learning_rate": 2.21105527638191e-05, |
|
"loss": 0.3665, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.14035087719298245, |
|
"eval_loss": 0.47331658005714417, |
|
"eval_runtime": 128.8126, |
|
"eval_samples_per_second": 3.493, |
|
"eval_steps_per_second": 0.443, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.14619883040935672, |
|
"grad_norm": 12.3046875, |
|
"learning_rate": 2.198492462311558e-05, |
|
"loss": 0.2409, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.15204678362573099, |
|
"grad_norm": 9.703125, |
|
"learning_rate": 2.1859296482412062e-05, |
|
"loss": 0.4553, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.15789473684210525, |
|
"grad_norm": 10.65625, |
|
"learning_rate": 2.1733668341708545e-05, |
|
"loss": 0.4242, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.16374269005847952, |
|
"grad_norm": 4.96484375, |
|
"learning_rate": 2.1608040201005025e-05, |
|
"loss": 0.4406, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.16374269005847952, |
|
"eval_loss": 0.45101454854011536, |
|
"eval_runtime": 129.1251, |
|
"eval_samples_per_second": 3.485, |
|
"eval_steps_per_second": 0.441, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.1695906432748538, |
|
"grad_norm": 6.20703125, |
|
"learning_rate": 2.1482412060301508e-05, |
|
"loss": 0.408, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.17543859649122806, |
|
"grad_norm": 9.875, |
|
"learning_rate": 2.135678391959799e-05, |
|
"loss": 0.5067, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.18128654970760233, |
|
"grad_norm": 7.1875, |
|
"learning_rate": 2.1231155778894474e-05, |
|
"loss": 0.4338, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.1871345029239766, |
|
"grad_norm": 12.765625, |
|
"learning_rate": 2.1105527638190957e-05, |
|
"loss": 0.4723, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.1871345029239766, |
|
"eval_loss": 0.4244983494281769, |
|
"eval_runtime": 127.3581, |
|
"eval_samples_per_second": 3.533, |
|
"eval_steps_per_second": 0.448, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.19298245614035087, |
|
"grad_norm": 26.03125, |
|
"learning_rate": 2.0979899497487437e-05, |
|
"loss": 0.4424, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.19883040935672514, |
|
"grad_norm": 5.234375, |
|
"learning_rate": 2.085427135678392e-05, |
|
"loss": 0.3599, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.2046783625730994, |
|
"grad_norm": 4.1875, |
|
"learning_rate": 2.0728643216080403e-05, |
|
"loss": 0.3399, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.21052631578947367, |
|
"grad_norm": 9.6640625, |
|
"learning_rate": 2.0603015075376886e-05, |
|
"loss": 0.4807, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.21052631578947367, |
|
"eval_loss": 0.41580215096473694, |
|
"eval_runtime": 128.5483, |
|
"eval_samples_per_second": 3.501, |
|
"eval_steps_per_second": 0.443, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.21637426900584794, |
|
"grad_norm": 9.03125, |
|
"learning_rate": 2.047738693467337e-05, |
|
"loss": 0.3341, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.2222222222222222, |
|
"grad_norm": 10.3515625, |
|
"learning_rate": 2.035175879396985e-05, |
|
"loss": 0.4407, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.22807017543859648, |
|
"grad_norm": 13.046875, |
|
"learning_rate": 2.0226130653266332e-05, |
|
"loss": 0.391, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.23391812865497075, |
|
"grad_norm": 13.0859375, |
|
"learning_rate": 2.0100502512562815e-05, |
|
"loss": 0.4196, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.23391812865497075, |
|
"eval_loss": 0.39709773659706116, |
|
"eval_runtime": 129.6284, |
|
"eval_samples_per_second": 3.471, |
|
"eval_steps_per_second": 0.44, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.23976608187134502, |
|
"grad_norm": 2.962890625, |
|
"learning_rate": 1.9974874371859298e-05, |
|
"loss": 0.3042, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.24561403508771928, |
|
"grad_norm": 11.9765625, |
|
"learning_rate": 1.984924623115578e-05, |
|
"loss": 0.3731, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.25146198830409355, |
|
"grad_norm": 5.42578125, |
|
"learning_rate": 1.972361809045226e-05, |
|
"loss": 0.3032, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.2573099415204678, |
|
"grad_norm": 8.9140625, |
|
"learning_rate": 1.9597989949748744e-05, |
|
"loss": 0.3443, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.2573099415204678, |
|
"eval_loss": 0.3737768530845642, |
|
"eval_runtime": 126.3694, |
|
"eval_samples_per_second": 3.561, |
|
"eval_steps_per_second": 0.451, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.2631578947368421, |
|
"grad_norm": 14.0625, |
|
"learning_rate": 1.9472361809045227e-05, |
|
"loss": 0.2804, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.26900584795321636, |
|
"grad_norm": 5.19921875, |
|
"learning_rate": 1.934673366834171e-05, |
|
"loss": 0.3257, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.27485380116959063, |
|
"grad_norm": 9.78125, |
|
"learning_rate": 1.9221105527638193e-05, |
|
"loss": 0.3625, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.2807017543859649, |
|
"grad_norm": 12.890625, |
|
"learning_rate": 1.9095477386934673e-05, |
|
"loss": 0.4133, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.2807017543859649, |
|
"eval_loss": 0.3631095290184021, |
|
"eval_runtime": 125.638, |
|
"eval_samples_per_second": 3.582, |
|
"eval_steps_per_second": 0.454, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.28654970760233917, |
|
"grad_norm": 7.71875, |
|
"learning_rate": 1.8969849246231156e-05, |
|
"loss": 0.5163, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.29239766081871343, |
|
"grad_norm": 7.25390625, |
|
"learning_rate": 1.884422110552764e-05, |
|
"loss": 0.2991, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.2982456140350877, |
|
"grad_norm": 6.5234375, |
|
"learning_rate": 1.8718592964824123e-05, |
|
"loss": 0.431, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.30409356725146197, |
|
"grad_norm": 3.8515625, |
|
"learning_rate": 1.8592964824120602e-05, |
|
"loss": 0.2838, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.30409356725146197, |
|
"eval_loss": 0.33338162302970886, |
|
"eval_runtime": 127.3115, |
|
"eval_samples_per_second": 3.535, |
|
"eval_steps_per_second": 0.448, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.30994152046783624, |
|
"grad_norm": 9.828125, |
|
"learning_rate": 1.8467336683417085e-05, |
|
"loss": 0.2653, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.3157894736842105, |
|
"grad_norm": 5.34765625, |
|
"learning_rate": 1.834170854271357e-05, |
|
"loss": 0.2515, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.3216374269005848, |
|
"grad_norm": 12.15625, |
|
"learning_rate": 1.821608040201005e-05, |
|
"loss": 0.2983, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.32748538011695905, |
|
"grad_norm": 9.171875, |
|
"learning_rate": 1.8090452261306535e-05, |
|
"loss": 0.4134, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.32748538011695905, |
|
"eval_loss": 0.32642024755477905, |
|
"eval_runtime": 127.196, |
|
"eval_samples_per_second": 3.538, |
|
"eval_steps_per_second": 0.448, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.3333333333333333, |
|
"grad_norm": 7.16015625, |
|
"learning_rate": 1.7964824120603014e-05, |
|
"loss": 0.3015, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.3391812865497076, |
|
"grad_norm": 15.6640625, |
|
"learning_rate": 1.7839195979899497e-05, |
|
"loss": 0.3098, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.34502923976608185, |
|
"grad_norm": 13.8984375, |
|
"learning_rate": 1.771356783919598e-05, |
|
"loss": 0.4022, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.3508771929824561, |
|
"grad_norm": 9.90625, |
|
"learning_rate": 1.7587939698492464e-05, |
|
"loss": 0.2838, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.3508771929824561, |
|
"eval_loss": 0.3124904930591583, |
|
"eval_runtime": 128.6196, |
|
"eval_samples_per_second": 3.499, |
|
"eval_steps_per_second": 0.443, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.3567251461988304, |
|
"grad_norm": 18.234375, |
|
"learning_rate": 1.7462311557788947e-05, |
|
"loss": 0.3792, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.36257309941520466, |
|
"grad_norm": 1.482421875, |
|
"learning_rate": 1.7336683417085427e-05, |
|
"loss": 0.4708, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.3684210526315789, |
|
"grad_norm": 10.8046875, |
|
"learning_rate": 1.721105527638191e-05, |
|
"loss": 0.3695, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.3742690058479532, |
|
"grad_norm": 13.6328125, |
|
"learning_rate": 1.7085427135678393e-05, |
|
"loss": 0.275, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.3742690058479532, |
|
"eval_loss": 0.294376403093338, |
|
"eval_runtime": 125.5965, |
|
"eval_samples_per_second": 3.583, |
|
"eval_steps_per_second": 0.454, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.38011695906432746, |
|
"grad_norm": 10.3359375, |
|
"learning_rate": 1.6959798994974876e-05, |
|
"loss": 0.3134, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.38596491228070173, |
|
"grad_norm": 4.671875, |
|
"learning_rate": 1.683417085427136e-05, |
|
"loss": 0.1961, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.391812865497076, |
|
"grad_norm": 3.734375, |
|
"learning_rate": 1.670854271356784e-05, |
|
"loss": 0.2686, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.39766081871345027, |
|
"grad_norm": 11.5859375, |
|
"learning_rate": 1.6582914572864322e-05, |
|
"loss": 0.4141, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.39766081871345027, |
|
"eval_loss": 0.2839096784591675, |
|
"eval_runtime": 126.4222, |
|
"eval_samples_per_second": 3.56, |
|
"eval_steps_per_second": 0.451, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.40350877192982454, |
|
"grad_norm": 12.0234375, |
|
"learning_rate": 1.6457286432160805e-05, |
|
"loss": 0.2948, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.4093567251461988, |
|
"grad_norm": 6.44140625, |
|
"learning_rate": 1.6331658291457288e-05, |
|
"loss": 0.3546, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.4152046783625731, |
|
"grad_norm": 7.5234375, |
|
"learning_rate": 1.620603015075377e-05, |
|
"loss": 0.3239, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.42105263157894735, |
|
"grad_norm": 4.30078125, |
|
"learning_rate": 1.608040201005025e-05, |
|
"loss": 0.2498, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.42105263157894735, |
|
"eval_loss": 0.2749183773994446, |
|
"eval_runtime": 126.776, |
|
"eval_samples_per_second": 3.55, |
|
"eval_steps_per_second": 0.45, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.4269005847953216, |
|
"grad_norm": 8.4375, |
|
"learning_rate": 1.5954773869346734e-05, |
|
"loss": 0.3026, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.4327485380116959, |
|
"grad_norm": 8.140625, |
|
"learning_rate": 1.5829145728643217e-05, |
|
"loss": 0.3848, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.43859649122807015, |
|
"grad_norm": 10.046875, |
|
"learning_rate": 1.57035175879397e-05, |
|
"loss": 0.2574, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 0.4444444444444444, |
|
"grad_norm": 9.1640625, |
|
"learning_rate": 1.5577889447236183e-05, |
|
"loss": 0.2817, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.4444444444444444, |
|
"eval_loss": 0.2605937719345093, |
|
"eval_runtime": 129.5889, |
|
"eval_samples_per_second": 3.473, |
|
"eval_steps_per_second": 0.44, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.4502923976608187, |
|
"grad_norm": 11.3828125, |
|
"learning_rate": 1.5452261306532663e-05, |
|
"loss": 0.376, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 0.45614035087719296, |
|
"grad_norm": 9.6953125, |
|
"learning_rate": 1.5326633165829146e-05, |
|
"loss": 0.2708, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.4619883040935672, |
|
"grad_norm": 10.0859375, |
|
"learning_rate": 1.5201005025125627e-05, |
|
"loss": 0.2509, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 0.4678362573099415, |
|
"grad_norm": 11.4921875, |
|
"learning_rate": 1.507537688442211e-05, |
|
"loss": 0.2899, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.4678362573099415, |
|
"eval_loss": 0.2526080906391144, |
|
"eval_runtime": 127.4275, |
|
"eval_samples_per_second": 3.531, |
|
"eval_steps_per_second": 0.447, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.47368421052631576, |
|
"grad_norm": 10.1796875, |
|
"learning_rate": 1.4949748743718595e-05, |
|
"loss": 0.3491, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 0.47953216374269003, |
|
"grad_norm": 4.72265625, |
|
"learning_rate": 1.4824120603015077e-05, |
|
"loss": 0.1925, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.4853801169590643, |
|
"grad_norm": 6.58203125, |
|
"learning_rate": 1.4698492462311558e-05, |
|
"loss": 0.2187, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 0.49122807017543857, |
|
"grad_norm": 11.4609375, |
|
"learning_rate": 1.457286432160804e-05, |
|
"loss": 0.2695, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.49122807017543857, |
|
"eval_loss": 0.2520677149295807, |
|
"eval_runtime": 129.417, |
|
"eval_samples_per_second": 3.477, |
|
"eval_steps_per_second": 0.44, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.49707602339181284, |
|
"grad_norm": 4.5, |
|
"learning_rate": 1.4447236180904523e-05, |
|
"loss": 0.1898, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 0.5029239766081871, |
|
"grad_norm": 7.640625, |
|
"learning_rate": 1.4321608040201007e-05, |
|
"loss": 0.3, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.5087719298245614, |
|
"grad_norm": 3.82421875, |
|
"learning_rate": 1.4195979899497489e-05, |
|
"loss": 0.2006, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 0.5146198830409356, |
|
"grad_norm": 3.509765625, |
|
"learning_rate": 1.407035175879397e-05, |
|
"loss": 0.2619, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.5146198830409356, |
|
"eval_loss": 0.24239127337932587, |
|
"eval_runtime": 125.4976, |
|
"eval_samples_per_second": 3.586, |
|
"eval_steps_per_second": 0.454, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.52046783625731, |
|
"grad_norm": 7.35546875, |
|
"learning_rate": 1.3944723618090452e-05, |
|
"loss": 0.3202, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 0.5263157894736842, |
|
"grad_norm": 5.37109375, |
|
"learning_rate": 1.3819095477386935e-05, |
|
"loss": 0.143, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.5321637426900585, |
|
"grad_norm": 13.203125, |
|
"learning_rate": 1.369346733668342e-05, |
|
"loss": 0.3726, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 0.5380116959064327, |
|
"grad_norm": 7.08984375, |
|
"learning_rate": 1.3567839195979901e-05, |
|
"loss": 0.2238, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.5380116959064327, |
|
"eval_loss": 0.2372661828994751, |
|
"eval_runtime": 125.8271, |
|
"eval_samples_per_second": 3.576, |
|
"eval_steps_per_second": 0.453, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.543859649122807, |
|
"grad_norm": 7.73046875, |
|
"learning_rate": 1.3442211055276382e-05, |
|
"loss": 0.2301, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 0.5497076023391813, |
|
"grad_norm": 1.9326171875, |
|
"learning_rate": 1.3316582914572864e-05, |
|
"loss": 0.2891, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.5555555555555556, |
|
"grad_norm": 5.25, |
|
"learning_rate": 1.3190954773869347e-05, |
|
"loss": 0.1861, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 0.5614035087719298, |
|
"grad_norm": 15.7265625, |
|
"learning_rate": 1.306532663316583e-05, |
|
"loss": 0.3049, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.5614035087719298, |
|
"eval_loss": 0.23007704317569733, |
|
"eval_runtime": 125.6345, |
|
"eval_samples_per_second": 3.582, |
|
"eval_steps_per_second": 0.454, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.5672514619883041, |
|
"grad_norm": 12.0078125, |
|
"learning_rate": 1.2939698492462313e-05, |
|
"loss": 0.2584, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 0.5730994152046783, |
|
"grad_norm": 11.8125, |
|
"learning_rate": 1.2814070351758795e-05, |
|
"loss": 0.29, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.5789473684210527, |
|
"grad_norm": 4.7578125, |
|
"learning_rate": 1.2688442211055276e-05, |
|
"loss": 0.2648, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 0.5847953216374269, |
|
"grad_norm": 10.53125, |
|
"learning_rate": 1.2562814070351759e-05, |
|
"loss": 0.1308, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.5847953216374269, |
|
"eval_loss": 0.2292058765888214, |
|
"eval_runtime": 127.785, |
|
"eval_samples_per_second": 3.522, |
|
"eval_steps_per_second": 0.446, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.5906432748538012, |
|
"grad_norm": 8.421875, |
|
"learning_rate": 1.2437185929648242e-05, |
|
"loss": 0.3249, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 0.5964912280701754, |
|
"grad_norm": 10.8828125, |
|
"learning_rate": 1.2311557788944725e-05, |
|
"loss": 0.2164, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.6023391812865497, |
|
"grad_norm": 7.421875, |
|
"learning_rate": 1.2185929648241207e-05, |
|
"loss": 0.2512, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 0.6081871345029239, |
|
"grad_norm": 9.6171875, |
|
"learning_rate": 1.2060301507537688e-05, |
|
"loss": 0.1936, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.6081871345029239, |
|
"eval_loss": 0.21902820467948914, |
|
"eval_runtime": 125.3191, |
|
"eval_samples_per_second": 3.591, |
|
"eval_steps_per_second": 0.455, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.6140350877192983, |
|
"grad_norm": 1.091796875, |
|
"learning_rate": 1.1934673366834171e-05, |
|
"loss": 0.189, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 0.6198830409356725, |
|
"grad_norm": 7.83203125, |
|
"learning_rate": 1.1809045226130654e-05, |
|
"loss": 0.2179, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.6257309941520468, |
|
"grad_norm": 8.2109375, |
|
"learning_rate": 1.1683417085427137e-05, |
|
"loss": 0.224, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 0.631578947368421, |
|
"grad_norm": 5.1640625, |
|
"learning_rate": 1.1557788944723619e-05, |
|
"loss": 0.2479, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.631578947368421, |
|
"eval_loss": 0.21907079219818115, |
|
"eval_runtime": 126.3038, |
|
"eval_samples_per_second": 3.563, |
|
"eval_steps_per_second": 0.451, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.6374269005847953, |
|
"grad_norm": 4.47265625, |
|
"learning_rate": 1.14321608040201e-05, |
|
"loss": 0.219, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 0.6432748538011696, |
|
"grad_norm": 3.30078125, |
|
"learning_rate": 1.1306532663316583e-05, |
|
"loss": 0.2741, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.6491228070175439, |
|
"grad_norm": 4.21484375, |
|
"learning_rate": 1.1180904522613066e-05, |
|
"loss": 0.1153, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 0.6549707602339181, |
|
"grad_norm": 5.125, |
|
"learning_rate": 1.105527638190955e-05, |
|
"loss": 0.1575, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.6549707602339181, |
|
"eval_loss": 0.2165260761976242, |
|
"eval_runtime": 125.8096, |
|
"eval_samples_per_second": 3.577, |
|
"eval_steps_per_second": 0.453, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.6608187134502924, |
|
"grad_norm": 0.83837890625, |
|
"learning_rate": 1.0929648241206031e-05, |
|
"loss": 0.2238, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 0.6666666666666666, |
|
"grad_norm": 6.61328125, |
|
"learning_rate": 1.0804020100502512e-05, |
|
"loss": 0.3554, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.672514619883041, |
|
"grad_norm": 12.875, |
|
"learning_rate": 1.0678391959798995e-05, |
|
"loss": 0.096, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 0.6783625730994152, |
|
"grad_norm": 3.28125, |
|
"learning_rate": 1.0552763819095479e-05, |
|
"loss": 0.193, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.6783625730994152, |
|
"eval_loss": 0.21065188944339752, |
|
"eval_runtime": 125.6087, |
|
"eval_samples_per_second": 3.583, |
|
"eval_steps_per_second": 0.454, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.6842105263157895, |
|
"grad_norm": 8.7265625, |
|
"learning_rate": 1.042713567839196e-05, |
|
"loss": 0.1936, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 0.6900584795321637, |
|
"grad_norm": 4.0859375, |
|
"learning_rate": 1.0301507537688443e-05, |
|
"loss": 0.2103, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.695906432748538, |
|
"grad_norm": 15.5703125, |
|
"learning_rate": 1.0175879396984924e-05, |
|
"loss": 0.2701, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 0.7017543859649122, |
|
"grad_norm": 3.73046875, |
|
"learning_rate": 1.0050251256281408e-05, |
|
"loss": 0.2526, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.7017543859649122, |
|
"eval_loss": 0.21144379675388336, |
|
"eval_runtime": 128.1153, |
|
"eval_samples_per_second": 3.512, |
|
"eval_steps_per_second": 0.445, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.7076023391812866, |
|
"grad_norm": 1.537109375, |
|
"learning_rate": 9.92462311557789e-06, |
|
"loss": 0.2087, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 0.7134502923976608, |
|
"grad_norm": 0.55908203125, |
|
"learning_rate": 9.798994974874372e-06, |
|
"loss": 0.1532, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.7192982456140351, |
|
"grad_norm": 9.7578125, |
|
"learning_rate": 9.673366834170855e-06, |
|
"loss": 0.1985, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 0.7251461988304093, |
|
"grad_norm": 9.390625, |
|
"learning_rate": 9.547738693467337e-06, |
|
"loss": 0.1574, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.7251461988304093, |
|
"eval_loss": 0.20868618786334991, |
|
"eval_runtime": 129.4237, |
|
"eval_samples_per_second": 3.477, |
|
"eval_steps_per_second": 0.44, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.7309941520467836, |
|
"grad_norm": 4.6171875, |
|
"learning_rate": 9.42211055276382e-06, |
|
"loss": 0.2921, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 0.7368421052631579, |
|
"grad_norm": 10.859375, |
|
"learning_rate": 9.296482412060301e-06, |
|
"loss": 0.1932, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.7426900584795322, |
|
"grad_norm": 5.97265625, |
|
"learning_rate": 9.170854271356784e-06, |
|
"loss": 0.2999, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 0.7485380116959064, |
|
"grad_norm": 4.85546875, |
|
"learning_rate": 9.045226130653267e-06, |
|
"loss": 0.1989, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.7485380116959064, |
|
"eval_loss": 0.20511174201965332, |
|
"eval_runtime": 127.5209, |
|
"eval_samples_per_second": 3.529, |
|
"eval_steps_per_second": 0.447, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.7543859649122807, |
|
"grad_norm": 8.7265625, |
|
"learning_rate": 8.919597989949749e-06, |
|
"loss": 0.1795, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 0.7602339181286549, |
|
"grad_norm": 3.490234375, |
|
"learning_rate": 8.793969849246232e-06, |
|
"loss": 0.2444, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.7660818713450293, |
|
"grad_norm": 6.671875, |
|
"learning_rate": 8.668341708542713e-06, |
|
"loss": 0.2775, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 0.7719298245614035, |
|
"grad_norm": 5.1796875, |
|
"learning_rate": 8.542713567839196e-06, |
|
"loss": 0.1761, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.7719298245614035, |
|
"eval_loss": 0.20133711397647858, |
|
"eval_runtime": 129.2927, |
|
"eval_samples_per_second": 3.48, |
|
"eval_steps_per_second": 0.441, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.7777777777777778, |
|
"grad_norm": 0.83935546875, |
|
"learning_rate": 8.41708542713568e-06, |
|
"loss": 0.216, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 0.783625730994152, |
|
"grad_norm": 2.580078125, |
|
"learning_rate": 8.291457286432161e-06, |
|
"loss": 0.0873, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.7894736842105263, |
|
"grad_norm": 5.87109375, |
|
"learning_rate": 8.165829145728644e-06, |
|
"loss": 0.1145, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 0.7953216374269005, |
|
"grad_norm": 7.125, |
|
"learning_rate": 8.040201005025125e-06, |
|
"loss": 0.2223, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.7953216374269005, |
|
"eval_loss": 0.19959186017513275, |
|
"eval_runtime": 125.4382, |
|
"eval_samples_per_second": 3.587, |
|
"eval_steps_per_second": 0.454, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.8011695906432749, |
|
"grad_norm": 3.94921875, |
|
"learning_rate": 7.914572864321608e-06, |
|
"loss": 0.1845, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 0.8070175438596491, |
|
"grad_norm": 5.99609375, |
|
"learning_rate": 7.788944723618092e-06, |
|
"loss": 0.104, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.8128654970760234, |
|
"grad_norm": 7.953125, |
|
"learning_rate": 7.663316582914573e-06, |
|
"loss": 0.1119, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 0.8187134502923976, |
|
"grad_norm": 7.5859375, |
|
"learning_rate": 7.537688442211055e-06, |
|
"loss": 0.2127, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.8187134502923976, |
|
"eval_loss": 0.1966124027967453, |
|
"eval_runtime": 126.749, |
|
"eval_samples_per_second": 3.55, |
|
"eval_steps_per_second": 0.45, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.8245614035087719, |
|
"grad_norm": 25.390625, |
|
"learning_rate": 7.412060301507538e-06, |
|
"loss": 0.1597, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 0.8304093567251462, |
|
"grad_norm": 7.55078125, |
|
"learning_rate": 7.28643216080402e-06, |
|
"loss": 0.0942, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.8362573099415205, |
|
"grad_norm": 6.5390625, |
|
"learning_rate": 7.160804020100504e-06, |
|
"loss": 0.236, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 0.8421052631578947, |
|
"grad_norm": 9.390625, |
|
"learning_rate": 7.035175879396985e-06, |
|
"loss": 0.2477, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.8421052631578947, |
|
"eval_loss": 0.1922728419303894, |
|
"eval_runtime": 125.5954, |
|
"eval_samples_per_second": 3.583, |
|
"eval_steps_per_second": 0.454, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.847953216374269, |
|
"grad_norm": 0.7783203125, |
|
"learning_rate": 6.909547738693467e-06, |
|
"loss": 0.3097, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 0.8538011695906432, |
|
"grad_norm": 5.91796875, |
|
"learning_rate": 6.7839195979899505e-06, |
|
"loss": 0.2097, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.8596491228070176, |
|
"grad_norm": 2.55078125, |
|
"learning_rate": 6.658291457286432e-06, |
|
"loss": 0.1837, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 0.8654970760233918, |
|
"grad_norm": 5.1015625, |
|
"learning_rate": 6.532663316582915e-06, |
|
"loss": 0.1931, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.8654970760233918, |
|
"eval_loss": 0.1908126324415207, |
|
"eval_runtime": 125.264, |
|
"eval_samples_per_second": 3.592, |
|
"eval_steps_per_second": 0.455, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.8713450292397661, |
|
"grad_norm": 1.0830078125, |
|
"learning_rate": 6.407035175879397e-06, |
|
"loss": 0.1688, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 0.8771929824561403, |
|
"grad_norm": 2.021484375, |
|
"learning_rate": 6.2814070351758795e-06, |
|
"loss": 0.1635, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.8830409356725146, |
|
"grad_norm": 24.90625, |
|
"learning_rate": 6.155778894472363e-06, |
|
"loss": 0.2588, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 0.8888888888888888, |
|
"grad_norm": 4.9453125, |
|
"learning_rate": 6.030150753768844e-06, |
|
"loss": 0.182, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.8888888888888888, |
|
"eval_loss": 0.18878790736198425, |
|
"eval_runtime": 126.7176, |
|
"eval_samples_per_second": 3.551, |
|
"eval_steps_per_second": 0.45, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.8947368421052632, |
|
"grad_norm": 7.25, |
|
"learning_rate": 5.904522613065327e-06, |
|
"loss": 0.2793, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 0.9005847953216374, |
|
"grad_norm": 9.59375, |
|
"learning_rate": 5.778894472361809e-06, |
|
"loss": 0.1608, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.9064327485380117, |
|
"grad_norm": 2.943359375, |
|
"learning_rate": 5.653266331658292e-06, |
|
"loss": 0.2136, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 0.9122807017543859, |
|
"grad_norm": 0.48681640625, |
|
"learning_rate": 5.527638190954775e-06, |
|
"loss": 0.1693, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.9122807017543859, |
|
"eval_loss": 0.18779730796813965, |
|
"eval_runtime": 126.6799, |
|
"eval_samples_per_second": 3.552, |
|
"eval_steps_per_second": 0.45, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.9181286549707602, |
|
"grad_norm": 3.2109375, |
|
"learning_rate": 5.402010050251256e-06, |
|
"loss": 0.0918, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 0.9239766081871345, |
|
"grad_norm": 1.541015625, |
|
"learning_rate": 5.276381909547739e-06, |
|
"loss": 0.2076, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.9298245614035088, |
|
"grad_norm": 4.7421875, |
|
"learning_rate": 5.1507537688442215e-06, |
|
"loss": 0.2429, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 0.935672514619883, |
|
"grad_norm": 4.375, |
|
"learning_rate": 5.025125628140704e-06, |
|
"loss": 0.1346, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.935672514619883, |
|
"eval_loss": 0.18527910113334656, |
|
"eval_runtime": 129.3266, |
|
"eval_samples_per_second": 3.48, |
|
"eval_steps_per_second": 0.441, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.9415204678362573, |
|
"grad_norm": 4.875, |
|
"learning_rate": 4.899497487437186e-06, |
|
"loss": 0.2457, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 0.9473684210526315, |
|
"grad_norm": 3.013671875, |
|
"learning_rate": 4.773869346733668e-06, |
|
"loss": 0.187, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.9532163742690059, |
|
"grad_norm": 4.3125, |
|
"learning_rate": 4.6482412060301506e-06, |
|
"loss": 0.1546, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 0.9590643274853801, |
|
"grad_norm": 2.5, |
|
"learning_rate": 4.522613065326634e-06, |
|
"loss": 0.1484, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.9590643274853801, |
|
"eval_loss": 0.18491099774837494, |
|
"eval_runtime": 126.5822, |
|
"eval_samples_per_second": 3.555, |
|
"eval_steps_per_second": 0.45, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.9649122807017544, |
|
"grad_norm": 5.03515625, |
|
"learning_rate": 4.396984924623116e-06, |
|
"loss": 0.1739, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 0.9707602339181286, |
|
"grad_norm": 7.1015625, |
|
"learning_rate": 4.271356783919598e-06, |
|
"loss": 0.3268, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.9766081871345029, |
|
"grad_norm": 0.397216796875, |
|
"learning_rate": 4.1457286432160804e-06, |
|
"loss": 0.2178, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 0.9824561403508771, |
|
"grad_norm": 5.00390625, |
|
"learning_rate": 4.020100502512563e-06, |
|
"loss": 0.1217, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.9824561403508771, |
|
"eval_loss": 0.1838068664073944, |
|
"eval_runtime": 127.461, |
|
"eval_samples_per_second": 3.53, |
|
"eval_steps_per_second": 0.447, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.9883040935672515, |
|
"grad_norm": 0.1112060546875, |
|
"learning_rate": 3.894472361809046e-06, |
|
"loss": 0.0894, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 0.9941520467836257, |
|
"grad_norm": 1.748046875, |
|
"learning_rate": 3.7688442211055276e-06, |
|
"loss": 0.1733, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 6.53125, |
|
"learning_rate": 3.64321608040201e-06, |
|
"loss": 0.1281, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 1.0058479532163742, |
|
"grad_norm": 0.9111328125, |
|
"learning_rate": 3.5175879396984926e-06, |
|
"loss": 0.0669, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.0058479532163742, |
|
"eval_loss": 0.18437370657920837, |
|
"eval_runtime": 128.6228, |
|
"eval_samples_per_second": 3.499, |
|
"eval_steps_per_second": 0.443, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.0116959064327484, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 3.3919597989949752e-06, |
|
"loss": 0.0748, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 1.0175438596491229, |
|
"grad_norm": 7.703125, |
|
"learning_rate": 3.2663316582914575e-06, |
|
"loss": 0.071, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 1.023391812865497, |
|
"grad_norm": 8.9765625, |
|
"learning_rate": 3.1407035175879398e-06, |
|
"loss": 0.1588, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 1.0292397660818713, |
|
"grad_norm": 0.38427734375, |
|
"learning_rate": 3.015075376884422e-06, |
|
"loss": 0.1292, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.0292397660818713, |
|
"eval_loss": 0.18767422437667847, |
|
"eval_runtime": 126.9252, |
|
"eval_samples_per_second": 3.545, |
|
"eval_steps_per_second": 0.449, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.0350877192982457, |
|
"grad_norm": 6.4375, |
|
"learning_rate": 2.8894472361809047e-06, |
|
"loss": 0.1419, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 1.04093567251462, |
|
"grad_norm": 13.3515625, |
|
"learning_rate": 2.7638190954773874e-06, |
|
"loss": 0.084, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 1.0467836257309941, |
|
"grad_norm": 9.203125, |
|
"learning_rate": 2.6381909547738696e-06, |
|
"loss": 0.072, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 1.0526315789473684, |
|
"grad_norm": 2.439453125, |
|
"learning_rate": 2.512562814070352e-06, |
|
"loss": 0.1106, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.0526315789473684, |
|
"eval_loss": 0.18756501376628876, |
|
"eval_runtime": 128.4204, |
|
"eval_samples_per_second": 3.504, |
|
"eval_steps_per_second": 0.444, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.0584795321637426, |
|
"grad_norm": 4.671875, |
|
"learning_rate": 2.386934673366834e-06, |
|
"loss": 0.1226, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 1.064327485380117, |
|
"grad_norm": 2.62890625, |
|
"learning_rate": 2.261306532663317e-06, |
|
"loss": 0.123, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 1.0701754385964912, |
|
"grad_norm": 16.640625, |
|
"learning_rate": 2.135678391959799e-06, |
|
"loss": 0.0593, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 1.0760233918128654, |
|
"grad_norm": 6.8828125, |
|
"learning_rate": 2.0100502512562813e-06, |
|
"loss": 0.0828, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.0760233918128654, |
|
"eval_loss": 0.1875353455543518, |
|
"eval_runtime": 126.5483, |
|
"eval_samples_per_second": 3.556, |
|
"eval_steps_per_second": 0.45, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.0818713450292399, |
|
"grad_norm": 4.55078125, |
|
"learning_rate": 1.8844221105527638e-06, |
|
"loss": 0.1481, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 1.087719298245614, |
|
"grad_norm": 3.6796875, |
|
"learning_rate": 1.7587939698492463e-06, |
|
"loss": 0.1275, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 1.0935672514619883, |
|
"grad_norm": 6.02734375, |
|
"learning_rate": 1.6331658291457288e-06, |
|
"loss": 0.1274, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 1.0994152046783625, |
|
"grad_norm": 1.4111328125, |
|
"learning_rate": 1.507537688442211e-06, |
|
"loss": 0.0485, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.0994152046783625, |
|
"eval_loss": 0.18709704279899597, |
|
"eval_runtime": 129.1721, |
|
"eval_samples_per_second": 3.484, |
|
"eval_steps_per_second": 0.441, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.1052631578947367, |
|
"grad_norm": 0.74853515625, |
|
"learning_rate": 1.3819095477386937e-06, |
|
"loss": 0.0483, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 1.1111111111111112, |
|
"grad_norm": 1.044921875, |
|
"learning_rate": 1.256281407035176e-06, |
|
"loss": 0.0799, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 1.1169590643274854, |
|
"grad_norm": 0.171142578125, |
|
"learning_rate": 1.1306532663316584e-06, |
|
"loss": 0.1273, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 1.1228070175438596, |
|
"grad_norm": 0.346923828125, |
|
"learning_rate": 1.0050251256281407e-06, |
|
"loss": 0.0624, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.1228070175438596, |
|
"eval_loss": 0.1874168962240219, |
|
"eval_runtime": 128.2151, |
|
"eval_samples_per_second": 3.51, |
|
"eval_steps_per_second": 0.445, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.128654970760234, |
|
"grad_norm": 0.1552734375, |
|
"learning_rate": 8.793969849246231e-07, |
|
"loss": 0.1017, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 1.1345029239766082, |
|
"grad_norm": 1.619140625, |
|
"learning_rate": 7.537688442211055e-07, |
|
"loss": 0.1556, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 1.1403508771929824, |
|
"grad_norm": 1.958984375, |
|
"learning_rate": 6.28140703517588e-07, |
|
"loss": 0.1113, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 1.1461988304093567, |
|
"grad_norm": 5.375, |
|
"learning_rate": 5.025125628140703e-07, |
|
"loss": 0.0895, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 1.1461988304093567, |
|
"eval_loss": 0.1870710253715515, |
|
"eval_runtime": 126.862, |
|
"eval_samples_per_second": 3.547, |
|
"eval_steps_per_second": 0.449, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 1.1520467836257309, |
|
"grad_norm": 5.984375, |
|
"learning_rate": 3.7688442211055275e-07, |
|
"loss": 0.1076, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 1.1578947368421053, |
|
"grad_norm": 1.9619140625, |
|
"learning_rate": 2.5125628140703517e-07, |
|
"loss": 0.1024, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 1.1637426900584795, |
|
"grad_norm": 6.17578125, |
|
"learning_rate": 1.2562814070351758e-07, |
|
"loss": 0.0698, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 1.1695906432748537, |
|
"grad_norm": 7.06640625, |
|
"learning_rate": 0.0, |
|
"loss": 0.1, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.1695906432748537, |
|
"eval_loss": 0.1871223896741867, |
|
"eval_runtime": 128.9773, |
|
"eval_samples_per_second": 3.489, |
|
"eval_steps_per_second": 0.442, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.1695906432748537, |
|
"step": 5000, |
|
"total_flos": 2.536648947447456e+17, |
|
"train_loss": 0.27304353165626527, |
|
"train_runtime": 21409.3382, |
|
"train_samples_per_second": 0.934, |
|
"train_steps_per_second": 0.234 |
|
} |
|
], |
|
"logging_steps": 25, |
|
"max_steps": 5000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.536648947447456e+17, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|