|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 50.0, |
|
"eval_steps": 500, |
|
"global_step": 12950, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 1.9305019305019306, |
|
"grad_norm": 2.738110065460205, |
|
"learning_rate": 9.613899613899614e-06, |
|
"loss": 0.0888, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.9305019305019306, |
|
"eval_runtime": 667.5755, |
|
"eval_samples_per_second": 264.418, |
|
"eval_steps_per_second": 16.527, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 3.861003861003861, |
|
"grad_norm": 4.450708389282227, |
|
"learning_rate": 9.227799227799229e-06, |
|
"loss": 0.0558, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 3.861003861003861, |
|
"eval_runtime": 651.8038, |
|
"eval_samples_per_second": 270.816, |
|
"eval_steps_per_second": 16.927, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 5.7915057915057915, |
|
"grad_norm": 5.049165725708008, |
|
"learning_rate": 8.841698841698842e-06, |
|
"loss": 0.0409, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 5.7915057915057915, |
|
"eval_runtime": 650.4821, |
|
"eval_samples_per_second": 271.366, |
|
"eval_steps_per_second": 16.961, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 7.722007722007722, |
|
"grad_norm": 3.824125051498413, |
|
"learning_rate": 8.455598455598457e-06, |
|
"loss": 0.0319, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 7.722007722007722, |
|
"eval_runtime": 652.7164, |
|
"eval_samples_per_second": 270.438, |
|
"eval_steps_per_second": 16.903, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 9.652509652509652, |
|
"grad_norm": 3.7455878257751465, |
|
"learning_rate": 8.06949806949807e-06, |
|
"loss": 0.0277, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 9.652509652509652, |
|
"eval_runtime": 652.469, |
|
"eval_samples_per_second": 270.54, |
|
"eval_steps_per_second": 16.91, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 11.583011583011583, |
|
"grad_norm": 2.825524091720581, |
|
"learning_rate": 7.683397683397685e-06, |
|
"loss": 0.0249, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 11.583011583011583, |
|
"eval_runtime": 724.8262, |
|
"eval_samples_per_second": 243.533, |
|
"eval_steps_per_second": 15.222, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 13.513513513513514, |
|
"grad_norm": 1.5855698585510254, |
|
"learning_rate": 7.297297297297298e-06, |
|
"loss": 0.0233, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 13.513513513513514, |
|
"eval_runtime": 721.7805, |
|
"eval_samples_per_second": 244.56, |
|
"eval_steps_per_second": 15.286, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 15.444015444015443, |
|
"grad_norm": 2.516690254211426, |
|
"learning_rate": 6.911196911196911e-06, |
|
"loss": 0.0217, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 15.444015444015443, |
|
"eval_runtime": 723.8804, |
|
"eval_samples_per_second": 243.851, |
|
"eval_steps_per_second": 15.241, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 17.374517374517374, |
|
"grad_norm": 2.6874454021453857, |
|
"learning_rate": 6.525096525096526e-06, |
|
"loss": 0.0207, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 17.374517374517374, |
|
"eval_runtime": 726.2467, |
|
"eval_samples_per_second": 243.057, |
|
"eval_steps_per_second": 15.192, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 19.305019305019304, |
|
"grad_norm": 2.5405662059783936, |
|
"learning_rate": 6.13899613899614e-06, |
|
"loss": 0.0198, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 19.305019305019304, |
|
"eval_runtime": 725.831, |
|
"eval_samples_per_second": 243.196, |
|
"eval_steps_per_second": 15.201, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 21.235521235521237, |
|
"grad_norm": 2.5149359703063965, |
|
"learning_rate": 5.752895752895753e-06, |
|
"loss": 0.0191, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 21.235521235521237, |
|
"eval_runtime": 724.4353, |
|
"eval_samples_per_second": 243.664, |
|
"eval_steps_per_second": 15.23, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 23.166023166023166, |
|
"grad_norm": 1.5971148014068604, |
|
"learning_rate": 5.366795366795368e-06, |
|
"loss": 0.0183, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 23.166023166023166, |
|
"eval_runtime": 725.1569, |
|
"eval_samples_per_second": 243.422, |
|
"eval_steps_per_second": 15.215, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 25.096525096525095, |
|
"grad_norm": 2.0284860134124756, |
|
"learning_rate": 4.980694980694981e-06, |
|
"loss": 0.0178, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 25.096525096525095, |
|
"eval_runtime": 726.8832, |
|
"eval_samples_per_second": 242.844, |
|
"eval_steps_per_second": 15.179, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 27.027027027027028, |
|
"grad_norm": 2.005959987640381, |
|
"learning_rate": 4.594594594594596e-06, |
|
"loss": 0.0174, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 27.027027027027028, |
|
"eval_runtime": 725.3449, |
|
"eval_samples_per_second": 243.359, |
|
"eval_steps_per_second": 15.211, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 28.957528957528957, |
|
"grad_norm": 1.96770441532135, |
|
"learning_rate": 4.208494208494209e-06, |
|
"loss": 0.0168, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 28.957528957528957, |
|
"eval_runtime": 765.8326, |
|
"eval_samples_per_second": 230.493, |
|
"eval_steps_per_second": 14.407, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 30.888030888030887, |
|
"grad_norm": 1.844897747039795, |
|
"learning_rate": 3.822393822393823e-06, |
|
"loss": 0.0165, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 30.888030888030887, |
|
"eval_runtime": 819.9764, |
|
"eval_samples_per_second": 215.273, |
|
"eval_steps_per_second": 13.455, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 32.818532818532816, |
|
"grad_norm": 1.615881323814392, |
|
"learning_rate": 3.4362934362934363e-06, |
|
"loss": 0.0162, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 32.818532818532816, |
|
"eval_runtime": 821.2267, |
|
"eval_samples_per_second": 214.946, |
|
"eval_steps_per_second": 13.435, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 34.74903474903475, |
|
"grad_norm": 1.6435213088989258, |
|
"learning_rate": 3.0501930501930503e-06, |
|
"loss": 0.0158, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 34.74903474903475, |
|
"eval_runtime": 819.0429, |
|
"eval_samples_per_second": 215.519, |
|
"eval_steps_per_second": 13.471, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 36.67953667953668, |
|
"grad_norm": 1.7934831380844116, |
|
"learning_rate": 2.6640926640926647e-06, |
|
"loss": 0.0155, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 36.67953667953668, |
|
"eval_runtime": 820.1169, |
|
"eval_samples_per_second": 215.236, |
|
"eval_steps_per_second": 13.453, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 38.61003861003861, |
|
"grad_norm": 1.1239484548568726, |
|
"learning_rate": 2.2779922779922782e-06, |
|
"loss": 0.0152, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 38.61003861003861, |
|
"eval_runtime": 821.2669, |
|
"eval_samples_per_second": 214.935, |
|
"eval_steps_per_second": 13.434, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 40.54054054054054, |
|
"grad_norm": 1.256516456604004, |
|
"learning_rate": 1.8918918918918922e-06, |
|
"loss": 0.015, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 40.54054054054054, |
|
"eval_runtime": 818.9208, |
|
"eval_samples_per_second": 215.551, |
|
"eval_steps_per_second": 13.473, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 42.47104247104247, |
|
"grad_norm": 0.4177967607975006, |
|
"learning_rate": 1.505791505791506e-06, |
|
"loss": 0.0148, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 42.47104247104247, |
|
"eval_runtime": 823.9199, |
|
"eval_samples_per_second": 214.243, |
|
"eval_steps_per_second": 13.391, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 44.4015444015444, |
|
"grad_norm": 0.3761753737926483, |
|
"learning_rate": 1.1196911196911197e-06, |
|
"loss": 0.0146, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 44.4015444015444, |
|
"eval_runtime": 821.1987, |
|
"eval_samples_per_second": 214.953, |
|
"eval_steps_per_second": 13.435, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 46.33204633204633, |
|
"grad_norm": 0.4973774552345276, |
|
"learning_rate": 7.335907335907337e-07, |
|
"loss": 0.0145, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 46.33204633204633, |
|
"eval_runtime": 821.7593, |
|
"eval_samples_per_second": 214.806, |
|
"eval_steps_per_second": 13.426, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 48.262548262548265, |
|
"grad_norm": 0.3773488998413086, |
|
"learning_rate": 3.474903474903475e-07, |
|
"loss": 0.0144, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 48.262548262548265, |
|
"eval_runtime": 824.3743, |
|
"eval_samples_per_second": 214.125, |
|
"eval_steps_per_second": 13.383, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"step": 12950, |
|
"total_flos": 5.544539360447693e+19, |
|
"train_loss": 0.023562463738283135, |
|
"train_runtime": 98807.7466, |
|
"train_samples_per_second": 267.972, |
|
"train_steps_per_second": 0.131 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 12950, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 50, |
|
"save_steps": 100, |
|
"total_flos": 5.544539360447693e+19, |
|
"train_batch_size": 2048, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|