|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.3, |
|
"eval_steps": 2000, |
|
"global_step": 6000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 12.253900527954102, |
|
"learning_rate": 9.8e-07, |
|
"loss": 0.3335, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 9.193947792053223, |
|
"learning_rate": 9.9010101010101e-07, |
|
"loss": 0.2105, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 5.168125152587891, |
|
"learning_rate": 9.8e-07, |
|
"loss": 0.1618, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 2.163803815841675, |
|
"learning_rate": 9.698989898989898e-07, |
|
"loss": 0.1418, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 10.751598358154297, |
|
"learning_rate": 9.597979797979797e-07, |
|
"loss": 0.1215, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 7.5342793464660645, |
|
"learning_rate": 9.496969696969696e-07, |
|
"loss": 0.1142, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 8.430835723876953, |
|
"learning_rate": 9.395959595959596e-07, |
|
"loss": 0.121, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 1.8440964221954346, |
|
"learning_rate": 9.295959595959596e-07, |
|
"loss": 0.0969, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 2.699073553085327, |
|
"learning_rate": 9.194949494949495e-07, |
|
"loss": 0.1024, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 11.251789093017578, |
|
"learning_rate": 9.093939393939394e-07, |
|
"loss": 0.0899, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 7.728929042816162, |
|
"learning_rate": 8.992929292929292e-07, |
|
"loss": 0.0837, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 1.7401416301727295, |
|
"learning_rate": 8.891919191919191e-07, |
|
"loss": 0.0914, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 3.5504443645477295, |
|
"learning_rate": 8.790909090909091e-07, |
|
"loss": 0.0754, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 5.9004316329956055, |
|
"learning_rate": 8.68989898989899e-07, |
|
"loss": 0.0791, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 2.9171862602233887, |
|
"learning_rate": 8.588888888888888e-07, |
|
"loss": 0.0866, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 5.907050132751465, |
|
"learning_rate": 8.487878787878787e-07, |
|
"loss": 0.0768, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 4.856807708740234, |
|
"learning_rate": 8.386868686868687e-07, |
|
"loss": 0.0745, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 2.4457929134368896, |
|
"learning_rate": 8.285858585858585e-07, |
|
"loss": 0.0808, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 3.084287643432617, |
|
"learning_rate": 8.184848484848484e-07, |
|
"loss": 0.0743, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 2.936805486679077, |
|
"learning_rate": 8.083838383838384e-07, |
|
"loss": 0.0675, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"eval_loss": 0.07925247400999069, |
|
"eval_runtime": 204.8299, |
|
"eval_samples_per_second": 4.882, |
|
"eval_steps_per_second": 1.221, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 13.029026985168457, |
|
"learning_rate": 7.982828282828282e-07, |
|
"loss": 0.0681, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 10.308150291442871, |
|
"learning_rate": 7.881818181818182e-07, |
|
"loss": 0.0798, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 5.64780330657959, |
|
"learning_rate": 7.78080808080808e-07, |
|
"loss": 0.0648, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 6.982935905456543, |
|
"learning_rate": 7.679797979797979e-07, |
|
"loss": 0.0731, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 3.5423080921173096, |
|
"learning_rate": 7.578787878787879e-07, |
|
"loss": 0.0817, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 3.585963249206543, |
|
"learning_rate": 7.477777777777778e-07, |
|
"loss": 0.072, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 7.086956977844238, |
|
"learning_rate": 7.376767676767676e-07, |
|
"loss": 0.0577, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 4.985733509063721, |
|
"learning_rate": 7.275757575757575e-07, |
|
"loss": 0.0672, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 1.4518052339553833, |
|
"learning_rate": 7.174747474747475e-07, |
|
"loss": 0.0616, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 5.120323657989502, |
|
"learning_rate": 7.073737373737373e-07, |
|
"loss": 0.0625, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 1.0982943773269653, |
|
"learning_rate": 6.972727272727273e-07, |
|
"loss": 0.0666, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 1.8519880771636963, |
|
"learning_rate": 6.871717171717171e-07, |
|
"loss": 0.0786, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 3.394327163696289, |
|
"learning_rate": 6.77070707070707e-07, |
|
"loss": 0.0786, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 9.026866912841797, |
|
"learning_rate": 6.66969696969697e-07, |
|
"loss": 0.0677, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 3.718776226043701, |
|
"learning_rate": 6.568686868686868e-07, |
|
"loss": 0.0576, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.7873659729957581, |
|
"learning_rate": 6.467676767676767e-07, |
|
"loss": 0.0559, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 5.039899826049805, |
|
"learning_rate": 6.366666666666667e-07, |
|
"loss": 0.0668, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 2.4250056743621826, |
|
"learning_rate": 6.266666666666667e-07, |
|
"loss": 0.0524, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 7.657347202301025, |
|
"learning_rate": 6.165656565656565e-07, |
|
"loss": 0.0665, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 3.9768455028533936, |
|
"learning_rate": 6.064646464646465e-07, |
|
"loss": 0.0649, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"eval_loss": 0.06388480961322784, |
|
"eval_runtime": 200.3994, |
|
"eval_samples_per_second": 4.99, |
|
"eval_steps_per_second": 1.248, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 6.998288154602051, |
|
"learning_rate": 5.963636363636363e-07, |
|
"loss": 0.0488, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 2.007672071456909, |
|
"learning_rate": 5.862626262626262e-07, |
|
"loss": 0.0683, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.5769347548484802, |
|
"learning_rate": 5.761616161616162e-07, |
|
"loss": 0.0544, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.005931401159614325, |
|
"learning_rate": 5.660606060606061e-07, |
|
"loss": 0.0603, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 8.401103973388672, |
|
"learning_rate": 5.559595959595959e-07, |
|
"loss": 0.0531, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 3.780881404876709, |
|
"learning_rate": 5.459595959595959e-07, |
|
"loss": 0.0729, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 8.754217147827148, |
|
"learning_rate": 5.358585858585858e-07, |
|
"loss": 0.0766, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 2.9296135902404785, |
|
"learning_rate": 5.257575757575757e-07, |
|
"loss": 0.0651, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 4.884037494659424, |
|
"learning_rate": 5.156565656565657e-07, |
|
"loss": 0.0623, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 1.247363567352295, |
|
"learning_rate": 5.055555555555555e-07, |
|
"loss": 0.0498, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 5.299038887023926, |
|
"learning_rate": 4.954545454545454e-07, |
|
"loss": 0.0558, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 3.8896191120147705, |
|
"learning_rate": 4.853535353535353e-07, |
|
"loss": 0.0733, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 2.1511714458465576, |
|
"learning_rate": 4.752525252525252e-07, |
|
"loss": 0.0647, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 3.958899736404419, |
|
"learning_rate": 4.6515151515151513e-07, |
|
"loss": 0.0594, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 7.097384929656982, |
|
"learning_rate": 4.55050505050505e-07, |
|
"loss": 0.0623, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 6.70374870300293, |
|
"learning_rate": 4.449494949494949e-07, |
|
"loss": 0.0669, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 1.3009291887283325, |
|
"learning_rate": 4.3484848484848483e-07, |
|
"loss": 0.0528, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 4.70461368560791, |
|
"learning_rate": 4.2474747474747474e-07, |
|
"loss": 0.0668, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.03208538889884949, |
|
"learning_rate": 4.1464646464646466e-07, |
|
"loss": 0.0641, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.0016973208403214812, |
|
"learning_rate": 4.045454545454545e-07, |
|
"loss": 0.064, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"eval_loss": 0.06021637097001076, |
|
"eval_runtime": 200.1582, |
|
"eval_samples_per_second": 4.996, |
|
"eval_steps_per_second": 1.249, |
|
"step": 6000 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 10000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 2000, |
|
"total_flos": 5.65692592029696e+17, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|