|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.45599635202918376, |
|
"eval_steps": 500, |
|
"global_step": 1000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.022799817601459188, |
|
"grad_norm": 4.15625, |
|
"learning_rate": 3.3333333333333335e-05, |
|
"loss": 5.7819, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.045599635202918376, |
|
"grad_norm": 2.28125, |
|
"learning_rate": 6.666666666666667e-05, |
|
"loss": 2.6693, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.06839945280437756, |
|
"grad_norm": 0.021484375, |
|
"learning_rate": 0.0001, |
|
"loss": 0.0543, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.09119927040583675, |
|
"grad_norm": 0.02978515625, |
|
"learning_rate": 0.00013333333333333334, |
|
"loss": 0.0068, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.11399908800729594, |
|
"grad_norm": 0.072265625, |
|
"learning_rate": 0.0001666666666666667, |
|
"loss": 0.003, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.13679890560875513, |
|
"grad_norm": 0.007110595703125, |
|
"learning_rate": 0.0002, |
|
"loss": 0.0028, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.15959872321021432, |
|
"grad_norm": 0.012939453125, |
|
"learning_rate": 0.00019840738971173754, |
|
"loss": 0.001, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.1823985408116735, |
|
"grad_norm": 0.00201416015625, |
|
"learning_rate": 0.00019681477942347508, |
|
"loss": 0.0009, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.2051983584131327, |
|
"grad_norm": 0.005706787109375, |
|
"learning_rate": 0.00019522216913521264, |
|
"loss": 0.0023, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.22799817601459188, |
|
"grad_norm": 0.0269775390625, |
|
"learning_rate": 0.00019362955884695015, |
|
"loss": 0.0022, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.2507979936160511, |
|
"grad_norm": 0.0015106201171875, |
|
"learning_rate": 0.0001920369485586877, |
|
"loss": 0.0021, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.27359781121751026, |
|
"grad_norm": 0.023193359375, |
|
"learning_rate": 0.00019044433827042522, |
|
"loss": 0.0015, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.29639762881896947, |
|
"grad_norm": 0.07421875, |
|
"learning_rate": 0.00018885172798216278, |
|
"loss": 0.0011, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.31919744642042863, |
|
"grad_norm": 0.00177001953125, |
|
"learning_rate": 0.0001872591176939003, |
|
"loss": 0.0019, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.34199726402188785, |
|
"grad_norm": 0.0230712890625, |
|
"learning_rate": 0.00018566650740563785, |
|
"loss": 0.0008, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.364797081623347, |
|
"grad_norm": 0.00101470947265625, |
|
"learning_rate": 0.0001840738971173754, |
|
"loss": 0.001, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.3875968992248062, |
|
"grad_norm": 0.0216064453125, |
|
"learning_rate": 0.00018248128682911291, |
|
"loss": 0.0011, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.4103967168262654, |
|
"grad_norm": 7.43865966796875e-05, |
|
"learning_rate": 0.00018088867654085048, |
|
"loss": 0.001, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.4331965344277246, |
|
"grad_norm": 4.315376281738281e-05, |
|
"learning_rate": 0.00017929606625258798, |
|
"loss": 0.0009, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.45599635202918376, |
|
"grad_norm": 0.0012969970703125, |
|
"learning_rate": 0.00017770345596432554, |
|
"loss": 0.0008, |
|
"step": 1000 |
|
} |
|
], |
|
"logging_steps": 50, |
|
"max_steps": 6579, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 200, |
|
"total_flos": 0.0, |
|
"train_batch_size": 128, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|