{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.2338062924120914, "eval_steps": 500, "global_step": 2000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.030845157310302282, "grad_norm": 2.8206074237823486, "learning_rate": 1.0277492291880782e-05, "loss": 1.8082, "step": 50 }, { "epoch": 0.061690314620604564, "grad_norm": 3.4183013439178467, "learning_rate": 2.0554984583761563e-05, "loss": 0.6538, "step": 100 }, { "epoch": 0.09253547193090685, "grad_norm": 2.170591354370117, "learning_rate": 3.083247687564235e-05, "loss": 0.4563, "step": 150 }, { "epoch": 0.12338062924120913, "grad_norm": 1.4687080383300781, "learning_rate": 4.110996916752313e-05, "loss": 0.4263, "step": 200 }, { "epoch": 0.15422578655151142, "grad_norm": 1.836676836013794, "learning_rate": 5.1387461459403907e-05, "loss": 0.3994, "step": 250 }, { "epoch": 0.1850709438618137, "grad_norm": 1.2718663215637207, "learning_rate": 6.16649537512847e-05, "loss": 0.3665, "step": 300 }, { "epoch": 0.215916101172116, "grad_norm": 1.6945191621780396, "learning_rate": 7.194244604316547e-05, "loss": 0.3577, "step": 350 }, { "epoch": 0.24676125848241826, "grad_norm": 1.2829898595809937, "learning_rate": 8.221993833504625e-05, "loss": 0.347, "step": 400 }, { "epoch": 0.27760641579272055, "grad_norm": 1.01521635055542, "learning_rate": 9.249743062692704e-05, "loss": 0.3288, "step": 450 }, { "epoch": 0.30845157310302285, "grad_norm": 1.522111415863037, "learning_rate": 0.00010277492291880781, "loss": 0.3267, "step": 500 }, { "epoch": 0.3392967304133251, "grad_norm": 0.9678927659988403, "learning_rate": 0.00011305241521068859, "loss": 0.3198, "step": 550 }, { "epoch": 0.3701418877236274, "grad_norm": 1.2144405841827393, "learning_rate": 0.0001233299075025694, "loss": 0.3099, "step": 600 }, { "epoch": 0.4009870450339297, "grad_norm": 1.3122639656066895, "learning_rate": 0.00013360739979445017, "loss": 0.2929, "step": 650 }, { "epoch": 0.431832202344232, "grad_norm": 1.0934101343154907, "learning_rate": 0.00014388489208633093, "loss": 0.3003, "step": 700 }, { "epoch": 0.4626773596545342, "grad_norm": 0.7938969731330872, "learning_rate": 0.00015416238437821172, "loss": 0.2956, "step": 750 }, { "epoch": 0.4935225169648365, "grad_norm": 0.6571168303489685, "learning_rate": 0.0001644398766700925, "loss": 0.2736, "step": 800 }, { "epoch": 0.5243676742751388, "grad_norm": 1.0073938369750977, "learning_rate": 0.0001747173689619733, "loss": 0.2892, "step": 850 }, { "epoch": 0.5552128315854411, "grad_norm": 0.9874083399772644, "learning_rate": 0.00018499486125385408, "loss": 0.2723, "step": 900 }, { "epoch": 0.5860579888957433, "grad_norm": 1.1770968437194824, "learning_rate": 0.00019527235354573487, "loss": 0.2855, "step": 950 }, { "epoch": 0.6169031462060457, "grad_norm": 1.00326669216156, "learning_rate": 0.00019997622717095418, "loss": 0.2587, "step": 1000 }, { "epoch": 0.6477483035163479, "grad_norm": 1.0380828380584717, "learning_rate": 0.0001998067088192682, "loss": 0.2764, "step": 1050 }, { "epoch": 0.6785934608266502, "grad_norm": 1.430301547050476, "learning_rate": 0.00019947447034120033, "loss": 0.2565, "step": 1100 }, { "epoch": 0.7094386181369525, "grad_norm": 1.0970648527145386, "learning_rate": 0.00019898005340261433, "loss": 0.2685, "step": 1150 }, { "epoch": 0.7402837754472548, "grad_norm": 0.9110261797904968, "learning_rate": 0.0001983242640774473, "loss": 0.2489, "step": 1200 }, { "epoch": 0.7711289327575571, "grad_norm": 1.0265332460403442, "learning_rate": 0.00019750817153352506, "loss": 0.2425, "step": 1250 }, { "epoch": 0.8019740900678594, "grad_norm": 0.8820884823799133, "learning_rate": 0.00019653310628944164, "loss": 0.2402, "step": 1300 }, { "epoch": 0.8328192473781616, "grad_norm": 0.933083713054657, "learning_rate": 0.00019540065804534467, "loss": 0.2444, "step": 1350 }, { "epoch": 0.863664404688464, "grad_norm": 0.8229042887687683, "learning_rate": 0.00019411267309116375, "loss": 0.2172, "step": 1400 }, { "epoch": 0.8945095619987662, "grad_norm": 0.9641085267066956, "learning_rate": 0.00019267125129650688, "loss": 0.2501, "step": 1450 }, { "epoch": 0.9253547193090684, "grad_norm": 0.6993410587310791, "learning_rate": 0.00019107874268713254, "loss": 0.2246, "step": 1500 }, { "epoch": 0.9561998766193708, "grad_norm": 1.0574674606323242, "learning_rate": 0.00018933774361357917, "loss": 0.2265, "step": 1550 }, { "epoch": 0.987045033929673, "grad_norm": 1.0701500177383423, "learning_rate": 0.0001874510925181983, "loss": 0.2116, "step": 1600 }, { "epoch": 1.0178901912399754, "grad_norm": 0.44411325454711914, "learning_rate": 0.0001854218653074927, "loss": 0.1978, "step": 1650 }, { "epoch": 1.0487353485502775, "grad_norm": 0.8408161401748657, "learning_rate": 0.0001832533703373043, "loss": 0.1701, "step": 1700 }, { "epoch": 1.0795805058605799, "grad_norm": 0.7417210936546326, "learning_rate": 0.0001809491430190276, "loss": 0.1873, "step": 1750 }, { "epoch": 1.1104256631708822, "grad_norm": 0.859341025352478, "learning_rate": 0.00017851294005564254, "loss": 0.1717, "step": 1800 }, { "epoch": 1.1412708204811843, "grad_norm": 0.6794934272766113, "learning_rate": 0.0001759487333169642, "loss": 0.1732, "step": 1850 }, { "epoch": 1.1721159777914867, "grad_norm": 0.5129622220993042, "learning_rate": 0.00017326070336409427, "loss": 0.172, "step": 1900 }, { "epoch": 1.202961135101789, "grad_norm": 0.5867941379547119, "learning_rate": 0.00017045323263363272, "loss": 0.1724, "step": 1950 }, { "epoch": 1.2338062924120914, "grad_norm": 0.887195348739624, "learning_rate": 0.0001675308982927608, "loss": 0.1655, "step": 2000 } ], "logging_steps": 50, "max_steps": 4863, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.4223873847328768e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }