{ "best_metric": null, "best_model_checkpoint": null, "epoch": 30.0, "eval_steps": 500, "global_step": 37620, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.39872408293460926, "grad_norm": 0.7819421887397766, "learning_rate": 4.800637958532696e-05, "loss": 8.5954, "step": 500 }, { "epoch": 0.7974481658692185, "grad_norm": 1.4368996620178223, "learning_rate": 4.6012759170653905e-05, "loss": 7.6895, "step": 1000 }, { "epoch": 1.1961722488038278, "grad_norm": 1.0771726369857788, "learning_rate": 4.401913875598087e-05, "loss": 7.5338, "step": 1500 }, { "epoch": 1.594896331738437, "grad_norm": 1.4085040092468262, "learning_rate": 4.2025518341307815e-05, "loss": 7.4062, "step": 2000 }, { "epoch": 1.9936204146730463, "grad_norm": 1.2863935232162476, "learning_rate": 4.003189792663477e-05, "loss": 7.3026, "step": 2500 }, { "epoch": 2.3923444976076556, "grad_norm": 1.343314528465271, "learning_rate": 3.8038277511961725e-05, "loss": 7.2067, "step": 3000 }, { "epoch": 2.7910685805422646, "grad_norm": 1.8292430639266968, "learning_rate": 3.604465709728867e-05, "loss": 7.1446, "step": 3500 }, { "epoch": 3.189792663476874, "grad_norm": 1.3264261484146118, "learning_rate": 3.4051036682615634e-05, "loss": 7.075, "step": 4000 }, { "epoch": 3.588516746411483, "grad_norm": 1.511952519416809, "learning_rate": 3.205741626794259e-05, "loss": 7.0362, "step": 4500 }, { "epoch": 3.9872408293460926, "grad_norm": 1.9909762144088745, "learning_rate": 3.0063795853269537e-05, "loss": 6.9639, "step": 5000 }, { "epoch": 4.385964912280702, "grad_norm": 1.8488012552261353, "learning_rate": 2.8070175438596492e-05, "loss": 6.9264, "step": 5500 }, { "epoch": 4.784688995215311, "grad_norm": 1.4738715887069702, "learning_rate": 2.6076555023923443e-05, "loss": 6.9039, "step": 6000 }, { "epoch": 5.18341307814992, "grad_norm": 1.7991809844970703, "learning_rate": 2.4082934609250398e-05, "loss": 6.8603, "step": 6500 }, { "epoch": 5.582137161084529, "grad_norm": 1.811631441116333, "learning_rate": 2.2089314194577353e-05, "loss": 6.8354, "step": 7000 }, { "epoch": 5.980861244019139, "grad_norm": 1.732555627822876, "learning_rate": 2.0095693779904308e-05, "loss": 6.8048, "step": 7500 }, { "epoch": 6.379585326953748, "grad_norm": 1.7755110263824463, "learning_rate": 1.8102073365231263e-05, "loss": 6.7829, "step": 8000 }, { "epoch": 6.778309409888357, "grad_norm": 2.225064277648926, "learning_rate": 1.611244019138756e-05, "loss": 6.7659, "step": 8500 }, { "epoch": 7.177033492822966, "grad_norm": 1.872410774230957, "learning_rate": 1.4118819776714515e-05, "loss": 6.7374, "step": 9000 }, { "epoch": 7.575757575757576, "grad_norm": 1.7238253355026245, "learning_rate": 1.2125199362041468e-05, "loss": 6.7349, "step": 9500 }, { "epoch": 7.974481658692185, "grad_norm": 1.8047776222229004, "learning_rate": 1.0131578947368421e-05, "loss": 6.7204, "step": 10000 }, { "epoch": 8.373205741626794, "grad_norm": 2.0625622272491455, "learning_rate": 8.145933014354067e-06, "loss": 6.6957, "step": 10500 }, { "epoch": 8.771929824561404, "grad_norm": 1.8150039911270142, "learning_rate": 6.152312599681021e-06, "loss": 6.7014, "step": 11000 }, { "epoch": 9.170653907496012, "grad_norm": 2.004660129547119, "learning_rate": 4.158692185007974e-06, "loss": 6.6855, "step": 11500 }, { "epoch": 9.569377990430622, "grad_norm": 2.1859395503997803, "learning_rate": 2.1650717703349284e-06, "loss": 6.6813, "step": 12000 }, { "epoch": 9.96810207336523, "grad_norm": 2.1167070865631104, "learning_rate": 1.7145135566188198e-07, "loss": 6.6833, "step": 12500 }, { "epoch": 10.0, "step": 12540, "total_flos": 5910924298141440.0, "train_loss": 7.017849675472083, "train_runtime": 3142.2268, "train_samples_per_second": 510.596, "train_steps_per_second": 3.991 }, { "epoch": 10.36682615629984, "grad_norm": 2.343043327331543, "learning_rate": 4.87772461456672e-05, "loss": 6.6945, "step": 13000 }, { "epoch": 10.76555023923445, "grad_norm": 2.0774621963500977, "learning_rate": 4.7448165869218504e-05, "loss": 6.6465, "step": 13500 }, { "epoch": 11.164274322169058, "grad_norm": 2.5551674365997314, "learning_rate": 4.611908559276981e-05, "loss": 6.6086, "step": 14000 }, { "epoch": 11.562998405103668, "grad_norm": 2.335573196411133, "learning_rate": 4.479000531632111e-05, "loss": 6.5677, "step": 14500 }, { "epoch": 11.961722488038278, "grad_norm": 2.7488040924072266, "learning_rate": 4.346092503987241e-05, "loss": 6.5439, "step": 15000 }, { "epoch": 12.360446570972886, "grad_norm": 2.1827917098999023, "learning_rate": 4.213184476342371e-05, "loss": 6.4975, "step": 15500 }, { "epoch": 12.759170653907496, "grad_norm": 2.1983468532562256, "learning_rate": 4.080276448697501e-05, "loss": 6.4953, "step": 16000 }, { "epoch": 13.157894736842104, "grad_norm": 2.4948813915252686, "learning_rate": 3.9473684210526316e-05, "loss": 6.4546, "step": 16500 }, { "epoch": 13.556618819776714, "grad_norm": 2.0236194133758545, "learning_rate": 3.814460393407762e-05, "loss": 6.432, "step": 17000 }, { "epoch": 13.955342902711324, "grad_norm": 2.3203213214874268, "learning_rate": 3.681552365762892e-05, "loss": 6.4038, "step": 17500 }, { "epoch": 14.354066985645932, "grad_norm": 2.347102165222168, "learning_rate": 3.5486443381180226e-05, "loss": 6.3872, "step": 18000 }, { "epoch": 14.752791068580542, "grad_norm": 2.5316317081451416, "learning_rate": 3.415736310473153e-05, "loss": 6.3687, "step": 18500 }, { "epoch": 15.0, "step": 18810, "total_flos": 8866386447212160.0, "train_loss": 2.1665261722384206, "train_runtime": 1599.9695, "train_samples_per_second": 1504.163, "train_steps_per_second": 11.756 }, { "epoch": 15.151515151515152, "grad_norm": 2.2735719680786133, "learning_rate": 4.974747474747475e-05, "loss": 6.3524, "step": 19000 }, { "epoch": 15.55023923444976, "grad_norm": 3.0628674030303955, "learning_rate": 4.9082934609250406e-05, "loss": 6.3525, "step": 19500 }, { "epoch": 15.94896331738437, "grad_norm": 2.572157382965088, "learning_rate": 4.841839447102605e-05, "loss": 6.3283, "step": 20000 }, { "epoch": 16.34768740031898, "grad_norm": 2.8556103706359863, "learning_rate": 4.775385433280171e-05, "loss": 6.3006, "step": 20500 }, { "epoch": 16.74641148325359, "grad_norm": 2.7908451557159424, "learning_rate": 4.7089314194577354e-05, "loss": 6.2868, "step": 21000 }, { "epoch": 17.1451355661882, "grad_norm": 2.637225389480591, "learning_rate": 4.6424774056353006e-05, "loss": 6.2645, "step": 21500 }, { "epoch": 17.54385964912281, "grad_norm": 2.9248273372650146, "learning_rate": 4.576023391812866e-05, "loss": 6.24, "step": 22000 }, { "epoch": 17.942583732057415, "grad_norm": 2.293976306915283, "learning_rate": 4.509569377990431e-05, "loss": 6.2331, "step": 22500 }, { "epoch": 18.341307814992025, "grad_norm": 2.7412922382354736, "learning_rate": 4.443115364167996e-05, "loss": 6.1999, "step": 23000 }, { "epoch": 18.740031897926634, "grad_norm": 3.194049119949341, "learning_rate": 4.376661350345561e-05, "loss": 6.1953, "step": 23500 }, { "epoch": 19.138755980861244, "grad_norm": 2.366976022720337, "learning_rate": 4.310207336523126e-05, "loss": 6.1713, "step": 24000 }, { "epoch": 19.537480063795854, "grad_norm": 2.459841728210449, "learning_rate": 4.2437533227006915e-05, "loss": 6.1535, "step": 24500 }, { "epoch": 19.93620414673046, "grad_norm": 2.5043203830718994, "learning_rate": 4.177432216905902e-05, "loss": 6.1481, "step": 25000 }, { "epoch": 20.33492822966507, "grad_norm": 2.6015625, "learning_rate": 4.1109782030834664e-05, "loss": 6.1228, "step": 25500 }, { "epoch": 20.73365231259968, "grad_norm": 3.39577579498291, "learning_rate": 4.0445241892610315e-05, "loss": 6.1113, "step": 26000 }, { "epoch": 21.13237639553429, "grad_norm": 2.7234418392181396, "learning_rate": 3.978070175438597e-05, "loss": 6.1074, "step": 26500 }, { "epoch": 21.5311004784689, "grad_norm": 2.6176741123199463, "learning_rate": 3.911749069643807e-05, "loss": 6.0859, "step": 27000 }, { "epoch": 21.92982456140351, "grad_norm": 2.5121538639068604, "learning_rate": 3.845427963849017e-05, "loss": 6.0804, "step": 27500 }, { "epoch": 22.328548644338117, "grad_norm": 2.72497296333313, "learning_rate": 3.778973950026582e-05, "loss": 6.0623, "step": 28000 }, { "epoch": 22.727272727272727, "grad_norm": 2.9392683506011963, "learning_rate": 3.712519936204147e-05, "loss": 6.0536, "step": 28500 }, { "epoch": 23.125996810207337, "grad_norm": 2.5510990619659424, "learning_rate": 3.646065922381712e-05, "loss": 6.0508, "step": 29000 }, { "epoch": 23.524720893141946, "grad_norm": 2.9767627716064453, "learning_rate": 3.5796119085592774e-05, "loss": 6.029, "step": 29500 }, { "epoch": 23.923444976076556, "grad_norm": 2.644033908843994, "learning_rate": 3.513157894736842e-05, "loss": 6.0329, "step": 30000 }, { "epoch": 24.322169059011163, "grad_norm": 2.623633861541748, "learning_rate": 3.446836788942052e-05, "loss": 6.003, "step": 30500 }, { "epoch": 24.720893141945773, "grad_norm": 2.6472909450531006, "learning_rate": 3.3803827751196174e-05, "loss": 6.0, "step": 31000 }, { "epoch": 25.119617224880383, "grad_norm": 3.0670645236968994, "learning_rate": 3.3139287612971825e-05, "loss": 5.9948, "step": 31500 }, { "epoch": 25.518341307814993, "grad_norm": 2.6866748332977295, "learning_rate": 3.247474747474748e-05, "loss": 5.9883, "step": 32000 }, { "epoch": 25.917065390749602, "grad_norm": 2.8670027256011963, "learning_rate": 3.181020733652313e-05, "loss": 5.9801, "step": 32500 }, { "epoch": 26.31578947368421, "grad_norm": 2.797853708267212, "learning_rate": 3.114566719829878e-05, "loss": 5.9665, "step": 33000 }, { "epoch": 26.71451355661882, "grad_norm": 3.076382637023926, "learning_rate": 3.0481127060074432e-05, "loss": 5.9655, "step": 33500 }, { "epoch": 27.11323763955343, "grad_norm": 2.7438435554504395, "learning_rate": 2.981791600212653e-05, "loss": 5.9445, "step": 34000 }, { "epoch": 27.51196172248804, "grad_norm": 3.119704484939575, "learning_rate": 2.9153375863902184e-05, "loss": 5.9319, "step": 34500 }, { "epoch": 27.91068580542265, "grad_norm": 2.594749927520752, "learning_rate": 2.8488835725677832e-05, "loss": 5.9327, "step": 35000 }, { "epoch": 28.30940988835726, "grad_norm": 2.6697604656219482, "learning_rate": 2.7824295587453487e-05, "loss": 5.9261, "step": 35500 }, { "epoch": 28.708133971291865, "grad_norm": 3.4312615394592285, "learning_rate": 2.7159755449229135e-05, "loss": 5.9146, "step": 36000 }, { "epoch": 29.106858054226475, "grad_norm": 3.0306286811828613, "learning_rate": 2.6495215311004783e-05, "loss": 5.9243, "step": 36500 }, { "epoch": 29.505582137161085, "grad_norm": 2.841744899749756, "learning_rate": 2.583067517278044e-05, "loss": 5.9046, "step": 37000 }, { "epoch": 29.904306220095695, "grad_norm": 2.9730348587036133, "learning_rate": 2.5166135034556087e-05, "loss": 5.8976, "step": 37500 }, { "epoch": 30.0, "step": 37620, "total_flos": 1.773277289442432e+16, "train_loss": 3.0397992164007466, "train_runtime": 4647.9256, "train_samples_per_second": 1035.565, "train_steps_per_second": 8.094 } ], "logging_steps": 500, "max_steps": 37620, "num_input_tokens_seen": 0, "num_train_epochs": 30, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.773277289442432e+16, "train_batch_size": 128, "trial_name": null, "trial_params": null }