| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.0, |
| "eval_steps": 500, |
| "global_step": 346, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.01447178002894356, |
| "grad_norm": 1.2414709329605103, |
| "learning_rate": 1.3793103448275862e-06, |
| "loss": 1.2575, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.02894356005788712, |
| "grad_norm": 0.995053231716156, |
| "learning_rate": 3.103448275862069e-06, |
| "loss": 1.2964, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.04341534008683068, |
| "grad_norm": 0.6870602369308472, |
| "learning_rate": 4.827586206896552e-06, |
| "loss": 1.2136, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.05788712011577424, |
| "grad_norm": 0.7477986216545105, |
| "learning_rate": 6.551724137931035e-06, |
| "loss": 1.208, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.0723589001447178, |
| "grad_norm": 0.503740131855011, |
| "learning_rate": 8.275862068965517e-06, |
| "loss": 1.262, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.08683068017366136, |
| "grad_norm": 0.5463724732398987, |
| "learning_rate": 9.999999999999999e-06, |
| "loss": 1.1591, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.10130246020260492, |
| "grad_norm": 0.5826733112335205, |
| "learning_rate": 1.1724137931034483e-05, |
| "loss": 1.1859, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.11577424023154848, |
| "grad_norm": 0.478679895401001, |
| "learning_rate": 1.3448275862068966e-05, |
| "loss": 1.1185, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.13024602026049203, |
| "grad_norm": 0.45855218172073364, |
| "learning_rate": 1.517241379310345e-05, |
| "loss": 1.0916, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.1447178002894356, |
| "grad_norm": 0.4334068298339844, |
| "learning_rate": 1.6896551724137932e-05, |
| "loss": 1.1103, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.15918958031837915, |
| "grad_norm": 0.49706199765205383, |
| "learning_rate": 1.8620689655172415e-05, |
| "loss": 1.1351, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.1736613603473227, |
| "grad_norm": 0.5013290643692017, |
| "learning_rate": 2.0344827586206894e-05, |
| "loss": 1.1267, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.18813314037626627, |
| "grad_norm": 0.4919586181640625, |
| "learning_rate": 2.206896551724138e-05, |
| "loss": 1.0824, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.20260492040520983, |
| "grad_norm": 0.4955993890762329, |
| "learning_rate": 2.3793103448275862e-05, |
| "loss": 1.088, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.2170767004341534, |
| "grad_norm": 0.47736915946006775, |
| "learning_rate": 2.5517241379310345e-05, |
| "loss": 1.097, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.23154848046309695, |
| "grad_norm": 0.4829305112361908, |
| "learning_rate": 2.7241379310344827e-05, |
| "loss": 1.0582, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.2460202604920405, |
| "grad_norm": 0.5232416391372681, |
| "learning_rate": 2.8965517241379313e-05, |
| "loss": 1.0447, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.26049204052098407, |
| "grad_norm": 0.5268356800079346, |
| "learning_rate": 2.999989031547876e-05, |
| "loss": 1.0425, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.27496382054992763, |
| "grad_norm": 0.6044050455093384, |
| "learning_rate": 2.9998656383036702e-05, |
| "loss": 0.9967, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.2894356005788712, |
| "grad_norm": 0.5722519755363464, |
| "learning_rate": 2.9996051525662343e-05, |
| "loss": 0.9925, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.30390738060781475, |
| "grad_norm": 0.5845656991004944, |
| "learning_rate": 2.9992075981447947e-05, |
| "loss": 0.9589, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.3183791606367583, |
| "grad_norm": 0.6687687635421753, |
| "learning_rate": 2.9986730113770898e-05, |
| "loss": 0.9607, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.33285094066570187, |
| "grad_norm": 0.5653572678565979, |
| "learning_rate": 2.9980014411260523e-05, |
| "loss": 0.9765, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.3473227206946454, |
| "grad_norm": 0.6110051274299622, |
| "learning_rate": 2.9971929487753402e-05, |
| "loss": 0.9347, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.361794500723589, |
| "grad_norm": 0.6042832136154175, |
| "learning_rate": 2.9962476082237285e-05, |
| "loss": 0.8882, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.37626628075253254, |
| "grad_norm": 0.6558710932731628, |
| "learning_rate": 2.9951655058783517e-05, |
| "loss": 0.9611, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.3907380607814761, |
| "grad_norm": 0.6648174524307251, |
| "learning_rate": 2.993946740646809e-05, |
| "loss": 0.9514, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.40520984081041966, |
| "grad_norm": 0.6010047793388367, |
| "learning_rate": 2.992591423928121e-05, |
| "loss": 0.8977, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.4196816208393632, |
| "grad_norm": 0.594023585319519, |
| "learning_rate": 2.99109967960255e-05, |
| "loss": 0.9169, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.4341534008683068, |
| "grad_norm": 0.6303651928901672, |
| "learning_rate": 2.9894716440202756e-05, |
| "loss": 0.8779, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.44862518089725034, |
| "grad_norm": 0.6532771587371826, |
| "learning_rate": 2.9877074659889316e-05, |
| "loss": 0.8661, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.4630969609261939, |
| "grad_norm": 0.7434287071228027, |
| "learning_rate": 2.9858073067600054e-05, |
| "loss": 0.8627, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.47756874095513746, |
| "grad_norm": 0.7219240069389343, |
| "learning_rate": 2.983771340014098e-05, |
| "loss": 0.8855, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.492040520984081, |
| "grad_norm": 0.7225666046142578, |
| "learning_rate": 2.981599751845051e-05, |
| "loss": 0.8531, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.5065123010130246, |
| "grad_norm": 0.7195663452148438, |
| "learning_rate": 2.9792927407429344e-05, |
| "loss": 0.8174, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.5209840810419681, |
| "grad_norm": 0.7346758842468262, |
| "learning_rate": 2.976850517575906e-05, |
| "loss": 0.8491, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.5354558610709117, |
| "grad_norm": 0.7614021897315979, |
| "learning_rate": 2.9742733055709366e-05, |
| "loss": 0.8102, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.5499276410998553, |
| "grad_norm": 0.7299443483352661, |
| "learning_rate": 2.9715613402934064e-05, |
| "loss": 0.7749, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.5643994211287988, |
| "grad_norm": 0.7258509993553162, |
| "learning_rate": 2.9687148696255737e-05, |
| "loss": 0.8408, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.5788712011577424, |
| "grad_norm": 0.8355885148048401, |
| "learning_rate": 2.9657341537439176e-05, |
| "loss": 0.7793, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.5933429811866859, |
| "grad_norm": 0.7875196933746338, |
| "learning_rate": 2.9626194650953563e-05, |
| "loss": 0.8168, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.6078147612156295, |
| "grad_norm": 1.2102893590927124, |
| "learning_rate": 2.9593710883723465e-05, |
| "loss": 0.7554, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.622286541244573, |
| "grad_norm": 0.9075624346733093, |
| "learning_rate": 2.95598932048686e-05, |
| "loss": 0.7782, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.6367583212735166, |
| "grad_norm": 0.9814792275428772, |
| "learning_rate": 2.9524744705432446e-05, |
| "loss": 0.7679, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.6512301013024602, |
| "grad_norm": 0.8673927187919617, |
| "learning_rate": 2.9488268598099735e-05, |
| "loss": 0.7131, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.6657018813314037, |
| "grad_norm": 1.2335319519042969, |
| "learning_rate": 2.945046821690277e-05, |
| "loss": 0.7368, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.6801736613603473, |
| "grad_norm": 1.0025559663772583, |
| "learning_rate": 2.9411347016916704e-05, |
| "loss": 0.7447, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.6946454413892909, |
| "grad_norm": 0.8002074360847473, |
| "learning_rate": 2.9370908573943737e-05, |
| "loss": 0.7396, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.7091172214182344, |
| "grad_norm": 0.7924606800079346, |
| "learning_rate": 2.9329156584186267e-05, |
| "loss": 0.7821, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.723589001447178, |
| "grad_norm": 0.8294292092323303, |
| "learning_rate": 2.9286094863909053e-05, |
| "loss": 0.7546, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.7380607814761215, |
| "grad_norm": 0.8875476121902466, |
| "learning_rate": 2.924172734909038e-05, |
| "loss": 0.7541, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.7525325615050651, |
| "grad_norm": 1.0516374111175537, |
| "learning_rate": 2.919605809506233e-05, |
| "loss": 0.7339, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.7670043415340086, |
| "grad_norm": 0.9758723378181458, |
| "learning_rate": 2.9149091276140066e-05, |
| "loss": 0.7286, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.7814761215629522, |
| "grad_norm": 0.8721650838851929, |
| "learning_rate": 2.910083118524034e-05, |
| "loss": 0.7224, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.7959479015918958, |
| "grad_norm": 0.8282304406166077, |
| "learning_rate": 2.9051282233489065e-05, |
| "loss": 0.6857, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.8104196816208393, |
| "grad_norm": 0.9779431223869324, |
| "learning_rate": 2.900044894981813e-05, |
| "loss": 0.6938, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.8248914616497829, |
| "grad_norm": 0.842149019241333, |
| "learning_rate": 2.894833598055147e-05, |
| "loss": 0.6996, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.8393632416787264, |
| "grad_norm": 0.8657100200653076, |
| "learning_rate": 2.8894948088980338e-05, |
| "loss": 0.6343, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.85383502170767, |
| "grad_norm": 0.8805454969406128, |
| "learning_rate": 2.8840290154927955e-05, |
| "loss": 0.6506, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.8683068017366136, |
| "grad_norm": 0.8671079874038696, |
| "learning_rate": 2.878436717430346e-05, |
| "loss": 0.7051, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.8827785817655571, |
| "grad_norm": 0.9610555768013, |
| "learning_rate": 2.8727184258645276e-05, |
| "loss": 0.6888, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.8972503617945007, |
| "grad_norm": 0.9625582695007324, |
| "learning_rate": 2.8668746634653908e-05, |
| "loss": 0.6669, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.9117221418234442, |
| "grad_norm": 0.8685387372970581, |
| "learning_rate": 2.860905964371418e-05, |
| "loss": 0.645, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.9261939218523878, |
| "grad_norm": 1.1596406698226929, |
| "learning_rate": 2.8548128741407043e-05, |
| "loss": 0.6432, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.9406657018813314, |
| "grad_norm": 0.9292768239974976, |
| "learning_rate": 2.8485959497010906e-05, |
| "loss": 0.6599, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.9551374819102749, |
| "grad_norm": 1.0150357484817505, |
| "learning_rate": 2.8422557592992584e-05, |
| "loss": 0.5829, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.9696092619392185, |
| "grad_norm": 1.0730253458023071, |
| "learning_rate": 2.83579288244879e-05, |
| "loss": 0.615, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.984081041968162, |
| "grad_norm": 1.0182080268859863, |
| "learning_rate": 2.829207909877201e-05, |
| "loss": 0.6109, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.9985528219971056, |
| "grad_norm": 0.9661368131637573, |
| "learning_rate": 2.8225014434719423e-05, |
| "loss": 0.5798, |
| "step": 345 |
| } |
| ], |
| "logging_steps": 5, |
| "max_steps": 1730, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 5, |
| "save_steps": 2000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 5.213914221804257e+17, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|