{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.71976401179941, "eval_steps": 100, "global_step": 600, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.03933136676499508, "grad_norm": 0.481609046459198, "learning_rate": 0.00015, "loss": 2.0722, "step": 5 }, { "epoch": 0.07866273352999016, "grad_norm": 0.15720224380493164, "learning_rate": 0.0003, "loss": 1.4825, "step": 10 }, { "epoch": 0.11799410029498525, "grad_norm": 0.06716315448284149, "learning_rate": 0.00029759999999999997, "loss": 1.3333, "step": 15 }, { "epoch": 0.15732546705998032, "grad_norm": 0.06133478134870529, "learning_rate": 0.00029519999999999997, "loss": 1.2341, "step": 20 }, { "epoch": 0.19665683382497542, "grad_norm": 0.07264667749404907, "learning_rate": 0.00029279999999999996, "loss": 1.1756, "step": 25 }, { "epoch": 0.2359882005899705, "grad_norm": 0.07928217202425003, "learning_rate": 0.00029039999999999996, "loss": 1.1197, "step": 30 }, { "epoch": 0.2753195673549656, "grad_norm": 0.09420346468687057, "learning_rate": 0.00028799999999999995, "loss": 1.0834, "step": 35 }, { "epoch": 0.31465093411996065, "grad_norm": 0.0862259566783905, "learning_rate": 0.00028559999999999995, "loss": 1.044, "step": 40 }, { "epoch": 0.35398230088495575, "grad_norm": 0.09086894243955612, "learning_rate": 0.00028319999999999994, "loss": 1.0205, "step": 45 }, { "epoch": 0.39331366764995085, "grad_norm": 0.08469890058040619, "learning_rate": 0.0002808, "loss": 0.9798, "step": 50 }, { "epoch": 0.4326450344149459, "grad_norm": 0.10012397915124893, "learning_rate": 0.0002784, "loss": 0.9811, "step": 55 }, { "epoch": 0.471976401179941, "grad_norm": 0.08633492887020111, "learning_rate": 0.000276, "loss": 0.9556, "step": 60 }, { "epoch": 0.511307767944936, "grad_norm": 0.09879346191883087, "learning_rate": 0.0002736, "loss": 0.9446, "step": 65 }, { "epoch": 0.5506391347099312, "grad_norm": 0.08795857429504395, "learning_rate": 0.0002712, "loss": 0.9228, "step": 70 }, { "epoch": 0.5899705014749262, "grad_norm": 0.0837111845612526, "learning_rate": 0.0002688, "loss": 0.9279, "step": 75 }, { "epoch": 0.6293018682399213, "grad_norm": 0.08551318198442459, "learning_rate": 0.00026639999999999997, "loss": 0.9267, "step": 80 }, { "epoch": 0.6686332350049164, "grad_norm": 0.08481767773628235, "learning_rate": 0.00026399999999999997, "loss": 0.9082, "step": 85 }, { "epoch": 0.7079646017699115, "grad_norm": 0.100365050137043, "learning_rate": 0.00026159999999999996, "loss": 0.9028, "step": 90 }, { "epoch": 0.7472959685349065, "grad_norm": 0.08463772386312485, "learning_rate": 0.00025919999999999996, "loss": 0.8866, "step": 95 }, { "epoch": 0.7866273352999017, "grad_norm": 0.09628409892320633, "learning_rate": 0.00025679999999999995, "loss": 0.8787, "step": 100 }, { "epoch": 0.7866273352999017, "eval_loss": 0.8853636980056763, "eval_runtime": 24.3719, "eval_samples_per_second": 6.729, "eval_steps_per_second": 0.862, "step": 100 }, { "epoch": 0.8259587020648967, "grad_norm": 0.08835043758153915, "learning_rate": 0.00025439999999999995, "loss": 0.8786, "step": 105 }, { "epoch": 0.8652900688298918, "grad_norm": 0.09190791845321655, "learning_rate": 0.00025199999999999995, "loss": 0.8693, "step": 110 }, { "epoch": 0.904621435594887, "grad_norm": 0.08965795487165451, "learning_rate": 0.00024959999999999994, "loss": 0.8772, "step": 115 }, { "epoch": 0.943952802359882, "grad_norm": 0.09055910259485245, "learning_rate": 0.0002472, "loss": 0.867, "step": 120 }, { "epoch": 0.983284169124877, "grad_norm": 0.09172637015581131, "learning_rate": 0.0002448, "loss": 0.8536, "step": 125 }, { "epoch": 1.022615535889872, "grad_norm": 0.10374542325735092, "learning_rate": 0.00024239999999999998, "loss": 0.9888, "step": 130 }, { "epoch": 1.0619469026548674, "grad_norm": 0.08842068910598755, "learning_rate": 0.00023999999999999998, "loss": 0.8443, "step": 135 }, { "epoch": 1.1012782694198624, "grad_norm": 0.0736837387084961, "learning_rate": 0.0002376, "loss": 0.8457, "step": 140 }, { "epoch": 1.1406096361848574, "grad_norm": 0.07575016468763351, "learning_rate": 0.0002352, "loss": 0.8335, "step": 145 }, { "epoch": 1.1799410029498525, "grad_norm": 0.07092955708503723, "learning_rate": 0.0002328, "loss": 0.8246, "step": 150 }, { "epoch": 1.2192723697148475, "grad_norm": 0.077423095703125, "learning_rate": 0.0002304, "loss": 0.823, "step": 155 }, { "epoch": 1.2586037364798428, "grad_norm": 0.07389391213655472, "learning_rate": 0.00022799999999999999, "loss": 0.819, "step": 160 }, { "epoch": 1.2979351032448379, "grad_norm": 0.08229434490203857, "learning_rate": 0.00022559999999999998, "loss": 0.8181, "step": 165 }, { "epoch": 1.337266470009833, "grad_norm": 0.07665972411632538, "learning_rate": 0.00022319999999999998, "loss": 0.8118, "step": 170 }, { "epoch": 1.376597836774828, "grad_norm": 0.09001573175191879, "learning_rate": 0.00022079999999999997, "loss": 0.8157, "step": 175 }, { "epoch": 1.415929203539823, "grad_norm": 0.07965826243162155, "learning_rate": 0.00021839999999999997, "loss": 0.8111, "step": 180 }, { "epoch": 1.455260570304818, "grad_norm": 0.08642959594726562, "learning_rate": 0.00021599999999999996, "loss": 0.8003, "step": 185 }, { "epoch": 1.494591937069813, "grad_norm": 0.0749087929725647, "learning_rate": 0.00021359999999999996, "loss": 0.7975, "step": 190 }, { "epoch": 1.5339233038348081, "grad_norm": 0.08575734496116638, "learning_rate": 0.00021119999999999996, "loss": 0.7888, "step": 195 }, { "epoch": 1.5732546705998034, "grad_norm": 0.0887129157781601, "learning_rate": 0.00020879999999999998, "loss": 0.7857, "step": 200 }, { "epoch": 1.5732546705998034, "eval_loss": 0.8026237487792969, "eval_runtime": 24.2397, "eval_samples_per_second": 6.766, "eval_steps_per_second": 0.866, "step": 200 }, { "epoch": 1.6125860373647984, "grad_norm": 0.0926935002207756, "learning_rate": 0.00020639999999999998, "loss": 0.7877, "step": 205 }, { "epoch": 1.6519174041297935, "grad_norm": 0.08537031710147858, "learning_rate": 0.000204, "loss": 0.7767, "step": 210 }, { "epoch": 1.6912487708947888, "grad_norm": 0.0766814798116684, "learning_rate": 0.0002016, "loss": 0.785, "step": 215 }, { "epoch": 1.7305801376597838, "grad_norm": 0.08394207805395126, "learning_rate": 0.0001992, "loss": 0.7832, "step": 220 }, { "epoch": 1.7699115044247788, "grad_norm": 0.0813060775399208, "learning_rate": 0.00019679999999999999, "loss": 0.7766, "step": 225 }, { "epoch": 1.809242871189774, "grad_norm": 0.08242856711149216, "learning_rate": 0.00019439999999999998, "loss": 0.7775, "step": 230 }, { "epoch": 1.848574237954769, "grad_norm": 0.07610878348350525, "learning_rate": 0.00019199999999999998, "loss": 0.7736, "step": 235 }, { "epoch": 1.887905604719764, "grad_norm": 0.08326178044080734, "learning_rate": 0.00018959999999999997, "loss": 0.7753, "step": 240 }, { "epoch": 1.927236971484759, "grad_norm": 0.09425383061170578, "learning_rate": 0.0001872, "loss": 0.7577, "step": 245 }, { "epoch": 1.966568338249754, "grad_norm": 0.08694498240947723, "learning_rate": 0.0001848, "loss": 0.7606, "step": 250 }, { "epoch": 2.005899705014749, "grad_norm": 0.22805309295654297, "learning_rate": 0.0001824, "loss": 0.8871, "step": 255 }, { "epoch": 2.045231071779744, "grad_norm": 0.09610473364591599, "learning_rate": 0.00017999999999999998, "loss": 0.7315, "step": 260 }, { "epoch": 2.084562438544739, "grad_norm": 0.09666857868432999, "learning_rate": 0.00017759999999999998, "loss": 0.7315, "step": 265 }, { "epoch": 2.1238938053097347, "grad_norm": 0.09328849613666534, "learning_rate": 0.00017519999999999998, "loss": 0.7344, "step": 270 }, { "epoch": 2.1632251720747298, "grad_norm": 0.08137473464012146, "learning_rate": 0.00017279999999999997, "loss": 0.7347, "step": 275 }, { "epoch": 2.202556538839725, "grad_norm": 0.08166103810071945, "learning_rate": 0.00017039999999999997, "loss": 0.7281, "step": 280 }, { "epoch": 2.24188790560472, "grad_norm": 0.08074019104242325, "learning_rate": 0.000168, "loss": 0.7345, "step": 285 }, { "epoch": 2.281219272369715, "grad_norm": 0.08479057997465134, "learning_rate": 0.0001656, "loss": 0.726, "step": 290 }, { "epoch": 2.32055063913471, "grad_norm": 0.08091601729393005, "learning_rate": 0.0001632, "loss": 0.7184, "step": 295 }, { "epoch": 2.359882005899705, "grad_norm": 0.08470489084720612, "learning_rate": 0.0001608, "loss": 0.7233, "step": 300 }, { "epoch": 2.359882005899705, "eval_loss": 0.7612683176994324, "eval_runtime": 24.27, "eval_samples_per_second": 6.757, "eval_steps_per_second": 0.865, "step": 300 }, { "epoch": 2.3992133726647, "grad_norm": 0.08677177131175995, "learning_rate": 0.0001584, "loss": 0.721, "step": 305 }, { "epoch": 2.438544739429695, "grad_norm": 0.08474377542734146, "learning_rate": 0.000156, "loss": 0.7141, "step": 310 }, { "epoch": 2.47787610619469, "grad_norm": 0.08565227687358856, "learning_rate": 0.0001536, "loss": 0.7173, "step": 315 }, { "epoch": 2.5172074729596856, "grad_norm": 0.08714301139116287, "learning_rate": 0.0001512, "loss": 0.7274, "step": 320 }, { "epoch": 2.5565388397246807, "grad_norm": 0.0934271439909935, "learning_rate": 0.00014879999999999998, "loss": 0.7263, "step": 325 }, { "epoch": 2.5958702064896757, "grad_norm": 0.08581375330686569, "learning_rate": 0.00014639999999999998, "loss": 0.7248, "step": 330 }, { "epoch": 2.6352015732546707, "grad_norm": 0.08378680050373077, "learning_rate": 0.00014399999999999998, "loss": 0.721, "step": 335 }, { "epoch": 2.674532940019666, "grad_norm": 0.08449660986661911, "learning_rate": 0.00014159999999999997, "loss": 0.7156, "step": 340 }, { "epoch": 2.713864306784661, "grad_norm": 0.08646751940250397, "learning_rate": 0.0001392, "loss": 0.7094, "step": 345 }, { "epoch": 2.753195673549656, "grad_norm": 0.08911272883415222, "learning_rate": 0.0001368, "loss": 0.709, "step": 350 }, { "epoch": 2.792527040314651, "grad_norm": 0.0970829427242279, "learning_rate": 0.0001344, "loss": 0.7107, "step": 355 }, { "epoch": 2.831858407079646, "grad_norm": 0.0854572132229805, "learning_rate": 0.00013199999999999998, "loss": 0.7148, "step": 360 }, { "epoch": 2.871189773844641, "grad_norm": 0.08210612088441849, "learning_rate": 0.00012959999999999998, "loss": 0.7132, "step": 365 }, { "epoch": 2.910521140609636, "grad_norm": 0.0925467386841774, "learning_rate": 0.00012719999999999997, "loss": 0.7201, "step": 370 }, { "epoch": 2.949852507374631, "grad_norm": 0.09149914979934692, "learning_rate": 0.00012479999999999997, "loss": 0.7086, "step": 375 }, { "epoch": 2.989183874139626, "grad_norm": 0.0827464610338211, "learning_rate": 0.0001224, "loss": 0.7102, "step": 380 }, { "epoch": 3.0285152409046217, "grad_norm": 0.09861475974321365, "learning_rate": 0.00011999999999999999, "loss": 0.8086, "step": 385 }, { "epoch": 3.0678466076696167, "grad_norm": 0.09810496121644974, "learning_rate": 0.0001176, "loss": 0.6784, "step": 390 }, { "epoch": 3.1071779744346117, "grad_norm": 0.08657824248075485, "learning_rate": 0.0001152, "loss": 0.6818, "step": 395 }, { "epoch": 3.146509341199607, "grad_norm": 0.08861815184354782, "learning_rate": 0.00011279999999999999, "loss": 0.6755, "step": 400 }, { "epoch": 3.146509341199607, "eval_loss": 0.7408613562583923, "eval_runtime": 24.2895, "eval_samples_per_second": 6.752, "eval_steps_per_second": 0.865, "step": 400 }, { "epoch": 3.185840707964602, "grad_norm": 0.09166675060987473, "learning_rate": 0.00011039999999999999, "loss": 0.6797, "step": 405 }, { "epoch": 3.225172074729597, "grad_norm": 0.08929497748613358, "learning_rate": 0.00010799999999999998, "loss": 0.6803, "step": 410 }, { "epoch": 3.264503441494592, "grad_norm": 0.08206567913293839, "learning_rate": 0.00010559999999999998, "loss": 0.674, "step": 415 }, { "epoch": 3.303834808259587, "grad_norm": 0.08605019748210907, "learning_rate": 0.00010319999999999999, "loss": 0.671, "step": 420 }, { "epoch": 3.343166175024582, "grad_norm": 0.09013310074806213, "learning_rate": 0.0001008, "loss": 0.6741, "step": 425 }, { "epoch": 3.382497541789577, "grad_norm": 0.08961386978626251, "learning_rate": 9.839999999999999e-05, "loss": 0.6788, "step": 430 }, { "epoch": 3.421828908554572, "grad_norm": 0.08613158762454987, "learning_rate": 9.599999999999999e-05, "loss": 0.6681, "step": 435 }, { "epoch": 3.4611602753195676, "grad_norm": 0.08968936651945114, "learning_rate": 9.36e-05, "loss": 0.6726, "step": 440 }, { "epoch": 3.5004916420845626, "grad_norm": 0.0898197814822197, "learning_rate": 9.12e-05, "loss": 0.6693, "step": 445 }, { "epoch": 3.5398230088495577, "grad_norm": 0.09239344298839569, "learning_rate": 8.879999999999999e-05, "loss": 0.6648, "step": 450 }, { "epoch": 3.5791543756145527, "grad_norm": 0.08533533662557602, "learning_rate": 8.639999999999999e-05, "loss": 0.6659, "step": 455 }, { "epoch": 3.618485742379548, "grad_norm": 0.08410927653312683, "learning_rate": 8.4e-05, "loss": 0.6725, "step": 460 }, { "epoch": 3.657817109144543, "grad_norm": 0.08881025016307831, "learning_rate": 8.16e-05, "loss": 0.67, "step": 465 }, { "epoch": 3.697148475909538, "grad_norm": 0.08324339985847473, "learning_rate": 7.92e-05, "loss": 0.6722, "step": 470 }, { "epoch": 3.736479842674533, "grad_norm": 0.09193731844425201, "learning_rate": 7.68e-05, "loss": 0.6689, "step": 475 }, { "epoch": 3.775811209439528, "grad_norm": 0.0902351662516594, "learning_rate": 7.439999999999999e-05, "loss": 0.6784, "step": 480 }, { "epoch": 3.815142576204523, "grad_norm": 0.08959916979074478, "learning_rate": 7.199999999999999e-05, "loss": 0.6808, "step": 485 }, { "epoch": 3.854473942969518, "grad_norm": 0.0888456478714943, "learning_rate": 6.96e-05, "loss": 0.6703, "step": 490 }, { "epoch": 3.893805309734513, "grad_norm": 0.08982842415571213, "learning_rate": 6.72e-05, "loss": 0.675, "step": 495 }, { "epoch": 3.933136676499508, "grad_norm": 0.0874968022108078, "learning_rate": 6.479999999999999e-05, "loss": 0.6673, "step": 500 }, { "epoch": 3.933136676499508, "eval_loss": 0.726395308971405, "eval_runtime": 24.2191, "eval_samples_per_second": 6.772, "eval_steps_per_second": 0.867, "step": 500 }, { "epoch": 3.972468043264503, "grad_norm": 0.08432712405920029, "learning_rate": 6.239999999999999e-05, "loss": 0.6747, "step": 505 }, { "epoch": 4.011799410029498, "grad_norm": 0.09498456120491028, "learning_rate": 5.9999999999999995e-05, "loss": 0.7806, "step": 510 }, { "epoch": 4.051130776794493, "grad_norm": 0.09371250122785568, "learning_rate": 5.76e-05, "loss": 0.6417, "step": 515 }, { "epoch": 4.090462143559488, "grad_norm": 0.09341807663440704, "learning_rate": 5.519999999999999e-05, "loss": 0.6399, "step": 520 }, { "epoch": 4.129793510324483, "grad_norm": 0.08755338191986084, "learning_rate": 5.279999999999999e-05, "loss": 0.6408, "step": 525 }, { "epoch": 4.169124877089478, "grad_norm": 0.08822200447320938, "learning_rate": 5.04e-05, "loss": 0.6407, "step": 530 }, { "epoch": 4.208456243854474, "grad_norm": 0.08952053636312485, "learning_rate": 4.7999999999999994e-05, "loss": 0.645, "step": 535 }, { "epoch": 4.247787610619469, "grad_norm": 0.0895063579082489, "learning_rate": 4.56e-05, "loss": 0.6409, "step": 540 }, { "epoch": 4.2871189773844645, "grad_norm": 0.08820494264364243, "learning_rate": 4.319999999999999e-05, "loss": 0.6422, "step": 545 }, { "epoch": 4.3264503441494595, "grad_norm": 0.08639927208423615, "learning_rate": 4.08e-05, "loss": 0.6476, "step": 550 }, { "epoch": 4.3657817109144545, "grad_norm": 0.08915071934461594, "learning_rate": 3.84e-05, "loss": 0.6389, "step": 555 }, { "epoch": 4.40511307767945, "grad_norm": 0.08612582087516785, "learning_rate": 3.5999999999999994e-05, "loss": 0.6401, "step": 560 }, { "epoch": 4.444444444444445, "grad_norm": 0.08792046457529068, "learning_rate": 3.36e-05, "loss": 0.6479, "step": 565 }, { "epoch": 4.48377581120944, "grad_norm": 0.08458750694990158, "learning_rate": 3.119999999999999e-05, "loss": 0.6448, "step": 570 }, { "epoch": 4.523107177974435, "grad_norm": 0.0890946015715599, "learning_rate": 2.88e-05, "loss": 0.6441, "step": 575 }, { "epoch": 4.56243854473943, "grad_norm": 0.08621781319379807, "learning_rate": 2.6399999999999995e-05, "loss": 0.6422, "step": 580 }, { "epoch": 4.601769911504425, "grad_norm": 0.08367058634757996, "learning_rate": 2.3999999999999997e-05, "loss": 0.6479, "step": 585 }, { "epoch": 4.64110127826942, "grad_norm": 0.08786690980195999, "learning_rate": 2.1599999999999996e-05, "loss": 0.6346, "step": 590 }, { "epoch": 4.680432645034415, "grad_norm": 0.08560498058795929, "learning_rate": 1.92e-05, "loss": 0.6413, "step": 595 }, { "epoch": 4.71976401179941, "grad_norm": 0.08328807353973389, "learning_rate": 1.68e-05, "loss": 0.6399, "step": 600 }, { "epoch": 4.71976401179941, "eval_loss": 0.7229037880897522, "eval_runtime": 24.3039, "eval_samples_per_second": 6.748, "eval_steps_per_second": 0.864, "step": 600 } ], "logging_steps": 5, "max_steps": 635, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 3.043845109381595e+19, "train_batch_size": 4, "trial_name": null, "trial_params": null }