diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,5519 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.9998921074234783, + "eval_steps": 400, + "global_step": 6661, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0015011141081271255, + "grad_norm": 0.8901742100715637, + "learning_rate": 1.1976047904191619e-06, + "loss": 2.5666, + "num_input_tokens_seen": 5242880, + "step": 10 + }, + { + "epoch": 0.003002228216254251, + "grad_norm": 0.8234608769416809, + "learning_rate": 2.3952095808383237e-06, + "loss": 2.5448, + "num_input_tokens_seen": 10485760, + "step": 20 + }, + { + "epoch": 0.004503342324381376, + "grad_norm": 0.6512044668197632, + "learning_rate": 3.592814371257485e-06, + "loss": 2.4883, + "num_input_tokens_seen": 15728640, + "step": 30 + }, + { + "epoch": 0.006004456432508502, + "grad_norm": 0.6285302639007568, + "learning_rate": 4.7904191616766475e-06, + "loss": 2.4675, + "num_input_tokens_seen": 20971520, + "step": 40 + }, + { + "epoch": 0.007505570540635628, + "grad_norm": 0.5729206204414368, + "learning_rate": 5.9880239520958085e-06, + "loss": 2.4332, + "num_input_tokens_seen": 26214400, + "step": 50 + }, + { + "epoch": 0.009006684648762753, + "grad_norm": 0.54844731092453, + "learning_rate": 7.18562874251497e-06, + "loss": 2.4374, + "num_input_tokens_seen": 31457280, + "step": 60 + }, + { + "epoch": 0.01050779875688988, + "grad_norm": 0.5901165008544922, + "learning_rate": 8.383233532934131e-06, + "loss": 2.4057, + "num_input_tokens_seen": 36700160, + "step": 70 + }, + { + "epoch": 0.012008912865017004, + "grad_norm": 0.6103231906890869, + "learning_rate": 9.580838323353295e-06, + "loss": 2.3765, + "num_input_tokens_seen": 41943040, + "step": 80 + }, + { + "epoch": 0.013510026973144131, + "grad_norm": 0.502941906452179, + "learning_rate": 1.0778443113772457e-05, + "loss": 2.3877, + "num_input_tokens_seen": 47185920, + "step": 90 + }, + { + "epoch": 0.015011141081271256, + "grad_norm": 0.6203625202178955, + "learning_rate": 1.1976047904191617e-05, + "loss": 2.3622, + "num_input_tokens_seen": 52428800, + "step": 100 + }, + { + "epoch": 0.016512255189398382, + "grad_norm": 0.658467173576355, + "learning_rate": 1.3173652694610779e-05, + "loss": 2.3567, + "num_input_tokens_seen": 57671680, + "step": 110 + }, + { + "epoch": 0.018013369297525506, + "grad_norm": 0.5568761229515076, + "learning_rate": 1.437125748502994e-05, + "loss": 2.3679, + "num_input_tokens_seen": 62914560, + "step": 120 + }, + { + "epoch": 0.019514483405652632, + "grad_norm": 0.6238291263580322, + "learning_rate": 1.5568862275449103e-05, + "loss": 2.3829, + "num_input_tokens_seen": 68157440, + "step": 130 + }, + { + "epoch": 0.02101559751377976, + "grad_norm": 0.6000885367393494, + "learning_rate": 1.6766467065868263e-05, + "loss": 2.3472, + "num_input_tokens_seen": 73400320, + "step": 140 + }, + { + "epoch": 0.022516711621906885, + "grad_norm": 0.6424698829650879, + "learning_rate": 1.7964071856287426e-05, + "loss": 2.3285, + "num_input_tokens_seen": 78643200, + "step": 150 + }, + { + "epoch": 0.02401782573003401, + "grad_norm": 0.6143007874488831, + "learning_rate": 1.916167664670659e-05, + "loss": 2.3151, + "num_input_tokens_seen": 83886080, + "step": 160 + }, + { + "epoch": 0.025518939838161135, + "grad_norm": 0.6206329464912415, + "learning_rate": 2.035928143712575e-05, + "loss": 2.3368, + "num_input_tokens_seen": 89128960, + "step": 170 + }, + { + "epoch": 0.027020053946288262, + "grad_norm": 0.6878198385238647, + "learning_rate": 2.1556886227544914e-05, + "loss": 2.3367, + "num_input_tokens_seen": 94371840, + "step": 180 + }, + { + "epoch": 0.028521168054415385, + "grad_norm": 0.8266871571540833, + "learning_rate": 2.2754491017964074e-05, + "loss": 2.3286, + "num_input_tokens_seen": 99614720, + "step": 190 + }, + { + "epoch": 0.03002228216254251, + "grad_norm": 0.5854138135910034, + "learning_rate": 2.3952095808383234e-05, + "loss": 2.303, + "num_input_tokens_seen": 104857600, + "step": 200 + }, + { + "epoch": 0.031523396270669635, + "grad_norm": 0.7110793590545654, + "learning_rate": 2.5149700598802394e-05, + "loss": 2.2985, + "num_input_tokens_seen": 110100480, + "step": 210 + }, + { + "epoch": 0.033024510378796765, + "grad_norm": 0.6561000943183899, + "learning_rate": 2.6347305389221558e-05, + "loss": 2.3154, + "num_input_tokens_seen": 115343360, + "step": 220 + }, + { + "epoch": 0.03452562448692389, + "grad_norm": 0.7813587188720703, + "learning_rate": 2.754491017964072e-05, + "loss": 2.3104, + "num_input_tokens_seen": 120586240, + "step": 230 + }, + { + "epoch": 0.03602673859505101, + "grad_norm": 0.7437533140182495, + "learning_rate": 2.874251497005988e-05, + "loss": 2.3098, + "num_input_tokens_seen": 125829120, + "step": 240 + }, + { + "epoch": 0.03752785270317814, + "grad_norm": 0.8770821690559387, + "learning_rate": 2.9940119760479045e-05, + "loss": 2.3061, + "num_input_tokens_seen": 131072000, + "step": 250 + }, + { + "epoch": 0.039028966811305264, + "grad_norm": 0.6238060593605042, + "learning_rate": 3.1137724550898205e-05, + "loss": 2.3052, + "num_input_tokens_seen": 136314880, + "step": 260 + }, + { + "epoch": 0.040530080919432394, + "grad_norm": 0.678520679473877, + "learning_rate": 3.233532934131737e-05, + "loss": 2.2934, + "num_input_tokens_seen": 141557760, + "step": 270 + }, + { + "epoch": 0.04203119502755952, + "grad_norm": 0.7275491952896118, + "learning_rate": 3.3532934131736525e-05, + "loss": 2.2938, + "num_input_tokens_seen": 146800640, + "step": 280 + }, + { + "epoch": 0.04353230913568664, + "grad_norm": 0.6745551824569702, + "learning_rate": 3.473053892215569e-05, + "loss": 2.3137, + "num_input_tokens_seen": 152043520, + "step": 290 + }, + { + "epoch": 0.04503342324381377, + "grad_norm": 0.8107547163963318, + "learning_rate": 3.592814371257485e-05, + "loss": 2.2833, + "num_input_tokens_seen": 157286400, + "step": 300 + }, + { + "epoch": 0.046534537351940894, + "grad_norm": 0.9518006443977356, + "learning_rate": 3.7125748502994016e-05, + "loss": 2.2883, + "num_input_tokens_seen": 162529280, + "step": 310 + }, + { + "epoch": 0.04803565146006802, + "grad_norm": 0.7946741580963135, + "learning_rate": 3.832335329341318e-05, + "loss": 2.292, + "num_input_tokens_seen": 167772160, + "step": 320 + }, + { + "epoch": 0.04953676556819515, + "grad_norm": 0.8819884657859802, + "learning_rate": 3.9520958083832336e-05, + "loss": 2.2874, + "num_input_tokens_seen": 173015040, + "step": 330 + }, + { + "epoch": 0.05103787967632227, + "grad_norm": 0.8569954633712769, + "learning_rate": 3.9645487837613415e-05, + "loss": 2.2701, + "num_input_tokens_seen": 178257920, + "step": 340 + }, + { + "epoch": 0.05253899378444939, + "grad_norm": 0.6152970194816589, + "learning_rate": 3.907501942235145e-05, + "loss": 2.2995, + "num_input_tokens_seen": 183500800, + "step": 350 + }, + { + "epoch": 0.054040107892576524, + "grad_norm": 0.6690587997436523, + "learning_rate": 3.852848873813304e-05, + "loss": 2.2663, + "num_input_tokens_seen": 188743680, + "step": 360 + }, + { + "epoch": 0.05554122200070365, + "grad_norm": 0.7796323895454407, + "learning_rate": 3.800426718572961e-05, + "loss": 2.2597, + "num_input_tokens_seen": 193986560, + "step": 370 + }, + { + "epoch": 0.05704233610883077, + "grad_norm": 0.9035688638687134, + "learning_rate": 3.7500877182723136e-05, + "loss": 2.2369, + "num_input_tokens_seen": 199229440, + "step": 380 + }, + { + "epoch": 0.0585434502169579, + "grad_norm": 0.7932724356651306, + "learning_rate": 3.701697462322401e-05, + "loss": 2.2514, + "num_input_tokens_seen": 204472320, + "step": 390 + }, + { + "epoch": 0.06004456432508502, + "grad_norm": 0.6233640909194946, + "learning_rate": 3.6551333764994134e-05, + "loss": 2.2572, + "num_input_tokens_seen": 209715200, + "step": 400 + }, + { + "epoch": 0.06004456432508502, + "eval_accuracy": 0.5491493691493692, + "eval_loss": 2.2461655139923096, + "eval_runtime": 95.7686, + "eval_samples_per_second": 3.133, + "eval_steps_per_second": 0.783, + "num_input_tokens_seen": 209715200, + "step": 400 + }, + { + "epoch": 0.061545678433212146, + "grad_norm": 0.5706452131271362, + "learning_rate": 3.610283415670218e-05, + "loss": 2.2674, + "num_input_tokens_seen": 214958080, + "step": 410 + }, + { + "epoch": 0.06304679254133927, + "grad_norm": 0.6556965112686157, + "learning_rate": 3.567044928762396e-05, + "loss": 2.2898, + "num_input_tokens_seen": 220200960, + "step": 420 + }, + { + "epoch": 0.0645479066494664, + "grad_norm": 0.7172439098358154, + "learning_rate": 3.5253236697846886e-05, + "loss": 2.2644, + "num_input_tokens_seen": 225443840, + "step": 430 + }, + { + "epoch": 0.06604902075759353, + "grad_norm": 0.6364426016807556, + "learning_rate": 3.485032933195115e-05, + "loss": 2.2551, + "num_input_tokens_seen": 230686720, + "step": 440 + }, + { + "epoch": 0.06755013486572065, + "grad_norm": 0.7587452530860901, + "learning_rate": 3.446092795552023e-05, + "loss": 2.2576, + "num_input_tokens_seen": 235929600, + "step": 450 + }, + { + "epoch": 0.06905124897384778, + "grad_norm": 0.6086819767951965, + "learning_rate": 3.408429448345356e-05, + "loss": 2.2768, + "num_input_tokens_seen": 241172480, + "step": 460 + }, + { + "epoch": 0.0705523630819749, + "grad_norm": 0.681690514087677, + "learning_rate": 3.371974609328701e-05, + "loss": 2.2744, + "num_input_tokens_seen": 246415360, + "step": 470 + }, + { + "epoch": 0.07205347719010202, + "grad_norm": 0.8943272233009338, + "learning_rate": 3.336665001664587e-05, + "loss": 2.2536, + "num_input_tokens_seen": 251658240, + "step": 480 + }, + { + "epoch": 0.07355459129822915, + "grad_norm": 1.1301355361938477, + "learning_rate": 3.3024418918399747e-05, + "loss": 2.2329, + "num_input_tokens_seen": 256901120, + "step": 490 + }, + { + "epoch": 0.07505570540635628, + "grad_norm": 0.7338290214538574, + "learning_rate": 3.269250678672408e-05, + "loss": 2.2592, + "num_input_tokens_seen": 262144000, + "step": 500 + }, + { + "epoch": 0.07655681951448341, + "grad_norm": 0.7819264531135559, + "learning_rate": 3.23704052686231e-05, + "loss": 2.2453, + "num_input_tokens_seen": 267386880, + "step": 510 + }, + { + "epoch": 0.07805793362261053, + "grad_norm": 0.6750678420066833, + "learning_rate": 3.205764039495589e-05, + "loss": 2.2514, + "num_input_tokens_seen": 272629760, + "step": 520 + }, + { + "epoch": 0.07955904773073766, + "grad_norm": 0.5943534970283508, + "learning_rate": 3.175376964696401e-05, + "loss": 2.2346, + "num_input_tokens_seen": 277872640, + "step": 530 + }, + { + "epoch": 0.08106016183886479, + "grad_norm": 0.7100771069526672, + "learning_rate": 3.145837932299803e-05, + "loss": 2.2316, + "num_input_tokens_seen": 283115520, + "step": 540 + }, + { + "epoch": 0.0825612759469919, + "grad_norm": 0.6773257851600647, + "learning_rate": 3.117108216979904e-05, + "loss": 2.2343, + "num_input_tokens_seen": 288358400, + "step": 550 + }, + { + "epoch": 0.08406239005511904, + "grad_norm": 0.8613666296005249, + "learning_rate": 3.089151524748689e-05, + "loss": 2.253, + "num_input_tokens_seen": 293601280, + "step": 560 + }, + { + "epoch": 0.08556350416324617, + "grad_norm": 0.5922672748565674, + "learning_rate": 3.0619338001484006e-05, + "loss": 2.2032, + "num_input_tokens_seen": 298844160, + "step": 570 + }, + { + "epoch": 0.08706461827137328, + "grad_norm": 0.8399538397789001, + "learning_rate": 3.0354230518081458e-05, + "loss": 2.2445, + "num_input_tokens_seen": 304087040, + "step": 580 + }, + { + "epoch": 0.08856573237950041, + "grad_norm": 0.8287783265113831, + "learning_rate": 3.0095891943326868e-05, + "loss": 2.2473, + "num_input_tokens_seen": 309329920, + "step": 590 + }, + { + "epoch": 0.09006684648762754, + "grad_norm": 0.6455395221710205, + "learning_rate": 2.9844039047465857e-05, + "loss": 2.2276, + "num_input_tokens_seen": 314572800, + "step": 600 + }, + { + "epoch": 0.09156796059575466, + "grad_norm": 0.6312438249588013, + "learning_rate": 2.959840491936165e-05, + "loss": 2.2383, + "num_input_tokens_seen": 319815680, + "step": 610 + }, + { + "epoch": 0.09306907470388179, + "grad_norm": 0.5632859468460083, + "learning_rate": 2.9358737777209838e-05, + "loss": 2.2342, + "num_input_tokens_seen": 325058560, + "step": 620 + }, + { + "epoch": 0.09457018881200892, + "grad_norm": 0.6137943863868713, + "learning_rate": 2.9124799883500802e-05, + "loss": 2.2058, + "num_input_tokens_seen": 330301440, + "step": 630 + }, + { + "epoch": 0.09607130292013603, + "grad_norm": 0.7079706788063049, + "learning_rate": 2.8896366553599782e-05, + "loss": 2.2235, + "num_input_tokens_seen": 335544320, + "step": 640 + }, + { + "epoch": 0.09757241702826316, + "grad_norm": 0.6246356964111328, + "learning_rate": 2.8673225248545833e-05, + "loss": 2.2363, + "num_input_tokens_seen": 340787200, + "step": 650 + }, + { + "epoch": 0.0990735311363903, + "grad_norm": 0.7458660006523132, + "learning_rate": 2.8455174743743358e-05, + "loss": 2.2199, + "num_input_tokens_seen": 346030080, + "step": 660 + }, + { + "epoch": 0.10057464524451741, + "grad_norm": 0.6742898225784302, + "learning_rate": 2.8242024366155264e-05, + "loss": 2.2293, + "num_input_tokens_seen": 351272960, + "step": 670 + }, + { + "epoch": 0.10207575935264454, + "grad_norm": 0.5833361148834229, + "learning_rate": 2.8033593293425236e-05, + "loss": 2.2346, + "num_input_tokens_seen": 356515840, + "step": 680 + }, + { + "epoch": 0.10357687346077167, + "grad_norm": 0.6706376075744629, + "learning_rate": 2.7829709909073584e-05, + "loss": 2.2353, + "num_input_tokens_seen": 361758720, + "step": 690 + }, + { + "epoch": 0.10507798756889879, + "grad_norm": 0.7505273222923279, + "learning_rate": 2.763021120854076e-05, + "loss": 2.2463, + "num_input_tokens_seen": 367001600, + "step": 700 + }, + { + "epoch": 0.10657910167702592, + "grad_norm": 0.6324784755706787, + "learning_rate": 2.743494225140684e-05, + "loss": 2.2208, + "num_input_tokens_seen": 372244480, + "step": 710 + }, + { + "epoch": 0.10808021578515305, + "grad_norm": 0.6054283380508423, + "learning_rate": 2.7243755655603398e-05, + "loss": 2.214, + "num_input_tokens_seen": 377487360, + "step": 720 + }, + { + "epoch": 0.10958132989328016, + "grad_norm": 0.6269551515579224, + "learning_rate": 2.7056511129865727e-05, + "loss": 2.2426, + "num_input_tokens_seen": 382730240, + "step": 730 + }, + { + "epoch": 0.1110824440014073, + "grad_norm": 0.8448464870452881, + "learning_rate": 2.6873075041054796e-05, + "loss": 2.2214, + "num_input_tokens_seen": 387973120, + "step": 740 + }, + { + "epoch": 0.11258355810953442, + "grad_norm": 0.7521204948425293, + "learning_rate": 2.669332001331669e-05, + "loss": 2.2033, + "num_input_tokens_seen": 393216000, + "step": 750 + }, + { + "epoch": 0.11408467221766154, + "grad_norm": 0.5531004071235657, + "learning_rate": 2.6517124556347397e-05, + "loss": 2.2177, + "num_input_tokens_seen": 398458880, + "step": 760 + }, + { + "epoch": 0.11558578632578867, + "grad_norm": 0.5678554773330688, + "learning_rate": 2.6344372720297863e-05, + "loss": 2.2527, + "num_input_tokens_seen": 403701760, + "step": 770 + }, + { + "epoch": 0.1170869004339158, + "grad_norm": 0.5761403441429138, + "learning_rate": 2.6174953775092044e-05, + "loss": 2.1891, + "num_input_tokens_seen": 408944640, + "step": 780 + }, + { + "epoch": 0.11858801454204292, + "grad_norm": 0.614530622959137, + "learning_rate": 2.6008761912142832e-05, + "loss": 2.2389, + "num_input_tokens_seen": 414187520, + "step": 790 + }, + { + "epoch": 0.12008912865017005, + "grad_norm": 0.8077418208122253, + "learning_rate": 2.584569596664017e-05, + "loss": 2.2173, + "num_input_tokens_seen": 419430400, + "step": 800 + }, + { + "epoch": 0.12008912865017005, + "eval_accuracy": 0.5564314204314205, + "eval_loss": 2.1939218044281006, + "eval_runtime": 94.2566, + "eval_samples_per_second": 3.183, + "eval_steps_per_second": 0.796, + "num_input_tokens_seen": 419430400, + "step": 800 + }, + { + "epoch": 0.12159024275829718, + "grad_norm": 0.5838737487792969, + "learning_rate": 2.5685659158755362e-05, + "loss": 2.2407, + "num_input_tokens_seen": 424673280, + "step": 810 + }, + { + "epoch": 0.12309135686642429, + "grad_norm": 0.5616093873977661, + "learning_rate": 2.5528558852257423e-05, + "loss": 2.2387, + "num_input_tokens_seen": 429916160, + "step": 820 + }, + { + "epoch": 0.12459247097455142, + "grad_norm": 0.5916436910629272, + "learning_rate": 2.537430632917375e-05, + "loss": 2.1901, + "num_input_tokens_seen": 435159040, + "step": 830 + }, + { + "epoch": 0.12609358508267854, + "grad_norm": 0.7512922286987305, + "learning_rate": 2.5222816579249755e-05, + "loss": 2.2354, + "num_input_tokens_seen": 440401920, + "step": 840 + }, + { + "epoch": 0.12759469919080568, + "grad_norm": 0.6344349384307861, + "learning_rate": 2.507400810307242e-05, + "loss": 2.18, + "num_input_tokens_seen": 445644800, + "step": 850 + }, + { + "epoch": 0.1290958132989328, + "grad_norm": 0.6193192005157471, + "learning_rate": 2.4927802727821985e-05, + "loss": 2.2427, + "num_input_tokens_seen": 450887680, + "step": 860 + }, + { + "epoch": 0.13059692740705992, + "grad_norm": 0.5937076807022095, + "learning_rate": 2.4784125434705546e-05, + "loss": 2.2096, + "num_input_tokens_seen": 456130560, + "step": 870 + }, + { + "epoch": 0.13209804151518706, + "grad_norm": 0.6233764886856079, + "learning_rate": 2.4642904197207104e-05, + "loss": 2.2124, + "num_input_tokens_seen": 461373440, + "step": 880 + }, + { + "epoch": 0.13359915562331418, + "grad_norm": 0.7363754510879517, + "learning_rate": 2.4504069829361966e-05, + "loss": 2.1925, + "num_input_tokens_seen": 466616320, + "step": 890 + }, + { + "epoch": 0.1351002697314413, + "grad_norm": 0.5889034867286682, + "learning_rate": 2.4367555843329425e-05, + "loss": 2.2139, + "num_input_tokens_seen": 471859200, + "step": 900 + }, + { + "epoch": 0.13660138383956844, + "grad_norm": 0.5365704894065857, + "learning_rate": 2.4233298315597637e-05, + "loss": 2.2096, + "num_input_tokens_seen": 477102080, + "step": 910 + }, + { + "epoch": 0.13810249794769555, + "grad_norm": 0.5648247599601746, + "learning_rate": 2.4101235761209246e-05, + "loss": 2.1803, + "num_input_tokens_seen": 482344960, + "step": 920 + }, + { + "epoch": 0.13960361205582267, + "grad_norm": 0.6017926931381226, + "learning_rate": 2.3971309015445497e-05, + "loss": 2.1787, + "num_input_tokens_seen": 487587840, + "step": 930 + }, + { + "epoch": 0.1411047261639498, + "grad_norm": 0.6408767700195312, + "learning_rate": 2.384346112245184e-05, + "loss": 2.1998, + "num_input_tokens_seen": 492830720, + "step": 940 + }, + { + "epoch": 0.14260584027207693, + "grad_norm": 0.635210394859314, + "learning_rate": 2.3717637230328695e-05, + "loss": 2.2327, + "num_input_tokens_seen": 498073600, + "step": 950 + }, + { + "epoch": 0.14410695438020404, + "grad_norm": 0.6507686972618103, + "learning_rate": 2.3593784492248522e-05, + "loss": 2.2162, + "num_input_tokens_seen": 503316480, + "step": 960 + }, + { + "epoch": 0.1456080684883312, + "grad_norm": 0.6332815885543823, + "learning_rate": 2.3471851973194328e-05, + "loss": 2.2287, + "num_input_tokens_seen": 508559360, + "step": 970 + }, + { + "epoch": 0.1471091825964583, + "grad_norm": 0.6600320339202881, + "learning_rate": 2.3351790561945772e-05, + "loss": 2.1837, + "num_input_tokens_seen": 513802240, + "step": 980 + }, + { + "epoch": 0.14861029670458545, + "grad_norm": 0.5689282417297363, + "learning_rate": 2.3233552887967433e-05, + "loss": 2.1926, + "num_input_tokens_seen": 519045120, + "step": 990 + }, + { + "epoch": 0.15011141081271256, + "grad_norm": 0.762301504611969, + "learning_rate": 2.3117093242879823e-05, + "loss": 2.2092, + "num_input_tokens_seen": 524288000, + "step": 1000 + }, + { + "epoch": 0.15161252492083968, + "grad_norm": 0.5970498919487, + "learning_rate": 2.3002367506217465e-05, + "loss": 2.2208, + "num_input_tokens_seen": 529530880, + "step": 1010 + }, + { + "epoch": 0.15311363902896682, + "grad_norm": 0.7378530502319336, + "learning_rate": 2.288933307520014e-05, + "loss": 2.193, + "num_input_tokens_seen": 534773760, + "step": 1020 + }, + { + "epoch": 0.15461475313709394, + "grad_norm": 0.6843045949935913, + "learning_rate": 2.2777948798263435e-05, + "loss": 2.201, + "num_input_tokens_seen": 540016640, + "step": 1030 + }, + { + "epoch": 0.15611586724522106, + "grad_norm": 0.6419305205345154, + "learning_rate": 2.2668174912113104e-05, + "loss": 2.2008, + "num_input_tokens_seen": 545259520, + "step": 1040 + }, + { + "epoch": 0.1576169813533482, + "grad_norm": 0.6570454239845276, + "learning_rate": 2.2559972982084464e-05, + "loss": 2.1878, + "num_input_tokens_seen": 550502400, + "step": 1050 + }, + { + "epoch": 0.15911809546147532, + "grad_norm": 0.6204289197921753, + "learning_rate": 2.2453305845603817e-05, + "loss": 2.1965, + "num_input_tokens_seen": 555745280, + "step": 1060 + }, + { + "epoch": 0.16061920956960243, + "grad_norm": 0.7957982420921326, + "learning_rate": 2.234813755856288e-05, + "loss": 2.1886, + "num_input_tokens_seen": 560988160, + "step": 1070 + }, + { + "epoch": 0.16212032367772958, + "grad_norm": 0.6827828288078308, + "learning_rate": 2.2244433344430573e-05, + "loss": 2.1646, + "num_input_tokens_seen": 566231040, + "step": 1080 + }, + { + "epoch": 0.1636214377858567, + "grad_norm": 0.8159109354019165, + "learning_rate": 2.2142159545938563e-05, + "loss": 2.1966, + "num_input_tokens_seen": 571473920, + "step": 1090 + }, + { + "epoch": 0.1651225518939838, + "grad_norm": 0.6716334223747253, + "learning_rate": 2.2041283579187985e-05, + "loss": 2.173, + "num_input_tokens_seen": 576716800, + "step": 1100 + }, + { + "epoch": 0.16662366600211095, + "grad_norm": 0.5386361479759216, + "learning_rate": 2.1941773890035455e-05, + "loss": 2.1958, + "num_input_tokens_seen": 581959680, + "step": 1110 + }, + { + "epoch": 0.16812478011023807, + "grad_norm": 0.5818248391151428, + "learning_rate": 2.1843599912625603e-05, + "loss": 2.2007, + "num_input_tokens_seen": 587202560, + "step": 1120 + }, + { + "epoch": 0.1696258942183652, + "grad_norm": 0.6868605613708496, + "learning_rate": 2.17467320299465e-05, + "loss": 2.2067, + "num_input_tokens_seen": 592445440, + "step": 1130 + }, + { + "epoch": 0.17112700832649233, + "grad_norm": 0.6540196537971497, + "learning_rate": 2.1651141536292296e-05, + "loss": 2.17, + "num_input_tokens_seen": 597688320, + "step": 1140 + }, + { + "epoch": 0.17262812243461945, + "grad_norm": 0.5799123644828796, + "learning_rate": 2.1556800601525106e-05, + "loss": 2.1932, + "num_input_tokens_seen": 602931200, + "step": 1150 + }, + { + "epoch": 0.17412923654274656, + "grad_norm": 0.7334414124488831, + "learning_rate": 2.1463682237035052e-05, + "loss": 2.1795, + "num_input_tokens_seen": 608174080, + "step": 1160 + }, + { + "epoch": 0.1756303506508737, + "grad_norm": 0.5947369337081909, + "learning_rate": 2.137176026330393e-05, + "loss": 2.1544, + "num_input_tokens_seen": 613416960, + "step": 1170 + }, + { + "epoch": 0.17713146475900082, + "grad_norm": 0.6859150528907776, + "learning_rate": 2.128100927898401e-05, + "loss": 2.1821, + "num_input_tokens_seen": 618659840, + "step": 1180 + }, + { + "epoch": 0.17863257886712794, + "grad_norm": 0.6155475378036499, + "learning_rate": 2.1191404631408953e-05, + "loss": 2.1643, + "num_input_tokens_seen": 623902720, + "step": 1190 + }, + { + "epoch": 0.18013369297525508, + "grad_norm": 0.5512512922286987, + "learning_rate": 2.110292238845922e-05, + "loss": 2.1992, + "num_input_tokens_seen": 629145600, + "step": 1200 + }, + { + "epoch": 0.18013369297525508, + "eval_accuracy": 0.5604013024013024, + "eval_loss": 2.1688950061798096, + "eval_runtime": 95.4401, + "eval_samples_per_second": 3.143, + "eval_steps_per_second": 0.786, + "num_input_tokens_seen": 629145600, + "step": 1200 + }, + { + "epoch": 0.1816348070833822, + "grad_norm": 0.6315280795097351, + "learning_rate": 2.1015539311708933e-05, + "loss": 2.1888, + "num_input_tokens_seen": 634388480, + "step": 1210 + }, + { + "epoch": 0.18313592119150932, + "grad_norm": 0.6135600209236145, + "learning_rate": 2.092923283078589e-05, + "loss": 2.2094, + "num_input_tokens_seen": 639631360, + "step": 1220 + }, + { + "epoch": 0.18463703529963646, + "grad_norm": 0.7772483229637146, + "learning_rate": 2.084398101888042e-05, + "loss": 2.1959, + "num_input_tokens_seen": 644874240, + "step": 1230 + }, + { + "epoch": 0.18613814940776358, + "grad_norm": 0.7412639856338501, + "learning_rate": 2.0759762569342743e-05, + "loss": 2.1783, + "num_input_tokens_seen": 650117120, + "step": 1240 + }, + { + "epoch": 0.1876392635158907, + "grad_norm": 0.8701518177986145, + "learning_rate": 2.0676556773312137e-05, + "loss": 2.1231, + "num_input_tokens_seen": 655360000, + "step": 1250 + }, + { + "epoch": 0.18914037762401784, + "grad_norm": 0.7169715762138367, + "learning_rate": 2.059434349832459e-05, + "loss": 2.1964, + "num_input_tokens_seen": 660602880, + "step": 1260 + }, + { + "epoch": 0.19064149173214495, + "grad_norm": 0.5910638570785522, + "learning_rate": 2.0513103167848672e-05, + "loss": 2.1915, + "num_input_tokens_seen": 665845760, + "step": 1270 + }, + { + "epoch": 0.19214260584027207, + "grad_norm": 0.5827318429946899, + "learning_rate": 2.0432816741702553e-05, + "loss": 2.2007, + "num_input_tokens_seen": 671088640, + "step": 1280 + }, + { + "epoch": 0.1936437199483992, + "grad_norm": 0.6388292908668518, + "learning_rate": 2.0353465697307495e-05, + "loss": 2.1615, + "num_input_tokens_seen": 676331520, + "step": 1290 + }, + { + "epoch": 0.19514483405652633, + "grad_norm": 0.6320261359214783, + "learning_rate": 2.0275032011736088e-05, + "loss": 2.1983, + "num_input_tokens_seen": 681574400, + "step": 1300 + }, + { + "epoch": 0.19664594816465344, + "grad_norm": 0.6461578011512756, + "learning_rate": 2.019749814451557e-05, + "loss": 2.177, + "num_input_tokens_seen": 686817280, + "step": 1310 + }, + { + "epoch": 0.1981470622727806, + "grad_norm": 0.7476032376289368, + "learning_rate": 2.0120847021149106e-05, + "loss": 2.1549, + "num_input_tokens_seen": 692060160, + "step": 1320 + }, + { + "epoch": 0.1996481763809077, + "grad_norm": 0.6539933681488037, + "learning_rate": 2.004506201731977e-05, + "loss": 2.1761, + "num_input_tokens_seen": 697303040, + "step": 1330 + }, + { + "epoch": 0.20114929048903482, + "grad_norm": 0.5645061135292053, + "learning_rate": 1.9970126943744094e-05, + "loss": 2.1721, + "num_input_tokens_seen": 702545920, + "step": 1340 + }, + { + "epoch": 0.20265040459716196, + "grad_norm": 0.6610363125801086, + "learning_rate": 1.9896026031643902e-05, + "loss": 2.1748, + "num_input_tokens_seen": 707788800, + "step": 1350 + }, + { + "epoch": 0.20415151870528908, + "grad_norm": 0.6691358089447021, + "learning_rate": 1.9822743918806708e-05, + "loss": 2.1685, + "num_input_tokens_seen": 713031680, + "step": 1360 + }, + { + "epoch": 0.2056526328134162, + "grad_norm": 0.6636555790901184, + "learning_rate": 1.9750265636206772e-05, + "loss": 2.1817, + "num_input_tokens_seen": 718274560, + "step": 1370 + }, + { + "epoch": 0.20715374692154334, + "grad_norm": 0.7346318960189819, + "learning_rate": 1.9678576595160388e-05, + "loss": 2.1833, + "num_input_tokens_seen": 723517440, + "step": 1380 + }, + { + "epoch": 0.20865486102967046, + "grad_norm": 0.5943858623504639, + "learning_rate": 1.960766257499033e-05, + "loss": 2.2494, + "num_input_tokens_seen": 728760320, + "step": 1390 + }, + { + "epoch": 0.21015597513779757, + "grad_norm": 0.7019902467727661, + "learning_rate": 1.9537509711175725e-05, + "loss": 2.2019, + "num_input_tokens_seen": 734003200, + "step": 1400 + }, + { + "epoch": 0.21165708924592472, + "grad_norm": 0.5532287955284119, + "learning_rate": 1.946810448396509e-05, + "loss": 2.1975, + "num_input_tokens_seen": 739246080, + "step": 1410 + }, + { + "epoch": 0.21315820335405183, + "grad_norm": 0.5484923720359802, + "learning_rate": 1.939943370743111e-05, + "loss": 2.1773, + "num_input_tokens_seen": 744488960, + "step": 1420 + }, + { + "epoch": 0.21465931746217895, + "grad_norm": 0.6313416957855225, + "learning_rate": 1.9331484518947162e-05, + "loss": 2.1919, + "num_input_tokens_seen": 749731840, + "step": 1430 + }, + { + "epoch": 0.2161604315703061, + "grad_norm": 0.5967773795127869, + "learning_rate": 1.926424436906652e-05, + "loss": 2.1957, + "num_input_tokens_seen": 754974720, + "step": 1440 + }, + { + "epoch": 0.2176615456784332, + "grad_norm": 0.5972282290458679, + "learning_rate": 1.919770101178605e-05, + "loss": 2.1647, + "num_input_tokens_seen": 760217600, + "step": 1450 + }, + { + "epoch": 0.21916265978656033, + "grad_norm": 0.5539584755897522, + "learning_rate": 1.9131842495177355e-05, + "loss": 2.1815, + "num_input_tokens_seen": 765460480, + "step": 1460 + }, + { + "epoch": 0.22066377389468747, + "grad_norm": 0.7532745003700256, + "learning_rate": 1.9066657152369064e-05, + "loss": 2.2143, + "num_input_tokens_seen": 770703360, + "step": 1470 + }, + { + "epoch": 0.2221648880028146, + "grad_norm": 0.537327229976654, + "learning_rate": 1.9002133592864807e-05, + "loss": 2.1541, + "num_input_tokens_seen": 775946240, + "step": 1480 + }, + { + "epoch": 0.2236660021109417, + "grad_norm": 0.7467483878135681, + "learning_rate": 1.893826069418217e-05, + "loss": 2.1638, + "num_input_tokens_seen": 781189120, + "step": 1490 + }, + { + "epoch": 0.22516711621906885, + "grad_norm": 0.5528019070625305, + "learning_rate": 1.8875027593798816e-05, + "loss": 2.1865, + "num_input_tokens_seen": 786432000, + "step": 1500 + }, + { + "epoch": 0.22666823032719596, + "grad_norm": 0.6948214173316956, + "learning_rate": 1.8812423681392357e-05, + "loss": 2.166, + "num_input_tokens_seen": 791674880, + "step": 1510 + }, + { + "epoch": 0.22816934443532308, + "grad_norm": 0.6632937788963318, + "learning_rate": 1.8750438591361568e-05, + "loss": 2.1864, + "num_input_tokens_seen": 796917760, + "step": 1520 + }, + { + "epoch": 0.22967045854345022, + "grad_norm": 0.5619977116584778, + "learning_rate": 1.8689062195616828e-05, + "loss": 2.192, + "num_input_tokens_seen": 802160640, + "step": 1530 + }, + { + "epoch": 0.23117157265157734, + "grad_norm": 0.6643936038017273, + "learning_rate": 1.8628284596628512e-05, + "loss": 2.1749, + "num_input_tokens_seen": 807403520, + "step": 1540 + }, + { + "epoch": 0.23267268675970446, + "grad_norm": 0.6980964541435242, + "learning_rate": 1.8568096120722424e-05, + "loss": 2.1662, + "num_input_tokens_seen": 812646400, + "step": 1550 + }, + { + "epoch": 0.2341738008678316, + "grad_norm": 0.8099634647369385, + "learning_rate": 1.8508487311612005e-05, + "loss": 2.1778, + "num_input_tokens_seen": 817889280, + "step": 1560 + }, + { + "epoch": 0.23567491497595872, + "grad_norm": 0.5385667085647583, + "learning_rate": 1.844944892415748e-05, + "loss": 2.1536, + "num_input_tokens_seen": 823132160, + "step": 1570 + }, + { + "epoch": 0.23717602908408583, + "grad_norm": 0.70493483543396, + "learning_rate": 1.8390971918342592e-05, + "loss": 2.192, + "num_input_tokens_seen": 828375040, + "step": 1580 + }, + { + "epoch": 0.23867714319221298, + "grad_norm": 0.5577861666679382, + "learning_rate": 1.833304745346004e-05, + "loss": 2.1941, + "num_input_tokens_seen": 833617920, + "step": 1590 + }, + { + "epoch": 0.2401782573003401, + "grad_norm": 0.5603963136672974, + "learning_rate": 1.8275666882497067e-05, + "loss": 2.1543, + "num_input_tokens_seen": 838860800, + "step": 1600 + }, + { + "epoch": 0.2401782573003401, + "eval_accuracy": 0.5632421652421652, + "eval_loss": 2.1520919799804688, + "eval_runtime": 94.9668, + "eval_samples_per_second": 3.159, + "eval_steps_per_second": 0.79, + "num_input_tokens_seen": 838860800, + "step": 1600 + }, + { + "epoch": 0.2416793714084672, + "grad_norm": 0.6568350791931152, + "learning_rate": 1.821882174671319e-05, + "loss": 2.1729, + "num_input_tokens_seen": 844103680, + "step": 1610 + }, + { + "epoch": 0.24318048551659435, + "grad_norm": 0.6658337712287903, + "learning_rate": 1.8162503770402265e-05, + "loss": 2.1954, + "num_input_tokens_seen": 849346560, + "step": 1620 + }, + { + "epoch": 0.24468159962472147, + "grad_norm": 0.6133268475532532, + "learning_rate": 1.8106704855831622e-05, + "loss": 2.1887, + "num_input_tokens_seen": 854589440, + "step": 1630 + }, + { + "epoch": 0.24618271373284858, + "grad_norm": 0.6844907402992249, + "learning_rate": 1.805141707835109e-05, + "loss": 2.1772, + "num_input_tokens_seen": 859832320, + "step": 1640 + }, + { + "epoch": 0.24768382784097573, + "grad_norm": 0.6765184998512268, + "learning_rate": 1.799663268166542e-05, + "loss": 2.1631, + "num_input_tokens_seen": 865075200, + "step": 1650 + }, + { + "epoch": 0.24918494194910285, + "grad_norm": 0.7173587083816528, + "learning_rate": 1.7942344073263488e-05, + "loss": 2.1347, + "num_input_tokens_seen": 870318080, + "step": 1660 + }, + { + "epoch": 0.25068605605723, + "grad_norm": 0.907135546207428, + "learning_rate": 1.788854381999832e-05, + "loss": 2.1809, + "num_input_tokens_seen": 875560960, + "step": 1670 + }, + { + "epoch": 0.2521871701653571, + "grad_norm": 0.6652698516845703, + "learning_rate": 1.783522464381198e-05, + "loss": 2.1558, + "num_input_tokens_seen": 880803840, + "step": 1680 + }, + { + "epoch": 0.2536882842734842, + "grad_norm": 0.5732602477073669, + "learning_rate": 1.7782379417599864e-05, + "loss": 2.1788, + "num_input_tokens_seen": 886046720, + "step": 1690 + }, + { + "epoch": 0.25518939838161137, + "grad_norm": 0.5569264888763428, + "learning_rate": 1.7730001161208948e-05, + "loss": 2.1375, + "num_input_tokens_seen": 891289600, + "step": 1700 + }, + { + "epoch": 0.25669051248973845, + "grad_norm": 0.6948941946029663, + "learning_rate": 1.767808303756493e-05, + "loss": 2.148, + "num_input_tokens_seen": 896532480, + "step": 1710 + }, + { + "epoch": 0.2581916265978656, + "grad_norm": 0.5902568101882935, + "learning_rate": 1.7626618348923443e-05, + "loss": 2.1651, + "num_input_tokens_seen": 901775360, + "step": 1720 + }, + { + "epoch": 0.25969274070599274, + "grad_norm": 0.6689503192901611, + "learning_rate": 1.757560053324057e-05, + "loss": 2.1761, + "num_input_tokens_seen": 907018240, + "step": 1730 + }, + { + "epoch": 0.26119385481411983, + "grad_norm": 0.6513310074806213, + "learning_rate": 1.7525023160658283e-05, + "loss": 2.195, + "num_input_tokens_seen": 912261120, + "step": 1740 + }, + { + "epoch": 0.262694968922247, + "grad_norm": 0.5774445533752441, + "learning_rate": 1.7474879930100483e-05, + "loss": 2.1681, + "num_input_tokens_seen": 917504000, + "step": 1750 + }, + { + "epoch": 0.2641960830303741, + "grad_norm": 0.583354115486145, + "learning_rate": 1.7425164665975574e-05, + "loss": 2.1619, + "num_input_tokens_seen": 922746880, + "step": 1760 + }, + { + "epoch": 0.2656971971385012, + "grad_norm": 0.6275432109832764, + "learning_rate": 1.7375871314981654e-05, + "loss": 2.1819, + "num_input_tokens_seen": 927989760, + "step": 1770 + }, + { + "epoch": 0.26719831124662835, + "grad_norm": 0.6457045674324036, + "learning_rate": 1.732699394301053e-05, + "loss": 2.135, + "num_input_tokens_seen": 933232640, + "step": 1780 + }, + { + "epoch": 0.2686994253547555, + "grad_norm": 0.6066982746124268, + "learning_rate": 1.7278526732147035e-05, + "loss": 2.1782, + "num_input_tokens_seen": 938475520, + "step": 1790 + }, + { + "epoch": 0.2702005394628826, + "grad_norm": 0.6707212328910828, + "learning_rate": 1.7230463977760115e-05, + "loss": 2.2126, + "num_input_tokens_seen": 943718400, + "step": 1800 + }, + { + "epoch": 0.2717016535710097, + "grad_norm": 0.6359196305274963, + "learning_rate": 1.71828000856825e-05, + "loss": 2.1631, + "num_input_tokens_seen": 948961280, + "step": 1810 + }, + { + "epoch": 0.27320276767913687, + "grad_norm": 0.5916318297386169, + "learning_rate": 1.713552956947563e-05, + "loss": 2.1535, + "num_input_tokens_seen": 954204160, + "step": 1820 + }, + { + "epoch": 0.27470388178726396, + "grad_norm": 0.7036842107772827, + "learning_rate": 1.708864704777699e-05, + "loss": 2.153, + "num_input_tokens_seen": 959447040, + "step": 1830 + }, + { + "epoch": 0.2762049958953911, + "grad_norm": 0.6982750296592712, + "learning_rate": 1.704214724172678e-05, + "loss": 2.1806, + "num_input_tokens_seen": 964689920, + "step": 1840 + }, + { + "epoch": 0.27770611000351825, + "grad_norm": 0.6548587083816528, + "learning_rate": 1.699602497247121e-05, + "loss": 2.1547, + "num_input_tokens_seen": 969932800, + "step": 1850 + }, + { + "epoch": 0.27920722411164534, + "grad_norm": 0.6462974548339844, + "learning_rate": 1.6950275158739732e-05, + "loss": 2.1697, + "num_input_tokens_seen": 975175680, + "step": 1860 + }, + { + "epoch": 0.2807083382197725, + "grad_norm": 0.5359413623809814, + "learning_rate": 1.690489281449364e-05, + "loss": 2.158, + "num_input_tokens_seen": 980418560, + "step": 1870 + }, + { + "epoch": 0.2822094523278996, + "grad_norm": 0.6209542155265808, + "learning_rate": 1.6859873046643506e-05, + "loss": 2.1345, + "num_input_tokens_seen": 985661440, + "step": 1880 + }, + { + "epoch": 0.2837105664360267, + "grad_norm": 0.602796733379364, + "learning_rate": 1.681521105283317e-05, + "loss": 2.1774, + "num_input_tokens_seen": 990904320, + "step": 1890 + }, + { + "epoch": 0.28521168054415386, + "grad_norm": 0.6523065567016602, + "learning_rate": 1.6770902119287942e-05, + "loss": 2.2011, + "num_input_tokens_seen": 996147200, + "step": 1900 + }, + { + "epoch": 0.286712794652281, + "grad_norm": 0.5961717963218689, + "learning_rate": 1.672694161872488e-05, + "loss": 2.1722, + "num_input_tokens_seen": 1001390080, + "step": 1910 + }, + { + "epoch": 0.2882139087604081, + "grad_norm": 0.5899094343185425, + "learning_rate": 1.6683325008322934e-05, + "loss": 2.1353, + "num_input_tokens_seen": 1006632960, + "step": 1920 + }, + { + "epoch": 0.28971502286853523, + "grad_norm": 0.590152382850647, + "learning_rate": 1.6640047827751115e-05, + "loss": 2.186, + "num_input_tokens_seen": 1011875840, + "step": 1930 + }, + { + "epoch": 0.2912161369766624, + "grad_norm": 0.6230789422988892, + "learning_rate": 1.6597105697252553e-05, + "loss": 2.1719, + "num_input_tokens_seen": 1017118720, + "step": 1940 + }, + { + "epoch": 0.29271725108478946, + "grad_norm": 0.6433463096618652, + "learning_rate": 1.655449431578271e-05, + "loss": 2.1711, + "num_input_tokens_seen": 1022361600, + "step": 1950 + }, + { + "epoch": 0.2942183651929166, + "grad_norm": 0.6224901080131531, + "learning_rate": 1.6512209459199873e-05, + "loss": 2.1473, + "num_input_tokens_seen": 1027604480, + "step": 1960 + }, + { + "epoch": 0.29571947930104375, + "grad_norm": 0.6015456914901733, + "learning_rate": 1.6470246978506274e-05, + "loss": 2.1788, + "num_input_tokens_seen": 1032847360, + "step": 1970 + }, + { + "epoch": 0.2972205934091709, + "grad_norm": 0.7380375266075134, + "learning_rate": 1.6428602798138068e-05, + "loss": 2.1566, + "num_input_tokens_seen": 1038090240, + "step": 1980 + }, + { + "epoch": 0.298721707517298, + "grad_norm": 0.5713395476341248, + "learning_rate": 1.638727291430271e-05, + "loss": 2.1596, + "num_input_tokens_seen": 1043333120, + "step": 1990 + }, + { + "epoch": 0.30022282162542513, + "grad_norm": 0.6106094717979431, + "learning_rate": 1.634625339336204e-05, + "loss": 2.1532, + "num_input_tokens_seen": 1048576000, + "step": 2000 + }, + { + "epoch": 0.30022282162542513, + "eval_accuracy": 0.565023199023199, + "eval_loss": 2.1400792598724365, + "eval_runtime": 95.5183, + "eval_samples_per_second": 3.141, + "eval_steps_per_second": 0.785, + "num_input_tokens_seen": 1048576000, + "step": 2000 + }, + { + "epoch": 0.3017239357335523, + "grad_norm": 0.6882668733596802, + "learning_rate": 1.6305540370259715e-05, + "loss": 2.1681, + "num_input_tokens_seen": 1053818880, + "step": 2010 + }, + { + "epoch": 0.30322504984167936, + "grad_norm": 0.5827280282974243, + "learning_rate": 1.6265130046991463e-05, + "loss": 2.1699, + "num_input_tokens_seen": 1059061760, + "step": 2020 + }, + { + "epoch": 0.3047261639498065, + "grad_norm": 0.5182669162750244, + "learning_rate": 1.6225018691116927e-05, + "loss": 2.1495, + "num_input_tokens_seen": 1064304640, + "step": 2030 + }, + { + "epoch": 0.30622727805793365, + "grad_norm": 0.8182618618011475, + "learning_rate": 1.618520263431155e-05, + "loss": 2.1837, + "num_input_tokens_seen": 1069547520, + "step": 2040 + }, + { + "epoch": 0.30772839216606074, + "grad_norm": 0.5096232891082764, + "learning_rate": 1.6145678270957475e-05, + "loss": 2.1489, + "num_input_tokens_seen": 1074790400, + "step": 2050 + }, + { + "epoch": 0.3092295062741879, + "grad_norm": 0.6253472566604614, + "learning_rate": 1.6106442056772048e-05, + "loss": 2.1699, + "num_input_tokens_seen": 1080033280, + "step": 2060 + }, + { + "epoch": 0.310730620382315, + "grad_norm": 0.5389710068702698, + "learning_rate": 1.6067490507472832e-05, + "loss": 2.1715, + "num_input_tokens_seen": 1085276160, + "step": 2070 + }, + { + "epoch": 0.3122317344904421, + "grad_norm": 0.6460773348808289, + "learning_rate": 1.6028820197477945e-05, + "loss": 2.1601, + "num_input_tokens_seen": 1090519040, + "step": 2080 + }, + { + "epoch": 0.31373284859856926, + "grad_norm": 0.7538013458251953, + "learning_rate": 1.5990427758640674e-05, + "loss": 2.1677, + "num_input_tokens_seen": 1095761920, + "step": 2090 + }, + { + "epoch": 0.3152339627066964, + "grad_norm": 0.6462552547454834, + "learning_rate": 1.5952309879017224e-05, + "loss": 2.1805, + "num_input_tokens_seen": 1101004800, + "step": 2100 + }, + { + "epoch": 0.3167350768148235, + "grad_norm": 0.6027917861938477, + "learning_rate": 1.5914463301666658e-05, + "loss": 2.1415, + "num_input_tokens_seen": 1106247680, + "step": 2110 + }, + { + "epoch": 0.31823619092295063, + "grad_norm": 0.6895632147789001, + "learning_rate": 1.5876884823482006e-05, + "loss": 2.1587, + "num_input_tokens_seen": 1111490560, + "step": 2120 + }, + { + "epoch": 0.3197373050310778, + "grad_norm": 0.6652107834815979, + "learning_rate": 1.5839571294051578e-05, + "loss": 2.1444, + "num_input_tokens_seen": 1116733440, + "step": 2130 + }, + { + "epoch": 0.32123841913920487, + "grad_norm": 0.7466081380844116, + "learning_rate": 1.580251961454958e-05, + "loss": 2.1251, + "num_input_tokens_seen": 1121976320, + "step": 2140 + }, + { + "epoch": 0.322739533247332, + "grad_norm": 0.5937127470970154, + "learning_rate": 1.5765726736655174e-05, + "loss": 2.1489, + "num_input_tokens_seen": 1127219200, + "step": 2150 + }, + { + "epoch": 0.32424064735545916, + "grad_norm": 0.6133046746253967, + "learning_rate": 1.5729189661499015e-05, + "loss": 2.1507, + "num_input_tokens_seen": 1132462080, + "step": 2160 + }, + { + "epoch": 0.32574176146358624, + "grad_norm": 0.7476202845573425, + "learning_rate": 1.5692905438636622e-05, + "loss": 2.1691, + "num_input_tokens_seen": 1137704960, + "step": 2170 + }, + { + "epoch": 0.3272428755717134, + "grad_norm": 0.7822461724281311, + "learning_rate": 1.5656871165047605e-05, + "loss": 2.1522, + "num_input_tokens_seen": 1142947840, + "step": 2180 + }, + { + "epoch": 0.32874398967984053, + "grad_norm": 0.6776607632637024, + "learning_rate": 1.5621083984160086e-05, + "loss": 2.142, + "num_input_tokens_seen": 1148190720, + "step": 2190 + }, + { + "epoch": 0.3302451037879676, + "grad_norm": 0.7508281469345093, + "learning_rate": 1.558554108489952e-05, + "loss": 2.1723, + "num_input_tokens_seen": 1153433600, + "step": 2200 + }, + { + "epoch": 0.33174621789609476, + "grad_norm": 0.7543405890464783, + "learning_rate": 1.5550239700761246e-05, + "loss": 2.1442, + "num_input_tokens_seen": 1158676480, + "step": 2210 + }, + { + "epoch": 0.3332473320042219, + "grad_norm": 0.9116769433021545, + "learning_rate": 1.5515177108906e-05, + "loss": 2.1916, + "num_input_tokens_seen": 1163919360, + "step": 2220 + }, + { + "epoch": 0.334748446112349, + "grad_norm": 0.7372170686721802, + "learning_rate": 1.5480350629277787e-05, + "loss": 2.1546, + "num_input_tokens_seen": 1169162240, + "step": 2230 + }, + { + "epoch": 0.33624956022047614, + "grad_norm": 0.5467498302459717, + "learning_rate": 1.5445757623743444e-05, + "loss": 2.1919, + "num_input_tokens_seen": 1174405120, + "step": 2240 + }, + { + "epoch": 0.3377506743286033, + "grad_norm": 0.5791755318641663, + "learning_rate": 1.5411395495253218e-05, + "loss": 2.1782, + "num_input_tokens_seen": 1179648000, + "step": 2250 + }, + { + "epoch": 0.3392517884367304, + "grad_norm": 0.7308974862098694, + "learning_rate": 1.5377261687021863e-05, + "loss": 2.1682, + "num_input_tokens_seen": 1184890880, + "step": 2260 + }, + { + "epoch": 0.3407529025448575, + "grad_norm": 0.6968861222267151, + "learning_rate": 1.5343353681729532e-05, + "loss": 2.1724, + "num_input_tokens_seen": 1190133760, + "step": 2270 + }, + { + "epoch": 0.34225401665298466, + "grad_norm": 0.6837549805641174, + "learning_rate": 1.5309669000742003e-05, + "loss": 2.1783, + "num_input_tokens_seen": 1195376640, + "step": 2280 + }, + { + "epoch": 0.34375513076111175, + "grad_norm": 0.6059392094612122, + "learning_rate": 1.5276205203349658e-05, + "loss": 2.1454, + "num_input_tokens_seen": 1200619520, + "step": 2290 + }, + { + "epoch": 0.3452562448692389, + "grad_norm": 0.5880239605903625, + "learning_rate": 1.5242959886024648e-05, + "loss": 2.1428, + "num_input_tokens_seen": 1205862400, + "step": 2300 + }, + { + "epoch": 0.34675735897736604, + "grad_norm": 0.6651424169540405, + "learning_rate": 1.5209930681695804e-05, + "loss": 2.1518, + "num_input_tokens_seen": 1211105280, + "step": 2310 + }, + { + "epoch": 0.3482584730854931, + "grad_norm": 0.5734732151031494, + "learning_rate": 1.5177115259040729e-05, + "loss": 2.1327, + "num_input_tokens_seen": 1216348160, + "step": 2320 + }, + { + "epoch": 0.34975958719362027, + "grad_norm": 0.6250523328781128, + "learning_rate": 1.514451132179463e-05, + "loss": 2.1459, + "num_input_tokens_seen": 1221591040, + "step": 2330 + }, + { + "epoch": 0.3512607013017474, + "grad_norm": 0.6869091391563416, + "learning_rate": 1.5112116608075403e-05, + "loss": 2.1352, + "num_input_tokens_seen": 1226833920, + "step": 2340 + }, + { + "epoch": 0.3527618154098745, + "grad_norm": 0.8655930161476135, + "learning_rate": 1.5079928889724547e-05, + "loss": 2.1581, + "num_input_tokens_seen": 1232076800, + "step": 2350 + }, + { + "epoch": 0.35426292951800165, + "grad_norm": 0.6587986946105957, + "learning_rate": 1.5047945971663434e-05, + "loss": 2.1783, + "num_input_tokens_seen": 1237319680, + "step": 2360 + }, + { + "epoch": 0.3557640436261288, + "grad_norm": 0.6270188689231873, + "learning_rate": 1.501616569126455e-05, + "loss": 2.1665, + "num_input_tokens_seen": 1242562560, + "step": 2370 + }, + { + "epoch": 0.3572651577342559, + "grad_norm": 0.6337898969650269, + "learning_rate": 1.4984585917737282e-05, + "loss": 2.177, + "num_input_tokens_seen": 1247805440, + "step": 2380 + }, + { + "epoch": 0.358766271842383, + "grad_norm": 0.6973663568496704, + "learning_rate": 1.4953204551527831e-05, + "loss": 2.1493, + "num_input_tokens_seen": 1253048320, + "step": 2390 + }, + { + "epoch": 0.36026738595051017, + "grad_norm": 0.5605215430259705, + "learning_rate": 1.4922019523732929e-05, + "loss": 2.1688, + "num_input_tokens_seen": 1258291200, + "step": 2400 + }, + { + "epoch": 0.36026738595051017, + "eval_accuracy": 0.5663190883190883, + "eval_loss": 2.1306827068328857, + "eval_runtime": 94.9906, + "eval_samples_per_second": 3.158, + "eval_steps_per_second": 0.79, + "num_input_tokens_seen": 1258291200, + "step": 2400 + }, + { + "epoch": 0.36176850005863725, + "grad_norm": 0.7090189456939697, + "learning_rate": 1.4891028795526912e-05, + "loss": 2.1446, + "num_input_tokens_seen": 1263534080, + "step": 2410 + }, + { + "epoch": 0.3632696141667644, + "grad_norm": 0.7129746079444885, + "learning_rate": 1.4860230357601855e-05, + "loss": 2.178, + "num_input_tokens_seen": 1268776960, + "step": 2420 + }, + { + "epoch": 0.36477072827489154, + "grad_norm": 0.5609286427497864, + "learning_rate": 1.4829622229620384e-05, + "loss": 2.1846, + "num_input_tokens_seen": 1274019840, + "step": 2430 + }, + { + "epoch": 0.36627184238301863, + "grad_norm": 0.6806447505950928, + "learning_rate": 1.4799202459680824e-05, + "loss": 2.156, + "num_input_tokens_seen": 1279262720, + "step": 2440 + }, + { + "epoch": 0.3677729564911458, + "grad_norm": 0.5787296295166016, + "learning_rate": 1.4768969123794386e-05, + "loss": 2.128, + "num_input_tokens_seen": 1284505600, + "step": 2450 + }, + { + "epoch": 0.3692740705992729, + "grad_norm": 0.7626485824584961, + "learning_rate": 1.4738920325374027e-05, + "loss": 2.1607, + "num_input_tokens_seen": 1289748480, + "step": 2460 + }, + { + "epoch": 0.3707751847074, + "grad_norm": 0.6390326619148254, + "learning_rate": 1.4709054194734743e-05, + "loss": 2.1821, + "num_input_tokens_seen": 1294991360, + "step": 2470 + }, + { + "epoch": 0.37227629881552715, + "grad_norm": 0.5435773730278015, + "learning_rate": 1.4679368888604919e-05, + "loss": 2.1762, + "num_input_tokens_seen": 1300234240, + "step": 2480 + }, + { + "epoch": 0.3737774129236543, + "grad_norm": 0.5910453200340271, + "learning_rate": 1.464986258964849e-05, + "loss": 2.1586, + "num_input_tokens_seen": 1305477120, + "step": 2490 + }, + { + "epoch": 0.3752785270317814, + "grad_norm": 0.5591862797737122, + "learning_rate": 1.4620533505997654e-05, + "loss": 2.1512, + "num_input_tokens_seen": 1310720000, + "step": 2500 + }, + { + "epoch": 0.3767796411399085, + "grad_norm": 0.6790704727172852, + "learning_rate": 1.459137987079579e-05, + "loss": 2.1696, + "num_input_tokens_seen": 1315962880, + "step": 2510 + }, + { + "epoch": 0.37828075524803567, + "grad_norm": 0.649533212184906, + "learning_rate": 1.4562399941750401e-05, + "loss": 2.1357, + "num_input_tokens_seen": 1321205760, + "step": 2520 + }, + { + "epoch": 0.37978186935616276, + "grad_norm": 0.5689984560012817, + "learning_rate": 1.453359200069576e-05, + "loss": 2.1972, + "num_input_tokens_seen": 1326448640, + "step": 2530 + }, + { + "epoch": 0.3812829834642899, + "grad_norm": 0.7337691187858582, + "learning_rate": 1.4504954353165044e-05, + "loss": 2.1198, + "num_input_tokens_seen": 1331691520, + "step": 2540 + }, + { + "epoch": 0.38278409757241705, + "grad_norm": 0.550014853477478, + "learning_rate": 1.447648532797172e-05, + "loss": 2.1267, + "num_input_tokens_seen": 1336934400, + "step": 2550 + }, + { + "epoch": 0.38428521168054414, + "grad_norm": 0.6532608270645142, + "learning_rate": 1.4448183276799891e-05, + "loss": 2.1735, + "num_input_tokens_seen": 1342177280, + "step": 2560 + }, + { + "epoch": 0.3857863257886713, + "grad_norm": 0.5801606774330139, + "learning_rate": 1.4420046573803481e-05, + "loss": 2.113, + "num_input_tokens_seen": 1347420160, + "step": 2570 + }, + { + "epoch": 0.3872874398967984, + "grad_norm": 0.6582128405570984, + "learning_rate": 1.4392073615213913e-05, + "loss": 2.1431, + "num_input_tokens_seen": 1352663040, + "step": 2580 + }, + { + "epoch": 0.3887885540049255, + "grad_norm": 0.6207413077354431, + "learning_rate": 1.436426281895616e-05, + "loss": 2.1683, + "num_input_tokens_seen": 1357905920, + "step": 2590 + }, + { + "epoch": 0.39028966811305266, + "grad_norm": 0.5679153800010681, + "learning_rate": 1.4336612624272917e-05, + "loss": 2.1194, + "num_input_tokens_seen": 1363148800, + "step": 2600 + }, + { + "epoch": 0.3917907822211798, + "grad_norm": 0.6869505643844604, + "learning_rate": 1.4309121491356698e-05, + "loss": 2.1033, + "num_input_tokens_seen": 1368391680, + "step": 2610 + }, + { + "epoch": 0.3932918963293069, + "grad_norm": 0.661869466304779, + "learning_rate": 1.4281787900989672e-05, + "loss": 2.161, + "num_input_tokens_seen": 1373634560, + "step": 2620 + }, + { + "epoch": 0.39479301043743403, + "grad_norm": 0.6266052722930908, + "learning_rate": 1.4254610354191023e-05, + "loss": 2.1838, + "num_input_tokens_seen": 1378877440, + "step": 2630 + }, + { + "epoch": 0.3962941245455612, + "grad_norm": 0.6248527765274048, + "learning_rate": 1.4227587371871679e-05, + "loss": 2.1255, + "num_input_tokens_seen": 1384120320, + "step": 2640 + }, + { + "epoch": 0.39779523865368827, + "grad_norm": 0.7318665385246277, + "learning_rate": 1.4200717494496206e-05, + "loss": 2.1487, + "num_input_tokens_seen": 1389363200, + "step": 2650 + }, + { + "epoch": 0.3992963527618154, + "grad_norm": 0.7174627780914307, + "learning_rate": 1.4173999281751702e-05, + "loss": 2.1375, + "num_input_tokens_seen": 1394606080, + "step": 2660 + }, + { + "epoch": 0.40079746686994255, + "grad_norm": 0.6472954154014587, + "learning_rate": 1.4147431312223518e-05, + "loss": 2.1546, + "num_input_tokens_seen": 1399848960, + "step": 2670 + }, + { + "epoch": 0.40229858097806964, + "grad_norm": 0.588234543800354, + "learning_rate": 1.4121012183077632e-05, + "loss": 2.1437, + "num_input_tokens_seen": 1405091840, + "step": 2680 + }, + { + "epoch": 0.4037996950861968, + "grad_norm": 0.5872112512588501, + "learning_rate": 1.4094740509749542e-05, + "loss": 2.1672, + "num_input_tokens_seen": 1410334720, + "step": 2690 + }, + { + "epoch": 0.40530080919432393, + "grad_norm": 0.6200748682022095, + "learning_rate": 1.406861492563948e-05, + "loss": 2.1434, + "num_input_tokens_seen": 1415577600, + "step": 2700 + }, + { + "epoch": 0.406801923302451, + "grad_norm": 0.7556484937667847, + "learning_rate": 1.4042634081813838e-05, + "loss": 2.1603, + "num_input_tokens_seen": 1420820480, + "step": 2710 + }, + { + "epoch": 0.40830303741057816, + "grad_norm": 0.6454352736473083, + "learning_rate": 1.4016796646712618e-05, + "loss": 2.1783, + "num_input_tokens_seen": 1426063360, + "step": 2720 + }, + { + "epoch": 0.4098041515187053, + "grad_norm": 0.5394317507743835, + "learning_rate": 1.3991101305862803e-05, + "loss": 2.1646, + "num_input_tokens_seen": 1431306240, + "step": 2730 + }, + { + "epoch": 0.4113052656268324, + "grad_norm": 0.6699331998825073, + "learning_rate": 1.396554676159745e-05, + "loss": 2.1597, + "num_input_tokens_seen": 1436549120, + "step": 2740 + }, + { + "epoch": 0.41280637973495954, + "grad_norm": 0.6491063237190247, + "learning_rate": 1.3940131732780461e-05, + "loss": 2.1236, + "num_input_tokens_seen": 1441792000, + "step": 2750 + }, + { + "epoch": 0.4143074938430867, + "grad_norm": 0.6439197063446045, + "learning_rate": 1.3914854954536792e-05, + "loss": 2.164, + "num_input_tokens_seen": 1447034880, + "step": 2760 + }, + { + "epoch": 0.41580860795121377, + "grad_norm": 0.7333915829658508, + "learning_rate": 1.3889715177988056e-05, + "loss": 2.1417, + "num_input_tokens_seen": 1452277760, + "step": 2770 + }, + { + "epoch": 0.4173097220593409, + "grad_norm": 0.5554391145706177, + "learning_rate": 1.386471116999334e-05, + "loss": 2.1821, + "num_input_tokens_seen": 1457520640, + "step": 2780 + }, + { + "epoch": 0.41881083616746806, + "grad_norm": 0.6607913374900818, + "learning_rate": 1.3839841712895161e-05, + "loss": 2.1341, + "num_input_tokens_seen": 1462763520, + "step": 2790 + }, + { + "epoch": 0.42031195027559515, + "grad_norm": 0.6506746411323547, + "learning_rate": 1.381510560427038e-05, + "loss": 2.1443, + "num_input_tokens_seen": 1468006400, + "step": 2800 + }, + { + "epoch": 0.42031195027559515, + "eval_accuracy": 0.5676076516076516, + "eval_loss": 2.1226789951324463, + "eval_runtime": 94.7272, + "eval_samples_per_second": 3.167, + "eval_steps_per_second": 0.792, + "num_input_tokens_seen": 1468006400, + "step": 2800 + }, + { + "epoch": 0.4218130643837223, + "grad_norm": 0.6999175548553467, + "learning_rate": 1.3790501656686045e-05, + "loss": 2.1222, + "num_input_tokens_seen": 1473249280, + "step": 2810 + }, + { + "epoch": 0.42331417849184944, + "grad_norm": 0.6637996435165405, + "learning_rate": 1.3766028697459948e-05, + "loss": 2.1752, + "num_input_tokens_seen": 1478492160, + "step": 2820 + }, + { + "epoch": 0.4248152925999765, + "grad_norm": 0.675205409526825, + "learning_rate": 1.3741685568425882e-05, + "loss": 2.1379, + "num_input_tokens_seen": 1483735040, + "step": 2830 + }, + { + "epoch": 0.42631640670810367, + "grad_norm": 0.7251035571098328, + "learning_rate": 1.371747112570342e-05, + "loss": 2.1499, + "num_input_tokens_seen": 1488977920, + "step": 2840 + }, + { + "epoch": 0.4278175208162308, + "grad_norm": 0.8489077687263489, + "learning_rate": 1.3693384239472163e-05, + "loss": 2.1291, + "num_input_tokens_seen": 1494220800, + "step": 2850 + }, + { + "epoch": 0.4293186349243579, + "grad_norm": 0.5617297887802124, + "learning_rate": 1.3669423793750302e-05, + "loss": 2.145, + "num_input_tokens_seen": 1499463680, + "step": 2860 + }, + { + "epoch": 0.43081974903248504, + "grad_norm": 0.5285384654998779, + "learning_rate": 1.3645588686177466e-05, + "loss": 2.1492, + "num_input_tokens_seen": 1504706560, + "step": 2870 + }, + { + "epoch": 0.4323208631406122, + "grad_norm": 0.6254436373710632, + "learning_rate": 1.3621877827801699e-05, + "loss": 2.1196, + "num_input_tokens_seen": 1509949440, + "step": 2880 + }, + { + "epoch": 0.4338219772487393, + "grad_norm": 0.5884484052658081, + "learning_rate": 1.3598290142870485e-05, + "loss": 2.1484, + "num_input_tokens_seen": 1515192320, + "step": 2890 + }, + { + "epoch": 0.4353230913568664, + "grad_norm": 0.7654128074645996, + "learning_rate": 1.357482456862576e-05, + "loss": 2.1568, + "num_input_tokens_seen": 1520435200, + "step": 2900 + }, + { + "epoch": 0.43682420546499356, + "grad_norm": 0.839947521686554, + "learning_rate": 1.3551480055102794e-05, + "loss": 2.1651, + "num_input_tokens_seen": 1525678080, + "step": 2910 + }, + { + "epoch": 0.43832531957312065, + "grad_norm": 0.828296959400177, + "learning_rate": 1.3528255564932864e-05, + "loss": 2.1342, + "num_input_tokens_seen": 1530920960, + "step": 2920 + }, + { + "epoch": 0.4398264336812478, + "grad_norm": 0.5922654271125793, + "learning_rate": 1.3505150073149634e-05, + "loss": 2.1632, + "num_input_tokens_seen": 1536163840, + "step": 2930 + }, + { + "epoch": 0.44132754778937494, + "grad_norm": 0.6538131833076477, + "learning_rate": 1.3482162566999154e-05, + "loss": 2.1633, + "num_input_tokens_seen": 1541406720, + "step": 2940 + }, + { + "epoch": 0.44282866189750203, + "grad_norm": 0.7411973476409912, + "learning_rate": 1.3459292045753423e-05, + "loss": 2.1349, + "num_input_tokens_seen": 1546649600, + "step": 2950 + }, + { + "epoch": 0.4443297760056292, + "grad_norm": 0.6776240468025208, + "learning_rate": 1.3436537520527398e-05, + "loss": 2.1368, + "num_input_tokens_seen": 1551892480, + "step": 2960 + }, + { + "epoch": 0.4458308901137563, + "grad_norm": 0.6425411105155945, + "learning_rate": 1.3413898014099404e-05, + "loss": 2.1358, + "num_input_tokens_seen": 1557135360, + "step": 2970 + }, + { + "epoch": 0.4473320042218834, + "grad_norm": 0.6736078858375549, + "learning_rate": 1.3391372560734868e-05, + "loss": 2.1154, + "num_input_tokens_seen": 1562378240, + "step": 2980 + }, + { + "epoch": 0.44883311833001055, + "grad_norm": 0.6204555630683899, + "learning_rate": 1.3368960206013277e-05, + "loss": 2.1744, + "num_input_tokens_seen": 1567621120, + "step": 2990 + }, + { + "epoch": 0.4503342324381377, + "grad_norm": 0.6431263089179993, + "learning_rate": 1.3346660006658346e-05, + "loss": 2.1463, + "num_input_tokens_seen": 1572864000, + "step": 3000 + }, + { + "epoch": 0.4518353465462648, + "grad_norm": 0.6702350378036499, + "learning_rate": 1.332447103037125e-05, + "loss": 2.1422, + "num_input_tokens_seen": 1578106880, + "step": 3010 + }, + { + "epoch": 0.4533364606543919, + "grad_norm": 0.5336691737174988, + "learning_rate": 1.330239235566693e-05, + "loss": 2.1515, + "num_input_tokens_seen": 1583349760, + "step": 3020 + }, + { + "epoch": 0.45483757476251907, + "grad_norm": 0.6110213398933411, + "learning_rate": 1.3280423071713355e-05, + "loss": 2.1564, + "num_input_tokens_seen": 1588592640, + "step": 3030 + }, + { + "epoch": 0.45633868887064616, + "grad_norm": 0.6357016563415527, + "learning_rate": 1.3258562278173699e-05, + "loss": 2.1591, + "num_input_tokens_seen": 1593835520, + "step": 3040 + }, + { + "epoch": 0.4578398029787733, + "grad_norm": 0.609986424446106, + "learning_rate": 1.3236809085051366e-05, + "loss": 2.1862, + "num_input_tokens_seen": 1599078400, + "step": 3050 + }, + { + "epoch": 0.45934091708690045, + "grad_norm": 0.8125211000442505, + "learning_rate": 1.3215162612537805e-05, + "loss": 2.121, + "num_input_tokens_seen": 1604321280, + "step": 3060 + }, + { + "epoch": 0.46084203119502753, + "grad_norm": 0.5937681198120117, + "learning_rate": 1.3193621990863051e-05, + "loss": 2.1548, + "num_input_tokens_seen": 1609564160, + "step": 3070 + }, + { + "epoch": 0.4623431453031547, + "grad_norm": 0.6638383865356445, + "learning_rate": 1.3172186360148932e-05, + "loss": 2.1251, + "num_input_tokens_seen": 1614807040, + "step": 3080 + }, + { + "epoch": 0.4638442594112818, + "grad_norm": 0.6757727861404419, + "learning_rate": 1.3150854870264907e-05, + "loss": 2.1388, + "num_input_tokens_seen": 1620049920, + "step": 3090 + }, + { + "epoch": 0.4653453735194089, + "grad_norm": 0.5711122155189514, + "learning_rate": 1.3129626680686454e-05, + "loss": 2.1498, + "num_input_tokens_seen": 1625292800, + "step": 3100 + }, + { + "epoch": 0.46684648762753606, + "grad_norm": 0.6421942114830017, + "learning_rate": 1.3108500960355947e-05, + "loss": 2.149, + "num_input_tokens_seen": 1630535680, + "step": 3110 + }, + { + "epoch": 0.4683476017356632, + "grad_norm": 0.7078666090965271, + "learning_rate": 1.3087476887546022e-05, + "loss": 2.148, + "num_input_tokens_seen": 1635778560, + "step": 3120 + }, + { + "epoch": 0.4698487158437903, + "grad_norm": 0.69465172290802, + "learning_rate": 1.3066553649725324e-05, + "loss": 2.1375, + "num_input_tokens_seen": 1641021440, + "step": 3130 + }, + { + "epoch": 0.47134982995191743, + "grad_norm": 0.6873762011528015, + "learning_rate": 1.304573044342661e-05, + "loss": 2.1458, + "num_input_tokens_seen": 1646264320, + "step": 3140 + }, + { + "epoch": 0.4728509440600446, + "grad_norm": 0.6284067034721375, + "learning_rate": 1.302500647411715e-05, + "loss": 2.1442, + "num_input_tokens_seen": 1651507200, + "step": 3150 + }, + { + "epoch": 0.47435205816817166, + "grad_norm": 0.8749895095825195, + "learning_rate": 1.3004380956071416e-05, + "loss": 2.1304, + "num_input_tokens_seen": 1656750080, + "step": 3160 + }, + { + "epoch": 0.4758531722762988, + "grad_norm": 0.6920912265777588, + "learning_rate": 1.298385311224594e-05, + "loss": 2.1086, + "num_input_tokens_seen": 1661992960, + "step": 3170 + }, + { + "epoch": 0.47735428638442595, + "grad_norm": 0.6951104402542114, + "learning_rate": 1.296342217415636e-05, + "loss": 2.1271, + "num_input_tokens_seen": 1667235840, + "step": 3180 + }, + { + "epoch": 0.47885540049255304, + "grad_norm": 0.784321665763855, + "learning_rate": 1.2943087381756598e-05, + "loss": 2.1074, + "num_input_tokens_seen": 1672478720, + "step": 3190 + }, + { + "epoch": 0.4803565146006802, + "grad_norm": 0.6487578749656677, + "learning_rate": 1.2922847983320086e-05, + "loss": 2.1105, + "num_input_tokens_seen": 1677721600, + "step": 3200 + }, + { + "epoch": 0.4803565146006802, + "eval_accuracy": 0.5689149369149369, + "eval_loss": 2.115828514099121, + "eval_runtime": 94.253, + "eval_samples_per_second": 3.183, + "eval_steps_per_second": 0.796, + "num_input_tokens_seen": 1677721600, + "step": 3200 + }, + { + "epoch": 0.48185762870880733, + "grad_norm": 0.7569764852523804, + "learning_rate": 1.2902703235323062e-05, + "loss": 2.1195, + "num_input_tokens_seen": 1682964480, + "step": 3210 + }, + { + "epoch": 0.4833587428169344, + "grad_norm": 0.5782780051231384, + "learning_rate": 1.2882652402329833e-05, + "loss": 2.1186, + "num_input_tokens_seen": 1688207360, + "step": 3220 + }, + { + "epoch": 0.48485985692506156, + "grad_norm": 0.6713874936103821, + "learning_rate": 1.286269475688002e-05, + "loss": 2.1446, + "num_input_tokens_seen": 1693450240, + "step": 3230 + }, + { + "epoch": 0.4863609710331887, + "grad_norm": 0.6497519016265869, + "learning_rate": 1.2842829579377681e-05, + "loss": 2.1245, + "num_input_tokens_seen": 1698693120, + "step": 3240 + }, + { + "epoch": 0.4878620851413158, + "grad_norm": 0.6722620129585266, + "learning_rate": 1.2823056157982359e-05, + "loss": 2.108, + "num_input_tokens_seen": 1703936000, + "step": 3250 + }, + { + "epoch": 0.48936319924944294, + "grad_norm": 0.6298756003379822, + "learning_rate": 1.2803373788501927e-05, + "loss": 2.1412, + "num_input_tokens_seen": 1709178880, + "step": 3260 + }, + { + "epoch": 0.4908643133575701, + "grad_norm": 0.6180531978607178, + "learning_rate": 1.2783781774287272e-05, + "loss": 2.16, + "num_input_tokens_seen": 1714421760, + "step": 3270 + }, + { + "epoch": 0.49236542746569717, + "grad_norm": 0.6927182674407959, + "learning_rate": 1.2764279426128711e-05, + "loss": 2.1604, + "num_input_tokens_seen": 1719664640, + "step": 3280 + }, + { + "epoch": 0.4938665415738243, + "grad_norm": 0.8382894396781921, + "learning_rate": 1.2744866062154176e-05, + "loss": 2.1509, + "num_input_tokens_seen": 1724907520, + "step": 3290 + }, + { + "epoch": 0.49536765568195146, + "grad_norm": 0.634811282157898, + "learning_rate": 1.2725541007729061e-05, + "loss": 2.1651, + "num_input_tokens_seen": 1730150400, + "step": 3300 + }, + { + "epoch": 0.4968687697900786, + "grad_norm": 0.5942947268486023, + "learning_rate": 1.2706303595357772e-05, + "loss": 2.158, + "num_input_tokens_seen": 1735393280, + "step": 3310 + }, + { + "epoch": 0.4983698838982057, + "grad_norm": 0.6313604712486267, + "learning_rate": 1.2687153164586875e-05, + "loss": 2.1394, + "num_input_tokens_seen": 1740636160, + "step": 3320 + }, + { + "epoch": 0.49987099800633283, + "grad_norm": 0.604307234287262, + "learning_rate": 1.2668089061909872e-05, + "loss": 2.1611, + "num_input_tokens_seen": 1745879040, + "step": 3330 + }, + { + "epoch": 0.50137211211446, + "grad_norm": 0.6554628610610962, + "learning_rate": 1.2649110640673518e-05, + "loss": 2.1072, + "num_input_tokens_seen": 1751121920, + "step": 3340 + }, + { + "epoch": 0.5028732262225871, + "grad_norm": 0.7124990820884705, + "learning_rate": 1.2630217260985716e-05, + "loss": 2.1346, + "num_input_tokens_seen": 1756364800, + "step": 3350 + }, + { + "epoch": 0.5043743403307142, + "grad_norm": 0.784233033657074, + "learning_rate": 1.2611408289624877e-05, + "loss": 2.1628, + "num_input_tokens_seen": 1761607680, + "step": 3360 + }, + { + "epoch": 0.5058754544388413, + "grad_norm": 0.8534424901008606, + "learning_rate": 1.2592683099950795e-05, + "loss": 2.1577, + "num_input_tokens_seen": 1766850560, + "step": 3370 + }, + { + "epoch": 0.5073765685469684, + "grad_norm": 0.7110067009925842, + "learning_rate": 1.2574041071816953e-05, + "loss": 2.1333, + "num_input_tokens_seen": 1772093440, + "step": 3380 + }, + { + "epoch": 0.5088776826550956, + "grad_norm": 0.7778078317642212, + "learning_rate": 1.2555481591484266e-05, + "loss": 2.1136, + "num_input_tokens_seen": 1777336320, + "step": 3390 + }, + { + "epoch": 0.5103787967632227, + "grad_norm": 0.7516434192657471, + "learning_rate": 1.253700405153621e-05, + "loss": 2.1456, + "num_input_tokens_seen": 1782579200, + "step": 3400 + }, + { + "epoch": 0.5118799108713499, + "grad_norm": 0.8273611664772034, + "learning_rate": 1.2518607850795318e-05, + "loss": 2.1022, + "num_input_tokens_seen": 1787822080, + "step": 3410 + }, + { + "epoch": 0.5133810249794769, + "grad_norm": 0.6412326097488403, + "learning_rate": 1.2500292394241044e-05, + "loss": 2.1175, + "num_input_tokens_seen": 1793064960, + "step": 3420 + }, + { + "epoch": 0.514882139087604, + "grad_norm": 0.6667389273643494, + "learning_rate": 1.2482057092928916e-05, + "loss": 2.1253, + "num_input_tokens_seen": 1798307840, + "step": 3430 + }, + { + "epoch": 0.5163832531957312, + "grad_norm": 0.8034870624542236, + "learning_rate": 1.2463901363910993e-05, + "loss": 2.1233, + "num_input_tokens_seen": 1803550720, + "step": 3440 + }, + { + "epoch": 0.5178843673038583, + "grad_norm": 0.646260678768158, + "learning_rate": 1.2445824630157606e-05, + "loss": 2.0914, + "num_input_tokens_seen": 1808793600, + "step": 3450 + }, + { + "epoch": 0.5193854814119855, + "grad_norm": 0.7674567699432373, + "learning_rate": 1.2427826320480309e-05, + "loss": 2.0866, + "num_input_tokens_seen": 1814036480, + "step": 3460 + }, + { + "epoch": 0.5208865955201126, + "grad_norm": 0.7650302648544312, + "learning_rate": 1.2409905869456093e-05, + "loss": 2.1287, + "num_input_tokens_seen": 1819279360, + "step": 3470 + }, + { + "epoch": 0.5223877096282397, + "grad_norm": 0.6644052267074585, + "learning_rate": 1.2392062717352773e-05, + "loss": 2.122, + "num_input_tokens_seen": 1824522240, + "step": 3480 + }, + { + "epoch": 0.5238888237363668, + "grad_norm": 0.8389644026756287, + "learning_rate": 1.2374296310055552e-05, + "loss": 2.1453, + "num_input_tokens_seen": 1829765120, + "step": 3490 + }, + { + "epoch": 0.525389937844494, + "grad_norm": 0.7249365448951721, + "learning_rate": 1.2356606098994752e-05, + "loss": 2.1418, + "num_input_tokens_seen": 1835008000, + "step": 3500 + }, + { + "epoch": 0.5268910519526211, + "grad_norm": 0.7120906710624695, + "learning_rate": 1.2338991541074669e-05, + "loss": 2.1318, + "num_input_tokens_seen": 1840250880, + "step": 3510 + }, + { + "epoch": 0.5283921660607482, + "grad_norm": 0.7992458343505859, + "learning_rate": 1.2321452098603552e-05, + "loss": 2.144, + "num_input_tokens_seen": 1845493760, + "step": 3520 + }, + { + "epoch": 0.5298932801688754, + "grad_norm": 0.8479806780815125, + "learning_rate": 1.2303987239224644e-05, + "loss": 2.1429, + "num_input_tokens_seen": 1850736640, + "step": 3530 + }, + { + "epoch": 0.5313943942770024, + "grad_norm": 0.670669436454773, + "learning_rate": 1.2286596435848341e-05, + "loss": 2.1395, + "num_input_tokens_seen": 1855979520, + "step": 3540 + }, + { + "epoch": 0.5328955083851296, + "grad_norm": 0.7732242345809937, + "learning_rate": 1.2269279166585364e-05, + "loss": 2.1563, + "num_input_tokens_seen": 1861222400, + "step": 3550 + }, + { + "epoch": 0.5343966224932567, + "grad_norm": 0.9772087931632996, + "learning_rate": 1.2252034914680983e-05, + "loss": 2.1375, + "num_input_tokens_seen": 1866465280, + "step": 3560 + }, + { + "epoch": 0.5358977366013838, + "grad_norm": 0.5742955207824707, + "learning_rate": 1.2234863168450241e-05, + "loss": 2.1077, + "num_input_tokens_seen": 1871708160, + "step": 3570 + }, + { + "epoch": 0.537398850709511, + "grad_norm": 0.7332693934440613, + "learning_rate": 1.2217763421214203e-05, + "loss": 2.1338, + "num_input_tokens_seen": 1876951040, + "step": 3580 + }, + { + "epoch": 0.5388999648176381, + "grad_norm": 0.6025305390357971, + "learning_rate": 1.2200735171237142e-05, + "loss": 2.1233, + "num_input_tokens_seen": 1882193920, + "step": 3590 + }, + { + "epoch": 0.5404010789257652, + "grad_norm": 0.6781802773475647, + "learning_rate": 1.2183777921664712e-05, + "loss": 2.1045, + "num_input_tokens_seen": 1887436800, + "step": 3600 + }, + { + "epoch": 0.5404010789257652, + "eval_accuracy": 0.56999592999593, + "eval_loss": 2.108959674835205, + "eval_runtime": 94.9286, + "eval_samples_per_second": 3.16, + "eval_steps_per_second": 0.79, + "num_input_tokens_seen": 1887436800, + "step": 3600 + }, + { + "epoch": 0.5419021930338923, + "grad_norm": 0.6946921944618225, + "learning_rate": 1.2166891180463065e-05, + "loss": 2.1645, + "num_input_tokens_seen": 1892679680, + "step": 3610 + }, + { + "epoch": 0.5434033071420195, + "grad_norm": 0.7196519374847412, + "learning_rate": 1.2150074460358885e-05, + "loss": 2.1177, + "num_input_tokens_seen": 1897922560, + "step": 3620 + }, + { + "epoch": 0.5449044212501466, + "grad_norm": 0.6712445616722107, + "learning_rate": 1.2133327278780315e-05, + "loss": 2.118, + "num_input_tokens_seen": 1903165440, + "step": 3630 + }, + { + "epoch": 0.5464055353582737, + "grad_norm": 0.6647999286651611, + "learning_rate": 1.2116649157798819e-05, + "loss": 2.1208, + "num_input_tokens_seen": 1908408320, + "step": 3640 + }, + { + "epoch": 0.5479066494664009, + "grad_norm": 0.663698136806488, + "learning_rate": 1.2100039624071883e-05, + "loss": 2.1387, + "num_input_tokens_seen": 1913651200, + "step": 3650 + }, + { + "epoch": 0.5494077635745279, + "grad_norm": 0.873149037361145, + "learning_rate": 1.2083498208786586e-05, + "loss": 2.1205, + "num_input_tokens_seen": 1918894080, + "step": 3660 + }, + { + "epoch": 0.5509088776826551, + "grad_norm": 0.6879292726516724, + "learning_rate": 1.2067024447604041e-05, + "loss": 2.1089, + "num_input_tokens_seen": 1924136960, + "step": 3670 + }, + { + "epoch": 0.5524099917907822, + "grad_norm": 0.7777056694030762, + "learning_rate": 1.2050617880604623e-05, + "loss": 2.1087, + "num_input_tokens_seen": 1929379840, + "step": 3680 + }, + { + "epoch": 0.5539111058989094, + "grad_norm": 0.7917245030403137, + "learning_rate": 1.2034278052234061e-05, + "loss": 2.094, + "num_input_tokens_seen": 1934622720, + "step": 3690 + }, + { + "epoch": 0.5554122200070365, + "grad_norm": 0.7892484664916992, + "learning_rate": 1.2018004511250296e-05, + "loss": 2.111, + "num_input_tokens_seen": 1939865600, + "step": 3700 + }, + { + "epoch": 0.5569133341151636, + "grad_norm": 0.8272449374198914, + "learning_rate": 1.2001796810671147e-05, + "loss": 2.1574, + "num_input_tokens_seen": 1945108480, + "step": 3710 + }, + { + "epoch": 0.5584144482232907, + "grad_norm": 0.6054387092590332, + "learning_rate": 1.1985654507722748e-05, + "loss": 2.1226, + "num_input_tokens_seen": 1950351360, + "step": 3720 + }, + { + "epoch": 0.5599155623314178, + "grad_norm": 0.5401285886764526, + "learning_rate": 1.1969577163788735e-05, + "loss": 2.1093, + "num_input_tokens_seen": 1955594240, + "step": 3730 + }, + { + "epoch": 0.561416676439545, + "grad_norm": 0.6505089998245239, + "learning_rate": 1.1953564344360196e-05, + "loss": 2.1177, + "num_input_tokens_seen": 1960837120, + "step": 3740 + }, + { + "epoch": 0.5629177905476721, + "grad_norm": 0.5896223783493042, + "learning_rate": 1.1937615618986343e-05, + "loss": 2.1246, + "num_input_tokens_seen": 1966080000, + "step": 3750 + }, + { + "epoch": 0.5644189046557992, + "grad_norm": 0.9334360361099243, + "learning_rate": 1.192173056122592e-05, + "loss": 2.0941, + "num_input_tokens_seen": 1971322880, + "step": 3760 + }, + { + "epoch": 0.5659200187639264, + "grad_norm": 0.8824881315231323, + "learning_rate": 1.1905908748599307e-05, + "loss": 2.135, + "num_input_tokens_seen": 1976565760, + "step": 3770 + }, + { + "epoch": 0.5674211328720534, + "grad_norm": 0.9339269399642944, + "learning_rate": 1.1890149762541318e-05, + "loss": 2.1136, + "num_input_tokens_seen": 1981808640, + "step": 3780 + }, + { + "epoch": 0.5689222469801806, + "grad_norm": 0.7937804460525513, + "learning_rate": 1.1874453188354716e-05, + "loss": 2.1191, + "num_input_tokens_seen": 1987051520, + "step": 3790 + }, + { + "epoch": 0.5704233610883077, + "grad_norm": 0.7529720067977905, + "learning_rate": 1.1858818615164347e-05, + "loss": 2.1283, + "num_input_tokens_seen": 1992294400, + "step": 3800 + }, + { + "epoch": 0.5719244751964349, + "grad_norm": 0.5535932779312134, + "learning_rate": 1.1843245635871996e-05, + "loss": 2.1311, + "num_input_tokens_seen": 1997537280, + "step": 3810 + }, + { + "epoch": 0.573425589304562, + "grad_norm": 0.6932185888290405, + "learning_rate": 1.1827733847111846e-05, + "loss": 2.1332, + "num_input_tokens_seen": 2002780160, + "step": 3820 + }, + { + "epoch": 0.5749267034126891, + "grad_norm": 0.7156121730804443, + "learning_rate": 1.181228284920661e-05, + "loss": 2.1264, + "num_input_tokens_seen": 2008023040, + "step": 3830 + }, + { + "epoch": 0.5764278175208162, + "grad_norm": 0.6193491220474243, + "learning_rate": 1.1796892246124261e-05, + "loss": 2.1376, + "num_input_tokens_seen": 2013265920, + "step": 3840 + }, + { + "epoch": 0.5779289316289433, + "grad_norm": 0.6163820624351501, + "learning_rate": 1.1781561645435414e-05, + "loss": 2.0823, + "num_input_tokens_seen": 2018508800, + "step": 3850 + }, + { + "epoch": 0.5794300457370705, + "grad_norm": 0.7165153622627258, + "learning_rate": 1.1766290658271293e-05, + "loss": 2.1194, + "num_input_tokens_seen": 2023751680, + "step": 3860 + }, + { + "epoch": 0.5809311598451976, + "grad_norm": 0.6811047196388245, + "learning_rate": 1.1751078899282295e-05, + "loss": 2.1564, + "num_input_tokens_seen": 2028994560, + "step": 3870 + }, + { + "epoch": 0.5824322739533248, + "grad_norm": 0.9540090560913086, + "learning_rate": 1.1735925986597164e-05, + "loss": 2.1241, + "num_input_tokens_seen": 2034237440, + "step": 3880 + }, + { + "epoch": 0.5839333880614519, + "grad_norm": 0.6102780103683472, + "learning_rate": 1.1720831541782706e-05, + "loss": 2.0992, + "num_input_tokens_seen": 2039480320, + "step": 3890 + }, + { + "epoch": 0.5854345021695789, + "grad_norm": 0.6293305158615112, + "learning_rate": 1.170579518980411e-05, + "loss": 2.1285, + "num_input_tokens_seen": 2044723200, + "step": 3900 + }, + { + "epoch": 0.5869356162777061, + "grad_norm": 0.7680484056472778, + "learning_rate": 1.169081655898581e-05, + "loss": 2.0774, + "num_input_tokens_seen": 2049966080, + "step": 3910 + }, + { + "epoch": 0.5884367303858332, + "grad_norm": 0.7843759059906006, + "learning_rate": 1.1675895280972886e-05, + "loss": 2.1271, + "num_input_tokens_seen": 2055208960, + "step": 3920 + }, + { + "epoch": 0.5899378444939604, + "grad_norm": 0.8438634276390076, + "learning_rate": 1.1661030990693031e-05, + "loss": 2.1214, + "num_input_tokens_seen": 2060451840, + "step": 3930 + }, + { + "epoch": 0.5914389586020875, + "grad_norm": 0.7281661033630371, + "learning_rate": 1.1646223326319031e-05, + "loss": 2.1074, + "num_input_tokens_seen": 2065694720, + "step": 3940 + }, + { + "epoch": 0.5929400727102146, + "grad_norm": 0.7339427471160889, + "learning_rate": 1.163147192923176e-05, + "loss": 2.1571, + "num_input_tokens_seen": 2070937600, + "step": 3950 + }, + { + "epoch": 0.5944411868183418, + "grad_norm": 0.583240807056427, + "learning_rate": 1.1616776443983717e-05, + "loss": 2.1111, + "num_input_tokens_seen": 2076180480, + "step": 3960 + }, + { + "epoch": 0.5959423009264688, + "grad_norm": 0.5651365518569946, + "learning_rate": 1.1602136518263054e-05, + "loss": 2.0996, + "num_input_tokens_seen": 2081423360, + "step": 3970 + }, + { + "epoch": 0.597443415034596, + "grad_norm": 0.6529115438461304, + "learning_rate": 1.1587551802858083e-05, + "loss": 2.1478, + "num_input_tokens_seen": 2086666240, + "step": 3980 + }, + { + "epoch": 0.5989445291427231, + "grad_norm": 0.8366259932518005, + "learning_rate": 1.1573021951622309e-05, + "loss": 2.1437, + "num_input_tokens_seen": 2091909120, + "step": 3990 + }, + { + "epoch": 0.6004456432508503, + "grad_norm": 0.5824588537216187, + "learning_rate": 1.1558546621439912e-05, + "loss": 2.1181, + "num_input_tokens_seen": 2097152000, + "step": 4000 + }, + { + "epoch": 0.6004456432508503, + "eval_accuracy": 0.5708310948310948, + "eval_loss": 2.104544162750244, + "eval_runtime": 94.4692, + "eval_samples_per_second": 3.176, + "eval_steps_per_second": 0.794, + "num_input_tokens_seen": 2097152000, + "step": 4000 + }, + { + "epoch": 0.6019467573589774, + "grad_norm": 0.6183150410652161, + "learning_rate": 1.1544125472191724e-05, + "loss": 2.1217, + "num_input_tokens_seen": 2102394880, + "step": 4010 + }, + { + "epoch": 0.6034478714671045, + "grad_norm": 0.6359286904335022, + "learning_rate": 1.152975816672165e-05, + "loss": 2.0985, + "num_input_tokens_seen": 2107637760, + "step": 4020 + }, + { + "epoch": 0.6049489855752316, + "grad_norm": 0.6601809859275818, + "learning_rate": 1.1515444370803586e-05, + "loss": 2.1115, + "num_input_tokens_seen": 2112880640, + "step": 4030 + }, + { + "epoch": 0.6064500996833587, + "grad_norm": 0.6836331486701965, + "learning_rate": 1.1501183753108733e-05, + "loss": 2.1012, + "num_input_tokens_seen": 2118123520, + "step": 4040 + }, + { + "epoch": 0.6079512137914859, + "grad_norm": 0.6830666661262512, + "learning_rate": 1.148697598517341e-05, + "loss": 2.1594, + "num_input_tokens_seen": 2123366400, + "step": 4050 + }, + { + "epoch": 0.609452327899613, + "grad_norm": 0.6801722049713135, + "learning_rate": 1.147282074136726e-05, + "loss": 2.1227, + "num_input_tokens_seen": 2128609280, + "step": 4060 + }, + { + "epoch": 0.6109534420077402, + "grad_norm": 0.764336347579956, + "learning_rate": 1.1458717698861917e-05, + "loss": 2.1268, + "num_input_tokens_seen": 2133852160, + "step": 4070 + }, + { + "epoch": 0.6124545561158673, + "grad_norm": 1.0238054990768433, + "learning_rate": 1.144466653760007e-05, + "loss": 2.1217, + "num_input_tokens_seen": 2139095040, + "step": 4080 + }, + { + "epoch": 0.6139556702239943, + "grad_norm": 0.6901516914367676, + "learning_rate": 1.1430666940264972e-05, + "loss": 2.1304, + "num_input_tokens_seen": 2144337920, + "step": 4090 + }, + { + "epoch": 0.6154567843321215, + "grad_norm": 0.6896687150001526, + "learning_rate": 1.141671859225032e-05, + "loss": 2.1239, + "num_input_tokens_seen": 2149580800, + "step": 4100 + }, + { + "epoch": 0.6169578984402486, + "grad_norm": 0.8825215101242065, + "learning_rate": 1.1402821181630592e-05, + "loss": 2.1061, + "num_input_tokens_seen": 2154823680, + "step": 4110 + }, + { + "epoch": 0.6184590125483758, + "grad_norm": 0.7571860551834106, + "learning_rate": 1.1388974399131718e-05, + "loss": 2.1222, + "num_input_tokens_seen": 2160066560, + "step": 4120 + }, + { + "epoch": 0.6199601266565029, + "grad_norm": 0.6392367482185364, + "learning_rate": 1.1375177938102187e-05, + "loss": 2.156, + "num_input_tokens_seen": 2165309440, + "step": 4130 + }, + { + "epoch": 0.62146124076463, + "grad_norm": 0.7527381777763367, + "learning_rate": 1.136143149448452e-05, + "loss": 2.1233, + "num_input_tokens_seen": 2170552320, + "step": 4140 + }, + { + "epoch": 0.6229623548727571, + "grad_norm": 0.7589773535728455, + "learning_rate": 1.1347734766787132e-05, + "loss": 2.1081, + "num_input_tokens_seen": 2175795200, + "step": 4150 + }, + { + "epoch": 0.6244634689808842, + "grad_norm": 0.8642693161964417, + "learning_rate": 1.1334087456056552e-05, + "loss": 2.1131, + "num_input_tokens_seen": 2181038080, + "step": 4160 + }, + { + "epoch": 0.6259645830890114, + "grad_norm": 0.7018624544143677, + "learning_rate": 1.1320489265850016e-05, + "loss": 2.1194, + "num_input_tokens_seen": 2186280960, + "step": 4170 + }, + { + "epoch": 0.6274656971971385, + "grad_norm": 0.632854163646698, + "learning_rate": 1.1306939902208426e-05, + "loss": 2.1336, + "num_input_tokens_seen": 2191523840, + "step": 4180 + }, + { + "epoch": 0.6289668113052657, + "grad_norm": 0.8238286972045898, + "learning_rate": 1.1293439073629661e-05, + "loss": 2.1341, + "num_input_tokens_seen": 2196766720, + "step": 4190 + }, + { + "epoch": 0.6304679254133928, + "grad_norm": 0.8351567387580872, + "learning_rate": 1.1279986491042232e-05, + "loss": 2.0898, + "num_input_tokens_seen": 2202009600, + "step": 4200 + }, + { + "epoch": 0.6319690395215198, + "grad_norm": 0.5938683748245239, + "learning_rate": 1.1266581867779284e-05, + "loss": 2.1127, + "num_input_tokens_seen": 2207252480, + "step": 4210 + }, + { + "epoch": 0.633470153629647, + "grad_norm": 0.7919456958770752, + "learning_rate": 1.1253224919552947e-05, + "loss": 2.1014, + "num_input_tokens_seen": 2212495360, + "step": 4220 + }, + { + "epoch": 0.6349712677377741, + "grad_norm": 0.6826079487800598, + "learning_rate": 1.1239915364429005e-05, + "loss": 2.121, + "num_input_tokens_seen": 2217738240, + "step": 4230 + }, + { + "epoch": 0.6364723818459013, + "grad_norm": 0.7500185966491699, + "learning_rate": 1.1226652922801909e-05, + "loss": 2.123, + "num_input_tokens_seen": 2222981120, + "step": 4240 + }, + { + "epoch": 0.6379734959540284, + "grad_norm": 0.6904823780059814, + "learning_rate": 1.1213437317370097e-05, + "loss": 2.1258, + "num_input_tokens_seen": 2228224000, + "step": 4250 + }, + { + "epoch": 0.6394746100621556, + "grad_norm": 0.6400703191757202, + "learning_rate": 1.1200268273111648e-05, + "loss": 2.1147, + "num_input_tokens_seen": 2233466880, + "step": 4260 + }, + { + "epoch": 0.6409757241702826, + "grad_norm": 0.805190920829773, + "learning_rate": 1.1187145517260246e-05, + "loss": 2.1117, + "num_input_tokens_seen": 2238709760, + "step": 4270 + }, + { + "epoch": 0.6424768382784097, + "grad_norm": 0.6888982653617859, + "learning_rate": 1.117406877928144e-05, + "loss": 2.1171, + "num_input_tokens_seen": 2243952640, + "step": 4280 + }, + { + "epoch": 0.6439779523865369, + "grad_norm": 0.7432284355163574, + "learning_rate": 1.1161037790849228e-05, + "loss": 2.1098, + "num_input_tokens_seen": 2249195520, + "step": 4290 + }, + { + "epoch": 0.645479066494664, + "grad_norm": 0.6752641201019287, + "learning_rate": 1.114805228582293e-05, + "loss": 2.1208, + "num_input_tokens_seen": 2254438400, + "step": 4300 + }, + { + "epoch": 0.6469801806027912, + "grad_norm": 0.8564686179161072, + "learning_rate": 1.1135112000224364e-05, + "loss": 2.1125, + "num_input_tokens_seen": 2259681280, + "step": 4310 + }, + { + "epoch": 0.6484812947109183, + "grad_norm": 0.9626083970069885, + "learning_rate": 1.1122216672215286e-05, + "loss": 2.1215, + "num_input_tokens_seen": 2264924160, + "step": 4320 + }, + { + "epoch": 0.6499824088190453, + "grad_norm": 0.5604343414306641, + "learning_rate": 1.1109366042075172e-05, + "loss": 2.1667, + "num_input_tokens_seen": 2270167040, + "step": 4330 + }, + { + "epoch": 0.6514835229271725, + "grad_norm": 0.5957362651824951, + "learning_rate": 1.1096559852179205e-05, + "loss": 2.0962, + "num_input_tokens_seen": 2275409920, + "step": 4340 + }, + { + "epoch": 0.6529846370352996, + "grad_norm": 0.6427695751190186, + "learning_rate": 1.1083797846976627e-05, + "loss": 2.1203, + "num_input_tokens_seen": 2280652800, + "step": 4350 + }, + { + "epoch": 0.6544857511434268, + "grad_norm": 0.6548131108283997, + "learning_rate": 1.1071079772969281e-05, + "loss": 2.1172, + "num_input_tokens_seen": 2285895680, + "step": 4360 + }, + { + "epoch": 0.6559868652515539, + "grad_norm": 0.6767981052398682, + "learning_rate": 1.1058405378690493e-05, + "loss": 2.1267, + "num_input_tokens_seen": 2291138560, + "step": 4370 + }, + { + "epoch": 0.6574879793596811, + "grad_norm": 0.7053226232528687, + "learning_rate": 1.1045774414684167e-05, + "loss": 2.1183, + "num_input_tokens_seen": 2296381440, + "step": 4380 + }, + { + "epoch": 0.6589890934678081, + "grad_norm": 0.7884171605110168, + "learning_rate": 1.1033186633484174e-05, + "loss": 2.1103, + "num_input_tokens_seen": 2301624320, + "step": 4390 + }, + { + "epoch": 0.6604902075759352, + "grad_norm": 0.6177472472190857, + "learning_rate": 1.1020641789593993e-05, + "loss": 2.127, + "num_input_tokens_seen": 2306867200, + "step": 4400 + }, + { + "epoch": 0.6604902075759352, + "eval_accuracy": 0.5715555555555556, + "eval_loss": 2.099358081817627, + "eval_runtime": 93.9705, + "eval_samples_per_second": 3.192, + "eval_steps_per_second": 0.798, + "num_input_tokens_seen": 2306867200, + "step": 4400 + }, + { + "epoch": 0.6619913216840624, + "grad_norm": 0.6775453090667725, + "learning_rate": 1.1008139639466584e-05, + "loss": 2.1329, + "num_input_tokens_seen": 2312110080, + "step": 4410 + }, + { + "epoch": 0.6634924357921895, + "grad_norm": 0.6154870986938477, + "learning_rate": 1.0995679941484547e-05, + "loss": 2.0883, + "num_input_tokens_seen": 2317352960, + "step": 4420 + }, + { + "epoch": 0.6649935499003167, + "grad_norm": 0.7121848464012146, + "learning_rate": 1.0983262455940505e-05, + "loss": 2.1364, + "num_input_tokens_seen": 2322595840, + "step": 4430 + }, + { + "epoch": 0.6664946640084438, + "grad_norm": 0.7415539622306824, + "learning_rate": 1.0970886945017727e-05, + "loss": 2.1281, + "num_input_tokens_seen": 2327838720, + "step": 4440 + }, + { + "epoch": 0.6679957781165708, + "grad_norm": 0.678644597530365, + "learning_rate": 1.0958553172771004e-05, + "loss": 2.0817, + "num_input_tokens_seen": 2333081600, + "step": 4450 + }, + { + "epoch": 0.669496892224698, + "grad_norm": 0.6753328442573547, + "learning_rate": 1.0946260905107762e-05, + "loss": 2.1169, + "num_input_tokens_seen": 2338324480, + "step": 4460 + }, + { + "epoch": 0.6709980063328251, + "grad_norm": 0.6162213087081909, + "learning_rate": 1.0934009909769385e-05, + "loss": 2.1551, + "num_input_tokens_seen": 2343567360, + "step": 4470 + }, + { + "epoch": 0.6724991204409523, + "grad_norm": 0.7211172580718994, + "learning_rate": 1.0921799956312801e-05, + "loss": 2.1153, + "num_input_tokens_seen": 2348810240, + "step": 4480 + }, + { + "epoch": 0.6740002345490794, + "grad_norm": 0.8724449872970581, + "learning_rate": 1.090963081609226e-05, + "loss": 2.1171, + "num_input_tokens_seen": 2354053120, + "step": 4490 + }, + { + "epoch": 0.6755013486572066, + "grad_norm": 0.8566068410873413, + "learning_rate": 1.0897502262241359e-05, + "loss": 2.1138, + "num_input_tokens_seen": 2359296000, + "step": 4500 + }, + { + "epoch": 0.6770024627653336, + "grad_norm": 0.6851250529289246, + "learning_rate": 1.0885414069655284e-05, + "loss": 2.1129, + "num_input_tokens_seen": 2364538880, + "step": 4510 + }, + { + "epoch": 0.6785035768734607, + "grad_norm": 0.6527068614959717, + "learning_rate": 1.087336601497325e-05, + "loss": 2.1354, + "num_input_tokens_seen": 2369781760, + "step": 4520 + }, + { + "epoch": 0.6800046909815879, + "grad_norm": 0.6223852038383484, + "learning_rate": 1.086135787656117e-05, + "loss": 2.1055, + "num_input_tokens_seen": 2375024640, + "step": 4530 + }, + { + "epoch": 0.681505805089715, + "grad_norm": 0.7803609371185303, + "learning_rate": 1.0849389434494532e-05, + "loss": 2.1206, + "num_input_tokens_seen": 2380267520, + "step": 4540 + }, + { + "epoch": 0.6830069191978422, + "grad_norm": 0.8386520743370056, + "learning_rate": 1.0837460470541495e-05, + "loss": 2.1396, + "num_input_tokens_seen": 2385510400, + "step": 4550 + }, + { + "epoch": 0.6845080333059693, + "grad_norm": 0.8144930601119995, + "learning_rate": 1.0825570768146148e-05, + "loss": 2.1149, + "num_input_tokens_seen": 2390753280, + "step": 4560 + }, + { + "epoch": 0.6860091474140964, + "grad_norm": 0.8149564266204834, + "learning_rate": 1.081372011241202e-05, + "loss": 2.1056, + "num_input_tokens_seen": 2395996160, + "step": 4570 + }, + { + "epoch": 0.6875102615222235, + "grad_norm": 0.5903097987174988, + "learning_rate": 1.0801908290085765e-05, + "loss": 2.123, + "num_input_tokens_seen": 2401239040, + "step": 4580 + }, + { + "epoch": 0.6890113756303506, + "grad_norm": 0.6024293899536133, + "learning_rate": 1.0790135089541035e-05, + "loss": 2.1504, + "num_input_tokens_seen": 2406481920, + "step": 4590 + }, + { + "epoch": 0.6905124897384778, + "grad_norm": 1.085522174835205, + "learning_rate": 1.0778400300762553e-05, + "loss": 2.1158, + "num_input_tokens_seen": 2411724800, + "step": 4600 + }, + { + "epoch": 0.6920136038466049, + "grad_norm": 0.6899899840354919, + "learning_rate": 1.0766703715330396e-05, + "loss": 2.0817, + "num_input_tokens_seen": 2416967680, + "step": 4610 + }, + { + "epoch": 0.6935147179547321, + "grad_norm": 0.732567310333252, + "learning_rate": 1.075504512640443e-05, + "loss": 2.1013, + "num_input_tokens_seen": 2422210560, + "step": 4620 + }, + { + "epoch": 0.6950158320628591, + "grad_norm": 0.7568468451499939, + "learning_rate": 1.0743424328708955e-05, + "loss": 2.1554, + "num_input_tokens_seen": 2427453440, + "step": 4630 + }, + { + "epoch": 0.6965169461709863, + "grad_norm": 0.7178658843040466, + "learning_rate": 1.0731841118517526e-05, + "loss": 2.0937, + "num_input_tokens_seen": 2432696320, + "step": 4640 + }, + { + "epoch": 0.6980180602791134, + "grad_norm": 0.8591102361679077, + "learning_rate": 1.072029529363794e-05, + "loss": 2.11, + "num_input_tokens_seen": 2437939200, + "step": 4650 + }, + { + "epoch": 0.6995191743872405, + "grad_norm": 0.776842474937439, + "learning_rate": 1.0708786653397427e-05, + "loss": 2.1335, + "num_input_tokens_seen": 2443182080, + "step": 4660 + }, + { + "epoch": 0.7010202884953677, + "grad_norm": 0.6280748248100281, + "learning_rate": 1.0697314998628e-05, + "loss": 2.0915, + "num_input_tokens_seen": 2448424960, + "step": 4670 + }, + { + "epoch": 0.7025214026034948, + "grad_norm": 0.6168721318244934, + "learning_rate": 1.0685880131651965e-05, + "loss": 2.0978, + "num_input_tokens_seen": 2453667840, + "step": 4680 + }, + { + "epoch": 0.7040225167116219, + "grad_norm": 0.7713338732719421, + "learning_rate": 1.067448185626763e-05, + "loss": 2.0992, + "num_input_tokens_seen": 2458910720, + "step": 4690 + }, + { + "epoch": 0.705523630819749, + "grad_norm": 0.9560806751251221, + "learning_rate": 1.0663119977735152e-05, + "loss": 2.1351, + "num_input_tokens_seen": 2464153600, + "step": 4700 + }, + { + "epoch": 0.7070247449278761, + "grad_norm": 0.7966359257698059, + "learning_rate": 1.0651794302762573e-05, + "loss": 2.1285, + "num_input_tokens_seen": 2469396480, + "step": 4710 + }, + { + "epoch": 0.7085258590360033, + "grad_norm": 0.7075307965278625, + "learning_rate": 1.0640504639492005e-05, + "loss": 2.1108, + "num_input_tokens_seen": 2474639360, + "step": 4720 + }, + { + "epoch": 0.7100269731441304, + "grad_norm": 0.7893349528312683, + "learning_rate": 1.0629250797485977e-05, + "loss": 2.1133, + "num_input_tokens_seen": 2479882240, + "step": 4730 + }, + { + "epoch": 0.7115280872522576, + "grad_norm": 0.6469416618347168, + "learning_rate": 1.0618032587713944e-05, + "loss": 2.1259, + "num_input_tokens_seen": 2485125120, + "step": 4740 + }, + { + "epoch": 0.7130292013603846, + "grad_norm": 0.6949597001075745, + "learning_rate": 1.0606849822538959e-05, + "loss": 2.1272, + "num_input_tokens_seen": 2490368000, + "step": 4750 + }, + { + "epoch": 0.7145303154685118, + "grad_norm": 0.6256146430969238, + "learning_rate": 1.0595702315704477e-05, + "loss": 2.1132, + "num_input_tokens_seen": 2495610880, + "step": 4760 + }, + { + "epoch": 0.7160314295766389, + "grad_norm": 0.644726037979126, + "learning_rate": 1.0584589882321336e-05, + "loss": 2.1304, + "num_input_tokens_seen": 2500853760, + "step": 4770 + }, + { + "epoch": 0.717532543684766, + "grad_norm": 0.9808601140975952, + "learning_rate": 1.0573512338854876e-05, + "loss": 2.1061, + "num_input_tokens_seen": 2506096640, + "step": 4780 + }, + { + "epoch": 0.7190336577928932, + "grad_norm": 0.829621434211731, + "learning_rate": 1.0562469503112205e-05, + "loss": 2.1288, + "num_input_tokens_seen": 2511339520, + "step": 4790 + }, + { + "epoch": 0.7205347719010203, + "grad_norm": 1.08700692653656, + "learning_rate": 1.055146119422961e-05, + "loss": 2.1265, + "num_input_tokens_seen": 2516582400, + "step": 4800 + }, + { + "epoch": 0.7205347719010203, + "eval_accuracy": 0.571921855921856, + "eval_loss": 2.095761299133301, + "eval_runtime": 95.3789, + "eval_samples_per_second": 3.145, + "eval_steps_per_second": 0.786, + "num_input_tokens_seen": 2516582400, + "step": 4800 + }, + { + "epoch": 0.7220358860091474, + "grad_norm": 0.7580691576004028, + "learning_rate": 1.0540487232660126e-05, + "loss": 2.118, + "num_input_tokens_seen": 2521825280, + "step": 4810 + }, + { + "epoch": 0.7235370001172745, + "grad_norm": 0.7149703502655029, + "learning_rate": 1.0529547440161225e-05, + "loss": 2.0822, + "num_input_tokens_seen": 2527068160, + "step": 4820 + }, + { + "epoch": 0.7250381142254017, + "grad_norm": 0.6624520421028137, + "learning_rate": 1.0518641639782667e-05, + "loss": 2.1325, + "num_input_tokens_seen": 2532311040, + "step": 4830 + }, + { + "epoch": 0.7265392283335288, + "grad_norm": 0.7207245826721191, + "learning_rate": 1.0507769655854467e-05, + "loss": 2.1231, + "num_input_tokens_seen": 2537553920, + "step": 4840 + }, + { + "epoch": 0.7280403424416559, + "grad_norm": 0.6605533361434937, + "learning_rate": 1.0496931313975018e-05, + "loss": 2.0944, + "num_input_tokens_seen": 2542796800, + "step": 4850 + }, + { + "epoch": 0.7295414565497831, + "grad_norm": 0.6779858469963074, + "learning_rate": 1.0486126440999343e-05, + "loss": 2.1355, + "num_input_tokens_seen": 2548039680, + "step": 4860 + }, + { + "epoch": 0.7310425706579101, + "grad_norm": 0.9669274091720581, + "learning_rate": 1.047535486502747e-05, + "loss": 2.1198, + "num_input_tokens_seen": 2553282560, + "step": 4870 + }, + { + "epoch": 0.7325436847660373, + "grad_norm": 0.6352229714393616, + "learning_rate": 1.0464616415392945e-05, + "loss": 2.0968, + "num_input_tokens_seen": 2558525440, + "step": 4880 + }, + { + "epoch": 0.7340447988741644, + "grad_norm": 0.9474151134490967, + "learning_rate": 1.0453910922651489e-05, + "loss": 2.1213, + "num_input_tokens_seen": 2563768320, + "step": 4890 + }, + { + "epoch": 0.7355459129822915, + "grad_norm": 0.6039556264877319, + "learning_rate": 1.0443238218569753e-05, + "loss": 2.1092, + "num_input_tokens_seen": 2569011200, + "step": 4900 + }, + { + "epoch": 0.7370470270904187, + "grad_norm": 0.6552874445915222, + "learning_rate": 1.0432598136114214e-05, + "loss": 2.1292, + "num_input_tokens_seen": 2574254080, + "step": 4910 + }, + { + "epoch": 0.7385481411985458, + "grad_norm": 0.6924867033958435, + "learning_rate": 1.042199050944021e-05, + "loss": 2.1396, + "num_input_tokens_seen": 2579496960, + "step": 4920 + }, + { + "epoch": 0.7400492553066729, + "grad_norm": 0.7396571636199951, + "learning_rate": 1.0411415173881066e-05, + "loss": 2.1028, + "num_input_tokens_seen": 2584739840, + "step": 4930 + }, + { + "epoch": 0.7415503694148, + "grad_norm": 0.737847626209259, + "learning_rate": 1.0400871965937369e-05, + "loss": 2.0786, + "num_input_tokens_seen": 2589982720, + "step": 4940 + }, + { + "epoch": 0.7430514835229272, + "grad_norm": 0.7622154355049133, + "learning_rate": 1.0390360723266348e-05, + "loss": 2.1172, + "num_input_tokens_seen": 2595225600, + "step": 4950 + }, + { + "epoch": 0.7445525976310543, + "grad_norm": 0.9400896430015564, + "learning_rate": 1.0379881284671372e-05, + "loss": 2.1024, + "num_input_tokens_seen": 2600468480, + "step": 4960 + }, + { + "epoch": 0.7460537117391814, + "grad_norm": 0.8766177296638489, + "learning_rate": 1.0369433490091569e-05, + "loss": 2.1389, + "num_input_tokens_seen": 2605711360, + "step": 4970 + }, + { + "epoch": 0.7475548258473086, + "grad_norm": 0.9279113411903381, + "learning_rate": 1.0359017180591565e-05, + "loss": 2.1417, + "num_input_tokens_seen": 2610954240, + "step": 4980 + }, + { + "epoch": 0.7490559399554356, + "grad_norm": 0.6392689347267151, + "learning_rate": 1.0348632198351312e-05, + "loss": 2.1454, + "num_input_tokens_seen": 2616197120, + "step": 4990 + }, + { + "epoch": 0.7505570540635628, + "grad_norm": 0.6666054129600525, + "learning_rate": 1.0338278386656068e-05, + "loss": 2.0914, + "num_input_tokens_seen": 2621440000, + "step": 5000 + }, + { + "epoch": 0.7520581681716899, + "grad_norm": 0.8447917103767395, + "learning_rate": 1.0327955589886445e-05, + "loss": 2.112, + "num_input_tokens_seen": 2626682880, + "step": 5010 + }, + { + "epoch": 0.753559282279817, + "grad_norm": 0.7507166862487793, + "learning_rate": 1.0317663653508592e-05, + "loss": 2.0921, + "num_input_tokens_seen": 2631925760, + "step": 5020 + }, + { + "epoch": 0.7550603963879442, + "grad_norm": 0.7831541299819946, + "learning_rate": 1.0307402424064481e-05, + "loss": 2.1418, + "num_input_tokens_seen": 2637168640, + "step": 5030 + }, + { + "epoch": 0.7565615104960713, + "grad_norm": 0.6328726410865784, + "learning_rate": 1.0297171749162294e-05, + "loss": 2.1253, + "num_input_tokens_seen": 2642411520, + "step": 5040 + }, + { + "epoch": 0.7580626246041984, + "grad_norm": 0.8324751257896423, + "learning_rate": 1.0286971477466913e-05, + "loss": 2.1222, + "num_input_tokens_seen": 2647654400, + "step": 5050 + }, + { + "epoch": 0.7595637387123255, + "grad_norm": 0.7374483942985535, + "learning_rate": 1.0276801458690536e-05, + "loss": 2.1117, + "num_input_tokens_seen": 2652897280, + "step": 5060 + }, + { + "epoch": 0.7610648528204527, + "grad_norm": 0.787074625492096, + "learning_rate": 1.0266661543583343e-05, + "loss": 2.0947, + "num_input_tokens_seen": 2658140160, + "step": 5070 + }, + { + "epoch": 0.7625659669285798, + "grad_norm": 0.6950966119766235, + "learning_rate": 1.0256551583924336e-05, + "loss": 2.1237, + "num_input_tokens_seen": 2663383040, + "step": 5080 + }, + { + "epoch": 0.764067081036707, + "grad_norm": 0.7359448671340942, + "learning_rate": 1.024647143251221e-05, + "loss": 2.1298, + "num_input_tokens_seen": 2668625920, + "step": 5090 + }, + { + "epoch": 0.7655681951448341, + "grad_norm": 0.7270290851593018, + "learning_rate": 1.0236420943156362e-05, + "loss": 2.1171, + "num_input_tokens_seen": 2673868800, + "step": 5100 + }, + { + "epoch": 0.7670693092529611, + "grad_norm": 0.9731076955795288, + "learning_rate": 1.022639997066799e-05, + "loss": 2.0728, + "num_input_tokens_seen": 2679111680, + "step": 5110 + }, + { + "epoch": 0.7685704233610883, + "grad_norm": 0.7202324867248535, + "learning_rate": 1.0216408370851276e-05, + "loss": 2.1083, + "num_input_tokens_seen": 2684354560, + "step": 5120 + }, + { + "epoch": 0.7700715374692154, + "grad_norm": 0.7496753334999084, + "learning_rate": 1.020644600049467e-05, + "loss": 2.105, + "num_input_tokens_seen": 2689597440, + "step": 5130 + }, + { + "epoch": 0.7715726515773426, + "grad_norm": 0.663766622543335, + "learning_rate": 1.0196512717362284e-05, + "loss": 2.1095, + "num_input_tokens_seen": 2694840320, + "step": 5140 + }, + { + "epoch": 0.7730737656854697, + "grad_norm": 0.697460949420929, + "learning_rate": 1.0186608380185336e-05, + "loss": 2.115, + "num_input_tokens_seen": 2700083200, + "step": 5150 + }, + { + "epoch": 0.7745748797935968, + "grad_norm": 0.7035109400749207, + "learning_rate": 1.0176732848653748e-05, + "loss": 2.1035, + "num_input_tokens_seen": 2705326080, + "step": 5160 + }, + { + "epoch": 0.7760759939017239, + "grad_norm": 0.7405036091804504, + "learning_rate": 1.0166885983407759e-05, + "loss": 2.1061, + "num_input_tokens_seen": 2710568960, + "step": 5170 + }, + { + "epoch": 0.777577108009851, + "grad_norm": 0.7046319842338562, + "learning_rate": 1.0157067646029694e-05, + "loss": 2.1098, + "num_input_tokens_seen": 2715811840, + "step": 5180 + }, + { + "epoch": 0.7790782221179782, + "grad_norm": 0.7800982594490051, + "learning_rate": 1.0147277699035774e-05, + "loss": 2.0871, + "num_input_tokens_seen": 2721054720, + "step": 5190 + }, + { + "epoch": 0.7805793362261053, + "grad_norm": 0.6889191269874573, + "learning_rate": 1.0137516005868044e-05, + "loss": 2.0951, + "num_input_tokens_seen": 2726297600, + "step": 5200 + }, + { + "epoch": 0.7805793362261053, + "eval_accuracy": 0.5728083028083029, + "eval_loss": 2.090898036956787, + "eval_runtime": 93.559, + "eval_samples_per_second": 3.207, + "eval_steps_per_second": 0.802, + "num_input_tokens_seen": 2726297600, + "step": 5200 + }, + { + "epoch": 0.7820804503342325, + "grad_norm": 0.8273455500602722, + "learning_rate": 1.0127782430886362e-05, + "loss": 2.0874, + "num_input_tokens_seen": 2731540480, + "step": 5210 + }, + { + "epoch": 0.7835815644423596, + "grad_norm": 0.7020549178123474, + "learning_rate": 1.0118076839360487e-05, + "loss": 2.1437, + "num_input_tokens_seen": 2736783360, + "step": 5220 + }, + { + "epoch": 0.7850826785504866, + "grad_norm": 0.6244192719459534, + "learning_rate": 1.0108399097462245e-05, + "loss": 2.114, + "num_input_tokens_seen": 2742026240, + "step": 5230 + }, + { + "epoch": 0.7865837926586138, + "grad_norm": 0.6874685883522034, + "learning_rate": 1.0098749072257785e-05, + "loss": 2.0951, + "num_input_tokens_seen": 2747269120, + "step": 5240 + }, + { + "epoch": 0.7880849067667409, + "grad_norm": 0.9486780762672424, + "learning_rate": 1.0089126631699902e-05, + "loss": 2.1273, + "num_input_tokens_seen": 2752512000, + "step": 5250 + }, + { + "epoch": 0.7895860208748681, + "grad_norm": 0.8076638579368591, + "learning_rate": 1.0079531644620446e-05, + "loss": 2.0947, + "num_input_tokens_seen": 2757754880, + "step": 5260 + }, + { + "epoch": 0.7910871349829952, + "grad_norm": 0.7314202785491943, + "learning_rate": 1.0069963980722823e-05, + "loss": 2.1266, + "num_input_tokens_seen": 2762997760, + "step": 5270 + }, + { + "epoch": 0.7925882490911224, + "grad_norm": 0.6520930528640747, + "learning_rate": 1.0060423510574553e-05, + "loss": 2.1178, + "num_input_tokens_seen": 2768240640, + "step": 5280 + }, + { + "epoch": 0.7940893631992495, + "grad_norm": 0.7945283651351929, + "learning_rate": 1.0050910105599924e-05, + "loss": 2.1345, + "num_input_tokens_seen": 2773483520, + "step": 5290 + }, + { + "epoch": 0.7955904773073765, + "grad_norm": 0.9129828810691833, + "learning_rate": 1.0041423638072707e-05, + "loss": 2.1111, + "num_input_tokens_seen": 2778726400, + "step": 5300 + }, + { + "epoch": 0.7970915914155037, + "grad_norm": 0.7378903031349182, + "learning_rate": 1.0031963981108954e-05, + "loss": 2.1075, + "num_input_tokens_seen": 2783969280, + "step": 5310 + }, + { + "epoch": 0.7985927055236308, + "grad_norm": 0.8121985197067261, + "learning_rate": 1.0022531008659884e-05, + "loss": 2.105, + "num_input_tokens_seen": 2789212160, + "step": 5320 + }, + { + "epoch": 0.800093819631758, + "grad_norm": 0.9001929759979248, + "learning_rate": 1.0013124595504802e-05, + "loss": 2.1389, + "num_input_tokens_seen": 2794455040, + "step": 5330 + }, + { + "epoch": 0.8015949337398851, + "grad_norm": 0.6680558323860168, + "learning_rate": 1.0003744617244145e-05, + "loss": 2.1057, + "num_input_tokens_seen": 2799697920, + "step": 5340 + }, + { + "epoch": 0.8030960478480123, + "grad_norm": 0.7327044606208801, + "learning_rate": 9.994390950292557e-06, + "loss": 2.1291, + "num_input_tokens_seen": 2804940800, + "step": 5350 + }, + { + "epoch": 0.8045971619561393, + "grad_norm": 0.8424970507621765, + "learning_rate": 9.985063471872047e-06, + "loss": 2.1063, + "num_input_tokens_seen": 2810183680, + "step": 5360 + }, + { + "epoch": 0.8060982760642664, + "grad_norm": 0.8721758127212524, + "learning_rate": 9.975762060005233e-06, + "loss": 2.1093, + "num_input_tokens_seen": 2815426560, + "step": 5370 + }, + { + "epoch": 0.8075993901723936, + "grad_norm": 0.8354282379150391, + "learning_rate": 9.966486593508638e-06, + "loss": 2.1093, + "num_input_tokens_seen": 2820669440, + "step": 5380 + }, + { + "epoch": 0.8091005042805207, + "grad_norm": 0.7554700970649719, + "learning_rate": 9.957236951986044e-06, + "loss": 2.1185, + "num_input_tokens_seen": 2825912320, + "step": 5390 + }, + { + "epoch": 0.8106016183886479, + "grad_norm": 0.6626455783843994, + "learning_rate": 9.948013015821951e-06, + "loss": 2.0802, + "num_input_tokens_seen": 2831155200, + "step": 5400 + }, + { + "epoch": 0.812102732496775, + "grad_norm": 0.7216958403587341, + "learning_rate": 9.938814666175072e-06, + "loss": 2.0879, + "num_input_tokens_seen": 2836398080, + "step": 5410 + }, + { + "epoch": 0.813603846604902, + "grad_norm": 0.7840401530265808, + "learning_rate": 9.929641784971893e-06, + "loss": 2.1033, + "num_input_tokens_seen": 2841640960, + "step": 5420 + }, + { + "epoch": 0.8151049607130292, + "grad_norm": 0.8764855265617371, + "learning_rate": 9.920494254900313e-06, + "loss": 2.0822, + "num_input_tokens_seen": 2846883840, + "step": 5430 + }, + { + "epoch": 0.8166060748211563, + "grad_norm": 0.8843982219696045, + "learning_rate": 9.911371959403354e-06, + "loss": 2.1132, + "num_input_tokens_seen": 2852126720, + "step": 5440 + }, + { + "epoch": 0.8181071889292835, + "grad_norm": 0.7158339023590088, + "learning_rate": 9.902274782672901e-06, + "loss": 2.1019, + "num_input_tokens_seen": 2857369600, + "step": 5450 + }, + { + "epoch": 0.8196083030374106, + "grad_norm": 0.6980221271514893, + "learning_rate": 9.893202609643547e-06, + "loss": 2.0935, + "num_input_tokens_seen": 2862612480, + "step": 5460 + }, + { + "epoch": 0.8211094171455378, + "grad_norm": 0.6455651521682739, + "learning_rate": 9.88415532598647e-06, + "loss": 2.1212, + "num_input_tokens_seen": 2867855360, + "step": 5470 + }, + { + "epoch": 0.8226105312536648, + "grad_norm": 0.6872067451477051, + "learning_rate": 9.875132818103386e-06, + "loss": 2.0931, + "num_input_tokens_seen": 2873098240, + "step": 5480 + }, + { + "epoch": 0.8241116453617919, + "grad_norm": 0.8125963807106018, + "learning_rate": 9.86613497312055e-06, + "loss": 2.1199, + "num_input_tokens_seen": 2878341120, + "step": 5490 + }, + { + "epoch": 0.8256127594699191, + "grad_norm": 0.840162992477417, + "learning_rate": 9.857161678882842e-06, + "loss": 2.131, + "num_input_tokens_seen": 2883584000, + "step": 5500 + }, + { + "epoch": 0.8271138735780462, + "grad_norm": 0.774443507194519, + "learning_rate": 9.848212823947872e-06, + "loss": 2.0972, + "num_input_tokens_seen": 2888826880, + "step": 5510 + }, + { + "epoch": 0.8286149876861734, + "grad_norm": 0.7709413766860962, + "learning_rate": 9.839288297580194e-06, + "loss": 2.1215, + "num_input_tokens_seen": 2894069760, + "step": 5520 + }, + { + "epoch": 0.8301161017943005, + "grad_norm": 0.738002598285675, + "learning_rate": 9.830387989745525e-06, + "loss": 2.0852, + "num_input_tokens_seen": 2899312640, + "step": 5530 + }, + { + "epoch": 0.8316172159024275, + "grad_norm": 0.803095817565918, + "learning_rate": 9.821511791105069e-06, + "loss": 2.1179, + "num_input_tokens_seen": 2904555520, + "step": 5540 + }, + { + "epoch": 0.8331183300105547, + "grad_norm": 0.808690071105957, + "learning_rate": 9.812659593009853e-06, + "loss": 2.1157, + "num_input_tokens_seen": 2909798400, + "step": 5550 + }, + { + "epoch": 0.8346194441186818, + "grad_norm": 0.7011695504188538, + "learning_rate": 9.803831287495165e-06, + "loss": 2.1104, + "num_input_tokens_seen": 2915041280, + "step": 5560 + }, + { + "epoch": 0.836120558226809, + "grad_norm": 0.8309350609779358, + "learning_rate": 9.795026767275002e-06, + "loss": 2.0879, + "num_input_tokens_seen": 2920284160, + "step": 5570 + }, + { + "epoch": 0.8376216723349361, + "grad_norm": 0.7645972371101379, + "learning_rate": 9.78624592573661e-06, + "loss": 2.083, + "num_input_tokens_seen": 2925527040, + "step": 5580 + }, + { + "epoch": 0.8391227864430633, + "grad_norm": 0.770995557308197, + "learning_rate": 9.777488656935059e-06, + "loss": 2.1311, + "num_input_tokens_seen": 2930769920, + "step": 5590 + }, + { + "epoch": 0.8406239005511903, + "grad_norm": 0.7518852949142456, + "learning_rate": 9.768754855587863e-06, + "loss": 2.0951, + "num_input_tokens_seen": 2936012800, + "step": 5600 + }, + { + "epoch": 0.8406239005511903, + "eval_accuracy": 0.5732901912901913, + "eval_loss": 2.0876457691192627, + "eval_runtime": 94.3441, + "eval_samples_per_second": 3.18, + "eval_steps_per_second": 0.795, + "num_input_tokens_seen": 2936012800, + "step": 5600 + }, + { + "epoch": 0.8421250146593174, + "grad_norm": 0.8328472375869751, + "learning_rate": 9.760044417069675e-06, + "loss": 2.1032, + "num_input_tokens_seen": 2941255680, + "step": 5610 + }, + { + "epoch": 0.8436261287674446, + "grad_norm": 0.803054690361023, + "learning_rate": 9.75135723740702e-06, + "loss": 2.0893, + "num_input_tokens_seen": 2946498560, + "step": 5620 + }, + { + "epoch": 0.8451272428755717, + "grad_norm": 1.0215765237808228, + "learning_rate": 9.742693213273084e-06, + "loss": 2.091, + "num_input_tokens_seen": 2951741440, + "step": 5630 + }, + { + "epoch": 0.8466283569836989, + "grad_norm": 0.669141948223114, + "learning_rate": 9.734052241982545e-06, + "loss": 2.1025, + "num_input_tokens_seen": 2956984320, + "step": 5640 + }, + { + "epoch": 0.848129471091826, + "grad_norm": 0.7541179656982422, + "learning_rate": 9.725434221486473e-06, + "loss": 2.1037, + "num_input_tokens_seen": 2962227200, + "step": 5650 + }, + { + "epoch": 0.849630585199953, + "grad_norm": 0.8103038668632507, + "learning_rate": 9.716839050367259e-06, + "loss": 2.1022, + "num_input_tokens_seen": 2967470080, + "step": 5660 + }, + { + "epoch": 0.8511316993080802, + "grad_norm": 0.7163017988204956, + "learning_rate": 9.7082666278336e-06, + "loss": 2.1019, + "num_input_tokens_seen": 2972712960, + "step": 5670 + }, + { + "epoch": 0.8526328134162073, + "grad_norm": 0.7300000786781311, + "learning_rate": 9.699716853715554e-06, + "loss": 2.118, + "num_input_tokens_seen": 2977955840, + "step": 5680 + }, + { + "epoch": 0.8541339275243345, + "grad_norm": 0.7246140241622925, + "learning_rate": 9.691189628459591e-06, + "loss": 2.1528, + "num_input_tokens_seen": 2983198720, + "step": 5690 + }, + { + "epoch": 0.8556350416324616, + "grad_norm": 0.8242561221122742, + "learning_rate": 9.682684853123759e-06, + "loss": 2.1482, + "num_input_tokens_seen": 2988441600, + "step": 5700 + }, + { + "epoch": 0.8571361557405888, + "grad_norm": 0.7430049777030945, + "learning_rate": 9.674202429372843e-06, + "loss": 2.0871, + "num_input_tokens_seen": 2993684480, + "step": 5710 + }, + { + "epoch": 0.8586372698487158, + "grad_norm": 0.7430338263511658, + "learning_rate": 9.665742259473581e-06, + "loss": 2.0706, + "num_input_tokens_seen": 2998927360, + "step": 5720 + }, + { + "epoch": 0.860138383956843, + "grad_norm": 0.7350336313247681, + "learning_rate": 9.657304246289963e-06, + "loss": 2.1035, + "num_input_tokens_seen": 3004170240, + "step": 5730 + }, + { + "epoch": 0.8616394980649701, + "grad_norm": 0.5950428247451782, + "learning_rate": 9.64888829327852e-06, + "loss": 2.1055, + "num_input_tokens_seen": 3009413120, + "step": 5740 + }, + { + "epoch": 0.8631406121730972, + "grad_norm": 0.7408620715141296, + "learning_rate": 9.640494304483698e-06, + "loss": 2.1078, + "num_input_tokens_seen": 3014656000, + "step": 5750 + }, + { + "epoch": 0.8646417262812244, + "grad_norm": 0.8998221158981323, + "learning_rate": 9.63212218453326e-06, + "loss": 2.1249, + "num_input_tokens_seen": 3019898880, + "step": 5760 + }, + { + "epoch": 0.8661428403893515, + "grad_norm": 0.8028159141540527, + "learning_rate": 9.623771838633735e-06, + "loss": 2.0808, + "num_input_tokens_seen": 3025141760, + "step": 5770 + }, + { + "epoch": 0.8676439544974786, + "grad_norm": 0.7718355655670166, + "learning_rate": 9.615443172565908e-06, + "loss": 2.1335, + "num_input_tokens_seen": 3030384640, + "step": 5780 + }, + { + "epoch": 0.8691450686056057, + "grad_norm": 0.6967669129371643, + "learning_rate": 9.607136092680355e-06, + "loss": 2.1056, + "num_input_tokens_seen": 3035627520, + "step": 5790 + }, + { + "epoch": 0.8706461827137328, + "grad_norm": 0.8609460592269897, + "learning_rate": 9.598850505893025e-06, + "loss": 2.1219, + "num_input_tokens_seen": 3040870400, + "step": 5800 + }, + { + "epoch": 0.87214729682186, + "grad_norm": 0.7492549419403076, + "learning_rate": 9.590586319680857e-06, + "loss": 2.0746, + "num_input_tokens_seen": 3046113280, + "step": 5810 + }, + { + "epoch": 0.8736484109299871, + "grad_norm": 0.7335068583488464, + "learning_rate": 9.582343442077434e-06, + "loss": 2.1379, + "num_input_tokens_seen": 3051356160, + "step": 5820 + }, + { + "epoch": 0.8751495250381143, + "grad_norm": 0.6927693486213684, + "learning_rate": 9.574121781668698e-06, + "loss": 2.1052, + "num_input_tokens_seen": 3056599040, + "step": 5830 + }, + { + "epoch": 0.8766506391462413, + "grad_norm": 0.7068751454353333, + "learning_rate": 9.565921247588678e-06, + "loss": 2.0933, + "num_input_tokens_seen": 3061841920, + "step": 5840 + }, + { + "epoch": 0.8781517532543685, + "grad_norm": 0.7502044439315796, + "learning_rate": 9.557741749515278e-06, + "loss": 2.0825, + "num_input_tokens_seen": 3067084800, + "step": 5850 + }, + { + "epoch": 0.8796528673624956, + "grad_norm": 0.6573644876480103, + "learning_rate": 9.549583197666103e-06, + "loss": 2.1222, + "num_input_tokens_seen": 3072327680, + "step": 5860 + }, + { + "epoch": 0.8811539814706227, + "grad_norm": 0.8025186657905579, + "learning_rate": 9.541445502794315e-06, + "loss": 2.0886, + "num_input_tokens_seen": 3077570560, + "step": 5870 + }, + { + "epoch": 0.8826550955787499, + "grad_norm": 0.8089122176170349, + "learning_rate": 9.533328576184532e-06, + "loss": 2.1055, + "num_input_tokens_seen": 3082813440, + "step": 5880 + }, + { + "epoch": 0.884156209686877, + "grad_norm": 0.8042004704475403, + "learning_rate": 9.525232329648768e-06, + "loss": 2.1178, + "num_input_tokens_seen": 3088056320, + "step": 5890 + }, + { + "epoch": 0.8856573237950041, + "grad_norm": 0.8820722699165344, + "learning_rate": 9.517156675522405e-06, + "loss": 2.0989, + "num_input_tokens_seen": 3093299200, + "step": 5900 + }, + { + "epoch": 0.8871584379031312, + "grad_norm": 0.731745719909668, + "learning_rate": 9.509101526660216e-06, + "loss": 2.1112, + "num_input_tokens_seen": 3098542080, + "step": 5910 + }, + { + "epoch": 0.8886595520112583, + "grad_norm": 0.8687134981155396, + "learning_rate": 9.501066796432403e-06, + "loss": 2.0962, + "num_input_tokens_seen": 3103784960, + "step": 5920 + }, + { + "epoch": 0.8901606661193855, + "grad_norm": 0.7933871150016785, + "learning_rate": 9.493052398720693e-06, + "loss": 2.1376, + "num_input_tokens_seen": 3109027840, + "step": 5930 + }, + { + "epoch": 0.8916617802275126, + "grad_norm": 0.8090782165527344, + "learning_rate": 9.485058247914453e-06, + "loss": 2.1105, + "num_input_tokens_seen": 3114270720, + "step": 5940 + }, + { + "epoch": 0.8931628943356398, + "grad_norm": 0.7016152739524841, + "learning_rate": 9.477084258906861e-06, + "loss": 2.0717, + "num_input_tokens_seen": 3119513600, + "step": 5950 + }, + { + "epoch": 0.8946640084437668, + "grad_norm": 0.7488442659378052, + "learning_rate": 9.469130347091085e-06, + "loss": 2.1083, + "num_input_tokens_seen": 3124756480, + "step": 5960 + }, + { + "epoch": 0.896165122551894, + "grad_norm": 0.8162315487861633, + "learning_rate": 9.461196428356533e-06, + "loss": 2.0822, + "num_input_tokens_seen": 3129999360, + "step": 5970 + }, + { + "epoch": 0.8976662366600211, + "grad_norm": 0.6994791030883789, + "learning_rate": 9.453282419085091e-06, + "loss": 2.1665, + "num_input_tokens_seen": 3135242240, + "step": 5980 + }, + { + "epoch": 0.8991673507681482, + "grad_norm": 0.7622292041778564, + "learning_rate": 9.445388236147448e-06, + "loss": 2.0986, + "num_input_tokens_seen": 3140485120, + "step": 5990 + }, + { + "epoch": 0.9006684648762754, + "grad_norm": 0.7731397747993469, + "learning_rate": 9.437513796899408e-06, + "loss": 2.1335, + "num_input_tokens_seen": 3145728000, + "step": 6000 + }, + { + "epoch": 0.9006684648762754, + "eval_accuracy": 0.5739422059422059, + "eval_loss": 2.0838024616241455, + "eval_runtime": 93.3731, + "eval_samples_per_second": 3.213, + "eval_steps_per_second": 0.803, + "num_input_tokens_seen": 3145728000, + "step": 6000 + }, + { + "epoch": 0.9021695789844025, + "grad_norm": 0.7499359250068665, + "learning_rate": 9.429659019178268e-06, + "loss": 2.1091, + "num_input_tokens_seen": 3150970880, + "step": 6010 + }, + { + "epoch": 0.9036706930925296, + "grad_norm": 1.0717604160308838, + "learning_rate": 9.421823821299214e-06, + "loss": 2.0924, + "num_input_tokens_seen": 3156213760, + "step": 6020 + }, + { + "epoch": 0.9051718072006567, + "grad_norm": 0.8998821973800659, + "learning_rate": 9.414008122051756e-06, + "loss": 2.108, + "num_input_tokens_seen": 3161456640, + "step": 6030 + }, + { + "epoch": 0.9066729213087839, + "grad_norm": 0.9980584383010864, + "learning_rate": 9.406211840696178e-06, + "loss": 2.0913, + "num_input_tokens_seen": 3166699520, + "step": 6040 + }, + { + "epoch": 0.908174035416911, + "grad_norm": 0.8425881862640381, + "learning_rate": 9.398434896960062e-06, + "loss": 2.1174, + "num_input_tokens_seen": 3171942400, + "step": 6050 + }, + { + "epoch": 0.9096751495250381, + "grad_norm": 0.7694622874259949, + "learning_rate": 9.390677211034795e-06, + "loss": 2.1214, + "num_input_tokens_seen": 3177185280, + "step": 6060 + }, + { + "epoch": 0.9111762636331653, + "grad_norm": 0.8509136438369751, + "learning_rate": 9.382938703572126e-06, + "loss": 2.1062, + "num_input_tokens_seen": 3182428160, + "step": 6070 + }, + { + "epoch": 0.9126773777412923, + "grad_norm": 0.688194990158081, + "learning_rate": 9.375219295680784e-06, + "loss": 2.0959, + "num_input_tokens_seen": 3187671040, + "step": 6080 + }, + { + "epoch": 0.9141784918494195, + "grad_norm": 0.9447828531265259, + "learning_rate": 9.367518908923069e-06, + "loss": 2.1215, + "num_input_tokens_seen": 3192913920, + "step": 6090 + }, + { + "epoch": 0.9156796059575466, + "grad_norm": 0.7632228136062622, + "learning_rate": 9.35983746531152e-06, + "loss": 2.0893, + "num_input_tokens_seen": 3198156800, + "step": 6100 + }, + { + "epoch": 0.9171807200656737, + "grad_norm": 0.6607562899589539, + "learning_rate": 9.352174887305604e-06, + "loss": 2.1476, + "num_input_tokens_seen": 3203399680, + "step": 6110 + }, + { + "epoch": 0.9186818341738009, + "grad_norm": 0.7273072004318237, + "learning_rate": 9.344531097808414e-06, + "loss": 2.0986, + "num_input_tokens_seen": 3208642560, + "step": 6120 + }, + { + "epoch": 0.920182948281928, + "grad_norm": 0.7447370290756226, + "learning_rate": 9.336906020163414e-06, + "loss": 2.1146, + "num_input_tokens_seen": 3213885440, + "step": 6130 + }, + { + "epoch": 0.9216840623900551, + "grad_norm": 0.6748001575469971, + "learning_rate": 9.329299578151221e-06, + "loss": 2.0587, + "num_input_tokens_seen": 3219128320, + "step": 6140 + }, + { + "epoch": 0.9231851764981822, + "grad_norm": 0.874092698097229, + "learning_rate": 9.321711695986389e-06, + "loss": 2.1088, + "num_input_tokens_seen": 3224371200, + "step": 6150 + }, + { + "epoch": 0.9246862906063094, + "grad_norm": 0.7109676003456116, + "learning_rate": 9.314142298314256e-06, + "loss": 2.1308, + "num_input_tokens_seen": 3229614080, + "step": 6160 + }, + { + "epoch": 0.9261874047144365, + "grad_norm": 0.9448009133338928, + "learning_rate": 9.306591310207784e-06, + "loss": 2.1008, + "num_input_tokens_seen": 3234856960, + "step": 6170 + }, + { + "epoch": 0.9276885188225636, + "grad_norm": 1.0548337697982788, + "learning_rate": 9.29905865716445e-06, + "loss": 2.1113, + "num_input_tokens_seen": 3240099840, + "step": 6180 + }, + { + "epoch": 0.9291896329306908, + "grad_norm": 0.8548759818077087, + "learning_rate": 9.291544265103168e-06, + "loss": 2.0949, + "num_input_tokens_seen": 3245342720, + "step": 6190 + }, + { + "epoch": 0.9306907470388178, + "grad_norm": 0.7196878790855408, + "learning_rate": 9.284048060361212e-06, + "loss": 2.1348, + "num_input_tokens_seen": 3250585600, + "step": 6200 + }, + { + "epoch": 0.932191861146945, + "grad_norm": 0.970755934715271, + "learning_rate": 9.276569969691194e-06, + "loss": 2.1198, + "num_input_tokens_seen": 3255828480, + "step": 6210 + }, + { + "epoch": 0.9336929752550721, + "grad_norm": 0.7153322100639343, + "learning_rate": 9.26910992025806e-06, + "loss": 2.1145, + "num_input_tokens_seen": 3261071360, + "step": 6220 + }, + { + "epoch": 0.9351940893631993, + "grad_norm": 0.9428135752677917, + "learning_rate": 9.2616678396361e-06, + "loss": 2.0888, + "num_input_tokens_seen": 3266314240, + "step": 6230 + }, + { + "epoch": 0.9366952034713264, + "grad_norm": 0.8652244806289673, + "learning_rate": 9.254243655806003e-06, + "loss": 2.0739, + "num_input_tokens_seen": 3271557120, + "step": 6240 + }, + { + "epoch": 0.9381963175794535, + "grad_norm": 1.0319374799728394, + "learning_rate": 9.24683729715193e-06, + "loss": 2.127, + "num_input_tokens_seen": 3276800000, + "step": 6250 + }, + { + "epoch": 0.9396974316875806, + "grad_norm": 1.1255468130111694, + "learning_rate": 9.239448692458609e-06, + "loss": 2.1033, + "num_input_tokens_seen": 3282042880, + "step": 6260 + }, + { + "epoch": 0.9411985457957077, + "grad_norm": 0.7438716888427734, + "learning_rate": 9.232077770908458e-06, + "loss": 2.1018, + "num_input_tokens_seen": 3287285760, + "step": 6270 + }, + { + "epoch": 0.9426996599038349, + "grad_norm": 0.7200952768325806, + "learning_rate": 9.22472446207874e-06, + "loss": 2.1389, + "num_input_tokens_seen": 3292528640, + "step": 6280 + }, + { + "epoch": 0.944200774011962, + "grad_norm": 0.6740290522575378, + "learning_rate": 9.21738869593873e-06, + "loss": 2.0873, + "num_input_tokens_seen": 3297771520, + "step": 6290 + }, + { + "epoch": 0.9457018881200892, + "grad_norm": 0.7916491031646729, + "learning_rate": 9.210070402846921e-06, + "loss": 2.0943, + "num_input_tokens_seen": 3303014400, + "step": 6300 + }, + { + "epoch": 0.9472030022282163, + "grad_norm": 0.9598779678344727, + "learning_rate": 9.202769513548237e-06, + "loss": 2.0996, + "num_input_tokens_seen": 3308257280, + "step": 6310 + }, + { + "epoch": 0.9487041163363433, + "grad_norm": 0.8254228234291077, + "learning_rate": 9.195485959171296e-06, + "loss": 2.0863, + "num_input_tokens_seen": 3313500160, + "step": 6320 + }, + { + "epoch": 0.9502052304444705, + "grad_norm": 0.6629865169525146, + "learning_rate": 9.188219671225665e-06, + "loss": 2.0822, + "num_input_tokens_seen": 3318743040, + "step": 6330 + }, + { + "epoch": 0.9517063445525976, + "grad_norm": 0.7633962035179138, + "learning_rate": 9.180970581599163e-06, + "loss": 2.1215, + "num_input_tokens_seen": 3323985920, + "step": 6340 + }, + { + "epoch": 0.9532074586607248, + "grad_norm": 0.7902926206588745, + "learning_rate": 9.17373862255518e-06, + "loss": 2.1073, + "num_input_tokens_seen": 3329228800, + "step": 6350 + }, + { + "epoch": 0.9547085727688519, + "grad_norm": 0.7041093111038208, + "learning_rate": 9.16652372673002e-06, + "loss": 2.1416, + "num_input_tokens_seen": 3334471680, + "step": 6360 + }, + { + "epoch": 0.956209686876979, + "grad_norm": 0.7147159576416016, + "learning_rate": 9.159325827130255e-06, + "loss": 2.0744, + "num_input_tokens_seen": 3339714560, + "step": 6370 + }, + { + "epoch": 0.9577108009851061, + "grad_norm": 0.7527651786804199, + "learning_rate": 9.152144857130128e-06, + "loss": 2.1002, + "num_input_tokens_seen": 3344957440, + "step": 6380 + }, + { + "epoch": 0.9592119150932332, + "grad_norm": 0.7740973830223083, + "learning_rate": 9.144980750468947e-06, + "loss": 2.0972, + "num_input_tokens_seen": 3350200320, + "step": 6390 + }, + { + "epoch": 0.9607130292013604, + "grad_norm": 0.8114108443260193, + "learning_rate": 9.137833441248534e-06, + "loss": 2.0731, + "num_input_tokens_seen": 3355443200, + "step": 6400 + }, + { + "epoch": 0.9607130292013604, + "eval_accuracy": 0.5744249084249085, + "eval_loss": 2.080249071121216, + "eval_runtime": 93.7101, + "eval_samples_per_second": 3.201, + "eval_steps_per_second": 0.8, + "num_input_tokens_seen": 3355443200, + "step": 6400 + }, + { + "epoch": 0.9622141433094875, + "grad_norm": 0.7784215211868286, + "learning_rate": 9.130702863930661e-06, + "loss": 2.0991, + "num_input_tokens_seen": 3360686080, + "step": 6410 + }, + { + "epoch": 0.9637152574176147, + "grad_norm": 0.883834958076477, + "learning_rate": 9.123588953334543e-06, + "loss": 2.127, + "num_input_tokens_seen": 3365928960, + "step": 6420 + }, + { + "epoch": 0.9652163715257418, + "grad_norm": 0.6606909036636353, + "learning_rate": 9.116491644634323e-06, + "loss": 2.0898, + "num_input_tokens_seen": 3371171840, + "step": 6430 + }, + { + "epoch": 0.9667174856338688, + "grad_norm": 0.8438331484794617, + "learning_rate": 9.109410873356595e-06, + "loss": 2.097, + "num_input_tokens_seen": 3376414720, + "step": 6440 + }, + { + "epoch": 0.968218599741996, + "grad_norm": 0.7696127891540527, + "learning_rate": 9.102346575377945e-06, + "loss": 2.0882, + "num_input_tokens_seen": 3381657600, + "step": 6450 + }, + { + "epoch": 0.9697197138501231, + "grad_norm": 0.7825368642807007, + "learning_rate": 9.095298686922513e-06, + "loss": 2.099, + "num_input_tokens_seen": 3386900480, + "step": 6460 + }, + { + "epoch": 0.9712208279582503, + "grad_norm": 1.1647248268127441, + "learning_rate": 9.088267144559575e-06, + "loss": 2.0652, + "num_input_tokens_seen": 3392143360, + "step": 6470 + }, + { + "epoch": 0.9727219420663774, + "grad_norm": 0.8124563694000244, + "learning_rate": 9.081251885201133e-06, + "loss": 2.0914, + "num_input_tokens_seen": 3397386240, + "step": 6480 + }, + { + "epoch": 0.9742230561745046, + "grad_norm": 0.8535528779029846, + "learning_rate": 9.074252846099567e-06, + "loss": 2.0706, + "num_input_tokens_seen": 3402629120, + "step": 6490 + }, + { + "epoch": 0.9757241702826316, + "grad_norm": 0.8512585759162903, + "learning_rate": 9.067269964845241e-06, + "loss": 2.1627, + "num_input_tokens_seen": 3407872000, + "step": 6500 + }, + { + "epoch": 0.9772252843907587, + "grad_norm": 0.7466162443161011, + "learning_rate": 9.060303179364195e-06, + "loss": 2.0776, + "num_input_tokens_seen": 3413114880, + "step": 6510 + }, + { + "epoch": 0.9787263984988859, + "grad_norm": 0.8283933997154236, + "learning_rate": 9.053352427915811e-06, + "loss": 2.0799, + "num_input_tokens_seen": 3418357760, + "step": 6520 + }, + { + "epoch": 0.980227512607013, + "grad_norm": 0.6737175583839417, + "learning_rate": 9.04641764909052e-06, + "loss": 2.0664, + "num_input_tokens_seen": 3423600640, + "step": 6530 + }, + { + "epoch": 0.9817286267151402, + "grad_norm": 0.6306325197219849, + "learning_rate": 9.039498781807522e-06, + "loss": 2.0973, + "num_input_tokens_seen": 3428843520, + "step": 6540 + }, + { + "epoch": 0.9832297408232673, + "grad_norm": 0.7176161408424377, + "learning_rate": 9.032595765312539e-06, + "loss": 2.0812, + "num_input_tokens_seen": 3434086400, + "step": 6550 + }, + { + "epoch": 0.9847308549313943, + "grad_norm": 0.7962211966514587, + "learning_rate": 9.025708539175545e-06, + "loss": 2.084, + "num_input_tokens_seen": 3439329280, + "step": 6560 + }, + { + "epoch": 0.9862319690395215, + "grad_norm": 0.9422264099121094, + "learning_rate": 9.018837043288575e-06, + "loss": 2.1091, + "num_input_tokens_seen": 3444572160, + "step": 6570 + }, + { + "epoch": 0.9877330831476486, + "grad_norm": 0.7303385734558105, + "learning_rate": 9.011981217863507e-06, + "loss": 2.113, + "num_input_tokens_seen": 3449815040, + "step": 6580 + }, + { + "epoch": 0.9892341972557758, + "grad_norm": 0.8498770594596863, + "learning_rate": 9.005141003429877e-06, + "loss": 2.1057, + "num_input_tokens_seen": 3455057920, + "step": 6590 + }, + { + "epoch": 0.9907353113639029, + "grad_norm": 0.945923924446106, + "learning_rate": 8.99831634083271e-06, + "loss": 2.0949, + "num_input_tokens_seen": 3460300800, + "step": 6600 + }, + { + "epoch": 0.9922364254720301, + "grad_norm": 0.8327839374542236, + "learning_rate": 8.991507171230386e-06, + "loss": 2.1105, + "num_input_tokens_seen": 3465543680, + "step": 6610 + }, + { + "epoch": 0.9937375395801572, + "grad_norm": 0.8914210200309753, + "learning_rate": 8.98471343609249e-06, + "loss": 2.1146, + "num_input_tokens_seen": 3470786560, + "step": 6620 + }, + { + "epoch": 0.9952386536882842, + "grad_norm": 0.8601353168487549, + "learning_rate": 8.977935077197712e-06, + "loss": 2.1059, + "num_input_tokens_seen": 3476029440, + "step": 6630 + }, + { + "epoch": 0.9967397677964114, + "grad_norm": 0.8136289715766907, + "learning_rate": 8.971172036631744e-06, + "loss": 2.0737, + "num_input_tokens_seen": 3481272320, + "step": 6640 + }, + { + "epoch": 0.9982408819045385, + "grad_norm": 0.980107843875885, + "learning_rate": 8.964424256785211e-06, + "loss": 2.0816, + "num_input_tokens_seen": 3486515200, + "step": 6650 + }, + { + "epoch": 0.9997419960126657, + "grad_norm": 0.812223494052887, + "learning_rate": 8.9576916803516e-06, + "loss": 2.0981, + "num_input_tokens_seen": 3491758080, + "step": 6660 + }, + { + "epoch": 0.9998921074234783, + "num_input_tokens_seen": 3492282368, + "step": 6661, + "total_flos": 4.5723602603621745e+18, + "train_loss": 2.154333936901462, + "train_runtime": 99522.0612, + "train_samples_per_second": 8.568, + "train_steps_per_second": 0.067 + } + ], + "logging_steps": 10, + "max_steps": 6661, + "num_input_tokens_seen": 3492282368, + "num_train_epochs": 1, + "save_steps": 200, + "total_flos": 4.5723602603621745e+18, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}