{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.526447931526391, "eval_steps": 500, "global_step": 79326, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.028530670470756064, "grad_norm": 4.224736213684082, "learning_rate": 9.936968963517636e-05, "loss": 7.9606, "step": 500 }, { "epoch": 0.05706134094151213, "grad_norm": 3.775930881500244, "learning_rate": 9.873937927035273e-05, "loss": 7.4609, "step": 1000 }, { "epoch": 0.08559201141226819, "grad_norm": 4.54003381729126, "learning_rate": 9.810906890552909e-05, "loss": 7.2264, "step": 1500 }, { "epoch": 0.11412268188302425, "grad_norm": 4.762158393859863, "learning_rate": 9.747875854070544e-05, "loss": 7.1154, "step": 2000 }, { "epoch": 0.14265335235378032, "grad_norm": 4.3742995262146, "learning_rate": 9.684844817588181e-05, "loss": 7.0396, "step": 2500 }, { "epoch": 0.17118402282453637, "grad_norm": 6.319996356964111, "learning_rate": 9.621813781105817e-05, "loss": 6.9091, "step": 3000 }, { "epoch": 0.19971469329529243, "grad_norm": 5.134556770324707, "learning_rate": 9.558782744623454e-05, "loss": 6.8436, "step": 3500 }, { "epoch": 0.2282453637660485, "grad_norm": 5.821471691131592, "learning_rate": 9.49575170814109e-05, "loss": 6.7296, "step": 4000 }, { "epoch": 0.25677603423680456, "grad_norm": 5.649785041809082, "learning_rate": 9.432720671658725e-05, "loss": 6.6347, "step": 4500 }, { "epoch": 0.28530670470756064, "grad_norm": 4.85307502746582, "learning_rate": 9.369689635176361e-05, "loss": 6.5734, "step": 5000 }, { "epoch": 0.31383737517831667, "grad_norm": 4.924481391906738, "learning_rate": 9.306658598693997e-05, "loss": 6.4956, "step": 5500 }, { "epoch": 0.34236804564907275, "grad_norm": 5.568727970123291, "learning_rate": 9.243627562211632e-05, "loss": 6.4667, "step": 6000 }, { "epoch": 0.37089871611982883, "grad_norm": 5.307117938995361, "learning_rate": 9.18059652572927e-05, "loss": 6.4108, "step": 6500 }, { "epoch": 0.39942938659058486, "grad_norm": 5.159631729125977, "learning_rate": 9.117565489246905e-05, "loss": 6.3248, "step": 7000 }, { "epoch": 0.42796005706134094, "grad_norm": 5.529886722564697, "learning_rate": 9.054534452764542e-05, "loss": 6.3069, "step": 7500 }, { "epoch": 0.456490727532097, "grad_norm": 9.25552749633789, "learning_rate": 8.991503416282178e-05, "loss": 6.2188, "step": 8000 }, { "epoch": 0.48502139800285304, "grad_norm": 7.253251075744629, "learning_rate": 8.928472379799813e-05, "loss": 6.0356, "step": 8500 }, { "epoch": 0.5135520684736091, "grad_norm": 8.699528694152832, "learning_rate": 8.86544134331745e-05, "loss": 6.0202, "step": 9000 }, { "epoch": 0.5420827389443652, "grad_norm": 7.144530296325684, "learning_rate": 8.802410306835086e-05, "loss": 5.8857, "step": 9500 }, { "epoch": 0.5706134094151213, "grad_norm": 7.254436492919922, "learning_rate": 8.739379270352722e-05, "loss": 5.8363, "step": 10000 }, { "epoch": 0.5991440798858774, "grad_norm": 7.687765121459961, "learning_rate": 8.676348233870359e-05, "loss": 5.7397, "step": 10500 }, { "epoch": 0.6276747503566333, "grad_norm": 6.547056674957275, "learning_rate": 8.613317197387994e-05, "loss": 5.6764, "step": 11000 }, { "epoch": 0.6562054208273894, "grad_norm": 9.949172973632812, "learning_rate": 8.55028616090563e-05, "loss": 5.6305, "step": 11500 }, { "epoch": 0.6847360912981455, "grad_norm": 8.107893943786621, "learning_rate": 8.487255124423267e-05, "loss": 5.5956, "step": 12000 }, { "epoch": 0.7132667617689016, "grad_norm": 6.340846538543701, "learning_rate": 8.424224087940903e-05, "loss": 5.5119, "step": 12500 }, { "epoch": 0.7417974322396577, "grad_norm": 7.98532247543335, "learning_rate": 8.361193051458538e-05, "loss": 5.4685, "step": 13000 }, { "epoch": 0.7703281027104137, "grad_norm": 8.349475860595703, "learning_rate": 8.298162014976175e-05, "loss": 5.4685, "step": 13500 }, { "epoch": 0.7988587731811697, "grad_norm": 8.270821571350098, "learning_rate": 8.23513097849381e-05, "loss": 5.4529, "step": 14000 }, { "epoch": 0.8273894436519258, "grad_norm": 7.967436790466309, "learning_rate": 8.172099942011447e-05, "loss": 5.369, "step": 14500 }, { "epoch": 0.8559201141226819, "grad_norm": 6.520368576049805, "learning_rate": 8.109068905529082e-05, "loss": 5.3417, "step": 15000 }, { "epoch": 0.884450784593438, "grad_norm": 7.0926713943481445, "learning_rate": 8.046037869046718e-05, "loss": 5.3051, "step": 15500 }, { "epoch": 0.912981455064194, "grad_norm": 7.306639194488525, "learning_rate": 7.983006832564355e-05, "loss": 5.275, "step": 16000 }, { "epoch": 0.9415121255349501, "grad_norm": 7.494438171386719, "learning_rate": 7.919975796081991e-05, "loss": 5.2383, "step": 16500 }, { "epoch": 0.9700427960057061, "grad_norm": 7.95282506942749, "learning_rate": 7.856944759599628e-05, "loss": 5.2275, "step": 17000 }, { "epoch": 0.9985734664764622, "grad_norm": 8.143848419189453, "learning_rate": 7.793913723117263e-05, "loss": 5.1984, "step": 17500 }, { "epoch": 1.0, "eval_loss": 5.088438034057617, "eval_runtime": 60.5124, "eval_samples_per_second": 817.683, "eval_steps_per_second": 51.113, "step": 17525 }, { "epoch": 1.0271041369472182, "grad_norm": 7.758360862731934, "learning_rate": 7.730882686634899e-05, "loss": 5.1319, "step": 18000 }, { "epoch": 1.0556348074179742, "grad_norm": 6.7139177322387695, "learning_rate": 7.667851650152536e-05, "loss": 5.1087, "step": 18500 }, { "epoch": 1.0841654778887304, "grad_norm": 6.584017753601074, "learning_rate": 7.604820613670172e-05, "loss": 5.0938, "step": 19000 }, { "epoch": 1.1126961483594864, "grad_norm": 7.792596817016602, "learning_rate": 7.541789577187807e-05, "loss": 5.0448, "step": 19500 }, { "epoch": 1.1412268188302426, "grad_norm": 7.243467807769775, "learning_rate": 7.478758540705444e-05, "loss": 5.0423, "step": 20000 }, { "epoch": 1.1697574893009985, "grad_norm": 8.1492338180542, "learning_rate": 7.41572750422308e-05, "loss": 4.9913, "step": 20500 }, { "epoch": 1.1982881597717547, "grad_norm": 7.754183292388916, "learning_rate": 7.352696467740716e-05, "loss": 4.9983, "step": 21000 }, { "epoch": 1.2268188302425107, "grad_norm": 7.245093822479248, "learning_rate": 7.289665431258353e-05, "loss": 4.9729, "step": 21500 }, { "epoch": 1.2553495007132667, "grad_norm": 7.720768451690674, "learning_rate": 7.226634394775988e-05, "loss": 4.9499, "step": 22000 }, { "epoch": 1.2838801711840229, "grad_norm": 7.034431457519531, "learning_rate": 7.163603358293624e-05, "loss": 4.9368, "step": 22500 }, { "epoch": 1.3124108416547788, "grad_norm": 9.341329574584961, "learning_rate": 7.100572321811261e-05, "loss": 4.8745, "step": 23000 }, { "epoch": 1.340941512125535, "grad_norm": 6.874212265014648, "learning_rate": 7.037541285328895e-05, "loss": 4.877, "step": 23500 }, { "epoch": 1.369472182596291, "grad_norm": 8.832662582397461, "learning_rate": 6.974510248846532e-05, "loss": 4.8839, "step": 24000 }, { "epoch": 1.3980028530670472, "grad_norm": 6.884407997131348, "learning_rate": 6.911479212364168e-05, "loss": 4.8456, "step": 24500 }, { "epoch": 1.4265335235378032, "grad_norm": 8.421597480773926, "learning_rate": 6.848448175881804e-05, "loss": 4.8491, "step": 25000 }, { "epoch": 1.4550641940085591, "grad_norm": 8.932446479797363, "learning_rate": 6.78541713939944e-05, "loss": 4.7936, "step": 25500 }, { "epoch": 1.4835948644793153, "grad_norm": 9.981544494628906, "learning_rate": 6.722386102917076e-05, "loss": 4.816, "step": 26000 }, { "epoch": 1.5121255349500713, "grad_norm": 7.8625922203063965, "learning_rate": 6.659355066434713e-05, "loss": 4.7983, "step": 26500 }, { "epoch": 1.5406562054208273, "grad_norm": 9.113426208496094, "learning_rate": 6.596324029952349e-05, "loss": 4.7438, "step": 27000 }, { "epoch": 1.5691868758915835, "grad_norm": 8.4766206741333, "learning_rate": 6.533292993469985e-05, "loss": 4.7544, "step": 27500 }, { "epoch": 1.5977175463623396, "grad_norm": 9.448740005493164, "learning_rate": 6.470261956987622e-05, "loss": 4.7244, "step": 28000 }, { "epoch": 1.6262482168330956, "grad_norm": 7.334268093109131, "learning_rate": 6.407230920505257e-05, "loss": 4.7225, "step": 28500 }, { "epoch": 1.6547788873038516, "grad_norm": 8.931607246398926, "learning_rate": 6.344199884022893e-05, "loss": 4.6495, "step": 29000 }, { "epoch": 1.6833095577746078, "grad_norm": 7.774843215942383, "learning_rate": 6.28116884754053e-05, "loss": 4.7006, "step": 29500 }, { "epoch": 1.7118402282453637, "grad_norm": 7.911451816558838, "learning_rate": 6.218137811058166e-05, "loss": 4.6525, "step": 30000 }, { "epoch": 1.7403708987161197, "grad_norm": 7.005200386047363, "learning_rate": 6.155106774575801e-05, "loss": 4.6765, "step": 30500 }, { "epoch": 1.768901569186876, "grad_norm": 8.889507293701172, "learning_rate": 6.0920757380934376e-05, "loss": 4.6397, "step": 31000 }, { "epoch": 1.797432239657632, "grad_norm": 8.807517051696777, "learning_rate": 6.029044701611074e-05, "loss": 4.6291, "step": 31500 }, { "epoch": 1.825962910128388, "grad_norm": 10.036596298217773, "learning_rate": 5.96601366512871e-05, "loss": 4.6033, "step": 32000 }, { "epoch": 1.854493580599144, "grad_norm": 8.668290138244629, "learning_rate": 5.902982628646345e-05, "loss": 4.617, "step": 32500 }, { "epoch": 1.8830242510699002, "grad_norm": 7.631191253662109, "learning_rate": 5.8399515921639816e-05, "loss": 4.591, "step": 33000 }, { "epoch": 1.9115549215406562, "grad_norm": 7.901882171630859, "learning_rate": 5.776920555681617e-05, "loss": 4.593, "step": 33500 }, { "epoch": 1.9400855920114122, "grad_norm": 8.088603973388672, "learning_rate": 5.7138895191992536e-05, "loss": 4.575, "step": 34000 }, { "epoch": 1.9686162624821684, "grad_norm": 13.519835472106934, "learning_rate": 5.65085848271689e-05, "loss": 4.5674, "step": 34500 }, { "epoch": 1.9971469329529246, "grad_norm": 7.497794151306152, "learning_rate": 5.587827446234526e-05, "loss": 4.5896, "step": 35000 }, { "epoch": 2.0, "eval_loss": 4.483283996582031, "eval_runtime": 55.5401, "eval_samples_per_second": 890.889, "eval_steps_per_second": 55.69, "step": 35050 }, { "epoch": 2.0256776034236803, "grad_norm": 9.086140632629395, "learning_rate": 5.524796409752162e-05, "loss": 4.4731, "step": 35500 }, { "epoch": 2.0542082738944365, "grad_norm": 7.284142017364502, "learning_rate": 5.461765373269798e-05, "loss": 4.5119, "step": 36000 }, { "epoch": 2.0827389443651927, "grad_norm": 9.4158296585083, "learning_rate": 5.3987343367874346e-05, "loss": 4.4564, "step": 36500 }, { "epoch": 2.1112696148359484, "grad_norm": 10.107216835021973, "learning_rate": 5.33570330030507e-05, "loss": 4.5029, "step": 37000 }, { "epoch": 2.1398002853067046, "grad_norm": 8.506832122802734, "learning_rate": 5.2726722638227066e-05, "loss": 4.4487, "step": 37500 }, { "epoch": 2.168330955777461, "grad_norm": 10.495549201965332, "learning_rate": 5.209641227340343e-05, "loss": 4.433, "step": 38000 }, { "epoch": 2.196861626248217, "grad_norm": 8.59913444519043, "learning_rate": 5.146610190857979e-05, "loss": 4.4377, "step": 38500 }, { "epoch": 2.2253922967189728, "grad_norm": 7.879130840301514, "learning_rate": 5.083579154375615e-05, "loss": 4.3888, "step": 39000 }, { "epoch": 2.253922967189729, "grad_norm": 8.134923934936523, "learning_rate": 5.020548117893251e-05, "loss": 4.4215, "step": 39500 }, { "epoch": 2.282453637660485, "grad_norm": 8.942688941955566, "learning_rate": 4.957517081410887e-05, "loss": 4.4249, "step": 40000 }, { "epoch": 2.310984308131241, "grad_norm": 12.744963645935059, "learning_rate": 4.894486044928523e-05, "loss": 4.3957, "step": 40500 }, { "epoch": 2.339514978601997, "grad_norm": 10.202946662902832, "learning_rate": 4.831455008446159e-05, "loss": 4.4239, "step": 41000 }, { "epoch": 2.3680456490727533, "grad_norm": 7.9722418785095215, "learning_rate": 4.768423971963795e-05, "loss": 4.4196, "step": 41500 }, { "epoch": 2.3965763195435095, "grad_norm": 8.571127891540527, "learning_rate": 4.7053929354814315e-05, "loss": 4.3274, "step": 42000 }, { "epoch": 2.425106990014265, "grad_norm": 9.048084259033203, "learning_rate": 4.642361898999067e-05, "loss": 4.3422, "step": 42500 }, { "epoch": 2.4536376604850214, "grad_norm": 9.029339790344238, "learning_rate": 4.5793308625167035e-05, "loss": 4.3474, "step": 43000 }, { "epoch": 2.4821683309557776, "grad_norm": 8.281103134155273, "learning_rate": 4.51629982603434e-05, "loss": 4.3849, "step": 43500 }, { "epoch": 2.5106990014265333, "grad_norm": 7.619591236114502, "learning_rate": 4.4532687895519755e-05, "loss": 4.3154, "step": 44000 }, { "epoch": 2.5392296718972895, "grad_norm": 10.685304641723633, "learning_rate": 4.390237753069611e-05, "loss": 4.2803, "step": 44500 }, { "epoch": 2.5677603423680457, "grad_norm": 8.622846603393555, "learning_rate": 4.3272067165872475e-05, "loss": 4.2911, "step": 45000 }, { "epoch": 2.596291012838802, "grad_norm": 10.364981651306152, "learning_rate": 4.264175680104884e-05, "loss": 4.3172, "step": 45500 }, { "epoch": 2.6248216833095577, "grad_norm": 6.996392726898193, "learning_rate": 4.20114464362252e-05, "loss": 4.3123, "step": 46000 }, { "epoch": 2.653352353780314, "grad_norm": 9.445304870605469, "learning_rate": 4.138113607140156e-05, "loss": 4.2621, "step": 46500 }, { "epoch": 2.68188302425107, "grad_norm": 9.185279846191406, "learning_rate": 4.075082570657792e-05, "loss": 4.2983, "step": 47000 }, { "epoch": 2.710413694721826, "grad_norm": 7.654629230499268, "learning_rate": 4.0120515341754285e-05, "loss": 4.2994, "step": 47500 }, { "epoch": 2.738944365192582, "grad_norm": 8.687335014343262, "learning_rate": 3.949020497693065e-05, "loss": 4.2667, "step": 48000 }, { "epoch": 2.767475035663338, "grad_norm": 9.457321166992188, "learning_rate": 3.8859894612107e-05, "loss": 4.2709, "step": 48500 }, { "epoch": 2.7960057061340944, "grad_norm": 9.891905784606934, "learning_rate": 3.822958424728336e-05, "loss": 4.217, "step": 49000 }, { "epoch": 2.82453637660485, "grad_norm": 7.958982467651367, "learning_rate": 3.7599273882459725e-05, "loss": 4.219, "step": 49500 }, { "epoch": 2.8530670470756063, "grad_norm": 9.362072944641113, "learning_rate": 3.696896351763609e-05, "loss": 4.2212, "step": 50000 }, { "epoch": 2.881597717546362, "grad_norm": 9.81021499633789, "learning_rate": 3.6338653152812445e-05, "loss": 4.1786, "step": 50500 }, { "epoch": 2.9101283880171183, "grad_norm": 9.377098083496094, "learning_rate": 3.570834278798881e-05, "loss": 4.1874, "step": 51000 }, { "epoch": 2.9386590584878745, "grad_norm": 7.9157562255859375, "learning_rate": 3.507803242316517e-05, "loss": 4.2021, "step": 51500 }, { "epoch": 2.9671897289586306, "grad_norm": 9.986918449401855, "learning_rate": 3.444772205834153e-05, "loss": 4.2068, "step": 52000 }, { "epoch": 2.995720399429387, "grad_norm": 9.590192794799805, "learning_rate": 3.381741169351789e-05, "loss": 4.2007, "step": 52500 }, { "epoch": 3.0, "eval_loss": 4.12980842590332, "eval_runtime": 54.7837, "eval_samples_per_second": 903.188, "eval_steps_per_second": 56.458, "step": 52575 }, { "epoch": 3.0242510699001426, "grad_norm": 10.125175476074219, "learning_rate": 3.3187101328694254e-05, "loss": 4.1689, "step": 53000 }, { "epoch": 3.0527817403708988, "grad_norm": 9.876838684082031, "learning_rate": 3.255679096387061e-05, "loss": 4.1606, "step": 53500 }, { "epoch": 3.081312410841655, "grad_norm": 8.361190795898438, "learning_rate": 3.192648059904697e-05, "loss": 4.1792, "step": 54000 }, { "epoch": 3.1098430813124107, "grad_norm": 11.448074340820312, "learning_rate": 3.129617023422333e-05, "loss": 4.1463, "step": 54500 }, { "epoch": 3.138373751783167, "grad_norm": 9.612030982971191, "learning_rate": 3.0665859869399694e-05, "loss": 4.1013, "step": 55000 }, { "epoch": 3.166904422253923, "grad_norm": 8.74341106414795, "learning_rate": 3.0035549504576054e-05, "loss": 4.1097, "step": 55500 }, { "epoch": 3.195435092724679, "grad_norm": 8.951421737670898, "learning_rate": 2.9405239139752418e-05, "loss": 4.1154, "step": 56000 }, { "epoch": 3.223965763195435, "grad_norm": 9.474740028381348, "learning_rate": 2.8774928774928778e-05, "loss": 4.1132, "step": 56500 }, { "epoch": 3.2524964336661912, "grad_norm": 10.41540241241455, "learning_rate": 2.8144618410105137e-05, "loss": 4.1416, "step": 57000 }, { "epoch": 3.281027104136947, "grad_norm": 10.558377265930176, "learning_rate": 2.75143080452815e-05, "loss": 4.1209, "step": 57500 }, { "epoch": 3.309557774607703, "grad_norm": 8.718255996704102, "learning_rate": 2.6883997680457857e-05, "loss": 4.0549, "step": 58000 }, { "epoch": 3.3380884450784594, "grad_norm": 8.449244499206543, "learning_rate": 2.6253687315634217e-05, "loss": 4.1046, "step": 58500 }, { "epoch": 3.3666191155492156, "grad_norm": 10.529504776000977, "learning_rate": 2.562337695081058e-05, "loss": 4.1058, "step": 59000 }, { "epoch": 3.3951497860199713, "grad_norm": 8.345918655395508, "learning_rate": 2.499306658598694e-05, "loss": 4.0668, "step": 59500 }, { "epoch": 3.4236804564907275, "grad_norm": 10.951416015625, "learning_rate": 2.43627562211633e-05, "loss": 4.1145, "step": 60000 }, { "epoch": 3.4522111269614837, "grad_norm": 8.29075813293457, "learning_rate": 2.3732445856339664e-05, "loss": 4.0765, "step": 60500 }, { "epoch": 3.4807417974322394, "grad_norm": 9.023518562316895, "learning_rate": 2.3102135491516024e-05, "loss": 4.0608, "step": 61000 }, { "epoch": 3.5092724679029956, "grad_norm": 10.263666152954102, "learning_rate": 2.2471825126692384e-05, "loss": 4.0361, "step": 61500 }, { "epoch": 3.537803138373752, "grad_norm": 10.308072090148926, "learning_rate": 2.1841514761868744e-05, "loss": 4.0801, "step": 62000 }, { "epoch": 3.566333808844508, "grad_norm": 11.750104904174805, "learning_rate": 2.1211204397045107e-05, "loss": 4.0166, "step": 62500 }, { "epoch": 3.5948644793152638, "grad_norm": 9.711214065551758, "learning_rate": 2.0580894032221467e-05, "loss": 4.1, "step": 63000 }, { "epoch": 3.62339514978602, "grad_norm": 9.461030006408691, "learning_rate": 1.9950583667397827e-05, "loss": 4.0546, "step": 63500 }, { "epoch": 3.651925820256776, "grad_norm": 9.695231437683105, "learning_rate": 1.9320273302574187e-05, "loss": 3.9823, "step": 64000 }, { "epoch": 3.680456490727532, "grad_norm": 8.181650161743164, "learning_rate": 1.868996293775055e-05, "loss": 4.0467, "step": 64500 }, { "epoch": 3.708987161198288, "grad_norm": 10.602164268493652, "learning_rate": 1.805965257292691e-05, "loss": 4.0006, "step": 65000 }, { "epoch": 3.7375178316690443, "grad_norm": 8.176977157592773, "learning_rate": 1.7429342208103274e-05, "loss": 4.0594, "step": 65500 }, { "epoch": 3.7660485021398005, "grad_norm": 11.314457893371582, "learning_rate": 1.679903184327963e-05, "loss": 3.9929, "step": 66000 }, { "epoch": 3.794579172610556, "grad_norm": 10.995957374572754, "learning_rate": 1.6168721478455993e-05, "loss": 4.0191, "step": 66500 }, { "epoch": 3.8231098430813124, "grad_norm": 8.929314613342285, "learning_rate": 1.5538411113632353e-05, "loss": 4.0284, "step": 67000 }, { "epoch": 3.8516405135520686, "grad_norm": 9.887913703918457, "learning_rate": 1.4908100748808715e-05, "loss": 3.9552, "step": 67500 }, { "epoch": 3.8801711840228243, "grad_norm": 12.043901443481445, "learning_rate": 1.4277790383985077e-05, "loss": 4.0139, "step": 68000 }, { "epoch": 3.9087018544935805, "grad_norm": 8.444652557373047, "learning_rate": 1.3647480019161435e-05, "loss": 3.9746, "step": 68500 }, { "epoch": 3.9372325249643367, "grad_norm": 9.180087089538574, "learning_rate": 1.3017169654337797e-05, "loss": 3.9663, "step": 69000 }, { "epoch": 3.965763195435093, "grad_norm": 11.611577987670898, "learning_rate": 1.2386859289514158e-05, "loss": 3.939, "step": 69500 }, { "epoch": 3.9942938659058487, "grad_norm": 9.529749870300293, "learning_rate": 1.1756548924690518e-05, "loss": 3.9911, "step": 70000 }, { "epoch": 4.0, "eval_loss": 3.9361355304718018, "eval_runtime": 55.3368, "eval_samples_per_second": 894.161, "eval_steps_per_second": 55.894, "step": 70100 }, { "epoch": 4.022824536376604, "grad_norm": 10.018484115600586, "learning_rate": 1.112623855986688e-05, "loss": 3.9744, "step": 70500 }, { "epoch": 4.051355206847361, "grad_norm": 11.023027420043945, "learning_rate": 1.049592819504324e-05, "loss": 3.9143, "step": 71000 }, { "epoch": 4.079885877318117, "grad_norm": 11.09890079498291, "learning_rate": 9.8656178302196e-06, "loss": 3.967, "step": 71500 }, { "epoch": 4.108416547788873, "grad_norm": 11.603713035583496, "learning_rate": 9.235307465395961e-06, "loss": 3.9525, "step": 72000 }, { "epoch": 4.136947218259629, "grad_norm": 8.631877899169922, "learning_rate": 8.604997100572321e-06, "loss": 3.9658, "step": 72500 }, { "epoch": 4.165477888730385, "grad_norm": 7.5662407875061035, "learning_rate": 7.974686735748683e-06, "loss": 3.9118, "step": 73000 }, { "epoch": 4.194008559201142, "grad_norm": 10.707962036132812, "learning_rate": 7.344376370925044e-06, "loss": 3.948, "step": 73500 }, { "epoch": 4.222539229671897, "grad_norm": 11.31724739074707, "learning_rate": 6.714066006101405e-06, "loss": 3.9319, "step": 74000 }, { "epoch": 4.251069900142653, "grad_norm": 9.398261070251465, "learning_rate": 6.083755641277765e-06, "loss": 3.9621, "step": 74500 }, { "epoch": 4.279600570613409, "grad_norm": 11.119857788085938, "learning_rate": 5.453445276454126e-06, "loss": 3.9429, "step": 75000 }, { "epoch": 4.3081312410841655, "grad_norm": 11.433586120605469, "learning_rate": 4.823134911630487e-06, "loss": 3.9126, "step": 75500 }, { "epoch": 4.336661911554922, "grad_norm": 10.043067932128906, "learning_rate": 4.192824546806848e-06, "loss": 3.9218, "step": 76000 }, { "epoch": 4.365192582025678, "grad_norm": 9.148512840270996, "learning_rate": 3.5625141819832086e-06, "loss": 3.9124, "step": 76500 }, { "epoch": 4.393723252496434, "grad_norm": 9.347081184387207, "learning_rate": 2.9322038171595694e-06, "loss": 3.8917, "step": 77000 }, { "epoch": 4.422253922967189, "grad_norm": 8.787015914916992, "learning_rate": 2.30189345233593e-06, "loss": 3.9091, "step": 77500 }, { "epoch": 4.4507845934379455, "grad_norm": 8.788630485534668, "learning_rate": 1.671583087512291e-06, "loss": 3.9211, "step": 78000 }, { "epoch": 4.479315263908702, "grad_norm": 8.639442443847656, "learning_rate": 1.0412727226886518e-06, "loss": 3.9139, "step": 78500 }, { "epoch": 4.507845934379458, "grad_norm": 9.101580619812012, "learning_rate": 4.1096235786501275e-07, "loss": 3.911, "step": 79000 } ], "logging_steps": 500, "max_steps": 79326, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 8192, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.6441172712886272e+16, "train_batch_size": 16, "trial_name": null, "trial_params": null }