{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.9976558837318334, "eval_steps": 500, "global_step": 2665, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.009376465072667605, "grad_norm": 0.20790895819664001, "learning_rate": 4.999956573574533e-05, "loss": 0.6093, "step": 5 }, { "epoch": 0.01875293014533521, "grad_norm": 0.22529356181621552, "learning_rate": 4.999826295806815e-05, "loss": 0.5708, "step": 10 }, { "epoch": 0.02812939521800281, "grad_norm": 0.19532211124897003, "learning_rate": 4.999609171222846e-05, "loss": 0.5403, "step": 15 }, { "epoch": 0.03750586029067042, "grad_norm": 0.17083343863487244, "learning_rate": 4.99930520736578e-05, "loss": 0.559, "step": 20 }, { "epoch": 0.04688232536333802, "grad_norm": 0.14475470781326294, "learning_rate": 4.998914414795668e-05, "loss": 0.5298, "step": 25 }, { "epoch": 0.05625879043600562, "grad_norm": 0.1513879895210266, "learning_rate": 4.99843680708909e-05, "loss": 0.5082, "step": 30 }, { "epoch": 0.06563525550867323, "grad_norm": 0.2249026745557785, "learning_rate": 4.997872400838682e-05, "loss": 0.5165, "step": 35 }, { "epoch": 0.07501172058134084, "grad_norm": 0.16140912473201752, "learning_rate": 4.997221215652562e-05, "loss": 0.5224, "step": 40 }, { "epoch": 0.08438818565400844, "grad_norm": 0.14945466816425323, "learning_rate": 4.9964832741536444e-05, "loss": 0.5322, "step": 45 }, { "epoch": 0.09376465072667604, "grad_norm": 0.12949012219905853, "learning_rate": 4.9956586019788584e-05, "loss": 0.4886, "step": 50 }, { "epoch": 0.10314111579934365, "grad_norm": 0.1388949751853943, "learning_rate": 4.9947472277782584e-05, "loss": 0.487, "step": 55 }, { "epoch": 0.11251758087201125, "grad_norm": 0.1725090742111206, "learning_rate": 4.993749183214021e-05, "loss": 0.4749, "step": 60 }, { "epoch": 0.12189404594467886, "grad_norm": 0.1453619748353958, "learning_rate": 4.992664502959351e-05, "loss": 0.4769, "step": 65 }, { "epoch": 0.13127051101734646, "grad_norm": 0.1829695701599121, "learning_rate": 4.991493224697281e-05, "loss": 0.4587, "step": 70 }, { "epoch": 0.14064697609001406, "grad_norm": 0.20768924057483673, "learning_rate": 4.990235389119352e-05, "loss": 0.4858, "step": 75 }, { "epoch": 0.15002344116268168, "grad_norm": 0.18472038209438324, "learning_rate": 4.9888910399242065e-05, "loss": 0.4485, "step": 80 }, { "epoch": 0.15939990623534928, "grad_norm": 0.20636047422885895, "learning_rate": 4.987460223816067e-05, "loss": 0.4629, "step": 85 }, { "epoch": 0.16877637130801687, "grad_norm": 0.22416724264621735, "learning_rate": 4.985942990503119e-05, "loss": 0.4683, "step": 90 }, { "epoch": 0.1781528363806845, "grad_norm": 0.17546997964382172, "learning_rate": 4.984339392695777e-05, "loss": 0.4625, "step": 95 }, { "epoch": 0.1875293014533521, "grad_norm": 0.18117469549179077, "learning_rate": 4.9826494861048576e-05, "loss": 0.4346, "step": 100 }, { "epoch": 0.19690576652601968, "grad_norm": 0.18969941139221191, "learning_rate": 4.980873329439644e-05, "loss": 0.4871, "step": 105 }, { "epoch": 0.2062822315986873, "grad_norm": 0.1979440301656723, "learning_rate": 4.979010984405842e-05, "loss": 0.4564, "step": 110 }, { "epoch": 0.2156586966713549, "grad_norm": 0.2334485948085785, "learning_rate": 4.9770625157034436e-05, "loss": 0.4801, "step": 115 }, { "epoch": 0.2250351617440225, "grad_norm": 0.2206389456987381, "learning_rate": 4.975027991024473e-05, "loss": 0.4678, "step": 120 }, { "epoch": 0.23441162681669012, "grad_norm": 0.19880324602127075, "learning_rate": 4.972907481050637e-05, "loss": 0.4794, "step": 125 }, { "epoch": 0.2437880918893577, "grad_norm": 0.22166623175144196, "learning_rate": 4.970701059450872e-05, "loss": 0.4729, "step": 130 }, { "epoch": 0.25316455696202533, "grad_norm": 0.19174839556217194, "learning_rate": 4.968408802878778e-05, "loss": 0.4754, "step": 135 }, { "epoch": 0.26254102203469293, "grad_norm": 0.21328027546405792, "learning_rate": 4.9660307909699645e-05, "loss": 0.4374, "step": 140 }, { "epoch": 0.2719174871073605, "grad_norm": 0.21817001700401306, "learning_rate": 4.963567106339276e-05, "loss": 0.4539, "step": 145 }, { "epoch": 0.2812939521800281, "grad_norm": 0.22513534128665924, "learning_rate": 4.961017834577927e-05, "loss": 0.4713, "step": 150 }, { "epoch": 0.2906704172526957, "grad_norm": 0.22900743782520294, "learning_rate": 4.958383064250525e-05, "loss": 0.4394, "step": 155 }, { "epoch": 0.30004688232536336, "grad_norm": 0.2164304554462433, "learning_rate": 4.955662886891995e-05, "loss": 0.4462, "step": 160 }, { "epoch": 0.30942334739803096, "grad_norm": 0.18812324106693268, "learning_rate": 4.952857397004401e-05, "loss": 0.475, "step": 165 }, { "epoch": 0.31879981247069855, "grad_norm": 0.21450480818748474, "learning_rate": 4.949966692053663e-05, "loss": 0.462, "step": 170 }, { "epoch": 0.32817627754336615, "grad_norm": 0.19328171014785767, "learning_rate": 4.946990872466164e-05, "loss": 0.4543, "step": 175 }, { "epoch": 0.33755274261603374, "grad_norm": 0.2076808661222458, "learning_rate": 4.943930041625272e-05, "loss": 0.442, "step": 180 }, { "epoch": 0.34692920768870134, "grad_norm": 0.207493394613266, "learning_rate": 4.940784305867741e-05, "loss": 0.4445, "step": 185 }, { "epoch": 0.356305672761369, "grad_norm": 0.2177191525697708, "learning_rate": 4.937553774480018e-05, "loss": 0.4327, "step": 190 }, { "epoch": 0.3656821378340366, "grad_norm": 0.22189687192440033, "learning_rate": 4.934238559694448e-05, "loss": 0.4804, "step": 195 }, { "epoch": 0.3750586029067042, "grad_norm": 0.23572292923927307, "learning_rate": 4.9308387766853725e-05, "loss": 0.4701, "step": 200 }, { "epoch": 0.38443506797937177, "grad_norm": 0.2647152841091156, "learning_rate": 4.92735454356513e-05, "loss": 0.4513, "step": 205 }, { "epoch": 0.39381153305203936, "grad_norm": 0.2548702657222748, "learning_rate": 4.9237859813799535e-05, "loss": 0.4733, "step": 210 }, { "epoch": 0.40318799812470696, "grad_norm": 0.22933240234851837, "learning_rate": 4.9201332141057623e-05, "loss": 0.4399, "step": 215 }, { "epoch": 0.4125644631973746, "grad_norm": 0.24353943765163422, "learning_rate": 4.9163963686438575e-05, "loss": 0.4483, "step": 220 }, { "epoch": 0.4219409282700422, "grad_norm": 0.23032227158546448, "learning_rate": 4.912575574816511e-05, "loss": 0.4415, "step": 225 }, { "epoch": 0.4313173933427098, "grad_norm": 0.27580568194389343, "learning_rate": 4.908670965362457e-05, "loss": 0.4167, "step": 230 }, { "epoch": 0.4406938584153774, "grad_norm": 0.20118704438209534, "learning_rate": 4.9046826759322825e-05, "loss": 0.4625, "step": 235 }, { "epoch": 0.450070323488045, "grad_norm": 0.27328649163246155, "learning_rate": 4.9006108450837095e-05, "loss": 0.4473, "step": 240 }, { "epoch": 0.45944678856071264, "grad_norm": 0.24165798723697662, "learning_rate": 4.8964556142767845e-05, "loss": 0.4512, "step": 245 }, { "epoch": 0.46882325363338023, "grad_norm": 0.23875291645526886, "learning_rate": 4.892217127868965e-05, "loss": 0.4469, "step": 250 }, { "epoch": 0.4781997187060478, "grad_norm": 0.2750314176082611, "learning_rate": 4.8878955331101026e-05, "loss": 0.4655, "step": 255 }, { "epoch": 0.4875761837787154, "grad_norm": 0.23661385476589203, "learning_rate": 4.8834909801373264e-05, "loss": 0.4181, "step": 260 }, { "epoch": 0.496952648851383, "grad_norm": 0.2660558223724365, "learning_rate": 4.879003621969831e-05, "loss": 0.4295, "step": 265 }, { "epoch": 0.5063291139240507, "grad_norm": 0.26039206981658936, "learning_rate": 4.874433614503554e-05, "loss": 0.4398, "step": 270 }, { "epoch": 0.5157055789967182, "grad_norm": 0.23488390445709229, "learning_rate": 4.869781116505768e-05, "loss": 0.4303, "step": 275 }, { "epoch": 0.5250820440693859, "grad_norm": 0.25998032093048096, "learning_rate": 4.8650462896095597e-05, "loss": 0.4462, "step": 280 }, { "epoch": 0.5344585091420534, "grad_norm": 0.2225557565689087, "learning_rate": 4.860229298308213e-05, "loss": 0.4321, "step": 285 }, { "epoch": 0.543834974214721, "grad_norm": 0.2588254511356354, "learning_rate": 4.8553303099495e-05, "loss": 0.4253, "step": 290 }, { "epoch": 0.5532114392873887, "grad_norm": 0.23887360095977783, "learning_rate": 4.8503494947298634e-05, "loss": 0.4438, "step": 295 }, { "epoch": 0.5625879043600562, "grad_norm": 0.24014081060886383, "learning_rate": 4.845287025688503e-05, "loss": 0.4449, "step": 300 }, { "epoch": 0.5719643694327239, "grad_norm": 0.23950184881687164, "learning_rate": 4.8401430787013666e-05, "loss": 0.4372, "step": 305 }, { "epoch": 0.5813408345053914, "grad_norm": 0.23509977757930756, "learning_rate": 4.8349178324750387e-05, "loss": 0.4308, "step": 310 }, { "epoch": 0.5907172995780591, "grad_norm": 0.25797203183174133, "learning_rate": 4.8296114685405324e-05, "loss": 0.429, "step": 315 }, { "epoch": 0.6000937646507267, "grad_norm": 0.4092111885547638, "learning_rate": 4.824224171246981e-05, "loss": 0.4755, "step": 320 }, { "epoch": 0.6094702297233943, "grad_norm": 0.2742483913898468, "learning_rate": 4.8187561277552374e-05, "loss": 0.4391, "step": 325 }, { "epoch": 0.6188466947960619, "grad_norm": 0.26388296484947205, "learning_rate": 4.813207528031366e-05, "loss": 0.4184, "step": 330 }, { "epoch": 0.6282231598687295, "grad_norm": 0.2813189923763275, "learning_rate": 4.807578564840051e-05, "loss": 0.4677, "step": 335 }, { "epoch": 0.6375996249413971, "grad_norm": 0.24881592392921448, "learning_rate": 4.801869433737891e-05, "loss": 0.4366, "step": 340 }, { "epoch": 0.6469760900140648, "grad_norm": 0.24432657659053802, "learning_rate": 4.796080333066613e-05, "loss": 0.4536, "step": 345 }, { "epoch": 0.6563525550867323, "grad_norm": 0.26776596903800964, "learning_rate": 4.790211463946174e-05, "loss": 0.4678, "step": 350 }, { "epoch": 0.6657290201593999, "grad_norm": 0.2862718403339386, "learning_rate": 4.784263030267781e-05, "loss": 0.4178, "step": 355 }, { "epoch": 0.6751054852320675, "grad_norm": 0.25835174322128296, "learning_rate": 4.7782352386868035e-05, "loss": 0.457, "step": 360 }, { "epoch": 0.6844819503047351, "grad_norm": 0.245767742395401, "learning_rate": 4.7721282986155945e-05, "loss": 0.4098, "step": 365 }, { "epoch": 0.6938584153774027, "grad_norm": 0.31210237741470337, "learning_rate": 4.7659424222162165e-05, "loss": 0.4263, "step": 370 }, { "epoch": 0.7032348804500703, "grad_norm": 0.27777954936027527, "learning_rate": 4.7596778243930694e-05, "loss": 0.4433, "step": 375 }, { "epoch": 0.712611345522738, "grad_norm": 0.2539767324924469, "learning_rate": 4.7533347227854265e-05, "loss": 0.4369, "step": 380 }, { "epoch": 0.7219878105954055, "grad_norm": 0.3035132586956024, "learning_rate": 4.7469133377598695e-05, "loss": 0.4206, "step": 385 }, { "epoch": 0.7313642756680732, "grad_norm": 0.3070785105228424, "learning_rate": 4.740413892402639e-05, "loss": 0.4311, "step": 390 }, { "epoch": 0.7407407407407407, "grad_norm": 0.29886579513549805, "learning_rate": 4.7338366125118775e-05, "loss": 0.4471, "step": 395 }, { "epoch": 0.7501172058134083, "grad_norm": 0.4168696105480194, "learning_rate": 4.727181726589789e-05, "loss": 0.4567, "step": 400 }, { "epoch": 0.759493670886076, "grad_norm": 0.4680975377559662, "learning_rate": 4.7204494658346996e-05, "loss": 0.4307, "step": 405 }, { "epoch": 0.7688701359587435, "grad_norm": 0.2854766249656677, "learning_rate": 4.713640064133025e-05, "loss": 0.4303, "step": 410 }, { "epoch": 0.7782466010314112, "grad_norm": 0.29096436500549316, "learning_rate": 4.706753758051145e-05, "loss": 0.4266, "step": 415 }, { "epoch": 0.7876230661040787, "grad_norm": 0.2967214286327362, "learning_rate": 4.699790786827188e-05, "loss": 0.4401, "step": 420 }, { "epoch": 0.7969995311767464, "grad_norm": 0.30379167199134827, "learning_rate": 4.6927513923627124e-05, "loss": 0.4343, "step": 425 }, { "epoch": 0.8063759962494139, "grad_norm": 0.28345710039138794, "learning_rate": 4.68563581921431e-05, "loss": 0.4205, "step": 430 }, { "epoch": 0.8157524613220816, "grad_norm": 0.3067210912704468, "learning_rate": 4.6784443145851074e-05, "loss": 0.3997, "step": 435 }, { "epoch": 0.8251289263947492, "grad_norm": 0.2822760343551636, "learning_rate": 4.671177128316176e-05, "loss": 0.4103, "step": 440 }, { "epoch": 0.8345053914674168, "grad_norm": 0.3110799491405487, "learning_rate": 4.663834512877853e-05, "loss": 0.4417, "step": 445 }, { "epoch": 0.8438818565400844, "grad_norm": 0.3088453710079193, "learning_rate": 4.6564167233609736e-05, "loss": 0.4285, "step": 450 }, { "epoch": 0.853258321612752, "grad_norm": 0.2757161259651184, "learning_rate": 4.648924017468003e-05, "loss": 0.4207, "step": 455 }, { "epoch": 0.8626347866854196, "grad_norm": 0.3142268657684326, "learning_rate": 4.6413566555040896e-05, "loss": 0.4621, "step": 460 }, { "epoch": 0.8720112517580872, "grad_norm": 0.2766186594963074, "learning_rate": 4.633714900368018e-05, "loss": 0.4201, "step": 465 }, { "epoch": 0.8813877168307548, "grad_norm": 0.2814094126224518, "learning_rate": 4.625999017543075e-05, "loss": 0.4435, "step": 470 }, { "epoch": 0.8907641819034224, "grad_norm": 0.29491695761680603, "learning_rate": 4.618209275087829e-05, "loss": 0.4527, "step": 475 }, { "epoch": 0.90014064697609, "grad_norm": 0.30693337321281433, "learning_rate": 4.610345943626817e-05, "loss": 0.4268, "step": 480 }, { "epoch": 0.9095171120487576, "grad_norm": 0.29456719756126404, "learning_rate": 4.602409296341141e-05, "loss": 0.4094, "step": 485 }, { "epoch": 0.9188935771214253, "grad_norm": 0.2991769015789032, "learning_rate": 4.5943996089589775e-05, "loss": 0.4333, "step": 490 }, { "epoch": 0.9282700421940928, "grad_norm": 0.3011597692966461, "learning_rate": 4.586317159746001e-05, "loss": 0.4821, "step": 495 }, { "epoch": 0.9376465072667605, "grad_norm": 0.33786553144454956, "learning_rate": 4.5781622294957136e-05, "loss": 0.4228, "step": 500 }, { "epoch": 0.947022972339428, "grad_norm": 0.28369849920272827, "learning_rate": 4.569935101519692e-05, "loss": 0.4333, "step": 505 }, { "epoch": 0.9563994374120957, "grad_norm": 0.27438902854919434, "learning_rate": 4.561636061637745e-05, "loss": 0.4698, "step": 510 }, { "epoch": 0.9657759024847632, "grad_norm": 0.26311373710632324, "learning_rate": 4.553265398167981e-05, "loss": 0.4211, "step": 515 }, { "epoch": 0.9751523675574308, "grad_norm": 0.3168148994445801, "learning_rate": 4.5448234019167945e-05, "loss": 0.4183, "step": 520 }, { "epoch": 0.9845288326300985, "grad_norm": 0.3082910478115082, "learning_rate": 4.536310366168763e-05, "loss": 0.4251, "step": 525 }, { "epoch": 0.993905297702766, "grad_norm": 0.30115172266960144, "learning_rate": 4.5277265866764565e-05, "loss": 0.4292, "step": 530 }, { "epoch": 1.0032817627754336, "grad_norm": 0.3527440130710602, "learning_rate": 4.519072361650163e-05, "loss": 0.4051, "step": 535 }, { "epoch": 1.0126582278481013, "grad_norm": 0.3075549006462097, "learning_rate": 4.5103479917475286e-05, "loss": 0.4027, "step": 540 }, { "epoch": 1.0220346929207689, "grad_norm": 0.3359437584877014, "learning_rate": 4.501553780063113e-05, "loss": 0.4008, "step": 545 }, { "epoch": 1.0314111579934364, "grad_norm": 0.29366499185562134, "learning_rate": 4.4926900321178595e-05, "loss": 0.4421, "step": 550 }, { "epoch": 1.0407876230661042, "grad_norm": 0.3159714937210083, "learning_rate": 4.483757055848479e-05, "loss": 0.4124, "step": 555 }, { "epoch": 1.0501640881387717, "grad_norm": 0.30557411909103394, "learning_rate": 4.4747551615967534e-05, "loss": 0.4023, "step": 560 }, { "epoch": 1.0595405532114393, "grad_norm": 0.3260650336742401, "learning_rate": 4.4656846620987557e-05, "loss": 0.4376, "step": 565 }, { "epoch": 1.0689170182841068, "grad_norm": 0.3325200378894806, "learning_rate": 4.4565458724739825e-05, "loss": 0.4421, "step": 570 }, { "epoch": 1.0782934833567746, "grad_norm": 0.3431427776813507, "learning_rate": 4.447339110214405e-05, "loss": 0.4119, "step": 575 }, { "epoch": 1.087669948429442, "grad_norm": 0.3403518497943878, "learning_rate": 4.438064695173446e-05, "loss": 0.4276, "step": 580 }, { "epoch": 1.0970464135021096, "grad_norm": 0.35619255900382996, "learning_rate": 4.428722949554857e-05, "loss": 0.4258, "step": 585 }, { "epoch": 1.1064228785747774, "grad_norm": 0.2965317964553833, "learning_rate": 4.419314197901537e-05, "loss": 0.41, "step": 590 }, { "epoch": 1.115799343647445, "grad_norm": 0.31006523966789246, "learning_rate": 4.4098387670842466e-05, "loss": 0.4292, "step": 595 }, { "epoch": 1.1251758087201125, "grad_norm": 0.3238849639892578, "learning_rate": 4.400296986290258e-05, "loss": 0.4079, "step": 600 }, { "epoch": 1.1345522737927802, "grad_norm": 0.33775433897972107, "learning_rate": 4.390689187011917e-05, "loss": 0.4123, "step": 605 }, { "epoch": 1.1439287388654478, "grad_norm": 0.32464325428009033, "learning_rate": 4.3810157030351276e-05, "loss": 0.3983, "step": 610 }, { "epoch": 1.1533052039381153, "grad_norm": 0.36141082644462585, "learning_rate": 4.371276870427753e-05, "loss": 0.3923, "step": 615 }, { "epoch": 1.1626816690107828, "grad_norm": 0.3391363322734833, "learning_rate": 4.3614730275279457e-05, "loss": 0.402, "step": 620 }, { "epoch": 1.1720581340834506, "grad_norm": 0.3580019772052765, "learning_rate": 4.351604514932387e-05, "loss": 0.4138, "step": 625 }, { "epoch": 1.1814345991561181, "grad_norm": 0.3803277611732483, "learning_rate": 4.341671675484459e-05, "loss": 0.4254, "step": 630 }, { "epoch": 1.1908110642287857, "grad_norm": 0.31499192118644714, "learning_rate": 4.331674854262331e-05, "loss": 0.4174, "step": 635 }, { "epoch": 1.2001875293014534, "grad_norm": 0.34143874049186707, "learning_rate": 4.321614398566972e-05, "loss": 0.3954, "step": 640 }, { "epoch": 1.209563994374121, "grad_norm": 0.328500360250473, "learning_rate": 4.3114906579100853e-05, "loss": 0.3944, "step": 645 }, { "epoch": 1.2189404594467885, "grad_norm": 0.34014376997947693, "learning_rate": 4.301303984001967e-05, "loss": 0.4265, "step": 650 }, { "epoch": 1.228316924519456, "grad_norm": 0.33687010407447815, "learning_rate": 4.291054730739286e-05, "loss": 0.4255, "step": 655 }, { "epoch": 1.2376933895921238, "grad_norm": 0.3507842719554901, "learning_rate": 4.2807432541927865e-05, "loss": 0.424, "step": 660 }, { "epoch": 1.2470698546647914, "grad_norm": 0.35059022903442383, "learning_rate": 4.2703699125949245e-05, "loss": 0.4227, "step": 665 }, { "epoch": 1.256446319737459, "grad_norm": 0.381517618894577, "learning_rate": 4.259935066327415e-05, "loss": 0.4674, "step": 670 }, { "epoch": 1.2658227848101267, "grad_norm": 0.309592604637146, "learning_rate": 4.2494390779087187e-05, "loss": 0.4365, "step": 675 }, { "epoch": 1.2751992498827942, "grad_norm": 0.33022499084472656, "learning_rate": 4.238882311981441e-05, "loss": 0.3918, "step": 680 }, { "epoch": 1.2845757149554617, "grad_norm": 0.37556394934654236, "learning_rate": 4.228265135299669e-05, "loss": 0.4257, "step": 685 }, { "epoch": 1.2939521800281293, "grad_norm": 0.32622891664505005, "learning_rate": 4.2175879167162304e-05, "loss": 0.4442, "step": 690 }, { "epoch": 1.303328645100797, "grad_norm": 0.3617080748081207, "learning_rate": 4.206851027169871e-05, "loss": 0.4292, "step": 695 }, { "epoch": 1.3127051101734646, "grad_norm": 0.3132336735725403, "learning_rate": 4.196054839672382e-05, "loss": 0.4135, "step": 700 }, { "epoch": 1.3220815752461323, "grad_norm": 0.315346360206604, "learning_rate": 4.1851997292956255e-05, "loss": 0.4163, "step": 705 }, { "epoch": 1.3314580403187999, "grad_norm": 0.3571698069572449, "learning_rate": 4.174286073158516e-05, "loss": 0.4144, "step": 710 }, { "epoch": 1.3408345053914674, "grad_norm": 0.35016143321990967, "learning_rate": 4.163314250413913e-05, "loss": 0.3898, "step": 715 }, { "epoch": 1.350210970464135, "grad_norm": 0.34223607182502747, "learning_rate": 4.152284642235452e-05, "loss": 0.422, "step": 720 }, { "epoch": 1.3595874355368025, "grad_norm": 0.3561420142650604, "learning_rate": 4.141197631804298e-05, "loss": 0.4279, "step": 725 }, { "epoch": 1.3689639006094703, "grad_norm": 0.3987065255641937, "learning_rate": 4.1300536042958354e-05, "loss": 0.4233, "step": 730 }, { "epoch": 1.3783403656821378, "grad_norm": 0.34799569845199585, "learning_rate": 4.118852946866291e-05, "loss": 0.4214, "step": 735 }, { "epoch": 1.3877168307548056, "grad_norm": 0.4023985266685486, "learning_rate": 4.107596048639274e-05, "loss": 0.394, "step": 740 }, { "epoch": 1.397093295827473, "grad_norm": 0.3554348647594452, "learning_rate": 4.0962833006922675e-05, "loss": 0.4283, "step": 745 }, { "epoch": 1.4064697609001406, "grad_norm": 0.33551228046417236, "learning_rate": 4.0849150960430356e-05, "loss": 0.4164, "step": 750 }, { "epoch": 1.4158462259728082, "grad_norm": 0.3325096368789673, "learning_rate": 4.0734918296359716e-05, "loss": 0.4242, "step": 755 }, { "epoch": 1.4252226910454757, "grad_norm": 0.32375457882881165, "learning_rate": 4.0620138983283785e-05, "loss": 0.4356, "step": 760 }, { "epoch": 1.4345991561181435, "grad_norm": 0.3310219943523407, "learning_rate": 4.050481700876677e-05, "loss": 0.4283, "step": 765 }, { "epoch": 1.443975621190811, "grad_norm": 0.3990609347820282, "learning_rate": 4.038895637922559e-05, "loss": 0.4069, "step": 770 }, { "epoch": 1.4533520862634788, "grad_norm": 0.35942110419273376, "learning_rate": 4.027256111979063e-05, "loss": 0.4103, "step": 775 }, { "epoch": 1.4627285513361463, "grad_norm": 0.3540840744972229, "learning_rate": 4.015563527416595e-05, "loss": 0.4243, "step": 780 }, { "epoch": 1.4721050164088139, "grad_norm": 0.3438728153705597, "learning_rate": 4.003818290448876e-05, "loss": 0.4222, "step": 785 }, { "epoch": 1.4814814814814814, "grad_norm": 0.3385811150074005, "learning_rate": 3.992020809118832e-05, "loss": 0.4597, "step": 790 }, { "epoch": 1.4908579465541492, "grad_norm": 0.3348815441131592, "learning_rate": 3.980171493284418e-05, "loss": 0.4049, "step": 795 }, { "epoch": 1.5002344116268167, "grad_norm": 0.3877304494380951, "learning_rate": 3.9682707546043785e-05, "loss": 0.4573, "step": 800 }, { "epoch": 1.5096108766994845, "grad_norm": 0.372578889131546, "learning_rate": 3.9563190065239474e-05, "loss": 0.4064, "step": 805 }, { "epoch": 1.518987341772152, "grad_norm": 0.4140545725822449, "learning_rate": 3.9443166642604814e-05, "loss": 0.4125, "step": 810 }, { "epoch": 1.5283638068448195, "grad_norm": 0.30370739102363586, "learning_rate": 3.932264144789038e-05, "loss": 0.3997, "step": 815 }, { "epoch": 1.537740271917487, "grad_norm": 0.3417491018772125, "learning_rate": 3.920161866827889e-05, "loss": 0.4374, "step": 820 }, { "epoch": 1.5471167369901546, "grad_norm": 0.3708433210849762, "learning_rate": 3.908010250823972e-05, "loss": 0.4273, "step": 825 }, { "epoch": 1.5564932020628222, "grad_norm": 0.3969629108905792, "learning_rate": 3.895809718938283e-05, "loss": 0.4255, "step": 830 }, { "epoch": 1.56586966713549, "grad_norm": 0.344990998506546, "learning_rate": 3.883560695031213e-05, "loss": 0.3804, "step": 835 }, { "epoch": 1.5752461322081577, "grad_norm": 0.35197708010673523, "learning_rate": 3.871263604647822e-05, "loss": 0.4172, "step": 840 }, { "epoch": 1.5846225972808252, "grad_norm": 0.35443294048309326, "learning_rate": 3.858918875003053e-05, "loss": 0.4132, "step": 845 }, { "epoch": 1.5939990623534928, "grad_norm": 0.3435305953025818, "learning_rate": 3.846526934966891e-05, "loss": 0.4295, "step": 850 }, { "epoch": 1.6033755274261603, "grad_norm": 0.39062219858169556, "learning_rate": 3.834088215049464e-05, "loss": 0.4305, "step": 855 }, { "epoch": 1.6127519924988278, "grad_norm": 0.37184613943099976, "learning_rate": 3.821603147386088e-05, "loss": 0.4114, "step": 860 }, { "epoch": 1.6221284575714956, "grad_norm": 0.3640553057193756, "learning_rate": 3.80907216572225e-05, "loss": 0.3906, "step": 865 }, { "epoch": 1.6315049226441631, "grad_norm": 0.3880080580711365, "learning_rate": 3.796495705398544e-05, "loss": 0.4302, "step": 870 }, { "epoch": 1.640881387716831, "grad_norm": 0.3895851671695709, "learning_rate": 3.783874203335542e-05, "loss": 0.4044, "step": 875 }, { "epoch": 1.6502578527894984, "grad_norm": 0.4187454283237457, "learning_rate": 3.77120809801862e-05, "loss": 0.404, "step": 880 }, { "epoch": 1.659634317862166, "grad_norm": 0.32384535670280457, "learning_rate": 3.758497829482721e-05, "loss": 0.439, "step": 885 }, { "epoch": 1.6690107829348335, "grad_norm": 0.38500747084617615, "learning_rate": 3.7457438392970686e-05, "loss": 0.3843, "step": 890 }, { "epoch": 1.678387248007501, "grad_norm": 0.40994030237197876, "learning_rate": 3.732946570549825e-05, "loss": 0.4189, "step": 895 }, { "epoch": 1.6877637130801688, "grad_norm": 0.34361371397972107, "learning_rate": 3.720106467832701e-05, "loss": 0.4532, "step": 900 }, { "epoch": 1.6971401781528364, "grad_norm": 0.37997591495513916, "learning_rate": 3.707223977225507e-05, "loss": 0.4298, "step": 905 }, { "epoch": 1.7065166432255041, "grad_norm": 0.3335699439048767, "learning_rate": 3.694299546280657e-05, "loss": 0.4442, "step": 910 }, { "epoch": 1.7158931082981717, "grad_norm": 0.38952916860580444, "learning_rate": 3.681333624007623e-05, "loss": 0.4263, "step": 915 }, { "epoch": 1.7252695733708392, "grad_norm": 0.36404407024383545, "learning_rate": 3.6683266608573286e-05, "loss": 0.4905, "step": 920 }, { "epoch": 1.7346460384435067, "grad_norm": 0.3932304084300995, "learning_rate": 3.6552791087065075e-05, "loss": 0.3976, "step": 925 }, { "epoch": 1.7440225035161743, "grad_norm": 0.3757387697696686, "learning_rate": 3.642191420842e-05, "loss": 0.4259, "step": 930 }, { "epoch": 1.753398968588842, "grad_norm": 0.36103346943855286, "learning_rate": 3.6290640519450074e-05, "loss": 0.4184, "step": 935 }, { "epoch": 1.7627754336615096, "grad_norm": 0.3944946229457855, "learning_rate": 3.6158974580752954e-05, "loss": 0.451, "step": 940 }, { "epoch": 1.7721518987341773, "grad_norm": 0.4115646183490753, "learning_rate": 3.60269209665535e-05, "loss": 0.4075, "step": 945 }, { "epoch": 1.7815283638068449, "grad_norm": 0.36925530433654785, "learning_rate": 3.589448426454486e-05, "loss": 0.432, "step": 950 }, { "epoch": 1.7909048288795124, "grad_norm": 0.39060622453689575, "learning_rate": 3.5761669075729084e-05, "loss": 0.3979, "step": 955 }, { "epoch": 1.80028129395218, "grad_norm": 0.3404456377029419, "learning_rate": 3.562848001425729e-05, "loss": 0.458, "step": 960 }, { "epoch": 1.8096577590248475, "grad_norm": 0.3717631995677948, "learning_rate": 3.549492170726937e-05, "loss": 0.4111, "step": 965 }, { "epoch": 1.8190342240975153, "grad_norm": 0.3680458068847656, "learning_rate": 3.53609987947332e-05, "loss": 0.4276, "step": 970 }, { "epoch": 1.8284106891701828, "grad_norm": 0.3876858651638031, "learning_rate": 3.5226715929283506e-05, "loss": 0.3976, "step": 975 }, { "epoch": 1.8377871542428506, "grad_norm": 0.3375720679759979, "learning_rate": 3.509207777606013e-05, "loss": 0.3993, "step": 980 }, { "epoch": 1.847163619315518, "grad_norm": 0.3920380771160126, "learning_rate": 3.49570890125461e-05, "loss": 0.4019, "step": 985 }, { "epoch": 1.8565400843881856, "grad_norm": 0.35098692774772644, "learning_rate": 3.482175432840495e-05, "loss": 0.3951, "step": 990 }, { "epoch": 1.8659165494608532, "grad_norm": 0.3828470706939697, "learning_rate": 3.468607842531797e-05, "loss": 0.4119, "step": 995 }, { "epoch": 1.8752930145335207, "grad_norm": 0.3853389620780945, "learning_rate": 3.455006601682075e-05, "loss": 0.3997, "step": 1000 }, { "epoch": 1.8846694796061885, "grad_norm": 0.36217638850212097, "learning_rate": 3.441372182813946e-05, "loss": 0.3949, "step": 1005 }, { "epoch": 1.8940459446788562, "grad_norm": 0.38590380549430847, "learning_rate": 3.427705059602671e-05, "loss": 0.4041, "step": 1010 }, { "epoch": 1.9034224097515238, "grad_norm": 0.3818942606449127, "learning_rate": 3.414005706859693e-05, "loss": 0.4247, "step": 1015 }, { "epoch": 1.9127988748241913, "grad_norm": 0.36795973777770996, "learning_rate": 3.400274600516152e-05, "loss": 0.406, "step": 1020 }, { "epoch": 1.9221753398968588, "grad_norm": 0.37341007590293884, "learning_rate": 3.386512217606339e-05, "loss": 0.405, "step": 1025 }, { "epoch": 1.9315518049695264, "grad_norm": 0.3815973699092865, "learning_rate": 3.372719036251132e-05, "loss": 0.4093, "step": 1030 }, { "epoch": 1.9409282700421941, "grad_norm": 0.32503455877304077, "learning_rate": 3.3588955356413795e-05, "loss": 0.409, "step": 1035 }, { "epoch": 1.9503047351148617, "grad_norm": 0.41935157775878906, "learning_rate": 3.3450421960212566e-05, "loss": 0.4052, "step": 1040 }, { "epoch": 1.9596812001875294, "grad_norm": 0.3876263201236725, "learning_rate": 3.3311594986715814e-05, "loss": 0.4271, "step": 1045 }, { "epoch": 1.969057665260197, "grad_norm": 0.37812694907188416, "learning_rate": 3.317247925893089e-05, "loss": 0.3977, "step": 1050 }, { "epoch": 1.9784341303328645, "grad_norm": 0.39891424775123596, "learning_rate": 3.3033079609896834e-05, "loss": 0.3831, "step": 1055 }, { "epoch": 1.987810595405532, "grad_norm": 0.3980333209037781, "learning_rate": 3.289340088251642e-05, "loss": 0.3874, "step": 1060 }, { "epoch": 1.9971870604781996, "grad_norm": 0.3976622521877289, "learning_rate": 3.275344792938791e-05, "loss": 0.4092, "step": 1065 }, { "epoch": 2.006563525550867, "grad_norm": 0.34287935495376587, "learning_rate": 3.2613225612636525e-05, "loss": 0.3795, "step": 1070 }, { "epoch": 2.015939990623535, "grad_norm": 0.3522503674030304, "learning_rate": 3.247273880374542e-05, "loss": 0.3895, "step": 1075 }, { "epoch": 2.0253164556962027, "grad_norm": 0.4062464237213135, "learning_rate": 3.2331992383386566e-05, "loss": 0.4279, "step": 1080 }, { "epoch": 2.03469292076887, "grad_norm": 0.3659546375274658, "learning_rate": 3.21909912412511e-05, "loss": 0.4056, "step": 1085 }, { "epoch": 2.0440693858415377, "grad_norm": 0.3941885232925415, "learning_rate": 3.2049740275879493e-05, "loss": 0.4142, "step": 1090 }, { "epoch": 2.0534458509142053, "grad_norm": 0.43663620948791504, "learning_rate": 3.190824439449137e-05, "loss": 0.431, "step": 1095 }, { "epoch": 2.062822315986873, "grad_norm": 0.40177416801452637, "learning_rate": 3.176650851281499e-05, "loss": 0.3952, "step": 1100 }, { "epoch": 2.0721987810595404, "grad_norm": 0.3755819797515869, "learning_rate": 3.162453755491655e-05, "loss": 0.3717, "step": 1105 }, { "epoch": 2.0815752461322083, "grad_norm": 0.3558565676212311, "learning_rate": 3.1482336453028986e-05, "loss": 0.392, "step": 1110 }, { "epoch": 2.090951711204876, "grad_norm": 0.3933524191379547, "learning_rate": 3.133991014738076e-05, "loss": 0.4004, "step": 1115 }, { "epoch": 2.1003281762775434, "grad_norm": 0.3724795877933502, "learning_rate": 3.1197263586024155e-05, "loss": 0.4095, "step": 1120 }, { "epoch": 2.109704641350211, "grad_norm": 0.4012450873851776, "learning_rate": 3.105440172466337e-05, "loss": 0.4169, "step": 1125 }, { "epoch": 2.1190811064228785, "grad_norm": 0.43174034357070923, "learning_rate": 3.09113295264824e-05, "loss": 0.4105, "step": 1130 }, { "epoch": 2.128457571495546, "grad_norm": 0.37383630871772766, "learning_rate": 3.076805196197255e-05, "loss": 0.4085, "step": 1135 }, { "epoch": 2.1378340365682136, "grad_norm": 0.40011560916900635, "learning_rate": 3.0624574008759805e-05, "loss": 0.4704, "step": 1140 }, { "epoch": 2.1472105016408816, "grad_norm": 0.38523784279823303, "learning_rate": 3.0480900651431876e-05, "loss": 0.3913, "step": 1145 }, { "epoch": 2.156586966713549, "grad_norm": 0.39474254846572876, "learning_rate": 3.0337036881365045e-05, "loss": 0.3904, "step": 1150 }, { "epoch": 2.1659634317862166, "grad_norm": 0.3798459768295288, "learning_rate": 3.0192987696550746e-05, "loss": 0.3899, "step": 1155 }, { "epoch": 2.175339896858884, "grad_norm": 0.42982256412506104, "learning_rate": 3.0048758101421914e-05, "loss": 0.4025, "step": 1160 }, { "epoch": 2.1847163619315517, "grad_norm": 0.40922918915748596, "learning_rate": 2.9904353106679178e-05, "loss": 0.4038, "step": 1165 }, { "epoch": 2.1940928270042193, "grad_norm": 0.3726600408554077, "learning_rate": 2.975977772911671e-05, "loss": 0.4396, "step": 1170 }, { "epoch": 2.2034692920768872, "grad_norm": 0.4845339357852936, "learning_rate": 2.9615036991448015e-05, "loss": 0.4164, "step": 1175 }, { "epoch": 2.212845757149555, "grad_norm": 0.4207274317741394, "learning_rate": 2.947013592213137e-05, "loss": 0.4128, "step": 1180 }, { "epoch": 2.2222222222222223, "grad_norm": 0.3607243299484253, "learning_rate": 2.9325079555195163e-05, "loss": 0.4096, "step": 1185 }, { "epoch": 2.23159868729489, "grad_norm": 0.4174239933490753, "learning_rate": 2.9179872930063e-05, "loss": 0.3859, "step": 1190 }, { "epoch": 2.2409751523675574, "grad_norm": 0.46482330560684204, "learning_rate": 2.9034521091378635e-05, "loss": 0.3905, "step": 1195 }, { "epoch": 2.250351617440225, "grad_norm": 0.43692639470100403, "learning_rate": 2.8889029088830686e-05, "loss": 0.4121, "step": 1200 }, { "epoch": 2.2597280825128925, "grad_norm": 0.40547794103622437, "learning_rate": 2.8743401976977257e-05, "loss": 0.4, "step": 1205 }, { "epoch": 2.2691045475855605, "grad_norm": 0.3966444432735443, "learning_rate": 2.8597644815070263e-05, "loss": 0.4482, "step": 1210 }, { "epoch": 2.278481012658228, "grad_norm": 0.41211700439453125, "learning_rate": 2.845176266687974e-05, "loss": 0.3914, "step": 1215 }, { "epoch": 2.2878574777308955, "grad_norm": 0.45993107557296753, "learning_rate": 2.8305760600517862e-05, "loss": 0.3862, "step": 1220 }, { "epoch": 2.297233942803563, "grad_norm": 0.4274505078792572, "learning_rate": 2.815964368826292e-05, "loss": 0.4039, "step": 1225 }, { "epoch": 2.3066104078762306, "grad_norm": 0.4318040609359741, "learning_rate": 2.8013417006383076e-05, "loss": 0.3868, "step": 1230 }, { "epoch": 2.315986872948898, "grad_norm": 0.43052956461906433, "learning_rate": 2.7867085634960016e-05, "loss": 0.4118, "step": 1235 }, { "epoch": 2.3253633380215657, "grad_norm": 0.4302162528038025, "learning_rate": 2.772065465771244e-05, "loss": 0.407, "step": 1240 }, { "epoch": 2.3347398030942337, "grad_norm": 0.46215954422950745, "learning_rate": 2.7574129161819495e-05, "loss": 0.4101, "step": 1245 }, { "epoch": 2.344116268166901, "grad_norm": 0.37758246064186096, "learning_rate": 2.7427514237744e-05, "loss": 0.4205, "step": 1250 }, { "epoch": 2.3534927332395688, "grad_norm": 0.40978917479515076, "learning_rate": 2.7280814979055612e-05, "loss": 0.4116, "step": 1255 }, { "epoch": 2.3628691983122363, "grad_norm": 0.41880902647972107, "learning_rate": 2.713403648225388e-05, "loss": 0.3917, "step": 1260 }, { "epoch": 2.372245663384904, "grad_norm": 0.39450138807296753, "learning_rate": 2.698718384659114e-05, "loss": 0.41, "step": 1265 }, { "epoch": 2.3816221284575714, "grad_norm": 0.4140292704105377, "learning_rate": 2.684026217389544e-05, "loss": 0.4449, "step": 1270 }, { "epoch": 2.390998593530239, "grad_norm": 0.40103039145469666, "learning_rate": 2.6693276568393245e-05, "loss": 0.3831, "step": 1275 }, { "epoch": 2.400375058602907, "grad_norm": 0.4108443558216095, "learning_rate": 2.6546232136532083e-05, "loss": 0.4321, "step": 1280 }, { "epoch": 2.4097515236755744, "grad_norm": 0.39723923802375793, "learning_rate": 2.639913398680322e-05, "loss": 0.4236, "step": 1285 }, { "epoch": 2.419127988748242, "grad_norm": 0.36689993739128113, "learning_rate": 2.6251987229564123e-05, "loss": 0.4029, "step": 1290 }, { "epoch": 2.4285044538209095, "grad_norm": 0.44802841544151306, "learning_rate": 2.610479697686093e-05, "loss": 0.3794, "step": 1295 }, { "epoch": 2.437880918893577, "grad_norm": 0.4667953848838806, "learning_rate": 2.595756834225089e-05, "loss": 0.4178, "step": 1300 }, { "epoch": 2.4472573839662446, "grad_norm": 0.41082021594047546, "learning_rate": 2.5810306440624644e-05, "loss": 0.4496, "step": 1305 }, { "epoch": 2.456633849038912, "grad_norm": 0.4804684519767761, "learning_rate": 2.566301638802861e-05, "loss": 0.4215, "step": 1310 }, { "epoch": 2.46601031411158, "grad_norm": 0.4113605320453644, "learning_rate": 2.551570330148716e-05, "loss": 0.3964, "step": 1315 }, { "epoch": 2.4753867791842477, "grad_norm": 0.40913546085357666, "learning_rate": 2.5368372298824922e-05, "loss": 0.3871, "step": 1320 }, { "epoch": 2.484763244256915, "grad_norm": 0.45879778265953064, "learning_rate": 2.5221028498488947e-05, "loss": 0.4146, "step": 1325 }, { "epoch": 2.4941397093295827, "grad_norm": 0.46979445219039917, "learning_rate": 2.507367701937087e-05, "loss": 0.3892, "step": 1330 }, { "epoch": 2.5035161744022503, "grad_norm": 0.3892490565776825, "learning_rate": 2.492632298062913e-05, "loss": 0.3995, "step": 1335 }, { "epoch": 2.512892639474918, "grad_norm": 0.4790550768375397, "learning_rate": 2.4778971501511063e-05, "loss": 0.423, "step": 1340 }, { "epoch": 2.5222691045475853, "grad_norm": 0.41578853130340576, "learning_rate": 2.4631627701175084e-05, "loss": 0.3845, "step": 1345 }, { "epoch": 2.5316455696202533, "grad_norm": 0.457792192697525, "learning_rate": 2.448429669851285e-05, "loss": 0.3854, "step": 1350 }, { "epoch": 2.541022034692921, "grad_norm": 0.44336166977882385, "learning_rate": 2.43369836119714e-05, "loss": 0.4096, "step": 1355 }, { "epoch": 2.5503984997655884, "grad_norm": 0.4408751428127289, "learning_rate": 2.4189693559375365e-05, "loss": 0.4167, "step": 1360 }, { "epoch": 2.559774964838256, "grad_norm": 0.4089582860469818, "learning_rate": 2.4042431657749117e-05, "loss": 0.3829, "step": 1365 }, { "epoch": 2.5691514299109235, "grad_norm": 0.4729689657688141, "learning_rate": 2.3895203023139073e-05, "loss": 0.3861, "step": 1370 }, { "epoch": 2.578527894983591, "grad_norm": 0.3834232985973358, "learning_rate": 2.3748012770435883e-05, "loss": 0.408, "step": 1375 }, { "epoch": 2.5879043600562586, "grad_norm": 0.47135239839553833, "learning_rate": 2.3600866013196787e-05, "loss": 0.408, "step": 1380 }, { "epoch": 2.5972808251289266, "grad_norm": 0.4259538948535919, "learning_rate": 2.3453767863467923e-05, "loss": 0.3994, "step": 1385 }, { "epoch": 2.606657290201594, "grad_norm": 0.4320566654205322, "learning_rate": 2.3306723431606758e-05, "loss": 0.3801, "step": 1390 }, { "epoch": 2.6160337552742616, "grad_norm": 0.3922535479068756, "learning_rate": 2.3159737826104565e-05, "loss": 0.421, "step": 1395 }, { "epoch": 2.625410220346929, "grad_norm": 0.4332144558429718, "learning_rate": 2.3012816153408863e-05, "loss": 0.4058, "step": 1400 }, { "epoch": 2.6347866854195967, "grad_norm": 0.46418026089668274, "learning_rate": 2.286596351774613e-05, "loss": 0.3922, "step": 1405 }, { "epoch": 2.6441631504922647, "grad_norm": 0.4476454555988312, "learning_rate": 2.271918502094439e-05, "loss": 0.3822, "step": 1410 }, { "epoch": 2.653539615564932, "grad_norm": 0.4266027510166168, "learning_rate": 2.2572485762256005e-05, "loss": 0.382, "step": 1415 }, { "epoch": 2.6629160806375998, "grad_norm": 0.5318887233734131, "learning_rate": 2.2425870838180507e-05, "loss": 0.3881, "step": 1420 }, { "epoch": 2.6722925457102673, "grad_norm": 0.4814053177833557, "learning_rate": 2.2279345342287567e-05, "loss": 0.4004, "step": 1425 }, { "epoch": 2.681669010782935, "grad_norm": 0.43376603722572327, "learning_rate": 2.2132914365039993e-05, "loss": 0.4052, "step": 1430 }, { "epoch": 2.6910454758556024, "grad_norm": 0.4502093493938446, "learning_rate": 2.1986582993616926e-05, "loss": 0.4018, "step": 1435 }, { "epoch": 2.70042194092827, "grad_norm": 0.41802868247032166, "learning_rate": 2.1840356311737084e-05, "loss": 0.3969, "step": 1440 }, { "epoch": 2.709798406000938, "grad_norm": 0.37022143602371216, "learning_rate": 2.169423939948215e-05, "loss": 0.421, "step": 1445 }, { "epoch": 2.719174871073605, "grad_norm": 0.47152194380760193, "learning_rate": 2.154823733312027e-05, "loss": 0.4105, "step": 1450 }, { "epoch": 2.728551336146273, "grad_norm": 0.4552992582321167, "learning_rate": 2.140235518492975e-05, "loss": 0.3879, "step": 1455 }, { "epoch": 2.7379278012189405, "grad_norm": 0.42312508821487427, "learning_rate": 2.125659802302275e-05, "loss": 0.4014, "step": 1460 }, { "epoch": 2.747304266291608, "grad_norm": 0.37846195697784424, "learning_rate": 2.1110970911169316e-05, "loss": 0.4116, "step": 1465 }, { "epoch": 2.7566807313642756, "grad_norm": 0.42782410979270935, "learning_rate": 2.096547890862137e-05, "loss": 0.394, "step": 1470 }, { "epoch": 2.766057196436943, "grad_norm": 0.4843522310256958, "learning_rate": 2.0820127069937008e-05, "loss": 0.3984, "step": 1475 }, { "epoch": 2.775433661509611, "grad_norm": 0.46708783507347107, "learning_rate": 2.0674920444804847e-05, "loss": 0.3906, "step": 1480 }, { "epoch": 2.7848101265822782, "grad_norm": 0.4305874705314636, "learning_rate": 2.0529864077868643e-05, "loss": 0.4101, "step": 1485 }, { "epoch": 2.794186591654946, "grad_norm": 0.40540897846221924, "learning_rate": 2.0384963008551995e-05, "loss": 0.4056, "step": 1490 }, { "epoch": 2.8035630567276137, "grad_norm": 0.4375884532928467, "learning_rate": 2.0240222270883288e-05, "loss": 0.4285, "step": 1495 }, { "epoch": 2.8129395218002813, "grad_norm": 0.4903147220611572, "learning_rate": 2.0095646893320828e-05, "loss": 0.3852, "step": 1500 }, { "epoch": 2.822315986872949, "grad_norm": 0.4598534107208252, "learning_rate": 1.9951241898578085e-05, "loss": 0.3815, "step": 1505 }, { "epoch": 2.8316924519456164, "grad_norm": 0.4689639210700989, "learning_rate": 1.980701230344926e-05, "loss": 0.4003, "step": 1510 }, { "epoch": 2.8410689170182843, "grad_norm": 0.4126102924346924, "learning_rate": 1.9662963118634954e-05, "loss": 0.419, "step": 1515 }, { "epoch": 2.8504453820909514, "grad_norm": 0.5043439865112305, "learning_rate": 1.9519099348568127e-05, "loss": 0.3939, "step": 1520 }, { "epoch": 2.8598218471636194, "grad_norm": 1.107731580734253, "learning_rate": 1.93754259912402e-05, "loss": 0.4255, "step": 1525 }, { "epoch": 2.869198312236287, "grad_norm": 0.43347910046577454, "learning_rate": 1.9231948038027462e-05, "loss": 0.3898, "step": 1530 }, { "epoch": 2.8785747773089545, "grad_norm": 0.42569872736930847, "learning_rate": 1.9088670473517605e-05, "loss": 0.404, "step": 1535 }, { "epoch": 2.887951242381622, "grad_norm": 0.42402687668800354, "learning_rate": 1.8945598275336633e-05, "loss": 0.3834, "step": 1540 }, { "epoch": 2.8973277074542896, "grad_norm": 0.47469502687454224, "learning_rate": 1.8802736413975844e-05, "loss": 0.4378, "step": 1545 }, { "epoch": 2.9067041725269576, "grad_norm": 0.43439674377441406, "learning_rate": 1.866008985261924e-05, "loss": 0.3953, "step": 1550 }, { "epoch": 2.916080637599625, "grad_norm": 0.4604777693748474, "learning_rate": 1.8517663546971013e-05, "loss": 0.401, "step": 1555 }, { "epoch": 2.9254571026722926, "grad_norm": 0.46638715267181396, "learning_rate": 1.8375462445083464e-05, "loss": 0.4101, "step": 1560 }, { "epoch": 2.93483356774496, "grad_norm": 0.46090996265411377, "learning_rate": 1.8233491487185006e-05, "loss": 0.4172, "step": 1565 }, { "epoch": 2.9442100328176277, "grad_norm": 0.4011158049106598, "learning_rate": 1.8091755605508643e-05, "loss": 0.4161, "step": 1570 }, { "epoch": 2.9535864978902953, "grad_norm": 0.48108768463134766, "learning_rate": 1.7950259724120512e-05, "loss": 0.4291, "step": 1575 }, { "epoch": 2.962962962962963, "grad_norm": 0.4274104833602905, "learning_rate": 1.7809008758748913e-05, "loss": 0.4102, "step": 1580 }, { "epoch": 2.972339428035631, "grad_norm": 0.46605607867240906, "learning_rate": 1.766800761661344e-05, "loss": 0.3811, "step": 1585 }, { "epoch": 2.9817158931082983, "grad_norm": 0.5101811289787292, "learning_rate": 1.752726119625459e-05, "loss": 0.4292, "step": 1590 }, { "epoch": 2.991092358180966, "grad_norm": 0.45644745230674744, "learning_rate": 1.7386774387363484e-05, "loss": 0.4238, "step": 1595 }, { "epoch": 3.0004688232536334, "grad_norm": 0.4713057577610016, "learning_rate": 1.724655207061209e-05, "loss": 0.4033, "step": 1600 }, { "epoch": 3.009845288326301, "grad_norm": 0.43844765424728394, "learning_rate": 1.710659911748359e-05, "loss": 0.4157, "step": 1605 }, { "epoch": 3.0192217533989685, "grad_norm": 0.46663135290145874, "learning_rate": 1.696692039010317e-05, "loss": 0.39, "step": 1610 }, { "epoch": 3.028598218471636, "grad_norm": 0.433660626411438, "learning_rate": 1.6827520741069118e-05, "loss": 0.4134, "step": 1615 }, { "epoch": 3.037974683544304, "grad_norm": 0.4730021059513092, "learning_rate": 1.6688405013284192e-05, "loss": 0.4095, "step": 1620 }, { "epoch": 3.0473511486169715, "grad_norm": 0.3873763680458069, "learning_rate": 1.6549578039787436e-05, "loss": 0.406, "step": 1625 }, { "epoch": 3.056727613689639, "grad_norm": 0.4337627589702606, "learning_rate": 1.6411044643586204e-05, "loss": 0.3721, "step": 1630 }, { "epoch": 3.0661040787623066, "grad_norm": 0.4677801728248596, "learning_rate": 1.627280963748869e-05, "loss": 0.3739, "step": 1635 }, { "epoch": 3.075480543834974, "grad_norm": 0.4103066921234131, "learning_rate": 1.613487782393661e-05, "loss": 0.4147, "step": 1640 }, { "epoch": 3.0848570089076417, "grad_norm": 0.48415660858154297, "learning_rate": 1.5997253994838484e-05, "loss": 0.3684, "step": 1645 }, { "epoch": 3.0942334739803092, "grad_norm": 0.5216359496116638, "learning_rate": 1.5859942931403072e-05, "loss": 0.3977, "step": 1650 }, { "epoch": 3.103609939052977, "grad_norm": 0.46912798285484314, "learning_rate": 1.5722949403973308e-05, "loss": 0.4219, "step": 1655 }, { "epoch": 3.1129864041256448, "grad_norm": 0.5608705282211304, "learning_rate": 1.5586278171860546e-05, "loss": 0.3957, "step": 1660 }, { "epoch": 3.1223628691983123, "grad_norm": 0.44117632508277893, "learning_rate": 1.5449933983179256e-05, "loss": 0.3977, "step": 1665 }, { "epoch": 3.13173933427098, "grad_norm": 0.45745447278022766, "learning_rate": 1.5313921574682032e-05, "loss": 0.3842, "step": 1670 }, { "epoch": 3.1411157993436474, "grad_norm": 0.4728071689605713, "learning_rate": 1.517824567159506e-05, "loss": 0.3607, "step": 1675 }, { "epoch": 3.150492264416315, "grad_norm": 0.4618283808231354, "learning_rate": 1.5042910987453909e-05, "loss": 0.4065, "step": 1680 }, { "epoch": 3.1598687294889825, "grad_norm": 0.5012606978416443, "learning_rate": 1.4907922223939874e-05, "loss": 0.3856, "step": 1685 }, { "epoch": 3.1692451945616504, "grad_norm": 0.42908909916877747, "learning_rate": 1.4773284070716503e-05, "loss": 0.4035, "step": 1690 }, { "epoch": 3.178621659634318, "grad_norm": 0.4527268707752228, "learning_rate": 1.4639001205266803e-05, "loss": 0.3798, "step": 1695 }, { "epoch": 3.1879981247069855, "grad_norm": 0.4418729841709137, "learning_rate": 1.4505078292730632e-05, "loss": 0.4021, "step": 1700 }, { "epoch": 3.197374589779653, "grad_norm": 0.46943390369415283, "learning_rate": 1.4371519985742715e-05, "loss": 0.3648, "step": 1705 }, { "epoch": 3.2067510548523206, "grad_norm": 0.49398934841156006, "learning_rate": 1.4238330924270927e-05, "loss": 0.3895, "step": 1710 }, { "epoch": 3.216127519924988, "grad_norm": 0.47141122817993164, "learning_rate": 1.4105515735455149e-05, "loss": 0.4117, "step": 1715 }, { "epoch": 3.2255039849976557, "grad_norm": 0.4392653703689575, "learning_rate": 1.3973079033446501e-05, "loss": 0.4201, "step": 1720 }, { "epoch": 3.2348804500703237, "grad_norm": 0.5093908309936523, "learning_rate": 1.3841025419247045e-05, "loss": 0.3906, "step": 1725 }, { "epoch": 3.244256915142991, "grad_norm": 0.4423893094062805, "learning_rate": 1.3709359480549932e-05, "loss": 0.3695, "step": 1730 }, { "epoch": 3.2536333802156587, "grad_norm": 0.48355525732040405, "learning_rate": 1.3578085791580008e-05, "loss": 0.365, "step": 1735 }, { "epoch": 3.2630098452883263, "grad_norm": 0.41361674666404724, "learning_rate": 1.3447208912934927e-05, "loss": 0.4157, "step": 1740 }, { "epoch": 3.272386310360994, "grad_norm": 0.5110962390899658, "learning_rate": 1.3316733391426716e-05, "loss": 0.4005, "step": 1745 }, { "epoch": 3.2817627754336613, "grad_norm": 0.4553833603858948, "learning_rate": 1.3186663759923782e-05, "loss": 0.3792, "step": 1750 }, { "epoch": 3.291139240506329, "grad_norm": 0.5377179980278015, "learning_rate": 1.3057004537193423e-05, "loss": 0.4093, "step": 1755 }, { "epoch": 3.300515705578997, "grad_norm": 0.4575432240962982, "learning_rate": 1.2927760227744943e-05, "loss": 0.3647, "step": 1760 }, { "epoch": 3.3098921706516644, "grad_norm": 0.5318431854248047, "learning_rate": 1.2798935321673e-05, "loss": 0.374, "step": 1765 }, { "epoch": 3.319268635724332, "grad_norm": 0.4416927695274353, "learning_rate": 1.2670534294501756e-05, "loss": 0.3888, "step": 1770 }, { "epoch": 3.3286451007969995, "grad_norm": 0.5091699361801147, "learning_rate": 1.2542561607029322e-05, "loss": 0.3969, "step": 1775 }, { "epoch": 3.338021565869667, "grad_norm": 0.45796316862106323, "learning_rate": 1.2415021705172799e-05, "loss": 0.4196, "step": 1780 }, { "epoch": 3.3473980309423346, "grad_norm": 0.5065374970436096, "learning_rate": 1.2287919019813807e-05, "loss": 0.3885, "step": 1785 }, { "epoch": 3.356774496015002, "grad_norm": 0.5083961486816406, "learning_rate": 1.2161257966644588e-05, "loss": 0.4118, "step": 1790 }, { "epoch": 3.36615096108767, "grad_norm": 0.44367310404777527, "learning_rate": 1.2035042946014572e-05, "loss": 0.4124, "step": 1795 }, { "epoch": 3.3755274261603376, "grad_norm": 0.46685874462127686, "learning_rate": 1.1909278342777513e-05, "loss": 0.394, "step": 1800 }, { "epoch": 3.384903891233005, "grad_norm": 0.4649623930454254, "learning_rate": 1.1783968526139121e-05, "loss": 0.3996, "step": 1805 }, { "epoch": 3.3942803563056727, "grad_norm": 0.5142059922218323, "learning_rate": 1.1659117849505367e-05, "loss": 0.4036, "step": 1810 }, { "epoch": 3.4036568213783402, "grad_norm": 0.4626547396183014, "learning_rate": 1.1534730650331096e-05, "loss": 0.3976, "step": 1815 }, { "epoch": 3.413033286451008, "grad_norm": 0.4900161623954773, "learning_rate": 1.1410811249969475e-05, "loss": 0.4282, "step": 1820 }, { "epoch": 3.4224097515236753, "grad_norm": 0.43262672424316406, "learning_rate": 1.1287363953521779e-05, "loss": 0.3807, "step": 1825 }, { "epoch": 3.4317862165963433, "grad_norm": 0.46232739090919495, "learning_rate": 1.1164393049687868e-05, "loss": 0.3761, "step": 1830 }, { "epoch": 3.441162681669011, "grad_norm": 0.5012332201004028, "learning_rate": 1.104190281061718e-05, "loss": 0.4152, "step": 1835 }, { "epoch": 3.4505391467416784, "grad_norm": 0.4876830279827118, "learning_rate": 1.0919897491760279e-05, "loss": 0.363, "step": 1840 }, { "epoch": 3.459915611814346, "grad_norm": 0.44954073429107666, "learning_rate": 1.0798381331721109e-05, "loss": 0.3789, "step": 1845 }, { "epoch": 3.4692920768870135, "grad_norm": 0.44105660915374756, "learning_rate": 1.0677358552109618e-05, "loss": 0.413, "step": 1850 }, { "epoch": 3.4786685419596814, "grad_norm": 0.45533353090286255, "learning_rate": 1.0556833357395188e-05, "loss": 0.3812, "step": 1855 }, { "epoch": 3.488045007032349, "grad_norm": 0.4477691054344177, "learning_rate": 1.0436809934760527e-05, "loss": 0.3949, "step": 1860 }, { "epoch": 3.4974214721050165, "grad_norm": 0.4761820137500763, "learning_rate": 1.031729245395622e-05, "loss": 0.3649, "step": 1865 }, { "epoch": 3.506797937177684, "grad_norm": 0.48367804288864136, "learning_rate": 1.0198285067155827e-05, "loss": 0.3942, "step": 1870 }, { "epoch": 3.5161744022503516, "grad_norm": 0.46382734179496765, "learning_rate": 1.0079791908811683e-05, "loss": 0.4007, "step": 1875 }, { "epoch": 3.525550867323019, "grad_norm": 0.43440571427345276, "learning_rate": 9.961817095511242e-06, "loss": 0.3773, "step": 1880 }, { "epoch": 3.5349273323956867, "grad_norm": 0.45736581087112427, "learning_rate": 9.844364725834057e-06, "loss": 0.4196, "step": 1885 }, { "epoch": 3.5443037974683547, "grad_norm": 0.49210745096206665, "learning_rate": 9.727438880209366e-06, "loss": 0.3792, "step": 1890 }, { "epoch": 3.5536802625410218, "grad_norm": 0.5130970478057861, "learning_rate": 9.611043620774419e-06, "loss": 0.4036, "step": 1895 }, { "epoch": 3.5630567276136897, "grad_norm": 0.471587598323822, "learning_rate": 9.495182991233236e-06, "loss": 0.3955, "step": 1900 }, { "epoch": 3.5724331926863573, "grad_norm": 0.522293746471405, "learning_rate": 9.379861016716224e-06, "loss": 0.4056, "step": 1905 }, { "epoch": 3.581809657759025, "grad_norm": 0.4667339026927948, "learning_rate": 9.265081703640285e-06, "loss": 0.4144, "step": 1910 }, { "epoch": 3.5911861228316924, "grad_norm": 0.4434965252876282, "learning_rate": 9.150849039569655e-06, "loss": 0.39, "step": 1915 }, { "epoch": 3.60056258790436, "grad_norm": 0.46499741077423096, "learning_rate": 9.037166993077337e-06, "loss": 0.3976, "step": 1920 }, { "epoch": 3.609939052977028, "grad_norm": 0.47565406560897827, "learning_rate": 8.92403951360726e-06, "loss": 0.3774, "step": 1925 }, { "epoch": 3.6193155180496954, "grad_norm": 0.45665204524993896, "learning_rate": 8.811470531337102e-06, "loss": 0.4209, "step": 1930 }, { "epoch": 3.628691983122363, "grad_norm": 0.4445512890815735, "learning_rate": 8.699463957041649e-06, "loss": 0.428, "step": 1935 }, { "epoch": 3.6380684481950305, "grad_norm": 0.5019098520278931, "learning_rate": 8.588023681957028e-06, "loss": 0.4121, "step": 1940 }, { "epoch": 3.647444913267698, "grad_norm": 0.4833836555480957, "learning_rate": 8.477153577645481e-06, "loss": 0.4191, "step": 1945 }, { "epoch": 3.6568213783403656, "grad_norm": 0.8952960968017578, "learning_rate": 8.36685749586087e-06, "loss": 0.417, "step": 1950 }, { "epoch": 3.666197843413033, "grad_norm": 0.537151575088501, "learning_rate": 8.257139268414844e-06, "loss": 0.4047, "step": 1955 }, { "epoch": 3.675574308485701, "grad_norm": 0.5016284584999084, "learning_rate": 8.14800270704375e-06, "loss": 0.4119, "step": 1960 }, { "epoch": 3.6849507735583686, "grad_norm": 0.49896350502967834, "learning_rate": 8.039451603276185e-06, "loss": 0.3908, "step": 1965 }, { "epoch": 3.694327238631036, "grad_norm": 0.4538595974445343, "learning_rate": 7.931489728301292e-06, "loss": 0.4636, "step": 1970 }, { "epoch": 3.7037037037037037, "grad_norm": 0.5051478147506714, "learning_rate": 7.8241208328377e-06, "loss": 0.3925, "step": 1975 }, { "epoch": 3.7130801687763713, "grad_norm": 0.464751660823822, "learning_rate": 7.71734864700331e-06, "loss": 0.3957, "step": 1980 }, { "epoch": 3.722456633849039, "grad_norm": 0.4575950801372528, "learning_rate": 7.611176880185597e-06, "loss": 0.4004, "step": 1985 }, { "epoch": 3.7318330989217063, "grad_norm": 1.0100325345993042, "learning_rate": 7.505609220912821e-06, "loss": 0.3945, "step": 1990 }, { "epoch": 3.7412095639943743, "grad_norm": 0.47300100326538086, "learning_rate": 7.4006493367258515e-06, "loss": 0.4112, "step": 1995 }, { "epoch": 3.750586029067042, "grad_norm": 0.4961658716201782, "learning_rate": 7.2963008740507656e-06, "loss": 0.4123, "step": 2000 }, { "epoch": 3.7599624941397094, "grad_norm": 0.4959772229194641, "learning_rate": 7.192567458072138e-06, "loss": 0.4225, "step": 2005 }, { "epoch": 3.769338959212377, "grad_norm": 0.4662582576274872, "learning_rate": 7.089452692607146e-06, "loss": 0.3601, "step": 2010 }, { "epoch": 3.7787154242850445, "grad_norm": 0.4932651221752167, "learning_rate": 6.986960159980327e-06, "loss": 0.3654, "step": 2015 }, { "epoch": 3.788091889357712, "grad_norm": 0.4127211272716522, "learning_rate": 6.885093420899152e-06, "loss": 0.404, "step": 2020 }, { "epoch": 3.7974683544303796, "grad_norm": 0.5077162384986877, "learning_rate": 6.783856014330281e-06, "loss": 0.4015, "step": 2025 }, { "epoch": 3.8068448195030475, "grad_norm": 0.48392489552497864, "learning_rate": 6.68325145737669e-06, "loss": 0.3995, "step": 2030 }, { "epoch": 3.816221284575715, "grad_norm": 0.472433865070343, "learning_rate": 6.583283245155414e-06, "loss": 0.4102, "step": 2035 }, { "epoch": 3.8255977496483826, "grad_norm": 0.5179023742675781, "learning_rate": 6.483954850676133e-06, "loss": 0.3872, "step": 2040 }, { "epoch": 3.83497421472105, "grad_norm": 0.473317414522171, "learning_rate": 6.385269724720547e-06, "loss": 0.4137, "step": 2045 }, { "epoch": 3.8443506797937177, "grad_norm": 0.5055838227272034, "learning_rate": 6.28723129572247e-06, "loss": 0.3769, "step": 2050 }, { "epoch": 3.8537271448663852, "grad_norm": 0.7340875267982483, "learning_rate": 6.189842969648737e-06, "loss": 0.4172, "step": 2055 }, { "epoch": 3.8631036099390528, "grad_norm": 0.47141486406326294, "learning_rate": 6.0931081298808316e-06, "loss": 0.3964, "step": 2060 }, { "epoch": 3.8724800750117208, "grad_norm": 0.48367300629615784, "learning_rate": 5.997030137097426e-06, "loss": 0.3858, "step": 2065 }, { "epoch": 3.8818565400843883, "grad_norm": 0.4850892424583435, "learning_rate": 5.901612329157535e-06, "loss": 0.4061, "step": 2070 }, { "epoch": 3.891233005157056, "grad_norm": 0.45096391439437866, "learning_rate": 5.806858020984629e-06, "loss": 0.4316, "step": 2075 }, { "epoch": 3.9006094702297234, "grad_norm": 0.4780539274215698, "learning_rate": 5.712770504451426e-06, "loss": 0.398, "step": 2080 }, { "epoch": 3.909985935302391, "grad_norm": 0.5059804320335388, "learning_rate": 5.619353048265552e-06, "loss": 0.3819, "step": 2085 }, { "epoch": 3.9193624003750585, "grad_norm": 0.4585495889186859, "learning_rate": 5.526608897855953e-06, "loss": 0.3873, "step": 2090 }, { "epoch": 3.928738865447726, "grad_norm": 0.5159944295883179, "learning_rate": 5.434541275260182e-06, "loss": 0.421, "step": 2095 }, { "epoch": 3.938115330520394, "grad_norm": 0.4670829176902771, "learning_rate": 5.343153379012444e-06, "loss": 0.4135, "step": 2100 }, { "epoch": 3.9474917955930615, "grad_norm": 0.4940509796142578, "learning_rate": 5.252448384032471e-06, "loss": 0.391, "step": 2105 }, { "epoch": 3.956868260665729, "grad_norm": 0.4698173999786377, "learning_rate": 5.162429441515221e-06, "loss": 0.4018, "step": 2110 }, { "epoch": 3.9662447257383966, "grad_norm": 0.49286118149757385, "learning_rate": 5.073099678821413e-06, "loss": 0.386, "step": 2115 }, { "epoch": 3.975621190811064, "grad_norm": 0.4916246235370636, "learning_rate": 4.984462199368872e-06, "loss": 0.418, "step": 2120 }, { "epoch": 3.9849976558837317, "grad_norm": 0.5064667463302612, "learning_rate": 4.8965200825247245e-06, "loss": 0.3901, "step": 2125 }, { "epoch": 3.994374120956399, "grad_norm": 0.40669238567352295, "learning_rate": 4.809276383498376e-06, "loss": 0.3967, "step": 2130 }, { "epoch": 4.003750586029067, "grad_norm": 0.46348291635513306, "learning_rate": 4.722734133235438e-06, "loss": 0.4028, "step": 2135 }, { "epoch": 4.013127051101734, "grad_norm": 0.4899981617927551, "learning_rate": 4.636896338312374e-06, "loss": 0.4171, "step": 2140 }, { "epoch": 4.022503516174402, "grad_norm": 0.4799928665161133, "learning_rate": 4.551765980832059e-06, "loss": 0.3881, "step": 2145 }, { "epoch": 4.03187998124707, "grad_norm": 0.5409300327301025, "learning_rate": 4.467346018320198e-06, "loss": 0.4012, "step": 2150 }, { "epoch": 4.041256446319737, "grad_norm": 0.4798271358013153, "learning_rate": 4.383639383622557e-06, "loss": 0.3872, "step": 2155 }, { "epoch": 4.050632911392405, "grad_norm": 0.487051784992218, "learning_rate": 4.300648984803085e-06, "loss": 0.3946, "step": 2160 }, { "epoch": 4.060009376465072, "grad_norm": 0.4268711507320404, "learning_rate": 4.218377705042867e-06, "loss": 0.4011, "step": 2165 }, { "epoch": 4.06938584153774, "grad_norm": 0.44820597767829895, "learning_rate": 4.1368284025399965e-06, "loss": 0.4089, "step": 2170 }, { "epoch": 4.0787623066104075, "grad_norm": 0.42342251539230347, "learning_rate": 4.0560039104102305e-06, "loss": 0.3993, "step": 2175 }, { "epoch": 4.0881387716830755, "grad_norm": 0.45130422711372375, "learning_rate": 3.975907036588594e-06, "loss": 0.3773, "step": 2180 }, { "epoch": 4.0975152367557435, "grad_norm": 0.486767053604126, "learning_rate": 3.8965405637318294e-06, "loss": 0.3855, "step": 2185 }, { "epoch": 4.106891701828411, "grad_norm": 0.5100731253623962, "learning_rate": 3.817907249121713e-06, "loss": 0.4134, "step": 2190 }, { "epoch": 4.1162681669010786, "grad_norm": 0.46529749035835266, "learning_rate": 3.7400098245692572e-06, "loss": 0.4058, "step": 2195 }, { "epoch": 4.125644631973746, "grad_norm": 0.4404692053794861, "learning_rate": 3.662850996319825e-06, "loss": 0.3808, "step": 2200 }, { "epoch": 4.135021097046414, "grad_norm": 0.5206236839294434, "learning_rate": 3.586433444959103e-06, "loss": 0.3882, "step": 2205 }, { "epoch": 4.144397562119081, "grad_norm": 0.5646428465843201, "learning_rate": 3.5107598253199758e-06, "loss": 0.3846, "step": 2210 }, { "epoch": 4.153774027191749, "grad_norm": 0.45417729020118713, "learning_rate": 3.4358327663902677e-06, "loss": 0.3876, "step": 2215 }, { "epoch": 4.163150492264417, "grad_norm": 0.557522177696228, "learning_rate": 3.3616548712214756e-06, "loss": 0.3866, "step": 2220 }, { "epoch": 4.172526957337084, "grad_norm": 0.4840220510959625, "learning_rate": 3.288228716838246e-06, "loss": 0.3884, "step": 2225 }, { "epoch": 4.181903422409752, "grad_norm": 0.513700008392334, "learning_rate": 3.2155568541489268e-06, "loss": 0.4142, "step": 2230 }, { "epoch": 4.191279887482419, "grad_norm": 0.44611451029777527, "learning_rate": 3.143641807856898e-06, "loss": 0.4048, "step": 2235 }, { "epoch": 4.200656352555087, "grad_norm": 0.4795036017894745, "learning_rate": 3.0724860763728767e-06, "loss": 0.4244, "step": 2240 }, { "epoch": 4.210032817627754, "grad_norm": 0.4182005524635315, "learning_rate": 3.0020921317281264e-06, "loss": 0.3679, "step": 2245 }, { "epoch": 4.219409282700422, "grad_norm": 0.485270619392395, "learning_rate": 2.9324624194885436e-06, "loss": 0.4163, "step": 2250 }, { "epoch": 4.22878574777309, "grad_norm": 0.5020278096199036, "learning_rate": 2.8635993586697553e-06, "loss": 0.3861, "step": 2255 }, { "epoch": 4.238162212845757, "grad_norm": 0.449848473072052, "learning_rate": 2.795505341653007e-06, "loss": 0.3841, "step": 2260 }, { "epoch": 4.247538677918425, "grad_norm": 0.49279651045799255, "learning_rate": 2.728182734102111e-06, "loss": 0.3884, "step": 2265 }, { "epoch": 4.256915142991092, "grad_norm": 0.4899311363697052, "learning_rate": 2.6616338748812255e-06, "loss": 0.4325, "step": 2270 }, { "epoch": 4.26629160806376, "grad_norm": 0.4672718942165375, "learning_rate": 2.595861075973613e-06, "loss": 0.3927, "step": 2275 }, { "epoch": 4.275668073136427, "grad_norm": 0.4548036754131317, "learning_rate": 2.530866622401304e-06, "loss": 0.3879, "step": 2280 }, { "epoch": 4.285044538209095, "grad_norm": 0.4743858575820923, "learning_rate": 2.4666527721457416e-06, "loss": 0.4045, "step": 2285 }, { "epoch": 4.294421003281763, "grad_norm": 0.405335932970047, "learning_rate": 2.40322175606931e-06, "loss": 0.4215, "step": 2290 }, { "epoch": 4.30379746835443, "grad_norm": 0.5168406367301941, "learning_rate": 2.3405757778378445e-06, "loss": 0.3735, "step": 2295 }, { "epoch": 4.313173933427098, "grad_norm": 0.47989800572395325, "learning_rate": 2.278717013844059e-06, "loss": 0.3858, "step": 2300 }, { "epoch": 4.322550398499765, "grad_norm": 0.4565723240375519, "learning_rate": 2.2176476131319707e-06, "loss": 0.3568, "step": 2305 }, { "epoch": 4.331926863572433, "grad_norm": 0.5169614553451538, "learning_rate": 2.1573696973221922e-06, "loss": 0.385, "step": 2310 }, { "epoch": 4.3413033286451, "grad_norm": 0.487652987241745, "learning_rate": 2.0978853605382624e-06, "loss": 0.3884, "step": 2315 }, { "epoch": 4.350679793717768, "grad_norm": 0.46447455883026123, "learning_rate": 2.0391966693338733e-06, "loss": 0.369, "step": 2320 }, { "epoch": 4.360056258790436, "grad_norm": 0.4496999979019165, "learning_rate": 1.9813056626210886e-06, "loss": 0.368, "step": 2325 }, { "epoch": 4.369432723863103, "grad_norm": 0.4512302577495575, "learning_rate": 1.9242143515994933e-06, "loss": 0.3907, "step": 2330 }, { "epoch": 4.378809188935771, "grad_norm": 0.4967067837715149, "learning_rate": 1.8679247196863425e-06, "loss": 0.3816, "step": 2335 }, { "epoch": 4.3881856540084385, "grad_norm": 0.47957539558410645, "learning_rate": 1.8124387224476347e-06, "loss": 0.3577, "step": 2340 }, { "epoch": 4.3975621190811065, "grad_norm": 0.5050100684165955, "learning_rate": 1.757758287530195e-06, "loss": 0.3922, "step": 2345 }, { "epoch": 4.4069385841537745, "grad_norm": 0.4630352258682251, "learning_rate": 1.7038853145946804e-06, "loss": 0.387, "step": 2350 }, { "epoch": 4.416315049226442, "grad_norm": 0.47333136200904846, "learning_rate": 1.6508216752496141e-06, "loss": 0.4288, "step": 2355 }, { "epoch": 4.42569151429911, "grad_norm": 0.502042829990387, "learning_rate": 1.5985692129863395e-06, "loss": 0.3881, "step": 2360 }, { "epoch": 4.435067979371777, "grad_norm": 0.47139260172843933, "learning_rate": 1.547129743114978e-06, "loss": 0.4099, "step": 2365 }, { "epoch": 4.444444444444445, "grad_norm": 0.4579063653945923, "learning_rate": 1.496505052701372e-06, "loss": 0.3691, "step": 2370 }, { "epoch": 4.453820909517112, "grad_norm": 0.49161937832832336, "learning_rate": 1.4466969005050013e-06, "loss": 0.3807, "step": 2375 }, { "epoch": 4.46319737458978, "grad_norm": 0.4590076506137848, "learning_rate": 1.3977070169178763e-06, "loss": 0.3943, "step": 2380 }, { "epoch": 4.472573839662447, "grad_norm": 0.4571836590766907, "learning_rate": 1.349537103904408e-06, "loss": 0.3941, "step": 2385 }, { "epoch": 4.481950304735115, "grad_norm": 0.49009573459625244, "learning_rate": 1.3021888349423222e-06, "loss": 0.385, "step": 2390 }, { "epoch": 4.491326769807783, "grad_norm": 0.457081139087677, "learning_rate": 1.2556638549644644e-06, "loss": 0.4137, "step": 2395 }, { "epoch": 4.50070323488045, "grad_norm": 0.46972543001174927, "learning_rate": 1.2099637803016983e-06, "loss": 0.3827, "step": 2400 }, { "epoch": 4.510079699953118, "grad_norm": 0.44439297914505005, "learning_rate": 1.1650901986267365e-06, "loss": 0.3695, "step": 2405 }, { "epoch": 4.519456165025785, "grad_norm": 0.5357493162155151, "learning_rate": 1.1210446688989768e-06, "loss": 0.444, "step": 2410 }, { "epoch": 4.528832630098453, "grad_norm": 0.4965517818927765, "learning_rate": 1.0778287213103478e-06, "loss": 0.3816, "step": 2415 }, { "epoch": 4.538209095171121, "grad_norm": 0.5623694062232971, "learning_rate": 1.0354438572321546e-06, "loss": 0.3774, "step": 2420 }, { "epoch": 4.547585560243788, "grad_norm": 0.48698845505714417, "learning_rate": 9.938915491629063e-07, "loss": 0.4021, "step": 2425 }, { "epoch": 4.556962025316456, "grad_norm": 0.4880342483520508, "learning_rate": 9.531732406771771e-07, "loss": 0.3901, "step": 2430 }, { "epoch": 4.566338490389123, "grad_norm": 0.5080013871192932, "learning_rate": 9.132903463754256e-07, "loss": 0.4357, "step": 2435 }, { "epoch": 4.575714955461791, "grad_norm": 0.5003151893615723, "learning_rate": 8.742442518348965e-07, "loss": 0.3877, "step": 2440 }, { "epoch": 4.585091420534458, "grad_norm": 0.49231135845184326, "learning_rate": 8.360363135614307e-07, "loss": 0.3994, "step": 2445 }, { "epoch": 4.594467885607126, "grad_norm": 0.4957781732082367, "learning_rate": 7.986678589423758e-07, "loss": 0.4222, "step": 2450 }, { "epoch": 4.603844350679793, "grad_norm": 0.428396075963974, "learning_rate": 7.621401862004634e-07, "loss": 0.3991, "step": 2455 }, { "epoch": 4.613220815752461, "grad_norm": 0.49666279554367065, "learning_rate": 7.264545643486997e-07, "loss": 0.369, "step": 2460 }, { "epoch": 4.622597280825129, "grad_norm": 0.4541693925857544, "learning_rate": 6.916122331462799e-07, "loss": 0.3856, "step": 2465 }, { "epoch": 4.631973745897796, "grad_norm": 0.41727012395858765, "learning_rate": 6.576144030555259e-07, "loss": 0.4261, "step": 2470 }, { "epoch": 4.641350210970464, "grad_norm": 0.41427725553512573, "learning_rate": 6.244622551998203e-07, "loss": 0.3831, "step": 2475 }, { "epoch": 4.650726676043131, "grad_norm": 0.5246068239212036, "learning_rate": 5.921569413225913e-07, "loss": 0.4014, "step": 2480 }, { "epoch": 4.660103141115799, "grad_norm": 0.5102460384368896, "learning_rate": 5.606995837472817e-07, "loss": 0.3954, "step": 2485 }, { "epoch": 4.669479606188467, "grad_norm": 0.45667141675949097, "learning_rate": 5.300912753383625e-07, "loss": 0.4065, "step": 2490 }, { "epoch": 4.6788560712611345, "grad_norm": 0.4597122371196747, "learning_rate": 5.003330794633776e-07, "loss": 0.3947, "step": 2495 }, { "epoch": 4.688232536333802, "grad_norm": 0.5187001824378967, "learning_rate": 4.714260299559875e-07, "loss": 0.4113, "step": 2500 }, { "epoch": 4.6976090014064695, "grad_norm": 0.48345983028411865, "learning_rate": 4.4337113108005314e-07, "loss": 0.3972, "step": 2505 }, { "epoch": 4.7069854664791375, "grad_norm": 0.4524119198322296, "learning_rate": 4.161693574947556e-07, "loss": 0.4381, "step": 2510 }, { "epoch": 4.716361931551805, "grad_norm": 0.44738277792930603, "learning_rate": 3.8982165422073445e-07, "loss": 0.3794, "step": 2515 }, { "epoch": 4.725738396624473, "grad_norm": 0.44519349932670593, "learning_rate": 3.6432893660723886e-07, "loss": 0.4074, "step": 2520 }, { "epoch": 4.73511486169714, "grad_norm": 0.4711282253265381, "learning_rate": 3.396920903003559e-07, "loss": 0.3893, "step": 2525 }, { "epoch": 4.744491326769808, "grad_norm": 0.4641968905925751, "learning_rate": 3.1591197121222107e-07, "loss": 0.3983, "step": 2530 }, { "epoch": 4.753867791842476, "grad_norm": 0.4854121804237366, "learning_rate": 2.9298940549128964e-07, "loss": 0.3846, "step": 2535 }, { "epoch": 4.763244256915143, "grad_norm": 0.4864709973335266, "learning_rate": 2.7092518949362875e-07, "loss": 0.3793, "step": 2540 }, { "epoch": 4.772620721987811, "grad_norm": 0.4431411027908325, "learning_rate": 2.4972008975527593e-07, "loss": 0.4213, "step": 2545 }, { "epoch": 4.781997187060478, "grad_norm": 0.47686663269996643, "learning_rate": 2.2937484296556566e-07, "loss": 0.3905, "step": 2550 }, { "epoch": 4.791373652133146, "grad_norm": 0.4837285876274109, "learning_rate": 2.0989015594158058e-07, "loss": 0.4093, "step": 2555 }, { "epoch": 4.800750117205814, "grad_norm": 0.46795183420181274, "learning_rate": 1.9126670560356553e-07, "loss": 0.4119, "step": 2560 }, { "epoch": 4.810126582278481, "grad_norm": 0.49227696657180786, "learning_rate": 1.735051389514214e-07, "loss": 0.3885, "step": 2565 }, { "epoch": 4.819503047351149, "grad_norm": 0.45765581727027893, "learning_rate": 1.5660607304223141e-07, "loss": 0.3872, "step": 2570 }, { "epoch": 4.828879512423816, "grad_norm": 0.496937096118927, "learning_rate": 1.4057009496881158e-07, "loss": 0.404, "step": 2575 }, { "epoch": 4.838255977496484, "grad_norm": 0.5063027143478394, "learning_rate": 1.2539776183932982e-07, "loss": 0.4079, "step": 2580 }, { "epoch": 4.847632442569151, "grad_norm": 0.44478198885917664, "learning_rate": 1.1108960075794372e-07, "loss": 0.3629, "step": 2585 }, { "epoch": 4.857008907641819, "grad_norm": 0.5043444037437439, "learning_rate": 9.764610880648451e-08, "loss": 0.3688, "step": 2590 }, { "epoch": 4.866385372714487, "grad_norm": 0.47527217864990234, "learning_rate": 8.506775302719039e-08, "loss": 0.3782, "step": 2595 }, { "epoch": 4.875761837787154, "grad_norm": 0.46921849250793457, "learning_rate": 7.335497040648898e-08, "loss": 0.4003, "step": 2600 }, { "epoch": 4.885138302859822, "grad_norm": 0.4486777186393738, "learning_rate": 6.250816785980385e-08, "loss": 0.4175, "step": 2605 }, { "epoch": 4.894514767932489, "grad_norm": 0.5218645930290222, "learning_rate": 5.2527722217421416e-08, "loss": 0.3857, "step": 2610 }, { "epoch": 4.903891233005157, "grad_norm": 0.480591744184494, "learning_rate": 4.3413980211412516e-08, "loss": 0.3995, "step": 2615 }, { "epoch": 4.913267698077824, "grad_norm": 0.5079865455627441, "learning_rate": 3.516725846355873e-08, "loss": 0.369, "step": 2620 }, { "epoch": 4.922644163150492, "grad_norm": 0.5890299677848816, "learning_rate": 2.7787843474386123e-08, "loss": 0.3704, "step": 2625 }, { "epoch": 4.93202062822316, "grad_norm": 0.49126380681991577, "learning_rate": 2.127599161318161e-08, "loss": 0.4038, "step": 2630 }, { "epoch": 4.941397093295827, "grad_norm": 0.47775766253471375, "learning_rate": 1.5631929109102828e-08, "loss": 0.4069, "step": 2635 }, { "epoch": 4.950773558368495, "grad_norm": 0.48111775517463684, "learning_rate": 1.0855852043323289e-08, "loss": 0.4016, "step": 2640 }, { "epoch": 4.960150023441162, "grad_norm": 0.4834578335285187, "learning_rate": 6.947926342204536e-09, "loss": 0.3734, "step": 2645 }, { "epoch": 4.96952648851383, "grad_norm": 0.4474198818206787, "learning_rate": 3.908287771542396e-09, "loss": 0.3704, "step": 2650 }, { "epoch": 4.978902953586498, "grad_norm": 0.4454098045825958, "learning_rate": 1.737041931845762e-09, "loss": 0.4141, "step": 2655 }, { "epoch": 4.9882794186591655, "grad_norm": 0.48251986503601074, "learning_rate": 4.3426425467008035e-10, "loss": 0.4006, "step": 2660 }, { "epoch": 4.9976558837318334, "grad_norm": 0.4934045970439911, "learning_rate": 0.0, "loss": 0.3921, "step": 2665 }, { "epoch": 4.9976558837318334, "step": 2665, "total_flos": 1.9479197956994335e+18, "train_loss": 0.41369020724162375, "train_runtime": 63586.5105, "train_samples_per_second": 0.671, "train_steps_per_second": 0.042 } ], "logging_steps": 5, "max_steps": 2665, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.9479197956994335e+18, "train_batch_size": 2, "trial_name": null, "trial_params": null }