{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.16925467186292975, "eval_steps": 500, "global_step": 6500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0002603918028660458, "grad_norm": 0.4500846266746521, "learning_rate": 5.194805194805195e-06, "loss": 1.0381, "step": 10 }, { "epoch": 0.0005207836057320916, "grad_norm": 0.35188010334968567, "learning_rate": 1.038961038961039e-05, "loss": 1.0108, "step": 20 }, { "epoch": 0.0007811754085981374, "grad_norm": 0.2300374060869217, "learning_rate": 1.5584415584415583e-05, "loss": 0.9668, "step": 30 }, { "epoch": 0.0010415672114641832, "grad_norm": 0.16189467906951904, "learning_rate": 2.077922077922078e-05, "loss": 0.918, "step": 40 }, { "epoch": 0.001301959014330229, "grad_norm": 0.18843211233615875, "learning_rate": 2.5974025974025972e-05, "loss": 0.9265, "step": 50 }, { "epoch": 0.0015623508171962747, "grad_norm": 0.20334510505199432, "learning_rate": 3.1168831168831166e-05, "loss": 0.9234, "step": 60 }, { "epoch": 0.0018227426200623205, "grad_norm": 0.1745327115058899, "learning_rate": 3.6363636363636364e-05, "loss": 0.881, "step": 70 }, { "epoch": 0.0020831344229283663, "grad_norm": 0.18667331337928772, "learning_rate": 4.155844155844156e-05, "loss": 0.8592, "step": 80 }, { "epoch": 0.002343526225794412, "grad_norm": 0.1848158985376358, "learning_rate": 4.675324675324675e-05, "loss": 0.8537, "step": 90 }, { "epoch": 0.002603918028660458, "grad_norm": 0.17589879035949707, "learning_rate": 5.1948051948051944e-05, "loss": 0.8518, "step": 100 }, { "epoch": 0.0028643098315265037, "grad_norm": 0.2132624089717865, "learning_rate": 5.714285714285714e-05, "loss": 0.8511, "step": 110 }, { "epoch": 0.0031247016343925495, "grad_norm": 0.23070092499256134, "learning_rate": 6.233766233766233e-05, "loss": 0.7975, "step": 120 }, { "epoch": 0.0033850934372585953, "grad_norm": 0.25368157029151917, "learning_rate": 6.753246753246754e-05, "loss": 0.8134, "step": 130 }, { "epoch": 0.003645485240124641, "grad_norm": 0.22897231578826904, "learning_rate": 7.272727272727273e-05, "loss": 0.8322, "step": 140 }, { "epoch": 0.003905877042990687, "grad_norm": 0.19932536780834198, "learning_rate": 7.792207792207793e-05, "loss": 0.7959, "step": 150 }, { "epoch": 0.004166268845856733, "grad_norm": 0.21011792123317719, "learning_rate": 8.311688311688312e-05, "loss": 0.8102, "step": 160 }, { "epoch": 0.004426660648722778, "grad_norm": 0.20594824850559235, "learning_rate": 8.831168831168831e-05, "loss": 0.8128, "step": 170 }, { "epoch": 0.004687052451588824, "grad_norm": 0.20465536415576935, "learning_rate": 9.35064935064935e-05, "loss": 0.7989, "step": 180 }, { "epoch": 0.00494744425445487, "grad_norm": 0.4109392762184143, "learning_rate": 9.870129870129871e-05, "loss": 0.8108, "step": 190 }, { "epoch": 0.005207836057320916, "grad_norm": 0.4293076694011688, "learning_rate": 0.00010389610389610389, "loss": 0.8101, "step": 200 }, { "epoch": 0.005468227860186962, "grad_norm": 0.31628963351249695, "learning_rate": 0.00010909090909090909, "loss": 0.7989, "step": 210 }, { "epoch": 0.005728619663053007, "grad_norm": 0.24642810225486755, "learning_rate": 0.00011428571428571428, "loss": 0.7751, "step": 220 }, { "epoch": 0.005989011465919053, "grad_norm": 0.3599106967449188, "learning_rate": 0.00011948051948051949, "loss": 0.8063, "step": 230 }, { "epoch": 0.006249403268785099, "grad_norm": 0.17053447663784027, "learning_rate": 0.00012467532467532467, "loss": 0.7751, "step": 240 }, { "epoch": 0.006509795071651145, "grad_norm": 0.17303769290447235, "learning_rate": 0.00012987012987012987, "loss": 0.7883, "step": 250 }, { "epoch": 0.0067701868745171905, "grad_norm": 0.1815861016511917, "learning_rate": 0.00013506493506493507, "loss": 0.788, "step": 260 }, { "epoch": 0.007030578677383236, "grad_norm": 0.24125365912914276, "learning_rate": 0.00014025974025974028, "loss": 0.8018, "step": 270 }, { "epoch": 0.007290970480249282, "grad_norm": 0.19443446397781372, "learning_rate": 0.00014545454545454546, "loss": 0.7908, "step": 280 }, { "epoch": 0.007551362283115328, "grad_norm": 0.17829768359661102, "learning_rate": 0.00015064935064935066, "loss": 0.8033, "step": 290 }, { "epoch": 0.007811754085981374, "grad_norm": 0.19535653293132782, "learning_rate": 0.00015584415584415587, "loss": 0.7997, "step": 300 }, { "epoch": 0.008072145888847419, "grad_norm": 0.19930541515350342, "learning_rate": 0.00016103896103896104, "loss": 0.7945, "step": 310 }, { "epoch": 0.008332537691713465, "grad_norm": 0.2156297266483307, "learning_rate": 0.00016623376623376625, "loss": 0.8018, "step": 320 }, { "epoch": 0.00859292949457951, "grad_norm": 0.1924206018447876, "learning_rate": 0.00017142857142857143, "loss": 0.7746, "step": 330 }, { "epoch": 0.008853321297445557, "grad_norm": 0.2294880747795105, "learning_rate": 0.00017662337662337663, "loss": 0.8152, "step": 340 }, { "epoch": 0.009113713100311602, "grad_norm": 0.16817067563533783, "learning_rate": 0.00018181818181818183, "loss": 0.7972, "step": 350 }, { "epoch": 0.009374104903177648, "grad_norm": 0.18544812500476837, "learning_rate": 0.000187012987012987, "loss": 0.7801, "step": 360 }, { "epoch": 0.009634496706043693, "grad_norm": 0.19597066938877106, "learning_rate": 0.00019220779220779222, "loss": 0.7706, "step": 370 }, { "epoch": 0.00989488850890974, "grad_norm": 0.40291881561279297, "learning_rate": 0.00019740259740259742, "loss": 0.7911, "step": 380 }, { "epoch": 0.010155280311775785, "grad_norm": 0.23841074109077454, "learning_rate": 0.00019999996515752773, "loss": 0.7861, "step": 390 }, { "epoch": 0.010415672114641832, "grad_norm": 0.1675388514995575, "learning_rate": 0.00019999968641789507, "loss": 0.788, "step": 400 }, { "epoch": 0.010676063917507876, "grad_norm": 1.8860758543014526, "learning_rate": 0.0001999991289394067, "loss": 0.7632, "step": 410 }, { "epoch": 0.010936455720373923, "grad_norm": 0.17022117972373962, "learning_rate": 0.00019999829272361654, "loss": 0.784, "step": 420 }, { "epoch": 0.011196847523239968, "grad_norm": 0.21460269391536713, "learning_rate": 0.00019999717777285545, "loss": 0.761, "step": 430 }, { "epoch": 0.011457239326106015, "grad_norm": 0.19413785636425018, "learning_rate": 0.00019999578409023126, "loss": 0.7772, "step": 440 }, { "epoch": 0.01171763112897206, "grad_norm": 0.20223405957221985, "learning_rate": 0.00019999411167962868, "loss": 0.7811, "step": 450 }, { "epoch": 0.011978022931838106, "grad_norm": 0.15166303515434265, "learning_rate": 0.00019999216054570942, "loss": 0.7709, "step": 460 }, { "epoch": 0.012238414734704151, "grad_norm": 0.16307081282138824, "learning_rate": 0.00019998993069391205, "loss": 0.7811, "step": 470 }, { "epoch": 0.012498806537570198, "grad_norm": 0.15996049344539642, "learning_rate": 0.00019998742213045206, "loss": 0.7599, "step": 480 }, { "epoch": 0.012759198340436243, "grad_norm": 0.17560279369354248, "learning_rate": 0.00019998463486232179, "loss": 0.7572, "step": 490 }, { "epoch": 0.01301959014330229, "grad_norm": 0.17571642994880676, "learning_rate": 0.0001999815688972905, "loss": 0.7643, "step": 500 }, { "epoch": 0.013279981946168334, "grad_norm": 0.17719799280166626, "learning_rate": 0.00019997822424390422, "loss": 0.7923, "step": 510 }, { "epoch": 0.013540373749034381, "grad_norm": 0.19846616685390472, "learning_rate": 0.00019997460091148586, "loss": 0.7674, "step": 520 }, { "epoch": 0.013800765551900426, "grad_norm": 0.2715558111667633, "learning_rate": 0.00019997069891013503, "loss": 0.7421, "step": 530 }, { "epoch": 0.014061157354766473, "grad_norm": 0.1725197583436966, "learning_rate": 0.00019996651825072826, "loss": 0.7663, "step": 540 }, { "epoch": 0.014321549157632518, "grad_norm": 0.15060502290725708, "learning_rate": 0.00019996205894491856, "loss": 0.7794, "step": 550 }, { "epoch": 0.014581940960498564, "grad_norm": 0.16645808517932892, "learning_rate": 0.00019995732100513592, "loss": 0.752, "step": 560 }, { "epoch": 0.014842332763364609, "grad_norm": 0.1736789345741272, "learning_rate": 0.00019995230444458682, "loss": 0.7788, "step": 570 }, { "epoch": 0.015102724566230656, "grad_norm": 0.15416319668293, "learning_rate": 0.0001999470092772544, "loss": 0.7656, "step": 580 }, { "epoch": 0.0153631163690967, "grad_norm": 0.16610187292099, "learning_rate": 0.00019994143551789839, "loss": 0.7676, "step": 590 }, { "epoch": 0.015623508171962747, "grad_norm": 0.15843011438846588, "learning_rate": 0.00019993558318205507, "loss": 0.7746, "step": 600 }, { "epoch": 0.015883899974828794, "grad_norm": 0.26837801933288574, "learning_rate": 0.00019992945228603724, "loss": 0.7617, "step": 610 }, { "epoch": 0.016144291777694837, "grad_norm": 0.15099173784255981, "learning_rate": 0.0001999230428469341, "loss": 0.7601, "step": 620 }, { "epoch": 0.016404683580560884, "grad_norm": 0.15511856973171234, "learning_rate": 0.00019991635488261138, "loss": 0.7647, "step": 630 }, { "epoch": 0.01666507538342693, "grad_norm": 0.14919579029083252, "learning_rate": 0.00019990938841171104, "loss": 0.7692, "step": 640 }, { "epoch": 0.016925467186292977, "grad_norm": 0.15838642418384552, "learning_rate": 0.0001999021434536514, "loss": 0.7763, "step": 650 }, { "epoch": 0.01718585898915902, "grad_norm": 0.15956635773181915, "learning_rate": 0.00019989462002862704, "loss": 0.7598, "step": 660 }, { "epoch": 0.017446250792025067, "grad_norm": 0.1499069333076477, "learning_rate": 0.0001998868181576088, "loss": 0.7626, "step": 670 }, { "epoch": 0.017706642594891114, "grad_norm": 0.2170073390007019, "learning_rate": 0.00019987873786234348, "loss": 0.7569, "step": 680 }, { "epoch": 0.01796703439775716, "grad_norm": 0.17841948568820953, "learning_rate": 0.00019987037916535417, "loss": 0.7494, "step": 690 }, { "epoch": 0.018227426200623204, "grad_norm": 0.2066909372806549, "learning_rate": 0.0001998617420899398, "loss": 0.7609, "step": 700 }, { "epoch": 0.01848781800348925, "grad_norm": 0.17015361785888672, "learning_rate": 0.0001998528266601754, "loss": 0.7761, "step": 710 }, { "epoch": 0.018748209806355297, "grad_norm": 0.22166290879249573, "learning_rate": 0.0001998436329009118, "loss": 0.7573, "step": 720 }, { "epoch": 0.01900860160922134, "grad_norm": 0.15084640681743622, "learning_rate": 0.00019983416083777563, "loss": 0.7775, "step": 730 }, { "epoch": 0.019268993412087387, "grad_norm": 0.17800921201705933, "learning_rate": 0.0001998244104971693, "loss": 0.7359, "step": 740 }, { "epoch": 0.019529385214953433, "grad_norm": 0.17354707419872284, "learning_rate": 0.0001998143819062709, "loss": 0.7415, "step": 750 }, { "epoch": 0.01978977701781948, "grad_norm": 0.16408118605613708, "learning_rate": 0.00019980407509303413, "loss": 0.7708, "step": 760 }, { "epoch": 0.020050168820685523, "grad_norm": 0.16820089519023895, "learning_rate": 0.00019979349008618808, "loss": 0.791, "step": 770 }, { "epoch": 0.02031056062355157, "grad_norm": 0.15958388149738312, "learning_rate": 0.00019978262691523743, "loss": 0.7412, "step": 780 }, { "epoch": 0.020570952426417616, "grad_norm": 0.1646542251110077, "learning_rate": 0.00019977148561046217, "loss": 0.7529, "step": 790 }, { "epoch": 0.020831344229283663, "grad_norm": 0.17032025754451752, "learning_rate": 0.0001997600662029175, "loss": 0.7656, "step": 800 }, { "epoch": 0.021091736032149706, "grad_norm": 0.17189227044582367, "learning_rate": 0.00019974836872443388, "loss": 0.7433, "step": 810 }, { "epoch": 0.021352127835015753, "grad_norm": 0.16334249079227448, "learning_rate": 0.0001997363932076168, "loss": 0.7703, "step": 820 }, { "epoch": 0.0216125196378818, "grad_norm": 0.1676424890756607, "learning_rate": 0.00019972413968584682, "loss": 0.7603, "step": 830 }, { "epoch": 0.021872911440747846, "grad_norm": 0.16826209425926208, "learning_rate": 0.0001997116081932793, "loss": 0.7569, "step": 840 }, { "epoch": 0.02213330324361389, "grad_norm": 0.1876436173915863, "learning_rate": 0.0001996987987648446, "loss": 0.7553, "step": 850 }, { "epoch": 0.022393695046479936, "grad_norm": 0.17252250015735626, "learning_rate": 0.0001996857114362476, "loss": 0.7644, "step": 860 }, { "epoch": 0.022654086849345983, "grad_norm": 0.1632252335548401, "learning_rate": 0.00019967234624396793, "loss": 0.7568, "step": 870 }, { "epoch": 0.02291447865221203, "grad_norm": 0.1818259060382843, "learning_rate": 0.00019965870322525965, "loss": 0.7672, "step": 880 }, { "epoch": 0.023174870455078073, "grad_norm": 0.15418195724487305, "learning_rate": 0.0001996447824181513, "loss": 0.7642, "step": 890 }, { "epoch": 0.02343526225794412, "grad_norm": 0.17383505403995514, "learning_rate": 0.0001996305838614457, "loss": 0.7607, "step": 900 }, { "epoch": 0.023695654060810166, "grad_norm": 0.17794272303581238, "learning_rate": 0.00019961610759471984, "loss": 0.7588, "step": 910 }, { "epoch": 0.023956045863676213, "grad_norm": 0.1909121572971344, "learning_rate": 0.00019960135365832486, "loss": 0.7438, "step": 920 }, { "epoch": 0.024216437666542256, "grad_norm": 0.17758873105049133, "learning_rate": 0.00019958632209338587, "loss": 0.7323, "step": 930 }, { "epoch": 0.024476829469408302, "grad_norm": 0.15553662180900574, "learning_rate": 0.00019957101294180174, "loss": 0.7508, "step": 940 }, { "epoch": 0.02473722127227435, "grad_norm": 0.15310749411582947, "learning_rate": 0.00019955542624624522, "loss": 0.7563, "step": 950 }, { "epoch": 0.024997613075140396, "grad_norm": 0.1628728210926056, "learning_rate": 0.00019953956205016256, "loss": 0.7524, "step": 960 }, { "epoch": 0.02525800487800644, "grad_norm": 0.16211454570293427, "learning_rate": 0.00019952342039777362, "loss": 0.7564, "step": 970 }, { "epoch": 0.025518396680872486, "grad_norm": 0.15663012862205505, "learning_rate": 0.00019950700133407163, "loss": 0.7395, "step": 980 }, { "epoch": 0.025778788483738532, "grad_norm": 0.1684863567352295, "learning_rate": 0.00019949030490482296, "loss": 0.753, "step": 990 }, { "epoch": 0.02603918028660458, "grad_norm": 0.1561436653137207, "learning_rate": 0.0001994733311565673, "loss": 0.7409, "step": 1000 }, { "epoch": 0.026299572089470622, "grad_norm": 0.1781485229730606, "learning_rate": 0.0001994560801366171, "loss": 0.762, "step": 1010 }, { "epoch": 0.02655996389233667, "grad_norm": 0.15422071516513824, "learning_rate": 0.00019943855189305792, "loss": 0.7291, "step": 1020 }, { "epoch": 0.026820355695202715, "grad_norm": 0.17980527877807617, "learning_rate": 0.00019942074647474786, "loss": 0.7732, "step": 1030 }, { "epoch": 0.027080747498068762, "grad_norm": 0.15810626745224, "learning_rate": 0.00019940266393131775, "loss": 0.7764, "step": 1040 }, { "epoch": 0.027341139300934805, "grad_norm": 0.16385480761528015, "learning_rate": 0.00019938430431317081, "loss": 0.7404, "step": 1050 }, { "epoch": 0.027601531103800852, "grad_norm": 0.15134255588054657, "learning_rate": 0.00019936566767148257, "loss": 0.7506, "step": 1060 }, { "epoch": 0.0278619229066669, "grad_norm": 0.1592187136411667, "learning_rate": 0.00019934675405820077, "loss": 0.73, "step": 1070 }, { "epoch": 0.028122314709532945, "grad_norm": 0.16852422058582306, "learning_rate": 0.00019932756352604515, "loss": 0.7443, "step": 1080 }, { "epoch": 0.02838270651239899, "grad_norm": 0.15741507709026337, "learning_rate": 0.00019930809612850735, "loss": 0.7377, "step": 1090 }, { "epoch": 0.028643098315265035, "grad_norm": 0.22424879670143127, "learning_rate": 0.00019928835191985076, "loss": 0.7544, "step": 1100 }, { "epoch": 0.028903490118131082, "grad_norm": 0.2047310769557953, "learning_rate": 0.0001992683309551103, "loss": 0.7441, "step": 1110 }, { "epoch": 0.02916388192099713, "grad_norm": 0.16392463445663452, "learning_rate": 0.00019924803329009243, "loss": 0.7606, "step": 1120 }, { "epoch": 0.02942427372386317, "grad_norm": 0.16227149963378906, "learning_rate": 0.00019922745898137473, "loss": 0.736, "step": 1130 }, { "epoch": 0.029684665526729218, "grad_norm": 0.15652808547019958, "learning_rate": 0.00019920660808630598, "loss": 0.7513, "step": 1140 }, { "epoch": 0.029945057329595265, "grad_norm": 0.15162768959999084, "learning_rate": 0.00019918548066300592, "loss": 0.7303, "step": 1150 }, { "epoch": 0.03020544913246131, "grad_norm": 0.17650415003299713, "learning_rate": 0.0001991640767703651, "loss": 0.7254, "step": 1160 }, { "epoch": 0.030465840935327355, "grad_norm": 0.1594468355178833, "learning_rate": 0.00019914239646804462, "loss": 0.741, "step": 1170 }, { "epoch": 0.0307262327381934, "grad_norm": 0.17928367853164673, "learning_rate": 0.00019912043981647616, "loss": 0.7515, "step": 1180 }, { "epoch": 0.030986624541059448, "grad_norm": 0.17009998857975006, "learning_rate": 0.00019909820687686157, "loss": 0.7539, "step": 1190 }, { "epoch": 0.031247016343925495, "grad_norm": 0.16556763648986816, "learning_rate": 0.0001990756977111729, "loss": 0.7418, "step": 1200 }, { "epoch": 0.03150740814679154, "grad_norm": 0.1561640352010727, "learning_rate": 0.0001990529123821522, "loss": 0.7465, "step": 1210 }, { "epoch": 0.03176779994965759, "grad_norm": 0.15182287991046906, "learning_rate": 0.00019902985095331113, "loss": 0.7694, "step": 1220 }, { "epoch": 0.03202819175252363, "grad_norm": 0.15173685550689697, "learning_rate": 0.00019900651348893114, "loss": 0.7519, "step": 1230 }, { "epoch": 0.032288583555389674, "grad_norm": 0.16535787284374237, "learning_rate": 0.00019898290005406296, "loss": 0.7646, "step": 1240 }, { "epoch": 0.032548975358255725, "grad_norm": 0.19272534549236298, "learning_rate": 0.00019895901071452667, "loss": 0.7655, "step": 1250 }, { "epoch": 0.03280936716112177, "grad_norm": 0.1672705113887787, "learning_rate": 0.0001989348455369113, "loss": 0.7486, "step": 1260 }, { "epoch": 0.03306975896398781, "grad_norm": 0.1525493860244751, "learning_rate": 0.0001989104045885748, "loss": 0.7546, "step": 1270 }, { "epoch": 0.03333015076685386, "grad_norm": 0.16333037614822388, "learning_rate": 0.00019888568793764385, "loss": 0.7299, "step": 1280 }, { "epoch": 0.033590542569719904, "grad_norm": 0.1590205729007721, "learning_rate": 0.00019886069565301355, "loss": 0.762, "step": 1290 }, { "epoch": 0.033850934372585954, "grad_norm": 0.15006420016288757, "learning_rate": 0.00019883542780434733, "loss": 0.7531, "step": 1300 }, { "epoch": 0.034111326175452, "grad_norm": 0.18390792608261108, "learning_rate": 0.0001988098844620767, "loss": 0.7621, "step": 1310 }, { "epoch": 0.03437171797831804, "grad_norm": 0.17046166956424713, "learning_rate": 0.0001987840656974011, "loss": 0.7422, "step": 1320 }, { "epoch": 0.03463210978118409, "grad_norm": 0.15121813118457794, "learning_rate": 0.00019875797158228775, "loss": 0.7555, "step": 1330 }, { "epoch": 0.034892501584050134, "grad_norm": 0.16219307482242584, "learning_rate": 0.00019873160218947125, "loss": 0.7301, "step": 1340 }, { "epoch": 0.03515289338691618, "grad_norm": 0.1779986321926117, "learning_rate": 0.00019870495759245362, "loss": 0.7356, "step": 1350 }, { "epoch": 0.03541328518978223, "grad_norm": 0.16951359808444977, "learning_rate": 0.0001986780378655039, "loss": 0.7645, "step": 1360 }, { "epoch": 0.03567367699264827, "grad_norm": 0.16620802879333496, "learning_rate": 0.0001986508430836581, "loss": 0.7331, "step": 1370 }, { "epoch": 0.03593406879551432, "grad_norm": 0.1577858328819275, "learning_rate": 0.0001986233733227188, "loss": 0.7667, "step": 1380 }, { "epoch": 0.036194460598380364, "grad_norm": 0.1637091338634491, "learning_rate": 0.00019859562865925525, "loss": 0.7521, "step": 1390 }, { "epoch": 0.03645485240124641, "grad_norm": 0.15061691403388977, "learning_rate": 0.00019856760917060277, "loss": 0.744, "step": 1400 }, { "epoch": 0.03671524420411246, "grad_norm": 0.15373477339744568, "learning_rate": 0.00019853931493486287, "loss": 0.7677, "step": 1410 }, { "epoch": 0.0369756360069785, "grad_norm": 0.16468606889247894, "learning_rate": 0.00019851074603090277, "loss": 0.7179, "step": 1420 }, { "epoch": 0.037236027809844544, "grad_norm": 0.16084876656532288, "learning_rate": 0.00019848190253835536, "loss": 0.749, "step": 1430 }, { "epoch": 0.037496419612710594, "grad_norm": 0.16743004322052002, "learning_rate": 0.00019845278453761896, "loss": 0.7483, "step": 1440 }, { "epoch": 0.03775681141557664, "grad_norm": 0.17335088551044464, "learning_rate": 0.00019842339210985696, "loss": 0.735, "step": 1450 }, { "epoch": 0.03801720321844268, "grad_norm": 0.1546197235584259, "learning_rate": 0.00019839372533699774, "loss": 0.7549, "step": 1460 }, { "epoch": 0.03827759502130873, "grad_norm": 0.16218656301498413, "learning_rate": 0.00019836378430173438, "loss": 0.7425, "step": 1470 }, { "epoch": 0.03853798682417477, "grad_norm": 0.1712743639945984, "learning_rate": 0.0001983335690875245, "loss": 0.733, "step": 1480 }, { "epoch": 0.03879837862704082, "grad_norm": 0.15490613877773285, "learning_rate": 0.00019830307977858984, "loss": 0.7265, "step": 1490 }, { "epoch": 0.03905877042990687, "grad_norm": 0.1646670252084732, "learning_rate": 0.00019827231645991623, "loss": 0.7315, "step": 1500 }, { "epoch": 0.03931916223277291, "grad_norm": 0.1599082201719284, "learning_rate": 0.00019824127921725326, "loss": 0.7293, "step": 1510 }, { "epoch": 0.03957955403563896, "grad_norm": 0.1565747708082199, "learning_rate": 0.00019820996813711407, "loss": 0.7396, "step": 1520 }, { "epoch": 0.039839945838505, "grad_norm": 0.154826357960701, "learning_rate": 0.0001981783833067751, "loss": 0.7217, "step": 1530 }, { "epoch": 0.040100337641371046, "grad_norm": 0.16705222427845, "learning_rate": 0.0001981465248142758, "loss": 0.761, "step": 1540 }, { "epoch": 0.040360729444237096, "grad_norm": 0.15651623904705048, "learning_rate": 0.00019811439274841842, "loss": 0.7565, "step": 1550 }, { "epoch": 0.04062112124710314, "grad_norm": 0.16211090981960297, "learning_rate": 0.00019808198719876782, "loss": 0.7555, "step": 1560 }, { "epoch": 0.04088151304996919, "grad_norm": 0.16856881976127625, "learning_rate": 0.00019804930825565112, "loss": 0.7567, "step": 1570 }, { "epoch": 0.04114190485283523, "grad_norm": 0.1588718593120575, "learning_rate": 0.00019801635601015752, "loss": 0.729, "step": 1580 }, { "epoch": 0.041402296655701276, "grad_norm": 0.17078711092472076, "learning_rate": 0.00019798313055413808, "loss": 0.7418, "step": 1590 }, { "epoch": 0.041662688458567326, "grad_norm": 0.16652734577655792, "learning_rate": 0.00019794963198020525, "loss": 0.7341, "step": 1600 }, { "epoch": 0.04192308026143337, "grad_norm": 0.15535488724708557, "learning_rate": 0.00019791586038173296, "loss": 0.7396, "step": 1610 }, { "epoch": 0.04218347206429941, "grad_norm": 0.3506317734718323, "learning_rate": 0.00019788181585285602, "loss": 0.7345, "step": 1620 }, { "epoch": 0.04244386386716546, "grad_norm": 0.16875872015953064, "learning_rate": 0.00019784749848847003, "loss": 0.7214, "step": 1630 }, { "epoch": 0.042704255670031506, "grad_norm": 0.17675861716270447, "learning_rate": 0.0001978129083842312, "loss": 0.7431, "step": 1640 }, { "epoch": 0.042964647472897556, "grad_norm": 0.15601837635040283, "learning_rate": 0.00019777804563655583, "loss": 0.7215, "step": 1650 }, { "epoch": 0.0432250392757636, "grad_norm": 0.1874823123216629, "learning_rate": 0.00019774291034262026, "loss": 0.727, "step": 1660 }, { "epoch": 0.04348543107862964, "grad_norm": 0.17005637288093567, "learning_rate": 0.00019770750260036054, "loss": 0.7446, "step": 1670 }, { "epoch": 0.04374582288149569, "grad_norm": 0.17069579660892487, "learning_rate": 0.00019767182250847207, "loss": 0.7266, "step": 1680 }, { "epoch": 0.044006214684361736, "grad_norm": 0.16133156418800354, "learning_rate": 0.00019763587016640948, "loss": 0.7568, "step": 1690 }, { "epoch": 0.04426660648722778, "grad_norm": 0.16229428350925446, "learning_rate": 0.00019759964567438623, "loss": 0.7402, "step": 1700 }, { "epoch": 0.04452699829009383, "grad_norm": 0.1622512936592102, "learning_rate": 0.00019756314913337432, "loss": 0.7536, "step": 1710 }, { "epoch": 0.04478739009295987, "grad_norm": 0.2161218672990799, "learning_rate": 0.00019752638064510415, "loss": 0.723, "step": 1720 }, { "epoch": 0.04504778189582592, "grad_norm": 0.154169961810112, "learning_rate": 0.00019748934031206414, "loss": 0.7441, "step": 1730 }, { "epoch": 0.045308173698691966, "grad_norm": 0.15468057990074158, "learning_rate": 0.00019745202823750034, "loss": 0.7349, "step": 1740 }, { "epoch": 0.04556856550155801, "grad_norm": 0.2015281468629837, "learning_rate": 0.0001974144445254164, "loss": 0.726, "step": 1750 }, { "epoch": 0.04582895730442406, "grad_norm": 0.1931644082069397, "learning_rate": 0.00019737658928057302, "loss": 0.7604, "step": 1760 }, { "epoch": 0.0460893491072901, "grad_norm": 0.1528482288122177, "learning_rate": 0.00019733846260848776, "loss": 0.7408, "step": 1770 }, { "epoch": 0.046349740910156145, "grad_norm": 0.16370061039924622, "learning_rate": 0.0001973000646154349, "loss": 0.7647, "step": 1780 }, { "epoch": 0.046610132713022195, "grad_norm": 0.16271348297595978, "learning_rate": 0.00019726139540844484, "loss": 0.7212, "step": 1790 }, { "epoch": 0.04687052451588824, "grad_norm": 0.16218173503875732, "learning_rate": 0.00019722245509530401, "loss": 0.735, "step": 1800 }, { "epoch": 0.04713091631875429, "grad_norm": 0.17063820362091064, "learning_rate": 0.00019718324378455458, "loss": 0.7311, "step": 1810 }, { "epoch": 0.04739130812162033, "grad_norm": 0.1678459346294403, "learning_rate": 0.00019714376158549404, "loss": 0.7486, "step": 1820 }, { "epoch": 0.047651699924486375, "grad_norm": 0.15926459431648254, "learning_rate": 0.00019710400860817494, "loss": 0.743, "step": 1830 }, { "epoch": 0.047912091727352425, "grad_norm": 0.1775251179933548, "learning_rate": 0.00019706398496340463, "loss": 0.7512, "step": 1840 }, { "epoch": 0.04817248353021847, "grad_norm": 0.1572408229112625, "learning_rate": 0.00019702369076274494, "loss": 0.733, "step": 1850 }, { "epoch": 0.04843287533308451, "grad_norm": 0.29658186435699463, "learning_rate": 0.0001969831261185118, "loss": 0.7297, "step": 1860 }, { "epoch": 0.04869326713595056, "grad_norm": 0.16520118713378906, "learning_rate": 0.00019694229114377494, "loss": 0.721, "step": 1870 }, { "epoch": 0.048953658938816605, "grad_norm": 0.17762574553489685, "learning_rate": 0.00019690118595235774, "loss": 0.7304, "step": 1880 }, { "epoch": 0.049214050741682655, "grad_norm": 0.16636615991592407, "learning_rate": 0.00019685981065883663, "loss": 0.7257, "step": 1890 }, { "epoch": 0.0494744425445487, "grad_norm": 0.1622323989868164, "learning_rate": 0.00019681816537854102, "loss": 0.7353, "step": 1900 }, { "epoch": 0.04973483434741474, "grad_norm": 0.17419832944869995, "learning_rate": 0.00019677625022755289, "loss": 0.7452, "step": 1910 }, { "epoch": 0.04999522615028079, "grad_norm": 0.17460434138774872, "learning_rate": 0.00019673406532270634, "loss": 0.7391, "step": 1920 }, { "epoch": 0.050255617953146835, "grad_norm": 0.15844550728797913, "learning_rate": 0.00019669161078158753, "loss": 0.7327, "step": 1930 }, { "epoch": 0.05051600975601288, "grad_norm": 0.1638839989900589, "learning_rate": 0.0001966488867225341, "loss": 0.745, "step": 1940 }, { "epoch": 0.05077640155887893, "grad_norm": 0.1587786227464676, "learning_rate": 0.00019660589326463498, "loss": 0.7476, "step": 1950 }, { "epoch": 0.05103679336174497, "grad_norm": 0.15708380937576294, "learning_rate": 0.00019656263052773002, "loss": 0.7208, "step": 1960 }, { "epoch": 0.05129718516461102, "grad_norm": 0.15816234052181244, "learning_rate": 0.00019651909863240965, "loss": 0.7262, "step": 1970 }, { "epoch": 0.051557576967477065, "grad_norm": 0.16749270260334015, "learning_rate": 0.00019647529770001456, "loss": 0.7284, "step": 1980 }, { "epoch": 0.05181796877034311, "grad_norm": 0.16943767666816711, "learning_rate": 0.00019643122785263536, "loss": 0.7225, "step": 1990 }, { "epoch": 0.05207836057320916, "grad_norm": 0.42929205298423767, "learning_rate": 0.00019638688921311224, "loss": 0.7305, "step": 2000 }, { "epoch": 0.0523387523760752, "grad_norm": 0.15851692855358124, "learning_rate": 0.00019634228190503454, "loss": 0.7344, "step": 2010 }, { "epoch": 0.052599144178941244, "grad_norm": 0.16053883731365204, "learning_rate": 0.00019629740605274062, "loss": 0.7468, "step": 2020 }, { "epoch": 0.052859535981807294, "grad_norm": 0.16504009068012238, "learning_rate": 0.00019625226178131728, "loss": 0.7375, "step": 2030 }, { "epoch": 0.05311992778467334, "grad_norm": 0.1618044674396515, "learning_rate": 0.00019620684921659953, "loss": 0.7201, "step": 2040 }, { "epoch": 0.05338031958753939, "grad_norm": 0.15512776374816895, "learning_rate": 0.00019616116848517027, "loss": 0.7355, "step": 2050 }, { "epoch": 0.05364071139040543, "grad_norm": 0.17377036809921265, "learning_rate": 0.00019611521971435979, "loss": 0.7226, "step": 2060 }, { "epoch": 0.053901103193271474, "grad_norm": 0.1685250997543335, "learning_rate": 0.0001960690030322456, "loss": 0.7483, "step": 2070 }, { "epoch": 0.054161494996137524, "grad_norm": 0.18394522368907928, "learning_rate": 0.00019602251856765194, "loss": 0.7385, "step": 2080 }, { "epoch": 0.05442188679900357, "grad_norm": 0.1753673106431961, "learning_rate": 0.0001959757664501495, "loss": 0.7378, "step": 2090 }, { "epoch": 0.05468227860186961, "grad_norm": 0.1795465052127838, "learning_rate": 0.000195928746810055, "loss": 0.748, "step": 2100 }, { "epoch": 0.05494267040473566, "grad_norm": 0.16327305138111115, "learning_rate": 0.0001958814597784309, "loss": 0.7306, "step": 2110 }, { "epoch": 0.055203062207601704, "grad_norm": 0.15880291163921356, "learning_rate": 0.00019583390548708486, "loss": 0.7281, "step": 2120 }, { "epoch": 0.05546345401046775, "grad_norm": 0.1702323853969574, "learning_rate": 0.0001957860840685696, "loss": 0.7407, "step": 2130 }, { "epoch": 0.0557238458133338, "grad_norm": 0.16931670904159546, "learning_rate": 0.0001957379956561825, "loss": 0.7272, "step": 2140 }, { "epoch": 0.05598423761619984, "grad_norm": 0.15455976128578186, "learning_rate": 0.000195689640383965, "loss": 0.7398, "step": 2150 }, { "epoch": 0.05624462941906589, "grad_norm": 0.16061417758464813, "learning_rate": 0.0001956410183867024, "loss": 0.749, "step": 2160 }, { "epoch": 0.056505021221931934, "grad_norm": 0.14933143556118011, "learning_rate": 0.00019559212979992365, "loss": 0.7418, "step": 2170 }, { "epoch": 0.05676541302479798, "grad_norm": 0.1592816412448883, "learning_rate": 0.00019554297475990058, "loss": 0.7423, "step": 2180 }, { "epoch": 0.05702580482766403, "grad_norm": 0.1677238792181015, "learning_rate": 0.00019549355340364787, "loss": 0.7101, "step": 2190 }, { "epoch": 0.05728619663053007, "grad_norm": 0.3558599054813385, "learning_rate": 0.00019544386586892238, "loss": 0.725, "step": 2200 }, { "epoch": 0.05754658843339611, "grad_norm": 0.1746376007795334, "learning_rate": 0.00019539391229422313, "loss": 0.7479, "step": 2210 }, { "epoch": 0.057806980236262163, "grad_norm": 0.15979182720184326, "learning_rate": 0.00019534369281879049, "loss": 0.7352, "step": 2220 }, { "epoch": 0.05806737203912821, "grad_norm": 0.16173166036605835, "learning_rate": 0.0001952932075826061, "loss": 0.7364, "step": 2230 }, { "epoch": 0.05832776384199426, "grad_norm": 0.1514744907617569, "learning_rate": 0.00019524245672639245, "loss": 0.734, "step": 2240 }, { "epoch": 0.0585881556448603, "grad_norm": 0.16860373318195343, "learning_rate": 0.00019519144039161222, "loss": 0.7098, "step": 2250 }, { "epoch": 0.05884854744772634, "grad_norm": 0.16847743093967438, "learning_rate": 0.00019514015872046833, "loss": 0.7103, "step": 2260 }, { "epoch": 0.05910893925059239, "grad_norm": 0.16181516647338867, "learning_rate": 0.00019508861185590307, "loss": 0.7561, "step": 2270 }, { "epoch": 0.059369331053458436, "grad_norm": 0.16594484448432922, "learning_rate": 0.0001950367999415981, "loss": 0.7308, "step": 2280 }, { "epoch": 0.05962972285632448, "grad_norm": 0.166441410779953, "learning_rate": 0.00019498472312197375, "loss": 0.735, "step": 2290 }, { "epoch": 0.05989011465919053, "grad_norm": 0.16273920238018036, "learning_rate": 0.00019493238154218886, "loss": 0.7458, "step": 2300 }, { "epoch": 0.06015050646205657, "grad_norm": 0.16227276623249054, "learning_rate": 0.00019487977534814012, "loss": 0.7143, "step": 2310 }, { "epoch": 0.06041089826492262, "grad_norm": 0.1619606912136078, "learning_rate": 0.000194826904686462, "loss": 0.7285, "step": 2320 }, { "epoch": 0.060671290067788666, "grad_norm": 0.1596045345067978, "learning_rate": 0.00019477376970452603, "loss": 0.7513, "step": 2330 }, { "epoch": 0.06093168187065471, "grad_norm": 0.17504757642745972, "learning_rate": 0.00019472037055044044, "loss": 0.7376, "step": 2340 }, { "epoch": 0.06119207367352076, "grad_norm": 0.1559167355298996, "learning_rate": 0.00019466670737304992, "loss": 0.7339, "step": 2350 }, { "epoch": 0.0614524654763868, "grad_norm": 0.1624836027622223, "learning_rate": 0.0001946127803219351, "loss": 0.7258, "step": 2360 }, { "epoch": 0.061712857279252846, "grad_norm": 0.17907138168811798, "learning_rate": 0.00019455858954741206, "loss": 0.72, "step": 2370 }, { "epoch": 0.061973249082118896, "grad_norm": 0.15922705829143524, "learning_rate": 0.00019450413520053202, "loss": 0.7187, "step": 2380 }, { "epoch": 0.06223364088498494, "grad_norm": 0.1552513986825943, "learning_rate": 0.0001944494174330809, "loss": 0.7183, "step": 2390 }, { "epoch": 0.06249403268785099, "grad_norm": 0.16838514804840088, "learning_rate": 0.00019439443639757885, "loss": 0.7286, "step": 2400 }, { "epoch": 0.06275442449071703, "grad_norm": 0.17352423071861267, "learning_rate": 0.00019433919224727986, "loss": 0.7436, "step": 2410 }, { "epoch": 0.06301481629358308, "grad_norm": 0.17366603016853333, "learning_rate": 0.0001942836851361713, "loss": 0.7265, "step": 2420 }, { "epoch": 0.06327520809644913, "grad_norm": 0.14833413064479828, "learning_rate": 0.00019422791521897357, "loss": 0.7234, "step": 2430 }, { "epoch": 0.06353559989931518, "grad_norm": 0.16602723300457, "learning_rate": 0.00019417188265113958, "loss": 0.725, "step": 2440 }, { "epoch": 0.06379599170218121, "grad_norm": 0.17290353775024414, "learning_rate": 0.00019411558758885438, "loss": 0.7174, "step": 2450 }, { "epoch": 0.06405638350504726, "grad_norm": 0.16486665606498718, "learning_rate": 0.0001940590301890346, "loss": 0.7301, "step": 2460 }, { "epoch": 0.06431677530791331, "grad_norm": 0.16255232691764832, "learning_rate": 0.00019400221060932827, "loss": 0.7462, "step": 2470 }, { "epoch": 0.06457716711077935, "grad_norm": 0.16139757633209229, "learning_rate": 0.0001939451290081141, "loss": 0.7424, "step": 2480 }, { "epoch": 0.0648375589136454, "grad_norm": 0.165597602725029, "learning_rate": 0.00019388778554450117, "loss": 0.7426, "step": 2490 }, { "epoch": 0.06509795071651145, "grad_norm": 0.19819000363349915, "learning_rate": 0.00019383018037832854, "loss": 0.7356, "step": 2500 }, { "epoch": 0.06535834251937749, "grad_norm": 0.16469696164131165, "learning_rate": 0.00019377231367016467, "loss": 0.718, "step": 2510 }, { "epoch": 0.06561873432224354, "grad_norm": 0.1644965261220932, "learning_rate": 0.00019371418558130702, "loss": 0.7253, "step": 2520 }, { "epoch": 0.06587912612510959, "grad_norm": 0.15347526967525482, "learning_rate": 0.00019365579627378174, "loss": 0.7214, "step": 2530 }, { "epoch": 0.06613951792797562, "grad_norm": 0.1618672013282776, "learning_rate": 0.00019359714591034302, "loss": 0.7204, "step": 2540 }, { "epoch": 0.06639990973084167, "grad_norm": 0.17043665051460266, "learning_rate": 0.00019353823465447268, "loss": 0.7278, "step": 2550 }, { "epoch": 0.06666030153370772, "grad_norm": 0.15762579441070557, "learning_rate": 0.00019347906267037983, "loss": 0.7283, "step": 2560 }, { "epoch": 0.06692069333657376, "grad_norm": 0.1622801572084427, "learning_rate": 0.00019341963012300029, "loss": 0.7193, "step": 2570 }, { "epoch": 0.06718108513943981, "grad_norm": 0.16705769300460815, "learning_rate": 0.00019335993717799617, "loss": 0.7414, "step": 2580 }, { "epoch": 0.06744147694230586, "grad_norm": 0.15886452794075012, "learning_rate": 0.00019329998400175545, "loss": 0.7242, "step": 2590 }, { "epoch": 0.06770186874517191, "grad_norm": 0.17994090914726257, "learning_rate": 0.00019323977076139142, "loss": 0.7017, "step": 2600 }, { "epoch": 0.06796226054803794, "grad_norm": 0.1609068214893341, "learning_rate": 0.00019317929762474232, "loss": 0.7352, "step": 2610 }, { "epoch": 0.068222652350904, "grad_norm": 0.15605950355529785, "learning_rate": 0.0001931185647603708, "loss": 0.7249, "step": 2620 }, { "epoch": 0.06848304415377005, "grad_norm": 0.16057750582695007, "learning_rate": 0.00019305757233756352, "loss": 0.7521, "step": 2630 }, { "epoch": 0.06874343595663608, "grad_norm": 0.1703862100839615, "learning_rate": 0.00019299632052633054, "loss": 0.7245, "step": 2640 }, { "epoch": 0.06900382775950213, "grad_norm": 0.16324444115161896, "learning_rate": 0.00019293480949740505, "loss": 0.7395, "step": 2650 }, { "epoch": 0.06926421956236818, "grad_norm": 0.15283791720867157, "learning_rate": 0.00019287303942224266, "loss": 0.7158, "step": 2660 }, { "epoch": 0.06952461136523422, "grad_norm": 0.1882282942533493, "learning_rate": 0.00019281101047302114, "loss": 0.724, "step": 2670 }, { "epoch": 0.06978500316810027, "grad_norm": 0.16147953271865845, "learning_rate": 0.00019274872282263984, "loss": 0.7365, "step": 2680 }, { "epoch": 0.07004539497096632, "grad_norm": 0.1614103466272354, "learning_rate": 0.00019268617664471916, "loss": 0.7206, "step": 2690 }, { "epoch": 0.07030578677383235, "grad_norm": 0.16784432530403137, "learning_rate": 0.00019262337211360016, "loss": 0.7279, "step": 2700 }, { "epoch": 0.0705661785766984, "grad_norm": 0.15966112911701202, "learning_rate": 0.000192560309404344, "loss": 0.7274, "step": 2710 }, { "epoch": 0.07082657037956445, "grad_norm": 0.16970521211624146, "learning_rate": 0.0001924969886927315, "loss": 0.7038, "step": 2720 }, { "epoch": 0.07108696218243049, "grad_norm": 0.16143856942653656, "learning_rate": 0.00019243341015526272, "loss": 0.7097, "step": 2730 }, { "epoch": 0.07134735398529654, "grad_norm": 0.16041269898414612, "learning_rate": 0.00019236957396915623, "loss": 0.722, "step": 2740 }, { "epoch": 0.07160774578816259, "grad_norm": 0.15845969319343567, "learning_rate": 0.00019230548031234882, "loss": 0.7238, "step": 2750 }, { "epoch": 0.07186813759102864, "grad_norm": 0.14966030418872833, "learning_rate": 0.00019224112936349502, "loss": 0.7182, "step": 2760 }, { "epoch": 0.07212852939389468, "grad_norm": 0.16525116562843323, "learning_rate": 0.00019217652130196653, "loss": 0.7397, "step": 2770 }, { "epoch": 0.07238892119676073, "grad_norm": 0.18119119107723236, "learning_rate": 0.0001921116563078516, "loss": 0.7222, "step": 2780 }, { "epoch": 0.07264931299962678, "grad_norm": 0.1709197610616684, "learning_rate": 0.00019204653456195478, "loss": 0.7068, "step": 2790 }, { "epoch": 0.07290970480249281, "grad_norm": 0.16309161484241486, "learning_rate": 0.00019198115624579625, "loss": 0.7349, "step": 2800 }, { "epoch": 0.07317009660535886, "grad_norm": 0.1736750453710556, "learning_rate": 0.00019191552154161135, "loss": 0.7445, "step": 2810 }, { "epoch": 0.07343048840822491, "grad_norm": 0.15009112656116486, "learning_rate": 0.00019184963063235006, "loss": 0.7034, "step": 2820 }, { "epoch": 0.07369088021109095, "grad_norm": 0.17244628071784973, "learning_rate": 0.0001917834837016766, "loss": 0.7285, "step": 2830 }, { "epoch": 0.073951272013957, "grad_norm": 0.15991820394992828, "learning_rate": 0.00019171708093396861, "loss": 0.7096, "step": 2840 }, { "epoch": 0.07421166381682305, "grad_norm": 0.17037667334079742, "learning_rate": 0.0001916504225143171, "loss": 0.7177, "step": 2850 }, { "epoch": 0.07447205561968909, "grad_norm": 0.16700348258018494, "learning_rate": 0.00019158350862852553, "loss": 0.7453, "step": 2860 }, { "epoch": 0.07473244742255514, "grad_norm": 0.17683659493923187, "learning_rate": 0.00019151633946310948, "loss": 0.7331, "step": 2870 }, { "epoch": 0.07499283922542119, "grad_norm": 0.16364306211471558, "learning_rate": 0.00019144891520529608, "loss": 0.7347, "step": 2880 }, { "epoch": 0.07525323102828722, "grad_norm": 0.1781424731016159, "learning_rate": 0.00019138123604302355, "loss": 0.7169, "step": 2890 }, { "epoch": 0.07551362283115327, "grad_norm": 0.16007259488105774, "learning_rate": 0.00019131330216494064, "loss": 0.7269, "step": 2900 }, { "epoch": 0.07577401463401932, "grad_norm": 0.1604921519756317, "learning_rate": 0.00019124511376040598, "loss": 0.7094, "step": 2910 }, { "epoch": 0.07603440643688536, "grad_norm": 0.16649965941905975, "learning_rate": 0.00019117667101948782, "loss": 0.7271, "step": 2920 }, { "epoch": 0.07629479823975141, "grad_norm": 0.16084066033363342, "learning_rate": 0.0001911079741329632, "loss": 0.7239, "step": 2930 }, { "epoch": 0.07655519004261746, "grad_norm": 0.1651066243648529, "learning_rate": 0.0001910390232923177, "loss": 0.7304, "step": 2940 }, { "epoch": 0.07681558184548351, "grad_norm": 0.1528957635164261, "learning_rate": 0.00019096981868974467, "loss": 0.7068, "step": 2950 }, { "epoch": 0.07707597364834955, "grad_norm": 0.172830730676651, "learning_rate": 0.00019090036051814483, "loss": 0.7277, "step": 2960 }, { "epoch": 0.0773363654512156, "grad_norm": 0.15909147262573242, "learning_rate": 0.00019083064897112571, "loss": 0.7135, "step": 2970 }, { "epoch": 0.07759675725408165, "grad_norm": 0.16273066401481628, "learning_rate": 0.0001907606842430011, "loss": 0.7346, "step": 2980 }, { "epoch": 0.07785714905694768, "grad_norm": 0.1595291793346405, "learning_rate": 0.00019069046652879049, "loss": 0.7377, "step": 2990 }, { "epoch": 0.07811754085981373, "grad_norm": 0.15573470294475555, "learning_rate": 0.0001906199960242185, "loss": 0.7026, "step": 3000 }, { "epoch": 0.07837793266267978, "grad_norm": 0.1670667678117752, "learning_rate": 0.0001905492729257145, "loss": 0.7231, "step": 3010 }, { "epoch": 0.07863832446554582, "grad_norm": 0.17074571549892426, "learning_rate": 0.00019047829743041184, "loss": 0.7003, "step": 3020 }, { "epoch": 0.07889871626841187, "grad_norm": 0.16979442536830902, "learning_rate": 0.00019040706973614738, "loss": 0.7217, "step": 3030 }, { "epoch": 0.07915910807127792, "grad_norm": 0.15843816101551056, "learning_rate": 0.00019033559004146103, "loss": 0.7334, "step": 3040 }, { "epoch": 0.07941949987414396, "grad_norm": 0.1607016921043396, "learning_rate": 0.0001902638585455951, "loss": 0.7271, "step": 3050 }, { "epoch": 0.07967989167701, "grad_norm": 0.1619115173816681, "learning_rate": 0.0001901918754484938, "loss": 0.7144, "step": 3060 }, { "epoch": 0.07994028347987606, "grad_norm": 0.1638360172510147, "learning_rate": 0.00019011964095080254, "loss": 0.7149, "step": 3070 }, { "epoch": 0.08020067528274209, "grad_norm": 0.16503652930259705, "learning_rate": 0.00019004715525386764, "loss": 0.7011, "step": 3080 }, { "epoch": 0.08046106708560814, "grad_norm": 0.16763822734355927, "learning_rate": 0.00018997441855973552, "loss": 0.7145, "step": 3090 }, { "epoch": 0.08072145888847419, "grad_norm": 0.1621125340461731, "learning_rate": 0.0001899014310711522, "loss": 0.7318, "step": 3100 }, { "epoch": 0.08098185069134024, "grad_norm": 0.16480112075805664, "learning_rate": 0.0001898281929915629, "loss": 0.7145, "step": 3110 }, { "epoch": 0.08124224249420628, "grad_norm": 0.1805388629436493, "learning_rate": 0.00018975470452511112, "loss": 0.7102, "step": 3120 }, { "epoch": 0.08150263429707233, "grad_norm": 0.1902652531862259, "learning_rate": 0.00018968096587663853, "loss": 0.7281, "step": 3130 }, { "epoch": 0.08176302609993838, "grad_norm": 0.1732487976551056, "learning_rate": 0.00018960697725168397, "loss": 0.7434, "step": 3140 }, { "epoch": 0.08202341790280442, "grad_norm": 0.1662171334028244, "learning_rate": 0.00018953273885648314, "loss": 0.716, "step": 3150 }, { "epoch": 0.08228380970567047, "grad_norm": 0.16129222512245178, "learning_rate": 0.00018945825089796797, "loss": 0.7318, "step": 3160 }, { "epoch": 0.08254420150853652, "grad_norm": 0.16837772727012634, "learning_rate": 0.00018938351358376596, "loss": 0.7137, "step": 3170 }, { "epoch": 0.08280459331140255, "grad_norm": 0.1618524193763733, "learning_rate": 0.00018930852712219974, "loss": 0.7079, "step": 3180 }, { "epoch": 0.0830649851142686, "grad_norm": 0.16333432495594025, "learning_rate": 0.00018923329172228632, "loss": 0.7062, "step": 3190 }, { "epoch": 0.08332537691713465, "grad_norm": 0.15985700488090515, "learning_rate": 0.00018915780759373672, "loss": 0.7277, "step": 3200 }, { "epoch": 0.08358576872000069, "grad_norm": 0.16181236505508423, "learning_rate": 0.0001890820749469551, "loss": 0.7048, "step": 3210 }, { "epoch": 0.08384616052286674, "grad_norm": 0.1679672747850418, "learning_rate": 0.00018900609399303853, "loss": 0.73, "step": 3220 }, { "epoch": 0.08410655232573279, "grad_norm": 0.16680286824703217, "learning_rate": 0.00018892986494377606, "loss": 0.7169, "step": 3230 }, { "epoch": 0.08436694412859883, "grad_norm": 0.15980315208435059, "learning_rate": 0.00018885338801164834, "loss": 0.7346, "step": 3240 }, { "epoch": 0.08462733593146488, "grad_norm": 0.16863352060317993, "learning_rate": 0.00018877666340982695, "loss": 0.7256, "step": 3250 }, { "epoch": 0.08488772773433093, "grad_norm": 0.1657836139202118, "learning_rate": 0.0001886996913521739, "loss": 0.7219, "step": 3260 }, { "epoch": 0.08514811953719698, "grad_norm": 0.1683470606803894, "learning_rate": 0.00018862247205324087, "loss": 0.7279, "step": 3270 }, { "epoch": 0.08540851134006301, "grad_norm": 0.1686122715473175, "learning_rate": 0.00018854500572826867, "loss": 0.7178, "step": 3280 }, { "epoch": 0.08566890314292906, "grad_norm": 0.16122782230377197, "learning_rate": 0.00018846729259318682, "loss": 0.7289, "step": 3290 }, { "epoch": 0.08592929494579511, "grad_norm": 0.1920589804649353, "learning_rate": 0.0001883893328646126, "loss": 0.7264, "step": 3300 }, { "epoch": 0.08618968674866115, "grad_norm": 0.17415335774421692, "learning_rate": 0.00018831112675985083, "loss": 0.7378, "step": 3310 }, { "epoch": 0.0864500785515272, "grad_norm": 0.16903561353683472, "learning_rate": 0.00018823267449689292, "loss": 0.702, "step": 3320 }, { "epoch": 0.08671047035439325, "grad_norm": 0.17158570885658264, "learning_rate": 0.00018815397629441658, "loss": 0.7264, "step": 3330 }, { "epoch": 0.08697086215725928, "grad_norm": 0.17094087600708008, "learning_rate": 0.0001880750323717849, "loss": 0.741, "step": 3340 }, { "epoch": 0.08723125396012534, "grad_norm": 0.1618223935365677, "learning_rate": 0.000187995842949046, "loss": 0.7084, "step": 3350 }, { "epoch": 0.08749164576299139, "grad_norm": 0.1671626716852188, "learning_rate": 0.0001879164082469322, "loss": 0.7236, "step": 3360 }, { "epoch": 0.08775203756585742, "grad_norm": 0.1681569665670395, "learning_rate": 0.00018783672848685966, "loss": 0.7382, "step": 3370 }, { "epoch": 0.08801242936872347, "grad_norm": 0.16421955823898315, "learning_rate": 0.00018775680389092748, "loss": 0.717, "step": 3380 }, { "epoch": 0.08827282117158952, "grad_norm": 0.174809530377388, "learning_rate": 0.00018767663468191725, "loss": 0.7225, "step": 3390 }, { "epoch": 0.08853321297445556, "grad_norm": 0.16181902587413788, "learning_rate": 0.00018759622108329243, "loss": 0.7014, "step": 3400 }, { "epoch": 0.08879360477732161, "grad_norm": 0.15579254925251007, "learning_rate": 0.0001875155633191977, "loss": 0.7125, "step": 3410 }, { "epoch": 0.08905399658018766, "grad_norm": 0.16342496871948242, "learning_rate": 0.00018743466161445823, "loss": 0.7075, "step": 3420 }, { "epoch": 0.08931438838305371, "grad_norm": 0.17215611040592194, "learning_rate": 0.00018735351619457923, "loss": 0.7331, "step": 3430 }, { "epoch": 0.08957478018591974, "grad_norm": 0.1682904213666916, "learning_rate": 0.00018727212728574522, "loss": 0.734, "step": 3440 }, { "epoch": 0.0898351719887858, "grad_norm": 0.16969889402389526, "learning_rate": 0.00018719049511481948, "loss": 0.7224, "step": 3450 }, { "epoch": 0.09009556379165184, "grad_norm": 0.16607950627803802, "learning_rate": 0.00018710861990934324, "loss": 0.7218, "step": 3460 }, { "epoch": 0.09035595559451788, "grad_norm": 0.16665585339069366, "learning_rate": 0.00018702650189753525, "loss": 0.7152, "step": 3470 }, { "epoch": 0.09061634739738393, "grad_norm": 0.16812992095947266, "learning_rate": 0.00018694414130829103, "loss": 0.7097, "step": 3480 }, { "epoch": 0.09087673920024998, "grad_norm": 0.16855508089065552, "learning_rate": 0.00018686153837118224, "loss": 0.7268, "step": 3490 }, { "epoch": 0.09113713100311602, "grad_norm": 0.1634734570980072, "learning_rate": 0.00018677869331645613, "loss": 0.7485, "step": 3500 }, { "epoch": 0.09139752280598207, "grad_norm": 0.17040792107582092, "learning_rate": 0.00018669560637503478, "loss": 0.7529, "step": 3510 }, { "epoch": 0.09165791460884812, "grad_norm": 0.16575287282466888, "learning_rate": 0.00018661227777851447, "loss": 0.7053, "step": 3520 }, { "epoch": 0.09191830641171415, "grad_norm": 0.16195647418498993, "learning_rate": 0.00018652870775916515, "loss": 0.7498, "step": 3530 }, { "epoch": 0.0921786982145802, "grad_norm": 0.16672903299331665, "learning_rate": 0.00018644489654992956, "loss": 0.7168, "step": 3540 }, { "epoch": 0.09243909001744625, "grad_norm": 0.16135342419147491, "learning_rate": 0.00018636084438442298, "loss": 0.6868, "step": 3550 }, { "epoch": 0.09269948182031229, "grad_norm": 0.16677817702293396, "learning_rate": 0.0001862765514969321, "loss": 0.7128, "step": 3560 }, { "epoch": 0.09295987362317834, "grad_norm": 0.16983947157859802, "learning_rate": 0.00018619201812241474, "loss": 0.7189, "step": 3570 }, { "epoch": 0.09322026542604439, "grad_norm": 0.17054444551467896, "learning_rate": 0.00018610724449649895, "loss": 0.6916, "step": 3580 }, { "epoch": 0.09348065722891043, "grad_norm": 0.1672518253326416, "learning_rate": 0.0001860222308554825, "loss": 0.7002, "step": 3590 }, { "epoch": 0.09374104903177648, "grad_norm": 0.16726090013980865, "learning_rate": 0.00018593697743633225, "loss": 0.7153, "step": 3600 }, { "epoch": 0.09400144083464253, "grad_norm": 0.16627268493175507, "learning_rate": 0.0001858514844766833, "loss": 0.7121, "step": 3610 }, { "epoch": 0.09426183263750858, "grad_norm": 0.16322413086891174, "learning_rate": 0.00018576575221483852, "loss": 0.7018, "step": 3620 }, { "epoch": 0.09452222444037461, "grad_norm": 0.1642254739999771, "learning_rate": 0.00018567978088976776, "loss": 0.706, "step": 3630 }, { "epoch": 0.09478261624324066, "grad_norm": 0.17001990973949432, "learning_rate": 0.00018559357074110727, "loss": 0.7236, "step": 3640 }, { "epoch": 0.09504300804610671, "grad_norm": 0.166624054312706, "learning_rate": 0.00018550712200915897, "loss": 0.7129, "step": 3650 }, { "epoch": 0.09530339984897275, "grad_norm": 0.17228394746780396, "learning_rate": 0.00018542043493488986, "loss": 0.7096, "step": 3660 }, { "epoch": 0.0955637916518388, "grad_norm": 0.17133943736553192, "learning_rate": 0.0001853335097599312, "loss": 0.7164, "step": 3670 }, { "epoch": 0.09582418345470485, "grad_norm": 0.16100934147834778, "learning_rate": 0.00018524634672657805, "loss": 0.711, "step": 3680 }, { "epoch": 0.09608457525757089, "grad_norm": 0.17711074650287628, "learning_rate": 0.00018515894607778837, "loss": 0.7153, "step": 3690 }, { "epoch": 0.09634496706043694, "grad_norm": 0.20850953459739685, "learning_rate": 0.00018507130805718252, "loss": 0.7309, "step": 3700 }, { "epoch": 0.09660535886330299, "grad_norm": 0.15805912017822266, "learning_rate": 0.00018498343290904246, "loss": 0.72, "step": 3710 }, { "epoch": 0.09686575066616902, "grad_norm": 0.17850877344608307, "learning_rate": 0.00018489532087831114, "loss": 0.7235, "step": 3720 }, { "epoch": 0.09712614246903507, "grad_norm": 0.16705553233623505, "learning_rate": 0.0001848069722105918, "loss": 0.7233, "step": 3730 }, { "epoch": 0.09738653427190112, "grad_norm": 0.17024511098861694, "learning_rate": 0.0001847183871521473, "loss": 0.7127, "step": 3740 }, { "epoch": 0.09764692607476716, "grad_norm": 0.16217222809791565, "learning_rate": 0.0001846295659498994, "loss": 0.6844, "step": 3750 }, { "epoch": 0.09790731787763321, "grad_norm": 0.20443694293498993, "learning_rate": 0.00018454050885142803, "loss": 0.7327, "step": 3760 }, { "epoch": 0.09816770968049926, "grad_norm": 0.16370315849781036, "learning_rate": 0.00018445121610497072, "loss": 0.7191, "step": 3770 }, { "epoch": 0.09842810148336531, "grad_norm": 0.16387638449668884, "learning_rate": 0.00018436168795942187, "loss": 0.724, "step": 3780 }, { "epoch": 0.09868849328623135, "grad_norm": 0.1664215326309204, "learning_rate": 0.00018427192466433193, "loss": 0.6955, "step": 3790 }, { "epoch": 0.0989488850890974, "grad_norm": 0.17285983264446259, "learning_rate": 0.0001841819264699069, "loss": 0.7063, "step": 3800 }, { "epoch": 0.09920927689196345, "grad_norm": 0.1637914627790451, "learning_rate": 0.00018409169362700753, "loss": 0.7211, "step": 3810 }, { "epoch": 0.09946966869482948, "grad_norm": 0.1695072054862976, "learning_rate": 0.00018400122638714855, "loss": 0.7172, "step": 3820 }, { "epoch": 0.09973006049769553, "grad_norm": 0.16097985208034515, "learning_rate": 0.00018391052500249812, "loss": 0.7073, "step": 3830 }, { "epoch": 0.09999045230056158, "grad_norm": 0.16112865507602692, "learning_rate": 0.00018381958972587707, "loss": 0.7126, "step": 3840 }, { "epoch": 0.10025084410342762, "grad_norm": 0.16681082546710968, "learning_rate": 0.0001837284208107581, "loss": 0.7331, "step": 3850 }, { "epoch": 0.10051123590629367, "grad_norm": 0.16337643563747406, "learning_rate": 0.0001836370185112652, "loss": 0.7185, "step": 3860 }, { "epoch": 0.10077162770915972, "grad_norm": 0.16530117392539978, "learning_rate": 0.0001835453830821729, "loss": 0.7058, "step": 3870 }, { "epoch": 0.10103201951202576, "grad_norm": 0.1671489030122757, "learning_rate": 0.00018345351477890554, "loss": 0.7225, "step": 3880 }, { "epoch": 0.1012924113148918, "grad_norm": 0.1631312221288681, "learning_rate": 0.00018336141385753661, "loss": 0.7189, "step": 3890 }, { "epoch": 0.10155280311775786, "grad_norm": 0.16685089468955994, "learning_rate": 0.00018326908057478795, "loss": 0.7076, "step": 3900 }, { "epoch": 0.10181319492062389, "grad_norm": 0.16232247650623322, "learning_rate": 0.00018317651518802913, "loss": 0.7168, "step": 3910 }, { "epoch": 0.10207358672348994, "grad_norm": 0.16136577725410461, "learning_rate": 0.0001830837179552766, "loss": 0.7335, "step": 3920 }, { "epoch": 0.10233397852635599, "grad_norm": 0.17820622026920319, "learning_rate": 0.00018299068913519312, "loss": 0.7071, "step": 3930 }, { "epoch": 0.10259437032922204, "grad_norm": 0.15879027545452118, "learning_rate": 0.00018289742898708707, "loss": 0.7301, "step": 3940 }, { "epoch": 0.10285476213208808, "grad_norm": 0.1569281965494156, "learning_rate": 0.00018280393777091143, "loss": 0.7018, "step": 3950 }, { "epoch": 0.10311515393495413, "grad_norm": 0.15729983150959015, "learning_rate": 0.00018271021574726336, "loss": 0.717, "step": 3960 }, { "epoch": 0.10337554573782018, "grad_norm": 0.1715734452009201, "learning_rate": 0.00018261626317738348, "loss": 0.7008, "step": 3970 }, { "epoch": 0.10363593754068622, "grad_norm": 0.16048088669776917, "learning_rate": 0.00018252208032315482, "loss": 0.7108, "step": 3980 }, { "epoch": 0.10389632934355227, "grad_norm": 0.1652391403913498, "learning_rate": 0.00018242766744710248, "loss": 0.7164, "step": 3990 }, { "epoch": 0.10415672114641832, "grad_norm": 0.17617185413837433, "learning_rate": 0.0001823330248123926, "loss": 0.7152, "step": 4000 }, { "epoch": 0.10441711294928435, "grad_norm": 0.16789917647838593, "learning_rate": 0.00018223815268283186, "loss": 0.742, "step": 4010 }, { "epoch": 0.1046775047521504, "grad_norm": 0.15830789506435394, "learning_rate": 0.00018214305132286656, "loss": 0.7127, "step": 4020 }, { "epoch": 0.10493789655501645, "grad_norm": 0.16791877150535583, "learning_rate": 0.00018204772099758194, "loss": 0.7173, "step": 4030 }, { "epoch": 0.10519828835788249, "grad_norm": 0.1708458811044693, "learning_rate": 0.00018195216197270156, "loss": 0.7152, "step": 4040 }, { "epoch": 0.10545868016074854, "grad_norm": 0.16406644880771637, "learning_rate": 0.0001818563745145863, "loss": 0.709, "step": 4050 }, { "epoch": 0.10571907196361459, "grad_norm": 0.16491225361824036, "learning_rate": 0.00018176035889023396, "loss": 0.6957, "step": 4060 }, { "epoch": 0.10597946376648062, "grad_norm": 0.18035203218460083, "learning_rate": 0.00018166411536727812, "loss": 0.7166, "step": 4070 }, { "epoch": 0.10623985556934668, "grad_norm": 0.1827452927827835, "learning_rate": 0.0001815676442139878, "loss": 0.7359, "step": 4080 }, { "epoch": 0.10650024737221273, "grad_norm": 0.17651152610778809, "learning_rate": 0.00018147094569926644, "loss": 0.7055, "step": 4090 }, { "epoch": 0.10676063917507878, "grad_norm": 0.16017574071884155, "learning_rate": 0.00018137402009265114, "loss": 0.7175, "step": 4100 }, { "epoch": 0.10702103097794481, "grad_norm": 0.16183608770370483, "learning_rate": 0.00018127686766431213, "loss": 0.7099, "step": 4110 }, { "epoch": 0.10728142278081086, "grad_norm": 0.173137828707695, "learning_rate": 0.00018117948868505182, "loss": 0.7139, "step": 4120 }, { "epoch": 0.10754181458367691, "grad_norm": 0.1736387312412262, "learning_rate": 0.00018108188342630413, "loss": 0.7219, "step": 4130 }, { "epoch": 0.10780220638654295, "grad_norm": 0.1679358333349228, "learning_rate": 0.0001809840521601337, "loss": 0.7057, "step": 4140 }, { "epoch": 0.108062598189409, "grad_norm": 0.16037985682487488, "learning_rate": 0.00018088599515923503, "loss": 0.7111, "step": 4150 }, { "epoch": 0.10832298999227505, "grad_norm": 0.15554189682006836, "learning_rate": 0.00018078771269693209, "loss": 0.6978, "step": 4160 }, { "epoch": 0.10858338179514108, "grad_norm": 0.1662927269935608, "learning_rate": 0.00018068920504717704, "loss": 0.7016, "step": 4170 }, { "epoch": 0.10884377359800713, "grad_norm": 0.168484628200531, "learning_rate": 0.0001805904724845499, "loss": 0.695, "step": 4180 }, { "epoch": 0.10910416540087318, "grad_norm": 0.18269601464271545, "learning_rate": 0.0001804915152842575, "loss": 0.7194, "step": 4190 }, { "epoch": 0.10936455720373922, "grad_norm": 0.17641094326972961, "learning_rate": 0.00018039233372213294, "loss": 0.7209, "step": 4200 }, { "epoch": 0.10962494900660527, "grad_norm": 0.18733248114585876, "learning_rate": 0.00018029292807463456, "loss": 0.7089, "step": 4210 }, { "epoch": 0.10988534080947132, "grad_norm": 0.16925212740898132, "learning_rate": 0.00018019329861884543, "loss": 0.7097, "step": 4220 }, { "epoch": 0.11014573261233736, "grad_norm": 0.15558676421642303, "learning_rate": 0.00018009344563247245, "loss": 0.6766, "step": 4230 }, { "epoch": 0.11040612441520341, "grad_norm": 0.16530027985572815, "learning_rate": 0.00017999336939384555, "loss": 0.7141, "step": 4240 }, { "epoch": 0.11066651621806946, "grad_norm": 0.16560573875904083, "learning_rate": 0.00017989307018191693, "loss": 0.6989, "step": 4250 }, { "epoch": 0.1109269080209355, "grad_norm": 0.17466457188129425, "learning_rate": 0.00017979254827626037, "loss": 0.7206, "step": 4260 }, { "epoch": 0.11118729982380154, "grad_norm": 0.18086989223957062, "learning_rate": 0.00017969180395707038, "loss": 0.714, "step": 4270 }, { "epoch": 0.1114476916266676, "grad_norm": 1.1102476119995117, "learning_rate": 0.00017959083750516137, "loss": 0.723, "step": 4280 }, { "epoch": 0.11170808342953364, "grad_norm": 0.19353629648685455, "learning_rate": 0.00017948964920196697, "loss": 0.718, "step": 4290 }, { "epoch": 0.11196847523239968, "grad_norm": 0.2126789540052414, "learning_rate": 0.00017938823932953915, "loss": 0.7171, "step": 4300 }, { "epoch": 0.11222886703526573, "grad_norm": 0.16679035127162933, "learning_rate": 0.00017928660817054758, "loss": 0.7113, "step": 4310 }, { "epoch": 0.11248925883813178, "grad_norm": 0.1647113859653473, "learning_rate": 0.00017918475600827861, "loss": 0.7184, "step": 4320 }, { "epoch": 0.11274965064099782, "grad_norm": 0.16823258996009827, "learning_rate": 0.00017908268312663475, "loss": 0.6876, "step": 4330 }, { "epoch": 0.11301004244386387, "grad_norm": 0.17199194431304932, "learning_rate": 0.0001789803898101336, "loss": 0.6954, "step": 4340 }, { "epoch": 0.11327043424672992, "grad_norm": 0.16541388630867004, "learning_rate": 0.00017887787634390733, "loss": 0.709, "step": 4350 }, { "epoch": 0.11353082604959595, "grad_norm": 0.17238549888134003, "learning_rate": 0.00017877514301370168, "loss": 0.7056, "step": 4360 }, { "epoch": 0.113791217852462, "grad_norm": 0.18374156951904297, "learning_rate": 0.00017867219010587527, "loss": 0.6951, "step": 4370 }, { "epoch": 0.11405160965532805, "grad_norm": 0.15940402448177338, "learning_rate": 0.00017856901790739874, "loss": 0.6873, "step": 4380 }, { "epoch": 0.11431200145819409, "grad_norm": 0.16984274983406067, "learning_rate": 0.00017846562670585398, "loss": 0.7065, "step": 4390 }, { "epoch": 0.11457239326106014, "grad_norm": 0.9227787852287292, "learning_rate": 0.0001783620167894334, "loss": 0.7591, "step": 4400 }, { "epoch": 0.11483278506392619, "grad_norm": 0.1881164163351059, "learning_rate": 0.00017825818844693892, "loss": 0.7133, "step": 4410 }, { "epoch": 0.11509317686679223, "grad_norm": 0.20252032577991486, "learning_rate": 0.00017815414196778147, "loss": 0.7145, "step": 4420 }, { "epoch": 0.11535356866965828, "grad_norm": 0.23380456864833832, "learning_rate": 0.0001780498776419799, "loss": 0.7232, "step": 4430 }, { "epoch": 0.11561396047252433, "grad_norm": 0.1666252464056015, "learning_rate": 0.00017794539576016028, "loss": 0.7224, "step": 4440 }, { "epoch": 0.11587435227539038, "grad_norm": 0.2421305775642395, "learning_rate": 0.0001778406966135551, "loss": 0.7031, "step": 4450 }, { "epoch": 0.11613474407825641, "grad_norm": 0.17607644200325012, "learning_rate": 0.00017773578049400256, "loss": 0.7162, "step": 4460 }, { "epoch": 0.11639513588112246, "grad_norm": 0.1905539333820343, "learning_rate": 0.00017763064769394553, "loss": 0.7087, "step": 4470 }, { "epoch": 0.11665552768398851, "grad_norm": 0.17976076900959015, "learning_rate": 0.00017752529850643088, "loss": 0.7274, "step": 4480 }, { "epoch": 0.11691591948685455, "grad_norm": 0.16786770522594452, "learning_rate": 0.00017741973322510864, "loss": 0.7003, "step": 4490 }, { "epoch": 0.1171763112897206, "grad_norm": 0.1559310108423233, "learning_rate": 0.00017731395214423126, "loss": 0.7279, "step": 4500 }, { "epoch": 0.11743670309258665, "grad_norm": 0.16374662518501282, "learning_rate": 0.00017720795555865255, "loss": 0.6924, "step": 4510 }, { "epoch": 0.11769709489545269, "grad_norm": 0.1647823303937912, "learning_rate": 0.00017710174376382712, "loss": 0.7224, "step": 4520 }, { "epoch": 0.11795748669831874, "grad_norm": 0.18070833384990692, "learning_rate": 0.0001769953170558095, "loss": 0.7223, "step": 4530 }, { "epoch": 0.11821787850118479, "grad_norm": 0.16518941521644592, "learning_rate": 0.0001768886757312532, "loss": 0.7153, "step": 4540 }, { "epoch": 0.11847827030405082, "grad_norm": 0.17081689834594727, "learning_rate": 0.0001767818200874099, "loss": 0.7197, "step": 4550 }, { "epoch": 0.11873866210691687, "grad_norm": 0.16503626108169556, "learning_rate": 0.00017667475042212874, "loss": 0.7222, "step": 4560 }, { "epoch": 0.11899905390978292, "grad_norm": 0.17310748994350433, "learning_rate": 0.00017656746703385547, "loss": 0.7057, "step": 4570 }, { "epoch": 0.11925944571264896, "grad_norm": 0.1613416224718094, "learning_rate": 0.0001764599702216315, "loss": 0.7133, "step": 4580 }, { "epoch": 0.11951983751551501, "grad_norm": 0.184098020195961, "learning_rate": 0.00017635226028509313, "loss": 0.72, "step": 4590 }, { "epoch": 0.11978022931838106, "grad_norm": 0.1819763481616974, "learning_rate": 0.00017624433752447076, "loss": 0.7154, "step": 4600 }, { "epoch": 0.12004062112124711, "grad_norm": 0.17078550159931183, "learning_rate": 0.000176136202240588, "loss": 0.7205, "step": 4610 }, { "epoch": 0.12030101292411315, "grad_norm": 0.16961923241615295, "learning_rate": 0.00017602785473486084, "loss": 0.6997, "step": 4620 }, { "epoch": 0.1205614047269792, "grad_norm": 0.18945324420928955, "learning_rate": 0.00017591929530929684, "loss": 0.7295, "step": 4630 }, { "epoch": 0.12082179652984525, "grad_norm": 0.1690925508737564, "learning_rate": 0.00017581052426649424, "loss": 0.714, "step": 4640 }, { "epoch": 0.12108218833271128, "grad_norm": 0.16739274561405182, "learning_rate": 0.00017570154190964113, "loss": 0.7018, "step": 4650 }, { "epoch": 0.12134258013557733, "grad_norm": 0.17123165726661682, "learning_rate": 0.0001755923485425147, "loss": 0.7288, "step": 4660 }, { "epoch": 0.12160297193844338, "grad_norm": 0.17006556689739227, "learning_rate": 0.0001754829444694802, "loss": 0.7185, "step": 4670 }, { "epoch": 0.12186336374130942, "grad_norm": 0.17169633507728577, "learning_rate": 0.00017537332999549023, "loss": 0.7169, "step": 4680 }, { "epoch": 0.12212375554417547, "grad_norm": 0.17851942777633667, "learning_rate": 0.00017526350542608393, "loss": 0.7138, "step": 4690 }, { "epoch": 0.12238414734704152, "grad_norm": 0.16988542675971985, "learning_rate": 0.00017515347106738597, "loss": 0.7188, "step": 4700 }, { "epoch": 0.12264453914990756, "grad_norm": 0.16778436303138733, "learning_rate": 0.00017504322722610582, "loss": 0.7034, "step": 4710 }, { "epoch": 0.1229049309527736, "grad_norm": 0.1835828721523285, "learning_rate": 0.00017493277420953688, "loss": 0.7084, "step": 4720 }, { "epoch": 0.12316532275563966, "grad_norm": 0.16590015590190887, "learning_rate": 0.0001748221123255556, "loss": 0.7023, "step": 4730 }, { "epoch": 0.12342571455850569, "grad_norm": 0.16521744430065155, "learning_rate": 0.00017471124188262062, "loss": 0.7224, "step": 4740 }, { "epoch": 0.12368610636137174, "grad_norm": 0.17280088365077972, "learning_rate": 0.00017460016318977188, "loss": 0.7087, "step": 4750 }, { "epoch": 0.12394649816423779, "grad_norm": 0.17343877255916595, "learning_rate": 0.00017448887655662984, "loss": 0.7212, "step": 4760 }, { "epoch": 0.12420688996710384, "grad_norm": 0.17116770148277283, "learning_rate": 0.0001743773822933946, "loss": 0.7289, "step": 4770 }, { "epoch": 0.12446728176996988, "grad_norm": 0.17440655827522278, "learning_rate": 0.0001742656807108449, "loss": 0.7353, "step": 4780 }, { "epoch": 0.12472767357283593, "grad_norm": 0.17304064333438873, "learning_rate": 0.0001741537721203375, "loss": 0.7322, "step": 4790 }, { "epoch": 0.12498806537570198, "grad_norm": 0.17076118290424347, "learning_rate": 0.00017404165683380604, "loss": 0.7199, "step": 4800 }, { "epoch": 0.12524845717856803, "grad_norm": 0.17371006309986115, "learning_rate": 0.0001739293351637604, "loss": 0.7035, "step": 4810 }, { "epoch": 0.12550884898143405, "grad_norm": 0.16811051964759827, "learning_rate": 0.0001738168074232857, "loss": 0.7113, "step": 4820 }, { "epoch": 0.1257692407843001, "grad_norm": 0.17408473789691925, "learning_rate": 0.00017370407392604145, "loss": 0.7151, "step": 4830 }, { "epoch": 0.12602963258716615, "grad_norm": 0.16915513575077057, "learning_rate": 0.0001735911349862607, "loss": 0.7089, "step": 4840 }, { "epoch": 0.1262900243900322, "grad_norm": 0.17281986773014069, "learning_rate": 0.00017347799091874913, "loss": 0.6935, "step": 4850 }, { "epoch": 0.12655041619289825, "grad_norm": 0.16271603107452393, "learning_rate": 0.00017336464203888426, "loss": 0.7131, "step": 4860 }, { "epoch": 0.1268108079957643, "grad_norm": 0.1661551147699356, "learning_rate": 0.00017325108866261439, "loss": 0.6899, "step": 4870 }, { "epoch": 0.12707119979863035, "grad_norm": 0.16490161418914795, "learning_rate": 0.00017313733110645792, "loss": 0.7042, "step": 4880 }, { "epoch": 0.12733159160149637, "grad_norm": 0.17232102155685425, "learning_rate": 0.00017302336968750236, "loss": 0.6965, "step": 4890 }, { "epoch": 0.12759198340436242, "grad_norm": 0.17850111424922943, "learning_rate": 0.00017290920472340347, "loss": 0.7185, "step": 4900 }, { "epoch": 0.12785237520722847, "grad_norm": 0.16296492516994476, "learning_rate": 0.00017279483653238434, "loss": 0.7179, "step": 4910 }, { "epoch": 0.12811276701009452, "grad_norm": 0.1704237312078476, "learning_rate": 0.00017268026543323454, "loss": 0.7103, "step": 4920 }, { "epoch": 0.12837315881296057, "grad_norm": 0.18265944719314575, "learning_rate": 0.00017256549174530928, "loss": 0.6988, "step": 4930 }, { "epoch": 0.12863355061582663, "grad_norm": 0.17257574200630188, "learning_rate": 0.0001724505157885284, "loss": 0.7231, "step": 4940 }, { "epoch": 0.12889394241869265, "grad_norm": 0.1703302413225174, "learning_rate": 0.0001723353378833756, "loss": 0.7192, "step": 4950 }, { "epoch": 0.1291543342215587, "grad_norm": 0.16430538892745972, "learning_rate": 0.00017221995835089743, "loss": 0.7007, "step": 4960 }, { "epoch": 0.12941472602442475, "grad_norm": 0.162199467420578, "learning_rate": 0.0001721043775127025, "loss": 0.7123, "step": 4970 }, { "epoch": 0.1296751178272908, "grad_norm": 0.1680443435907364, "learning_rate": 0.00017198859569096057, "loss": 0.6976, "step": 4980 }, { "epoch": 0.12993550963015685, "grad_norm": 0.16768567264080048, "learning_rate": 0.0001718726132084015, "loss": 0.6942, "step": 4990 }, { "epoch": 0.1301959014330229, "grad_norm": 0.16759110987186432, "learning_rate": 0.00017175643038831463, "loss": 0.6995, "step": 5000 }, { "epoch": 0.13045629323588892, "grad_norm": 0.17288459837436676, "learning_rate": 0.00017164004755454754, "loss": 0.7193, "step": 5010 }, { "epoch": 0.13071668503875497, "grad_norm": 0.18251213431358337, "learning_rate": 0.00017152346503150552, "loss": 0.7156, "step": 5020 }, { "epoch": 0.13097707684162102, "grad_norm": 0.17177079617977142, "learning_rate": 0.0001714066831441503, "loss": 0.7354, "step": 5030 }, { "epoch": 0.13123746864448707, "grad_norm": 0.18522126972675323, "learning_rate": 0.00017128970221799947, "loss": 0.717, "step": 5040 }, { "epoch": 0.13149786044735312, "grad_norm": 0.17539465427398682, "learning_rate": 0.00017117252257912532, "loss": 0.7204, "step": 5050 }, { "epoch": 0.13175825225021917, "grad_norm": 0.16526655852794647, "learning_rate": 0.00017105514455415403, "loss": 0.7099, "step": 5060 }, { "epoch": 0.13201864405308522, "grad_norm": 0.1709279865026474, "learning_rate": 0.00017093756847026483, "loss": 0.7205, "step": 5070 }, { "epoch": 0.13227903585595124, "grad_norm": 0.16695177555084229, "learning_rate": 0.00017081979465518896, "loss": 0.7092, "step": 5080 }, { "epoch": 0.1325394276588173, "grad_norm": 0.1795840859413147, "learning_rate": 0.00017070182343720882, "loss": 0.6983, "step": 5090 }, { "epoch": 0.13279981946168334, "grad_norm": 0.17300280928611755, "learning_rate": 0.0001705836551451571, "loss": 0.7018, "step": 5100 }, { "epoch": 0.1330602112645494, "grad_norm": 0.16966886818408966, "learning_rate": 0.00017046529010841574, "loss": 0.708, "step": 5110 }, { "epoch": 0.13332060306741544, "grad_norm": 0.18426869809627533, "learning_rate": 0.00017034672865691515, "loss": 0.7118, "step": 5120 }, { "epoch": 0.1335809948702815, "grad_norm": 0.1808805912733078, "learning_rate": 0.00017022797112113314, "loss": 0.7041, "step": 5130 }, { "epoch": 0.13384138667314752, "grad_norm": 0.17788094282150269, "learning_rate": 0.00017010901783209427, "loss": 0.7032, "step": 5140 }, { "epoch": 0.13410177847601357, "grad_norm": 0.17427201569080353, "learning_rate": 0.00016998986912136852, "loss": 0.7194, "step": 5150 }, { "epoch": 0.13436217027887962, "grad_norm": 0.17783771455287933, "learning_rate": 0.0001698705253210707, "loss": 0.705, "step": 5160 }, { "epoch": 0.13462256208174567, "grad_norm": 0.16358746588230133, "learning_rate": 0.0001697509867638594, "loss": 0.7118, "step": 5170 }, { "epoch": 0.13488295388461172, "grad_norm": 0.17140185832977295, "learning_rate": 0.0001696312537829361, "loss": 0.6929, "step": 5180 }, { "epoch": 0.13514334568747777, "grad_norm": 0.1690370887517929, "learning_rate": 0.0001695113267120441, "loss": 0.7044, "step": 5190 }, { "epoch": 0.13540373749034382, "grad_norm": 0.16261689364910126, "learning_rate": 0.0001693912058854679, "loss": 0.6933, "step": 5200 }, { "epoch": 0.13566412929320984, "grad_norm": 0.16243134438991547, "learning_rate": 0.00016927089163803192, "loss": 0.7009, "step": 5210 }, { "epoch": 0.1359245210960759, "grad_norm": 0.17476530373096466, "learning_rate": 0.00016915038430509978, "loss": 0.7215, "step": 5220 }, { "epoch": 0.13618491289894194, "grad_norm": 0.167569100856781, "learning_rate": 0.00016902968422257325, "loss": 0.6975, "step": 5230 }, { "epoch": 0.136445304701808, "grad_norm": 0.1719943732023239, "learning_rate": 0.00016890879172689146, "loss": 0.7324, "step": 5240 }, { "epoch": 0.13670569650467404, "grad_norm": 0.167982816696167, "learning_rate": 0.00016878770715502984, "loss": 0.6884, "step": 5250 }, { "epoch": 0.1369660883075401, "grad_norm": 0.1711047738790512, "learning_rate": 0.00016866643084449914, "loss": 0.7164, "step": 5260 }, { "epoch": 0.1372264801104061, "grad_norm": 0.17784222960472107, "learning_rate": 0.00016854496313334466, "loss": 0.6987, "step": 5270 }, { "epoch": 0.13748687191327216, "grad_norm": 0.16513650119304657, "learning_rate": 0.00016842330436014517, "loss": 0.7078, "step": 5280 }, { "epoch": 0.1377472637161382, "grad_norm": 0.17001712322235107, "learning_rate": 0.000168301454864012, "loss": 0.7053, "step": 5290 }, { "epoch": 0.13800765551900426, "grad_norm": 0.1930648535490036, "learning_rate": 0.0001681794149845881, "loss": 0.6954, "step": 5300 }, { "epoch": 0.1382680473218703, "grad_norm": 0.1823079138994217, "learning_rate": 0.00016805718506204711, "loss": 0.7325, "step": 5310 }, { "epoch": 0.13852843912473636, "grad_norm": 0.16231094300746918, "learning_rate": 0.00016793476543709238, "loss": 0.7074, "step": 5320 }, { "epoch": 0.13878883092760239, "grad_norm": 0.16010914742946625, "learning_rate": 0.00016781215645095604, "loss": 0.6946, "step": 5330 }, { "epoch": 0.13904922273046844, "grad_norm": 0.17128457129001617, "learning_rate": 0.00016768935844539804, "loss": 0.7116, "step": 5340 }, { "epoch": 0.13930961453333449, "grad_norm": 0.17700603604316711, "learning_rate": 0.00016756637176270522, "loss": 0.7276, "step": 5350 }, { "epoch": 0.13957000633620054, "grad_norm": 0.16596612334251404, "learning_rate": 0.00016744319674569033, "loss": 0.6683, "step": 5360 }, { "epoch": 0.1398303981390666, "grad_norm": 0.17127594351768494, "learning_rate": 0.00016731983373769106, "loss": 0.7053, "step": 5370 }, { "epoch": 0.14009078994193264, "grad_norm": 0.17723415791988373, "learning_rate": 0.00016719628308256915, "loss": 0.7294, "step": 5380 }, { "epoch": 0.1403511817447987, "grad_norm": 0.16194766759872437, "learning_rate": 0.00016707254512470936, "loss": 0.7164, "step": 5390 }, { "epoch": 0.1406115735476647, "grad_norm": 0.17294222116470337, "learning_rate": 0.00016694862020901852, "loss": 0.7084, "step": 5400 }, { "epoch": 0.14087196535053076, "grad_norm": 0.17684406042099, "learning_rate": 0.00016682450868092463, "loss": 0.7141, "step": 5410 }, { "epoch": 0.1411323571533968, "grad_norm": 0.16746504604816437, "learning_rate": 0.00016670021088637579, "loss": 0.7336, "step": 5420 }, { "epoch": 0.14139274895626286, "grad_norm": 0.2371237874031067, "learning_rate": 0.00016657572717183937, "loss": 0.7001, "step": 5430 }, { "epoch": 0.1416531407591289, "grad_norm": 0.16818630695343018, "learning_rate": 0.00016645105788430095, "loss": 0.7124, "step": 5440 }, { "epoch": 0.14191353256199496, "grad_norm": 0.1719949096441269, "learning_rate": 0.00016632620337126333, "loss": 0.7169, "step": 5450 }, { "epoch": 0.14217392436486098, "grad_norm": 0.1754136085510254, "learning_rate": 0.00016620116398074567, "loss": 0.6966, "step": 5460 }, { "epoch": 0.14243431616772703, "grad_norm": 0.1630551517009735, "learning_rate": 0.0001660759400612824, "loss": 0.695, "step": 5470 }, { "epoch": 0.14269470797059308, "grad_norm": 0.16431647539138794, "learning_rate": 0.00016595053196192234, "loss": 0.6983, "step": 5480 }, { "epoch": 0.14295509977345913, "grad_norm": 0.1706673949956894, "learning_rate": 0.00016582494003222772, "loss": 0.7025, "step": 5490 }, { "epoch": 0.14321549157632518, "grad_norm": 0.1757555902004242, "learning_rate": 0.00016569916462227312, "loss": 0.7159, "step": 5500 }, { "epoch": 0.14347588337919123, "grad_norm": 0.1697273999452591, "learning_rate": 0.0001655732060826446, "loss": 0.7112, "step": 5510 }, { "epoch": 0.14373627518205728, "grad_norm": 0.17706693708896637, "learning_rate": 0.00016544706476443862, "loss": 0.746, "step": 5520 }, { "epoch": 0.1439966669849233, "grad_norm": 0.17265696823596954, "learning_rate": 0.00016532074101926117, "loss": 0.7276, "step": 5530 }, { "epoch": 0.14425705878778936, "grad_norm": 0.16773734986782074, "learning_rate": 0.00016519423519922668, "loss": 0.6944, "step": 5540 }, { "epoch": 0.1445174505906554, "grad_norm": 0.17664338648319244, "learning_rate": 0.0001650675476569572, "loss": 0.6953, "step": 5550 }, { "epoch": 0.14477784239352146, "grad_norm": 0.17331312596797943, "learning_rate": 0.00016494067874558117, "loss": 0.7113, "step": 5560 }, { "epoch": 0.1450382341963875, "grad_norm": 0.1759510487318039, "learning_rate": 0.00016481362881873272, "loss": 0.71, "step": 5570 }, { "epoch": 0.14529862599925356, "grad_norm": 0.18095999956130981, "learning_rate": 0.00016468639823055044, "loss": 0.7255, "step": 5580 }, { "epoch": 0.14555901780211958, "grad_norm": 0.17434170842170715, "learning_rate": 0.0001645589873356765, "loss": 0.6989, "step": 5590 }, { "epoch": 0.14581940960498563, "grad_norm": 0.16754081845283508, "learning_rate": 0.00016443139648925572, "loss": 0.7196, "step": 5600 }, { "epoch": 0.14607980140785168, "grad_norm": 0.17215152084827423, "learning_rate": 0.00016430362604693448, "loss": 0.7104, "step": 5610 }, { "epoch": 0.14634019321071773, "grad_norm": 0.17826251685619354, "learning_rate": 0.00016417567636485973, "loss": 0.7142, "step": 5620 }, { "epoch": 0.14660058501358378, "grad_norm": 0.17078974843025208, "learning_rate": 0.00016404754779967813, "loss": 0.717, "step": 5630 }, { "epoch": 0.14686097681644983, "grad_norm": 0.18085996806621552, "learning_rate": 0.0001639192407085348, "loss": 0.711, "step": 5640 }, { "epoch": 0.14712136861931585, "grad_norm": 0.19176393747329712, "learning_rate": 0.00016379075544907268, "loss": 0.699, "step": 5650 }, { "epoch": 0.1473817604221819, "grad_norm": 0.17717042565345764, "learning_rate": 0.00016366209237943122, "loss": 0.7069, "step": 5660 }, { "epoch": 0.14764215222504795, "grad_norm": 0.17190402746200562, "learning_rate": 0.0001635332518582455, "loss": 0.6988, "step": 5670 }, { "epoch": 0.147902544027914, "grad_norm": 0.18004834651947021, "learning_rate": 0.00016340423424464516, "loss": 0.7196, "step": 5680 }, { "epoch": 0.14816293583078005, "grad_norm": 0.17457543313503265, "learning_rate": 0.00016327503989825364, "loss": 0.7109, "step": 5690 }, { "epoch": 0.1484233276336461, "grad_norm": 0.1904648393392563, "learning_rate": 0.00016314566917918693, "loss": 0.6874, "step": 5700 }, { "epoch": 0.14868371943651215, "grad_norm": 0.18096795678138733, "learning_rate": 0.00016301612244805252, "loss": 0.7154, "step": 5710 }, { "epoch": 0.14894411123937817, "grad_norm": 0.17521658539772034, "learning_rate": 0.00016288640006594874, "loss": 0.703, "step": 5720 }, { "epoch": 0.14920450304224422, "grad_norm": 0.180609330534935, "learning_rate": 0.00016275650239446328, "loss": 0.7033, "step": 5730 }, { "epoch": 0.14946489484511027, "grad_norm": 0.19524458050727844, "learning_rate": 0.00016262642979567268, "loss": 0.6925, "step": 5740 }, { "epoch": 0.14972528664797632, "grad_norm": 0.16790035367012024, "learning_rate": 0.00016249618263214087, "loss": 0.7104, "step": 5750 }, { "epoch": 0.14998567845084237, "grad_norm": 0.16895075142383575, "learning_rate": 0.00016236576126691843, "loss": 0.7126, "step": 5760 }, { "epoch": 0.15024607025370842, "grad_norm": 0.18354611098766327, "learning_rate": 0.00016223516606354163, "loss": 0.7143, "step": 5770 }, { "epoch": 0.15050646205657445, "grad_norm": 0.1766786128282547, "learning_rate": 0.00016210439738603108, "loss": 0.7164, "step": 5780 }, { "epoch": 0.1507668538594405, "grad_norm": 0.17794294655323029, "learning_rate": 0.00016197345559889107, "loss": 0.6919, "step": 5790 }, { "epoch": 0.15102724566230655, "grad_norm": 0.1671450436115265, "learning_rate": 0.00016184234106710838, "loss": 0.7048, "step": 5800 }, { "epoch": 0.1512876374651726, "grad_norm": 0.16732144355773926, "learning_rate": 0.00016171105415615134, "loss": 0.7168, "step": 5810 }, { "epoch": 0.15154802926803865, "grad_norm": 0.16594409942626953, "learning_rate": 0.00016157959523196866, "loss": 0.6942, "step": 5820 }, { "epoch": 0.1518084210709047, "grad_norm": 0.173036128282547, "learning_rate": 0.00016144796466098865, "loss": 0.7093, "step": 5830 }, { "epoch": 0.15206881287377072, "grad_norm": 0.17697438597679138, "learning_rate": 0.00016131616281011798, "loss": 0.7122, "step": 5840 }, { "epoch": 0.15232920467663677, "grad_norm": 0.1677410751581192, "learning_rate": 0.0001611841900467408, "loss": 0.7027, "step": 5850 }, { "epoch": 0.15258959647950282, "grad_norm": 0.16899007558822632, "learning_rate": 0.0001610520467387176, "loss": 0.6949, "step": 5860 }, { "epoch": 0.15284998828236887, "grad_norm": 0.17043690383434296, "learning_rate": 0.00016091973325438428, "loss": 0.7052, "step": 5870 }, { "epoch": 0.15311038008523492, "grad_norm": 0.17432864010334015, "learning_rate": 0.00016078724996255114, "loss": 0.7026, "step": 5880 }, { "epoch": 0.15337077188810097, "grad_norm": 0.17059588432312012, "learning_rate": 0.00016065459723250164, "loss": 0.7234, "step": 5890 }, { "epoch": 0.15363116369096702, "grad_norm": 0.16677695512771606, "learning_rate": 0.0001605217754339918, "loss": 0.7281, "step": 5900 }, { "epoch": 0.15389155549383304, "grad_norm": 0.16831070184707642, "learning_rate": 0.00016038878493724858, "loss": 0.6934, "step": 5910 }, { "epoch": 0.1541519472966991, "grad_norm": 0.17344878613948822, "learning_rate": 0.00016025562611296946, "loss": 0.7096, "step": 5920 }, { "epoch": 0.15441233909956514, "grad_norm": 0.16978445649147034, "learning_rate": 0.0001601222993323209, "loss": 0.6851, "step": 5930 }, { "epoch": 0.1546727309024312, "grad_norm": 0.17398668825626373, "learning_rate": 0.00015998880496693766, "loss": 0.7022, "step": 5940 }, { "epoch": 0.15493312270529724, "grad_norm": 0.16338910162448883, "learning_rate": 0.00015985514338892154, "loss": 0.7257, "step": 5950 }, { "epoch": 0.1551935145081633, "grad_norm": 0.1754075437784195, "learning_rate": 0.0001597213149708405, "loss": 0.7012, "step": 5960 }, { "epoch": 0.15545390631102932, "grad_norm": 0.16998423635959625, "learning_rate": 0.00015958732008572744, "loss": 0.7313, "step": 5970 }, { "epoch": 0.15571429811389537, "grad_norm": 0.18673604726791382, "learning_rate": 0.00015945315910707945, "loss": 0.6923, "step": 5980 }, { "epoch": 0.15597468991676142, "grad_norm": 0.17040878534317017, "learning_rate": 0.0001593188324088564, "loss": 0.7009, "step": 5990 }, { "epoch": 0.15623508171962747, "grad_norm": 0.16569504141807556, "learning_rate": 0.00015918434036548017, "loss": 0.6919, "step": 6000 }, { "epoch": 0.15649547352249352, "grad_norm": 0.16722472012043, "learning_rate": 0.00015904968335183354, "loss": 0.7028, "step": 6010 }, { "epoch": 0.15675586532535957, "grad_norm": 0.16823537647724152, "learning_rate": 0.00015891486174325912, "loss": 0.6822, "step": 6020 }, { "epoch": 0.15701625712822562, "grad_norm": 0.1750400960445404, "learning_rate": 0.00015877987591555822, "loss": 0.7146, "step": 6030 }, { "epoch": 0.15727664893109164, "grad_norm": 0.1703154593706131, "learning_rate": 0.00015864472624499008, "loss": 0.7057, "step": 6040 }, { "epoch": 0.1575370407339577, "grad_norm": 0.17317266762256622, "learning_rate": 0.00015850941310827045, "loss": 0.692, "step": 6050 }, { "epoch": 0.15779743253682374, "grad_norm": 0.18270671367645264, "learning_rate": 0.00015837393688257083, "loss": 0.7279, "step": 6060 }, { "epoch": 0.1580578243396898, "grad_norm": 0.1647195667028427, "learning_rate": 0.00015823829794551724, "loss": 0.6999, "step": 6070 }, { "epoch": 0.15831821614255584, "grad_norm": 0.17559175193309784, "learning_rate": 0.00015810249667518936, "loss": 0.6952, "step": 6080 }, { "epoch": 0.1585786079454219, "grad_norm": 0.168562114238739, "learning_rate": 0.00015796653345011928, "loss": 0.7013, "step": 6090 }, { "epoch": 0.1588389997482879, "grad_norm": 0.18079742789268494, "learning_rate": 0.00015783040864929043, "loss": 0.7115, "step": 6100 }, { "epoch": 0.15909939155115396, "grad_norm": 0.17427009344100952, "learning_rate": 0.00015769412265213684, "loss": 0.6955, "step": 6110 }, { "epoch": 0.15935978335402, "grad_norm": 0.18136094510555267, "learning_rate": 0.00015755767583854163, "loss": 0.7129, "step": 6120 }, { "epoch": 0.15962017515688606, "grad_norm": 0.17283126711845398, "learning_rate": 0.00015742106858883633, "loss": 0.6993, "step": 6130 }, { "epoch": 0.1598805669597521, "grad_norm": 0.16747362911701202, "learning_rate": 0.00015728430128379963, "loss": 0.718, "step": 6140 }, { "epoch": 0.16014095876261816, "grad_norm": 0.17267820239067078, "learning_rate": 0.0001571473743046563, "loss": 0.6946, "step": 6150 }, { "epoch": 0.16040135056548419, "grad_norm": 0.17288938164710999, "learning_rate": 0.00015701028803307625, "loss": 0.7031, "step": 6160 }, { "epoch": 0.16066174236835024, "grad_norm": 0.16363868117332458, "learning_rate": 0.0001568730428511734, "loss": 0.7038, "step": 6170 }, { "epoch": 0.16092213417121629, "grad_norm": 0.17383398115634918, "learning_rate": 0.0001567356391415046, "loss": 0.695, "step": 6180 }, { "epoch": 0.16118252597408234, "grad_norm": 0.15880119800567627, "learning_rate": 0.00015659807728706854, "loss": 0.6977, "step": 6190 }, { "epoch": 0.16144291777694839, "grad_norm": 0.1778557300567627, "learning_rate": 0.0001564603576713048, "loss": 0.7319, "step": 6200 }, { "epoch": 0.16170330957981444, "grad_norm": 0.1838517040014267, "learning_rate": 0.00015632248067809265, "loss": 0.6999, "step": 6210 }, { "epoch": 0.16196370138268049, "grad_norm": 0.1765550673007965, "learning_rate": 0.00015618444669175004, "loss": 0.689, "step": 6220 }, { "epoch": 0.1622240931855465, "grad_norm": 0.17367658019065857, "learning_rate": 0.00015604625609703257, "loss": 0.7028, "step": 6230 }, { "epoch": 0.16248448498841256, "grad_norm": 0.1650581955909729, "learning_rate": 0.0001559079092791323, "loss": 0.6926, "step": 6240 }, { "epoch": 0.1627448767912786, "grad_norm": 0.16974525153636932, "learning_rate": 0.00015576940662367675, "loss": 0.711, "step": 6250 }, { "epoch": 0.16300526859414466, "grad_norm": 0.17168691754341125, "learning_rate": 0.00015563074851672787, "loss": 0.6949, "step": 6260 }, { "epoch": 0.1632656603970107, "grad_norm": 0.17578692734241486, "learning_rate": 0.00015549193534478094, "loss": 0.6858, "step": 6270 }, { "epoch": 0.16352605219987676, "grad_norm": 0.17432482540607452, "learning_rate": 0.00015535296749476335, "loss": 0.7025, "step": 6280 }, { "epoch": 0.16378644400274278, "grad_norm": 0.17605715990066528, "learning_rate": 0.00015521384535403376, "loss": 0.7021, "step": 6290 }, { "epoch": 0.16404683580560883, "grad_norm": 0.17800843715667725, "learning_rate": 0.0001550745693103808, "loss": 0.6998, "step": 6300 }, { "epoch": 0.16430722760847488, "grad_norm": 0.17110416293144226, "learning_rate": 0.0001549351397520222, "loss": 0.7116, "step": 6310 }, { "epoch": 0.16456761941134093, "grad_norm": 0.1723811775445938, "learning_rate": 0.00015479555706760344, "loss": 0.72, "step": 6320 }, { "epoch": 0.16482801121420698, "grad_norm": 0.1752457171678543, "learning_rate": 0.00015465582164619703, "loss": 0.6866, "step": 6330 }, { "epoch": 0.16508840301707303, "grad_norm": 0.1849747896194458, "learning_rate": 0.00015451593387730105, "loss": 0.7288, "step": 6340 }, { "epoch": 0.16534879481993905, "grad_norm": 0.16762009263038635, "learning_rate": 0.00015437589415083828, "loss": 0.6919, "step": 6350 }, { "epoch": 0.1656091866228051, "grad_norm": 0.17073680460453033, "learning_rate": 0.00015423570285715506, "loss": 0.709, "step": 6360 }, { "epoch": 0.16586957842567115, "grad_norm": 0.16971205174922943, "learning_rate": 0.00015409536038702028, "loss": 0.7156, "step": 6370 }, { "epoch": 0.1661299702285372, "grad_norm": 0.17233216762542725, "learning_rate": 0.00015395486713162414, "loss": 0.7187, "step": 6380 }, { "epoch": 0.16639036203140325, "grad_norm": 0.17737407982349396, "learning_rate": 0.00015381422348257715, "loss": 0.71, "step": 6390 }, { "epoch": 0.1666507538342693, "grad_norm": 0.17228469252586365, "learning_rate": 0.00015367342983190906, "loss": 0.7224, "step": 6400 }, { "epoch": 0.16691114563713536, "grad_norm": 0.17020554840564728, "learning_rate": 0.00015353248657206764, "loss": 0.7052, "step": 6410 }, { "epoch": 0.16717153744000138, "grad_norm": 0.1692919284105301, "learning_rate": 0.00015339139409591783, "loss": 0.7035, "step": 6420 }, { "epoch": 0.16743192924286743, "grad_norm": 0.17089873552322388, "learning_rate": 0.00015325015279674036, "loss": 0.7043, "step": 6430 }, { "epoch": 0.16769232104573348, "grad_norm": 0.16704905033111572, "learning_rate": 0.00015310876306823092, "loss": 0.6924, "step": 6440 }, { "epoch": 0.16795271284859953, "grad_norm": 0.1821179836988449, "learning_rate": 0.00015296722530449877, "loss": 0.6996, "step": 6450 }, { "epoch": 0.16821310465146558, "grad_norm": 0.17202268540859222, "learning_rate": 0.000152825539900066, "loss": 0.7164, "step": 6460 }, { "epoch": 0.16847349645433163, "grad_norm": 0.16377121210098267, "learning_rate": 0.00015268370724986601, "loss": 0.699, "step": 6470 }, { "epoch": 0.16873388825719765, "grad_norm": 0.17642685770988464, "learning_rate": 0.00015254172774924277, "loss": 0.7127, "step": 6480 }, { "epoch": 0.1689942800600637, "grad_norm": 0.1717827171087265, "learning_rate": 0.0001523996017939496, "loss": 0.6927, "step": 6490 }, { "epoch": 0.16925467186292975, "grad_norm": 0.1642349660396576, "learning_rate": 0.000152257329780148, "loss": 0.704, "step": 6500 } ], "logging_steps": 10, "max_steps": 19202, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 3.04958769463296e+18, "train_batch_size": 5, "trial_name": null, "trial_params": null }