{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.9990309034133737, "eval_steps": 500, "global_step": 6963, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.004307095940562076, "grad_norm": 9.035698890686035, "learning_rate": 5.730659025787966e-07, "loss": 1.2789, "step": 10 }, { "epoch": 0.008614191881124151, "grad_norm": 6.42362642288208, "learning_rate": 1.1461318051575932e-06, "loss": 1.1797, "step": 20 }, { "epoch": 0.012921287821686228, "grad_norm": 1.6813161373138428, "learning_rate": 1.7191977077363897e-06, "loss": 0.9736, "step": 30 }, { "epoch": 0.017228383762248303, "grad_norm": 0.9371042251586914, "learning_rate": 2.2922636103151864e-06, "loss": 0.871, "step": 40 }, { "epoch": 0.02153547970281038, "grad_norm": 0.7695503234863281, "learning_rate": 2.865329512893983e-06, "loss": 0.8025, "step": 50 }, { "epoch": 0.025842575643372456, "grad_norm": 0.716374397277832, "learning_rate": 3.4383954154727795e-06, "loss": 0.7698, "step": 60 }, { "epoch": 0.030149671583934532, "grad_norm": 0.6864265203475952, "learning_rate": 4.011461318051576e-06, "loss": 0.7701, "step": 70 }, { "epoch": 0.034456767524496605, "grad_norm": 0.7581704258918762, "learning_rate": 4.584527220630373e-06, "loss": 0.7619, "step": 80 }, { "epoch": 0.03876386346505868, "grad_norm": 0.6616791486740112, "learning_rate": 5.157593123209169e-06, "loss": 0.7296, "step": 90 }, { "epoch": 0.04307095940562076, "grad_norm": 0.6397051811218262, "learning_rate": 5.730659025787966e-06, "loss": 0.7453, "step": 100 }, { "epoch": 0.047378055346182835, "grad_norm": 0.6572911143302917, "learning_rate": 6.303724928366762e-06, "loss": 0.767, "step": 110 }, { "epoch": 0.05168515128674491, "grad_norm": 0.669222354888916, "learning_rate": 6.876790830945559e-06, "loss": 0.7369, "step": 120 }, { "epoch": 0.05599224722730699, "grad_norm": 0.6517964601516724, "learning_rate": 7.449856733524356e-06, "loss": 0.7186, "step": 130 }, { "epoch": 0.060299343167869064, "grad_norm": 0.6209223866462708, "learning_rate": 8.022922636103152e-06, "loss": 0.7155, "step": 140 }, { "epoch": 0.06460643910843114, "grad_norm": 0.6591508388519287, "learning_rate": 8.595988538681949e-06, "loss": 0.7289, "step": 150 }, { "epoch": 0.06891353504899321, "grad_norm": 0.5842370390892029, "learning_rate": 9.169054441260746e-06, "loss": 0.7183, "step": 160 }, { "epoch": 0.0732206309895553, "grad_norm": 0.7117204070091248, "learning_rate": 9.742120343839543e-06, "loss": 0.7192, "step": 170 }, { "epoch": 0.07752772693011736, "grad_norm": 0.6163178086280823, "learning_rate": 1.0315186246418338e-05, "loss": 0.7193, "step": 180 }, { "epoch": 0.08183482287067945, "grad_norm": 0.5932906270027161, "learning_rate": 1.0888252148997137e-05, "loss": 0.714, "step": 190 }, { "epoch": 0.08614191881124152, "grad_norm": 0.5982919335365295, "learning_rate": 1.1461318051575932e-05, "loss": 0.7058, "step": 200 }, { "epoch": 0.0904490147518036, "grad_norm": 0.6208463907241821, "learning_rate": 1.2034383954154729e-05, "loss": 0.7189, "step": 210 }, { "epoch": 0.09475611069236567, "grad_norm": 0.5887411236763, "learning_rate": 1.2607449856733524e-05, "loss": 0.7249, "step": 220 }, { "epoch": 0.09906320663292775, "grad_norm": 0.5963988900184631, "learning_rate": 1.3180515759312323e-05, "loss": 0.7293, "step": 230 }, { "epoch": 0.10337030257348982, "grad_norm": 0.5715692043304443, "learning_rate": 1.3753581661891118e-05, "loss": 0.6845, "step": 240 }, { "epoch": 0.1076773985140519, "grad_norm": 0.639398455619812, "learning_rate": 1.4326647564469915e-05, "loss": 0.6994, "step": 250 }, { "epoch": 0.11198449445461398, "grad_norm": 0.6884477734565735, "learning_rate": 1.4899713467048712e-05, "loss": 0.7126, "step": 260 }, { "epoch": 0.11629159039517606, "grad_norm": 0.6021578907966614, "learning_rate": 1.5472779369627507e-05, "loss": 0.7215, "step": 270 }, { "epoch": 0.12059868633573813, "grad_norm": 0.6716468930244446, "learning_rate": 1.6045845272206304e-05, "loss": 0.6969, "step": 280 }, { "epoch": 0.1249057822763002, "grad_norm": 0.5783571600914001, "learning_rate": 1.66189111747851e-05, "loss": 0.7111, "step": 290 }, { "epoch": 0.12921287821686228, "grad_norm": 0.5546681880950928, "learning_rate": 1.7191977077363898e-05, "loss": 0.7, "step": 300 }, { "epoch": 0.13351997415742436, "grad_norm": 0.5409330129623413, "learning_rate": 1.7765042979942695e-05, "loss": 0.696, "step": 310 }, { "epoch": 0.13782707009798642, "grad_norm": 0.5752865672111511, "learning_rate": 1.833810888252149e-05, "loss": 0.6883, "step": 320 }, { "epoch": 0.1421341660385485, "grad_norm": 0.6340565085411072, "learning_rate": 1.891117478510029e-05, "loss": 0.6881, "step": 330 }, { "epoch": 0.1464412619791106, "grad_norm": 0.5298891067504883, "learning_rate": 1.9484240687679085e-05, "loss": 0.6935, "step": 340 }, { "epoch": 0.15074835791967267, "grad_norm": 0.5659753680229187, "learning_rate": 1.9999998871916207e-05, "loss": 0.7103, "step": 350 }, { "epoch": 0.15505545386023473, "grad_norm": 0.6017744541168213, "learning_rate": 1.999986350216883e-05, "loss": 0.6855, "step": 360 }, { "epoch": 0.1593625498007968, "grad_norm": 0.5426760911941528, "learning_rate": 1.999950251916212e-05, "loss": 0.6914, "step": 370 }, { "epoch": 0.1636696457413589, "grad_norm": 0.5532637238502502, "learning_rate": 1.999891593104044e-05, "loss": 0.6895, "step": 380 }, { "epoch": 0.16797674168192098, "grad_norm": 0.5581168532371521, "learning_rate": 1.9998103751038177e-05, "loss": 0.6897, "step": 390 }, { "epoch": 0.17228383762248303, "grad_norm": 0.5208210945129395, "learning_rate": 1.9997065997479442e-05, "loss": 0.6889, "step": 400 }, { "epoch": 0.17659093356304512, "grad_norm": 0.5863595604896545, "learning_rate": 1.9995802693777644e-05, "loss": 0.6905, "step": 410 }, { "epoch": 0.1808980295036072, "grad_norm": 0.5605342984199524, "learning_rate": 1.9994313868434988e-05, "loss": 0.6815, "step": 420 }, { "epoch": 0.18520512544416926, "grad_norm": 0.5580301880836487, "learning_rate": 1.9992599555041798e-05, "loss": 0.7067, "step": 430 }, { "epoch": 0.18951222138473134, "grad_norm": 0.558312177658081, "learning_rate": 1.999065979227579e-05, "loss": 0.7061, "step": 440 }, { "epoch": 0.19381931732529342, "grad_norm": 0.5273975133895874, "learning_rate": 1.998849462390118e-05, "loss": 0.6905, "step": 450 }, { "epoch": 0.1981264132658555, "grad_norm": 0.4772217571735382, "learning_rate": 1.9986104098767703e-05, "loss": 0.686, "step": 460 }, { "epoch": 0.20243350920641756, "grad_norm": 0.5336763858795166, "learning_rate": 1.9983488270809515e-05, "loss": 0.6861, "step": 470 }, { "epoch": 0.20674060514697964, "grad_norm": 0.4961983859539032, "learning_rate": 1.9980647199043966e-05, "loss": 0.6882, "step": 480 }, { "epoch": 0.21104770108754173, "grad_norm": 0.5408128499984741, "learning_rate": 1.9977580947570275e-05, "loss": 0.7001, "step": 490 }, { "epoch": 0.2153547970281038, "grad_norm": 0.5350680351257324, "learning_rate": 1.997428958556809e-05, "loss": 0.6931, "step": 500 }, { "epoch": 0.21966189296866587, "grad_norm": 0.5455281734466553, "learning_rate": 1.9970773187295917e-05, "loss": 0.6919, "step": 510 }, { "epoch": 0.22396898890922795, "grad_norm": 0.524664580821991, "learning_rate": 1.9967031832089438e-05, "loss": 0.6738, "step": 520 }, { "epoch": 0.22827608484979003, "grad_norm": 0.48598727583885193, "learning_rate": 1.9963065604359746e-05, "loss": 0.6678, "step": 530 }, { "epoch": 0.23258318079035212, "grad_norm": 0.5560494065284729, "learning_rate": 1.9958874593591418e-05, "loss": 0.694, "step": 540 }, { "epoch": 0.23689027673091417, "grad_norm": 0.5516777038574219, "learning_rate": 1.99544588943405e-05, "loss": 0.6715, "step": 550 }, { "epoch": 0.24119737267147626, "grad_norm": 0.5097941756248474, "learning_rate": 1.9949818606232393e-05, "loss": 0.6782, "step": 560 }, { "epoch": 0.24550446861203834, "grad_norm": 0.5353350639343262, "learning_rate": 1.9944953833959567e-05, "loss": 0.6904, "step": 570 }, { "epoch": 0.2498115645526004, "grad_norm": 0.5160298943519592, "learning_rate": 1.9939864687279237e-05, "loss": 0.6756, "step": 580 }, { "epoch": 0.2541186604931625, "grad_norm": 0.5377163887023926, "learning_rate": 1.993455128101087e-05, "loss": 0.712, "step": 590 }, { "epoch": 0.25842575643372456, "grad_norm": 0.47318100929260254, "learning_rate": 1.992901373503359e-05, "loss": 0.6648, "step": 600 }, { "epoch": 0.2627328523742866, "grad_norm": 0.4977729916572571, "learning_rate": 1.992325217428348e-05, "loss": 0.6893, "step": 610 }, { "epoch": 0.26703994831484873, "grad_norm": 0.5569038391113281, "learning_rate": 1.991726672875077e-05, "loss": 0.6876, "step": 620 }, { "epoch": 0.2713470442554108, "grad_norm": 0.544884443283081, "learning_rate": 1.9911057533476884e-05, "loss": 0.6736, "step": 630 }, { "epoch": 0.27565414019597284, "grad_norm": 0.5159808993339539, "learning_rate": 1.9904624728551417e-05, "loss": 0.674, "step": 640 }, { "epoch": 0.27996123613653495, "grad_norm": 0.48680537939071655, "learning_rate": 1.989796845910896e-05, "loss": 0.6903, "step": 650 }, { "epoch": 0.284268332077097, "grad_norm": 0.527867317199707, "learning_rate": 1.9891088875325827e-05, "loss": 0.6693, "step": 660 }, { "epoch": 0.2885754280176591, "grad_norm": 0.5441365838050842, "learning_rate": 1.988398613241666e-05, "loss": 0.6721, "step": 670 }, { "epoch": 0.2928825239582212, "grad_norm": 0.5693966150283813, "learning_rate": 1.9876660390630954e-05, "loss": 0.6684, "step": 680 }, { "epoch": 0.29718961989878323, "grad_norm": 0.5607503652572632, "learning_rate": 1.986911181524941e-05, "loss": 0.6783, "step": 690 }, { "epoch": 0.30149671583934534, "grad_norm": 0.5421719551086426, "learning_rate": 1.9861340576580225e-05, "loss": 0.6658, "step": 700 }, { "epoch": 0.3058038117799074, "grad_norm": 0.497612863779068, "learning_rate": 1.9853346849955236e-05, "loss": 0.6816, "step": 710 }, { "epoch": 0.31011090772046945, "grad_norm": 0.5503632426261902, "learning_rate": 1.984513081572598e-05, "loss": 0.6663, "step": 720 }, { "epoch": 0.31441800366103156, "grad_norm": 0.5319767594337463, "learning_rate": 1.983669265925961e-05, "loss": 0.6513, "step": 730 }, { "epoch": 0.3187250996015936, "grad_norm": 0.5350950956344604, "learning_rate": 1.9828032570934726e-05, "loss": 0.6699, "step": 740 }, { "epoch": 0.3230321955421557, "grad_norm": 0.5330127477645874, "learning_rate": 1.9819150746137067e-05, "loss": 0.6786, "step": 750 }, { "epoch": 0.3273392914827178, "grad_norm": 0.4740910232067108, "learning_rate": 1.981004738525512e-05, "loss": 0.6867, "step": 760 }, { "epoch": 0.33164638742327984, "grad_norm": 0.5131900906562805, "learning_rate": 1.980072269367557e-05, "loss": 0.6618, "step": 770 }, { "epoch": 0.33595348336384195, "grad_norm": 0.4712623059749603, "learning_rate": 1.97911768817787e-05, "loss": 0.6863, "step": 780 }, { "epoch": 0.340260579304404, "grad_norm": 0.5240254998207092, "learning_rate": 1.9781410164933626e-05, "loss": 0.6941, "step": 790 }, { "epoch": 0.34456767524496607, "grad_norm": 0.5192612409591675, "learning_rate": 1.9771422763493434e-05, "loss": 0.6726, "step": 800 }, { "epoch": 0.3488747711855282, "grad_norm": 0.4864448010921478, "learning_rate": 1.9761214902790217e-05, "loss": 0.6541, "step": 810 }, { "epoch": 0.35318186712609023, "grad_norm": 0.5248873829841614, "learning_rate": 1.9750786813129995e-05, "loss": 0.6713, "step": 820 }, { "epoch": 0.3574889630666523, "grad_norm": 0.5010212659835815, "learning_rate": 1.9740138729787505e-05, "loss": 0.6793, "step": 830 }, { "epoch": 0.3617960590072144, "grad_norm": 0.4966225326061249, "learning_rate": 1.9729270893000913e-05, "loss": 0.6692, "step": 840 }, { "epoch": 0.36610315494777645, "grad_norm": 0.48576685786247253, "learning_rate": 1.9718183547966366e-05, "loss": 0.6812, "step": 850 }, { "epoch": 0.3704102508883385, "grad_norm": 0.5232109427452087, "learning_rate": 1.9706876944832486e-05, "loss": 0.6567, "step": 860 }, { "epoch": 0.3747173468289006, "grad_norm": 0.4847777485847473, "learning_rate": 1.9695351338694713e-05, "loss": 0.6638, "step": 870 }, { "epoch": 0.3790244427694627, "grad_norm": 0.49412795901298523, "learning_rate": 1.9683606989589553e-05, "loss": 0.6731, "step": 880 }, { "epoch": 0.3833315387100248, "grad_norm": 0.5143546462059021, "learning_rate": 1.9671644162488716e-05, "loss": 0.6779, "step": 890 }, { "epoch": 0.38763863465058684, "grad_norm": 0.5516107082366943, "learning_rate": 1.965946312729312e-05, "loss": 0.6798, "step": 900 }, { "epoch": 0.3919457305911489, "grad_norm": 0.5140990018844604, "learning_rate": 1.9647064158826825e-05, "loss": 0.6473, "step": 910 }, { "epoch": 0.396252826531711, "grad_norm": 0.4911974370479584, "learning_rate": 1.9634447536830815e-05, "loss": 0.6565, "step": 920 }, { "epoch": 0.40055992247227307, "grad_norm": 0.4995877742767334, "learning_rate": 1.9621613545956703e-05, "loss": 0.6514, "step": 930 }, { "epoch": 0.4048670184128351, "grad_norm": 0.48752328753471375, "learning_rate": 1.9608562475760287e-05, "loss": 0.6751, "step": 940 }, { "epoch": 0.40917411435339723, "grad_norm": 0.4956004321575165, "learning_rate": 1.9595294620695036e-05, "loss": 0.6492, "step": 950 }, { "epoch": 0.4134812102939593, "grad_norm": 0.48215603828430176, "learning_rate": 1.958181028010544e-05, "loss": 0.6741, "step": 960 }, { "epoch": 0.4177883062345214, "grad_norm": 0.48835939168930054, "learning_rate": 1.9568109758220253e-05, "loss": 0.6638, "step": 970 }, { "epoch": 0.42209540217508346, "grad_norm": 0.47754788398742676, "learning_rate": 1.9554193364145635e-05, "loss": 0.6657, "step": 980 }, { "epoch": 0.4264024981156455, "grad_norm": 0.5080917477607727, "learning_rate": 1.9540061411858172e-05, "loss": 0.6675, "step": 990 }, { "epoch": 0.4307095940562076, "grad_norm": 0.4634297788143158, "learning_rate": 1.9525714220197802e-05, "loss": 0.6693, "step": 1000 }, { "epoch": 0.4350166899967697, "grad_norm": 0.4760366678237915, "learning_rate": 1.951115211286061e-05, "loss": 0.6721, "step": 1010 }, { "epoch": 0.43932378593733173, "grad_norm": 0.5227916836738586, "learning_rate": 1.9496375418391525e-05, "loss": 0.6691, "step": 1020 }, { "epoch": 0.44363088187789385, "grad_norm": 0.5157990455627441, "learning_rate": 1.948138447017692e-05, "loss": 0.6774, "step": 1030 }, { "epoch": 0.4479379778184559, "grad_norm": 0.49596408009529114, "learning_rate": 1.9466179606437087e-05, "loss": 0.6313, "step": 1040 }, { "epoch": 0.45224507375901796, "grad_norm": 0.47041237354278564, "learning_rate": 1.945076117021859e-05, "loss": 0.6724, "step": 1050 }, { "epoch": 0.45655216969958007, "grad_norm": 0.5206364989280701, "learning_rate": 1.9435129509386538e-05, "loss": 0.6843, "step": 1060 }, { "epoch": 0.4608592656401421, "grad_norm": 0.5067657828330994, "learning_rate": 1.9419284976616745e-05, "loss": 0.6649, "step": 1070 }, { "epoch": 0.46516636158070424, "grad_norm": 1.3445152044296265, "learning_rate": 1.9403227929387756e-05, "loss": 0.6548, "step": 1080 }, { "epoch": 0.4694734575212663, "grad_norm": 0.5465224385261536, "learning_rate": 1.93869587299728e-05, "loss": 0.6427, "step": 1090 }, { "epoch": 0.47378055346182835, "grad_norm": 0.49137911200523376, "learning_rate": 1.9370477745431587e-05, "loss": 0.6519, "step": 1100 }, { "epoch": 0.47808764940239046, "grad_norm": 0.48190736770629883, "learning_rate": 1.935378534760206e-05, "loss": 0.6615, "step": 1110 }, { "epoch": 0.4823947453429525, "grad_norm": 0.4869353771209717, "learning_rate": 1.9336881913091992e-05, "loss": 0.65, "step": 1120 }, { "epoch": 0.48670184128351457, "grad_norm": 0.4473590552806854, "learning_rate": 1.931976782327048e-05, "loss": 0.6821, "step": 1130 }, { "epoch": 0.4910089372240767, "grad_norm": 0.4703207314014435, "learning_rate": 1.9302443464259352e-05, "loss": 0.657, "step": 1140 }, { "epoch": 0.49531603316463874, "grad_norm": 0.48172295093536377, "learning_rate": 1.9284909226924457e-05, "loss": 0.6581, "step": 1150 }, { "epoch": 0.4996231291052008, "grad_norm": 0.4986841082572937, "learning_rate": 1.9267165506866835e-05, "loss": 0.664, "step": 1160 }, { "epoch": 0.5039302250457629, "grad_norm": 0.4936910569667816, "learning_rate": 1.9249212704413803e-05, "loss": 0.6409, "step": 1170 }, { "epoch": 0.508237320986325, "grad_norm": 0.48618724942207336, "learning_rate": 1.9231051224609918e-05, "loss": 0.6566, "step": 1180 }, { "epoch": 0.512544416926887, "grad_norm": 0.5300356149673462, "learning_rate": 1.921268147720784e-05, "loss": 0.6533, "step": 1190 }, { "epoch": 0.5168515128674491, "grad_norm": 0.4799743890762329, "learning_rate": 1.919410387665908e-05, "loss": 0.6677, "step": 1200 }, { "epoch": 0.5211586088080112, "grad_norm": 0.5317394137382507, "learning_rate": 1.9175318842104667e-05, "loss": 0.6464, "step": 1210 }, { "epoch": 0.5254657047485732, "grad_norm": 0.49199768900871277, "learning_rate": 1.9156326797365665e-05, "loss": 0.6655, "step": 1220 }, { "epoch": 0.5297728006891353, "grad_norm": 0.4916874170303345, "learning_rate": 1.913712817093364e-05, "loss": 0.6372, "step": 1230 }, { "epoch": 0.5340798966296975, "grad_norm": 0.48562970757484436, "learning_rate": 1.9117723395960972e-05, "loss": 0.6639, "step": 1240 }, { "epoch": 0.5383869925702595, "grad_norm": 0.5152992010116577, "learning_rate": 1.909811291025109e-05, "loss": 0.6609, "step": 1250 }, { "epoch": 0.5426940885108216, "grad_norm": 0.48352181911468506, "learning_rate": 1.907829715624859e-05, "loss": 0.6726, "step": 1260 }, { "epoch": 0.5470011844513837, "grad_norm": 0.5064017176628113, "learning_rate": 1.905827658102926e-05, "loss": 0.6698, "step": 1270 }, { "epoch": 0.5513082803919457, "grad_norm": 0.46494290232658386, "learning_rate": 1.9038051636289997e-05, "loss": 0.68, "step": 1280 }, { "epoch": 0.5556153763325078, "grad_norm": 0.4788792133331299, "learning_rate": 1.9017622778338585e-05, "loss": 0.6501, "step": 1290 }, { "epoch": 0.5599224722730699, "grad_norm": 0.4712987542152405, "learning_rate": 1.8996990468083448e-05, "loss": 0.6488, "step": 1300 }, { "epoch": 0.5642295682136319, "grad_norm": 0.4997137784957886, "learning_rate": 1.8976155171023216e-05, "loss": 0.6518, "step": 1310 }, { "epoch": 0.568536664154194, "grad_norm": 0.5003030896186829, "learning_rate": 1.895511735723623e-05, "loss": 0.6317, "step": 1320 }, { "epoch": 0.5728437600947561, "grad_norm": 0.4551664888858795, "learning_rate": 1.8933877501369944e-05, "loss": 0.6634, "step": 1330 }, { "epoch": 0.5771508560353182, "grad_norm": 0.532534122467041, "learning_rate": 1.891243608263021e-05, "loss": 0.6656, "step": 1340 }, { "epoch": 0.5814579519758802, "grad_norm": 0.47166600823402405, "learning_rate": 1.889079358477047e-05, "loss": 0.657, "step": 1350 }, { "epoch": 0.5857650479164423, "grad_norm": 0.45552805066108704, "learning_rate": 1.8868950496080832e-05, "loss": 0.6652, "step": 1360 }, { "epoch": 0.5900721438570045, "grad_norm": 0.5267536044120789, "learning_rate": 1.884690730937707e-05, "loss": 0.6463, "step": 1370 }, { "epoch": 0.5943792397975665, "grad_norm": 0.49093228578567505, "learning_rate": 1.882466452198949e-05, "loss": 0.6604, "step": 1380 }, { "epoch": 0.5986863357381286, "grad_norm": 0.5105960369110107, "learning_rate": 1.880222263575172e-05, "loss": 0.6457, "step": 1390 }, { "epoch": 0.6029934316786907, "grad_norm": 0.47326135635375977, "learning_rate": 1.8779582156989384e-05, "loss": 0.6464, "step": 1400 }, { "epoch": 0.6073005276192527, "grad_norm": 0.4910115599632263, "learning_rate": 1.875674359650867e-05, "loss": 0.6547, "step": 1410 }, { "epoch": 0.6116076235598148, "grad_norm": 0.48352956771850586, "learning_rate": 1.873370746958482e-05, "loss": 0.654, "step": 1420 }, { "epoch": 0.6159147195003769, "grad_norm": 0.4722056984901428, "learning_rate": 1.871047429595049e-05, "loss": 0.6372, "step": 1430 }, { "epoch": 0.6202218154409389, "grad_norm": 0.4340212345123291, "learning_rate": 1.868704459978405e-05, "loss": 0.6507, "step": 1440 }, { "epoch": 0.624528911381501, "grad_norm": 0.48497867584228516, "learning_rate": 1.8663418909697723e-05, "loss": 0.6349, "step": 1450 }, { "epoch": 0.6288360073220631, "grad_norm": 0.4707370102405548, "learning_rate": 1.863959775872567e-05, "loss": 0.6445, "step": 1460 }, { "epoch": 0.6331431032626251, "grad_norm": 0.5151925683021545, "learning_rate": 1.861558168431199e-05, "loss": 0.6493, "step": 1470 }, { "epoch": 0.6374501992031872, "grad_norm": 0.47226110100746155, "learning_rate": 1.8591371228298554e-05, "loss": 0.6211, "step": 1480 }, { "epoch": 0.6417572951437494, "grad_norm": 0.48166829347610474, "learning_rate": 1.856696693691281e-05, "loss": 0.6476, "step": 1490 }, { "epoch": 0.6460643910843114, "grad_norm": 0.5039719343185425, "learning_rate": 1.8542369360755448e-05, "loss": 0.636, "step": 1500 }, { "epoch": 0.6503714870248735, "grad_norm": 0.45818519592285156, "learning_rate": 1.8517579054787974e-05, "loss": 0.658, "step": 1510 }, { "epoch": 0.6546785829654356, "grad_norm": 0.4803057014942169, "learning_rate": 1.8492596578320194e-05, "loss": 0.6468, "step": 1520 }, { "epoch": 0.6589856789059977, "grad_norm": 0.480227530002594, "learning_rate": 1.8467422494997593e-05, "loss": 0.641, "step": 1530 }, { "epoch": 0.6632927748465597, "grad_norm": 0.49187588691711426, "learning_rate": 1.844205737278863e-05, "loss": 0.6572, "step": 1540 }, { "epoch": 0.6675998707871218, "grad_norm": 0.49701517820358276, "learning_rate": 1.84165017839719e-05, "loss": 0.6567, "step": 1550 }, { "epoch": 0.6719069667276839, "grad_norm": 0.48368483781814575, "learning_rate": 1.8390756305123246e-05, "loss": 0.669, "step": 1560 }, { "epoch": 0.6762140626682459, "grad_norm": 0.5007254481315613, "learning_rate": 1.836482151710273e-05, "loss": 0.6448, "step": 1570 }, { "epoch": 0.680521158608808, "grad_norm": 0.44526585936546326, "learning_rate": 1.8338698005041556e-05, "loss": 0.6386, "step": 1580 }, { "epoch": 0.6848282545493701, "grad_norm": 0.4812663197517395, "learning_rate": 1.8312386358328828e-05, "loss": 0.6447, "step": 1590 }, { "epoch": 0.6891353504899321, "grad_norm": 0.4910503029823303, "learning_rate": 1.828588717059829e-05, "loss": 0.6449, "step": 1600 }, { "epoch": 0.6934424464304942, "grad_norm": 0.47431930899620056, "learning_rate": 1.8259201039714914e-05, "loss": 0.6372, "step": 1610 }, { "epoch": 0.6977495423710564, "grad_norm": 0.5024338364601135, "learning_rate": 1.8232328567761416e-05, "loss": 0.6433, "step": 1620 }, { "epoch": 0.7020566383116184, "grad_norm": 0.47510799765586853, "learning_rate": 1.820527036102467e-05, "loss": 0.6601, "step": 1630 }, { "epoch": 0.7063637342521805, "grad_norm": 0.47990313172340393, "learning_rate": 1.8178027029982027e-05, "loss": 0.6463, "step": 1640 }, { "epoch": 0.7106708301927426, "grad_norm": 0.5117030739784241, "learning_rate": 1.8150599189287553e-05, "loss": 0.6455, "step": 1650 }, { "epoch": 0.7149779261333046, "grad_norm": 0.4917861819267273, "learning_rate": 1.8122987457758147e-05, "loss": 0.6688, "step": 1660 }, { "epoch": 0.7192850220738667, "grad_norm": 0.49872297048568726, "learning_rate": 1.8095192458359588e-05, "loss": 0.6513, "step": 1670 }, { "epoch": 0.7235921180144288, "grad_norm": 0.47510796785354614, "learning_rate": 1.806721481819247e-05, "loss": 0.649, "step": 1680 }, { "epoch": 0.7278992139549908, "grad_norm": 0.4924173057079315, "learning_rate": 1.8039055168478074e-05, "loss": 0.6177, "step": 1690 }, { "epoch": 0.7322063098955529, "grad_norm": 0.4918348789215088, "learning_rate": 1.8010714144544104e-05, "loss": 0.6543, "step": 1700 }, { "epoch": 0.736513405836115, "grad_norm": 0.45298415422439575, "learning_rate": 1.7982192385810372e-05, "loss": 0.6367, "step": 1710 }, { "epoch": 0.740820501776677, "grad_norm": 0.46879851818084717, "learning_rate": 1.795349053577435e-05, "loss": 0.6414, "step": 1720 }, { "epoch": 0.7451275977172391, "grad_norm": 0.4573706388473511, "learning_rate": 1.7924609241996672e-05, "loss": 0.628, "step": 1730 }, { "epoch": 0.7494346936578012, "grad_norm": 0.46929094195365906, "learning_rate": 1.7895549156086514e-05, "loss": 0.6478, "step": 1740 }, { "epoch": 0.7537417895983634, "grad_norm": 0.5428628325462341, "learning_rate": 1.78663109336869e-05, "loss": 0.6405, "step": 1750 }, { "epoch": 0.7580488855389254, "grad_norm": 0.47853079438209534, "learning_rate": 1.78368952344599e-05, "loss": 0.6442, "step": 1760 }, { "epoch": 0.7623559814794875, "grad_norm": 0.46747061610221863, "learning_rate": 1.7807302722071742e-05, "loss": 0.6369, "step": 1770 }, { "epoch": 0.7666630774200496, "grad_norm": 0.5107671022415161, "learning_rate": 1.7777534064177864e-05, "loss": 0.6322, "step": 1780 }, { "epoch": 0.7709701733606116, "grad_norm": 0.5013517141342163, "learning_rate": 1.7747589932407826e-05, "loss": 0.6384, "step": 1790 }, { "epoch": 0.7752772693011737, "grad_norm": 0.5039073824882507, "learning_rate": 1.7717471002350162e-05, "loss": 0.6504, "step": 1800 }, { "epoch": 0.7795843652417358, "grad_norm": 0.4767347276210785, "learning_rate": 1.7687177953537148e-05, "loss": 0.645, "step": 1810 }, { "epoch": 0.7838914611822978, "grad_norm": 0.4766087532043457, "learning_rate": 1.7656711469429464e-05, "loss": 0.6249, "step": 1820 }, { "epoch": 0.7881985571228599, "grad_norm": 0.5031486749649048, "learning_rate": 1.7626072237400764e-05, "loss": 0.6263, "step": 1830 }, { "epoch": 0.792505653063422, "grad_norm": 0.444658488035202, "learning_rate": 1.759526094872219e-05, "loss": 0.6561, "step": 1840 }, { "epoch": 0.796812749003984, "grad_norm": 0.5070600509643555, "learning_rate": 1.7564278298546758e-05, "loss": 0.6477, "step": 1850 }, { "epoch": 0.8011198449445461, "grad_norm": 0.45487794280052185, "learning_rate": 1.753312498589367e-05, "loss": 0.6257, "step": 1860 }, { "epoch": 0.8054269408851082, "grad_norm": 0.4745471477508545, "learning_rate": 1.7501801713632568e-05, "loss": 0.6586, "step": 1870 }, { "epoch": 0.8097340368256702, "grad_norm": 0.4743909537792206, "learning_rate": 1.7470309188467645e-05, "loss": 0.6255, "step": 1880 }, { "epoch": 0.8140411327662324, "grad_norm": 0.5165956020355225, "learning_rate": 1.7438648120921736e-05, "loss": 0.6592, "step": 1890 }, { "epoch": 0.8183482287067945, "grad_norm": 0.455861359834671, "learning_rate": 1.740681922532025e-05, "loss": 0.6467, "step": 1900 }, { "epoch": 0.8226553246473565, "grad_norm": 0.468013733625412, "learning_rate": 1.7374823219775073e-05, "loss": 0.6382, "step": 1910 }, { "epoch": 0.8269624205879186, "grad_norm": 0.46119919419288635, "learning_rate": 1.7342660826168374e-05, "loss": 0.6437, "step": 1920 }, { "epoch": 0.8312695165284807, "grad_norm": 0.4399983286857605, "learning_rate": 1.73103327701363e-05, "loss": 0.6379, "step": 1930 }, { "epoch": 0.8355766124690428, "grad_norm": 0.46829739212989807, "learning_rate": 1.7277839781052617e-05, "loss": 0.6402, "step": 1940 }, { "epoch": 0.8398837084096048, "grad_norm": 0.5193459987640381, "learning_rate": 1.7245182592012248e-05, "loss": 0.6348, "step": 1950 }, { "epoch": 0.8441908043501669, "grad_norm": 0.5310715436935425, "learning_rate": 1.7212361939814735e-05, "loss": 0.6351, "step": 1960 }, { "epoch": 0.848497900290729, "grad_norm": 0.4883059561252594, "learning_rate": 1.7179378564947615e-05, "loss": 0.6401, "step": 1970 }, { "epoch": 0.852804996231291, "grad_norm": 0.5028474926948547, "learning_rate": 1.7146233211569723e-05, "loss": 0.6559, "step": 1980 }, { "epoch": 0.8571120921718531, "grad_norm": 0.48668941855430603, "learning_rate": 1.7112926627494385e-05, "loss": 0.6572, "step": 1990 }, { "epoch": 0.8614191881124152, "grad_norm": 0.4668605327606201, "learning_rate": 1.7079459564172555e-05, "loss": 0.6321, "step": 2000 }, { "epoch": 0.8657262840529772, "grad_norm": 0.4556910991668701, "learning_rate": 1.7045832776675863e-05, "loss": 0.6268, "step": 2010 }, { "epoch": 0.8700333799935394, "grad_norm": 0.45260846614837646, "learning_rate": 1.701204702367958e-05, "loss": 0.6271, "step": 2020 }, { "epoch": 0.8743404759341015, "grad_norm": 0.4828309714794159, "learning_rate": 1.6978103067445494e-05, "loss": 0.6351, "step": 2030 }, { "epoch": 0.8786475718746635, "grad_norm": 0.4691152274608612, "learning_rate": 1.6944001673804723e-05, "loss": 0.6512, "step": 2040 }, { "epoch": 0.8829546678152256, "grad_norm": 0.4812765419483185, "learning_rate": 1.6909743612140417e-05, "loss": 0.6335, "step": 2050 }, { "epoch": 0.8872617637557877, "grad_norm": 0.4415755867958069, "learning_rate": 1.687532965537043e-05, "loss": 0.6541, "step": 2060 }, { "epoch": 0.8915688596963497, "grad_norm": 0.4993227422237396, "learning_rate": 1.6840760579929846e-05, "loss": 0.6318, "step": 2070 }, { "epoch": 0.8958759556369118, "grad_norm": 0.4628779888153076, "learning_rate": 1.6806037165753498e-05, "loss": 0.6369, "step": 2080 }, { "epoch": 0.9001830515774739, "grad_norm": 0.5235878229141235, "learning_rate": 1.677116019625834e-05, "loss": 0.6415, "step": 2090 }, { "epoch": 0.9044901475180359, "grad_norm": 0.4750138819217682, "learning_rate": 1.6736130458325793e-05, "loss": 0.6101, "step": 2100 }, { "epoch": 0.908797243458598, "grad_norm": 0.5292583107948303, "learning_rate": 1.6700948742283977e-05, "loss": 0.6248, "step": 2110 }, { "epoch": 0.9131043393991601, "grad_norm": 0.45959070324897766, "learning_rate": 1.6665615841889885e-05, "loss": 0.6339, "step": 2120 }, { "epoch": 0.9174114353397222, "grad_norm": 0.48287901282310486, "learning_rate": 1.6630132554311486e-05, "loss": 0.6161, "step": 2130 }, { "epoch": 0.9217185312802842, "grad_norm": 0.4725618064403534, "learning_rate": 1.6594499680109722e-05, "loss": 0.627, "step": 2140 }, { "epoch": 0.9260256272208464, "grad_norm": 0.4820912778377533, "learning_rate": 1.6558718023220457e-05, "loss": 0.6399, "step": 2150 }, { "epoch": 0.9303327231614085, "grad_norm": 0.48815685510635376, "learning_rate": 1.6522788390936328e-05, "loss": 0.6437, "step": 2160 }, { "epoch": 0.9346398191019705, "grad_norm": 0.4747340679168701, "learning_rate": 1.648671159388855e-05, "loss": 0.6455, "step": 2170 }, { "epoch": 0.9389469150425326, "grad_norm": 0.4894673526287079, "learning_rate": 1.6450488446028612e-05, "loss": 0.6545, "step": 2180 }, { "epoch": 0.9432540109830947, "grad_norm": 0.4756160080432892, "learning_rate": 1.641411976460991e-05, "loss": 0.6498, "step": 2190 }, { "epoch": 0.9475611069236567, "grad_norm": 0.45228078961372375, "learning_rate": 1.637760637016932e-05, "loss": 0.6438, "step": 2200 }, { "epoch": 0.9518682028642188, "grad_norm": 0.49898287653923035, "learning_rate": 1.6340949086508676e-05, "loss": 0.6518, "step": 2210 }, { "epoch": 0.9561752988047809, "grad_norm": 0.4354493021965027, "learning_rate": 1.6304148740676204e-05, "loss": 0.6125, "step": 2220 }, { "epoch": 0.9604823947453429, "grad_norm": 0.45118704438209534, "learning_rate": 1.6267206162947823e-05, "loss": 0.6146, "step": 2230 }, { "epoch": 0.964789490685905, "grad_norm": 0.4822487533092499, "learning_rate": 1.6230122186808443e-05, "loss": 0.6425, "step": 2240 }, { "epoch": 0.9690965866264671, "grad_norm": 0.490903377532959, "learning_rate": 1.619289764893317e-05, "loss": 0.6353, "step": 2250 }, { "epoch": 0.9734036825670291, "grad_norm": 0.4738866686820984, "learning_rate": 1.615553338916839e-05, "loss": 0.6315, "step": 2260 }, { "epoch": 0.9777107785075912, "grad_norm": 0.46285027265548706, "learning_rate": 1.6118030250512863e-05, "loss": 0.6501, "step": 2270 }, { "epoch": 0.9820178744481534, "grad_norm": 0.46414172649383545, "learning_rate": 1.6080389079098657e-05, "loss": 0.6501, "step": 2280 }, { "epoch": 0.9863249703887154, "grad_norm": 0.5042113661766052, "learning_rate": 1.604261072417211e-05, "loss": 0.6319, "step": 2290 }, { "epoch": 0.9906320663292775, "grad_norm": 0.43653419613838196, "learning_rate": 1.600469603807464e-05, "loss": 0.6461, "step": 2300 }, { "epoch": 0.9949391622698396, "grad_norm": 0.4572006165981293, "learning_rate": 1.5966645876223505e-05, "loss": 0.6477, "step": 2310 }, { "epoch": 0.9992462582104016, "grad_norm": 0.43867436051368713, "learning_rate": 1.5928461097092532e-05, "loss": 0.6288, "step": 2320 }, { "epoch": 1.0035533541509638, "grad_norm": 0.5620077848434448, "learning_rate": 1.589014256219273e-05, "loss": 0.5378, "step": 2330 }, { "epoch": 1.0078604500915258, "grad_norm": 0.4836018681526184, "learning_rate": 1.5851691136052842e-05, "loss": 0.5421, "step": 2340 }, { "epoch": 1.0121675460320878, "grad_norm": 0.49632197618484497, "learning_rate": 1.581310768619988e-05, "loss": 0.5237, "step": 2350 }, { "epoch": 1.01647464197265, "grad_norm": 0.49445948004722595, "learning_rate": 1.5774393083139513e-05, "loss": 0.5313, "step": 2360 }, { "epoch": 1.020781737913212, "grad_norm": 0.5299666523933411, "learning_rate": 1.5735548200336435e-05, "loss": 0.5326, "step": 2370 }, { "epoch": 1.025088833853774, "grad_norm": 0.5012844204902649, "learning_rate": 1.569657391419468e-05, "loss": 0.5401, "step": 2380 }, { "epoch": 1.0293959297943363, "grad_norm": 0.4741289019584656, "learning_rate": 1.565747110403781e-05, "loss": 0.5052, "step": 2390 }, { "epoch": 1.0337030257348983, "grad_norm": 0.4950823485851288, "learning_rate": 1.5618240652089123e-05, "loss": 0.5294, "step": 2400 }, { "epoch": 1.0380101216754603, "grad_norm": 0.4934958517551422, "learning_rate": 1.557888344345171e-05, "loss": 0.5278, "step": 2410 }, { "epoch": 1.0423172176160225, "grad_norm": 0.467101514339447, "learning_rate": 1.5539400366088503e-05, "loss": 0.504, "step": 2420 }, { "epoch": 1.0466243135565845, "grad_norm": 0.5479716062545776, "learning_rate": 1.5499792310802238e-05, "loss": 0.5256, "step": 2430 }, { "epoch": 1.0509314094971465, "grad_norm": 0.4706737697124481, "learning_rate": 1.5460060171215362e-05, "loss": 0.5251, "step": 2440 }, { "epoch": 1.0552385054377087, "grad_norm": 0.5142565965652466, "learning_rate": 1.5420204843749857e-05, "loss": 0.5333, "step": 2450 }, { "epoch": 1.0595456013782707, "grad_norm": 0.5430694222450256, "learning_rate": 1.5380227227607032e-05, "loss": 0.5391, "step": 2460 }, { "epoch": 1.0638526973188327, "grad_norm": 0.4780258536338806, "learning_rate": 1.5340128224747225e-05, "loss": 0.5338, "step": 2470 }, { "epoch": 1.068159793259395, "grad_norm": 0.47647717595100403, "learning_rate": 1.5299908739869464e-05, "loss": 0.5178, "step": 2480 }, { "epoch": 1.072466889199957, "grad_norm": 0.5330241918563843, "learning_rate": 1.525956968039103e-05, "loss": 0.5027, "step": 2490 }, { "epoch": 1.076773985140519, "grad_norm": 0.4681854546070099, "learning_rate": 1.5219111956427027e-05, "loss": 0.5315, "step": 2500 }, { "epoch": 1.0810810810810811, "grad_norm": 0.5060921311378479, "learning_rate": 1.5178536480769803e-05, "loss": 0.5103, "step": 2510 }, { "epoch": 1.0853881770216431, "grad_norm": 0.497199147939682, "learning_rate": 1.5137844168868391e-05, "loss": 0.5302, "step": 2520 }, { "epoch": 1.0896952729622051, "grad_norm": 0.4658927321434021, "learning_rate": 1.5097035938807834e-05, "loss": 0.5196, "step": 2530 }, { "epoch": 1.0940023689027674, "grad_norm": 0.5109249353408813, "learning_rate": 1.5056112711288475e-05, "loss": 0.5099, "step": 2540 }, { "epoch": 1.0983094648433294, "grad_norm": 0.5212246775627136, "learning_rate": 1.5015075409605189e-05, "loss": 0.4911, "step": 2550 }, { "epoch": 1.1026165607838914, "grad_norm": 0.47850698232650757, "learning_rate": 1.497392495962656e-05, "loss": 0.5225, "step": 2560 }, { "epoch": 1.1069236567244536, "grad_norm": 0.4982755184173584, "learning_rate": 1.4932662289773969e-05, "loss": 0.5278, "step": 2570 }, { "epoch": 1.1112307526650156, "grad_norm": 0.49975791573524475, "learning_rate": 1.4891288331000668e-05, "loss": 0.5261, "step": 2580 }, { "epoch": 1.1155378486055776, "grad_norm": 0.5002388954162598, "learning_rate": 1.484980401677077e-05, "loss": 0.5313, "step": 2590 }, { "epoch": 1.1198449445461398, "grad_norm": 0.4950617253780365, "learning_rate": 1.4808210283038183e-05, "loss": 0.5286, "step": 2600 }, { "epoch": 1.1241520404867018, "grad_norm": 0.49831753969192505, "learning_rate": 1.47665080682255e-05, "loss": 0.5133, "step": 2610 }, { "epoch": 1.128459136427264, "grad_norm": 0.6730148792266846, "learning_rate": 1.4724698313202825e-05, "loss": 0.5224, "step": 2620 }, { "epoch": 1.132766232367826, "grad_norm": 0.5355139374732971, "learning_rate": 1.4682781961266546e-05, "loss": 0.5188, "step": 2630 }, { "epoch": 1.137073328308388, "grad_norm": 0.5199829936027527, "learning_rate": 1.4640759958118045e-05, "loss": 0.5121, "step": 2640 }, { "epoch": 1.14138042424895, "grad_norm": 0.5292408466339111, "learning_rate": 1.4598633251842373e-05, "loss": 0.5267, "step": 2650 }, { "epoch": 1.1456875201895123, "grad_norm": 0.5363121032714844, "learning_rate": 1.4556402792886856e-05, "loss": 0.5147, "step": 2660 }, { "epoch": 1.1499946161300743, "grad_norm": 0.5359490513801575, "learning_rate": 1.4514069534039649e-05, "loss": 0.5155, "step": 2670 }, { "epoch": 1.1543017120706365, "grad_norm": 0.4707220792770386, "learning_rate": 1.4471634430408244e-05, "loss": 0.5419, "step": 2680 }, { "epoch": 1.1586088080111985, "grad_norm": 0.4798811376094818, "learning_rate": 1.4429098439397901e-05, "loss": 0.5152, "step": 2690 }, { "epoch": 1.1629159039517605, "grad_norm": 0.4730081260204315, "learning_rate": 1.4386462520690087e-05, "loss": 0.5283, "step": 2700 }, { "epoch": 1.1672229998923225, "grad_norm": 0.524276614189148, "learning_rate": 1.4343727636220785e-05, "loss": 0.5087, "step": 2710 }, { "epoch": 1.1715300958328847, "grad_norm": 0.5093454122543335, "learning_rate": 1.430089475015882e-05, "loss": 0.5371, "step": 2720 }, { "epoch": 1.1758371917734467, "grad_norm": 0.5228180289268494, "learning_rate": 1.4257964828884077e-05, "loss": 0.5121, "step": 2730 }, { "epoch": 1.180144287714009, "grad_norm": 0.5263434052467346, "learning_rate": 1.4214938840965729e-05, "loss": 0.5104, "step": 2740 }, { "epoch": 1.184451383654571, "grad_norm": 0.5519675612449646, "learning_rate": 1.417181775714036e-05, "loss": 0.5081, "step": 2750 }, { "epoch": 1.188758479595133, "grad_norm": 0.48901626467704773, "learning_rate": 1.4128602550290078e-05, "loss": 0.5332, "step": 2760 }, { "epoch": 1.1930655755356951, "grad_norm": 0.5022098422050476, "learning_rate": 1.4085294195420563e-05, "loss": 0.5267, "step": 2770 }, { "epoch": 1.1973726714762571, "grad_norm": 0.5244942307472229, "learning_rate": 1.4041893669639053e-05, "loss": 0.5309, "step": 2780 }, { "epoch": 1.2016797674168191, "grad_norm": 0.5060109496116638, "learning_rate": 1.399840195213233e-05, "loss": 0.509, "step": 2790 }, { "epoch": 1.2059868633573814, "grad_norm": 0.48709142208099365, "learning_rate": 1.3954820024144595e-05, "loss": 0.5249, "step": 2800 }, { "epoch": 1.2102939592979434, "grad_norm": 0.48755279183387756, "learning_rate": 1.3911148868955357e-05, "loss": 0.5216, "step": 2810 }, { "epoch": 1.2146010552385054, "grad_norm": 0.4871668219566345, "learning_rate": 1.3867389471857229e-05, "loss": 0.5199, "step": 2820 }, { "epoch": 1.2189081511790676, "grad_norm": 0.5313363671302795, "learning_rate": 1.3823542820133706e-05, "loss": 0.5146, "step": 2830 }, { "epoch": 1.2232152471196296, "grad_norm": 0.48473960161209106, "learning_rate": 1.3779609903036894e-05, "loss": 0.5126, "step": 2840 }, { "epoch": 1.2275223430601916, "grad_norm": 0.5411814451217651, "learning_rate": 1.3735591711765189e-05, "loss": 0.5186, "step": 2850 }, { "epoch": 1.2318294390007538, "grad_norm": 0.5286210775375366, "learning_rate": 1.3691489239440899e-05, "loss": 0.513, "step": 2860 }, { "epoch": 1.2361365349413158, "grad_norm": 0.47112423181533813, "learning_rate": 1.3647303481087858e-05, "loss": 0.5268, "step": 2870 }, { "epoch": 1.2404436308818778, "grad_norm": 0.5465208888053894, "learning_rate": 1.3603035433608977e-05, "loss": 0.5109, "step": 2880 }, { "epoch": 1.24475072682244, "grad_norm": 0.4758882522583008, "learning_rate": 1.3558686095763732e-05, "loss": 0.5307, "step": 2890 }, { "epoch": 1.249057822763002, "grad_norm": 0.5721794962882996, "learning_rate": 1.3514256468145645e-05, "loss": 0.5104, "step": 2900 }, { "epoch": 1.2533649187035643, "grad_norm": 0.5125982761383057, "learning_rate": 1.3469747553159714e-05, "loss": 0.5278, "step": 2910 }, { "epoch": 1.2576720146441263, "grad_norm": 0.5272653698921204, "learning_rate": 1.342516035499978e-05, "loss": 0.5276, "step": 2920 }, { "epoch": 1.2619791105846883, "grad_norm": 0.5423816442489624, "learning_rate": 1.3380495879625884e-05, "loss": 0.5408, "step": 2930 }, { "epoch": 1.2662862065252503, "grad_norm": 0.4817509055137634, "learning_rate": 1.333575513474157e-05, "loss": 0.5152, "step": 2940 }, { "epoch": 1.2705933024658125, "grad_norm": 0.5113592147827148, "learning_rate": 1.3290939129771143e-05, "loss": 0.5397, "step": 2950 }, { "epoch": 1.2749003984063745, "grad_norm": 0.5106224417686462, "learning_rate": 1.3246048875836898e-05, "loss": 0.5269, "step": 2960 }, { "epoch": 1.2792074943469367, "grad_norm": 0.5446826219558716, "learning_rate": 1.3201085385736313e-05, "loss": 0.5252, "step": 2970 }, { "epoch": 1.2835145902874987, "grad_norm": 0.484943151473999, "learning_rate": 1.3156049673919184e-05, "loss": 0.525, "step": 2980 }, { "epoch": 1.2878216862280607, "grad_norm": 0.5692194700241089, "learning_rate": 1.3110942756464764e-05, "loss": 0.5197, "step": 2990 }, { "epoch": 1.2921287821686227, "grad_norm": 0.5009827017784119, "learning_rate": 1.3065765651058802e-05, "loss": 0.5325, "step": 3000 }, { "epoch": 1.296435878109185, "grad_norm": 0.4953298568725586, "learning_rate": 1.3020519376970613e-05, "loss": 0.5095, "step": 3010 }, { "epoch": 1.300742974049747, "grad_norm": 0.5116891264915466, "learning_rate": 1.2975204955030068e-05, "loss": 0.5263, "step": 3020 }, { "epoch": 1.3050500699903091, "grad_norm": 0.4844088554382324, "learning_rate": 1.2929823407604567e-05, "loss": 0.5113, "step": 3030 }, { "epoch": 1.3093571659308711, "grad_norm": 0.4732029438018799, "learning_rate": 1.2884375758575967e-05, "loss": 0.532, "step": 3040 }, { "epoch": 1.3136642618714331, "grad_norm": 0.5469485521316528, "learning_rate": 1.2838863033317484e-05, "loss": 0.519, "step": 3050 }, { "epoch": 1.3179713578119951, "grad_norm": 0.4888254702091217, "learning_rate": 1.2793286258670565e-05, "loss": 0.5097, "step": 3060 }, { "epoch": 1.3222784537525574, "grad_norm": 0.5359517335891724, "learning_rate": 1.2747646462921717e-05, "loss": 0.5246, "step": 3070 }, { "epoch": 1.3265855496931194, "grad_norm": 0.5013801455497742, "learning_rate": 1.2701944675779299e-05, "loss": 0.524, "step": 3080 }, { "epoch": 1.3308926456336816, "grad_norm": 0.49307557940483093, "learning_rate": 1.2656181928350301e-05, "loss": 0.5403, "step": 3090 }, { "epoch": 1.3351997415742436, "grad_norm": 0.47625210881233215, "learning_rate": 1.2610359253117078e-05, "loss": 0.5275, "step": 3100 }, { "epoch": 1.3395068375148056, "grad_norm": 0.5096368789672852, "learning_rate": 1.2564477683914053e-05, "loss": 0.5231, "step": 3110 }, { "epoch": 1.3438139334553676, "grad_norm": 0.4992668926715851, "learning_rate": 1.2518538255904389e-05, "loss": 0.5235, "step": 3120 }, { "epoch": 1.3481210293959298, "grad_norm": 0.491062194108963, "learning_rate": 1.2472542005556647e-05, "loss": 0.5432, "step": 3130 }, { "epoch": 1.3524281253364918, "grad_norm": 0.48666131496429443, "learning_rate": 1.2426489970621385e-05, "loss": 0.531, "step": 3140 }, { "epoch": 1.356735221277054, "grad_norm": 0.4706876575946808, "learning_rate": 1.2380383190107757e-05, "loss": 0.5188, "step": 3150 }, { "epoch": 1.361042317217616, "grad_norm": 0.4910385310649872, "learning_rate": 1.2334222704260063e-05, "loss": 0.5106, "step": 3160 }, { "epoch": 1.365349413158178, "grad_norm": 0.506514847278595, "learning_rate": 1.2288009554534291e-05, "loss": 0.5292, "step": 3170 }, { "epoch": 1.36965650909874, "grad_norm": 0.49671700596809387, "learning_rate": 1.2241744783574596e-05, "loss": 0.5284, "step": 3180 }, { "epoch": 1.3739636050393023, "grad_norm": 0.4892718195915222, "learning_rate": 1.219542943518981e-05, "loss": 0.5215, "step": 3190 }, { "epoch": 1.3782707009798643, "grad_norm": 0.5412102937698364, "learning_rate": 1.2149064554329864e-05, "loss": 0.5256, "step": 3200 }, { "epoch": 1.3825777969204265, "grad_norm": 0.4869970679283142, "learning_rate": 1.2102651187062227e-05, "loss": 0.5218, "step": 3210 }, { "epoch": 1.3868848928609885, "grad_norm": 0.5195066332817078, "learning_rate": 1.2056190380548299e-05, "loss": 0.5269, "step": 3220 }, { "epoch": 1.3911919888015505, "grad_norm": 0.5343438982963562, "learning_rate": 1.2009683183019788e-05, "loss": 0.5301, "step": 3230 }, { "epoch": 1.3954990847421127, "grad_norm": 0.522270679473877, "learning_rate": 1.1963130643755055e-05, "loss": 0.545, "step": 3240 }, { "epoch": 1.3998061806826747, "grad_norm": 0.501485288143158, "learning_rate": 1.191653381305545e-05, "loss": 0.5253, "step": 3250 }, { "epoch": 1.4041132766232367, "grad_norm": 0.5288712382316589, "learning_rate": 1.186989374222161e-05, "loss": 0.5181, "step": 3260 }, { "epoch": 1.408420372563799, "grad_norm": 0.5131502151489258, "learning_rate": 1.1823211483529733e-05, "loss": 0.5138, "step": 3270 }, { "epoch": 1.412727468504361, "grad_norm": 0.4853404462337494, "learning_rate": 1.1776488090207852e-05, "loss": 0.5319, "step": 3280 }, { "epoch": 1.417034564444923, "grad_norm": 0.5093010663986206, "learning_rate": 1.1729724616412062e-05, "loss": 0.5155, "step": 3290 }, { "epoch": 1.4213416603854852, "grad_norm": 0.5078168511390686, "learning_rate": 1.1682922117202736e-05, "loss": 0.5206, "step": 3300 }, { "epoch": 1.4256487563260472, "grad_norm": 0.5315324664115906, "learning_rate": 1.163608164852073e-05, "loss": 0.5314, "step": 3310 }, { "epoch": 1.4299558522666094, "grad_norm": 0.4705192446708679, "learning_rate": 1.1589204267163545e-05, "loss": 0.4966, "step": 3320 }, { "epoch": 1.4342629482071714, "grad_norm": 0.48757535219192505, "learning_rate": 1.15422910307615e-05, "loss": 0.5299, "step": 3330 }, { "epoch": 1.4385700441477334, "grad_norm": 0.5582148432731628, "learning_rate": 1.1495342997753864e-05, "loss": 0.5201, "step": 3340 }, { "epoch": 1.4428771400882954, "grad_norm": 0.5134326219558716, "learning_rate": 1.1448361227364963e-05, "loss": 0.5061, "step": 3350 }, { "epoch": 1.4471842360288576, "grad_norm": 0.5316387414932251, "learning_rate": 1.1401346779580303e-05, "loss": 0.5145, "step": 3360 }, { "epoch": 1.4514913319694196, "grad_norm": 0.5328738689422607, "learning_rate": 1.1354300715122637e-05, "loss": 0.5288, "step": 3370 }, { "epoch": 1.4557984279099818, "grad_norm": 0.5279168486595154, "learning_rate": 1.1307224095428058e-05, "loss": 0.5031, "step": 3380 }, { "epoch": 1.4601055238505438, "grad_norm": 0.5049686431884766, "learning_rate": 1.1260117982622021e-05, "loss": 0.5004, "step": 3390 }, { "epoch": 1.4644126197911058, "grad_norm": 0.47000184655189514, "learning_rate": 1.1212983439495392e-05, "loss": 0.5267, "step": 3400 }, { "epoch": 1.4687197157316678, "grad_norm": 0.49505382776260376, "learning_rate": 1.1165821529480483e-05, "loss": 0.5278, "step": 3410 }, { "epoch": 1.47302681167223, "grad_norm": 0.568454384803772, "learning_rate": 1.1118633316627037e-05, "loss": 0.5116, "step": 3420 }, { "epoch": 1.477333907612792, "grad_norm": 0.5094279646873474, "learning_rate": 1.1071419865578241e-05, "loss": 0.5181, "step": 3430 }, { "epoch": 1.4816410035533543, "grad_norm": 0.5605435371398926, "learning_rate": 1.1024182241546686e-05, "loss": 0.5191, "step": 3440 }, { "epoch": 1.4859480994939163, "grad_norm": 0.49941274523735046, "learning_rate": 1.097692151029036e-05, "loss": 0.5036, "step": 3450 }, { "epoch": 1.4902551954344783, "grad_norm": 0.5064433813095093, "learning_rate": 1.0929638738088571e-05, "loss": 0.5195, "step": 3460 }, { "epoch": 1.4945622913750403, "grad_norm": 0.5021061301231384, "learning_rate": 1.088233499171792e-05, "loss": 0.522, "step": 3470 }, { "epoch": 1.4988693873156025, "grad_norm": 0.5188096761703491, "learning_rate": 1.0835011338428217e-05, "loss": 0.5156, "step": 3480 }, { "epoch": 1.5031764832561645, "grad_norm": 0.6124559640884399, "learning_rate": 1.0787668845918393e-05, "loss": 0.5145, "step": 3490 }, { "epoch": 1.5074835791967267, "grad_norm": 0.48937344551086426, "learning_rate": 1.074030858231244e-05, "loss": 0.515, "step": 3500 }, { "epoch": 1.5117906751372887, "grad_norm": 0.518526017665863, "learning_rate": 1.0692931616135283e-05, "loss": 0.505, "step": 3510 }, { "epoch": 1.5160977710778507, "grad_norm": 0.5395667552947998, "learning_rate": 1.0645539016288686e-05, "loss": 0.5076, "step": 3520 }, { "epoch": 1.5204048670184127, "grad_norm": 0.495190292596817, "learning_rate": 1.059813185202714e-05, "loss": 0.523, "step": 3530 }, { "epoch": 1.524711962958975, "grad_norm": 0.49644342064857483, "learning_rate": 1.055071119293373e-05, "loss": 0.5038, "step": 3540 }, { "epoch": 1.5290190588995372, "grad_norm": 0.483696848154068, "learning_rate": 1.0503278108896e-05, "loss": 0.5103, "step": 3550 }, { "epoch": 1.5333261548400992, "grad_norm": 0.5149986147880554, "learning_rate": 1.0455833670081831e-05, "loss": 0.5402, "step": 3560 }, { "epoch": 1.5376332507806612, "grad_norm": 0.4734952449798584, "learning_rate": 1.0408378946915282e-05, "loss": 0.5292, "step": 3570 }, { "epoch": 1.5419403467212232, "grad_norm": 0.5490080118179321, "learning_rate": 1.0360915010052443e-05, "loss": 0.5155, "step": 3580 }, { "epoch": 1.5462474426617852, "grad_norm": 0.5176838636398315, "learning_rate": 1.0313442930357278e-05, "loss": 0.5111, "step": 3590 }, { "epoch": 1.5505545386023474, "grad_norm": 0.5659157633781433, "learning_rate": 1.026596377887747e-05, "loss": 0.5152, "step": 3600 }, { "epoch": 1.5548616345429096, "grad_norm": 0.5195504426956177, "learning_rate": 1.0218478626820256e-05, "loss": 0.5178, "step": 3610 }, { "epoch": 1.5591687304834716, "grad_norm": 0.533338189125061, "learning_rate": 1.0170988545528248e-05, "loss": 0.5138, "step": 3620 }, { "epoch": 1.5634758264240336, "grad_norm": 0.5108840465545654, "learning_rate": 1.0123494606455278e-05, "loss": 0.5273, "step": 3630 }, { "epoch": 1.5677829223645956, "grad_norm": 0.4785379469394684, "learning_rate": 1.0075997881142208e-05, "loss": 0.5071, "step": 3640 }, { "epoch": 1.5720900183051576, "grad_norm": 0.49497827887535095, "learning_rate": 1.0028499441192765e-05, "loss": 0.5132, "step": 3650 }, { "epoch": 1.5763971142457198, "grad_norm": 0.5214102864265442, "learning_rate": 9.981000358249368e-06, "loss": 0.5133, "step": 3660 }, { "epoch": 1.580704210186282, "grad_norm": 0.47462400794029236, "learning_rate": 9.933501703968928e-06, "loss": 0.5226, "step": 3670 }, { "epoch": 1.585011306126844, "grad_norm": 0.4743979275226593, "learning_rate": 9.8860045499987e-06, "loss": 0.5219, "step": 3680 }, { "epoch": 1.589318402067406, "grad_norm": 0.5265910625457764, "learning_rate": 9.838509967952076e-06, "loss": 0.4945, "step": 3690 }, { "epoch": 1.593625498007968, "grad_norm": 0.5075172185897827, "learning_rate": 9.791019029384437e-06, "loss": 0.5175, "step": 3700 }, { "epoch": 1.59793259394853, "grad_norm": 0.5206677913665771, "learning_rate": 9.743532805768948e-06, "loss": 0.5188, "step": 3710 }, { "epoch": 1.6022396898890923, "grad_norm": 0.4802674651145935, "learning_rate": 9.696052368472406e-06, "loss": 0.5064, "step": 3720 }, { "epoch": 1.6065467858296545, "grad_norm": 0.5289535522460938, "learning_rate": 9.648578788731044e-06, "loss": 0.5281, "step": 3730 }, { "epoch": 1.6108538817702165, "grad_norm": 0.47722700238227844, "learning_rate": 9.601113137626394e-06, "loss": 0.5151, "step": 3740 }, { "epoch": 1.6151609777107785, "grad_norm": 0.4994152784347534, "learning_rate": 9.553656486061098e-06, "loss": 0.52, "step": 3750 }, { "epoch": 1.6194680736513405, "grad_norm": 0.48130089044570923, "learning_rate": 9.506209904734753e-06, "loss": 0.5336, "step": 3760 }, { "epoch": 1.6237751695919027, "grad_norm": 0.48449528217315674, "learning_rate": 9.45877446411976e-06, "loss": 0.5252, "step": 3770 }, { "epoch": 1.6280822655324647, "grad_norm": 0.5411643981933594, "learning_rate": 9.411351234437163e-06, "loss": 0.5187, "step": 3780 }, { "epoch": 1.632389361473027, "grad_norm": 0.5133873820304871, "learning_rate": 9.363941285632507e-06, "loss": 0.5217, "step": 3790 }, { "epoch": 1.636696457413589, "grad_norm": 0.5814666748046875, "learning_rate": 9.3165456873517e-06, "loss": 0.5, "step": 3800 }, { "epoch": 1.641003553354151, "grad_norm": 0.52715665102005, "learning_rate": 9.269165508916883e-06, "loss": 0.5184, "step": 3810 }, { "epoch": 1.645310649294713, "grad_norm": 0.48196879029273987, "learning_rate": 9.221801819302288e-06, "loss": 0.5191, "step": 3820 }, { "epoch": 1.6496177452352752, "grad_norm": 0.49397778511047363, "learning_rate": 9.174455687110142e-06, "loss": 0.5013, "step": 3830 }, { "epoch": 1.6539248411758372, "grad_norm": 0.5037091970443726, "learning_rate": 9.127128180546548e-06, "loss": 0.5298, "step": 3840 }, { "epoch": 1.6582319371163994, "grad_norm": 0.5031833052635193, "learning_rate": 9.079820367397384e-06, "loss": 0.4929, "step": 3850 }, { "epoch": 1.6625390330569614, "grad_norm": 0.5380353927612305, "learning_rate": 9.032533315004207e-06, "loss": 0.4968, "step": 3860 }, { "epoch": 1.6668461289975234, "grad_norm": 0.5191226005554199, "learning_rate": 8.98526809024018e-06, "loss": 0.5267, "step": 3870 }, { "epoch": 1.6711532249380854, "grad_norm": 0.5179468393325806, "learning_rate": 8.938025759486007e-06, "loss": 0.5159, "step": 3880 }, { "epoch": 1.6754603208786476, "grad_norm": 0.4779166579246521, "learning_rate": 8.89080738860585e-06, "loss": 0.5211, "step": 3890 }, { "epoch": 1.6797674168192096, "grad_norm": 0.5136571526527405, "learning_rate": 8.843614042923318e-06, "loss": 0.5003, "step": 3900 }, { "epoch": 1.6840745127597718, "grad_norm": 0.540773332118988, "learning_rate": 8.796446787197383e-06, "loss": 0.5131, "step": 3910 }, { "epoch": 1.6883816087003338, "grad_norm": 0.5126665234565735, "learning_rate": 8.749306685598409e-06, "loss": 0.5093, "step": 3920 }, { "epoch": 1.6926887046408958, "grad_norm": 0.47659188508987427, "learning_rate": 8.702194801684112e-06, "loss": 0.5158, "step": 3930 }, { "epoch": 1.6969958005814578, "grad_norm": 0.47945475578308105, "learning_rate": 8.655112198375564e-06, "loss": 0.5026, "step": 3940 }, { "epoch": 1.70130289652202, "grad_norm": 0.4939498007297516, "learning_rate": 8.60805993793323e-06, "loss": 0.5099, "step": 3950 }, { "epoch": 1.7056099924625823, "grad_norm": 0.5328351259231567, "learning_rate": 8.561039081932975e-06, "loss": 0.52, "step": 3960 }, { "epoch": 1.7099170884031443, "grad_norm": 0.49865198135375977, "learning_rate": 8.514050691242145e-06, "loss": 0.5077, "step": 3970 }, { "epoch": 1.7142241843437063, "grad_norm": 0.49807870388031006, "learning_rate": 8.467095825995605e-06, "loss": 0.4976, "step": 3980 }, { "epoch": 1.7185312802842683, "grad_norm": 0.5023031234741211, "learning_rate": 8.420175545571837e-06, "loss": 0.5233, "step": 3990 }, { "epoch": 1.7228383762248303, "grad_norm": 0.49054110050201416, "learning_rate": 8.373290908569026e-06, "loss": 0.5115, "step": 4000 }, { "epoch": 1.7271454721653925, "grad_norm": 0.47637811303138733, "learning_rate": 8.32644297278119e-06, "loss": 0.5103, "step": 4010 }, { "epoch": 1.7314525681059547, "grad_norm": 0.5239661931991577, "learning_rate": 8.279632795174304e-06, "loss": 0.5161, "step": 4020 }, { "epoch": 1.7357596640465167, "grad_norm": 0.5000544190406799, "learning_rate": 8.232861431862457e-06, "loss": 0.5113, "step": 4030 }, { "epoch": 1.7400667599870787, "grad_norm": 0.5361005067825317, "learning_rate": 8.186129938084028e-06, "loss": 0.5137, "step": 4040 }, { "epoch": 1.7443738559276407, "grad_norm": 0.48270535469055176, "learning_rate": 8.139439368177868e-06, "loss": 0.5116, "step": 4050 }, { "epoch": 1.7486809518682027, "grad_norm": 0.48645904660224915, "learning_rate": 8.092790775559522e-06, "loss": 0.517, "step": 4060 }, { "epoch": 1.752988047808765, "grad_norm": 0.4865799844264984, "learning_rate": 8.046185212697459e-06, "loss": 0.5202, "step": 4070 }, { "epoch": 1.7572951437493272, "grad_norm": 0.5095897912979126, "learning_rate": 7.999623731089327e-06, "loss": 0.5186, "step": 4080 }, { "epoch": 1.7616022396898892, "grad_norm": 0.49918055534362793, "learning_rate": 7.953107381238226e-06, "loss": 0.5091, "step": 4090 }, { "epoch": 1.7659093356304512, "grad_norm": 0.5209227204322815, "learning_rate": 7.906637212629011e-06, "loss": 0.5098, "step": 4100 }, { "epoch": 1.7702164315710132, "grad_norm": 0.5320930480957031, "learning_rate": 7.860214273704614e-06, "loss": 0.5172, "step": 4110 }, { "epoch": 1.7745235275115752, "grad_norm": 0.4841155707836151, "learning_rate": 7.813839611842387e-06, "loss": 0.4851, "step": 4120 }, { "epoch": 1.7788306234521374, "grad_norm": 0.5300472378730774, "learning_rate": 7.767514273330473e-06, "loss": 0.4953, "step": 4130 }, { "epoch": 1.7831377193926996, "grad_norm": 0.5021957159042358, "learning_rate": 7.721239303344201e-06, "loss": 0.5112, "step": 4140 }, { "epoch": 1.7874448153332616, "grad_norm": 0.498737096786499, "learning_rate": 7.675015745922499e-06, "loss": 0.5045, "step": 4150 }, { "epoch": 1.7917519112738236, "grad_norm": 0.4690532684326172, "learning_rate": 7.628844643944349e-06, "loss": 0.5102, "step": 4160 }, { "epoch": 1.7960590072143856, "grad_norm": 0.5077162384986877, "learning_rate": 7.582727039105255e-06, "loss": 0.5105, "step": 4170 }, { "epoch": 1.8003661031549478, "grad_norm": 0.47492554783821106, "learning_rate": 7.536663971893724e-06, "loss": 0.5008, "step": 4180 }, { "epoch": 1.8046731990955098, "grad_norm": 0.5036799907684326, "learning_rate": 7.4906564815678205e-06, "loss": 0.5179, "step": 4190 }, { "epoch": 1.808980295036072, "grad_norm": 0.5044455528259277, "learning_rate": 7.444705606131697e-06, "loss": 0.5171, "step": 4200 }, { "epoch": 1.813287390976634, "grad_norm": 0.5645790696144104, "learning_rate": 7.39881238231218e-06, "loss": 0.5111, "step": 4210 }, { "epoch": 1.817594486917196, "grad_norm": 0.4966265857219696, "learning_rate": 7.352977845535387e-06, "loss": 0.5144, "step": 4220 }, { "epoch": 1.821901582857758, "grad_norm": 0.5225628614425659, "learning_rate": 7.307203029903354e-06, "loss": 0.5115, "step": 4230 }, { "epoch": 1.8262086787983203, "grad_norm": 0.5282090902328491, "learning_rate": 7.261488968170713e-06, "loss": 0.5251, "step": 4240 }, { "epoch": 1.8305157747388823, "grad_norm": 0.5346629023551941, "learning_rate": 7.21583669172139e-06, "loss": 0.5042, "step": 4250 }, { "epoch": 1.8348228706794445, "grad_norm": 0.5141210556030273, "learning_rate": 7.170247230545335e-06, "loss": 0.5199, "step": 4260 }, { "epoch": 1.8391299666200065, "grad_norm": 0.5251668691635132, "learning_rate": 7.124721613215275e-06, "loss": 0.4936, "step": 4270 }, { "epoch": 1.8434370625605685, "grad_norm": 0.5125293731689453, "learning_rate": 7.079260866863523e-06, "loss": 0.5161, "step": 4280 }, { "epoch": 1.8477441585011305, "grad_norm": 0.4881208837032318, "learning_rate": 7.033866017158797e-06, "loss": 0.5142, "step": 4290 }, { "epoch": 1.8520512544416927, "grad_norm": 0.5215027928352356, "learning_rate": 6.9885380882830735e-06, "loss": 0.5097, "step": 4300 }, { "epoch": 1.8563583503822547, "grad_norm": 0.4931368827819824, "learning_rate": 6.943278102908491e-06, "loss": 0.5123, "step": 4310 }, { "epoch": 1.860665446322817, "grad_norm": 0.5080362558364868, "learning_rate": 6.898087082174267e-06, "loss": 0.5093, "step": 4320 }, { "epoch": 1.864972542263379, "grad_norm": 0.537807285785675, "learning_rate": 6.852966045663671e-06, "loss": 0.5245, "step": 4330 }, { "epoch": 1.869279638203941, "grad_norm": 0.5395597815513611, "learning_rate": 6.807916011381008e-06, "loss": 0.5016, "step": 4340 }, { "epoch": 1.873586734144503, "grad_norm": 0.48623430728912354, "learning_rate": 6.762937995728663e-06, "loss": 0.4962, "step": 4350 }, { "epoch": 1.8778938300850652, "grad_norm": 0.5058403611183167, "learning_rate": 6.718033013484147e-06, "loss": 0.5401, "step": 4360 }, { "epoch": 1.8822009260256274, "grad_norm": 0.5220633149147034, "learning_rate": 6.673202077777239e-06, "loss": 0.5112, "step": 4370 }, { "epoch": 1.8865080219661894, "grad_norm": 0.5163370966911316, "learning_rate": 6.6284462000670924e-06, "loss": 0.5231, "step": 4380 }, { "epoch": 1.8908151179067514, "grad_norm": 0.508660614490509, "learning_rate": 6.583766390119437e-06, "loss": 0.5304, "step": 4390 }, { "epoch": 1.8951222138473134, "grad_norm": 0.568144679069519, "learning_rate": 6.539163655983786e-06, "loss": 0.5086, "step": 4400 }, { "epoch": 1.8994293097878754, "grad_norm": 0.5001341700553894, "learning_rate": 6.494639003970701e-06, "loss": 0.5084, "step": 4410 }, { "epoch": 1.9037364057284376, "grad_norm": 0.5228297710418701, "learning_rate": 6.450193438629078e-06, "loss": 0.504, "step": 4420 }, { "epoch": 1.9080435016689998, "grad_norm": 0.4816001057624817, "learning_rate": 6.40582796272349e-06, "loss": 0.5102, "step": 4430 }, { "epoch": 1.9123505976095618, "grad_norm": 0.5058324933052063, "learning_rate": 6.361543577211566e-06, "loss": 0.524, "step": 4440 }, { "epoch": 1.9166576935501238, "grad_norm": 0.5428106188774109, "learning_rate": 6.317341281221392e-06, "loss": 0.5082, "step": 4450 }, { "epoch": 1.9209647894906858, "grad_norm": 0.5131290555000305, "learning_rate": 6.273222072028991e-06, "loss": 0.5316, "step": 4460 }, { "epoch": 1.9252718854312478, "grad_norm": 0.5238609910011292, "learning_rate": 6.2291869450358074e-06, "loss": 0.5021, "step": 4470 }, { "epoch": 1.92957898137181, "grad_norm": 0.4843258261680603, "learning_rate": 6.1852368937462585e-06, "loss": 0.5048, "step": 4480 }, { "epoch": 1.9338860773123723, "grad_norm": 0.5138316750526428, "learning_rate": 6.141372909745307e-06, "loss": 0.5352, "step": 4490 }, { "epoch": 1.9381931732529343, "grad_norm": 0.49319642782211304, "learning_rate": 6.097595982676103e-06, "loss": 0.5065, "step": 4500 }, { "epoch": 1.9425002691934963, "grad_norm": 0.5176106095314026, "learning_rate": 6.053907100217648e-06, "loss": 0.5155, "step": 4510 }, { "epoch": 1.9468073651340583, "grad_norm": 0.4772352874279022, "learning_rate": 6.010307248062514e-06, "loss": 0.5056, "step": 4520 }, { "epoch": 1.9511144610746203, "grad_norm": 0.5366437435150146, "learning_rate": 5.966797409894607e-06, "loss": 0.4888, "step": 4530 }, { "epoch": 1.9554215570151825, "grad_norm": 0.4917809069156647, "learning_rate": 5.923378567366956e-06, "loss": 0.5221, "step": 4540 }, { "epoch": 1.9597286529557447, "grad_norm": 0.5597509741783142, "learning_rate": 5.880051700079596e-06, "loss": 0.5225, "step": 4550 }, { "epoch": 1.9640357488963067, "grad_norm": 0.5258151888847351, "learning_rate": 5.836817785557448e-06, "loss": 0.5031, "step": 4560 }, { "epoch": 1.9683428448368687, "grad_norm": 0.5679864287376404, "learning_rate": 5.7936777992282565e-06, "loss": 0.5074, "step": 4570 }, { "epoch": 1.9726499407774307, "grad_norm": 0.5309889912605286, "learning_rate": 5.750632714400607e-06, "loss": 0.521, "step": 4580 }, { "epoch": 1.976957036717993, "grad_norm": 0.5293132662773132, "learning_rate": 5.707683502241936e-06, "loss": 0.5133, "step": 4590 }, { "epoch": 1.981264132658555, "grad_norm": 0.5223381519317627, "learning_rate": 5.664831131756652e-06, "loss": 0.5129, "step": 4600 }, { "epoch": 1.9855712285991172, "grad_norm": 0.5365522503852844, "learning_rate": 5.622076569764247e-06, "loss": 0.504, "step": 4610 }, { "epoch": 1.9898783245396792, "grad_norm": 0.5084212422370911, "learning_rate": 5.5794207808774904e-06, "loss": 0.488, "step": 4620 }, { "epoch": 1.9941854204802412, "grad_norm": 0.4913804531097412, "learning_rate": 5.536864727480683e-06, "loss": 0.5098, "step": 4630 }, { "epoch": 1.9984925164208032, "grad_norm": 0.5197212100028992, "learning_rate": 5.4944093697079136e-06, "loss": 0.5066, "step": 4640 }, { "epoch": 2.002799612361365, "grad_norm": 0.51143479347229, "learning_rate": 5.45205566542143e-06, "loss": 0.4521, "step": 4650 }, { "epoch": 2.0071067083019276, "grad_norm": 0.5107315182685852, "learning_rate": 5.4098045701899934e-06, "loss": 0.3968, "step": 4660 }, { "epoch": 2.0114138042424896, "grad_norm": 0.5407351851463318, "learning_rate": 5.367657037267354e-06, "loss": 0.3933, "step": 4670 }, { "epoch": 2.0157209001830516, "grad_norm": 0.5835046172142029, "learning_rate": 5.325614017570712e-06, "loss": 0.3897, "step": 4680 }, { "epoch": 2.0200279961236136, "grad_norm": 0.5047739744186401, "learning_rate": 5.283676459659288e-06, "loss": 0.3992, "step": 4690 }, { "epoch": 2.0243350920641756, "grad_norm": 0.5422953963279724, "learning_rate": 5.241845309712921e-06, "loss": 0.4131, "step": 4700 }, { "epoch": 2.0286421880047376, "grad_norm": 0.5471384525299072, "learning_rate": 5.2001215115106814e-06, "loss": 0.3955, "step": 4710 }, { "epoch": 2.0329492839453, "grad_norm": 0.5800908803939819, "learning_rate": 5.158506006409644e-06, "loss": 0.397, "step": 4720 }, { "epoch": 2.037256379885862, "grad_norm": 0.5329377055168152, "learning_rate": 5.116999733323591e-06, "loss": 0.4017, "step": 4730 }, { "epoch": 2.041563475826424, "grad_norm": 0.556845486164093, "learning_rate": 5.075603628701869e-06, "loss": 0.4009, "step": 4740 }, { "epoch": 2.045870571766986, "grad_norm": 0.5501790642738342, "learning_rate": 5.034318626508223e-06, "loss": 0.3969, "step": 4750 }, { "epoch": 2.050177667707548, "grad_norm": 0.5467825531959534, "learning_rate": 4.993145658199766e-06, "loss": 0.3996, "step": 4760 }, { "epoch": 2.05448476364811, "grad_norm": 0.5644121766090393, "learning_rate": 4.952085652705938e-06, "loss": 0.3926, "step": 4770 }, { "epoch": 2.0587918595886725, "grad_norm": 0.5279033780097961, "learning_rate": 4.911139536407542e-06, "loss": 0.3742, "step": 4780 }, { "epoch": 2.0630989555292345, "grad_norm": 0.5283676981925964, "learning_rate": 4.870308233115876e-06, "loss": 0.3893, "step": 4790 }, { "epoch": 2.0674060514697965, "grad_norm": 0.5302291512489319, "learning_rate": 4.82959266405184e-06, "loss": 0.3956, "step": 4800 }, { "epoch": 2.0717131474103585, "grad_norm": 0.5381713509559631, "learning_rate": 4.788993747825209e-06, "loss": 0.4124, "step": 4810 }, { "epoch": 2.0760202433509205, "grad_norm": 0.5772622227668762, "learning_rate": 4.748512400413861e-06, "loss": 0.405, "step": 4820 }, { "epoch": 2.0803273392914825, "grad_norm": 0.5383191704750061, "learning_rate": 4.708149535143138e-06, "loss": 0.3874, "step": 4830 }, { "epoch": 2.084634435232045, "grad_norm": 0.5546970963478088, "learning_rate": 4.667906062665234e-06, "loss": 0.3994, "step": 4840 }, { "epoch": 2.088941531172607, "grad_norm": 0.5541481375694275, "learning_rate": 4.627782890938632e-06, "loss": 0.4073, "step": 4850 }, { "epoch": 2.093248627113169, "grad_norm": 0.5656886100769043, "learning_rate": 4.587780925207654e-06, "loss": 0.3986, "step": 4860 }, { "epoch": 2.097555723053731, "grad_norm": 0.5167860984802246, "learning_rate": 4.5479010679819965e-06, "loss": 0.3994, "step": 4870 }, { "epoch": 2.101862818994293, "grad_norm": 0.585415780544281, "learning_rate": 4.50814421901641e-06, "loss": 0.3959, "step": 4880 }, { "epoch": 2.1061699149348554, "grad_norm": 0.5390037894248962, "learning_rate": 4.46851127529035e-06, "loss": 0.393, "step": 4890 }, { "epoch": 2.1104770108754174, "grad_norm": 0.5685362815856934, "learning_rate": 4.42900313098779e-06, "loss": 0.4031, "step": 4900 }, { "epoch": 2.1147841068159794, "grad_norm": 0.5294394493103027, "learning_rate": 4.389620677477023e-06, "loss": 0.3926, "step": 4910 }, { "epoch": 2.1190912027565414, "grad_norm": 0.5693227648735046, "learning_rate": 4.3503648032905384e-06, "loss": 0.3909, "step": 4920 }, { "epoch": 2.1233982986971034, "grad_norm": 0.6294069886207581, "learning_rate": 4.311236394105006e-06, "loss": 0.3908, "step": 4930 }, { "epoch": 2.1277053946376654, "grad_norm": 0.566862165927887, "learning_rate": 4.27223633272126e-06, "loss": 0.4019, "step": 4940 }, { "epoch": 2.132012490578228, "grad_norm": 0.5680539608001709, "learning_rate": 4.233365499044416e-06, "loss": 0.3957, "step": 4950 }, { "epoch": 2.13631958651879, "grad_norm": 0.5697780251502991, "learning_rate": 4.194624770063985e-06, "loss": 0.3876, "step": 4960 }, { "epoch": 2.140626682459352, "grad_norm": 0.5857852697372437, "learning_rate": 4.1560150198341174e-06, "loss": 0.3986, "step": 4970 }, { "epoch": 2.144933778399914, "grad_norm": 0.5707722306251526, "learning_rate": 4.11753711945386e-06, "loss": 0.4165, "step": 4980 }, { "epoch": 2.149240874340476, "grad_norm": 0.5498836040496826, "learning_rate": 4.079191937047511e-06, "loss": 0.4236, "step": 4990 }, { "epoch": 2.153547970281038, "grad_norm": 0.6008414626121521, "learning_rate": 4.040980337745044e-06, "loss": 0.3955, "step": 5000 }, { "epoch": 2.1578550662216003, "grad_norm": 0.5871570110321045, "learning_rate": 4.002903183662566e-06, "loss": 0.3939, "step": 5010 }, { "epoch": 2.1621621621621623, "grad_norm": 0.5556260347366333, "learning_rate": 3.964961333882893e-06, "loss": 0.4005, "step": 5020 }, { "epoch": 2.1664692581027243, "grad_norm": 0.5592585206031799, "learning_rate": 3.927155644436144e-06, "loss": 0.4035, "step": 5030 }, { "epoch": 2.1707763540432863, "grad_norm": 0.5638931393623352, "learning_rate": 3.889486968280448e-06, "loss": 0.3961, "step": 5040 }, { "epoch": 2.1750834499838483, "grad_norm": 0.5473156571388245, "learning_rate": 3.851956155282682e-06, "loss": 0.3999, "step": 5050 }, { "epoch": 2.1793905459244103, "grad_norm": 0.7088154554367065, "learning_rate": 3.814564052199313e-06, "loss": 0.3919, "step": 5060 }, { "epoch": 2.1836976418649727, "grad_norm": 0.569315493106842, "learning_rate": 3.777311502657279e-06, "loss": 0.3924, "step": 5070 }, { "epoch": 2.1880047378055347, "grad_norm": 0.6128218770027161, "learning_rate": 3.7401993471349616e-06, "loss": 0.4094, "step": 5080 }, { "epoch": 2.1923118337460967, "grad_norm": 0.5971004962921143, "learning_rate": 3.7032284229432325e-06, "loss": 0.3786, "step": 5090 }, { "epoch": 2.1966189296866587, "grad_norm": 0.5701526999473572, "learning_rate": 3.666399564206541e-06, "loss": 0.3912, "step": 5100 }, { "epoch": 2.2009260256272207, "grad_norm": 0.5547009706497192, "learning_rate": 3.6297136018441215e-06, "loss": 0.3866, "step": 5110 }, { "epoch": 2.2052331215677827, "grad_norm": 0.5613463521003723, "learning_rate": 3.59317136355122e-06, "loss": 0.3926, "step": 5120 }, { "epoch": 2.209540217508345, "grad_norm": 0.6126610040664673, "learning_rate": 3.556773673780446e-06, "loss": 0.389, "step": 5130 }, { "epoch": 2.213847313448907, "grad_norm": 0.5699272751808167, "learning_rate": 3.520521353723142e-06, "loss": 0.3982, "step": 5140 }, { "epoch": 2.218154409389469, "grad_norm": 0.593333899974823, "learning_rate": 3.484415221290889e-06, "loss": 0.3826, "step": 5150 }, { "epoch": 2.222461505330031, "grad_norm": 0.6188777685165405, "learning_rate": 3.448456091097023e-06, "loss": 0.4, "step": 5160 }, { "epoch": 2.226768601270593, "grad_norm": 0.5949888825416565, "learning_rate": 3.4126447744382753e-06, "loss": 0.4062, "step": 5170 }, { "epoch": 2.231075697211155, "grad_norm": 0.5788257718086243, "learning_rate": 3.376982079276464e-06, "loss": 0.3881, "step": 5180 }, { "epoch": 2.2353827931517176, "grad_norm": 0.5726456642150879, "learning_rate": 3.3414688102202564e-06, "loss": 0.3968, "step": 5190 }, { "epoch": 2.2396898890922796, "grad_norm": 0.5855600833892822, "learning_rate": 3.3061057685070354e-06, "loss": 0.3925, "step": 5200 }, { "epoch": 2.2439969850328416, "grad_norm": 0.5823237299919128, "learning_rate": 3.2708937519847916e-06, "loss": 0.3875, "step": 5210 }, { "epoch": 2.2483040809734036, "grad_norm": 0.5852989554405212, "learning_rate": 3.23583355509416e-06, "loss": 0.3985, "step": 5220 }, { "epoch": 2.2526111769139656, "grad_norm": 0.5461825728416443, "learning_rate": 3.200925968850459e-06, "loss": 0.3917, "step": 5230 }, { "epoch": 2.256918272854528, "grad_norm": 0.5536659359931946, "learning_rate": 3.166171780825876e-06, "loss": 0.3963, "step": 5240 }, { "epoch": 2.26122536879509, "grad_norm": 0.5736192464828491, "learning_rate": 3.1315717751316755e-06, "loss": 0.4114, "step": 5250 }, { "epoch": 2.265532464735652, "grad_norm": 0.5808764100074768, "learning_rate": 3.097126732400515e-06, "loss": 0.3795, "step": 5260 }, { "epoch": 2.269839560676214, "grad_norm": 0.5790621042251587, "learning_rate": 3.0628374297688436e-06, "loss": 0.3991, "step": 5270 }, { "epoch": 2.274146656616776, "grad_norm": 0.5211635231971741, "learning_rate": 3.0287046408593478e-06, "loss": 0.3796, "step": 5280 }, { "epoch": 2.278453752557338, "grad_norm": 0.6152241230010986, "learning_rate": 2.994729135763522e-06, "loss": 0.3976, "step": 5290 }, { "epoch": 2.2827608484979, "grad_norm": 0.6017261147499084, "learning_rate": 2.9609116810242677e-06, "loss": 0.4031, "step": 5300 }, { "epoch": 2.2870679444384625, "grad_norm": 0.5612776279449463, "learning_rate": 2.9272530396186194e-06, "loss": 0.3985, "step": 5310 }, { "epoch": 2.2913750403790245, "grad_norm": 0.6065710186958313, "learning_rate": 2.893753970940525e-06, "loss": 0.3975, "step": 5320 }, { "epoch": 2.2956821363195865, "grad_norm": 0.5793972611427307, "learning_rate": 2.8604152307837064e-06, "loss": 0.3889, "step": 5330 }, { "epoch": 2.2999892322601485, "grad_norm": 0.5591062307357788, "learning_rate": 2.8272375713246125e-06, "loss": 0.3903, "step": 5340 }, { "epoch": 2.3042963282007105, "grad_norm": 0.5505937337875366, "learning_rate": 2.794221741105446e-06, "loss": 0.397, "step": 5350 }, { "epoch": 2.308603424141273, "grad_norm": 0.6174246668815613, "learning_rate": 2.7613684850172882e-06, "loss": 0.3966, "step": 5360 }, { "epoch": 2.312910520081835, "grad_norm": 0.6093124747276306, "learning_rate": 2.7286785442832685e-06, "loss": 0.3902, "step": 5370 }, { "epoch": 2.317217616022397, "grad_norm": 0.5350244045257568, "learning_rate": 2.696152656441868e-06, "loss": 0.3935, "step": 5380 }, { "epoch": 2.321524711962959, "grad_norm": 0.5422816276550293, "learning_rate": 2.663791555330255e-06, "loss": 0.3924, "step": 5390 }, { "epoch": 2.325831807903521, "grad_norm": 0.5582048892974854, "learning_rate": 2.6315959710677464e-06, "loss": 0.397, "step": 5400 }, { "epoch": 2.330138903844083, "grad_norm": 0.5601301789283752, "learning_rate": 2.599566630039332e-06, "loss": 0.3813, "step": 5410 }, { "epoch": 2.334445999784645, "grad_norm": 0.5601345896720886, "learning_rate": 2.567704254879274e-06, "loss": 0.3974, "step": 5420 }, { "epoch": 2.3387530957252074, "grad_norm": 0.614778459072113, "learning_rate": 2.536009564454817e-06, "loss": 0.3836, "step": 5430 }, { "epoch": 2.3430601916657694, "grad_norm": 0.5759994983673096, "learning_rate": 2.504483273849958e-06, "loss": 0.3949, "step": 5440 }, { "epoch": 2.3473672876063314, "grad_norm": 0.586625874042511, "learning_rate": 2.473126094349331e-06, "loss": 0.3829, "step": 5450 }, { "epoch": 2.3516743835468934, "grad_norm": 0.5470960736274719, "learning_rate": 2.4419387334221333e-06, "loss": 0.3881, "step": 5460 }, { "epoch": 2.3559814794874554, "grad_norm": 0.5486071705818176, "learning_rate": 2.4109218947061884e-06, "loss": 0.399, "step": 5470 }, { "epoch": 2.360288575428018, "grad_norm": 0.5942230820655823, "learning_rate": 2.3800762779920574e-06, "loss": 0.3921, "step": 5480 }, { "epoch": 2.36459567136858, "grad_norm": 0.5786502957344055, "learning_rate": 2.3494025792072474e-06, "loss": 0.3901, "step": 5490 }, { "epoch": 2.368902767309142, "grad_norm": 0.6082814931869507, "learning_rate": 2.3189014904005247e-06, "loss": 0.391, "step": 5500 }, { "epoch": 2.373209863249704, "grad_norm": 0.612694501876831, "learning_rate": 2.2885736997262863e-06, "loss": 0.3981, "step": 5510 }, { "epoch": 2.377516959190266, "grad_norm": 0.5050374865531921, "learning_rate": 2.2584198914290435e-06, "loss": 0.3951, "step": 5520 }, { "epoch": 2.381824055130828, "grad_norm": 0.5465214848518372, "learning_rate": 2.2284407458279743e-06, "loss": 0.4, "step": 5530 }, { "epoch": 2.3861311510713903, "grad_norm": 0.5544529557228088, "learning_rate": 2.1986369393015914e-06, "loss": 0.3836, "step": 5540 }, { "epoch": 2.3904382470119523, "grad_norm": 0.586337149143219, "learning_rate": 2.169009144272467e-06, "loss": 0.4139, "step": 5550 }, { "epoch": 2.3947453429525143, "grad_norm": 0.6219981908798218, "learning_rate": 2.1395580291920625e-06, "loss": 0.4011, "step": 5560 }, { "epoch": 2.3990524388930763, "grad_norm": 0.6941688060760498, "learning_rate": 2.110284258525658e-06, "loss": 0.405, "step": 5570 }, { "epoch": 2.4033595348336383, "grad_norm": 0.5210332274436951, "learning_rate": 2.081188492737345e-06, "loss": 0.4017, "step": 5580 }, { "epoch": 2.4076666307742007, "grad_norm": 0.5930879712104797, "learning_rate": 2.0522713882751445e-06, "loss": 0.3918, "step": 5590 }, { "epoch": 2.4119737267147627, "grad_norm": 0.5910641551017761, "learning_rate": 2.0235335975561775e-06, "loss": 0.3996, "step": 5600 }, { "epoch": 2.4162808226553247, "grad_norm": 0.5827698111534119, "learning_rate": 1.9949757689519555e-06, "loss": 0.3854, "step": 5610 }, { "epoch": 2.4205879185958867, "grad_norm": 0.5518185496330261, "learning_rate": 1.966598546773757e-06, "loss": 0.4077, "step": 5620 }, { "epoch": 2.4248950145364487, "grad_norm": 0.6005439162254333, "learning_rate": 1.938402571258073e-06, "loss": 0.4095, "step": 5630 }, { "epoch": 2.4292021104770107, "grad_norm": 0.5761522054672241, "learning_rate": 1.9103884785521887e-06, "loss": 0.3966, "step": 5640 }, { "epoch": 2.4335092064175727, "grad_norm": 0.5546764135360718, "learning_rate": 1.8825569006998012e-06, "loss": 0.395, "step": 5650 }, { "epoch": 2.437816302358135, "grad_norm": 0.5639533996582031, "learning_rate": 1.8549084656267846e-06, "loss": 0.3938, "step": 5660 }, { "epoch": 2.442123398298697, "grad_norm": 0.5662581324577332, "learning_rate": 1.8274437971270044e-06, "loss": 0.4004, "step": 5670 }, { "epoch": 2.446430494239259, "grad_norm": 0.5856819748878479, "learning_rate": 1.8001635148482621e-06, "loss": 0.3946, "step": 5680 }, { "epoch": 2.450737590179821, "grad_norm": 0.5766512751579285, "learning_rate": 1.7730682342782967e-06, "loss": 0.3931, "step": 5690 }, { "epoch": 2.455044686120383, "grad_norm": 0.6373909711837769, "learning_rate": 1.7461585667309045e-06, "loss": 0.4006, "step": 5700 }, { "epoch": 2.4593517820609456, "grad_norm": 0.5694748759269714, "learning_rate": 1.719435119332159e-06, "loss": 0.3989, "step": 5710 }, { "epoch": 2.4636588780015076, "grad_norm": 0.5339934229850769, "learning_rate": 1.6928984950066918e-06, "loss": 0.3966, "step": 5720 }, { "epoch": 2.4679659739420696, "grad_norm": 0.5888383388519287, "learning_rate": 1.6665492924641113e-06, "loss": 0.3833, "step": 5730 }, { "epoch": 2.4722730698826316, "grad_norm": 0.5573282241821289, "learning_rate": 1.6403881061854732e-06, "loss": 0.4, "step": 5740 }, { "epoch": 2.4765801658231936, "grad_norm": 0.5756634473800659, "learning_rate": 1.6144155264098883e-06, "loss": 0.3964, "step": 5750 }, { "epoch": 2.4808872617637556, "grad_norm": 0.5784355401992798, "learning_rate": 1.58863213912119e-06, "loss": 0.3762, "step": 5760 }, { "epoch": 2.4851943577043176, "grad_norm": 0.6090006828308105, "learning_rate": 1.563038526034727e-06, "loss": 0.3986, "step": 5770 }, { "epoch": 2.48950145364488, "grad_norm": 0.5565779209136963, "learning_rate": 1.5376352645842242e-06, "loss": 0.3916, "step": 5780 }, { "epoch": 2.493808549585442, "grad_norm": 0.6107103228569031, "learning_rate": 1.5124229279087655e-06, "loss": 0.4093, "step": 5790 }, { "epoch": 2.498115645526004, "grad_norm": 0.5300205945968628, "learning_rate": 1.487402084839864e-06, "loss": 0.4047, "step": 5800 }, { "epoch": 2.502422741466566, "grad_norm": 0.6008495688438416, "learning_rate": 1.4625732998886178e-06, "loss": 0.4023, "step": 5810 }, { "epoch": 2.5067298374071285, "grad_norm": 0.5560673475265503, "learning_rate": 1.437937133232985e-06, "loss": 0.3968, "step": 5820 }, { "epoch": 2.5110369333476905, "grad_norm": 0.5503118634223938, "learning_rate": 1.413494140705136e-06, "loss": 0.3876, "step": 5830 }, { "epoch": 2.5153440292882525, "grad_norm": 0.5559957027435303, "learning_rate": 1.3892448737789243e-06, "loss": 0.392, "step": 5840 }, { "epoch": 2.5196511252288145, "grad_norm": 0.5354902148246765, "learning_rate": 1.365189879557426e-06, "loss": 0.3988, "step": 5850 }, { "epoch": 2.5239582211693765, "grad_norm": 0.577046275138855, "learning_rate": 1.3413297007606196e-06, "loss": 0.3948, "step": 5860 }, { "epoch": 2.5282653171099385, "grad_norm": 0.5745800733566284, "learning_rate": 1.3176648757131205e-06, "loss": 0.395, "step": 5870 }, { "epoch": 2.5325724130505005, "grad_norm": 0.5721185207366943, "learning_rate": 1.2941959383320478e-06, "loss": 0.3918, "step": 5880 }, { "epoch": 2.5368795089910625, "grad_norm": 0.5935482978820801, "learning_rate": 1.2709234181149765e-06, "loss": 0.376, "step": 5890 }, { "epoch": 2.541186604931625, "grad_norm": 0.5709375143051147, "learning_rate": 1.2478478401279848e-06, "loss": 0.3881, "step": 5900 }, { "epoch": 2.545493700872187, "grad_norm": 0.5233684182167053, "learning_rate": 1.2249697249938197e-06, "loss": 0.3945, "step": 5910 }, { "epoch": 2.549800796812749, "grad_norm": 0.5812388062477112, "learning_rate": 1.2022895888801333e-06, "loss": 0.3984, "step": 5920 }, { "epoch": 2.554107892753311, "grad_norm": 0.560550332069397, "learning_rate": 1.1798079434878584e-06, "loss": 0.3942, "step": 5930 }, { "epoch": 2.5584149886938734, "grad_norm": 0.6010858416557312, "learning_rate": 1.1575252960396422e-06, "loss": 0.3851, "step": 5940 }, { "epoch": 2.5627220846344354, "grad_norm": 0.5857875347137451, "learning_rate": 1.1354421492684252e-06, "loss": 0.3993, "step": 5950 }, { "epoch": 2.5670291805749974, "grad_norm": 0.604179859161377, "learning_rate": 1.1135590014060772e-06, "loss": 0.388, "step": 5960 }, { "epoch": 2.5713362765155594, "grad_norm": 0.569106936454773, "learning_rate": 1.0918763461721648e-06, "loss": 0.4014, "step": 5970 }, { "epoch": 2.5756433724561214, "grad_norm": 0.5742547512054443, "learning_rate": 1.0703946727628234e-06, "loss": 0.3839, "step": 5980 }, { "epoch": 2.5799504683966834, "grad_norm": 0.5561407208442688, "learning_rate": 1.0491144658397e-06, "loss": 0.3853, "step": 5990 }, { "epoch": 2.5842575643372454, "grad_norm": 0.5482295155525208, "learning_rate": 1.0280362055190341e-06, "loss": 0.3876, "step": 6000 }, { "epoch": 2.588564660277808, "grad_norm": 0.5737982392311096, "learning_rate": 1.0071603673608176e-06, "loss": 0.4059, "step": 6010 }, { "epoch": 2.59287175621837, "grad_norm": 0.547715961933136, "learning_rate": 9.864874223580668e-07, "loss": 0.3837, "step": 6020 }, { "epoch": 2.597178852158932, "grad_norm": 0.607851505279541, "learning_rate": 9.66017836926203e-07, "loss": 0.3779, "step": 6030 }, { "epoch": 2.601485948099494, "grad_norm": 0.5557613968849182, "learning_rate": 9.457520728925151e-07, "loss": 0.3995, "step": 6040 }, { "epoch": 2.605793044040056, "grad_norm": 0.5470052361488342, "learning_rate": 9.256905874857535e-07, "loss": 0.3916, "step": 6050 }, { "epoch": 2.6101001399806183, "grad_norm": 0.5718830227851868, "learning_rate": 9.058338333258032e-07, "loss": 0.3997, "step": 6060 }, { "epoch": 2.6144072359211803, "grad_norm": 0.5838637948036194, "learning_rate": 8.861822584134882e-07, "loss": 0.39, "step": 6070 }, { "epoch": 2.6187143318617423, "grad_norm": 0.5819488763809204, "learning_rate": 8.667363061204415e-07, "loss": 0.4028, "step": 6080 }, { "epoch": 2.6230214278023043, "grad_norm": 0.5477743744850159, "learning_rate": 8.474964151791232e-07, "loss": 0.3979, "step": 6090 }, { "epoch": 2.6273285237428663, "grad_norm": 0.6217262744903564, "learning_rate": 8.284630196729059e-07, "loss": 0.3993, "step": 6100 }, { "epoch": 2.6316356196834283, "grad_norm": 0.5514227747917175, "learning_rate": 8.096365490262925e-07, "loss": 0.4058, "step": 6110 }, { "epoch": 2.6359427156239903, "grad_norm": 0.645946204662323, "learning_rate": 7.910174279952232e-07, "loss": 0.3992, "step": 6120 }, { "epoch": 2.6402498115645527, "grad_norm": 0.5741420984268188, "learning_rate": 7.726060766574883e-07, "loss": 0.3938, "step": 6130 }, { "epoch": 2.6445569075051147, "grad_norm": 0.5910946726799011, "learning_rate": 7.544029104032558e-07, "loss": 0.3898, "step": 6140 }, { "epoch": 2.6488640034456767, "grad_norm": 0.5803595185279846, "learning_rate": 7.364083399256971e-07, "loss": 0.388, "step": 6150 }, { "epoch": 2.6531710993862387, "grad_norm": 0.596809446811676, "learning_rate": 7.186227712117266e-07, "loss": 0.388, "step": 6160 }, { "epoch": 2.6574781953268007, "grad_norm": 0.6213387250900269, "learning_rate": 7.010466055328313e-07, "loss": 0.3839, "step": 6170 }, { "epoch": 2.661785291267363, "grad_norm": 0.5913180112838745, "learning_rate": 6.836802394360276e-07, "loss": 0.3989, "step": 6180 }, { "epoch": 2.666092387207925, "grad_norm": 0.6089721322059631, "learning_rate": 6.665240647349125e-07, "loss": 0.4039, "step": 6190 }, { "epoch": 2.670399483148487, "grad_norm": 0.5730729103088379, "learning_rate": 6.495784685008133e-07, "loss": 0.3951, "step": 6200 }, { "epoch": 2.674706579089049, "grad_norm": 0.5562758445739746, "learning_rate": 6.32843833054072e-07, "loss": 0.3837, "step": 6210 }, { "epoch": 2.679013675029611, "grad_norm": 0.5627213716506958, "learning_rate": 6.16320535955407e-07, "loss": 0.3712, "step": 6220 }, { "epoch": 2.683320770970173, "grad_norm": 0.559660017490387, "learning_rate": 6.000089499973971e-07, "loss": 0.3901, "step": 6230 }, { "epoch": 2.687627866910735, "grad_norm": 0.6018761992454529, "learning_rate": 5.839094431960713e-07, "loss": 0.383, "step": 6240 }, { "epoch": 2.6919349628512976, "grad_norm": 0.5534284710884094, "learning_rate": 5.680223787826089e-07, "loss": 0.3925, "step": 6250 }, { "epoch": 2.6962420587918596, "grad_norm": 0.5682888031005859, "learning_rate": 5.523481151951427e-07, "loss": 0.3929, "step": 6260 }, { "epoch": 2.7005491547324216, "grad_norm": 0.6271238923072815, "learning_rate": 5.368870060706677e-07, "loss": 0.3942, "step": 6270 }, { "epoch": 2.7048562506729836, "grad_norm": 0.5881267786026001, "learning_rate": 5.216394002370695e-07, "loss": 0.3876, "step": 6280 }, { "epoch": 2.709163346613546, "grad_norm": 0.6085900068283081, "learning_rate": 5.066056417052445e-07, "loss": 0.3958, "step": 6290 }, { "epoch": 2.713470442554108, "grad_norm": 0.5912172198295593, "learning_rate": 4.917860696613541e-07, "loss": 0.3887, "step": 6300 }, { "epoch": 2.71777753849467, "grad_norm": 0.6698789596557617, "learning_rate": 4.771810184591541e-07, "loss": 0.3899, "step": 6310 }, { "epoch": 2.722084634435232, "grad_norm": 0.5682712197303772, "learning_rate": 4.627908176124618e-07, "loss": 0.3826, "step": 6320 }, { "epoch": 2.726391730375794, "grad_norm": 0.5702280402183533, "learning_rate": 4.486157917877232e-07, "loss": 0.3908, "step": 6330 }, { "epoch": 2.730698826316356, "grad_norm": 0.5540564060211182, "learning_rate": 4.346562607966787e-07, "loss": 0.3962, "step": 6340 }, { "epoch": 2.735005922256918, "grad_norm": 0.6031074523925781, "learning_rate": 4.209125395891589e-07, "loss": 0.3791, "step": 6350 }, { "epoch": 2.73931301819748, "grad_norm": 0.5727553963661194, "learning_rate": 4.0738493824596715e-07, "loss": 0.4023, "step": 6360 }, { "epoch": 2.7436201141380425, "grad_norm": 0.5374717116355896, "learning_rate": 3.940737619718937e-07, "loss": 0.38, "step": 6370 }, { "epoch": 2.7479272100786045, "grad_norm": 0.5720168352127075, "learning_rate": 3.809793110888249e-07, "loss": 0.4011, "step": 6380 }, { "epoch": 2.7522343060191665, "grad_norm": 0.5751203894615173, "learning_rate": 3.6810188102896605e-07, "loss": 0.3941, "step": 6390 }, { "epoch": 2.7565414019597285, "grad_norm": 0.5838513970375061, "learning_rate": 3.554417623281825e-07, "loss": 0.3834, "step": 6400 }, { "epoch": 2.760848497900291, "grad_norm": 0.6204310059547424, "learning_rate": 3.429992406194338e-07, "loss": 0.3933, "step": 6410 }, { "epoch": 2.765155593840853, "grad_norm": 0.6237754225730896, "learning_rate": 3.3077459662634205e-07, "loss": 0.3911, "step": 6420 }, { "epoch": 2.769462689781415, "grad_norm": 0.561553418636322, "learning_rate": 3.1876810615684705e-07, "loss": 0.3847, "step": 6430 }, { "epoch": 2.773769785721977, "grad_norm": 0.568580150604248, "learning_rate": 3.069800400969947e-07, "loss": 0.3967, "step": 6440 }, { "epoch": 2.778076881662539, "grad_norm": 0.6103531122207642, "learning_rate": 2.954106644048127e-07, "loss": 0.3731, "step": 6450 }, { "epoch": 2.782383977603101, "grad_norm": 0.560199499130249, "learning_rate": 2.840602401043213e-07, "loss": 0.3889, "step": 6460 }, { "epoch": 2.786691073543663, "grad_norm": 0.5612174868583679, "learning_rate": 2.7292902327963776e-07, "loss": 0.3915, "step": 6470 }, { "epoch": 2.7909981694842254, "grad_norm": 0.5860500335693359, "learning_rate": 2.620172650692021e-07, "loss": 0.4063, "step": 6480 }, { "epoch": 2.7953052654247874, "grad_norm": 0.6044652462005615, "learning_rate": 2.513252116601062e-07, "loss": 0.39, "step": 6490 }, { "epoch": 2.7996123613653494, "grad_norm": 0.5966377258300781, "learning_rate": 2.408531042825446e-07, "loss": 0.3965, "step": 6500 }, { "epoch": 2.8039194573059114, "grad_norm": 0.5729289650917053, "learning_rate": 2.3060117920437164e-07, "loss": 0.3798, "step": 6510 }, { "epoch": 2.8082265532464734, "grad_norm": 0.6403810977935791, "learning_rate": 2.2056966772576626e-07, "loss": 0.4096, "step": 6520 }, { "epoch": 2.812533649187036, "grad_norm": 0.5852852463722229, "learning_rate": 2.1075879617401984e-07, "loss": 0.383, "step": 6530 }, { "epoch": 2.816840745127598, "grad_norm": 0.6858223080635071, "learning_rate": 2.0116878589842236e-07, "loss": 0.3763, "step": 6540 }, { "epoch": 2.82114784106816, "grad_norm": 0.5583459138870239, "learning_rate": 1.917998532652765e-07, "loss": 0.4007, "step": 6550 }, { "epoch": 2.825454937008722, "grad_norm": 0.6212313175201416, "learning_rate": 1.8265220965300812e-07, "loss": 0.3946, "step": 6560 }, { "epoch": 2.829762032949284, "grad_norm": 0.5777102112770081, "learning_rate": 1.7372606144740567e-07, "loss": 0.3908, "step": 6570 }, { "epoch": 2.834069128889846, "grad_norm": 0.5885289311408997, "learning_rate": 1.6502161003695615e-07, "loss": 0.4051, "step": 6580 }, { "epoch": 2.838376224830408, "grad_norm": 0.6133362054824829, "learning_rate": 1.5653905180830432e-07, "loss": 0.3909, "step": 6590 }, { "epoch": 2.8426833207709703, "grad_norm": 0.5662548542022705, "learning_rate": 1.48278578141825e-07, "loss": 0.3689, "step": 6600 }, { "epoch": 2.8469904167115323, "grad_norm": 0.5703479647636414, "learning_rate": 1.4024037540730006e-07, "loss": 0.3812, "step": 6610 }, { "epoch": 2.8512975126520943, "grad_norm": 0.5604844689369202, "learning_rate": 1.324246249597183e-07, "loss": 0.3992, "step": 6620 }, { "epoch": 2.8556046085926563, "grad_norm": 0.6033147573471069, "learning_rate": 1.2483150313517766e-07, "loss": 0.3937, "step": 6630 }, { "epoch": 2.8599117045332187, "grad_norm": 0.5846080780029297, "learning_rate": 1.1746118124691508e-07, "loss": 0.4123, "step": 6640 }, { "epoch": 2.8642188004737807, "grad_norm": 0.63025963306427, "learning_rate": 1.103138255814329e-07, "loss": 0.3998, "step": 6650 }, { "epoch": 2.8685258964143427, "grad_norm": 0.5580465197563171, "learning_rate": 1.0338959739475296e-07, "loss": 0.4007, "step": 6660 }, { "epoch": 2.8728329923549047, "grad_norm": 0.5767059326171875, "learning_rate": 9.66886529087785e-08, "loss": 0.4008, "step": 6670 }, { "epoch": 2.8771400882954667, "grad_norm": 0.583044707775116, "learning_rate": 9.021114330776348e-08, "loss": 0.403, "step": 6680 }, { "epoch": 2.8814471842360287, "grad_norm": 0.5440847873687744, "learning_rate": 8.395721473490992e-08, "loss": 0.3839, "step": 6690 }, { "epoch": 2.8857542801765907, "grad_norm": 0.55162513256073, "learning_rate": 7.792700828906374e-08, "loss": 0.4017, "step": 6700 }, { "epoch": 2.8900613761171527, "grad_norm": 0.5817933082580566, "learning_rate": 7.212066002153518e-08, "loss": 0.4009, "step": 6710 }, { "epoch": 2.894368472057715, "grad_norm": 0.6080750226974487, "learning_rate": 6.653830093302782e-08, "loss": 0.3964, "step": 6720 }, { "epoch": 2.898675567998277, "grad_norm": 0.5681482553482056, "learning_rate": 6.11800569706833e-08, "loss": 0.4003, "step": 6730 }, { "epoch": 2.902982663938839, "grad_norm": 0.5769705176353455, "learning_rate": 5.604604902524235e-08, "loss": 0.4017, "step": 6740 }, { "epoch": 2.907289759879401, "grad_norm": 0.546116828918457, "learning_rate": 5.113639292831152e-08, "loss": 0.3828, "step": 6750 }, { "epoch": 2.9115968558199636, "grad_norm": 0.590798020362854, "learning_rate": 4.645119944975296e-08, "loss": 0.3853, "step": 6760 }, { "epoch": 2.9159039517605256, "grad_norm": 0.5748469233512878, "learning_rate": 4.1990574295187606e-08, "loss": 0.4107, "step": 6770 }, { "epoch": 2.9202110477010876, "grad_norm": 0.5733410716056824, "learning_rate": 3.7754618103608144e-08, "loss": 0.4052, "step": 6780 }, { "epoch": 2.9245181436416496, "grad_norm": 0.5576743483543396, "learning_rate": 3.374342644510531e-08, "loss": 0.3846, "step": 6790 }, { "epoch": 2.9288252395822116, "grad_norm": 0.596834123134613, "learning_rate": 2.9957089818718476e-08, "loss": 0.4029, "step": 6800 }, { "epoch": 2.9331323355227736, "grad_norm": 0.5680873990058899, "learning_rate": 2.639569365038841e-08, "loss": 0.381, "step": 6810 }, { "epoch": 2.9374394314633356, "grad_norm": 0.5597060918807983, "learning_rate": 2.305931829102992e-08, "loss": 0.3974, "step": 6820 }, { "epoch": 2.941746527403898, "grad_norm": 0.5827191472053528, "learning_rate": 1.9948039014724417e-08, "loss": 0.3973, "step": 6830 }, { "epoch": 2.94605362334446, "grad_norm": 0.6119829416275024, "learning_rate": 1.706192601701462e-08, "loss": 0.3984, "step": 6840 }, { "epoch": 2.950360719285022, "grad_norm": 0.602497935295105, "learning_rate": 1.4401044413324682e-08, "loss": 0.4086, "step": 6850 }, { "epoch": 2.954667815225584, "grad_norm": 0.5783790349960327, "learning_rate": 1.1965454237493623e-08, "loss": 0.3945, "step": 6860 }, { "epoch": 2.958974911166146, "grad_norm": 0.5653091073036194, "learning_rate": 9.755210440413055e-09, "loss": 0.3938, "step": 6870 }, { "epoch": 2.9632820071067085, "grad_norm": 0.5716846585273743, "learning_rate": 7.770362888795957e-09, "loss": 0.3935, "step": 6880 }, { "epoch": 2.9675891030472705, "grad_norm": 0.6015262603759766, "learning_rate": 6.0109563640442515e-09, "loss": 0.3955, "step": 6890 }, { "epoch": 2.9718961989878325, "grad_norm": 0.5763514041900635, "learning_rate": 4.477030561246265e-09, "loss": 0.4069, "step": 6900 }, { "epoch": 2.9762032949283945, "grad_norm": 0.5644577741622925, "learning_rate": 3.168620088271901e-09, "loss": 0.3921, "step": 6910 }, { "epoch": 2.9805103908689565, "grad_norm": 0.5302848219871521, "learning_rate": 2.0857544650010332e-09, "loss": 0.404, "step": 6920 }, { "epoch": 2.9848174868095185, "grad_norm": 0.6025976538658142, "learning_rate": 1.2284581226507108e-09, "loss": 0.4037, "step": 6930 }, { "epoch": 2.9891245827500805, "grad_norm": 0.5681896805763245, "learning_rate": 5.967504032267091e-10, "loss": 0.4031, "step": 6940 }, { "epoch": 2.993431678690643, "grad_norm": 0.5708478093147278, "learning_rate": 1.906455590883205e-10, "loss": 0.4206, "step": 6950 }, { "epoch": 2.997738774631205, "grad_norm": 0.5966918468475342, "learning_rate": 1.015275262306048e-11, "loss": 0.4014, "step": 6960 } ], "logging_steps": 10, "max_steps": 6963, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 300, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 7523782707118080.0, "train_batch_size": 16, "trial_name": null, "trial_params": null }