{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 2179, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0004589261128958238, "grad_norm": 8.542048983401715, "learning_rate": 9.174311926605506e-08, "loss": 1.1206, "step": 1 }, { "epoch": 0.002294630564479119, "grad_norm": 8.730820381996422, "learning_rate": 4.587155963302753e-07, "loss": 1.1225, "step": 5 }, { "epoch": 0.004589261128958238, "grad_norm": 5.026231881904144, "learning_rate": 9.174311926605506e-07, "loss": 1.1138, "step": 10 }, { "epoch": 0.006883891693437357, "grad_norm": 3.214436315778165, "learning_rate": 1.3761467889908258e-06, "loss": 1.0308, "step": 15 }, { "epoch": 0.009178522257916476, "grad_norm": 2.9262839132141174, "learning_rate": 1.8348623853211011e-06, "loss": 0.9803, "step": 20 }, { "epoch": 0.011473152822395595, "grad_norm": 2.5816500702829965, "learning_rate": 2.2935779816513764e-06, "loss": 0.9955, "step": 25 }, { "epoch": 0.013767783386874713, "grad_norm": 2.62089651849134, "learning_rate": 2.7522935779816517e-06, "loss": 1.0214, "step": 30 }, { "epoch": 0.016062413951353834, "grad_norm": 2.522275443963982, "learning_rate": 3.211009174311927e-06, "loss": 0.9751, "step": 35 }, { "epoch": 0.018357044515832952, "grad_norm": 2.415617593047094, "learning_rate": 3.6697247706422022e-06, "loss": 0.995, "step": 40 }, { "epoch": 0.02065167508031207, "grad_norm": 2.525030642664337, "learning_rate": 4.128440366972478e-06, "loss": 1.0113, "step": 45 }, { "epoch": 0.02294630564479119, "grad_norm": 2.6665037719940954, "learning_rate": 4.587155963302753e-06, "loss": 0.9655, "step": 50 }, { "epoch": 0.025240936209270308, "grad_norm": 3.3232412426313385, "learning_rate": 5.045871559633028e-06, "loss": 0.9874, "step": 55 }, { "epoch": 0.027535566773749427, "grad_norm": 2.394624410644964, "learning_rate": 5.504587155963303e-06, "loss": 0.9921, "step": 60 }, { "epoch": 0.029830197338228545, "grad_norm": 2.6803374112332796, "learning_rate": 5.963302752293578e-06, "loss": 0.9968, "step": 65 }, { "epoch": 0.03212482790270767, "grad_norm": 2.695577014057411, "learning_rate": 6.422018348623854e-06, "loss": 0.9819, "step": 70 }, { "epoch": 0.03441945846718678, "grad_norm": 2.5003353553530308, "learning_rate": 6.880733944954129e-06, "loss": 1.0135, "step": 75 }, { "epoch": 0.036714089031665904, "grad_norm": 2.449197716174392, "learning_rate": 7.3394495412844045e-06, "loss": 0.9861, "step": 80 }, { "epoch": 0.03900871959614502, "grad_norm": 2.547680048482618, "learning_rate": 7.79816513761468e-06, "loss": 0.9886, "step": 85 }, { "epoch": 0.04130335016062414, "grad_norm": 2.738205329791908, "learning_rate": 8.256880733944956e-06, "loss": 0.9849, "step": 90 }, { "epoch": 0.04359798072510326, "grad_norm": 2.6040796770347256, "learning_rate": 8.71559633027523e-06, "loss": 1.0033, "step": 95 }, { "epoch": 0.04589261128958238, "grad_norm": 2.6163342737972495, "learning_rate": 9.174311926605506e-06, "loss": 0.9906, "step": 100 }, { "epoch": 0.048187241854061494, "grad_norm": 2.4805649474435536, "learning_rate": 9.633027522935781e-06, "loss": 0.9953, "step": 105 }, { "epoch": 0.050481872418540616, "grad_norm": 2.5415673257044604, "learning_rate": 1.0091743119266055e-05, "loss": 1.0171, "step": 110 }, { "epoch": 0.05277650298301973, "grad_norm": 2.4798532274771303, "learning_rate": 1.055045871559633e-05, "loss": 1.0314, "step": 115 }, { "epoch": 0.05507113354749885, "grad_norm": 2.8712967477215217, "learning_rate": 1.1009174311926607e-05, "loss": 0.9966, "step": 120 }, { "epoch": 0.05736576411197797, "grad_norm": 2.40307484944342, "learning_rate": 1.1467889908256882e-05, "loss": 0.9862, "step": 125 }, { "epoch": 0.05966039467645709, "grad_norm": 2.531947293972582, "learning_rate": 1.1926605504587156e-05, "loss": 1.0111, "step": 130 }, { "epoch": 0.06195502524093621, "grad_norm": 2.5630237672720355, "learning_rate": 1.238532110091743e-05, "loss": 1.0111, "step": 135 }, { "epoch": 0.06424965580541533, "grad_norm": 3.0796021235291504, "learning_rate": 1.2844036697247708e-05, "loss": 0.996, "step": 140 }, { "epoch": 0.06654428636989444, "grad_norm": 2.532220737324273, "learning_rate": 1.3302752293577984e-05, "loss": 1.0361, "step": 145 }, { "epoch": 0.06883891693437356, "grad_norm": 2.656655575725622, "learning_rate": 1.3761467889908258e-05, "loss": 1.0061, "step": 150 }, { "epoch": 0.07113354749885269, "grad_norm": 2.4257637087990194, "learning_rate": 1.4220183486238533e-05, "loss": 0.9896, "step": 155 }, { "epoch": 0.07342817806333181, "grad_norm": 2.482384770473286, "learning_rate": 1.4678899082568809e-05, "loss": 1.0321, "step": 160 }, { "epoch": 0.07572280862781092, "grad_norm": 2.9466970045401957, "learning_rate": 1.5137614678899085e-05, "loss": 1.0113, "step": 165 }, { "epoch": 0.07801743919229004, "grad_norm": 2.5190350812742364, "learning_rate": 1.559633027522936e-05, "loss": 1.0271, "step": 170 }, { "epoch": 0.08031206975676916, "grad_norm": 2.478301477100018, "learning_rate": 1.6055045871559634e-05, "loss": 1.0165, "step": 175 }, { "epoch": 0.08260670032124828, "grad_norm": 2.4464230838695826, "learning_rate": 1.6513761467889912e-05, "loss": 1.0646, "step": 180 }, { "epoch": 0.08490133088572739, "grad_norm": 2.5240574285299315, "learning_rate": 1.6972477064220186e-05, "loss": 1.0555, "step": 185 }, { "epoch": 0.08719596145020651, "grad_norm": 2.4185989752236505, "learning_rate": 1.743119266055046e-05, "loss": 1.0451, "step": 190 }, { "epoch": 0.08949059201468564, "grad_norm": 2.6482287747026954, "learning_rate": 1.7889908256880734e-05, "loss": 1.0325, "step": 195 }, { "epoch": 0.09178522257916476, "grad_norm": 2.4286821504213814, "learning_rate": 1.834862385321101e-05, "loss": 1.0269, "step": 200 }, { "epoch": 0.09407985314364388, "grad_norm": 2.193912154365014, "learning_rate": 1.8807339449541285e-05, "loss": 1.0551, "step": 205 }, { "epoch": 0.09637448370812299, "grad_norm": 2.695401581193139, "learning_rate": 1.9266055045871563e-05, "loss": 1.0456, "step": 210 }, { "epoch": 0.09866911427260211, "grad_norm": 2.6583940178739827, "learning_rate": 1.9724770642201837e-05, "loss": 1.0664, "step": 215 }, { "epoch": 0.10096374483708123, "grad_norm": 2.371077606257049, "learning_rate": 1.9999948669655127e-05, "loss": 1.0578, "step": 220 }, { "epoch": 0.10325837540156035, "grad_norm": 2.159798506956808, "learning_rate": 1.999937120932709e-05, "loss": 1.0421, "step": 225 }, { "epoch": 0.10555300596603946, "grad_norm": 2.7274671704918516, "learning_rate": 1.9998152162914807e-05, "loss": 1.0513, "step": 230 }, { "epoch": 0.10784763653051858, "grad_norm": 2.171023592361791, "learning_rate": 1.9996291608635527e-05, "loss": 1.0566, "step": 235 }, { "epoch": 0.1101422670949977, "grad_norm": 2.1701858629565445, "learning_rate": 1.9993789665867316e-05, "loss": 1.0371, "step": 240 }, { "epoch": 0.11243689765947683, "grad_norm": 2.2165028686751254, "learning_rate": 1.9990646495141445e-05, "loss": 1.0491, "step": 245 }, { "epoch": 0.11473152822395594, "grad_norm": 2.6072713574574466, "learning_rate": 1.998686229813205e-05, "loss": 1.0636, "step": 250 }, { "epoch": 0.11702615878843506, "grad_norm": 2.135919480089347, "learning_rate": 1.9982437317643218e-05, "loss": 1.05, "step": 255 }, { "epoch": 0.11932078935291418, "grad_norm": 2.4909036218466993, "learning_rate": 1.9977371837593382e-05, "loss": 1.0611, "step": 260 }, { "epoch": 0.1216154199173933, "grad_norm": 5.355161964355881, "learning_rate": 1.997166618299714e-05, "loss": 1.0253, "step": 265 }, { "epoch": 0.12391005048187242, "grad_norm": 2.374369183096664, "learning_rate": 1.9965320719944366e-05, "loss": 1.044, "step": 270 }, { "epoch": 0.12620468104635155, "grad_norm": 2.5178752868507144, "learning_rate": 1.995833585557674e-05, "loss": 1.0692, "step": 275 }, { "epoch": 0.12849931161083067, "grad_norm": 2.560582169592424, "learning_rate": 1.9950712038061617e-05, "loss": 1.0518, "step": 280 }, { "epoch": 0.13079394217530976, "grad_norm": 2.279796271646809, "learning_rate": 1.994244975656328e-05, "loss": 1.0627, "step": 285 }, { "epoch": 0.13308857273978889, "grad_norm": 2.428851471685207, "learning_rate": 1.993354954121155e-05, "loss": 1.0672, "step": 290 }, { "epoch": 0.135383203304268, "grad_norm": 2.253932739513235, "learning_rate": 1.9924011963067765e-05, "loss": 1.0572, "step": 295 }, { "epoch": 0.13767783386874713, "grad_norm": 2.2863725772380232, "learning_rate": 1.9913837634088143e-05, "loss": 1.0612, "step": 300 }, { "epoch": 0.13997246443322625, "grad_norm": 2.3216601188965433, "learning_rate": 1.9903027207084525e-05, "loss": 1.0711, "step": 305 }, { "epoch": 0.14226709499770537, "grad_norm": 2.4619963294994305, "learning_rate": 1.9891581375682472e-05, "loss": 1.091, "step": 310 }, { "epoch": 0.1445617255621845, "grad_norm": 2.6730364673716216, "learning_rate": 1.9879500874276788e-05, "loss": 1.0739, "step": 315 }, { "epoch": 0.14685635612666362, "grad_norm": 2.240223628972246, "learning_rate": 1.9866786477984357e-05, "loss": 1.0675, "step": 320 }, { "epoch": 0.14915098669114274, "grad_norm": 2.3152538012542436, "learning_rate": 1.985343900259446e-05, "loss": 1.0723, "step": 325 }, { "epoch": 0.15144561725562183, "grad_norm": 2.4361030726660364, "learning_rate": 1.983945930451639e-05, "loss": 1.0895, "step": 330 }, { "epoch": 0.15374024782010096, "grad_norm": 2.1416955206141086, "learning_rate": 1.982484828072452e-05, "loss": 1.0606, "step": 335 }, { "epoch": 0.15603487838458008, "grad_norm": 2.1520402841348054, "learning_rate": 1.9809606868700755e-05, "loss": 1.0751, "step": 340 }, { "epoch": 0.1583295089490592, "grad_norm": 2.198043038406304, "learning_rate": 1.9793736046374375e-05, "loss": 1.0642, "step": 345 }, { "epoch": 0.16062413951353832, "grad_norm": 2.310801939427329, "learning_rate": 1.977723683205928e-05, "loss": 1.0922, "step": 350 }, { "epoch": 0.16291877007801744, "grad_norm": 2.108829610277771, "learning_rate": 1.9760110284388667e-05, "loss": 1.0738, "step": 355 }, { "epoch": 0.16521340064249657, "grad_norm": 2.2185687059059602, "learning_rate": 1.9742357502247104e-05, "loss": 1.0636, "step": 360 }, { "epoch": 0.1675080312069757, "grad_norm": 2.3820797912728433, "learning_rate": 1.9723979624700004e-05, "loss": 1.0646, "step": 365 }, { "epoch": 0.16980266177145478, "grad_norm": 2.39166890987706, "learning_rate": 1.970497783092057e-05, "loss": 1.0826, "step": 370 }, { "epoch": 0.1720972923359339, "grad_norm": 2.067234800314795, "learning_rate": 1.9685353340114104e-05, "loss": 1.0675, "step": 375 }, { "epoch": 0.17439192290041303, "grad_norm": 2.16248649556658, "learning_rate": 1.9665107411439805e-05, "loss": 1.0955, "step": 380 }, { "epoch": 0.17668655346489215, "grad_norm": 2.290019980298114, "learning_rate": 1.9644241343929966e-05, "loss": 1.0591, "step": 385 }, { "epoch": 0.17898118402937127, "grad_norm": 2.2129439226232503, "learning_rate": 1.9622756476406625e-05, "loss": 1.0631, "step": 390 }, { "epoch": 0.1812758145938504, "grad_norm": 2.027387747335278, "learning_rate": 1.9600654187395666e-05, "loss": 1.0497, "step": 395 }, { "epoch": 0.18357044515832951, "grad_norm": 2.4021963976761995, "learning_rate": 1.9577935895038363e-05, "loss": 1.0738, "step": 400 }, { "epoch": 0.18586507572280864, "grad_norm": 2.066698083420903, "learning_rate": 1.9554603057000397e-05, "loss": 1.0708, "step": 405 }, { "epoch": 0.18815970628728776, "grad_norm": 2.2419541719123823, "learning_rate": 1.953065717037832e-05, "loss": 1.058, "step": 410 }, { "epoch": 0.19045433685176685, "grad_norm": 2.2301364115878823, "learning_rate": 1.9506099771603515e-05, "loss": 1.0654, "step": 415 }, { "epoch": 0.19274896741624598, "grad_norm": 2.259635192704883, "learning_rate": 1.9480932436343584e-05, "loss": 1.0608, "step": 420 }, { "epoch": 0.1950435979807251, "grad_norm": 2.1694738965454743, "learning_rate": 1.945515677940127e-05, "loss": 1.0795, "step": 425 }, { "epoch": 0.19733822854520422, "grad_norm": 24.332124680220836, "learning_rate": 1.9428774454610845e-05, "loss": 1.09, "step": 430 }, { "epoch": 0.19963285910968334, "grad_norm": 13.049567795778453, "learning_rate": 1.9401787154731993e-05, "loss": 1.6953, "step": 435 }, { "epoch": 0.20192748967416246, "grad_norm": 7.723175796818905, "learning_rate": 1.9374196611341212e-05, "loss": 1.1141, "step": 440 }, { "epoch": 0.20422212023864159, "grad_norm": 87.6420124445034, "learning_rate": 1.934600459472067e-05, "loss": 1.1117, "step": 445 }, { "epoch": 0.2065167508031207, "grad_norm": 2.9869822864961053, "learning_rate": 1.931721291374467e-05, "loss": 1.0953, "step": 450 }, { "epoch": 0.20881138136759983, "grad_norm": 2.4448207284406775, "learning_rate": 1.9287823415763552e-05, "loss": 1.0806, "step": 455 }, { "epoch": 0.21110601193207892, "grad_norm": 2.566404692048504, "learning_rate": 1.9257837986485187e-05, "loss": 1.0969, "step": 460 }, { "epoch": 0.21340064249655805, "grad_norm": 4.385856228808105, "learning_rate": 1.922725854985396e-05, "loss": 1.055, "step": 465 }, { "epoch": 0.21569527306103717, "grad_norm": 2.3775711238334507, "learning_rate": 1.9196087067927348e-05, "loss": 1.0698, "step": 470 }, { "epoch": 0.2179899036255163, "grad_norm": 2.2620590555710915, "learning_rate": 1.916432554075002e-05, "loss": 1.0825, "step": 475 }, { "epoch": 0.2202845341899954, "grad_norm": 2.7304207516254997, "learning_rate": 1.913197600622549e-05, "loss": 1.035, "step": 480 }, { "epoch": 0.22257916475447453, "grad_norm": 2.3652995101192382, "learning_rate": 1.9099040539985395e-05, "loss": 1.0598, "step": 485 }, { "epoch": 0.22487379531895366, "grad_norm": 2.3400675787952627, "learning_rate": 1.90655212552563e-05, "loss": 1.0507, "step": 490 }, { "epoch": 0.22716842588343278, "grad_norm": 2.1219643393114493, "learning_rate": 1.9031420302724093e-05, "loss": 1.0549, "step": 495 }, { "epoch": 0.22946305644791187, "grad_norm": 2.0597546067330454, "learning_rate": 1.8996739870396027e-05, "loss": 1.0545, "step": 500 }, { "epoch": 0.231757687012391, "grad_norm": 2.0948012744577467, "learning_rate": 1.896148218346028e-05, "loss": 1.052, "step": 505 }, { "epoch": 0.23405231757687012, "grad_norm": 2.1526817626887227, "learning_rate": 1.8925649504143244e-05, "loss": 1.0462, "step": 510 }, { "epoch": 0.23634694814134924, "grad_norm": 2.038989673045541, "learning_rate": 1.888924413156432e-05, "loss": 1.0651, "step": 515 }, { "epoch": 0.23864157870582836, "grad_norm": 2.1006533572356183, "learning_rate": 1.885226840158843e-05, "loss": 1.0644, "step": 520 }, { "epoch": 0.24093620927030748, "grad_norm": 2.3332747317371596, "learning_rate": 1.8814724686676133e-05, "loss": 1.0722, "step": 525 }, { "epoch": 0.2432308398347866, "grad_norm": 1.9851117275060848, "learning_rate": 1.8776615395731398e-05, "loss": 1.0596, "step": 530 }, { "epoch": 0.24552547039926573, "grad_norm": 2.030081193927863, "learning_rate": 1.8737942973947062e-05, "loss": 1.0759, "step": 535 }, { "epoch": 0.24782010096374485, "grad_norm": 2.2516155399349866, "learning_rate": 1.8698709902647903e-05, "loss": 1.0453, "step": 540 }, { "epoch": 0.25011473152822394, "grad_norm": 2.5068515840745067, "learning_rate": 1.865891869913147e-05, "loss": 1.0908, "step": 545 }, { "epoch": 0.2524093620927031, "grad_norm": 2.4364345139539307, "learning_rate": 1.8618571916506548e-05, "loss": 1.0747, "step": 550 }, { "epoch": 0.2547039926571822, "grad_norm": 2.174192040006664, "learning_rate": 1.8577672143529337e-05, "loss": 1.0695, "step": 555 }, { "epoch": 0.25699862322166134, "grad_norm": 2.1082629086312306, "learning_rate": 1.853622200443737e-05, "loss": 1.0678, "step": 560 }, { "epoch": 0.25929325378614043, "grad_norm": 2.026795919074908, "learning_rate": 1.849422415878112e-05, "loss": 1.0615, "step": 565 }, { "epoch": 0.2615878843506195, "grad_norm": 4.391522020875447, "learning_rate": 1.8451681301253363e-05, "loss": 1.0748, "step": 570 }, { "epoch": 0.2638825149150987, "grad_norm": 2.2545875042843218, "learning_rate": 1.840859616151627e-05, "loss": 1.0841, "step": 575 }, { "epoch": 0.26617714547957777, "grad_norm": 2.4383336966875806, "learning_rate": 1.8364971504026273e-05, "loss": 1.0683, "step": 580 }, { "epoch": 0.2684717760440569, "grad_norm": 2.032234747949604, "learning_rate": 1.8320810127856706e-05, "loss": 1.0503, "step": 585 }, { "epoch": 0.270766406608536, "grad_norm": 2.135201384180716, "learning_rate": 1.827611486651817e-05, "loss": 1.0702, "step": 590 }, { "epoch": 0.27306103717301516, "grad_norm": 3.134318242625564, "learning_rate": 1.8230888587776758e-05, "loss": 1.0808, "step": 595 }, { "epoch": 0.27535566773749426, "grad_norm": 2.098719263576858, "learning_rate": 1.8185134193470043e-05, "loss": 1.0639, "step": 600 }, { "epoch": 0.2776502983019734, "grad_norm": 2.167620655991022, "learning_rate": 1.8138854619320893e-05, "loss": 1.0679, "step": 605 }, { "epoch": 0.2799449288664525, "grad_norm": 2.110572139729922, "learning_rate": 1.8092052834749094e-05, "loss": 1.0481, "step": 610 }, { "epoch": 0.2822395594309316, "grad_norm": 2.111753085065974, "learning_rate": 1.804473184268084e-05, "loss": 1.0929, "step": 615 }, { "epoch": 0.28453418999541075, "grad_norm": 2.047982339354027, "learning_rate": 1.799689467935604e-05, "loss": 1.0675, "step": 620 }, { "epoch": 0.28682882055988984, "grad_norm": 1.9901275601162307, "learning_rate": 1.7948544414133534e-05, "loss": 1.0671, "step": 625 }, { "epoch": 0.289123451124369, "grad_norm": 2.355283701088896, "learning_rate": 1.7899684149294118e-05, "loss": 1.0831, "step": 630 }, { "epoch": 0.2914180816888481, "grad_norm": 2.034042330707897, "learning_rate": 1.7850317019841514e-05, "loss": 1.0551, "step": 635 }, { "epoch": 0.29371271225332723, "grad_norm": 2.0459885339908337, "learning_rate": 1.7800446193301225e-05, "loss": 1.0656, "step": 640 }, { "epoch": 0.29600734281780633, "grad_norm": 1.9809660700340832, "learning_rate": 1.7750074869517285e-05, "loss": 1.0547, "step": 645 }, { "epoch": 0.2983019733822855, "grad_norm": 2.1222169648334037, "learning_rate": 1.7699206280446955e-05, "loss": 1.0856, "step": 650 }, { "epoch": 0.3005966039467646, "grad_norm": 2.033320981208692, "learning_rate": 1.7647843689953352e-05, "loss": 1.0496, "step": 655 }, { "epoch": 0.30289123451124367, "grad_norm": 1.9724823263811897, "learning_rate": 1.759599039359603e-05, "loss": 1.058, "step": 660 }, { "epoch": 0.3051858650757228, "grad_norm": 2.081504542418749, "learning_rate": 1.754364971841952e-05, "loss": 1.0652, "step": 665 }, { "epoch": 0.3074804956402019, "grad_norm": 1.9226506132795518, "learning_rate": 1.749082502273988e-05, "loss": 1.0498, "step": 670 }, { "epoch": 0.30977512620468106, "grad_norm": 2.093061922783108, "learning_rate": 1.7437519695929194e-05, "loss": 1.0469, "step": 675 }, { "epoch": 0.31206975676916016, "grad_norm": 1.9436746042499786, "learning_rate": 1.738373715819811e-05, "loss": 1.0935, "step": 680 }, { "epoch": 0.3143643873336393, "grad_norm": 2.141922458996327, "learning_rate": 1.7329480860376392e-05, "loss": 1.0708, "step": 685 }, { "epoch": 0.3166590178981184, "grad_norm": 1.9342237159056526, "learning_rate": 1.7274754283691507e-05, "loss": 1.0576, "step": 690 }, { "epoch": 0.3189536484625975, "grad_norm": 2.3671090315479724, "learning_rate": 1.7219560939545246e-05, "loss": 1.0314, "step": 695 }, { "epoch": 0.32124827902707664, "grad_norm": 2.804036651572821, "learning_rate": 1.7163904369288443e-05, "loss": 1.037, "step": 700 }, { "epoch": 0.32354290959155574, "grad_norm": 2.130875477845555, "learning_rate": 1.710778814399374e-05, "loss": 1.0676, "step": 705 }, { "epoch": 0.3258375401560349, "grad_norm": 2.2917073480609145, "learning_rate": 1.705121586422647e-05, "loss": 1.0603, "step": 710 }, { "epoch": 0.328132170720514, "grad_norm": 1.9344196026563947, "learning_rate": 1.699419115981361e-05, "loss": 1.0563, "step": 715 }, { "epoch": 0.33042680128499313, "grad_norm": 2.588695773647949, "learning_rate": 1.69367176896109e-05, "loss": 1.0383, "step": 720 }, { "epoch": 0.3327214318494722, "grad_norm": 1.8338278514413, "learning_rate": 1.6878799141268107e-05, "loss": 1.0367, "step": 725 }, { "epoch": 0.3350160624139514, "grad_norm": 1.9232017037694178, "learning_rate": 1.682043923099234e-05, "loss": 1.0609, "step": 730 }, { "epoch": 0.33731069297843047, "grad_norm": 2.0707722390620638, "learning_rate": 1.6761641703309702e-05, "loss": 1.0544, "step": 735 }, { "epoch": 0.33960532354290957, "grad_norm": 1.863894286464851, "learning_rate": 1.6702410330824962e-05, "loss": 1.0459, "step": 740 }, { "epoch": 0.3418999541073887, "grad_norm": 2.295179291530668, "learning_rate": 1.6642748913979515e-05, "loss": 1.0664, "step": 745 }, { "epoch": 0.3441945846718678, "grad_norm": 1.9748901739443767, "learning_rate": 1.6582661280807553e-05, "loss": 1.0824, "step": 750 }, { "epoch": 0.34648921523634696, "grad_norm": 2.0941843827334976, "learning_rate": 1.652215128669042e-05, "loss": 1.0663, "step": 755 }, { "epoch": 0.34878384580082605, "grad_norm": 2.026330695493712, "learning_rate": 1.646122281410927e-05, "loss": 1.0582, "step": 760 }, { "epoch": 0.3510784763653052, "grad_norm": 1.814204512009467, "learning_rate": 1.6399879772395915e-05, "loss": 1.0664, "step": 765 }, { "epoch": 0.3533731069297843, "grad_norm": 2.0517724058062794, "learning_rate": 1.633812609748206e-05, "loss": 1.0479, "step": 770 }, { "epoch": 0.35566773749426345, "grad_norm": 1.917088322296075, "learning_rate": 1.6275965751646682e-05, "loss": 1.049, "step": 775 }, { "epoch": 0.35796236805874254, "grad_norm": 1.8734674623090466, "learning_rate": 1.6213402723261852e-05, "loss": 1.0544, "step": 780 }, { "epoch": 0.36025699862322164, "grad_norm": 2.1022753225906268, "learning_rate": 1.6150441026536827e-05, "loss": 1.0487, "step": 785 }, { "epoch": 0.3625516291877008, "grad_norm": 1.8722897359622224, "learning_rate": 1.6087084701260468e-05, "loss": 1.0349, "step": 790 }, { "epoch": 0.3648462597521799, "grad_norm": 4.170483033000324, "learning_rate": 1.6023337812542048e-05, "loss": 1.0416, "step": 795 }, { "epoch": 0.36714089031665903, "grad_norm": 1.9386838285813406, "learning_rate": 1.5959204450550427e-05, "loss": 1.0588, "step": 800 }, { "epoch": 0.3694355208811381, "grad_norm": 2.0117934404635136, "learning_rate": 1.5894688730251613e-05, "loss": 1.0454, "step": 805 }, { "epoch": 0.3717301514456173, "grad_norm": 2.121087436021138, "learning_rate": 1.5829794791144723e-05, "loss": 1.0388, "step": 810 }, { "epoch": 0.37402478201009637, "grad_norm": 2.0084458470640394, "learning_rate": 1.57645267969964e-05, "loss": 1.0425, "step": 815 }, { "epoch": 0.3763194125745755, "grad_norm": 1.9502413688336317, "learning_rate": 1.569888893557365e-05, "loss": 1.0352, "step": 820 }, { "epoch": 0.3786140431390546, "grad_norm": 1.9753852329094403, "learning_rate": 1.5632885418375136e-05, "loss": 1.0562, "step": 825 }, { "epoch": 0.3809086737035337, "grad_norm": 2.0333895336233745, "learning_rate": 1.556652048036096e-05, "loss": 1.0571, "step": 830 }, { "epoch": 0.38320330426801286, "grad_norm": 2.2000920597322464, "learning_rate": 1.549979837968094e-05, "loss": 1.0504, "step": 835 }, { "epoch": 0.38549793483249195, "grad_norm": 2.0041860947851777, "learning_rate": 1.5432723397401406e-05, "loss": 1.0483, "step": 840 }, { "epoch": 0.3877925653969711, "grad_norm": 2.063158568485873, "learning_rate": 1.5365299837230483e-05, "loss": 1.0479, "step": 845 }, { "epoch": 0.3900871959614502, "grad_norm": 2.0282677973594017, "learning_rate": 1.5297532025241993e-05, "loss": 1.03, "step": 850 }, { "epoch": 0.39238182652592934, "grad_norm": 1.8181342854668368, "learning_rate": 1.5229424309597853e-05, "loss": 1.0461, "step": 855 }, { "epoch": 0.39467645709040844, "grad_norm": 1.9188404459309296, "learning_rate": 1.5160981060269107e-05, "loss": 1.0926, "step": 860 }, { "epoch": 0.3969710876548876, "grad_norm": 1.8587214491979347, "learning_rate": 1.5092206668755518e-05, "loss": 1.023, "step": 865 }, { "epoch": 0.3992657182193667, "grad_norm": 2.0702564004428736, "learning_rate": 1.5023105547803807e-05, "loss": 1.0497, "step": 870 }, { "epoch": 0.4015603487838458, "grad_norm": 1.866283622621383, "learning_rate": 1.4953682131124527e-05, "loss": 1.0548, "step": 875 }, { "epoch": 0.4038549793483249, "grad_norm": 2.3748260414549613, "learning_rate": 1.4883940873107572e-05, "loss": 1.0354, "step": 880 }, { "epoch": 0.406149609912804, "grad_norm": 1.8622352558527133, "learning_rate": 1.4813886248536376e-05, "loss": 1.0272, "step": 885 }, { "epoch": 0.40844424047728317, "grad_norm": 1.8451716042859445, "learning_rate": 1.4743522752300793e-05, "loss": 1.0346, "step": 890 }, { "epoch": 0.41073887104176227, "grad_norm": 1.7791032929700201, "learning_rate": 1.467285489910872e-05, "loss": 1.0445, "step": 895 }, { "epoch": 0.4130335016062414, "grad_norm": 1.9142844645885038, "learning_rate": 1.4601887223196374e-05, "loss": 1.0367, "step": 900 }, { "epoch": 0.4153281321707205, "grad_norm": 1.9821960797296039, "learning_rate": 1.4530624278037406e-05, "loss": 1.0459, "step": 905 }, { "epoch": 0.41762276273519966, "grad_norm": 1.8030416883238183, "learning_rate": 1.4459070636050721e-05, "loss": 1.0395, "step": 910 }, { "epoch": 0.41991739329967875, "grad_norm": 1.8541049154816385, "learning_rate": 1.4387230888307098e-05, "loss": 1.0297, "step": 915 }, { "epoch": 0.42221202386415785, "grad_norm": 1.882570932670042, "learning_rate": 1.4315109644234619e-05, "loss": 1.0526, "step": 920 }, { "epoch": 0.424506654428637, "grad_norm": 2.104659835262144, "learning_rate": 1.4242711531322912e-05, "loss": 1.0602, "step": 925 }, { "epoch": 0.4268012849931161, "grad_norm": 1.9520034799103085, "learning_rate": 1.4170041194826247e-05, "loss": 1.0513, "step": 930 }, { "epoch": 0.42909591555759524, "grad_norm": 1.8636499197241878, "learning_rate": 1.4097103297465471e-05, "loss": 1.0313, "step": 935 }, { "epoch": 0.43139054612207434, "grad_norm": 1.866901038371879, "learning_rate": 1.402390251912885e-05, "loss": 1.0492, "step": 940 }, { "epoch": 0.4336851766865535, "grad_norm": 1.9095505177623704, "learning_rate": 1.395044355657178e-05, "loss": 1.0427, "step": 945 }, { "epoch": 0.4359798072510326, "grad_norm": 1.8833835923493178, "learning_rate": 1.387673112311545e-05, "loss": 1.0062, "step": 950 }, { "epoch": 0.4382744378155117, "grad_norm": 1.8503750464774162, "learning_rate": 1.3802769948344406e-05, "loss": 1.0584, "step": 955 }, { "epoch": 0.4405690683799908, "grad_norm": 1.9850425042063173, "learning_rate": 1.3728564777803089e-05, "loss": 1.0299, "step": 960 }, { "epoch": 0.4428636989444699, "grad_norm": 2.0470542749046308, "learning_rate": 1.3654120372691361e-05, "loss": 1.0211, "step": 965 }, { "epoch": 0.44515832950894907, "grad_norm": 1.963916772949296, "learning_rate": 1.3579441509559007e-05, "loss": 1.0453, "step": 970 }, { "epoch": 0.44745296007342816, "grad_norm": 1.889454130597838, "learning_rate": 1.350453297999925e-05, "loss": 1.0441, "step": 975 }, { "epoch": 0.4497475906379073, "grad_norm": 1.8300661545935728, "learning_rate": 1.3429399590341325e-05, "loss": 1.0109, "step": 980 }, { "epoch": 0.4520422212023864, "grad_norm": 1.8927209237187554, "learning_rate": 1.3354046161342087e-05, "loss": 1.0153, "step": 985 }, { "epoch": 0.45433685176686556, "grad_norm": 1.8419785511286724, "learning_rate": 1.327847752787669e-05, "loss": 1.0234, "step": 990 }, { "epoch": 0.45663148233134465, "grad_norm": 1.8479631910892635, "learning_rate": 1.3202698538628376e-05, "loss": 1.0515, "step": 995 }, { "epoch": 0.45892611289582375, "grad_norm": 1.8454383680403843, "learning_rate": 1.3126714055777378e-05, "loss": 1.0283, "step": 1000 }, { "epoch": 0.4612207434603029, "grad_norm": 1.8614938060407424, "learning_rate": 1.3050528954688932e-05, "loss": 1.028, "step": 1005 }, { "epoch": 0.463515374024782, "grad_norm": 2.138120924499351, "learning_rate": 1.2974148123600477e-05, "loss": 1.0154, "step": 1010 }, { "epoch": 0.46581000458926114, "grad_norm": 1.8576229088325686, "learning_rate": 1.2897576463307999e-05, "loss": 1.036, "step": 1015 }, { "epoch": 0.46810463515374023, "grad_norm": 2.0045877966342776, "learning_rate": 1.2820818886851599e-05, "loss": 1.0382, "step": 1020 }, { "epoch": 0.4703992657182194, "grad_norm": 1.8120186612041058, "learning_rate": 1.2743880319200241e-05, "loss": 1.0532, "step": 1025 }, { "epoch": 0.4726938962826985, "grad_norm": 1.8202058192088733, "learning_rate": 1.2666765696935773e-05, "loss": 1.0223, "step": 1030 }, { "epoch": 0.47498852684717763, "grad_norm": 1.8341249835092734, "learning_rate": 1.2589479967936163e-05, "loss": 1.0327, "step": 1035 }, { "epoch": 0.4772831574116567, "grad_norm": 1.890017754125909, "learning_rate": 1.2512028091058044e-05, "loss": 1.0282, "step": 1040 }, { "epoch": 0.4795777879761358, "grad_norm": 1.7963761077816227, "learning_rate": 1.2434415035818535e-05, "loss": 1.015, "step": 1045 }, { "epoch": 0.48187241854061497, "grad_norm": 1.89045217328483, "learning_rate": 1.2356645782076384e-05, "loss": 1.0189, "step": 1050 }, { "epoch": 0.48416704910509406, "grad_norm": 1.8184075309069563, "learning_rate": 1.2278725319712449e-05, "loss": 1.0388, "step": 1055 }, { "epoch": 0.4864616796695732, "grad_norm": 1.7960523305438247, "learning_rate": 1.2200658648309531e-05, "loss": 1.0447, "step": 1060 }, { "epoch": 0.4887563102340523, "grad_norm": 1.8467823592474473, "learning_rate": 1.2122450776831593e-05, "loss": 1.0122, "step": 1065 }, { "epoch": 0.49105094079853145, "grad_norm": 1.7814611078259526, "learning_rate": 1.2044106723302364e-05, "loss": 0.998, "step": 1070 }, { "epoch": 0.49334557136301055, "grad_norm": 1.8655763524159674, "learning_rate": 1.1965631514483376e-05, "loss": 1.0051, "step": 1075 }, { "epoch": 0.4956402019274897, "grad_norm": 1.8504586712684936, "learning_rate": 1.1887030185551427e-05, "loss": 1.0173, "step": 1080 }, { "epoch": 0.4979348324919688, "grad_norm": 1.8553598830430822, "learning_rate": 1.1808307779775518e-05, "loss": 1.0276, "step": 1085 }, { "epoch": 0.5002294630564479, "grad_norm": 1.7545364104117196, "learning_rate": 1.1729469348193263e-05, "loss": 1.0079, "step": 1090 }, { "epoch": 0.502524093620927, "grad_norm": 1.7459754103129477, "learning_rate": 1.1650519949286797e-05, "loss": 1.0049, "step": 1095 }, { "epoch": 0.5048187241854062, "grad_norm": 1.8061298693213486, "learning_rate": 1.1571464648658201e-05, "loss": 0.9806, "step": 1100 }, { "epoch": 0.5071133547498853, "grad_norm": 1.7062918430841572, "learning_rate": 1.1492308518704507e-05, "loss": 1.0133, "step": 1105 }, { "epoch": 0.5094079853143644, "grad_norm": 1.744436514313244, "learning_rate": 1.1413056638292215e-05, "loss": 0.9962, "step": 1110 }, { "epoch": 0.5117026158788435, "grad_norm": 1.7695374210335713, "learning_rate": 1.1333714092431423e-05, "loss": 1.0085, "step": 1115 }, { "epoch": 0.5139972464433227, "grad_norm": 1.9055267822840887, "learning_rate": 1.1254285971949574e-05, "loss": 1.0198, "step": 1120 }, { "epoch": 0.5162918770078018, "grad_norm": 1.7470563158478891, "learning_rate": 1.1174777373164797e-05, "loss": 1.005, "step": 1125 }, { "epoch": 0.5185865075722809, "grad_norm": 1.8402758989282078, "learning_rate": 1.109519339755893e-05, "loss": 1.0225, "step": 1130 }, { "epoch": 0.52088113813676, "grad_norm": 1.9185683165429528, "learning_rate": 1.1015539151450172e-05, "loss": 1.0202, "step": 1135 }, { "epoch": 0.523175768701239, "grad_norm": 1.7862218928745999, "learning_rate": 1.0935819745665477e-05, "loss": 0.9938, "step": 1140 }, { "epoch": 0.5254703992657183, "grad_norm": 1.7946257670444443, "learning_rate": 1.0856040295212614e-05, "loss": 1.0203, "step": 1145 }, { "epoch": 0.5277650298301974, "grad_norm": 1.8084692996883154, "learning_rate": 1.077620591895197e-05, "loss": 1.0132, "step": 1150 }, { "epoch": 0.5300596603946764, "grad_norm": 1.8496851492898159, "learning_rate": 1.069632173926812e-05, "loss": 1.009, "step": 1155 }, { "epoch": 0.5323542909591555, "grad_norm": 1.7965406450823929, "learning_rate": 1.0616392881741166e-05, "loss": 0.9898, "step": 1160 }, { "epoch": 0.5346489215236347, "grad_norm": 1.8696271364197514, "learning_rate": 1.0536424474817848e-05, "loss": 0.9949, "step": 1165 }, { "epoch": 0.5369435520881138, "grad_norm": 1.8512029165438884, "learning_rate": 1.0456421649482502e-05, "loss": 1.0065, "step": 1170 }, { "epoch": 0.5392381826525929, "grad_norm": 1.833426494602823, "learning_rate": 1.0376389538927841e-05, "loss": 1.0158, "step": 1175 }, { "epoch": 0.541532813217072, "grad_norm": 1.8287651297802663, "learning_rate": 1.0296333278225599e-05, "loss": 1.0064, "step": 1180 }, { "epoch": 0.5438274437815511, "grad_norm": 1.8964046455895252, "learning_rate": 1.0216258003997044e-05, "loss": 1.0198, "step": 1185 }, { "epoch": 0.5461220743460303, "grad_norm": 1.777279291853422, "learning_rate": 1.0136168854083401e-05, "loss": 0.9712, "step": 1190 }, { "epoch": 0.5484167049105094, "grad_norm": 1.8585809331304046, "learning_rate": 1.0056070967216199e-05, "loss": 0.9954, "step": 1195 }, { "epoch": 0.5507113354749885, "grad_norm": 1.9725460354320736, "learning_rate": 9.975969482687547e-06, "loss": 0.9835, "step": 1200 }, { "epoch": 0.5530059660394676, "grad_norm": 1.7245123642288822, "learning_rate": 9.8958695400204e-06, "loss": 1.0004, "step": 1205 }, { "epoch": 0.5553005966039468, "grad_norm": 1.8366846045184153, "learning_rate": 9.815776278638772e-06, "loss": 1.0265, "step": 1210 }, { "epoch": 0.5575952271684259, "grad_norm": 1.8649330489030282, "learning_rate": 9.735694837537993e-06, "loss": 0.9572, "step": 1215 }, { "epoch": 0.559889857732905, "grad_norm": 1.7668993960491821, "learning_rate": 9.655630354954974e-06, "loss": 0.9922, "step": 1220 }, { "epoch": 0.5621844882973841, "grad_norm": 1.9573923418916903, "learning_rate": 9.57558796803852e-06, "loss": 1.0109, "step": 1225 }, { "epoch": 0.5644791188618632, "grad_norm": 1.711437620336517, "learning_rate": 9.495572812519718e-06, "loss": 0.9946, "step": 1230 }, { "epoch": 0.5667737494263424, "grad_norm": 1.7691621774728759, "learning_rate": 9.41559002238242e-06, "loss": 1.0061, "step": 1235 }, { "epoch": 0.5690683799908215, "grad_norm": 1.8386891074186076, "learning_rate": 9.33564472953383e-06, "loss": 0.9886, "step": 1240 }, { "epoch": 0.5713630105553006, "grad_norm": 1.8757816038935415, "learning_rate": 9.255742063475228e-06, "loss": 0.9952, "step": 1245 }, { "epoch": 0.5736576411197797, "grad_norm": 1.7204727826291393, "learning_rate": 9.175887150972841e-06, "loss": 0.9922, "step": 1250 }, { "epoch": 0.5759522716842589, "grad_norm": 1.7936118442403097, "learning_rate": 9.096085115728902e-06, "loss": 0.982, "step": 1255 }, { "epoch": 0.578246902248738, "grad_norm": 1.6703981680215847, "learning_rate": 9.016341078052908e-06, "loss": 0.9828, "step": 1260 }, { "epoch": 0.5805415328132171, "grad_norm": 1.836996443217227, "learning_rate": 8.93666015453307e-06, "loss": 0.9882, "step": 1265 }, { "epoch": 0.5828361633776962, "grad_norm": 2.001392851330759, "learning_rate": 8.857047457708023e-06, "loss": 0.9923, "step": 1270 }, { "epoch": 0.5851307939421753, "grad_norm": 1.779792680703507, "learning_rate": 8.777508095738818e-06, "loss": 0.9821, "step": 1275 }, { "epoch": 0.5874254245066545, "grad_norm": 1.7551844873505347, "learning_rate": 8.698047172081129e-06, "loss": 0.9905, "step": 1280 }, { "epoch": 0.5897200550711336, "grad_norm": 1.934345593929013, "learning_rate": 8.618669785157825e-06, "loss": 0.9796, "step": 1285 }, { "epoch": 0.5920146856356127, "grad_norm": 1.8648278829036111, "learning_rate": 8.539381028031838e-06, "loss": 1.0053, "step": 1290 }, { "epoch": 0.5943093162000918, "grad_norm": 1.7236526799565135, "learning_rate": 8.46018598807938e-06, "loss": 1.0005, "step": 1295 }, { "epoch": 0.596603946764571, "grad_norm": 1.8020519118969505, "learning_rate": 8.381089746663517e-06, "loss": 0.9855, "step": 1300 }, { "epoch": 0.59889857732905, "grad_norm": 1.8759050818870677, "learning_rate": 8.302097378808147e-06, "loss": 0.9884, "step": 1305 }, { "epoch": 0.6011932078935291, "grad_norm": 1.7101070470722577, "learning_rate": 8.223213952872353e-06, "loss": 0.9939, "step": 1310 }, { "epoch": 0.6034878384580082, "grad_norm": 1.8363414696856302, "learning_rate": 8.144444530225237e-06, "loss": 1.006, "step": 1315 }, { "epoch": 0.6057824690224873, "grad_norm": 1.8189880317647966, "learning_rate": 8.065794164921128e-06, "loss": 1.0103, "step": 1320 }, { "epoch": 0.6080770995869665, "grad_norm": 1.7184329357048536, "learning_rate": 7.987267903375331e-06, "loss": 0.9802, "step": 1325 }, { "epoch": 0.6103717301514456, "grad_norm": 1.6760569997350725, "learning_rate": 7.90887078404033e-06, "loss": 0.9864, "step": 1330 }, { "epoch": 0.6126663607159247, "grad_norm": 1.759298284823385, "learning_rate": 7.830607837082494e-06, "loss": 0.979, "step": 1335 }, { "epoch": 0.6149609912804038, "grad_norm": 1.693614807322877, "learning_rate": 7.75248408405934e-06, "loss": 0.9865, "step": 1340 }, { "epoch": 0.617255621844883, "grad_norm": 1.7256883752598309, "learning_rate": 7.674504537597336e-06, "loss": 0.9883, "step": 1345 }, { "epoch": 0.6195502524093621, "grad_norm": 1.7555572901812775, "learning_rate": 7.596674201070282e-06, "loss": 0.9953, "step": 1350 }, { "epoch": 0.6218448829738412, "grad_norm": 1.7549420037924819, "learning_rate": 7.518998068278266e-06, "loss": 0.9761, "step": 1355 }, { "epoch": 0.6241395135383203, "grad_norm": 1.7638758955771752, "learning_rate": 7.441481123127257e-06, "loss": 0.9663, "step": 1360 }, { "epoch": 0.6264341441027994, "grad_norm": 1.9316469572623007, "learning_rate": 7.364128339309326e-06, "loss": 0.9912, "step": 1365 }, { "epoch": 0.6287287746672786, "grad_norm": 1.7094849210033014, "learning_rate": 7.286944679983521e-06, "loss": 0.9759, "step": 1370 }, { "epoch": 0.6310234052317577, "grad_norm": 1.6887590729536133, "learning_rate": 7.209935097457412e-06, "loss": 0.9899, "step": 1375 }, { "epoch": 0.6333180357962368, "grad_norm": 1.7727590940094575, "learning_rate": 7.133104532869342e-06, "loss": 1.0, "step": 1380 }, { "epoch": 0.6356126663607159, "grad_norm": 1.8172481765220854, "learning_rate": 7.056457915871399e-06, "loss": 0.9663, "step": 1385 }, { "epoch": 0.637907296925195, "grad_norm": 1.7974312188733144, "learning_rate": 6.980000164313093e-06, "loss": 0.9586, "step": 1390 }, { "epoch": 0.6402019274896742, "grad_norm": 1.7508328675764726, "learning_rate": 6.903736183925835e-06, "loss": 0.9719, "step": 1395 }, { "epoch": 0.6424965580541533, "grad_norm": 1.7326843284737148, "learning_rate": 6.82767086800817e-06, "loss": 0.9961, "step": 1400 }, { "epoch": 0.6447911886186324, "grad_norm": 1.7352259185492422, "learning_rate": 6.751809097111799e-06, "loss": 0.997, "step": 1405 }, { "epoch": 0.6470858191831115, "grad_norm": 1.8933368018227825, "learning_rate": 6.676155738728438e-06, "loss": 0.9779, "step": 1410 }, { "epoch": 0.6493804497475907, "grad_norm": 1.6974498786941927, "learning_rate": 6.600715646977503e-06, "loss": 0.9799, "step": 1415 }, { "epoch": 0.6516750803120698, "grad_norm": 1.7806592546745614, "learning_rate": 6.525493662294669e-06, "loss": 0.9718, "step": 1420 }, { "epoch": 0.6539697108765489, "grad_norm": 1.7658701508098122, "learning_rate": 6.450494611121274e-06, "loss": 0.9852, "step": 1425 }, { "epoch": 0.656264341441028, "grad_norm": 1.721774620040715, "learning_rate": 6.375723305594658e-06, "loss": 0.9766, "step": 1430 }, { "epoch": 0.6585589720055071, "grad_norm": 1.8456782805007963, "learning_rate": 6.301184543239398e-06, "loss": 0.9604, "step": 1435 }, { "epoch": 0.6608536025699863, "grad_norm": 1.6517747886851708, "learning_rate": 6.2268831066594846e-06, "loss": 0.9978, "step": 1440 }, { "epoch": 0.6631482331344654, "grad_norm": 1.7868240633431423, "learning_rate": 6.152823763231463e-06, "loss": 0.9659, "step": 1445 }, { "epoch": 0.6654428636989445, "grad_norm": 1.6864295399871028, "learning_rate": 6.079011264798534e-06, "loss": 0.97, "step": 1450 }, { "epoch": 0.6677374942634235, "grad_norm": 1.7298790044434966, "learning_rate": 6.005450347365687e-06, "loss": 0.9807, "step": 1455 }, { "epoch": 0.6700321248279028, "grad_norm": 1.7628376564735946, "learning_rate": 5.932145730795793e-06, "loss": 0.9686, "step": 1460 }, { "epoch": 0.6723267553923818, "grad_norm": 1.7343262868098874, "learning_rate": 5.859102118506787e-06, "loss": 0.9524, "step": 1465 }, { "epoch": 0.6746213859568609, "grad_norm": 1.9084160680148732, "learning_rate": 5.786324197169887e-06, "loss": 0.9818, "step": 1470 }, { "epoch": 0.67691601652134, "grad_norm": 1.735061984384823, "learning_rate": 5.713816636408871e-06, "loss": 0.9673, "step": 1475 }, { "epoch": 0.6792106470858191, "grad_norm": 1.7982234436842786, "learning_rate": 5.641584088500461e-06, "loss": 0.9694, "step": 1480 }, { "epoch": 0.6815052776502983, "grad_norm": 1.7104886881206862, "learning_rate": 5.569631188075842e-06, "loss": 0.9638, "step": 1485 }, { "epoch": 0.6837999082147774, "grad_norm": 1.7069667971132039, "learning_rate": 5.497962551823266e-06, "loss": 0.9746, "step": 1490 }, { "epoch": 0.6860945387792565, "grad_norm": 2.380253232983682, "learning_rate": 5.426582778191858e-06, "loss": 0.9898, "step": 1495 }, { "epoch": 0.6883891693437356, "grad_norm": 1.789361116223533, "learning_rate": 5.355496447096533e-06, "loss": 0.9813, "step": 1500 }, { "epoch": 0.6906837999082148, "grad_norm": 1.7257448134853217, "learning_rate": 5.284708119624173e-06, "loss": 0.9897, "step": 1505 }, { "epoch": 0.6929784304726939, "grad_norm": 1.695066389007114, "learning_rate": 5.2142223377409616e-06, "loss": 0.9662, "step": 1510 }, { "epoch": 0.695273061037173, "grad_norm": 1.8265038779545326, "learning_rate": 5.144043624000944e-06, "loss": 0.9221, "step": 1515 }, { "epoch": 0.6975676916016521, "grad_norm": 1.682425505945564, "learning_rate": 5.074176481255873e-06, "loss": 0.9665, "step": 1520 }, { "epoch": 0.6998623221661312, "grad_norm": 1.6947722881421843, "learning_rate": 5.00462539236628e-06, "loss": 0.9735, "step": 1525 }, { "epoch": 0.7021569527306104, "grad_norm": 2.0862442184881926, "learning_rate": 4.935394819913849e-06, "loss": 0.9462, "step": 1530 }, { "epoch": 0.7044515832950895, "grad_norm": 1.8473532390280518, "learning_rate": 4.866489205915072e-06, "loss": 0.9699, "step": 1535 }, { "epoch": 0.7067462138595686, "grad_norm": 1.748002794088265, "learning_rate": 4.7979129715362625e-06, "loss": 1.0002, "step": 1540 }, { "epoch": 0.7090408444240477, "grad_norm": 1.70600748304509, "learning_rate": 4.72967051680985e-06, "loss": 0.9638, "step": 1545 }, { "epoch": 0.7113354749885269, "grad_norm": 1.6249584663789558, "learning_rate": 4.661766220352098e-06, "loss": 0.9414, "step": 1550 }, { "epoch": 0.713630105553006, "grad_norm": 1.80174378572434, "learning_rate": 4.594204439082122e-06, "loss": 0.9727, "step": 1555 }, { "epoch": 0.7159247361174851, "grad_norm": 1.8374638729274564, "learning_rate": 4.526989507942374e-06, "loss": 0.9565, "step": 1560 }, { "epoch": 0.7182193666819642, "grad_norm": 1.6726191265269426, "learning_rate": 4.460125739620479e-06, "loss": 0.9487, "step": 1565 }, { "epoch": 0.7205139972464433, "grad_norm": 1.8264345398924386, "learning_rate": 4.393617424272527e-06, "loss": 0.9636, "step": 1570 }, { "epoch": 0.7228086278109225, "grad_norm": 1.6887334363300377, "learning_rate": 4.3274688292478105e-06, "loss": 0.961, "step": 1575 }, { "epoch": 0.7251032583754016, "grad_norm": 1.8470657094730811, "learning_rate": 4.261684198815004e-06, "loss": 0.9694, "step": 1580 }, { "epoch": 0.7273978889398807, "grad_norm": 1.7132408618736092, "learning_rate": 4.196267753889864e-06, "loss": 0.9581, "step": 1585 }, { "epoch": 0.7296925195043598, "grad_norm": 1.6900030303741553, "learning_rate": 4.131223691764384e-06, "loss": 0.9325, "step": 1590 }, { "epoch": 0.731987150068839, "grad_norm": 1.7522033167073814, "learning_rate": 4.066556185837494e-06, "loss": 0.9347, "step": 1595 }, { "epoch": 0.7342817806333181, "grad_norm": 1.755668968211341, "learning_rate": 4.002269385347289e-06, "loss": 0.9256, "step": 1600 }, { "epoch": 0.7365764111977972, "grad_norm": 1.713678331254988, "learning_rate": 3.9383674151047936e-06, "loss": 0.975, "step": 1605 }, { "epoch": 0.7388710417622762, "grad_norm": 1.6865904408801047, "learning_rate": 3.8748543752293e-06, "loss": 0.9729, "step": 1610 }, { "epoch": 0.7411656723267553, "grad_norm": 1.6694830205166722, "learning_rate": 3.8117343408853124e-06, "loss": 0.9554, "step": 1615 }, { "epoch": 0.7434603028912345, "grad_norm": 1.6703972771704265, "learning_rate": 3.7490113620210487e-06, "loss": 0.9297, "step": 1620 }, { "epoch": 0.7457549334557136, "grad_norm": 1.7865359308285997, "learning_rate": 3.686689463108608e-06, "loss": 0.952, "step": 1625 }, { "epoch": 0.7480495640201927, "grad_norm": 1.7448768340578766, "learning_rate": 3.6247726428857344e-06, "loss": 0.9286, "step": 1630 }, { "epoch": 0.7503441945846718, "grad_norm": 1.7530030106253895, "learning_rate": 3.563264874099258e-06, "loss": 0.9524, "step": 1635 }, { "epoch": 0.752638825149151, "grad_norm": 1.653777443656037, "learning_rate": 3.5021701032501777e-06, "loss": 0.9218, "step": 1640 }, { "epoch": 0.7549334557136301, "grad_norm": 3.331790675502453, "learning_rate": 3.441492250340461e-06, "loss": 0.9453, "step": 1645 }, { "epoch": 0.7572280862781092, "grad_norm": 1.7244365474975045, "learning_rate": 3.3812352086215216e-06, "loss": 0.9374, "step": 1650 }, { "epoch": 0.7595227168425883, "grad_norm": 1.766304020059291, "learning_rate": 3.3214028443444034e-06, "loss": 0.9394, "step": 1655 }, { "epoch": 0.7618173474070674, "grad_norm": 1.7263700470016823, "learning_rate": 3.261998996511736e-06, "loss": 0.9371, "step": 1660 }, { "epoch": 0.7641119779715466, "grad_norm": 1.7511251400040453, "learning_rate": 3.203027476631386e-06, "loss": 0.9535, "step": 1665 }, { "epoch": 0.7664066085360257, "grad_norm": 1.7194345747840845, "learning_rate": 3.1444920684719394e-06, "loss": 0.9399, "step": 1670 }, { "epoch": 0.7687012391005048, "grad_norm": 1.7100670269940494, "learning_rate": 3.086396527819876e-06, "loss": 0.9311, "step": 1675 }, { "epoch": 0.7709958696649839, "grad_norm": 1.7008517044769462, "learning_rate": 3.028744582238633e-06, "loss": 0.9418, "step": 1680 }, { "epoch": 0.7732905002294631, "grad_norm": 1.7289960064005703, "learning_rate": 2.9715399308294003e-06, "loss": 0.9748, "step": 1685 }, { "epoch": 0.7755851307939422, "grad_norm": 1.7377650207700148, "learning_rate": 2.914786243993808e-06, "loss": 0.936, "step": 1690 }, { "epoch": 0.7778797613584213, "grad_norm": 1.8858822369634092, "learning_rate": 2.858487163198389e-06, "loss": 0.9358, "step": 1695 }, { "epoch": 0.7801743919229004, "grad_norm": 1.7588532610850607, "learning_rate": 2.8026463007409665e-06, "loss": 0.9399, "step": 1700 }, { "epoch": 0.7824690224873795, "grad_norm": 1.8795956958488933, "learning_rate": 2.747267239518857e-06, "loss": 0.9642, "step": 1705 }, { "epoch": 0.7847636530518587, "grad_norm": 1.733996923587813, "learning_rate": 2.6923535327989925e-06, "loss": 0.9447, "step": 1710 }, { "epoch": 0.7870582836163378, "grad_norm": 1.7231127716770163, "learning_rate": 2.637908703989924e-06, "loss": 0.9615, "step": 1715 }, { "epoch": 0.7893529141808169, "grad_norm": 1.7382322874229845, "learning_rate": 2.5839362464157635e-06, "loss": 0.9523, "step": 1720 }, { "epoch": 0.791647544745296, "grad_norm": 1.76018582887562, "learning_rate": 2.5304396230920346e-06, "loss": 0.9661, "step": 1725 }, { "epoch": 0.7939421753097752, "grad_norm": 1.6962974121939103, "learning_rate": 2.477422266503473e-06, "loss": 0.9491, "step": 1730 }, { "epoch": 0.7962368058742543, "grad_norm": 1.6864869043430553, "learning_rate": 2.424887578383799e-06, "loss": 0.9495, "step": 1735 }, { "epoch": 0.7985314364387334, "grad_norm": 1.7487461110618132, "learning_rate": 2.3728389294974472e-06, "loss": 0.9497, "step": 1740 }, { "epoch": 0.8008260670032125, "grad_norm": 3.804635705666014, "learning_rate": 2.3212796594232947e-06, "loss": 0.9633, "step": 1745 }, { "epoch": 0.8031206975676916, "grad_norm": 1.7104227286473583, "learning_rate": 2.2702130763403674e-06, "loss": 0.9402, "step": 1750 }, { "epoch": 0.8054153281321708, "grad_norm": 1.703649935060647, "learning_rate": 2.2196424568156073e-06, "loss": 0.9493, "step": 1755 }, { "epoch": 0.8077099586966499, "grad_norm": 1.731876264748493, "learning_rate": 2.1695710455936115e-06, "loss": 0.949, "step": 1760 }, { "epoch": 0.810004589261129, "grad_norm": 1.6008487931711857, "learning_rate": 2.1200020553884603e-06, "loss": 0.931, "step": 1765 }, { "epoch": 0.812299219825608, "grad_norm": 1.7485750666616973, "learning_rate": 2.0709386666775732e-06, "loss": 0.9473, "step": 1770 }, { "epoch": 0.8145938503900872, "grad_norm": 1.907625017544001, "learning_rate": 2.0223840274976413e-06, "loss": 0.9921, "step": 1775 }, { "epoch": 0.8168884809545663, "grad_norm": 1.7187402463869024, "learning_rate": 1.9743412532426355e-06, "loss": 0.9147, "step": 1780 }, { "epoch": 0.8191831115190454, "grad_norm": 1.6617657428032502, "learning_rate": 1.9268134264639273e-06, "loss": 0.9457, "step": 1785 }, { "epoch": 0.8214777420835245, "grad_norm": 1.7007750210249686, "learning_rate": 1.879803596672497e-06, "loss": 0.9665, "step": 1790 }, { "epoch": 0.8237723726480036, "grad_norm": 1.7101753911626782, "learning_rate": 1.8333147801432616e-06, "loss": 0.929, "step": 1795 }, { "epoch": 0.8260670032124828, "grad_norm": 1.774895067675658, "learning_rate": 1.7873499597215604e-06, "loss": 0.9205, "step": 1800 }, { "epoch": 0.8283616337769619, "grad_norm": 1.7257248036119928, "learning_rate": 1.7419120846317462e-06, "loss": 0.9482, "step": 1805 }, { "epoch": 0.830656264341441, "grad_norm": 1.7250389140740945, "learning_rate": 1.697004070287982e-06, "loss": 0.907, "step": 1810 }, { "epoch": 0.8329508949059201, "grad_norm": 1.6576292112238928, "learning_rate": 1.6526287981071477e-06, "loss": 0.9169, "step": 1815 }, { "epoch": 0.8352455254703993, "grad_norm": 1.6244943528548752, "learning_rate": 1.6087891153239932e-06, "loss": 0.9707, "step": 1820 }, { "epoch": 0.8375401560348784, "grad_norm": 1.7684196725742007, "learning_rate": 1.5654878348084246e-06, "loss": 0.9597, "step": 1825 }, { "epoch": 0.8398347865993575, "grad_norm": 1.7498743654009963, "learning_rate": 1.5227277348850466e-06, "loss": 0.9438, "step": 1830 }, { "epoch": 0.8421294171638366, "grad_norm": 1.666533412682978, "learning_rate": 1.4805115591548746e-06, "loss": 0.9234, "step": 1835 }, { "epoch": 0.8444240477283157, "grad_norm": 1.6784896812471775, "learning_rate": 1.4388420163193217e-06, "loss": 0.9228, "step": 1840 }, { "epoch": 0.8467186782927949, "grad_norm": 1.782913468726149, "learning_rate": 1.3977217800063847e-06, "loss": 0.9525, "step": 1845 }, { "epoch": 0.849013308857274, "grad_norm": 1.8714100640537485, "learning_rate": 1.3571534885991044e-06, "loss": 0.9202, "step": 1850 }, { "epoch": 0.8513079394217531, "grad_norm": 1.6341170338063304, "learning_rate": 1.3171397450662716e-06, "loss": 0.9324, "step": 1855 }, { "epoch": 0.8536025699862322, "grad_norm": 1.7569629604181838, "learning_rate": 1.2776831167954252e-06, "loss": 0.92, "step": 1860 }, { "epoch": 0.8558972005507114, "grad_norm": 1.7383544701565197, "learning_rate": 1.2387861354281194e-06, "loss": 0.9415, "step": 1865 }, { "epoch": 0.8581918311151905, "grad_norm": 1.6322549841517964, "learning_rate": 1.2004512966974746e-06, "loss": 0.9377, "step": 1870 }, { "epoch": 0.8604864616796696, "grad_norm": 1.6131611632532987, "learning_rate": 1.162681060268065e-06, "loss": 0.9339, "step": 1875 }, { "epoch": 0.8627810922441487, "grad_norm": 1.825918080886602, "learning_rate": 1.1254778495780749e-06, "loss": 0.9435, "step": 1880 }, { "epoch": 0.8650757228086278, "grad_norm": 1.7362620117149066, "learning_rate": 1.0888440516838373e-06, "loss": 0.9375, "step": 1885 }, { "epoch": 0.867370353373107, "grad_norm": 1.6924752721498857, "learning_rate": 1.0527820171066372e-06, "loss": 0.9518, "step": 1890 }, { "epoch": 0.8696649839375861, "grad_norm": 1.7645375725424164, "learning_rate": 1.0172940596819258e-06, "loss": 0.9105, "step": 1895 }, { "epoch": 0.8719596145020652, "grad_norm": 1.7445241378126046, "learning_rate": 9.823824564108408e-07, "loss": 0.9534, "step": 1900 }, { "epoch": 0.8742542450665443, "grad_norm": 1.6812050382224737, "learning_rate": 9.480494473141189e-07, "loss": 0.9513, "step": 1905 }, { "epoch": 0.8765488756310233, "grad_norm": 1.7180889119146323, "learning_rate": 9.142972352883595e-07, "loss": 0.9352, "step": 1910 }, { "epoch": 0.8788435061955026, "grad_norm": 1.593136362005407, "learning_rate": 8.811279859646915e-07, "loss": 0.9135, "step": 1915 }, { "epoch": 0.8811381367599816, "grad_norm": 1.732265054877238, "learning_rate": 8.485438275698154e-07, "loss": 0.9325, "step": 1920 }, { "epoch": 0.8834327673244607, "grad_norm": 1.6465162040573431, "learning_rate": 8.165468507894514e-07, "loss": 0.9363, "step": 1925 }, { "epoch": 0.8857273978889398, "grad_norm": 1.5926308744021644, "learning_rate": 7.851391086341953e-07, "loss": 0.9251, "step": 1930 }, { "epoch": 0.888022028453419, "grad_norm": 1.7096508943146629, "learning_rate": 7.543226163077899e-07, "loss": 0.9471, "step": 1935 }, { "epoch": 0.8903166590178981, "grad_norm": 1.803335599460816, "learning_rate": 7.240993510778304e-07, "loss": 0.9328, "step": 1940 }, { "epoch": 0.8926112895823772, "grad_norm": 1.7212998187307902, "learning_rate": 6.944712521488884e-07, "loss": 0.9479, "step": 1945 }, { "epoch": 0.8949059201468563, "grad_norm": 1.735482839392802, "learning_rate": 6.654402205380961e-07, "loss": 0.9541, "step": 1950 }, { "epoch": 0.8972005507113354, "grad_norm": 1.65097416177555, "learning_rate": 6.370081189531707e-07, "loss": 0.9132, "step": 1955 }, { "epoch": 0.8994951812758146, "grad_norm": 1.7196384780834688, "learning_rate": 6.091767716728924e-07, "loss": 0.9341, "step": 1960 }, { "epoch": 0.9017898118402937, "grad_norm": 1.8341416005690085, "learning_rate": 5.819479644300563e-07, "loss": 0.928, "step": 1965 }, { "epoch": 0.9040844424047728, "grad_norm": 1.7371545663059045, "learning_rate": 5.553234442969014e-07, "loss": 0.9311, "step": 1970 }, { "epoch": 0.9063790729692519, "grad_norm": 1.6781199455045805, "learning_rate": 5.293049195730038e-07, "loss": 0.9394, "step": 1975 }, { "epoch": 0.9086737035337311, "grad_norm": 1.7582513777384394, "learning_rate": 5.038940596756747e-07, "loss": 0.9495, "step": 1980 }, { "epoch": 0.9109683340982102, "grad_norm": 1.6554350150475086, "learning_rate": 4.790924950328435e-07, "loss": 0.9491, "step": 1985 }, { "epoch": 0.9132629646626893, "grad_norm": 1.5696628957382233, "learning_rate": 4.5490181697844916e-07, "loss": 0.9047, "step": 1990 }, { "epoch": 0.9155575952271684, "grad_norm": 1.660006052531741, "learning_rate": 4.313235776503244e-07, "loss": 0.9266, "step": 1995 }, { "epoch": 0.9178522257916475, "grad_norm": 1.6789601308304245, "learning_rate": 4.08359289890623e-07, "loss": 0.9356, "step": 2000 }, { "epoch": 0.9201468563561267, "grad_norm": 1.6938274156881534, "learning_rate": 3.860104271487397e-07, "loss": 0.9501, "step": 2005 }, { "epoch": 0.9224414869206058, "grad_norm": 1.6613669924815342, "learning_rate": 3.6427842338677353e-07, "loss": 0.9178, "step": 2010 }, { "epoch": 0.9247361174850849, "grad_norm": 1.6863633601389076, "learning_rate": 3.4316467298752264e-07, "loss": 0.9283, "step": 2015 }, { "epoch": 0.927030748049564, "grad_norm": 1.6704738495679314, "learning_rate": 3.226705306650113e-07, "loss": 0.9397, "step": 2020 }, { "epoch": 0.9293253786140432, "grad_norm": 1.6915649797171424, "learning_rate": 3.027973113775795e-07, "loss": 0.9162, "step": 2025 }, { "epoch": 0.9316200091785223, "grad_norm": 1.624516748205876, "learning_rate": 2.835462902434971e-07, "loss": 0.9088, "step": 2030 }, { "epoch": 0.9339146397430014, "grad_norm": 1.7321228286245276, "learning_rate": 2.649187024591604e-07, "loss": 0.9291, "step": 2035 }, { "epoch": 0.9362092703074805, "grad_norm": 1.6981737432031856, "learning_rate": 2.4691574321983216e-07, "loss": 0.9455, "step": 2040 }, { "epoch": 0.9385039008719596, "grad_norm": 1.6209611667412176, "learning_rate": 2.2953856764295623e-07, "loss": 0.9427, "step": 2045 }, { "epoch": 0.9407985314364388, "grad_norm": 1.6075798817603826, "learning_rate": 2.1278829069404483e-07, "loss": 0.8938, "step": 2050 }, { "epoch": 0.9430931620009179, "grad_norm": 1.6694215956258427, "learning_rate": 1.9666598711513663e-07, "loss": 0.9425, "step": 2055 }, { "epoch": 0.945387792565397, "grad_norm": 1.6380060844501796, "learning_rate": 1.811726913558387e-07, "loss": 0.9344, "step": 2060 }, { "epoch": 0.947682423129876, "grad_norm": 1.7361661814337723, "learning_rate": 1.663093975069552e-07, "loss": 0.9264, "step": 2065 }, { "epoch": 0.9499770536943553, "grad_norm": 1.72352480478006, "learning_rate": 1.5207705923670158e-07, "loss": 0.9274, "step": 2070 }, { "epoch": 0.9522716842588343, "grad_norm": 1.7260384890222744, "learning_rate": 1.3847658972951482e-07, "loss": 0.9416, "step": 2075 }, { "epoch": 0.9545663148233134, "grad_norm": 1.678840542944468, "learning_rate": 1.2550886162746468e-07, "loss": 0.9476, "step": 2080 }, { "epoch": 0.9568609453877925, "grad_norm": 1.7327761448558552, "learning_rate": 1.1317470697425837e-07, "loss": 0.9433, "step": 2085 }, { "epoch": 0.9591555759522716, "grad_norm": 1.6188397369323666, "learning_rate": 1.0147491716185675e-07, "loss": 0.9306, "step": 2090 }, { "epoch": 0.9614502065167508, "grad_norm": 1.7303211630716537, "learning_rate": 9.041024287969491e-08, "loss": 0.951, "step": 2095 }, { "epoch": 0.9637448370812299, "grad_norm": 1.6406456484197742, "learning_rate": 7.99813940665195e-08, "loss": 0.9414, "step": 2100 }, { "epoch": 0.966039467645709, "grad_norm": 1.742920472563737, "learning_rate": 7.018903986483083e-08, "loss": 0.9331, "step": 2105 }, { "epoch": 0.9683340982101881, "grad_norm": 1.7699514892789145, "learning_rate": 6.103380857795604e-08, "loss": 0.9458, "step": 2110 }, { "epoch": 0.9706287287746673, "grad_norm": 1.6207392453935818, "learning_rate": 5.251628762972916e-08, "loss": 0.9286, "step": 2115 }, { "epoch": 0.9729233593391464, "grad_norm": 1.6654831370842644, "learning_rate": 4.4637023526807875e-08, "loss": 0.9207, "step": 2120 }, { "epoch": 0.9752179899036255, "grad_norm": 1.6652739914078956, "learning_rate": 3.739652182360054e-08, "loss": 0.9153, "step": 2125 }, { "epoch": 0.9775126204681046, "grad_norm": 1.7380457873464283, "learning_rate": 3.079524708983095e-08, "loss": 0.9193, "step": 2130 }, { "epoch": 0.9798072510325837, "grad_norm": 1.7928242828367038, "learning_rate": 2.483362288073443e-08, "loss": 0.9258, "step": 2135 }, { "epoch": 0.9821018815970629, "grad_norm": 1.6588826041799296, "learning_rate": 1.9512031709874037e-08, "loss": 0.9275, "step": 2140 }, { "epoch": 0.984396512161542, "grad_norm": 1.763891394623808, "learning_rate": 1.4830815024606815e-08, "loss": 0.9211, "step": 2145 }, { "epoch": 0.9866911427260211, "grad_norm": 1.782131811620528, "learning_rate": 1.0790273184164701e-08, "loss": 0.8983, "step": 2150 }, { "epoch": 0.9889857732905002, "grad_norm": 2.0241090268732402, "learning_rate": 7.390665440393241e-09, "loss": 0.9335, "step": 2155 }, { "epoch": 0.9912804038549794, "grad_norm": 1.6899996297251758, "learning_rate": 4.632209921107133e-09, "loss": 0.9444, "step": 2160 }, { "epoch": 0.9935750344194585, "grad_norm": 1.6477330125103407, "learning_rate": 2.5150836161058624e-09, "loss": 0.9399, "step": 2165 }, { "epoch": 0.9958696649839376, "grad_norm": 1.7301015675243123, "learning_rate": 1.03942236580723e-09, "loss": 0.9332, "step": 2170 }, { "epoch": 0.9981642955484167, "grad_norm": 1.7014766333026237, "learning_rate": 2.053208525365502e-10, "loss": 0.9409, "step": 2175 }, { "epoch": 1.0, "eval_loss": 0.9490520358085632, "eval_runtime": 758.2182, "eval_samples_per_second": 20.352, "eval_steps_per_second": 0.637, "step": 2179 }, { "epoch": 1.0, "step": 2179, "total_flos": 456238269726720.0, "train_loss": 1.0045825210718344, "train_runtime": 27286.1988, "train_samples_per_second": 5.11, "train_steps_per_second": 0.08 } ], "logging_steps": 5, "max_steps": 2179, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "total_flos": 456238269726720.0, "train_batch_size": 16, "trial_name": null, "trial_params": null }