{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.9971722884265803, "eval_steps": 500, "global_step": 1236, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.02, "grad_norm": 0.31760987639427185, "learning_rate": 5e-06, "loss": 3.2917, "step": 5 }, { "epoch": 0.03, "grad_norm": 0.24959272146224976, "learning_rate": 1e-05, "loss": 3.3179, "step": 10 }, { "epoch": 0.05, "grad_norm": 0.2924320101737976, "learning_rate": 1.5e-05, "loss": 3.3109, "step": 15 }, { "epoch": 0.06, "grad_norm": 0.22757597267627716, "learning_rate": 2e-05, "loss": 3.251, "step": 20 }, { "epoch": 0.08, "grad_norm": 0.1803845316171646, "learning_rate": 2.5e-05, "loss": 3.2489, "step": 25 }, { "epoch": 0.1, "grad_norm": 0.19006718695163727, "learning_rate": 3e-05, "loss": 3.2527, "step": 30 }, { "epoch": 0.11, "grad_norm": 0.17421793937683105, "learning_rate": 3.5e-05, "loss": 3.225, "step": 35 }, { "epoch": 0.13, "grad_norm": 0.16238895058631897, "learning_rate": 4e-05, "loss": 3.2243, "step": 40 }, { "epoch": 0.15, "grad_norm": 0.16933956742286682, "learning_rate": 4.5e-05, "loss": 3.2182, "step": 45 }, { "epoch": 0.16, "grad_norm": 0.16001147031784058, "learning_rate": 5e-05, "loss": 3.2524, "step": 50 }, { "epoch": 0.18, "grad_norm": 0.16399356722831726, "learning_rate": 4.999599004988406e-05, "loss": 3.2055, "step": 55 }, { "epoch": 0.19, "grad_norm": 0.17454692721366882, "learning_rate": 4.9983961485912235e-05, "loss": 3.2122, "step": 60 }, { "epoch": 0.21, "grad_norm": 0.17302922904491425, "learning_rate": 4.9963918166799836e-05, "loss": 3.2483, "step": 65 }, { "epoch": 0.23, "grad_norm": 0.16110074520111084, "learning_rate": 4.9935866522363665e-05, "loss": 3.2212, "step": 70 }, { "epoch": 0.24, "grad_norm": 0.16342906653881073, "learning_rate": 4.989981555145928e-05, "loss": 3.2017, "step": 75 }, { "epoch": 0.26, "grad_norm": 0.16569341719150543, "learning_rate": 4.985577681909431e-05, "loss": 3.1735, "step": 80 }, { "epoch": 0.27, "grad_norm": 0.16199316084384918, "learning_rate": 4.9803764452718335e-05, "loss": 3.2362, "step": 85 }, { "epoch": 0.29, "grad_norm": 0.163334921002388, "learning_rate": 4.974379513769093e-05, "loss": 3.1873, "step": 90 }, { "epoch": 0.31, "grad_norm": 0.17866787314414978, "learning_rate": 4.967588811192902e-05, "loss": 3.2065, "step": 95 }, { "epoch": 0.32, "grad_norm": 0.16487747430801392, "learning_rate": 4.960006515973548e-05, "loss": 3.2155, "step": 100 }, { "epoch": 0.34, "grad_norm": 0.1537981629371643, "learning_rate": 4.9516350604810793e-05, "loss": 3.2032, "step": 105 }, { "epoch": 0.36, "grad_norm": 0.16236944496631622, "learning_rate": 4.9424771302450084e-05, "loss": 3.1732, "step": 110 }, { "epoch": 0.37, "grad_norm": 0.18180155754089355, "learning_rate": 4.932535663092809e-05, "loss": 3.1764, "step": 115 }, { "epoch": 0.39, "grad_norm": 0.21276631951332092, "learning_rate": 4.921813848207469e-05, "loss": 3.1902, "step": 120 }, { "epoch": 0.4, "grad_norm": 0.18672068417072296, "learning_rate": 4.9103151251044174e-05, "loss": 3.1871, "step": 125 }, { "epoch": 0.42, "grad_norm": 0.1597897857427597, "learning_rate": 4.898043182528136e-05, "loss": 3.1487, "step": 130 }, { "epoch": 0.44, "grad_norm": 0.1650695651769638, "learning_rate": 4.88500195726883e-05, "loss": 3.1696, "step": 135 }, { "epoch": 0.45, "grad_norm": 0.19672705233097076, "learning_rate": 4.871195632899518e-05, "loss": 3.1775, "step": 140 }, { "epoch": 0.47, "grad_norm": 0.21495744585990906, "learning_rate": 4.856628638433962e-05, "loss": 3.1691, "step": 145 }, { "epoch": 0.48, "grad_norm": 0.2001628875732422, "learning_rate": 4.841305646905851e-05, "loss": 3.1456, "step": 150 }, { "epoch": 0.5, "grad_norm": 0.18947800993919373, "learning_rate": 4.825231573869721e-05, "loss": 3.1269, "step": 155 }, { "epoch": 0.52, "grad_norm": 0.17868556082248688, "learning_rate": 4.808411575824051e-05, "loss": 3.181, "step": 160 }, { "epoch": 0.53, "grad_norm": 0.16371792554855347, "learning_rate": 4.7908510485570925e-05, "loss": 3.145, "step": 165 }, { "epoch": 0.55, "grad_norm": 0.16323748230934143, "learning_rate": 4.772555625415912e-05, "loss": 3.1727, "step": 170 }, { "epoch": 0.57, "grad_norm": 0.1625365912914276, "learning_rate": 4.753531175499243e-05, "loss": 3.147, "step": 175 }, { "epoch": 0.58, "grad_norm": 0.1947379857301712, "learning_rate": 4.733783801774696e-05, "loss": 3.1401, "step": 180 }, { "epoch": 0.6, "grad_norm": 0.19056910276412964, "learning_rate": 4.7133198391209566e-05, "loss": 3.1175, "step": 185 }, { "epoch": 0.61, "grad_norm": 0.184840589761734, "learning_rate": 4.692145852295576e-05, "loss": 3.1431, "step": 190 }, { "epoch": 0.63, "grad_norm": 0.16869348287582397, "learning_rate": 4.670268633829031e-05, "loss": 3.1332, "step": 195 }, { "epoch": 0.65, "grad_norm": 0.18127334117889404, "learning_rate": 4.6476952018456974e-05, "loss": 3.0935, "step": 200 }, { "epoch": 0.66, "grad_norm": 0.19219477474689484, "learning_rate": 4.6244327978124734e-05, "loss": 3.1763, "step": 205 }, { "epoch": 0.68, "grad_norm": 0.17717011272907257, "learning_rate": 4.600488884215737e-05, "loss": 3.1111, "step": 210 }, { "epoch": 0.69, "grad_norm": 0.17769476771354675, "learning_rate": 4.5758711421674166e-05, "loss": 3.1324, "step": 215 }, { "epoch": 0.71, "grad_norm": 0.1706475168466568, "learning_rate": 4.55058746894092e-05, "loss": 3.1194, "step": 220 }, { "epoch": 0.73, "grad_norm": 0.17008131742477417, "learning_rate": 4.524645975437717e-05, "loss": 3.1486, "step": 225 }, { "epoch": 0.74, "grad_norm": 0.19806550443172455, "learning_rate": 4.498054983585398e-05, "loss": 3.1491, "step": 230 }, { "epoch": 0.76, "grad_norm": 0.17164206504821777, "learning_rate": 4.470823023668032e-05, "loss": 3.1492, "step": 235 }, { "epoch": 0.78, "grad_norm": 0.16124162077903748, "learning_rate": 4.442958831589686e-05, "loss": 3.1251, "step": 240 }, { "epoch": 0.79, "grad_norm": 0.17326141893863678, "learning_rate": 4.41447134607198e-05, "loss": 3.1005, "step": 245 }, { "epoch": 0.81, "grad_norm": 0.18267503380775452, "learning_rate": 4.3853697057865816e-05, "loss": 3.1378, "step": 250 }, { "epoch": 0.82, "grad_norm": 0.17902736365795135, "learning_rate": 4.3556632464235594e-05, "loss": 3.139, "step": 255 }, { "epoch": 0.84, "grad_norm": 0.18006114661693573, "learning_rate": 4.3253614976965254e-05, "loss": 3.1213, "step": 260 }, { "epoch": 0.86, "grad_norm": 0.18118861317634583, "learning_rate": 4.294474180285545e-05, "loss": 3.1044, "step": 265 }, { "epoch": 0.87, "grad_norm": 0.17744861543178558, "learning_rate": 4.2630112027187816e-05, "loss": 3.1343, "step": 270 }, { "epoch": 0.89, "grad_norm": 0.16853739321231842, "learning_rate": 4.230982658193879e-05, "loss": 3.1137, "step": 275 }, { "epoch": 0.9, "grad_norm": 0.19506467878818512, "learning_rate": 4.1983988213401024e-05, "loss": 3.1064, "step": 280 }, { "epoch": 0.92, "grad_norm": 0.17228995263576508, "learning_rate": 4.165270144922283e-05, "loss": 3.0868, "step": 285 }, { "epoch": 0.94, "grad_norm": 0.17022188007831573, "learning_rate": 4.131607256487606e-05, "loss": 3.124, "step": 290 }, { "epoch": 0.95, "grad_norm": 0.17359808087348938, "learning_rate": 4.097420954956345e-05, "loss": 3.0944, "step": 295 }, { "epoch": 0.97, "grad_norm": 0.16375815868377686, "learning_rate": 4.0627222071576e-05, "loss": 3.1295, "step": 300 }, { "epoch": 0.99, "grad_norm": 0.18526001274585724, "learning_rate": 4.027522144311193e-05, "loss": 3.1082, "step": 305 }, { "epoch": 1.0, "grad_norm": 0.19401881098747253, "learning_rate": 3.991832058456812e-05, "loss": 3.0918, "step": 310 }, { "epoch": 1.02, "grad_norm": 0.19234879314899445, "learning_rate": 3.955663398831569e-05, "loss": 3.0422, "step": 315 }, { "epoch": 1.03, "grad_norm": 0.18378770351409912, "learning_rate": 3.919027768197132e-05, "loss": 3.0276, "step": 320 }, { "epoch": 1.05, "grad_norm": 0.189378023147583, "learning_rate": 3.881936919117607e-05, "loss": 3.0004, "step": 325 }, { "epoch": 1.07, "grad_norm": 0.2043304294347763, "learning_rate": 3.84440275018936e-05, "loss": 3.019, "step": 330 }, { "epoch": 1.08, "grad_norm": 0.1825193613767624, "learning_rate": 3.806437302223992e-05, "loss": 3.0234, "step": 335 }, { "epoch": 1.1, "grad_norm": 0.20000045001506805, "learning_rate": 3.768052754385703e-05, "loss": 3.0334, "step": 340 }, { "epoch": 1.11, "grad_norm": 0.2001664638519287, "learning_rate": 3.7292614202842566e-05, "loss": 3.0101, "step": 345 }, { "epoch": 1.13, "grad_norm": 0.18154805898666382, "learning_rate": 3.690075744024826e-05, "loss": 3.0098, "step": 350 }, { "epoch": 1.15, "grad_norm": 0.19567397236824036, "learning_rate": 3.650508296215978e-05, "loss": 3.0064, "step": 355 }, { "epoch": 1.16, "grad_norm": 0.19580692052841187, "learning_rate": 3.610571769937066e-05, "loss": 2.9863, "step": 360 }, { "epoch": 1.18, "grad_norm": 0.1916712522506714, "learning_rate": 3.5702789766663435e-05, "loss": 3.0066, "step": 365 }, { "epoch": 1.2, "grad_norm": 0.19139952957630157, "learning_rate": 3.529642842171096e-05, "loss": 3.0356, "step": 370 }, { "epoch": 1.21, "grad_norm": 0.18503962457180023, "learning_rate": 3.488676402361101e-05, "loss": 2.9687, "step": 375 }, { "epoch": 1.23, "grad_norm": 0.18717491626739502, "learning_rate": 3.4473927991067636e-05, "loss": 3.0223, "step": 380 }, { "epoch": 1.24, "grad_norm": 0.19255763292312622, "learning_rate": 3.405805276023256e-05, "loss": 3.0031, "step": 385 }, { "epoch": 1.26, "grad_norm": 0.1963498741388321, "learning_rate": 3.363927174222021e-05, "loss": 3.0106, "step": 390 }, { "epoch": 1.28, "grad_norm": 0.19114218652248383, "learning_rate": 3.3217719280309905e-05, "loss": 2.9939, "step": 395 }, { "epoch": 1.29, "grad_norm": 0.18220435082912445, "learning_rate": 3.279353060684914e-05, "loss": 2.9843, "step": 400 }, { "epoch": 1.31, "grad_norm": 0.19181542098522186, "learning_rate": 3.236684179987152e-05, "loss": 2.967, "step": 405 }, { "epoch": 1.32, "grad_norm": 0.188537135720253, "learning_rate": 3.193778973944354e-05, "loss": 2.9838, "step": 410 }, { "epoch": 1.34, "grad_norm": 0.18813706934452057, "learning_rate": 3.150651206375397e-05, "loss": 3.0091, "step": 415 }, { "epoch": 1.36, "grad_norm": 0.212008997797966, "learning_rate": 3.1073147124960034e-05, "loss": 2.9995, "step": 420 }, { "epoch": 1.37, "grad_norm": 0.22185884416103363, "learning_rate": 3.063783394480467e-05, "loss": 3.0065, "step": 425 }, { "epoch": 1.39, "grad_norm": 0.19503167271614075, "learning_rate": 3.0200712170018852e-05, "loss": 2.976, "step": 430 }, { "epoch": 1.41, "grad_norm": 0.2213529348373413, "learning_rate": 2.9761922027523513e-05, "loss": 2.9819, "step": 435 }, { "epoch": 1.42, "grad_norm": 0.1939724087715149, "learning_rate": 2.9321604279445258e-05, "loss": 3.0025, "step": 440 }, { "epoch": 1.44, "grad_norm": 0.19026894867420197, "learning_rate": 2.8879900177960488e-05, "loss": 2.9868, "step": 445 }, { "epoch": 1.45, "grad_norm": 0.2038724273443222, "learning_rate": 2.843695141998225e-05, "loss": 2.9699, "step": 450 }, { "epoch": 1.47, "grad_norm": 0.19388242065906525, "learning_rate": 2.7992900101704406e-05, "loss": 3.0146, "step": 455 }, { "epoch": 1.49, "grad_norm": 0.21257171034812927, "learning_rate": 2.7547888673017775e-05, "loss": 2.9556, "step": 460 }, { "epoch": 1.5, "grad_norm": 0.204994797706604, "learning_rate": 2.7102059891812757e-05, "loss": 2.9929, "step": 465 }, { "epoch": 1.52, "grad_norm": 0.21233408153057098, "learning_rate": 2.6655556778183198e-05, "loss": 2.969, "step": 470 }, { "epoch": 1.54, "grad_norm": 0.20436938107013702, "learning_rate": 2.6208522568546063e-05, "loss": 2.9893, "step": 475 }, { "epoch": 1.55, "grad_norm": 0.196218803524971, "learning_rate": 2.5761100669691817e-05, "loss": 2.9813, "step": 480 }, { "epoch": 1.57, "grad_norm": 0.20421084761619568, "learning_rate": 2.5313434612780084e-05, "loss": 2.9651, "step": 485 }, { "epoch": 1.58, "grad_norm": 0.18784867227077484, "learning_rate": 2.4865668007295394e-05, "loss": 2.9521, "step": 490 }, { "epoch": 1.6, "grad_norm": 0.22264321148395538, "learning_rate": 2.4417944494977888e-05, "loss": 2.9846, "step": 495 }, { "epoch": 1.62, "grad_norm": 0.20101739466190338, "learning_rate": 2.3970407703743568e-05, "loss": 2.9799, "step": 500 }, { "epoch": 1.63, "grad_norm": 0.20549960434436798, "learning_rate": 2.352320120160907e-05, "loss": 2.9651, "step": 505 }, { "epoch": 1.65, "grad_norm": 0.20113161206245422, "learning_rate": 2.3076468450635593e-05, "loss": 3.0041, "step": 510 }, { "epoch": 1.66, "grad_norm": 0.19988353550434113, "learning_rate": 2.263035276090687e-05, "loss": 2.9815, "step": 515 }, { "epoch": 1.68, "grad_norm": 0.20471665263175964, "learning_rate": 2.218499724455584e-05, "loss": 3.0102, "step": 520 }, { "epoch": 1.7, "grad_norm": 0.1928517073392868, "learning_rate": 2.1740544769854852e-05, "loss": 2.9728, "step": 525 }, { "epoch": 1.71, "grad_norm": 0.19472594559192657, "learning_rate": 2.129713791538411e-05, "loss": 2.9326, "step": 530 }, { "epoch": 1.73, "grad_norm": 0.19601967930793762, "learning_rate": 2.0854918924293006e-05, "loss": 2.9677, "step": 535 }, { "epoch": 1.75, "grad_norm": 0.21098604798316956, "learning_rate": 2.041402965866911e-05, "loss": 2.9511, "step": 540 }, { "epoch": 1.76, "grad_norm": 0.2089085876941681, "learning_rate": 1.9974611554029374e-05, "loss": 2.9258, "step": 545 }, { "epoch": 1.78, "grad_norm": 0.19997884333133698, "learning_rate": 1.953680557394816e-05, "loss": 2.9502, "step": 550 }, { "epoch": 1.79, "grad_norm": 0.19092586636543274, "learning_rate": 1.910075216483672e-05, "loss": 2.9492, "step": 555 }, { "epoch": 1.81, "grad_norm": 0.2013268768787384, "learning_rate": 1.8666591210888524e-05, "loss": 2.9583, "step": 560 }, { "epoch": 1.83, "grad_norm": 0.19930323958396912, "learning_rate": 1.8234461989204984e-05, "loss": 2.9586, "step": 565 }, { "epoch": 1.84, "grad_norm": 0.20546257495880127, "learning_rate": 1.7804503125115908e-05, "loss": 2.9489, "step": 570 }, { "epoch": 1.86, "grad_norm": 0.2034834921360016, "learning_rate": 1.7376852547709042e-05, "loss": 2.9444, "step": 575 }, { "epoch": 1.87, "grad_norm": 0.2054486870765686, "learning_rate": 1.6951647445582997e-05, "loss": 2.9398, "step": 580 }, { "epoch": 1.89, "grad_norm": 0.2075914889574051, "learning_rate": 1.652902422283764e-05, "loss": 2.9243, "step": 585 }, { "epoch": 1.91, "grad_norm": 0.20934872329235077, "learning_rate": 1.6109118455316273e-05, "loss": 2.9431, "step": 590 }, { "epoch": 1.92, "grad_norm": 0.21601444482803345, "learning_rate": 1.5692064847113375e-05, "loss": 2.9359, "step": 595 }, { "epoch": 1.94, "grad_norm": 0.19840656220912933, "learning_rate": 1.5277997187362113e-05, "loss": 2.9224, "step": 600 }, { "epoch": 1.96, "grad_norm": 0.19977068901062012, "learning_rate": 1.4867048307315318e-05, "loss": 2.9303, "step": 605 }, { "epoch": 1.97, "grad_norm": 0.20031294226646423, "learning_rate": 1.445935003773371e-05, "loss": 2.965, "step": 610 }, { "epoch": 1.99, "grad_norm": 0.20860356092453003, "learning_rate": 1.4055033166595159e-05, "loss": 2.9572, "step": 615 }, { "epoch": 2.0, "grad_norm": 0.2589782774448395, "learning_rate": 1.3654227397138419e-05, "loss": 2.9187, "step": 620 }, { "epoch": 2.02, "grad_norm": 0.22443072497844696, "learning_rate": 1.3257061306254825e-05, "loss": 2.9113, "step": 625 }, { "epoch": 2.04, "grad_norm": 0.22083857655525208, "learning_rate": 1.2863662303241342e-05, "loss": 2.8691, "step": 630 }, { "epoch": 2.05, "grad_norm": 0.22610703110694885, "learning_rate": 1.2474156588928201e-05, "loss": 2.8508, "step": 635 }, { "epoch": 2.07, "grad_norm": 0.2190137356519699, "learning_rate": 1.208866911519414e-05, "loss": 2.8914, "step": 640 }, { "epoch": 2.08, "grad_norm": 0.22293512523174286, "learning_rate": 1.1707323544882348e-05, "loss": 2.8692, "step": 645 }, { "epoch": 2.1, "grad_norm": 0.21358615159988403, "learning_rate": 1.1330242212129946e-05, "loss": 2.8271, "step": 650 }, { "epoch": 2.12, "grad_norm": 0.21398493647575378, "learning_rate": 1.095754608312366e-05, "loss": 2.8766, "step": 655 }, { "epoch": 2.13, "grad_norm": 0.22630859911441803, "learning_rate": 1.0589354717294339e-05, "loss": 2.8624, "step": 660 }, { "epoch": 2.15, "grad_norm": 0.20600877702236176, "learning_rate": 1.0225786228962794e-05, "loss": 2.8448, "step": 665 }, { "epoch": 2.17, "grad_norm": 0.2191879153251648, "learning_rate": 9.866957249449168e-06, "loss": 2.8518, "step": 670 }, { "epoch": 2.18, "grad_norm": 0.22143206000328064, "learning_rate": 9.512982889658118e-06, "loss": 2.8477, "step": 675 }, { "epoch": 2.2, "grad_norm": 0.2108563482761383, "learning_rate": 9.163976703151645e-06, "loss": 2.856, "step": 680 }, { "epoch": 2.21, "grad_norm": 0.21850666403770447, "learning_rate": 8.820050649721581e-06, "loss": 2.8622, "step": 685 }, { "epoch": 2.23, "grad_norm": 0.2310490608215332, "learning_rate": 8.481315059473372e-06, "loss": 2.8614, "step": 690 }, { "epoch": 2.25, "grad_norm": 0.22149670124053955, "learning_rate": 8.14787859743256e-06, "loss": 2.8515, "step": 695 }, { "epoch": 2.26, "grad_norm": 0.2082003504037857, "learning_rate": 7.819848228685511e-06, "loss": 2.8315, "step": 700 }, { "epoch": 2.28, "grad_norm": 0.23224076628684998, "learning_rate": 7.497329184065438e-06, "loss": 2.8405, "step": 705 }, { "epoch": 2.29, "grad_norm": 0.2128145843744278, "learning_rate": 7.1804249263947825e-06, "loss": 2.8578, "step": 710 }, { "epoch": 2.31, "grad_norm": 0.21818694472312927, "learning_rate": 6.869237117294722e-06, "loss": 2.8221, "step": 715 }, { "epoch": 2.33, "grad_norm": 0.21370700001716614, "learning_rate": 6.56386558457254e-06, "loss": 2.8116, "step": 720 }, { "epoch": 2.34, "grad_norm": 0.2164902687072754, "learning_rate": 6.264408290197296e-06, "loss": 2.8525, "step": 725 }, { "epoch": 2.36, "grad_norm": 0.22312401235103607, "learning_rate": 5.9709612988739695e-06, "loss": 2.8241, "step": 730 }, { "epoch": 2.38, "grad_norm": 0.22373555600643158, "learning_rate": 5.683618747226302e-06, "loss": 2.8624, "step": 735 }, { "epoch": 2.39, "grad_norm": 0.22297130525112152, "learning_rate": 5.4024728135981695e-06, "loss": 2.8759, "step": 740 }, { "epoch": 2.41, "grad_norm": 0.20991022884845734, "learning_rate": 5.127613688483099e-06, "loss": 2.8545, "step": 745 }, { "epoch": 2.42, "grad_norm": 0.23325349390506744, "learning_rate": 4.8591295455915396e-06, "loss": 2.8478, "step": 750 }, { "epoch": 2.44, "grad_norm": 0.20747652649879456, "learning_rate": 4.5971065135650775e-06, "loss": 2.8373, "step": 755 }, { "epoch": 2.46, "grad_norm": 0.2196613997220993, "learning_rate": 4.341628648346738e-06, "loss": 2.8495, "step": 760 }, { "epoch": 2.47, "grad_norm": 0.21464018523693085, "learning_rate": 4.092777906216136e-06, "loss": 2.8437, "step": 765 }, { "epoch": 2.49, "grad_norm": 0.21386373043060303, "learning_rate": 3.850634117498248e-06, "loss": 2.822, "step": 770 }, { "epoch": 2.5, "grad_norm": 0.22314497828483582, "learning_rate": 3.615274960954171e-06, "loss": 2.8418, "step": 775 }, { "epoch": 2.52, "grad_norm": 0.21419712901115417, "learning_rate": 3.3867759388620717e-06, "loss": 2.8767, "step": 780 }, { "epoch": 2.54, "grad_norm": 0.2160976380109787, "learning_rate": 3.165210352796355e-06, "loss": 2.8508, "step": 785 }, { "epoch": 2.55, "grad_norm": 0.22127215564250946, "learning_rate": 2.9506492801128216e-06, "loss": 2.864, "step": 790 }, { "epoch": 2.57, "grad_norm": 0.22035780549049377, "learning_rate": 2.743161551147344e-06, "loss": 2.8496, "step": 795 }, { "epoch": 2.59, "grad_norm": 0.21932028234004974, "learning_rate": 2.542813727135343e-06, "loss": 2.8862, "step": 800 }, { "epoch": 2.6, "grad_norm": 0.22386135160923004, "learning_rate": 2.349670078859226e-06, "loss": 2.8799, "step": 805 }, { "epoch": 2.62, "grad_norm": 0.2204558104276657, "learning_rate": 2.1637925660305756e-06, "loss": 2.8744, "step": 810 }, { "epoch": 2.64, "grad_norm": 0.2307337075471878, "learning_rate": 1.9852408174137305e-06, "loss": 2.8959, "step": 815 }, { "epoch": 2.65, "grad_norm": 0.21677350997924805, "learning_rate": 1.8140721116970904e-06, "loss": 2.8413, "step": 820 }, { "epoch": 2.67, "grad_norm": 0.20689088106155396, "learning_rate": 1.6503413591183598e-06, "loss": 2.8479, "step": 825 }, { "epoch": 2.69, "grad_norm": 0.21706977486610413, "learning_rate": 1.494101083849564e-06, "loss": 2.8886, "step": 830 }, { "epoch": 2.7, "grad_norm": 0.20675326883792877, "learning_rate": 1.3454014071475008e-06, "loss": 2.8745, "step": 835 }, { "epoch": 2.72, "grad_norm": 0.232208713889122, "learning_rate": 1.2042900312750265e-06, "loss": 2.8619, "step": 840 }, { "epoch": 2.73, "grad_norm": 0.22064746916294098, "learning_rate": 1.0708122241983949e-06, "loss": 2.8327, "step": 845 }, { "epoch": 2.75, "grad_norm": 0.22476071119308472, "learning_rate": 9.45010805065441e-07, "loss": 2.8462, "step": 850 }, { "epoch": 2.77, "grad_norm": 0.21655958890914917, "learning_rate": 8.269261304693798e-07, "loss": 2.858, "step": 855 }, { "epoch": 2.78, "grad_norm": 0.2096630185842514, "learning_rate": 7.165960815025769e-07, "loss": 2.8449, "step": 860 }, { "epoch": 2.8, "grad_norm": 0.20629090070724487, "learning_rate": 6.14056051604453e-07, "loss": 2.8543, "step": 865 }, { "epoch": 2.81, "grad_norm": 0.2086755633354187, "learning_rate": 5.193389352073836e-07, "loss": 2.8563, "step": 870 }, { "epoch": 2.83, "grad_norm": 0.2304910272359848, "learning_rate": 4.3247511718431663e-07, "loss": 2.8676, "step": 875 }, { "epoch": 2.85, "grad_norm": 0.21316254138946533, "learning_rate": 3.5349246310142914e-07, "loss": 2.849, "step": 880 }, { "epoch": 2.86, "grad_norm": 0.22124738991260529, "learning_rate": 2.8241631027895454e-07, "loss": 2.8742, "step": 885 }, { "epoch": 2.88, "grad_norm": 0.21196460723876953, "learning_rate": 2.1926945966306568e-07, "loss": 2.8629, "step": 890 }, { "epoch": 2.9, "grad_norm": 0.21042677760124207, "learning_rate": 1.6407216851144491e-07, "loss": 2.8601, "step": 895 }, { "epoch": 2.91, "grad_norm": 0.2190413475036621, "learning_rate": 1.1684214389481307e-07, "loss": 2.8737, "step": 900 }, { "epoch": 2.93, "grad_norm": 0.2288431078195572, "learning_rate": 9.009358251619367e-06, "loss": 2.8718, "step": 905 }, { "epoch": 2.94, "grad_norm": 0.2152530401945114, "learning_rate": 8.756246682582565e-06, "loss": 2.89, "step": 910 }, { "epoch": 2.96, "grad_norm": 0.21837005019187927, "learning_rate": 8.505984493817182e-06, "loss": 2.8403, "step": 915 }, { "epoch": 2.98, "grad_norm": 0.21821348369121552, "learning_rate": 8.258615584792079e-06, "loss": 2.8578, "step": 920 }, { "epoch": 2.99, "grad_norm": 0.22208793461322784, "learning_rate": 8.014183347454662e-06, "loss": 2.8394, "step": 925 }, { "epoch": 3.01, "grad_norm": 0.21861128509044647, "learning_rate": 7.77273065861924e-06, "loss": 2.8518, "step": 930 }, { "epoch": 3.02, "grad_norm": 0.21865125000476837, "learning_rate": 7.534299872445855e-06, "loss": 2.8264, "step": 935 }, { "epoch": 3.04, "grad_norm": 0.22050639986991882, "learning_rate": 7.298932813010834e-06, "loss": 2.8247, "step": 940 }, { "epoch": 3.06, "grad_norm": 0.21507680416107178, "learning_rate": 7.066670766970201e-06, "loss": 2.8686, "step": 945 }, { "epoch": 3.07, "grad_norm": 0.22617888450622559, "learning_rate": 6.837554476317412e-06, "loss": 2.8339, "step": 950 }, { "epoch": 3.09, "grad_norm": 0.22186200320720673, "learning_rate": 6.611624131236724e-06, "loss": 2.8544, "step": 955 }, { "epoch": 3.11, "grad_norm": 0.22832706570625305, "learning_rate": 6.388919363053197e-06, "loss": 2.8552, "step": 960 }, { "epoch": 3.12, "grad_norm": 0.22354218363761902, "learning_rate": 6.169479237280823e-06, "loss": 2.8238, "step": 965 }, { "epoch": 3.14, "grad_norm": 0.2282090187072754, "learning_rate": 5.95334224676995e-06, "loss": 2.8314, "step": 970 }, { "epoch": 3.15, "grad_norm": 0.24048872292041779, "learning_rate": 5.740546304955024e-06, "loss": 2.8377, "step": 975 }, { "epoch": 3.17, "grad_norm": 0.22570078074932098, "learning_rate": 5.531128739204072e-06, "loss": 2.8314, "step": 980 }, { "epoch": 3.19, "grad_norm": 0.22778679430484772, "learning_rate": 5.325126284270998e-06, "loss": 2.8072, "step": 985 }, { "epoch": 3.2, "grad_norm": 0.2253977507352829, "learning_rate": 5.1225750758517545e-06, "loss": 2.8477, "step": 990 }, { "epoch": 3.22, "grad_norm": 0.21220557391643524, "learning_rate": 4.923510644245632e-06, "loss": 2.8143, "step": 995 }, { "epoch": 3.23, "grad_norm": 0.2333599179983139, "learning_rate": 4.727967908122788e-06, "loss": 2.8508, "step": 1000 }, { "epoch": 3.25, "grad_norm": 0.2198331207036972, "learning_rate": 4.5359811683990175e-06, "loss": 2.8267, "step": 1005 }, { "epoch": 3.27, "grad_norm": 0.22090409696102142, "learning_rate": 4.3475841022188604e-06, "loss": 2.8183, "step": 1010 }, { "epoch": 3.28, "grad_norm": 0.21586580574512482, "learning_rate": 4.162809757048197e-06, "loss": 2.799, "step": 1015 }, { "epoch": 3.3, "grad_norm": 0.22138214111328125, "learning_rate": 3.981690544877273e-06, "loss": 2.867, "step": 1020 }, { "epoch": 3.32, "grad_norm": 0.23438501358032227, "learning_rate": 3.80425823653513e-06, "loss": 2.8302, "step": 1025 }, { "epoch": 3.33, "grad_norm": 0.2235979288816452, "learning_rate": 3.630543956116614e-06, "loss": 2.8301, "step": 1030 }, { "epoch": 3.35, "grad_norm": 0.2205328643321991, "learning_rate": 3.4605781755227774e-06, "loss": 2.8322, "step": 1035 }, { "epoch": 3.36, "grad_norm": 0.226404070854187, "learning_rate": 3.294390709115622e-06, "loss": 2.8633, "step": 1040 }, { "epoch": 3.38, "grad_norm": 0.22632290422916412, "learning_rate": 3.132010708488317e-06, "loss": 2.8269, "step": 1045 }, { "epoch": 3.4, "grad_norm": 0.22171452641487122, "learning_rate": 2.9734666573515784e-06, "loss": 2.8466, "step": 1050 }, { "epoch": 3.41, "grad_norm": 0.2134045958518982, "learning_rate": 2.8187863665372017e-06, "loss": 2.8398, "step": 1055 }, { "epoch": 3.43, "grad_norm": 0.22553056478500366, "learning_rate": 2.6679969691196904e-06, "loss": 2.801, "step": 1060 }, { "epoch": 3.44, "grad_norm": 0.23389177024364471, "learning_rate": 2.52112491565675e-06, "loss": 2.8276, "step": 1065 }, { "epoch": 3.46, "grad_norm": 0.22804100811481476, "learning_rate": 2.3781959695494267e-06, "loss": 2.8394, "step": 1070 }, { "epoch": 3.48, "grad_norm": 0.2190062403678894, "learning_rate": 2.2392352025229184e-06, "loss": 2.831, "step": 1075 }, { "epoch": 3.49, "grad_norm": 0.21799859404563904, "learning_rate": 2.1042669902286204e-06, "loss": 2.8119, "step": 1080 }, { "epoch": 3.51, "grad_norm": 0.22122910618782043, "learning_rate": 1.973315007968274e-06, "loss": 2.8602, "step": 1085 }, { "epoch": 3.53, "grad_norm": 0.2231130748987198, "learning_rate": 1.846402226541019e-06, "loss": 2.8384, "step": 1090 }, { "epoch": 3.54, "grad_norm": 0.23154446482658386, "learning_rate": 1.723550908214e-06, "loss": 2.8427, "step": 1095 }, { "epoch": 3.56, "grad_norm": 0.21399781107902527, "learning_rate": 1.6047826028172202e-06, "loss": 2.8229, "step": 1100 }, { "epoch": 3.57, "grad_norm": 0.2319013625383377, "learning_rate": 1.4901181439634288e-06, "loss": 2.8229, "step": 1105 }, { "epoch": 3.59, "grad_norm": 0.22346845269203186, "learning_rate": 1.3795776453936287e-06, "loss": 2.8075, "step": 1110 }, { "epoch": 3.61, "grad_norm": 0.22235344350337982, "learning_rate": 1.2731804974487977e-06, "loss": 2.8388, "step": 1115 }, { "epoch": 3.62, "grad_norm": 0.22748830914497375, "learning_rate": 1.1709453636685964e-06, "loss": 2.8469, "step": 1120 }, { "epoch": 3.64, "grad_norm": 0.2320627123117447, "learning_rate": 1.0728901775175272e-06, "loss": 2.8171, "step": 1125 }, { "epoch": 3.65, "grad_norm": 0.2156900018453598, "learning_rate": 9.790321392391027e-07, "loss": 2.8332, "step": 1130 }, { "epoch": 3.67, "grad_norm": 0.21801438927650452, "learning_rate": 8.893877128387462e-07, "loss": 2.836, "step": 1135 }, { "epoch": 3.69, "grad_norm": 0.23140712082386017, "learning_rate": 8.039726231957479e-07, "loss": 2.804, "step": 1140 }, { "epoch": 3.7, "grad_norm": 0.22428776323795319, "learning_rate": 7.228018533048886e-07, "loss": 2.8004, "step": 1145 }, { "epoch": 3.72, "grad_norm": 0.22678926587104797, "learning_rate": 6.45889641648234e-07, "loss": 2.8458, "step": 1150 }, { "epoch": 3.74, "grad_norm": 0.23222048580646515, "learning_rate": 5.732494796975157e-07, "loss": 2.853, "step": 1155 }, { "epoch": 3.75, "grad_norm": 0.20704413950443268, "learning_rate": 5.048941095475074e-07, "loss": 2.8043, "step": 1160 }, { "epoch": 3.77, "grad_norm": 0.2371576577425003, "learning_rate": 4.408355216809129e-07, "loss": 2.867, "step": 1165 }, { "epoch": 3.78, "grad_norm": 0.21293425559997559, "learning_rate": 3.810849528650545e-07, "loss": 2.8478, "step": 1170 }, { "epoch": 3.8, "grad_norm": 0.25140973925590515, "learning_rate": 3.2565288418079096e-07, "loss": 2.8324, "step": 1175 }, { "epoch": 3.82, "grad_norm": 0.21989835798740387, "learning_rate": 2.7454903918399386e-07, "loss": 2.8174, "step": 1180 }, { "epoch": 3.83, "grad_norm": 0.22354339063167572, "learning_rate": 2.2778238219989533e-07, "loss": 2.8396, "step": 1185 }, { "epoch": 3.85, "grad_norm": 0.23933887481689453, "learning_rate": 1.8536111675062084e-07, "loss": 2.822, "step": 1190 }, { "epoch": 3.86, "grad_norm": 0.21597786247730255, "learning_rate": 1.4729268411618202e-07, "loss": 2.8396, "step": 1195 }, { "epoch": 3.88, "grad_norm": 0.2182658612728119, "learning_rate": 1.1358376202916798e-07, "loss": 2.8289, "step": 1200 }, { "epoch": 3.9, "grad_norm": 0.23224902153015137, "learning_rate": 8.42402635033851e-08, "loss": 2.8171, "step": 1205 }, { "epoch": 3.91, "grad_norm": 0.2257353514432907, "learning_rate": 5.926733579661725e-08, "loss": 2.7882, "step": 1210 }, { "epoch": 3.93, "grad_norm": 0.2192804217338562, "learning_rate": 3.866935950774253e-08, "loss": 2.8312, "step": 1215 }, { "epoch": 3.95, "grad_norm": 0.2221554070711136, "learning_rate": 2.2449947808309002e-08, "loss": 2.8118, "step": 1220 }, { "epoch": 3.96, "grad_norm": 0.2162608802318573, "learning_rate": 1.0611945808725043e-08, "loss": 2.8038, "step": 1225 }, { "epoch": 3.98, "grad_norm": 0.2273118495941162, "learning_rate": 3.157430059200217e-09, "loss": 2.8124, "step": 1230 }, { "epoch": 3.99, "grad_norm": 0.22424130141735077, "learning_rate": 8.770818548942128e-11, "loss": 2.8155, "step": 1235 }, { "epoch": 4.0, "step": 1236, "total_flos": 1.3106595430434079e+19, "train_loss": 0.7700594385850776, "train_runtime": 36129.1454, "train_samples_per_second": 4.385, "train_steps_per_second": 0.034 } ], "logging_steps": 5, "max_steps": 1236, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 100, "total_flos": 1.3106595430434079e+19, "train_batch_size": 8, "trial_name": null, "trial_params": null }