{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.07790039877585088, "eval_steps": 500, "global_step": 3000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 2.596679959195029e-05, "grad_norm": 10.167712131798243, "learning_rate": 0.0, "loss": 2.0279, "step": 1 }, { "epoch": 5.193359918390058e-05, "grad_norm": 12.71908591386877, "learning_rate": 9.82808161164113e-07, "loss": 2.051, "step": 2 }, { "epoch": 7.790039877585088e-05, "grad_norm": 31.52939070180938, "learning_rate": 1.5577140808478338e-06, "loss": 2.0822, "step": 3 }, { "epoch": 0.00010386719836780116, "grad_norm": 8.52302663566333, "learning_rate": 1.965616322328226e-06, "loss": 2.0849, "step": 4 }, { "epoch": 0.00012983399795975147, "grad_norm": 12.410036369227404, "learning_rate": 2.2820098812915408e-06, "loss": 2.0754, "step": 5 }, { "epoch": 0.00015580079755170176, "grad_norm": 8.92003156921548, "learning_rate": 2.540522242011947e-06, "loss": 2.0222, "step": 6 }, { "epoch": 0.00018176759714365204, "grad_norm": 13.189865984810673, "learning_rate": 2.759091328682456e-06, "loss": 2.0348, "step": 7 }, { "epoch": 0.00020773439673560232, "grad_norm": 12.068025420458316, "learning_rate": 2.948424483492339e-06, "loss": 2.0539, "step": 8 }, { "epoch": 0.00023370119632755264, "grad_norm": 8.673342798781551, "learning_rate": 3.1154281616956676e-06, "loss": 2.0554, "step": 9 }, { "epoch": 0.00025966799591950295, "grad_norm": 58.430923760988925, "learning_rate": 3.2648180424556543e-06, "loss": 2.065, "step": 10 }, { "epoch": 0.0002856347955114532, "grad_norm": 8.40080362744625, "learning_rate": 3.3999576277859138e-06, "loss": 2.039, "step": 11 }, { "epoch": 0.0003116015951034035, "grad_norm": 11.594089554495266, "learning_rate": 3.52333040317606e-06, "loss": 2.0788, "step": 12 }, { "epoch": 0.0003375683946953538, "grad_norm": 11.471317513183969, "learning_rate": 3.636822354884896e-06, "loss": 2.0432, "step": 13 }, { "epoch": 0.0003635351942873041, "grad_norm": 11.566295433866307, "learning_rate": 3.7418994898465686e-06, "loss": 2.0643, "step": 14 }, { "epoch": 0.0003895019938792544, "grad_norm": 10.147984542311645, "learning_rate": 3.839723962139374e-06, "loss": 2.0618, "step": 15 }, { "epoch": 0.00041546879347120465, "grad_norm": 7.1649375722205955, "learning_rate": 3.931232644656452e-06, "loss": 2.0061, "step": 16 }, { "epoch": 0.00044143559306315496, "grad_norm": 8.161162201638032, "learning_rate": 4.017191838835887e-06, "loss": 1.9924, "step": 17 }, { "epoch": 0.00046740239265510527, "grad_norm": 14.897103436436844, "learning_rate": 4.09823632285978e-06, "loss": 1.9987, "step": 18 }, { "epoch": 0.0004933691922470555, "grad_norm": 9.367730335735887, "learning_rate": 4.174897828245933e-06, "loss": 1.9955, "step": 19 }, { "epoch": 0.0005193359918390059, "grad_norm": 13.833683116734269, "learning_rate": 4.2476262036197665e-06, "loss": 1.9203, "step": 20 }, { "epoch": 0.0005453027914309561, "grad_norm": 9.656884033720331, "learning_rate": 4.3168054095302896e-06, "loss": 1.9405, "step": 21 }, { "epoch": 0.0005712695910229064, "grad_norm": 7.4336704685183275, "learning_rate": 4.382765788950027e-06, "loss": 1.9728, "step": 22 }, { "epoch": 0.0005972363906148568, "grad_norm": 10.20043087217532, "learning_rate": 4.445793607944332e-06, "loss": 1.9714, "step": 23 }, { "epoch": 0.000623203190206807, "grad_norm": 7.451984822133405, "learning_rate": 4.506138564340173e-06, "loss": 1.934, "step": 24 }, { "epoch": 0.0006491699897987573, "grad_norm": 9.163112529353, "learning_rate": 4.5640197625830816e-06, "loss": 1.8725, "step": 25 }, { "epoch": 0.0006751367893907076, "grad_norm": 7.889725254240628, "learning_rate": 4.619630516049009e-06, "loss": 1.9659, "step": 26 }, { "epoch": 0.0007011035889826579, "grad_norm": 14.977112665951129, "learning_rate": 4.673142242543501e-06, "loss": 1.8455, "step": 27 }, { "epoch": 0.0007270703885746082, "grad_norm": 11.452971786577518, "learning_rate": 4.724707651010682e-06, "loss": 1.8676, "step": 28 }, { "epoch": 0.0007530371881665584, "grad_norm": 8.541682198363038, "learning_rate": 4.774463368791538e-06, "loss": 1.8726, "step": 29 }, { "epoch": 0.0007790039877585088, "grad_norm": 8.082582524093958, "learning_rate": 4.822532123303487e-06, "loss": 1.8402, "step": 30 }, { "epoch": 0.000804970787350459, "grad_norm": 17.061179979845694, "learning_rate": 4.869024565857359e-06, "loss": 1.855, "step": 31 }, { "epoch": 0.0008309375869424093, "grad_norm": 10.305135790926464, "learning_rate": 4.914040805820566e-06, "loss": 1.8031, "step": 32 }, { "epoch": 0.0008569043865343597, "grad_norm": 17.521505166809924, "learning_rate": 4.957671708633747e-06, "loss": 1.7963, "step": 33 }, { "epoch": 0.0008828711861263099, "grad_norm": 27.768773782133074, "learning_rate": 5e-06, "loss": 1.797, "step": 34 }, { "epoch": 0.0009088379857182602, "grad_norm": 7.709614772647655, "learning_rate": 5.041101209973997e-06, "loss": 1.7693, "step": 35 }, { "epoch": 0.0009348047853102105, "grad_norm": 40.43078274482668, "learning_rate": 5.081044484023894e-06, "loss": 1.7098, "step": 36 }, { "epoch": 0.0009607715849021608, "grad_norm": 13.228036202417233, "learning_rate": 5.1198932829439884e-06, "loss": 1.7796, "step": 37 }, { "epoch": 0.000986738384494111, "grad_norm": 12.98342308108088, "learning_rate": 5.157705989410047e-06, "loss": 1.7911, "step": 38 }, { "epoch": 0.0010127051840860613, "grad_norm": 11.634216592772333, "learning_rate": 5.19453643573273e-06, "loss": 1.6965, "step": 39 }, { "epoch": 0.0010386719836780118, "grad_norm": 11.958818343868286, "learning_rate": 5.23043436478388e-06, "loss": 1.755, "step": 40 }, { "epoch": 0.001064638783269962, "grad_norm": 6.580766233685262, "learning_rate": 5.265445833999807e-06, "loss": 1.74, "step": 41 }, { "epoch": 0.0010906055828619123, "grad_norm": 21.555133478985347, "learning_rate": 5.299613570694403e-06, "loss": 1.6821, "step": 42 }, { "epoch": 0.0011165723824538626, "grad_norm": 12.351837985147316, "learning_rate": 5.332977285558406e-06, "loss": 1.6646, "step": 43 }, { "epoch": 0.0011425391820458128, "grad_norm": 7.967517952014531, "learning_rate": 5.365573950114139e-06, "loss": 1.6549, "step": 44 }, { "epoch": 0.001168505981637763, "grad_norm": 43.7458996404413, "learning_rate": 5.397438042987208e-06, "loss": 1.641, "step": 45 }, { "epoch": 0.0011944727812297135, "grad_norm": 10.5030422344766, "learning_rate": 5.428601769108445e-06, "loss": 1.6464, "step": 46 }, { "epoch": 0.0012204395808216638, "grad_norm": 10.938006405687435, "learning_rate": 5.4590952553399815e-06, "loss": 1.6533, "step": 47 }, { "epoch": 0.001246406380413614, "grad_norm": 7.8619150110899545, "learning_rate": 5.488946725504286e-06, "loss": 1.5973, "step": 48 }, { "epoch": 0.0012723731800055643, "grad_norm": 14.23050306233854, "learning_rate": 5.518182657364912e-06, "loss": 1.5932, "step": 49 }, { "epoch": 0.0012983399795975146, "grad_norm": 9.252604349903836, "learning_rate": 5.546827923747195e-06, "loss": 1.6129, "step": 50 }, { "epoch": 0.0013243067791894648, "grad_norm": 13.090581615282309, "learning_rate": 5.5749059196837205e-06, "loss": 1.5941, "step": 51 }, { "epoch": 0.0013502735787814153, "grad_norm": 7.915832020446602, "learning_rate": 5.602438677213122e-06, "loss": 1.6106, "step": 52 }, { "epoch": 0.0013762403783733656, "grad_norm": 9.105486764648631, "learning_rate": 5.62944696924357e-06, "loss": 1.5117, "step": 53 }, { "epoch": 0.0014022071779653158, "grad_norm": 7.126014771057802, "learning_rate": 5.655950403707615e-06, "loss": 1.5508, "step": 54 }, { "epoch": 0.001428173977557266, "grad_norm": 9.865708241007317, "learning_rate": 5.681967509077455e-06, "loss": 1.5662, "step": 55 }, { "epoch": 0.0014541407771492163, "grad_norm": 9.65133077507866, "learning_rate": 5.7075158121747964e-06, "loss": 1.4955, "step": 56 }, { "epoch": 0.0014801075767411666, "grad_norm": 8.443505078765376, "learning_rate": 5.732611909093767e-06, "loss": 1.528, "step": 57 }, { "epoch": 0.0015060743763331168, "grad_norm": 23.391536962432454, "learning_rate": 5.75727152995565e-06, "loss": 1.532, "step": 58 }, { "epoch": 0.0015320411759250673, "grad_norm": 10.05259226299638, "learning_rate": 5.781509598128162e-06, "loss": 1.5033, "step": 59 }, { "epoch": 0.0015580079755170176, "grad_norm": 7.054507704225335, "learning_rate": 5.805340284467601e-06, "loss": 1.4956, "step": 60 }, { "epoch": 0.0015839747751089678, "grad_norm": 7.676786819867121, "learning_rate": 5.828777057077528e-06, "loss": 1.5122, "step": 61 }, { "epoch": 0.001609941574700918, "grad_norm": 8.308300821998932, "learning_rate": 5.851832727021472e-06, "loss": 1.4783, "step": 62 }, { "epoch": 0.0016359083742928683, "grad_norm": 7.94511036723585, "learning_rate": 5.8745194903781236e-06, "loss": 1.4798, "step": 63 }, { "epoch": 0.0016618751738848186, "grad_norm": 10.036220566856906, "learning_rate": 5.896848966984678e-06, "loss": 1.4794, "step": 64 }, { "epoch": 0.001687841973476769, "grad_norm": 11.66207388439793, "learning_rate": 5.918832236176436e-06, "loss": 1.5142, "step": 65 }, { "epoch": 0.0017138087730687193, "grad_norm": 8.391434671275535, "learning_rate": 5.940479869797859e-06, "loss": 1.5192, "step": 66 }, { "epoch": 0.0017397755726606696, "grad_norm": 6.706762220689152, "learning_rate": 5.961801962731307e-06, "loss": 1.4689, "step": 67 }, { "epoch": 0.0017657423722526198, "grad_norm": 6.743265253698558, "learning_rate": 5.982808161164114e-06, "loss": 1.4538, "step": 68 }, { "epoch": 0.00179170917184457, "grad_norm": 8.66399777717597, "learning_rate": 6.0035076887921665e-06, "loss": 1.4629, "step": 69 }, { "epoch": 0.0018176759714365204, "grad_norm": 11.238826713278984, "learning_rate": 6.02390937113811e-06, "loss": 1.4716, "step": 70 }, { "epoch": 0.0018436427710284708, "grad_norm": 7.419337302584475, "learning_rate": 6.0440216581446974e-06, "loss": 1.473, "step": 71 }, { "epoch": 0.001869609570620421, "grad_norm": 7.494123063652215, "learning_rate": 6.063852645188006e-06, "loss": 1.4674, "step": 72 }, { "epoch": 0.0018955763702123713, "grad_norm": 18.654266507572654, "learning_rate": 6.083410092641338e-06, "loss": 1.4646, "step": 73 }, { "epoch": 0.0019215431698043216, "grad_norm": 6.675700678821861, "learning_rate": 6.102701444108102e-06, "loss": 1.4292, "step": 74 }, { "epoch": 0.0019475099693962719, "grad_norm": 16.474908419225734, "learning_rate": 6.121733843430915e-06, "loss": 1.4084, "step": 75 }, { "epoch": 0.001973476768988222, "grad_norm": 8.395304648237534, "learning_rate": 6.140514150574159e-06, "loss": 1.3988, "step": 76 }, { "epoch": 0.0019994435685801726, "grad_norm": 17.2232361088312, "learning_rate": 6.159048956468369e-06, "loss": 1.496, "step": 77 }, { "epoch": 0.0020254103681721226, "grad_norm": 17.05890100550516, "learning_rate": 6.177344596896842e-06, "loss": 1.4618, "step": 78 }, { "epoch": 0.002051377167764073, "grad_norm": 8.844316336355204, "learning_rate": 6.1954071654976755e-06, "loss": 1.4066, "step": 79 }, { "epoch": 0.0020773439673560236, "grad_norm": 7.807838941601465, "learning_rate": 6.2132425259479926e-06, "loss": 1.4033, "step": 80 }, { "epoch": 0.0021033107669479736, "grad_norm": 7.118514472206657, "learning_rate": 6.230856323391335e-06, "loss": 1.4293, "step": 81 }, { "epoch": 0.002129277566539924, "grad_norm": 7.4210338236309505, "learning_rate": 6.2482539951639196e-06, "loss": 1.4293, "step": 82 }, { "epoch": 0.002155244366131874, "grad_norm": 8.614845379522835, "learning_rate": 6.265440780870785e-06, "loss": 1.4266, "step": 83 }, { "epoch": 0.0021812111657238246, "grad_norm": 13.527631949897033, "learning_rate": 6.282421731858516e-06, "loss": 1.4275, "step": 84 }, { "epoch": 0.0022071779653157746, "grad_norm": 10.933003400461018, "learning_rate": 6.299201720127428e-06, "loss": 1.4149, "step": 85 }, { "epoch": 0.002233144764907725, "grad_norm": 8.947921545688304, "learning_rate": 6.315785446722519e-06, "loss": 1.4006, "step": 86 }, { "epoch": 0.0022591115644996756, "grad_norm": 6.766296776243929, "learning_rate": 6.332177449639372e-06, "loss": 1.4137, "step": 87 }, { "epoch": 0.0022850783640916256, "grad_norm": 23.311286465147226, "learning_rate": 6.348382111278253e-06, "loss": 1.3642, "step": 88 }, { "epoch": 0.002311045163683576, "grad_norm": 7.53677507397063, "learning_rate": 6.364403665477058e-06, "loss": 1.3805, "step": 89 }, { "epoch": 0.002337011963275526, "grad_norm": 7.973492576174175, "learning_rate": 6.380246204151321e-06, "loss": 1.3802, "step": 90 }, { "epoch": 0.0023629787628674766, "grad_norm": 5.992278216494705, "learning_rate": 6.395913683567352e-06, "loss": 1.404, "step": 91 }, { "epoch": 0.002388945562459427, "grad_norm": 9.849752479192926, "learning_rate": 6.411409930272558e-06, "loss": 1.3895, "step": 92 }, { "epoch": 0.002414912362051377, "grad_norm": 9.836286543151555, "learning_rate": 6.426738646705193e-06, "loss": 1.3441, "step": 93 }, { "epoch": 0.0024408791616433276, "grad_norm": 7.4713873256950505, "learning_rate": 6.441903416504094e-06, "loss": 1.3697, "step": 94 }, { "epoch": 0.0024668459612352776, "grad_norm": 6.267332534043902, "learning_rate": 6.4569077095374745e-06, "loss": 1.4325, "step": 95 }, { "epoch": 0.002492812760827228, "grad_norm": 8.904193841597902, "learning_rate": 6.471754886668398e-06, "loss": 1.4295, "step": 96 }, { "epoch": 0.002518779560419178, "grad_norm": 6.633396226338773, "learning_rate": 6.486448204273347e-06, "loss": 1.3815, "step": 97 }, { "epoch": 0.0025447463600111286, "grad_norm": 7.783640299045919, "learning_rate": 6.500990818529026e-06, "loss": 1.386, "step": 98 }, { "epoch": 0.002570713159603079, "grad_norm": 8.83256033148245, "learning_rate": 6.515385789481581e-06, "loss": 1.3658, "step": 99 }, { "epoch": 0.002596679959195029, "grad_norm": 17.54175403902429, "learning_rate": 6.5296360849113085e-06, "loss": 1.4168, "step": 100 }, { "epoch": 0.0026226467587869796, "grad_norm": 17.514768635677953, "learning_rate": 6.543744584005075e-06, "loss": 1.3831, "step": 101 }, { "epoch": 0.0026486135583789296, "grad_norm": 11.530537404938231, "learning_rate": 6.5577140808478336e-06, "loss": 1.3836, "step": 102 }, { "epoch": 0.00267458035797088, "grad_norm": 15.291566825908028, "learning_rate": 6.571547287743812e-06, "loss": 1.4025, "step": 103 }, { "epoch": 0.0027005471575628306, "grad_norm": 13.018564906569484, "learning_rate": 6.585246838377235e-06, "loss": 1.3979, "step": 104 }, { "epoch": 0.0027265139571547806, "grad_norm": 10.037675077400776, "learning_rate": 6.598815290821831e-06, "loss": 1.3604, "step": 105 }, { "epoch": 0.002752480756746731, "grad_norm": 8.77763102069398, "learning_rate": 6.612255130407681e-06, "loss": 1.3627, "step": 106 }, { "epoch": 0.002778447556338681, "grad_norm": 7.253699838457306, "learning_rate": 6.625568772453486e-06, "loss": 1.323, "step": 107 }, { "epoch": 0.0028044143559306316, "grad_norm": 6.200582833736466, "learning_rate": 6.638758564871728e-06, "loss": 1.4246, "step": 108 }, { "epoch": 0.0028303811555225817, "grad_norm": 9.93854330092373, "learning_rate": 6.6518267906537845e-06, "loss": 1.3579, "step": 109 }, { "epoch": 0.002856347955114532, "grad_norm": 9.143808289848597, "learning_rate": 6.664775670241568e-06, "loss": 1.4106, "step": 110 }, { "epoch": 0.0028823147547064826, "grad_norm": 8.970037387867794, "learning_rate": 6.677607363791822e-06, "loss": 1.3676, "step": 111 }, { "epoch": 0.0029082815542984326, "grad_norm": 10.658336804240012, "learning_rate": 6.690323973338908e-06, "loss": 1.3303, "step": 112 }, { "epoch": 0.002934248353890383, "grad_norm": 6.462273197879613, "learning_rate": 6.70292754486144e-06, "loss": 1.3445, "step": 113 }, { "epoch": 0.002960215153482333, "grad_norm": 6.922270683879563, "learning_rate": 6.715420070257881e-06, "loss": 1.3677, "step": 114 }, { "epoch": 0.0029861819530742836, "grad_norm": 12.297885024434843, "learning_rate": 6.727803489235872e-06, "loss": 1.3731, "step": 115 }, { "epoch": 0.0030121487526662337, "grad_norm": 9.740242795326719, "learning_rate": 6.740079691119763e-06, "loss": 1.4205, "step": 116 }, { "epoch": 0.003038115552258184, "grad_norm": 9.065516318478803, "learning_rate": 6.752250516580564e-06, "loss": 1.3875, "step": 117 }, { "epoch": 0.0030640823518501346, "grad_norm": 9.111923253942726, "learning_rate": 6.764317759292276e-06, "loss": 1.3462, "step": 118 }, { "epoch": 0.0030900491514420847, "grad_norm": 7.087398424715804, "learning_rate": 6.776283167518344e-06, "loss": 1.3987, "step": 119 }, { "epoch": 0.003116015951034035, "grad_norm": 6.798203893079039, "learning_rate": 6.7881484456317135e-06, "loss": 1.3966, "step": 120 }, { "epoch": 0.003141982750625985, "grad_norm": 8.597048816428273, "learning_rate": 6.7999152555718276e-06, "loss": 1.3295, "step": 121 }, { "epoch": 0.0031679495502179356, "grad_norm": 9.601550978303706, "learning_rate": 6.811585218241642e-06, "loss": 1.3596, "step": 122 }, { "epoch": 0.003193916349809886, "grad_norm": 7.453082006558386, "learning_rate": 6.8231599148476405e-06, "loss": 1.3269, "step": 123 }, { "epoch": 0.003219883149401836, "grad_norm": 11.568309087610185, "learning_rate": 6.834640888185586e-06, "loss": 1.3935, "step": 124 }, { "epoch": 0.0032458499489937866, "grad_norm": 7.52277172466284, "learning_rate": 6.846029643874624e-06, "loss": 1.3854, "step": 125 }, { "epoch": 0.0032718167485857367, "grad_norm": 7.104760049181486, "learning_rate": 6.8573276515422375e-06, "loss": 1.4156, "step": 126 }, { "epoch": 0.003297783548177687, "grad_norm": 10.389672257189655, "learning_rate": 6.868536345962348e-06, "loss": 1.3725, "step": 127 }, { "epoch": 0.003323750347769637, "grad_norm": 9.29401873140395, "learning_rate": 6.879657128148791e-06, "loss": 1.3771, "step": 128 }, { "epoch": 0.0033497171473615877, "grad_norm": 27.153211136070077, "learning_rate": 6.8906913664062406e-06, "loss": 1.3813, "step": 129 }, { "epoch": 0.003375683946953538, "grad_norm": 7.2659788640305765, "learning_rate": 6.90164039734055e-06, "loss": 1.362, "step": 130 }, { "epoch": 0.003401650746545488, "grad_norm": 7.806757410935308, "learning_rate": 6.9125055268303975e-06, "loss": 1.3908, "step": 131 }, { "epoch": 0.0034276175461374386, "grad_norm": 6.417780923910124, "learning_rate": 6.923288030961973e-06, "loss": 1.3829, "step": 132 }, { "epoch": 0.0034535843457293887, "grad_norm": 19.56526179211872, "learning_rate": 6.933989156928389e-06, "loss": 1.3482, "step": 133 }, { "epoch": 0.003479551145321339, "grad_norm": 17.573846701579736, "learning_rate": 6.94461012389542e-06, "loss": 1.3563, "step": 134 }, { "epoch": 0.0035055179449132896, "grad_norm": 16.164281505282915, "learning_rate": 6.955152123835043e-06, "loss": 1.3702, "step": 135 }, { "epoch": 0.0035314847445052397, "grad_norm": 7.006035246607669, "learning_rate": 6.9656163223282265e-06, "loss": 1.3861, "step": 136 }, { "epoch": 0.00355745154409719, "grad_norm": 22.67827996837985, "learning_rate": 6.976003859338315e-06, "loss": 1.358, "step": 137 }, { "epoch": 0.00358341834368914, "grad_norm": 10.845342777475386, "learning_rate": 6.986315849956279e-06, "loss": 1.3868, "step": 138 }, { "epoch": 0.0036093851432810907, "grad_norm": 11.680227046484237, "learning_rate": 6.996553385119069e-06, "loss": 1.3601, "step": 139 }, { "epoch": 0.0036353519428730407, "grad_norm": 8.494801446256888, "learning_rate": 7.006717532302223e-06, "loss": 1.3897, "step": 140 }, { "epoch": 0.003661318742464991, "grad_norm": 6.680234009989669, "learning_rate": 7.0168093361878155e-06, "loss": 1.3977, "step": 141 }, { "epoch": 0.0036872855420569416, "grad_norm": 7.6763028110811335, "learning_rate": 7.026829819308811e-06, "loss": 1.3739, "step": 142 }, { "epoch": 0.0037132523416488917, "grad_norm": 14.889423461983792, "learning_rate": 7.036779982670809e-06, "loss": 1.3987, "step": 143 }, { "epoch": 0.003739219141240842, "grad_norm": 9.744161569492924, "learning_rate": 7.04666080635212e-06, "loss": 1.376, "step": 144 }, { "epoch": 0.003765185940832792, "grad_norm": 14.935510478942511, "learning_rate": 7.056473250083078e-06, "loss": 1.396, "step": 145 }, { "epoch": 0.0037911527404247427, "grad_norm": 7.908486272406978, "learning_rate": 7.066218253805451e-06, "loss": 1.3675, "step": 146 }, { "epoch": 0.003817119540016693, "grad_norm": 6.528226745893618, "learning_rate": 7.075896738212745e-06, "loss": 1.3674, "step": 147 }, { "epoch": 0.003843086339608643, "grad_norm": 14.668321400554357, "learning_rate": 7.085509605272214e-06, "loss": 1.3038, "step": 148 }, { "epoch": 0.0038690531392005937, "grad_norm": 7.779697493136657, "learning_rate": 7.095057738729268e-06, "loss": 1.3498, "step": 149 }, { "epoch": 0.0038950199387925437, "grad_norm": 7.679658938225753, "learning_rate": 7.104542004595029e-06, "loss": 1.3304, "step": 150 }, { "epoch": 0.003920986738384494, "grad_norm": 7.0238160571298085, "learning_rate": 7.1139632516176825e-06, "loss": 1.3435, "step": 151 }, { "epoch": 0.003946953537976444, "grad_norm": 11.49497477989533, "learning_rate": 7.123322311738273e-06, "loss": 1.3641, "step": 152 }, { "epoch": 0.003972920337568395, "grad_norm": 11.703728002030285, "learning_rate": 7.1326200005315545e-06, "loss": 1.334, "step": 153 }, { "epoch": 0.003998887137160345, "grad_norm": 23.317689865090223, "learning_rate": 7.141857117632483e-06, "loss": 1.3823, "step": 154 }, { "epoch": 0.004024853936752296, "grad_norm": 8.019241558180987, "learning_rate": 7.1510344471489e-06, "loss": 1.298, "step": 155 }, { "epoch": 0.004050820736344245, "grad_norm": 9.57004742936486, "learning_rate": 7.160152758060956e-06, "loss": 1.401, "step": 156 }, { "epoch": 0.004076787535936196, "grad_norm": 12.620595635868442, "learning_rate": 7.169212804607765e-06, "loss": 1.3685, "step": 157 }, { "epoch": 0.004102754335528146, "grad_norm": 14.83889209159024, "learning_rate": 7.1782153266617894e-06, "loss": 1.3579, "step": 158 }, { "epoch": 0.004128721135120097, "grad_norm": 9.262507462595046, "learning_rate": 7.187161050091403e-06, "loss": 1.3487, "step": 159 }, { "epoch": 0.004154687934712047, "grad_norm": 10.32604682547935, "learning_rate": 7.1960506871121065e-06, "loss": 1.3609, "step": 160 }, { "epoch": 0.004180654734303997, "grad_norm": 10.239856818509438, "learning_rate": 7.204884936626788e-06, "loss": 1.351, "step": 161 }, { "epoch": 0.004206621533895947, "grad_norm": 10.286713513778382, "learning_rate": 7.213664484555448e-06, "loss": 1.3406, "step": 162 }, { "epoch": 0.004232588333487898, "grad_norm": 7.978487893923655, "learning_rate": 7.222390004154792e-06, "loss": 1.3988, "step": 163 }, { "epoch": 0.004258555133079848, "grad_norm": 9.492382810490653, "learning_rate": 7.2310621563280335e-06, "loss": 1.3728, "step": 164 }, { "epoch": 0.004284521932671798, "grad_norm": 22.24914549140385, "learning_rate": 7.239681589925288e-06, "loss": 1.3557, "step": 165 }, { "epoch": 0.004310488732263748, "grad_norm": 16.38696317979202, "learning_rate": 7.248248942034898e-06, "loss": 1.3206, "step": 166 }, { "epoch": 0.004336455531855699, "grad_norm": 7.292030834005163, "learning_rate": 7.256764838265993e-06, "loss": 1.3565, "step": 167 }, { "epoch": 0.004362422331447649, "grad_norm": 8.101374624900906, "learning_rate": 7.265229893022629e-06, "loss": 1.3202, "step": 168 }, { "epoch": 0.0043883891310396, "grad_norm": 18.62859376376098, "learning_rate": 7.273644709769792e-06, "loss": 1.2802, "step": 169 }, { "epoch": 0.004414355930631549, "grad_norm": 9.860671065516012, "learning_rate": 7.282009881291542e-06, "loss": 1.3646, "step": 170 }, { "epoch": 0.0044403227302235, "grad_norm": 7.165928322069725, "learning_rate": 7.290325989941602e-06, "loss": 1.3744, "step": 171 }, { "epoch": 0.00446628952981545, "grad_norm": 10.13862992279656, "learning_rate": 7.298593607886632e-06, "loss": 1.3551, "step": 172 }, { "epoch": 0.004492256329407401, "grad_norm": 23.03111484441347, "learning_rate": 7.306813297342459e-06, "loss": 1.4089, "step": 173 }, { "epoch": 0.004518223128999351, "grad_norm": 7.336396126837946, "learning_rate": 7.314985610803485e-06, "loss": 1.3533, "step": 174 }, { "epoch": 0.004544189928591301, "grad_norm": 10.975889311540966, "learning_rate": 7.323111091265538e-06, "loss": 1.3359, "step": 175 }, { "epoch": 0.004570156728183251, "grad_norm": 18.203166912505395, "learning_rate": 7.331190272442366e-06, "loss": 1.3938, "step": 176 }, { "epoch": 0.004596123527775202, "grad_norm": 10.626390433492434, "learning_rate": 7.339223678975995e-06, "loss": 1.3613, "step": 177 }, { "epoch": 0.004622090327367152, "grad_norm": 8.335399778374882, "learning_rate": 7.347211826641172e-06, "loss": 1.3547, "step": 178 }, { "epoch": 0.004648057126959103, "grad_norm": 7.513147401762808, "learning_rate": 7.355155222544062e-06, "loss": 1.3884, "step": 179 }, { "epoch": 0.004674023926551052, "grad_norm": 23.357641292268386, "learning_rate": 7.363054365315435e-06, "loss": 1.3547, "step": 180 }, { "epoch": 0.004699990726143003, "grad_norm": 7.548544878660804, "learning_rate": 7.370909745298483e-06, "loss": 1.3339, "step": 181 }, { "epoch": 0.004725957525734953, "grad_norm": 7.7003669056855815, "learning_rate": 7.378721844731465e-06, "loss": 1.2954, "step": 182 }, { "epoch": 0.004751924325326904, "grad_norm": 8.491756256299768, "learning_rate": 7.386491137925362e-06, "loss": 1.2771, "step": 183 }, { "epoch": 0.004777891124918854, "grad_norm": 11.947803783908952, "learning_rate": 7.394218091436671e-06, "loss": 1.3229, "step": 184 }, { "epoch": 0.004803857924510804, "grad_norm": 16.63183820244135, "learning_rate": 7.401903164235529e-06, "loss": 1.3632, "step": 185 }, { "epoch": 0.004829824724102754, "grad_norm": 16.49252890012579, "learning_rate": 7.409546807869306e-06, "loss": 1.3492, "step": 186 }, { "epoch": 0.004855791523694705, "grad_norm": 12.450384319455702, "learning_rate": 7.417149466621801e-06, "loss": 1.3701, "step": 187 }, { "epoch": 0.004881758323286655, "grad_norm": 7.413463057794066, "learning_rate": 7.4247115776682076e-06, "loss": 1.3011, "step": 188 }, { "epoch": 0.004907725122878605, "grad_norm": 8.050823420710756, "learning_rate": 7.4322335712259576e-06, "loss": 1.4044, "step": 189 }, { "epoch": 0.004933691922470555, "grad_norm": 10.789497383297318, "learning_rate": 7.439715870701588e-06, "loss": 1.3529, "step": 190 }, { "epoch": 0.004959658722062506, "grad_norm": 29.81939553541715, "learning_rate": 7.447158892833755e-06, "loss": 1.3545, "step": 191 }, { "epoch": 0.004985625521654456, "grad_norm": 8.831806218799514, "learning_rate": 7.454563047832512e-06, "loss": 1.382, "step": 192 }, { "epoch": 0.005011592321246407, "grad_norm": 26.37743435152005, "learning_rate": 7.461928739514972e-06, "loss": 1.3224, "step": 193 }, { "epoch": 0.005037559120838356, "grad_norm": 9.06957521160958, "learning_rate": 7.469256365437459e-06, "loss": 1.4097, "step": 194 }, { "epoch": 0.005063525920430307, "grad_norm": 8.583746841494522, "learning_rate": 7.4765463170242706e-06, "loss": 1.3359, "step": 195 }, { "epoch": 0.005089492720022257, "grad_norm": 9.050779142614127, "learning_rate": 7.483798979693137e-06, "loss": 1.3384, "step": 196 }, { "epoch": 0.005115459519614208, "grad_norm": 7.649394551172773, "learning_rate": 7.491014732977504e-06, "loss": 1.3715, "step": 197 }, { "epoch": 0.005141426319206158, "grad_norm": 9.36104389270932, "learning_rate": 7.498193950645695e-06, "loss": 1.3808, "step": 198 }, { "epoch": 0.005167393118798108, "grad_norm": 10.646616051572883, "learning_rate": 7.505337000817097e-06, "loss": 1.3675, "step": 199 }, { "epoch": 0.005193359918390058, "grad_norm": 10.842199209173675, "learning_rate": 7.512444246075421e-06, "loss": 1.3477, "step": 200 }, { "epoch": 0.005219326717982009, "grad_norm": 9.077553919887066, "learning_rate": 7.519516043579141e-06, "loss": 1.3332, "step": 201 }, { "epoch": 0.005245293517573959, "grad_norm": 12.194503571625965, "learning_rate": 7.526552745169188e-06, "loss": 1.3095, "step": 202 }, { "epoch": 0.00527126031716591, "grad_norm": 6.785703260369773, "learning_rate": 7.533554697473993e-06, "loss": 1.3439, "step": 203 }, { "epoch": 0.005297227116757859, "grad_norm": 9.0612556392459, "learning_rate": 7.5405222420119475e-06, "loss": 1.3357, "step": 204 }, { "epoch": 0.00532319391634981, "grad_norm": 10.883029122260794, "learning_rate": 7.5474557152913486e-06, "loss": 1.3381, "step": 205 }, { "epoch": 0.00534916071594176, "grad_norm": 7.539056175110689, "learning_rate": 7.5543554489079255e-06, "loss": 1.3359, "step": 206 }, { "epoch": 0.005375127515533711, "grad_norm": 9.94351378998889, "learning_rate": 7.561221769639999e-06, "loss": 1.3504, "step": 207 }, { "epoch": 0.005401094315125661, "grad_norm": 10.110755979382025, "learning_rate": 7.568054999541348e-06, "loss": 1.3391, "step": 208 }, { "epoch": 0.005427061114717611, "grad_norm": 8.226666369239807, "learning_rate": 7.574855456031846e-06, "loss": 1.3481, "step": 209 }, { "epoch": 0.005453027914309561, "grad_norm": 17.50960342664, "learning_rate": 7.581623451985943e-06, "loss": 1.3117, "step": 210 }, { "epoch": 0.005478994713901512, "grad_norm": 12.655596530197865, "learning_rate": 7.588359295819035e-06, "loss": 1.3926, "step": 211 }, { "epoch": 0.005504961513493462, "grad_norm": 22.713369537823286, "learning_rate": 7.595063291571795e-06, "loss": 1.3602, "step": 212 }, { "epoch": 0.005530928313085412, "grad_norm": 18.895159130693205, "learning_rate": 7.6017357389925314e-06, "loss": 1.3178, "step": 213 }, { "epoch": 0.005556895112677362, "grad_norm": 15.828232971003901, "learning_rate": 7.6083769336176e-06, "loss": 1.3411, "step": 214 }, { "epoch": 0.005582861912269313, "grad_norm": 7.981957355122498, "learning_rate": 7.614987166849946e-06, "loss": 1.3366, "step": 215 }, { "epoch": 0.005608828711861263, "grad_norm": 55.26763054743821, "learning_rate": 7.621566726035842e-06, "loss": 1.3349, "step": 216 }, { "epoch": 0.005634795511453214, "grad_norm": 9.36976619856241, "learning_rate": 7.628115894539815e-06, "loss": 1.317, "step": 217 }, { "epoch": 0.005660762311045163, "grad_norm": 13.451327142862034, "learning_rate": 7.634634951817898e-06, "loss": 1.3679, "step": 218 }, { "epoch": 0.005686729110637114, "grad_norm": 123.50038746143673, "learning_rate": 7.641124173489172e-06, "loss": 1.3204, "step": 219 }, { "epoch": 0.005712695910229064, "grad_norm": 7.227459826046035, "learning_rate": 7.647583831405682e-06, "loss": 1.3339, "step": 220 }, { "epoch": 0.005738662709821015, "grad_norm": 20.714984810593634, "learning_rate": 7.654014193720782e-06, "loss": 1.3468, "step": 221 }, { "epoch": 0.005764629509412965, "grad_norm": 8.575404335031518, "learning_rate": 7.660415524955935e-06, "loss": 1.3678, "step": 222 }, { "epoch": 0.005790596309004915, "grad_norm": 7.160803266678416, "learning_rate": 7.66678808606599e-06, "loss": 1.3147, "step": 223 }, { "epoch": 0.005816563108596865, "grad_norm": 14.710955651708133, "learning_rate": 7.67313213450302e-06, "loss": 1.3462, "step": 224 }, { "epoch": 0.005842529908188816, "grad_norm": 9.41572169961027, "learning_rate": 7.679447924278749e-06, "loss": 1.31, "step": 225 }, { "epoch": 0.005868496707780766, "grad_norm": 7.9713889146903405, "learning_rate": 7.685735706025554e-06, "loss": 1.3749, "step": 226 }, { "epoch": 0.005894463507372717, "grad_norm": 6.4611887208376215, "learning_rate": 7.691995727056155e-06, "loss": 1.3525, "step": 227 }, { "epoch": 0.005920430306964666, "grad_norm": 10.622313395401079, "learning_rate": 7.698228231421993e-06, "loss": 1.3047, "step": 228 }, { "epoch": 0.005946397106556617, "grad_norm": 7.760333647486589, "learning_rate": 7.704433459970307e-06, "loss": 1.3254, "step": 229 }, { "epoch": 0.005972363906148567, "grad_norm": 5.8583065432084735, "learning_rate": 7.710611650399987e-06, "loss": 1.3149, "step": 230 }, { "epoch": 0.005998330705740518, "grad_norm": 8.20501605806982, "learning_rate": 7.716763037316203e-06, "loss": 1.2932, "step": 231 }, { "epoch": 0.006024297505332467, "grad_norm": 6.875382291395296, "learning_rate": 7.722887852283877e-06, "loss": 1.3203, "step": 232 }, { "epoch": 0.006050264304924418, "grad_norm": 26.869264922229217, "learning_rate": 7.72898632387997e-06, "loss": 1.336, "step": 233 }, { "epoch": 0.006076231104516368, "grad_norm": 9.363350102088319, "learning_rate": 7.735058677744676e-06, "loss": 1.4033, "step": 234 }, { "epoch": 0.006102197904108319, "grad_norm": 9.617671802514312, "learning_rate": 7.741105136631522e-06, "loss": 1.3055, "step": 235 }, { "epoch": 0.006128164703700269, "grad_norm": 8.310921655400623, "learning_rate": 7.747125920456388e-06, "loss": 1.3452, "step": 236 }, { "epoch": 0.006154131503292219, "grad_norm": 67.017805324173, "learning_rate": 7.75312124634551e-06, "loss": 1.3431, "step": 237 }, { "epoch": 0.006180098302884169, "grad_norm": 9.122703367743568, "learning_rate": 7.759091328682457e-06, "loss": 1.3231, "step": 238 }, { "epoch": 0.00620606510247612, "grad_norm": 6.3775782140035275, "learning_rate": 7.765036379154136e-06, "loss": 1.3677, "step": 239 }, { "epoch": 0.00623203190206807, "grad_norm": 17.411403803883104, "learning_rate": 7.770956606795828e-06, "loss": 1.2984, "step": 240 }, { "epoch": 0.006257998701660021, "grad_norm": 10.143552007993412, "learning_rate": 7.776852218035288e-06, "loss": 1.3445, "step": 241 }, { "epoch": 0.00628396550125197, "grad_norm": 10.123678238860833, "learning_rate": 7.782723416735941e-06, "loss": 1.375, "step": 242 }, { "epoch": 0.006309932300843921, "grad_norm": 6.371850526022763, "learning_rate": 7.788570404239169e-06, "loss": 1.3653, "step": 243 }, { "epoch": 0.006335899100435871, "grad_norm": 5.89328177767566, "learning_rate": 7.794393379405754e-06, "loss": 1.3096, "step": 244 }, { "epoch": 0.006361865900027822, "grad_norm": 8.603886511231103, "learning_rate": 7.800192538656454e-06, "loss": 1.3078, "step": 245 }, { "epoch": 0.006387832699619772, "grad_norm": 8.645650716178183, "learning_rate": 7.805968076011753e-06, "loss": 1.2719, "step": 246 }, { "epoch": 0.006413799499211722, "grad_norm": 8.08649558385069, "learning_rate": 7.81172018313083e-06, "loss": 1.3295, "step": 247 }, { "epoch": 0.006439766298803672, "grad_norm": 11.904470395833384, "learning_rate": 7.817449049349698e-06, "loss": 1.2976, "step": 248 }, { "epoch": 0.006465733098395623, "grad_norm": 13.380287337475172, "learning_rate": 7.823154861718619e-06, "loss": 1.3463, "step": 249 }, { "epoch": 0.006491699897987573, "grad_norm": 10.258324568925755, "learning_rate": 7.828837805038734e-06, "loss": 1.3248, "step": 250 }, { "epoch": 0.006517666697579524, "grad_norm": 6.645810245804659, "learning_rate": 7.834498061897997e-06, "loss": 1.3229, "step": 251 }, { "epoch": 0.006543633497171473, "grad_norm": 12.370636758412791, "learning_rate": 7.84013581270635e-06, "loss": 1.3401, "step": 252 }, { "epoch": 0.006569600296763424, "grad_norm": 6.798062205493258, "learning_rate": 7.845751235730246e-06, "loss": 1.3291, "step": 253 }, { "epoch": 0.006595567096355374, "grad_norm": 8.803854739622139, "learning_rate": 7.851344507126461e-06, "loss": 1.3397, "step": 254 }, { "epoch": 0.006621533895947325, "grad_norm": 13.118091327184736, "learning_rate": 7.856915800975262e-06, "loss": 1.3658, "step": 255 }, { "epoch": 0.006647500695539274, "grad_norm": 8.944820367193921, "learning_rate": 7.862465289312904e-06, "loss": 1.3595, "step": 256 }, { "epoch": 0.006673467495131225, "grad_norm": 9.043443730776387, "learning_rate": 7.867993142163519e-06, "loss": 1.3599, "step": 257 }, { "epoch": 0.006699434294723175, "grad_norm": 9.929264902986116, "learning_rate": 7.873499527570354e-06, "loss": 1.3672, "step": 258 }, { "epoch": 0.006725401094315126, "grad_norm": 11.242870190789944, "learning_rate": 7.878984611626443e-06, "loss": 1.3362, "step": 259 }, { "epoch": 0.006751367893907076, "grad_norm": 9.896732415297485, "learning_rate": 7.884448558504662e-06, "loss": 1.4016, "step": 260 }, { "epoch": 0.006777334693499026, "grad_norm": 8.787737747383957, "learning_rate": 7.889891530487207e-06, "loss": 1.3494, "step": 261 }, { "epoch": 0.006803301493090976, "grad_norm": 9.270952156819869, "learning_rate": 7.895313687994511e-06, "loss": 1.2866, "step": 262 }, { "epoch": 0.006829268292682927, "grad_norm": 8.802471108707424, "learning_rate": 7.90071518961364e-06, "loss": 1.372, "step": 263 }, { "epoch": 0.006855235092274877, "grad_norm": 9.18034073473634, "learning_rate": 7.906096192126086e-06, "loss": 1.3346, "step": 264 }, { "epoch": 0.006881201891866828, "grad_norm": 6.511979434831468, "learning_rate": 7.91145685053511e-06, "loss": 1.3295, "step": 265 }, { "epoch": 0.006907168691458777, "grad_norm": 7.1220536414217674, "learning_rate": 7.916797318092504e-06, "loss": 1.3226, "step": 266 }, { "epoch": 0.006933135491050728, "grad_norm": 14.632575787260487, "learning_rate": 7.922117746324892e-06, "loss": 1.2882, "step": 267 }, { "epoch": 0.006959102290642678, "grad_norm": 12.03047252966854, "learning_rate": 7.927418285059532e-06, "loss": 1.3179, "step": 268 }, { "epoch": 0.006985069090234629, "grad_norm": 7.132617559202905, "learning_rate": 7.932699082449625e-06, "loss": 1.3161, "step": 269 }, { "epoch": 0.007011035889826579, "grad_norm": 8.880372046847457, "learning_rate": 7.937960284999156e-06, "loss": 1.3506, "step": 270 }, { "epoch": 0.007037002689418529, "grad_norm": 8.621517196923953, "learning_rate": 7.943202037587298e-06, "loss": 1.3627, "step": 271 }, { "epoch": 0.007062969489010479, "grad_norm": 18.460590347984372, "learning_rate": 7.948424483492341e-06, "loss": 1.3132, "step": 272 }, { "epoch": 0.00708893628860243, "grad_norm": 21.72006586054306, "learning_rate": 7.953627764415187e-06, "loss": 1.2672, "step": 273 }, { "epoch": 0.00711490308819438, "grad_norm": 6.496909066863696, "learning_rate": 7.958812020502428e-06, "loss": 1.302, "step": 274 }, { "epoch": 0.00714086988778633, "grad_norm": 7.720008407746318, "learning_rate": 7.963977390368995e-06, "loss": 1.3412, "step": 275 }, { "epoch": 0.00716683668737828, "grad_norm": 7.97355596688705, "learning_rate": 7.969124011120391e-06, "loss": 1.2785, "step": 276 }, { "epoch": 0.007192803486970231, "grad_norm": 7.065621825415309, "learning_rate": 7.974252018374531e-06, "loss": 1.312, "step": 277 }, { "epoch": 0.007218770286562181, "grad_norm": 6.136903678686395, "learning_rate": 7.979361546283181e-06, "loss": 1.3367, "step": 278 }, { "epoch": 0.007244737086154132, "grad_norm": 6.6185127021245185, "learning_rate": 7.984452727553026e-06, "loss": 1.3175, "step": 279 }, { "epoch": 0.007270703885746081, "grad_norm": 18.039243516307025, "learning_rate": 7.989525693466335e-06, "loss": 1.3621, "step": 280 }, { "epoch": 0.007296670685338032, "grad_norm": 6.549995352474049, "learning_rate": 7.99458057390129e-06, "loss": 1.2807, "step": 281 }, { "epoch": 0.007322637484929982, "grad_norm": 6.850537422427106, "learning_rate": 7.99961749735193e-06, "loss": 1.3491, "step": 282 }, { "epoch": 0.007348604284521933, "grad_norm": 35.012267660332284, "learning_rate": 8.004636590947738e-06, "loss": 1.3712, "step": 283 }, { "epoch": 0.007374571084113883, "grad_norm": 8.012232012142825, "learning_rate": 8.009637980472924e-06, "loss": 1.3338, "step": 284 }, { "epoch": 0.007400537883705833, "grad_norm": 7.53472668188132, "learning_rate": 8.014621790385308e-06, "loss": 1.3488, "step": 285 }, { "epoch": 0.007426504683297783, "grad_norm": 6.701185821812214, "learning_rate": 8.019588143834921e-06, "loss": 1.3261, "step": 286 }, { "epoch": 0.007452471482889734, "grad_norm": 13.09963310834476, "learning_rate": 8.024537162682263e-06, "loss": 1.3009, "step": 287 }, { "epoch": 0.007478438282481684, "grad_norm": 8.611267214218124, "learning_rate": 8.029468967516232e-06, "loss": 1.3634, "step": 288 }, { "epoch": 0.007504405082073635, "grad_norm": 10.418050770674501, "learning_rate": 8.034383677671775e-06, "loss": 1.3336, "step": 289 }, { "epoch": 0.007530371881665584, "grad_norm": 8.163813449401028, "learning_rate": 8.039281411247192e-06, "loss": 1.3103, "step": 290 }, { "epoch": 0.007556338681257535, "grad_norm": 10.877005556582644, "learning_rate": 8.044162285121181e-06, "loss": 1.2917, "step": 291 }, { "epoch": 0.007582305480849485, "grad_norm": 11.176676879963095, "learning_rate": 8.049026414969564e-06, "loss": 1.337, "step": 292 }, { "epoch": 0.007608272280441436, "grad_norm": 9.174352671891471, "learning_rate": 8.053873915281741e-06, "loss": 1.2822, "step": 293 }, { "epoch": 0.007634239080033386, "grad_norm": 9.914802397405095, "learning_rate": 8.05870489937686e-06, "loss": 1.2873, "step": 294 }, { "epoch": 0.007660205879625336, "grad_norm": 7.7322134369007856, "learning_rate": 8.063519479419703e-06, "loss": 1.3384, "step": 295 }, { "epoch": 0.007686172679217286, "grad_norm": 28.19253393589449, "learning_rate": 8.068317766436328e-06, "loss": 1.3465, "step": 296 }, { "epoch": 0.007712139478809237, "grad_norm": 13.530092960779823, "learning_rate": 8.073099870329416e-06, "loss": 1.2604, "step": 297 }, { "epoch": 0.007738106278401187, "grad_norm": 48.151400530481716, "learning_rate": 8.077865899893382e-06, "loss": 1.3166, "step": 298 }, { "epoch": 0.007764073077993137, "grad_norm": 21.928781543007297, "learning_rate": 8.082615962829228e-06, "loss": 1.3289, "step": 299 }, { "epoch": 0.007790039877585087, "grad_norm": 7.46148813822125, "learning_rate": 8.087350165759142e-06, "loss": 1.3655, "step": 300 }, { "epoch": 0.007816006677177039, "grad_norm": 9.11497594950522, "learning_rate": 8.092068614240862e-06, "loss": 1.3542, "step": 301 }, { "epoch": 0.007841973476768987, "grad_norm": 30.764909804726695, "learning_rate": 8.096771412781796e-06, "loss": 1.3151, "step": 302 }, { "epoch": 0.007867940276360938, "grad_norm": 16.4965650277518, "learning_rate": 8.101458664852908e-06, "loss": 1.3254, "step": 303 }, { "epoch": 0.007893907075952888, "grad_norm": 7.479352376716694, "learning_rate": 8.106130472902386e-06, "loss": 1.2903, "step": 304 }, { "epoch": 0.007919873875544839, "grad_norm": 7.482740506462893, "learning_rate": 8.11078693836907e-06, "loss": 1.341, "step": 305 }, { "epoch": 0.00794584067513679, "grad_norm": 8.554184934581409, "learning_rate": 8.115428161695668e-06, "loss": 1.2741, "step": 306 }, { "epoch": 0.00797180747472874, "grad_norm": 12.29465302649864, "learning_rate": 8.120054242341762e-06, "loss": 1.4047, "step": 307 }, { "epoch": 0.00799777427432069, "grad_norm": 8.519258560371146, "learning_rate": 8.124665278796595e-06, "loss": 1.3364, "step": 308 }, { "epoch": 0.00802374107391264, "grad_norm": 9.562780438468483, "learning_rate": 8.129261368591647e-06, "loss": 1.3074, "step": 309 }, { "epoch": 0.008049707873504591, "grad_norm": 9.465584908945512, "learning_rate": 8.133842608313013e-06, "loss": 1.3066, "step": 310 }, { "epoch": 0.00807567467309654, "grad_norm": 11.084808442659224, "learning_rate": 8.138409093613595e-06, "loss": 1.3473, "step": 311 }, { "epoch": 0.00810164147268849, "grad_norm": 7.739093366253686, "learning_rate": 8.14296091922507e-06, "loss": 1.3603, "step": 312 }, { "epoch": 0.008127608272280441, "grad_norm": 7.3436679267355265, "learning_rate": 8.147498178969687e-06, "loss": 1.3074, "step": 313 }, { "epoch": 0.008153575071872391, "grad_norm": 6.684921765893247, "learning_rate": 8.152020965771878e-06, "loss": 1.284, "step": 314 }, { "epoch": 0.008179541871464342, "grad_norm": 9.769524612027933, "learning_rate": 8.156529371669665e-06, "loss": 1.3036, "step": 315 }, { "epoch": 0.008205508671056292, "grad_norm": 6.773991088068427, "learning_rate": 8.161023487825902e-06, "loss": 1.3137, "step": 316 }, { "epoch": 0.008231475470648243, "grad_norm": 12.179017230193878, "learning_rate": 8.16550340453934e-06, "loss": 1.3228, "step": 317 }, { "epoch": 0.008257442270240193, "grad_norm": 14.197717424068946, "learning_rate": 8.169969211255516e-06, "loss": 1.347, "step": 318 }, { "epoch": 0.008283409069832144, "grad_norm": 11.358211662403935, "learning_rate": 8.174420996577452e-06, "loss": 1.3583, "step": 319 }, { "epoch": 0.008309375869424094, "grad_norm": 16.21989586812, "learning_rate": 8.17885884827622e-06, "loss": 1.2989, "step": 320 }, { "epoch": 0.008335342669016043, "grad_norm": 15.520815746459554, "learning_rate": 8.18328285330132e-06, "loss": 1.3354, "step": 321 }, { "epoch": 0.008361309468607993, "grad_norm": 10.815957988296267, "learning_rate": 8.187693097790901e-06, "loss": 1.368, "step": 322 }, { "epoch": 0.008387276268199944, "grad_norm": 10.876482983027103, "learning_rate": 8.192089667081821e-06, "loss": 1.3192, "step": 323 }, { "epoch": 0.008413243067791894, "grad_norm": 16.671686459219753, "learning_rate": 8.19647264571956e-06, "loss": 1.3481, "step": 324 }, { "epoch": 0.008439209867383845, "grad_norm": 12.076490126254866, "learning_rate": 8.200842117467977e-06, "loss": 1.4066, "step": 325 }, { "epoch": 0.008465176666975795, "grad_norm": 8.878343409074304, "learning_rate": 8.205198165318907e-06, "loss": 1.3239, "step": 326 }, { "epoch": 0.008491143466567746, "grad_norm": 17.014450214840362, "learning_rate": 8.20954087150162e-06, "loss": 1.368, "step": 327 }, { "epoch": 0.008517110266159696, "grad_norm": 7.500569385531857, "learning_rate": 8.213870317492146e-06, "loss": 1.2655, "step": 328 }, { "epoch": 0.008543077065751647, "grad_norm": 8.38835513990228, "learning_rate": 8.218186584022438e-06, "loss": 1.2958, "step": 329 }, { "epoch": 0.008569043865343596, "grad_norm": 11.806358072766674, "learning_rate": 8.222489751089401e-06, "loss": 1.3349, "step": 330 }, { "epoch": 0.008595010664935546, "grad_norm": 11.669761537862971, "learning_rate": 8.2267798979638e-06, "loss": 1.2796, "step": 331 }, { "epoch": 0.008620977464527496, "grad_norm": 6.457679305236312, "learning_rate": 8.23105710319901e-06, "loss": 1.3004, "step": 332 }, { "epoch": 0.008646944264119447, "grad_norm": 7.702922454709103, "learning_rate": 8.235321444639656e-06, "loss": 1.3333, "step": 333 }, { "epoch": 0.008672911063711397, "grad_norm": 9.694037612593371, "learning_rate": 8.239572999430105e-06, "loss": 1.3713, "step": 334 }, { "epoch": 0.008698877863303348, "grad_norm": 6.260321067888362, "learning_rate": 8.243811844022847e-06, "loss": 1.3499, "step": 335 }, { "epoch": 0.008724844662895298, "grad_norm": 6.271144043922839, "learning_rate": 8.248038054186743e-06, "loss": 1.3211, "step": 336 }, { "epoch": 0.008750811462487249, "grad_norm": 8.871865162133503, "learning_rate": 8.252251705015143e-06, "loss": 1.3437, "step": 337 }, { "epoch": 0.0087767782620792, "grad_norm": 8.427454867540591, "learning_rate": 8.256452870933905e-06, "loss": 1.3057, "step": 338 }, { "epoch": 0.00880274506167115, "grad_norm": 5.303857392881011, "learning_rate": 8.260641625709273e-06, "loss": 1.3295, "step": 339 }, { "epoch": 0.008828711861263099, "grad_norm": 10.905614796151166, "learning_rate": 8.264818042455655e-06, "loss": 1.3395, "step": 340 }, { "epoch": 0.008854678660855049, "grad_norm": 6.666194453999609, "learning_rate": 8.268982193643272e-06, "loss": 1.3235, "step": 341 }, { "epoch": 0.008880645460447, "grad_norm": 6.029204501033549, "learning_rate": 8.273134151105714e-06, "loss": 1.3478, "step": 342 }, { "epoch": 0.00890661226003895, "grad_norm": 7.000323696826566, "learning_rate": 8.277273986047367e-06, "loss": 1.3205, "step": 343 }, { "epoch": 0.0089325790596309, "grad_norm": 8.103916925996776, "learning_rate": 8.281401769050745e-06, "loss": 1.3429, "step": 344 }, { "epoch": 0.008958545859222851, "grad_norm": 7.185255147907229, "learning_rate": 8.285517570083706e-06, "loss": 1.2911, "step": 345 }, { "epoch": 0.008984512658814801, "grad_norm": 7.380007837803098, "learning_rate": 8.289621458506573e-06, "loss": 1.3415, "step": 346 }, { "epoch": 0.009010479458406752, "grad_norm": 8.760110735607036, "learning_rate": 8.293713503079141e-06, "loss": 1.3316, "step": 347 }, { "epoch": 0.009036446257998702, "grad_norm": 12.26091295835125, "learning_rate": 8.297793771967598e-06, "loss": 1.3194, "step": 348 }, { "epoch": 0.009062413057590653, "grad_norm": 7.346531377794021, "learning_rate": 8.301862332751332e-06, "loss": 1.3705, "step": 349 }, { "epoch": 0.009088379857182602, "grad_norm": 6.2407475458986985, "learning_rate": 8.30591925242965e-06, "loss": 1.2924, "step": 350 }, { "epoch": 0.009114346656774552, "grad_norm": 11.724512255188712, "learning_rate": 8.309964597428397e-06, "loss": 1.3795, "step": 351 }, { "epoch": 0.009140313456366502, "grad_norm": 12.390461212268669, "learning_rate": 8.313998433606479e-06, "loss": 1.2961, "step": 352 }, { "epoch": 0.009166280255958453, "grad_norm": 8.084909449290242, "learning_rate": 8.3180208262623e-06, "loss": 1.3388, "step": 353 }, { "epoch": 0.009192247055550403, "grad_norm": 8.476629104001004, "learning_rate": 8.322031840140109e-06, "loss": 1.3583, "step": 354 }, { "epoch": 0.009218213855142354, "grad_norm": 11.72318340617448, "learning_rate": 8.326031539436239e-06, "loss": 1.3318, "step": 355 }, { "epoch": 0.009244180654734304, "grad_norm": 17.19596471247188, "learning_rate": 8.330019987805285e-06, "loss": 1.3348, "step": 356 }, { "epoch": 0.009270147454326255, "grad_norm": 9.533497029568581, "learning_rate": 8.333997248366176e-06, "loss": 1.2795, "step": 357 }, { "epoch": 0.009296114253918205, "grad_norm": 25.8323416570833, "learning_rate": 8.337963383708176e-06, "loss": 1.3164, "step": 358 }, { "epoch": 0.009322081053510154, "grad_norm": 10.029549617981397, "learning_rate": 8.341918455896776e-06, "loss": 1.2734, "step": 359 }, { "epoch": 0.009348047853102105, "grad_norm": 8.751897987510427, "learning_rate": 8.345862526479548e-06, "loss": 1.3325, "step": 360 }, { "epoch": 0.009374014652694055, "grad_norm": 8.871129011103413, "learning_rate": 8.349795656491867e-06, "loss": 1.345, "step": 361 }, { "epoch": 0.009399981452286005, "grad_norm": 7.675148570458639, "learning_rate": 8.353717906462596e-06, "loss": 1.3233, "step": 362 }, { "epoch": 0.009425948251877956, "grad_norm": 11.736448224183725, "learning_rate": 8.35762933641966e-06, "loss": 1.2794, "step": 363 }, { "epoch": 0.009451915051469906, "grad_norm": 6.15664160397889, "learning_rate": 8.361530005895578e-06, "loss": 1.2959, "step": 364 }, { "epoch": 0.009477881851061857, "grad_norm": 7.896612150756339, "learning_rate": 8.365419973932877e-06, "loss": 1.3352, "step": 365 }, { "epoch": 0.009503848650653807, "grad_norm": 5.725559870234239, "learning_rate": 8.369299299089475e-06, "loss": 1.312, "step": 366 }, { "epoch": 0.009529815450245758, "grad_norm": 9.00910489156751, "learning_rate": 8.373168039443953e-06, "loss": 1.2983, "step": 367 }, { "epoch": 0.009555782249837708, "grad_norm": 6.023136670706045, "learning_rate": 8.377026252600784e-06, "loss": 1.3382, "step": 368 }, { "epoch": 0.009581749049429657, "grad_norm": 6.440249550290697, "learning_rate": 8.380873995695474e-06, "loss": 1.3093, "step": 369 }, { "epoch": 0.009607715849021608, "grad_norm": 6.385236114222794, "learning_rate": 8.384711325399641e-06, "loss": 1.3444, "step": 370 }, { "epoch": 0.009633682648613558, "grad_norm": 9.37516316589781, "learning_rate": 8.388538297926024e-06, "loss": 1.2719, "step": 371 }, { "epoch": 0.009659649448205508, "grad_norm": 10.452128318632534, "learning_rate": 8.392354969033419e-06, "loss": 1.2962, "step": 372 }, { "epoch": 0.009685616247797459, "grad_norm": 7.880888901741465, "learning_rate": 8.396161394031556e-06, "loss": 1.3398, "step": 373 }, { "epoch": 0.00971158304738941, "grad_norm": 6.493042082837747, "learning_rate": 8.399957627785914e-06, "loss": 1.3446, "step": 374 }, { "epoch": 0.00973754984698136, "grad_norm": 11.569726283891926, "learning_rate": 8.403743724722457e-06, "loss": 1.3247, "step": 375 }, { "epoch": 0.00976351664657331, "grad_norm": 6.853232029167367, "learning_rate": 8.40751973883232e-06, "loss": 1.2906, "step": 376 }, { "epoch": 0.00978948344616526, "grad_norm": 13.805150074486752, "learning_rate": 8.411285723676433e-06, "loss": 1.2665, "step": 377 }, { "epoch": 0.00981545024575721, "grad_norm": 17.736827852123373, "learning_rate": 8.41504173239007e-06, "loss": 1.3635, "step": 378 }, { "epoch": 0.00984141704534916, "grad_norm": 10.441537772740675, "learning_rate": 8.418787817687356e-06, "loss": 1.3385, "step": 379 }, { "epoch": 0.00986738384494111, "grad_norm": 7.712042361242046, "learning_rate": 8.422524031865701e-06, "loss": 1.3786, "step": 380 }, { "epoch": 0.009893350644533061, "grad_norm": 5.769924307976265, "learning_rate": 8.426250426810182e-06, "loss": 1.3124, "step": 381 }, { "epoch": 0.009919317444125011, "grad_norm": 6.186067049192837, "learning_rate": 8.429967053997868e-06, "loss": 1.3281, "step": 382 }, { "epoch": 0.009945284243716962, "grad_norm": 8.321126720723807, "learning_rate": 8.433673964502083e-06, "loss": 1.3269, "step": 383 }, { "epoch": 0.009971251043308912, "grad_norm": 5.898701036641689, "learning_rate": 8.437371208996625e-06, "loss": 1.3661, "step": 384 }, { "epoch": 0.009997217842900863, "grad_norm": 7.703732108745755, "learning_rate": 8.441058837759911e-06, "loss": 1.2866, "step": 385 }, { "epoch": 0.010023184642492813, "grad_norm": 6.797432010671182, "learning_rate": 8.444736900679085e-06, "loss": 1.2959, "step": 386 }, { "epoch": 0.010049151442084764, "grad_norm": 8.701968213769632, "learning_rate": 8.448405447254075e-06, "loss": 1.3062, "step": 387 }, { "epoch": 0.010075118241676713, "grad_norm": 17.26909078620039, "learning_rate": 8.452064526601574e-06, "loss": 1.3395, "step": 388 }, { "epoch": 0.010101085041268663, "grad_norm": 11.113619694978292, "learning_rate": 8.455714187458997e-06, "loss": 1.2838, "step": 389 }, { "epoch": 0.010127051840860614, "grad_norm": 14.090985147489079, "learning_rate": 8.459354478188384e-06, "loss": 1.305, "step": 390 }, { "epoch": 0.010153018640452564, "grad_norm": 8.184098097525583, "learning_rate": 8.462985446780219e-06, "loss": 1.3261, "step": 391 }, { "epoch": 0.010178985440044514, "grad_norm": 7.603606181630663, "learning_rate": 8.466607140857251e-06, "loss": 1.3311, "step": 392 }, { "epoch": 0.010204952239636465, "grad_norm": 8.722347050750185, "learning_rate": 8.470219607678232e-06, "loss": 1.3135, "step": 393 }, { "epoch": 0.010230919039228415, "grad_norm": 20.83685875119346, "learning_rate": 8.473822894141617e-06, "loss": 1.302, "step": 394 }, { "epoch": 0.010256885838820366, "grad_norm": 6.976740477065645, "learning_rate": 8.477417046789217e-06, "loss": 1.4206, "step": 395 }, { "epoch": 0.010282852638412316, "grad_norm": 10.933475473301844, "learning_rate": 8.481002111809807e-06, "loss": 1.2899, "step": 396 }, { "epoch": 0.010308819438004265, "grad_norm": 8.75623034325952, "learning_rate": 8.484578135042691e-06, "loss": 1.3136, "step": 397 }, { "epoch": 0.010334786237596216, "grad_norm": 7.781121310353383, "learning_rate": 8.48814516198121e-06, "loss": 1.2797, "step": 398 }, { "epoch": 0.010360753037188166, "grad_norm": 7.018769305127057, "learning_rate": 8.491703237776225e-06, "loss": 1.3303, "step": 399 }, { "epoch": 0.010386719836780117, "grad_norm": 8.044161025944726, "learning_rate": 8.495252407239533e-06, "loss": 1.318, "step": 400 }, { "epoch": 0.010412686636372067, "grad_norm": 6.36698619313673, "learning_rate": 8.498792714847276e-06, "loss": 1.2821, "step": 401 }, { "epoch": 0.010438653435964017, "grad_norm": 10.383729869767452, "learning_rate": 8.502324204743254e-06, "loss": 1.3111, "step": 402 }, { "epoch": 0.010464620235555968, "grad_norm": 19.929967719021775, "learning_rate": 8.505846920742254e-06, "loss": 1.3135, "step": 403 }, { "epoch": 0.010490587035147918, "grad_norm": 20.099977468601335, "learning_rate": 8.509360906333301e-06, "loss": 1.328, "step": 404 }, { "epoch": 0.010516553834739869, "grad_norm": 9.359546589751165, "learning_rate": 8.512866204682875e-06, "loss": 1.292, "step": 405 }, { "epoch": 0.01054252063433182, "grad_norm": 21.01337624469642, "learning_rate": 8.516362858638106e-06, "loss": 1.3853, "step": 406 }, { "epoch": 0.010568487433923768, "grad_norm": 7.672721986961927, "learning_rate": 8.5198509107299e-06, "loss": 1.3317, "step": 407 }, { "epoch": 0.010594454233515719, "grad_norm": 5.951840399766034, "learning_rate": 8.52333040317606e-06, "loss": 1.3225, "step": 408 }, { "epoch": 0.010620421033107669, "grad_norm": 7.012137924883387, "learning_rate": 8.526801377884334e-06, "loss": 1.3224, "step": 409 }, { "epoch": 0.01064638783269962, "grad_norm": 7.990295007103124, "learning_rate": 8.53026387645546e-06, "loss": 1.3227, "step": 410 }, { "epoch": 0.01067235463229157, "grad_norm": 7.019355716537877, "learning_rate": 8.533717940186149e-06, "loss": 1.3206, "step": 411 }, { "epoch": 0.01069832143188352, "grad_norm": 13.27508760987304, "learning_rate": 8.537163610072039e-06, "loss": 1.3406, "step": 412 }, { "epoch": 0.010724288231475471, "grad_norm": 7.280744150296732, "learning_rate": 8.54060092681062e-06, "loss": 1.3355, "step": 413 }, { "epoch": 0.010750255031067421, "grad_norm": 10.454443057316437, "learning_rate": 8.544029930804112e-06, "loss": 1.3307, "step": 414 }, { "epoch": 0.010776221830659372, "grad_norm": 7.501350789816125, "learning_rate": 8.547450662162324e-06, "loss": 1.3074, "step": 415 }, { "epoch": 0.010802188630251322, "grad_norm": 6.930846726954727, "learning_rate": 8.550863160705462e-06, "loss": 1.3242, "step": 416 }, { "epoch": 0.010828155429843271, "grad_norm": 9.40454288846183, "learning_rate": 8.554267465966904e-06, "loss": 1.3099, "step": 417 }, { "epoch": 0.010854122229435222, "grad_norm": 9.64250948647588, "learning_rate": 8.557663617195961e-06, "loss": 1.3177, "step": 418 }, { "epoch": 0.010880089029027172, "grad_norm": 10.216995321923326, "learning_rate": 8.561051653360584e-06, "loss": 1.3154, "step": 419 }, { "epoch": 0.010906055828619123, "grad_norm": 10.533806646151053, "learning_rate": 8.564431613150058e-06, "loss": 1.2982, "step": 420 }, { "epoch": 0.010932022628211073, "grad_norm": 8.249024304579242, "learning_rate": 8.567803534977628e-06, "loss": 1.3643, "step": 421 }, { "epoch": 0.010957989427803023, "grad_norm": 6.821933199834766, "learning_rate": 8.571167456983147e-06, "loss": 1.265, "step": 422 }, { "epoch": 0.010983956227394974, "grad_norm": 11.64364539778733, "learning_rate": 8.574523417035649e-06, "loss": 1.3597, "step": 423 }, { "epoch": 0.011009923026986924, "grad_norm": 8.547013360111805, "learning_rate": 8.577871452735907e-06, "loss": 1.3389, "step": 424 }, { "epoch": 0.011035889826578875, "grad_norm": 12.51575728023884, "learning_rate": 8.58121160141897e-06, "loss": 1.2996, "step": 425 }, { "epoch": 0.011061856626170824, "grad_norm": 6.503887187941285, "learning_rate": 8.584543900156646e-06, "loss": 1.3073, "step": 426 }, { "epoch": 0.011087823425762774, "grad_norm": 8.662711145599271, "learning_rate": 8.587868385759984e-06, "loss": 1.3097, "step": 427 }, { "epoch": 0.011113790225354725, "grad_norm": 6.834107455932946, "learning_rate": 8.591185094781712e-06, "loss": 1.2679, "step": 428 }, { "epoch": 0.011139757024946675, "grad_norm": 6.4298619261729435, "learning_rate": 8.594494063518644e-06, "loss": 1.2943, "step": 429 }, { "epoch": 0.011165723824538626, "grad_norm": 8.70481409635774, "learning_rate": 8.59779532801406e-06, "loss": 1.326, "step": 430 }, { "epoch": 0.011191690624130576, "grad_norm": 5.903954438258581, "learning_rate": 8.601088924060076e-06, "loss": 1.3182, "step": 431 }, { "epoch": 0.011217657423722526, "grad_norm": 11.8824664284336, "learning_rate": 8.604374887199955e-06, "loss": 1.3358, "step": 432 }, { "epoch": 0.011243624223314477, "grad_norm": 12.624693611884025, "learning_rate": 8.60765325273042e-06, "loss": 1.3101, "step": 433 }, { "epoch": 0.011269591022906427, "grad_norm": 7.643104946463594, "learning_rate": 8.610924055703927e-06, "loss": 1.2687, "step": 434 }, { "epoch": 0.011295557822498378, "grad_norm": 13.344452671962902, "learning_rate": 8.614187330930911e-06, "loss": 1.3315, "step": 435 }, { "epoch": 0.011321524622090327, "grad_norm": 49.021965241837, "learning_rate": 8.61744311298201e-06, "loss": 1.2778, "step": 436 }, { "epoch": 0.011347491421682277, "grad_norm": 7.14513216573154, "learning_rate": 8.620691436190264e-06, "loss": 1.2894, "step": 437 }, { "epoch": 0.011373458221274228, "grad_norm": 8.229940360352616, "learning_rate": 8.623932334653285e-06, "loss": 1.3516, "step": 438 }, { "epoch": 0.011399425020866178, "grad_norm": 11.419731972349986, "learning_rate": 8.627165842235402e-06, "loss": 1.2959, "step": 439 }, { "epoch": 0.011425391820458129, "grad_norm": 7.209055811079275, "learning_rate": 8.630391992569792e-06, "loss": 1.3563, "step": 440 }, { "epoch": 0.011451358620050079, "grad_norm": 6.40050088591324, "learning_rate": 8.633610819060579e-06, "loss": 1.3164, "step": 441 }, { "epoch": 0.01147732541964203, "grad_norm": 6.273048322299633, "learning_rate": 8.636822354884896e-06, "loss": 1.3187, "step": 442 }, { "epoch": 0.01150329221923398, "grad_norm": 8.258419942032788, "learning_rate": 8.640026632994951e-06, "loss": 1.3302, "step": 443 }, { "epoch": 0.01152925901882593, "grad_norm": 16.629842640860005, "learning_rate": 8.643223686120049e-06, "loss": 1.3237, "step": 444 }, { "epoch": 0.01155522581841788, "grad_norm": 7.6472244313807245, "learning_rate": 8.6464135467686e-06, "loss": 1.2647, "step": 445 }, { "epoch": 0.01158119261800983, "grad_norm": 8.715713464418599, "learning_rate": 8.649596247230101e-06, "loss": 1.3273, "step": 446 }, { "epoch": 0.01160715941760178, "grad_norm": 11.869932358250956, "learning_rate": 8.652771819577103e-06, "loss": 1.2864, "step": 447 }, { "epoch": 0.01163312621719373, "grad_norm": 7.558537982271662, "learning_rate": 8.655940295667134e-06, "loss": 1.2688, "step": 448 }, { "epoch": 0.011659093016785681, "grad_norm": 6.43557888364574, "learning_rate": 8.659101707144642e-06, "loss": 1.3129, "step": 449 }, { "epoch": 0.011685059816377632, "grad_norm": 10.57662975838232, "learning_rate": 8.662256085442863e-06, "loss": 1.3073, "step": 450 }, { "epoch": 0.011711026615969582, "grad_norm": 9.285370705085466, "learning_rate": 8.66540346178572e-06, "loss": 1.2818, "step": 451 }, { "epoch": 0.011736993415561532, "grad_norm": 7.901559757822838, "learning_rate": 8.668543867189666e-06, "loss": 1.3301, "step": 452 }, { "epoch": 0.011762960215153483, "grad_norm": 8.429572103734019, "learning_rate": 8.671677332465516e-06, "loss": 1.3729, "step": 453 }, { "epoch": 0.011788927014745433, "grad_norm": 6.463813409208414, "learning_rate": 8.674803888220268e-06, "loss": 1.2708, "step": 454 }, { "epoch": 0.011814893814337382, "grad_norm": 16.758537280363814, "learning_rate": 8.677923564858893e-06, "loss": 1.3361, "step": 455 }, { "epoch": 0.011840860613929333, "grad_norm": 28.670275202302843, "learning_rate": 8.681036392586107e-06, "loss": 1.3323, "step": 456 }, { "epoch": 0.011866827413521283, "grad_norm": 8.377091331702639, "learning_rate": 8.684142401408131e-06, "loss": 1.3586, "step": 457 }, { "epoch": 0.011892794213113234, "grad_norm": 12.474321158062757, "learning_rate": 8.687241621134421e-06, "loss": 1.3726, "step": 458 }, { "epoch": 0.011918761012705184, "grad_norm": 11.219645599915147, "learning_rate": 8.690334081379388e-06, "loss": 1.2291, "step": 459 }, { "epoch": 0.011944727812297135, "grad_norm": 9.00940806011971, "learning_rate": 8.693419811564099e-06, "loss": 1.2881, "step": 460 }, { "epoch": 0.011970694611889085, "grad_norm": 12.080513741387797, "learning_rate": 8.696498840917945e-06, "loss": 1.2973, "step": 461 }, { "epoch": 0.011996661411481035, "grad_norm": 10.166901716006194, "learning_rate": 8.699571198480317e-06, "loss": 1.3592, "step": 462 }, { "epoch": 0.012022628211072986, "grad_norm": 24.717390676514963, "learning_rate": 8.702636913102236e-06, "loss": 1.3226, "step": 463 }, { "epoch": 0.012048595010664935, "grad_norm": 29.858104572986647, "learning_rate": 8.705696013447989e-06, "loss": 1.2975, "step": 464 }, { "epoch": 0.012074561810256885, "grad_norm": 8.549380310088178, "learning_rate": 8.708748527996734e-06, "loss": 1.2562, "step": 465 }, { "epoch": 0.012100528609848836, "grad_norm": 14.784844572342324, "learning_rate": 8.711794485044082e-06, "loss": 1.2887, "step": 466 }, { "epoch": 0.012126495409440786, "grad_norm": 8.702597692873313, "learning_rate": 8.714833912703688e-06, "loss": 1.324, "step": 467 }, { "epoch": 0.012152462209032737, "grad_norm": 8.740970421172273, "learning_rate": 8.71786683890879e-06, "loss": 1.2953, "step": 468 }, { "epoch": 0.012178429008624687, "grad_norm": 14.943312786858824, "learning_rate": 8.720893291413763e-06, "loss": 1.334, "step": 469 }, { "epoch": 0.012204395808216638, "grad_norm": 22.592041288689337, "learning_rate": 8.723913297795636e-06, "loss": 1.259, "step": 470 }, { "epoch": 0.012230362607808588, "grad_norm": 7.939431263423161, "learning_rate": 8.726926885455599e-06, "loss": 1.3223, "step": 471 }, { "epoch": 0.012256329407400538, "grad_norm": 14.375975467922059, "learning_rate": 8.729934081620502e-06, "loss": 1.3075, "step": 472 }, { "epoch": 0.012282296206992489, "grad_norm": 7.938068279438846, "learning_rate": 8.732934913344318e-06, "loss": 1.2826, "step": 473 }, { "epoch": 0.012308263006584438, "grad_norm": 11.73295298162169, "learning_rate": 8.735929407509624e-06, "loss": 1.3223, "step": 474 }, { "epoch": 0.012334229806176388, "grad_norm": 34.04614716408818, "learning_rate": 8.738917590829017e-06, "loss": 1.3326, "step": 475 }, { "epoch": 0.012360196605768339, "grad_norm": 7.837383711102742, "learning_rate": 8.74189948984657e-06, "loss": 1.289, "step": 476 }, { "epoch": 0.012386163405360289, "grad_norm": 97.16839643436028, "learning_rate": 8.744875130939237e-06, "loss": 1.293, "step": 477 }, { "epoch": 0.01241213020495224, "grad_norm": 7.972486868913614, "learning_rate": 8.74784454031825e-06, "loss": 1.3219, "step": 478 }, { "epoch": 0.01243809700454419, "grad_norm": 9.091655686672258, "learning_rate": 8.750807744030504e-06, "loss": 1.3014, "step": 479 }, { "epoch": 0.01246406380413614, "grad_norm": 12.356849383421743, "learning_rate": 8.75376476795994e-06, "loss": 1.3343, "step": 480 }, { "epoch": 0.012490030603728091, "grad_norm": 11.799624576092189, "learning_rate": 8.756715637828884e-06, "loss": 1.3273, "step": 481 }, { "epoch": 0.012515997403320041, "grad_norm": 11.711360300281424, "learning_rate": 8.7596603791994e-06, "loss": 1.3486, "step": 482 }, { "epoch": 0.012541964202911992, "grad_norm": 6.304515963918772, "learning_rate": 8.76259901747462e-06, "loss": 1.3623, "step": 483 }, { "epoch": 0.01256793100250394, "grad_norm": 10.630146889304193, "learning_rate": 8.765531577900054e-06, "loss": 1.3145, "step": 484 }, { "epoch": 0.012593897802095891, "grad_norm": 8.559846118720351, "learning_rate": 8.768458085564887e-06, "loss": 1.2686, "step": 485 }, { "epoch": 0.012619864601687842, "grad_norm": 12.44509425813039, "learning_rate": 8.771378565403283e-06, "loss": 1.2983, "step": 486 }, { "epoch": 0.012645831401279792, "grad_norm": 7.424097888745131, "learning_rate": 8.77429304219564e-06, "loss": 1.3874, "step": 487 }, { "epoch": 0.012671798200871743, "grad_norm": 19.253763202577762, "learning_rate": 8.777201540569868e-06, "loss": 1.363, "step": 488 }, { "epoch": 0.012697765000463693, "grad_norm": 12.012439002796162, "learning_rate": 8.780104085002626e-06, "loss": 1.2677, "step": 489 }, { "epoch": 0.012723731800055644, "grad_norm": 8.46542371902764, "learning_rate": 8.783000699820565e-06, "loss": 1.318, "step": 490 }, { "epoch": 0.012749698599647594, "grad_norm": 31.918480033920495, "learning_rate": 8.78589140920155e-06, "loss": 1.2861, "step": 491 }, { "epoch": 0.012775665399239544, "grad_norm": 9.0983791039578, "learning_rate": 8.788776237175867e-06, "loss": 1.3323, "step": 492 }, { "epoch": 0.012801632198831493, "grad_norm": 34.21629022266386, "learning_rate": 8.791655207627424e-06, "loss": 1.257, "step": 493 }, { "epoch": 0.012827598998423444, "grad_norm": 8.234301340544596, "learning_rate": 8.794528344294943e-06, "loss": 1.344, "step": 494 }, { "epoch": 0.012853565798015394, "grad_norm": 8.278718842245176, "learning_rate": 8.797395670773122e-06, "loss": 1.3074, "step": 495 }, { "epoch": 0.012879532597607345, "grad_norm": 8.319601825576319, "learning_rate": 8.800257210513812e-06, "loss": 1.3175, "step": 496 }, { "epoch": 0.012905499397199295, "grad_norm": 7.7405290877214155, "learning_rate": 8.803112986827153e-06, "loss": 1.3357, "step": 497 }, { "epoch": 0.012931466196791246, "grad_norm": 8.526207855661813, "learning_rate": 8.805963022882731e-06, "loss": 1.289, "step": 498 }, { "epoch": 0.012957432996383196, "grad_norm": 16.773957840089313, "learning_rate": 8.808807341710687e-06, "loss": 1.3642, "step": 499 }, { "epoch": 0.012983399795975147, "grad_norm": 21.823192697264833, "learning_rate": 8.811645966202848e-06, "loss": 1.3052, "step": 500 }, { "epoch": 0.013009366595567097, "grad_norm": 5.908800008589215, "learning_rate": 8.814478919113826e-06, "loss": 1.3261, "step": 501 }, { "epoch": 0.013035333395159047, "grad_norm": 7.658593142325667, "learning_rate": 8.81730622306211e-06, "loss": 1.3202, "step": 502 }, { "epoch": 0.013061300194750996, "grad_norm": 9.109407244931138, "learning_rate": 8.82012790053116e-06, "loss": 1.3604, "step": 503 }, { "epoch": 0.013087266994342947, "grad_norm": 8.423662155288833, "learning_rate": 8.822943973870464e-06, "loss": 1.3056, "step": 504 }, { "epoch": 0.013113233793934897, "grad_norm": 6.343488716043891, "learning_rate": 8.825754465296615e-06, "loss": 1.3373, "step": 505 }, { "epoch": 0.013139200593526848, "grad_norm": 6.476887325520665, "learning_rate": 8.828559396894359e-06, "loss": 1.2722, "step": 506 }, { "epoch": 0.013165167393118798, "grad_norm": 8.869875865871887, "learning_rate": 8.831358790617626e-06, "loss": 1.361, "step": 507 }, { "epoch": 0.013191134192710749, "grad_norm": 20.233810169481412, "learning_rate": 8.834152668290574e-06, "loss": 1.3159, "step": 508 }, { "epoch": 0.013217100992302699, "grad_norm": 8.093723135060385, "learning_rate": 8.836941051608605e-06, "loss": 1.3335, "step": 509 }, { "epoch": 0.01324306779189465, "grad_norm": 9.801160735852644, "learning_rate": 8.839723962139376e-06, "loss": 1.3044, "step": 510 }, { "epoch": 0.0132690345914866, "grad_norm": 9.961510783135234, "learning_rate": 8.842501421323793e-06, "loss": 1.3411, "step": 511 }, { "epoch": 0.013295001391078549, "grad_norm": 9.041207391234236, "learning_rate": 8.845273450477018e-06, "loss": 1.322, "step": 512 }, { "epoch": 0.0133209681906705, "grad_norm": 7.125103063005283, "learning_rate": 8.848040070789435e-06, "loss": 1.283, "step": 513 }, { "epoch": 0.01334693499026245, "grad_norm": 9.695936110710935, "learning_rate": 8.850801303327631e-06, "loss": 1.3237, "step": 514 }, { "epoch": 0.0133729017898544, "grad_norm": 7.0853662902004295, "learning_rate": 8.853557169035352e-06, "loss": 1.2671, "step": 515 }, { "epoch": 0.01339886858944635, "grad_norm": 16.186120876640057, "learning_rate": 8.856307688734466e-06, "loss": 1.3301, "step": 516 }, { "epoch": 0.013424835389038301, "grad_norm": 21.04460710683975, "learning_rate": 8.859052883125895e-06, "loss": 1.3258, "step": 517 }, { "epoch": 0.013450802188630252, "grad_norm": 11.375608323148247, "learning_rate": 8.861792772790557e-06, "loss": 1.325, "step": 518 }, { "epoch": 0.013476768988222202, "grad_norm": 15.634870417551685, "learning_rate": 8.864527378190292e-06, "loss": 1.3483, "step": 519 }, { "epoch": 0.013502735787814153, "grad_norm": 17.707983012006636, "learning_rate": 8.867256719668776e-06, "loss": 1.3044, "step": 520 }, { "epoch": 0.013528702587406103, "grad_norm": 17.390560071195054, "learning_rate": 8.869980817452429e-06, "loss": 1.3198, "step": 521 }, { "epoch": 0.013554669386998052, "grad_norm": 24.013850983599028, "learning_rate": 8.872699691651319e-06, "loss": 1.315, "step": 522 }, { "epoch": 0.013580636186590002, "grad_norm": 13.636292423871524, "learning_rate": 8.875413362260044e-06, "loss": 1.3022, "step": 523 }, { "epoch": 0.013606602986181953, "grad_norm": 13.332696772320904, "learning_rate": 8.878121849158624e-06, "loss": 1.3065, "step": 524 }, { "epoch": 0.013632569785773903, "grad_norm": 12.810938595789073, "learning_rate": 8.880825172113371e-06, "loss": 1.3111, "step": 525 }, { "epoch": 0.013658536585365854, "grad_norm": 11.527270946059486, "learning_rate": 8.883523350777752e-06, "loss": 1.3245, "step": 526 }, { "epoch": 0.013684503384957804, "grad_norm": 8.155897344806606, "learning_rate": 8.886216404693247e-06, "loss": 1.3301, "step": 527 }, { "epoch": 0.013710470184549755, "grad_norm": 8.968821455508033, "learning_rate": 8.8889043532902e-06, "loss": 1.3234, "step": 528 }, { "epoch": 0.013736436984141705, "grad_norm": 9.720529473094802, "learning_rate": 8.891587215888663e-06, "loss": 1.3125, "step": 529 }, { "epoch": 0.013762403783733656, "grad_norm": 17.16864199831793, "learning_rate": 8.894265011699222e-06, "loss": 1.288, "step": 530 }, { "epoch": 0.013788370583325604, "grad_norm": 49.356201920009504, "learning_rate": 8.89693775982383e-06, "loss": 1.2833, "step": 531 }, { "epoch": 0.013814337382917555, "grad_norm": 11.400028402728118, "learning_rate": 8.899605479256616e-06, "loss": 1.3399, "step": 532 }, { "epoch": 0.013840304182509505, "grad_norm": 9.460091538097275, "learning_rate": 8.902268188884702e-06, "loss": 1.345, "step": 533 }, { "epoch": 0.013866270982101456, "grad_norm": 6.814165944077656, "learning_rate": 8.904925907489006e-06, "loss": 1.3619, "step": 534 }, { "epoch": 0.013892237781693406, "grad_norm": 15.578239601195818, "learning_rate": 8.907578653745027e-06, "loss": 1.2911, "step": 535 }, { "epoch": 0.013918204581285357, "grad_norm": 10.999393446411709, "learning_rate": 8.910226446223646e-06, "loss": 1.3281, "step": 536 }, { "epoch": 0.013944171380877307, "grad_norm": 15.10227179205996, "learning_rate": 8.912869303391895e-06, "loss": 1.3268, "step": 537 }, { "epoch": 0.013970138180469258, "grad_norm": 7.852517626917796, "learning_rate": 8.915507243613737e-06, "loss": 1.3016, "step": 538 }, { "epoch": 0.013996104980061208, "grad_norm": 12.065995890764157, "learning_rate": 8.918140285150826e-06, "loss": 1.3555, "step": 539 }, { "epoch": 0.014022071779653159, "grad_norm": 8.603419187527049, "learning_rate": 8.92076844616327e-06, "loss": 1.2917, "step": 540 }, { "epoch": 0.014048038579245107, "grad_norm": 11.660028658221508, "learning_rate": 8.923391744710377e-06, "loss": 1.3089, "step": 541 }, { "epoch": 0.014074005378837058, "grad_norm": 18.65551797270926, "learning_rate": 8.926010198751412e-06, "loss": 1.336, "step": 542 }, { "epoch": 0.014099972178429008, "grad_norm": 7.000661818689587, "learning_rate": 8.928623826146315e-06, "loss": 1.3511, "step": 543 }, { "epoch": 0.014125938978020959, "grad_norm": 15.827903753543781, "learning_rate": 8.931232644656452e-06, "loss": 1.2936, "step": 544 }, { "epoch": 0.01415190577761291, "grad_norm": 13.601446388971663, "learning_rate": 8.933836671945326e-06, "loss": 1.3131, "step": 545 }, { "epoch": 0.01417787257720486, "grad_norm": 83.19046705358583, "learning_rate": 8.936435925579299e-06, "loss": 1.2495, "step": 546 }, { "epoch": 0.01420383937679681, "grad_norm": 23.934707837318854, "learning_rate": 8.9390304230283e-06, "loss": 1.3317, "step": 547 }, { "epoch": 0.01422980617638876, "grad_norm": 11.469646699447907, "learning_rate": 8.941620181666542e-06, "loss": 1.2713, "step": 548 }, { "epoch": 0.014255772975980711, "grad_norm": 9.729578002686013, "learning_rate": 8.944205218773196e-06, "loss": 1.311, "step": 549 }, { "epoch": 0.01428173977557266, "grad_norm": 12.578228114872365, "learning_rate": 8.946785551533108e-06, "loss": 1.286, "step": 550 }, { "epoch": 0.01430770657516461, "grad_norm": 38.60564088073788, "learning_rate": 8.949361197037471e-06, "loss": 1.3017, "step": 551 }, { "epoch": 0.01433367337475656, "grad_norm": 11.08377839153315, "learning_rate": 8.951932172284505e-06, "loss": 1.3343, "step": 552 }, { "epoch": 0.014359640174348511, "grad_norm": 19.78722957732343, "learning_rate": 8.954498494180131e-06, "loss": 1.2972, "step": 553 }, { "epoch": 0.014385606973940462, "grad_norm": 16.061551908425017, "learning_rate": 8.957060179538645e-06, "loss": 1.2834, "step": 554 }, { "epoch": 0.014411573773532412, "grad_norm": 10.370988920727264, "learning_rate": 8.959617245083364e-06, "loss": 1.2948, "step": 555 }, { "epoch": 0.014437540573124363, "grad_norm": 11.34506344087767, "learning_rate": 8.962169707447295e-06, "loss": 1.3255, "step": 556 }, { "epoch": 0.014463507372716313, "grad_norm": 9.757952845643032, "learning_rate": 8.964717583173782e-06, "loss": 1.2764, "step": 557 }, { "epoch": 0.014489474172308264, "grad_norm": 13.268599788841524, "learning_rate": 8.967260888717138e-06, "loss": 1.3444, "step": 558 }, { "epoch": 0.014515440971900214, "grad_norm": 19.686153692806446, "learning_rate": 8.969799640443301e-06, "loss": 1.3064, "step": 559 }, { "epoch": 0.014541407771492163, "grad_norm": 22.32061397788173, "learning_rate": 8.972333854630449e-06, "loss": 1.3004, "step": 560 }, { "epoch": 0.014567374571084113, "grad_norm": 12.073207076065684, "learning_rate": 8.974863547469635e-06, "loss": 1.3336, "step": 561 }, { "epoch": 0.014593341370676064, "grad_norm": 8.93231477576882, "learning_rate": 8.977388735065404e-06, "loss": 1.3105, "step": 562 }, { "epoch": 0.014619308170268014, "grad_norm": 9.706787204552938, "learning_rate": 8.979909433436415e-06, "loss": 1.337, "step": 563 }, { "epoch": 0.014645274969859965, "grad_norm": 8.49702940250251, "learning_rate": 8.982425658516042e-06, "loss": 1.3846, "step": 564 }, { "epoch": 0.014671241769451915, "grad_norm": 12.46072932819845, "learning_rate": 8.984937426152981e-06, "loss": 1.302, "step": 565 }, { "epoch": 0.014697208569043866, "grad_norm": 11.22518048045766, "learning_rate": 8.987444752111852e-06, "loss": 1.3049, "step": 566 }, { "epoch": 0.014723175368635816, "grad_norm": 9.653347134589065, "learning_rate": 8.98994765207379e-06, "loss": 1.3146, "step": 567 }, { "epoch": 0.014749142168227767, "grad_norm": 12.618630190351858, "learning_rate": 8.992446141637037e-06, "loss": 1.3258, "step": 568 }, { "epoch": 0.014775108967819717, "grad_norm": 9.694785252081882, "learning_rate": 8.994940236317517e-06, "loss": 1.3053, "step": 569 }, { "epoch": 0.014801075767411666, "grad_norm": 14.601681630733923, "learning_rate": 8.997429951549422e-06, "loss": 1.2976, "step": 570 }, { "epoch": 0.014827042567003616, "grad_norm": 8.100937645412614, "learning_rate": 8.999915302685781e-06, "loss": 1.2719, "step": 571 }, { "epoch": 0.014853009366595567, "grad_norm": 9.764920287073041, "learning_rate": 9.002396304999035e-06, "loss": 1.3126, "step": 572 }, { "epoch": 0.014878976166187517, "grad_norm": 8.66681871525339, "learning_rate": 9.004872973681589e-06, "loss": 1.3199, "step": 573 }, { "epoch": 0.014904942965779468, "grad_norm": 12.930710637389216, "learning_rate": 9.007345323846377e-06, "loss": 1.2854, "step": 574 }, { "epoch": 0.014930909765371418, "grad_norm": 14.495094258691779, "learning_rate": 9.009813370527414e-06, "loss": 1.2773, "step": 575 }, { "epoch": 0.014956876564963369, "grad_norm": 7.94574363536512, "learning_rate": 9.012277128680346e-06, "loss": 1.2841, "step": 576 }, { "epoch": 0.014982843364555319, "grad_norm": 7.950306810900519, "learning_rate": 9.014736613182994e-06, "loss": 1.274, "step": 577 }, { "epoch": 0.01500881016414727, "grad_norm": 10.119029981239827, "learning_rate": 9.017191838835889e-06, "loss": 1.3142, "step": 578 }, { "epoch": 0.015034776963739218, "grad_norm": 9.057933840910332, "learning_rate": 9.019642820362806e-06, "loss": 1.3321, "step": 579 }, { "epoch": 0.015060743763331169, "grad_norm": 8.03784099229282, "learning_rate": 9.022089572411304e-06, "loss": 1.2983, "step": 580 }, { "epoch": 0.01508671056292312, "grad_norm": 10.008949090171678, "learning_rate": 9.02453210955324e-06, "loss": 1.3003, "step": 581 }, { "epoch": 0.01511267736251507, "grad_norm": 10.371568594944012, "learning_rate": 9.026970446285293e-06, "loss": 1.3236, "step": 582 }, { "epoch": 0.01513864416210702, "grad_norm": 11.90331674723314, "learning_rate": 9.029404597029482e-06, "loss": 1.2916, "step": 583 }, { "epoch": 0.01516461096169897, "grad_norm": 11.661498169222565, "learning_rate": 9.031834576133676e-06, "loss": 1.2984, "step": 584 }, { "epoch": 0.015190577761290921, "grad_norm": 10.018358164149243, "learning_rate": 9.034260397872104e-06, "loss": 1.3306, "step": 585 }, { "epoch": 0.015216544560882872, "grad_norm": 13.27678693393332, "learning_rate": 9.036682076445854e-06, "loss": 1.3203, "step": 586 }, { "epoch": 0.015242511360474822, "grad_norm": 15.120002866875916, "learning_rate": 9.039099625983376e-06, "loss": 1.3284, "step": 587 }, { "epoch": 0.015268478160066773, "grad_norm": 17.973215307948085, "learning_rate": 9.041513060540972e-06, "loss": 1.3476, "step": 588 }, { "epoch": 0.015294444959658721, "grad_norm": 8.089184036506083, "learning_rate": 9.043922394103292e-06, "loss": 1.3123, "step": 589 }, { "epoch": 0.015320411759250672, "grad_norm": 9.2648085544856, "learning_rate": 9.046327640583815e-06, "loss": 1.2739, "step": 590 }, { "epoch": 0.015346378558842622, "grad_norm": 9.459103545927496, "learning_rate": 9.048728813825338e-06, "loss": 1.334, "step": 591 }, { "epoch": 0.015372345358434573, "grad_norm": 23.601991782755988, "learning_rate": 9.05112592760044e-06, "loss": 1.3188, "step": 592 }, { "epoch": 0.015398312158026523, "grad_norm": 7.51043991811877, "learning_rate": 9.053518995611976e-06, "loss": 1.3281, "step": 593 }, { "epoch": 0.015424278957618474, "grad_norm": 22.389406478532152, "learning_rate": 9.055908031493528e-06, "loss": 1.3311, "step": 594 }, { "epoch": 0.015450245757210424, "grad_norm": 10.556096229314086, "learning_rate": 9.058293048809884e-06, "loss": 1.298, "step": 595 }, { "epoch": 0.015476212556802375, "grad_norm": 7.762648075184148, "learning_rate": 9.060674061057494e-06, "loss": 1.3115, "step": 596 }, { "epoch": 0.015502179356394325, "grad_norm": 47.315583102154356, "learning_rate": 9.06305108166493e-06, "loss": 1.328, "step": 597 }, { "epoch": 0.015528146155986274, "grad_norm": 18.333856933514937, "learning_rate": 9.06542412399334e-06, "loss": 1.2809, "step": 598 }, { "epoch": 0.015554112955578224, "grad_norm": 7.2518300344724675, "learning_rate": 9.067793201336898e-06, "loss": 1.2767, "step": 599 }, { "epoch": 0.015580079755170175, "grad_norm": 8.965675172502522, "learning_rate": 9.070158326923256e-06, "loss": 1.3126, "step": 600 }, { "epoch": 0.015606046554762125, "grad_norm": 15.710259573265331, "learning_rate": 9.072519513913973e-06, "loss": 1.2567, "step": 601 }, { "epoch": 0.015632013354354077, "grad_norm": 8.087489000457406, "learning_rate": 9.074876775404974e-06, "loss": 1.2956, "step": 602 }, { "epoch": 0.015657980153946024, "grad_norm": 7.332846220773334, "learning_rate": 9.077230124426974e-06, "loss": 1.2562, "step": 603 }, { "epoch": 0.015683946953537975, "grad_norm": 8.517217812618046, "learning_rate": 9.07957957394591e-06, "loss": 1.3486, "step": 604 }, { "epoch": 0.015709913753129925, "grad_norm": 16.28920207524134, "learning_rate": 9.081925136863369e-06, "loss": 1.3155, "step": 605 }, { "epoch": 0.015735880552721876, "grad_norm": 17.08201204478111, "learning_rate": 9.08426682601702e-06, "loss": 1.2697, "step": 606 }, { "epoch": 0.015761847352313826, "grad_norm": 7.118860376857067, "learning_rate": 9.086604654181038e-06, "loss": 1.3118, "step": 607 }, { "epoch": 0.015787814151905777, "grad_norm": 11.266422407523432, "learning_rate": 9.0889386340665e-06, "loss": 1.2964, "step": 608 }, { "epoch": 0.015813780951497727, "grad_norm": 11.018286676992332, "learning_rate": 9.091268778321827e-06, "loss": 1.3387, "step": 609 }, { "epoch": 0.015839747751089678, "grad_norm": 7.319764009851145, "learning_rate": 9.093595099533182e-06, "loss": 1.3338, "step": 610 }, { "epoch": 0.01586571455068163, "grad_norm": 9.811724784017573, "learning_rate": 9.095917610224879e-06, "loss": 1.3304, "step": 611 }, { "epoch": 0.01589168135027358, "grad_norm": 10.221337413733616, "learning_rate": 9.098236322859781e-06, "loss": 1.3093, "step": 612 }, { "epoch": 0.01591764814986553, "grad_norm": 9.241624906931149, "learning_rate": 9.100551249839722e-06, "loss": 1.3273, "step": 613 }, { "epoch": 0.01594361494945748, "grad_norm": 8.14333821073124, "learning_rate": 9.102862403505876e-06, "loss": 1.2895, "step": 614 }, { "epoch": 0.01596958174904943, "grad_norm": 8.330253744118858, "learning_rate": 9.105169796139182e-06, "loss": 1.3284, "step": 615 }, { "epoch": 0.01599554854864138, "grad_norm": 14.354473286063564, "learning_rate": 9.107473439960709e-06, "loss": 1.2391, "step": 616 }, { "epoch": 0.01602151534823333, "grad_norm": 9.313910665492, "learning_rate": 9.109773347132062e-06, "loss": 1.3036, "step": 617 }, { "epoch": 0.01604748214782528, "grad_norm": 14.74312004370333, "learning_rate": 9.112069529755758e-06, "loss": 1.2741, "step": 618 }, { "epoch": 0.016073448947417232, "grad_norm": 8.355429486525326, "learning_rate": 9.114361999875616e-06, "loss": 1.3031, "step": 619 }, { "epoch": 0.016099415747009183, "grad_norm": 80.22214640258325, "learning_rate": 9.116650769477127e-06, "loss": 1.324, "step": 620 }, { "epoch": 0.016125382546601133, "grad_norm": 11.7324230129952, "learning_rate": 9.118935850487833e-06, "loss": 1.3167, "step": 621 }, { "epoch": 0.01615134934619308, "grad_norm": 10.691009480672081, "learning_rate": 9.121217254777707e-06, "loss": 1.3022, "step": 622 }, { "epoch": 0.01617731614578503, "grad_norm": 11.96406579049354, "learning_rate": 9.123494994159514e-06, "loss": 1.3093, "step": 623 }, { "epoch": 0.01620328294537698, "grad_norm": 8.40092759233295, "learning_rate": 9.125769080389182e-06, "loss": 1.2821, "step": 624 }, { "epoch": 0.01622924974496893, "grad_norm": 12.362364820892633, "learning_rate": 9.128039525166163e-06, "loss": 1.2593, "step": 625 }, { "epoch": 0.016255216544560882, "grad_norm": 11.000279509721205, "learning_rate": 9.130306340133801e-06, "loss": 1.2966, "step": 626 }, { "epoch": 0.016281183344152832, "grad_norm": 10.213677410110082, "learning_rate": 9.132569536879682e-06, "loss": 1.2504, "step": 627 }, { "epoch": 0.016307150143744783, "grad_norm": 18.052368279211493, "learning_rate": 9.134829126935992e-06, "loss": 1.263, "step": 628 }, { "epoch": 0.016333116943336733, "grad_norm": 11.542408245697757, "learning_rate": 9.137085121779877e-06, "loss": 1.2882, "step": 629 }, { "epoch": 0.016359083742928684, "grad_norm": 25.364351791379864, "learning_rate": 9.139337532833777e-06, "loss": 1.3332, "step": 630 }, { "epoch": 0.016385050542520634, "grad_norm": 11.710337475247231, "learning_rate": 9.141586371465793e-06, "loss": 1.2768, "step": 631 }, { "epoch": 0.016411017342112585, "grad_norm": 7.828728262371525, "learning_rate": 9.143831648990016e-06, "loss": 1.2837, "step": 632 }, { "epoch": 0.016436984141704535, "grad_norm": 7.066480207082579, "learning_rate": 9.146073376666868e-06, "loss": 1.2865, "step": 633 }, { "epoch": 0.016462950941296486, "grad_norm": 8.120703867204377, "learning_rate": 9.148311565703455e-06, "loss": 1.2979, "step": 634 }, { "epoch": 0.016488917740888436, "grad_norm": 27.393370156218914, "learning_rate": 9.150546227253889e-06, "loss": 1.3217, "step": 635 }, { "epoch": 0.016514884540480387, "grad_norm": 8.034100334728, "learning_rate": 9.152777372419628e-06, "loss": 1.2805, "step": 636 }, { "epoch": 0.016540851340072337, "grad_norm": 12.914565245620869, "learning_rate": 9.155005012249807e-06, "loss": 1.2925, "step": 637 }, { "epoch": 0.016566818139664288, "grad_norm": 10.3796041912254, "learning_rate": 9.157229157741564e-06, "loss": 1.2607, "step": 638 }, { "epoch": 0.016592784939256238, "grad_norm": 13.795358645791548, "learning_rate": 9.159449819840365e-06, "loss": 1.3279, "step": 639 }, { "epoch": 0.01661875173884819, "grad_norm": 9.413164911177299, "learning_rate": 9.161667009440333e-06, "loss": 1.2971, "step": 640 }, { "epoch": 0.016644718538440136, "grad_norm": 9.72507361847537, "learning_rate": 9.16388073738456e-06, "loss": 1.3089, "step": 641 }, { "epoch": 0.016670685338032086, "grad_norm": 7.724865919017422, "learning_rate": 9.166091014465433e-06, "loss": 1.2698, "step": 642 }, { "epoch": 0.016696652137624036, "grad_norm": 7.844480928811063, "learning_rate": 9.168297851424947e-06, "loss": 1.2959, "step": 643 }, { "epoch": 0.016722618937215987, "grad_norm": 10.170541702292283, "learning_rate": 9.170501258955015e-06, "loss": 1.2719, "step": 644 }, { "epoch": 0.016748585736807937, "grad_norm": 8.59401056068061, "learning_rate": 9.172701247697781e-06, "loss": 1.3337, "step": 645 }, { "epoch": 0.016774552536399888, "grad_norm": 8.043001950210739, "learning_rate": 9.174897828245934e-06, "loss": 1.2882, "step": 646 }, { "epoch": 0.01680051933599184, "grad_norm": 8.777179901535902, "learning_rate": 9.177091011143006e-06, "loss": 1.3435, "step": 647 }, { "epoch": 0.01682648613558379, "grad_norm": 33.96567553071501, "learning_rate": 9.179280806883674e-06, "loss": 1.2949, "step": 648 }, { "epoch": 0.01685245293517574, "grad_norm": 28.6817187631937, "learning_rate": 9.181467225914077e-06, "loss": 1.3144, "step": 649 }, { "epoch": 0.01687841973476769, "grad_norm": 9.929303321639692, "learning_rate": 9.18365027863209e-06, "loss": 1.3016, "step": 650 }, { "epoch": 0.01690438653435964, "grad_norm": 14.84143845077309, "learning_rate": 9.185829975387648e-06, "loss": 1.3121, "step": 651 }, { "epoch": 0.01693035333395159, "grad_norm": 11.151662700855525, "learning_rate": 9.188006326483019e-06, "loss": 1.3064, "step": 652 }, { "epoch": 0.01695632013354354, "grad_norm": 11.958299661161398, "learning_rate": 9.190179342173104e-06, "loss": 1.3112, "step": 653 }, { "epoch": 0.01698228693313549, "grad_norm": 9.915681891604242, "learning_rate": 9.192349032665733e-06, "loss": 1.3032, "step": 654 }, { "epoch": 0.017008253732727442, "grad_norm": 37.723450476067434, "learning_rate": 9.194515408121939e-06, "loss": 1.267, "step": 655 }, { "epoch": 0.017034220532319393, "grad_norm": 9.993901109537108, "learning_rate": 9.19667847865626e-06, "loss": 1.2801, "step": 656 }, { "epoch": 0.017060187331911343, "grad_norm": 9.149354805633838, "learning_rate": 9.198838254337004e-06, "loss": 1.2861, "step": 657 }, { "epoch": 0.017086154131503294, "grad_norm": 9.15138047440944, "learning_rate": 9.200994745186552e-06, "loss": 1.2721, "step": 658 }, { "epoch": 0.017112120931095244, "grad_norm": 19.171036108887836, "learning_rate": 9.20314796118161e-06, "loss": 1.2741, "step": 659 }, { "epoch": 0.01713808773068719, "grad_norm": 27.297866637325686, "learning_rate": 9.205297912253515e-06, "loss": 1.2858, "step": 660 }, { "epoch": 0.01716405453027914, "grad_norm": 21.454637800585143, "learning_rate": 9.207444608288486e-06, "loss": 1.3048, "step": 661 }, { "epoch": 0.017190021329871092, "grad_norm": 6.771246074909877, "learning_rate": 9.209588059127914e-06, "loss": 1.3081, "step": 662 }, { "epoch": 0.017215988129463042, "grad_norm": 12.409302919592031, "learning_rate": 9.211728274568617e-06, "loss": 1.289, "step": 663 }, { "epoch": 0.017241954929054993, "grad_norm": 14.705951674354077, "learning_rate": 9.213865264363124e-06, "loss": 1.2987, "step": 664 }, { "epoch": 0.017267921728646943, "grad_norm": 8.296533380002053, "learning_rate": 9.215999038219931e-06, "loss": 1.269, "step": 665 }, { "epoch": 0.017293888528238894, "grad_norm": 10.64632044882743, "learning_rate": 9.21812960580377e-06, "loss": 1.3025, "step": 666 }, { "epoch": 0.017319855327830844, "grad_norm": 13.388991408705243, "learning_rate": 9.220256976735869e-06, "loss": 1.2994, "step": 667 }, { "epoch": 0.017345822127422795, "grad_norm": 10.900458341988639, "learning_rate": 9.222381160594218e-06, "loss": 1.2921, "step": 668 }, { "epoch": 0.017371788927014745, "grad_norm": 9.746180642147722, "learning_rate": 9.224502166913824e-06, "loss": 1.3148, "step": 669 }, { "epoch": 0.017397755726606696, "grad_norm": 10.674548058691956, "learning_rate": 9.226620005186961e-06, "loss": 1.2596, "step": 670 }, { "epoch": 0.017423722526198646, "grad_norm": 40.8701437768911, "learning_rate": 9.228734684863441e-06, "loss": 1.3363, "step": 671 }, { "epoch": 0.017449689325790597, "grad_norm": 6.747756019871416, "learning_rate": 9.230846215350855e-06, "loss": 1.304, "step": 672 }, { "epoch": 0.017475656125382547, "grad_norm": 19.778278434979182, "learning_rate": 9.232954606014825e-06, "loss": 1.3071, "step": 673 }, { "epoch": 0.017501622924974498, "grad_norm": 8.453818132395858, "learning_rate": 9.235059866179255e-06, "loss": 1.3355, "step": 674 }, { "epoch": 0.017527589724566448, "grad_norm": 6.875251904120836, "learning_rate": 9.237162005126584e-06, "loss": 1.3299, "step": 675 }, { "epoch": 0.0175535565241584, "grad_norm": 7.727957604396621, "learning_rate": 9.239261032098019e-06, "loss": 1.2462, "step": 676 }, { "epoch": 0.01757952332375035, "grad_norm": 10.502004522900542, "learning_rate": 9.24135695629379e-06, "loss": 1.2725, "step": 677 }, { "epoch": 0.0176054901233423, "grad_norm": 9.042296310041701, "learning_rate": 9.243449786873387e-06, "loss": 1.3239, "step": 678 }, { "epoch": 0.01763145692293425, "grad_norm": 15.823222147554505, "learning_rate": 9.245539532955803e-06, "loss": 1.3487, "step": 679 }, { "epoch": 0.017657423722526197, "grad_norm": 8.707504403362098, "learning_rate": 9.247626203619767e-06, "loss": 1.2611, "step": 680 }, { "epoch": 0.017683390522118148, "grad_norm": 20.351384806086063, "learning_rate": 9.249709807903988e-06, "loss": 1.2986, "step": 681 }, { "epoch": 0.017709357321710098, "grad_norm": 8.237904322517537, "learning_rate": 9.251790354807385e-06, "loss": 1.2947, "step": 682 }, { "epoch": 0.01773532412130205, "grad_norm": 13.219234232228562, "learning_rate": 9.253867853289318e-06, "loss": 1.2485, "step": 683 }, { "epoch": 0.017761290920894, "grad_norm": 6.886971056336054, "learning_rate": 9.255942312269827e-06, "loss": 1.2678, "step": 684 }, { "epoch": 0.01778725772048595, "grad_norm": 22.48223718337504, "learning_rate": 9.258013740629857e-06, "loss": 1.3017, "step": 685 }, { "epoch": 0.0178132245200779, "grad_norm": 8.939413226027959, "learning_rate": 9.26008214721148e-06, "loss": 1.307, "step": 686 }, { "epoch": 0.01783919131966985, "grad_norm": 7.990898677240369, "learning_rate": 9.26214754081814e-06, "loss": 1.3052, "step": 687 }, { "epoch": 0.0178651581192618, "grad_norm": 9.292740923673259, "learning_rate": 9.264209930214859e-06, "loss": 1.2895, "step": 688 }, { "epoch": 0.01789112491885375, "grad_norm": 44.645794295120204, "learning_rate": 9.266269324128465e-06, "loss": 1.2751, "step": 689 }, { "epoch": 0.017917091718445702, "grad_norm": 16.521185019342873, "learning_rate": 9.26832573124782e-06, "loss": 1.3404, "step": 690 }, { "epoch": 0.017943058518037652, "grad_norm": 15.21818751366837, "learning_rate": 9.270379160224033e-06, "loss": 1.3005, "step": 691 }, { "epoch": 0.017969025317629603, "grad_norm": 15.675672715640966, "learning_rate": 9.272429619670684e-06, "loss": 1.325, "step": 692 }, { "epoch": 0.017994992117221553, "grad_norm": 14.719263152877996, "learning_rate": 9.274477118164037e-06, "loss": 1.2533, "step": 693 }, { "epoch": 0.018020958916813504, "grad_norm": 12.439264066951218, "learning_rate": 9.276521664243253e-06, "loss": 1.3126, "step": 694 }, { "epoch": 0.018046925716405454, "grad_norm": 9.623182248959004, "learning_rate": 9.27856326641061e-06, "loss": 1.3116, "step": 695 }, { "epoch": 0.018072892515997405, "grad_norm": 9.49080373465731, "learning_rate": 9.280601933131712e-06, "loss": 1.2609, "step": 696 }, { "epoch": 0.018098859315589355, "grad_norm": 12.38593648980537, "learning_rate": 9.282637672835695e-06, "loss": 1.308, "step": 697 }, { "epoch": 0.018124826115181306, "grad_norm": 13.301384095185593, "learning_rate": 9.284670493915446e-06, "loss": 1.3239, "step": 698 }, { "epoch": 0.018150792914773253, "grad_norm": 15.319515193907645, "learning_rate": 9.286700404727803e-06, "loss": 1.2822, "step": 699 }, { "epoch": 0.018176759714365203, "grad_norm": 33.45554419673965, "learning_rate": 9.288727413593764e-06, "loss": 1.29, "step": 700 }, { "epoch": 0.018202726513957154, "grad_norm": 8.482947354132778, "learning_rate": 9.29075152879869e-06, "loss": 1.2667, "step": 701 }, { "epoch": 0.018228693313549104, "grad_norm": 8.608671349988299, "learning_rate": 9.29277275859251e-06, "loss": 1.299, "step": 702 }, { "epoch": 0.018254660113141054, "grad_norm": 8.489157549394257, "learning_rate": 9.294791111189922e-06, "loss": 1.2708, "step": 703 }, { "epoch": 0.018280626912733005, "grad_norm": 14.702670958883468, "learning_rate": 9.296806594770593e-06, "loss": 1.273, "step": 704 }, { "epoch": 0.018306593712324955, "grad_norm": 8.877941235228574, "learning_rate": 9.298819217479357e-06, "loss": 1.3226, "step": 705 }, { "epoch": 0.018332560511916906, "grad_norm": 13.776428901792105, "learning_rate": 9.300828987426414e-06, "loss": 1.2911, "step": 706 }, { "epoch": 0.018358527311508856, "grad_norm": 30.326219108713296, "learning_rate": 9.30283591268753e-06, "loss": 1.2732, "step": 707 }, { "epoch": 0.018384494111100807, "grad_norm": 10.18200797671993, "learning_rate": 9.304840001304221e-06, "loss": 1.3039, "step": 708 }, { "epoch": 0.018410460910692757, "grad_norm": 14.94280324351539, "learning_rate": 9.30684126128396e-06, "loss": 1.3127, "step": 709 }, { "epoch": 0.018436427710284708, "grad_norm": 9.642062038358937, "learning_rate": 9.308839700600353e-06, "loss": 1.2957, "step": 710 }, { "epoch": 0.01846239450987666, "grad_norm": 88.29279238703462, "learning_rate": 9.310835327193344e-06, "loss": 1.2355, "step": 711 }, { "epoch": 0.01848836130946861, "grad_norm": 111.35942652061387, "learning_rate": 9.312828148969399e-06, "loss": 1.2957, "step": 712 }, { "epoch": 0.01851432810906056, "grad_norm": 11.701025658669575, "learning_rate": 9.31481817380169e-06, "loss": 1.3128, "step": 713 }, { "epoch": 0.01854029490865251, "grad_norm": 11.232881534293707, "learning_rate": 9.31680540953029e-06, "loss": 1.2767, "step": 714 }, { "epoch": 0.01856626170824446, "grad_norm": 10.595722026192819, "learning_rate": 9.31878986396235e-06, "loss": 1.36, "step": 715 }, { "epoch": 0.01859222850783641, "grad_norm": 8.54474531067582, "learning_rate": 9.320771544872288e-06, "loss": 1.2775, "step": 716 }, { "epoch": 0.01861819530742836, "grad_norm": 13.221015754031889, "learning_rate": 9.322750460001969e-06, "loss": 1.3249, "step": 717 }, { "epoch": 0.018644162107020308, "grad_norm": 8.739058711083347, "learning_rate": 9.32472661706089e-06, "loss": 1.3474, "step": 718 }, { "epoch": 0.01867012890661226, "grad_norm": 6.732240272155417, "learning_rate": 9.326700023726357e-06, "loss": 1.2855, "step": 719 }, { "epoch": 0.01869609570620421, "grad_norm": 9.009528308507436, "learning_rate": 9.328670687643661e-06, "loss": 1.3067, "step": 720 }, { "epoch": 0.01872206250579616, "grad_norm": 10.285638029330517, "learning_rate": 9.330638616426268e-06, "loss": 1.3101, "step": 721 }, { "epoch": 0.01874802930538811, "grad_norm": 11.14549715955254, "learning_rate": 9.33260381765598e-06, "loss": 1.3528, "step": 722 }, { "epoch": 0.01877399610498006, "grad_norm": 12.534631591544372, "learning_rate": 9.334566298883122e-06, "loss": 1.312, "step": 723 }, { "epoch": 0.01879996290457201, "grad_norm": 13.551120748305516, "learning_rate": 9.336526067626708e-06, "loss": 1.3116, "step": 724 }, { "epoch": 0.01882592970416396, "grad_norm": 8.36327532037973, "learning_rate": 9.33848313137462e-06, "loss": 1.3082, "step": 725 }, { "epoch": 0.018851896503755912, "grad_norm": 9.228729122861095, "learning_rate": 9.340437497583775e-06, "loss": 1.252, "step": 726 }, { "epoch": 0.018877863303347862, "grad_norm": 39.62267713025901, "learning_rate": 9.342389173680298e-06, "loss": 1.2739, "step": 727 }, { "epoch": 0.018903830102939813, "grad_norm": 48.02572196961879, "learning_rate": 9.344338167059692e-06, "loss": 1.2428, "step": 728 }, { "epoch": 0.018929796902531763, "grad_norm": 7.551323821362591, "learning_rate": 9.346284485087002e-06, "loss": 1.2825, "step": 729 }, { "epoch": 0.018955763702123714, "grad_norm": 10.56927597929785, "learning_rate": 9.348228135096991e-06, "loss": 1.3145, "step": 730 }, { "epoch": 0.018981730501715664, "grad_norm": 9.27424125755419, "learning_rate": 9.350169124394292e-06, "loss": 1.3003, "step": 731 }, { "epoch": 0.019007697301307615, "grad_norm": 9.001938928918618, "learning_rate": 9.352107460253587e-06, "loss": 1.3159, "step": 732 }, { "epoch": 0.019033664100899565, "grad_norm": 13.985923775095209, "learning_rate": 9.354043149919761e-06, "loss": 1.315, "step": 733 }, { "epoch": 0.019059630900491516, "grad_norm": 17.833782344881563, "learning_rate": 9.355976200608065e-06, "loss": 1.3294, "step": 734 }, { "epoch": 0.019085597700083466, "grad_norm": 7.499492119441192, "learning_rate": 9.357906619504287e-06, "loss": 1.225, "step": 735 }, { "epoch": 0.019111564499675417, "grad_norm": 8.104860199293856, "learning_rate": 9.359834413764898e-06, "loss": 1.3128, "step": 736 }, { "epoch": 0.019137531299267364, "grad_norm": 38.36440267459048, "learning_rate": 9.36175959051722e-06, "loss": 1.2852, "step": 737 }, { "epoch": 0.019163498098859314, "grad_norm": 8.564558909338148, "learning_rate": 9.363682156859588e-06, "loss": 1.2998, "step": 738 }, { "epoch": 0.019189464898451265, "grad_norm": 15.0431498098361, "learning_rate": 9.365602119861494e-06, "loss": 1.3233, "step": 739 }, { "epoch": 0.019215431698043215, "grad_norm": 7.23419861547596, "learning_rate": 9.367519486563755e-06, "loss": 1.3532, "step": 740 }, { "epoch": 0.019241398497635166, "grad_norm": 20.118620246309995, "learning_rate": 9.369434263978663e-06, "loss": 1.279, "step": 741 }, { "epoch": 0.019267365297227116, "grad_norm": 22.442236100300715, "learning_rate": 9.371346459090138e-06, "loss": 1.2863, "step": 742 }, { "epoch": 0.019293332096819066, "grad_norm": 10.269761922689858, "learning_rate": 9.373256078853882e-06, "loss": 1.2896, "step": 743 }, { "epoch": 0.019319298896411017, "grad_norm": 12.349139077762226, "learning_rate": 9.375163130197531e-06, "loss": 1.3074, "step": 744 }, { "epoch": 0.019345265696002967, "grad_norm": 12.604691591707839, "learning_rate": 9.37706762002081e-06, "loss": 1.261, "step": 745 }, { "epoch": 0.019371232495594918, "grad_norm": 6.887276525731672, "learning_rate": 9.378969555195668e-06, "loss": 1.2556, "step": 746 }, { "epoch": 0.01939719929518687, "grad_norm": 8.542389133572351, "learning_rate": 9.380868942566452e-06, "loss": 1.3205, "step": 747 }, { "epoch": 0.01942316609477882, "grad_norm": 8.700608278253945, "learning_rate": 9.382765788950028e-06, "loss": 1.2898, "step": 748 }, { "epoch": 0.01944913289437077, "grad_norm": 8.489947489473856, "learning_rate": 9.384660101135941e-06, "loss": 1.2905, "step": 749 }, { "epoch": 0.01947509969396272, "grad_norm": 10.502281304443157, "learning_rate": 9.38655188588657e-06, "loss": 1.292, "step": 750 }, { "epoch": 0.01950106649355467, "grad_norm": 6.105573715508463, "learning_rate": 9.388441149937248e-06, "loss": 1.2616, "step": 751 }, { "epoch": 0.01952703329314662, "grad_norm": 10.081980548836095, "learning_rate": 9.390327899996435e-06, "loss": 1.2787, "step": 752 }, { "epoch": 0.01955300009273857, "grad_norm": 13.091469796050344, "learning_rate": 9.392212142745832e-06, "loss": 1.3429, "step": 753 }, { "epoch": 0.01957896689233052, "grad_norm": 13.511628405667896, "learning_rate": 9.394093884840547e-06, "loss": 1.3442, "step": 754 }, { "epoch": 0.019604933691922472, "grad_norm": 7.892695330962823, "learning_rate": 9.395973132909223e-06, "loss": 1.292, "step": 755 }, { "epoch": 0.01963090049151442, "grad_norm": 7.252237698092822, "learning_rate": 9.397849893554185e-06, "loss": 1.2646, "step": 756 }, { "epoch": 0.01965686729110637, "grad_norm": 6.032199923169485, "learning_rate": 9.399724173351567e-06, "loss": 1.3044, "step": 757 }, { "epoch": 0.01968283409069832, "grad_norm": 7.917219069097807, "learning_rate": 9.40159597885147e-06, "loss": 1.316, "step": 758 }, { "epoch": 0.01970880089029027, "grad_norm": 33.126442701483775, "learning_rate": 9.40346531657808e-06, "loss": 1.3299, "step": 759 }, { "epoch": 0.01973476768988222, "grad_norm": 6.957160927079609, "learning_rate": 9.405332193029815e-06, "loss": 1.2675, "step": 760 }, { "epoch": 0.01976073448947417, "grad_norm": 9.926834236747744, "learning_rate": 9.407196614679458e-06, "loss": 1.3629, "step": 761 }, { "epoch": 0.019786701289066122, "grad_norm": 13.584940684571292, "learning_rate": 9.409058587974296e-06, "loss": 1.3248, "step": 762 }, { "epoch": 0.019812668088658072, "grad_norm": 10.016061853650482, "learning_rate": 9.41091811933624e-06, "loss": 1.2911, "step": 763 }, { "epoch": 0.019838634888250023, "grad_norm": 8.734101156401493, "learning_rate": 9.412775215161982e-06, "loss": 1.2621, "step": 764 }, { "epoch": 0.019864601687841973, "grad_norm": 9.316102016127902, "learning_rate": 9.414629881823095e-06, "loss": 1.3703, "step": 765 }, { "epoch": 0.019890568487433924, "grad_norm": 8.310801397027836, "learning_rate": 9.416482125666197e-06, "loss": 1.3004, "step": 766 }, { "epoch": 0.019916535287025874, "grad_norm": 9.479428152872343, "learning_rate": 9.418331953013058e-06, "loss": 1.3436, "step": 767 }, { "epoch": 0.019942502086617825, "grad_norm": 13.372233167016153, "learning_rate": 9.42017937016074e-06, "loss": 1.2542, "step": 768 }, { "epoch": 0.019968468886209775, "grad_norm": 8.900459414851019, "learning_rate": 9.42202438338172e-06, "loss": 1.2621, "step": 769 }, { "epoch": 0.019994435685801726, "grad_norm": 9.76160893983355, "learning_rate": 9.423866998924024e-06, "loss": 1.2915, "step": 770 }, { "epoch": 0.020020402485393676, "grad_norm": 6.760180108193483, "learning_rate": 9.425707223011352e-06, "loss": 1.2811, "step": 771 }, { "epoch": 0.020046369284985627, "grad_norm": 46.14461441439429, "learning_rate": 9.427545061843199e-06, "loss": 1.2612, "step": 772 }, { "epoch": 0.020072336084577577, "grad_norm": 6.536094219709947, "learning_rate": 9.429380521594988e-06, "loss": 1.2701, "step": 773 }, { "epoch": 0.020098302884169528, "grad_norm": 8.885132928561234, "learning_rate": 9.431213608418187e-06, "loss": 1.3699, "step": 774 }, { "epoch": 0.020124269683761475, "grad_norm": 10.123906205989652, "learning_rate": 9.43304432844044e-06, "loss": 1.2437, "step": 775 }, { "epoch": 0.020150236483353425, "grad_norm": 7.299895810990617, "learning_rate": 9.434872687765686e-06, "loss": 1.2887, "step": 776 }, { "epoch": 0.020176203282945376, "grad_norm": 8.515268412832219, "learning_rate": 9.436698692474278e-06, "loss": 1.352, "step": 777 }, { "epoch": 0.020202170082537326, "grad_norm": 11.817555470974964, "learning_rate": 9.438522348623111e-06, "loss": 1.2841, "step": 778 }, { "epoch": 0.020228136882129277, "grad_norm": 7.456795674235343, "learning_rate": 9.44034366224574e-06, "loss": 1.2854, "step": 779 }, { "epoch": 0.020254103681721227, "grad_norm": 12.625504587495392, "learning_rate": 9.442162639352497e-06, "loss": 1.2714, "step": 780 }, { "epoch": 0.020280070481313178, "grad_norm": 7.162390327612572, "learning_rate": 9.44397928593061e-06, "loss": 1.3103, "step": 781 }, { "epoch": 0.020306037280905128, "grad_norm": 9.687976850033104, "learning_rate": 9.445793607944333e-06, "loss": 1.3012, "step": 782 }, { "epoch": 0.02033200408049708, "grad_norm": 11.083621105005188, "learning_rate": 9.44760561133504e-06, "loss": 1.284, "step": 783 }, { "epoch": 0.02035797088008903, "grad_norm": 12.402071847413175, "learning_rate": 9.449415302021363e-06, "loss": 1.3215, "step": 784 }, { "epoch": 0.02038393767968098, "grad_norm": 8.763750878920135, "learning_rate": 9.451222685899305e-06, "loss": 1.2608, "step": 785 }, { "epoch": 0.02040990447927293, "grad_norm": 7.151611690672944, "learning_rate": 9.453027768842346e-06, "loss": 1.366, "step": 786 }, { "epoch": 0.02043587127886488, "grad_norm": 10.316221363356519, "learning_rate": 9.454830556701558e-06, "loss": 1.277, "step": 787 }, { "epoch": 0.02046183807845683, "grad_norm": 5.82706979807524, "learning_rate": 9.45663105530573e-06, "loss": 1.3201, "step": 788 }, { "epoch": 0.02048780487804878, "grad_norm": 9.118958445853728, "learning_rate": 9.45842927046147e-06, "loss": 1.3263, "step": 789 }, { "epoch": 0.020513771677640732, "grad_norm": 13.75415309260886, "learning_rate": 9.460225207953329e-06, "loss": 1.3388, "step": 790 }, { "epoch": 0.020539738477232682, "grad_norm": 7.998784926645695, "learning_rate": 9.462018873543895e-06, "loss": 1.2874, "step": 791 }, { "epoch": 0.020565705276824633, "grad_norm": 11.592095516635329, "learning_rate": 9.463810272973921e-06, "loss": 1.3074, "step": 792 }, { "epoch": 0.020591672076416583, "grad_norm": 11.969750106102428, "learning_rate": 9.465599411962424e-06, "loss": 1.2917, "step": 793 }, { "epoch": 0.02061763887600853, "grad_norm": 15.218198041678331, "learning_rate": 9.467386296206804e-06, "loss": 1.2511, "step": 794 }, { "epoch": 0.02064360567560048, "grad_norm": 6.054787990597522, "learning_rate": 9.469170931382943e-06, "loss": 1.2647, "step": 795 }, { "epoch": 0.02066957247519243, "grad_norm": 6.2885017097606095, "learning_rate": 9.470953323145324e-06, "loss": 1.2985, "step": 796 }, { "epoch": 0.02069553927478438, "grad_norm": 12.999428956621218, "learning_rate": 9.472733477127124e-06, "loss": 1.3291, "step": 797 }, { "epoch": 0.020721506074376332, "grad_norm": 6.862496953800269, "learning_rate": 9.474511398940337e-06, "loss": 1.2704, "step": 798 }, { "epoch": 0.020747472873968283, "grad_norm": 6.035594581068338, "learning_rate": 9.47628709417587e-06, "loss": 1.2695, "step": 799 }, { "epoch": 0.020773439673560233, "grad_norm": 13.75671389264251, "learning_rate": 9.478060568403647e-06, "loss": 1.3492, "step": 800 }, { "epoch": 0.020799406473152184, "grad_norm": 7.501462920882109, "learning_rate": 9.479831827172727e-06, "loss": 1.2786, "step": 801 }, { "epoch": 0.020825373272744134, "grad_norm": 8.663390151577905, "learning_rate": 9.481600876011389e-06, "loss": 1.3202, "step": 802 }, { "epoch": 0.020851340072336084, "grad_norm": 8.690349742668722, "learning_rate": 9.48336772042725e-06, "loss": 1.3365, "step": 803 }, { "epoch": 0.020877306871928035, "grad_norm": 15.078268194753988, "learning_rate": 9.485132365907367e-06, "loss": 1.2758, "step": 804 }, { "epoch": 0.020903273671519985, "grad_norm": 7.777259876023124, "learning_rate": 9.486894817918328e-06, "loss": 1.3495, "step": 805 }, { "epoch": 0.020929240471111936, "grad_norm": 7.724683203266029, "learning_rate": 9.488655081906368e-06, "loss": 1.3174, "step": 806 }, { "epoch": 0.020955207270703886, "grad_norm": 11.227759523758062, "learning_rate": 9.490413163297458e-06, "loss": 1.2436, "step": 807 }, { "epoch": 0.020981174070295837, "grad_norm": 9.121196633520393, "learning_rate": 9.492169067497413e-06, "loss": 1.3519, "step": 808 }, { "epoch": 0.021007140869887787, "grad_norm": 6.272152817849625, "learning_rate": 9.493922799891992e-06, "loss": 1.3085, "step": 809 }, { "epoch": 0.021033107669479738, "grad_norm": 12.326125917211503, "learning_rate": 9.495674365846988e-06, "loss": 1.3724, "step": 810 }, { "epoch": 0.02105907446907169, "grad_norm": 5.911233081531251, "learning_rate": 9.497423770708341e-06, "loss": 1.2948, "step": 811 }, { "epoch": 0.02108504126866364, "grad_norm": 7.176595626080198, "learning_rate": 9.49917101980222e-06, "loss": 1.2966, "step": 812 }, { "epoch": 0.02111100806825559, "grad_norm": 11.241405340559592, "learning_rate": 9.500916118435131e-06, "loss": 1.3274, "step": 813 }, { "epoch": 0.021136974867847536, "grad_norm": 7.827668136696391, "learning_rate": 9.502659071894014e-06, "loss": 1.2669, "step": 814 }, { "epoch": 0.021162941667439487, "grad_norm": 7.770025652790313, "learning_rate": 9.504399885446334e-06, "loss": 1.3116, "step": 815 }, { "epoch": 0.021188908467031437, "grad_norm": 8.230935929431967, "learning_rate": 9.506138564340174e-06, "loss": 1.3104, "step": 816 }, { "epoch": 0.021214875266623388, "grad_norm": 6.358543847808392, "learning_rate": 9.50787511380434e-06, "loss": 1.3019, "step": 817 }, { "epoch": 0.021240842066215338, "grad_norm": 5.515899219126165, "learning_rate": 9.509609539048448e-06, "loss": 1.2849, "step": 818 }, { "epoch": 0.02126680886580729, "grad_norm": 6.932183780494624, "learning_rate": 9.51134184526302e-06, "loss": 1.2885, "step": 819 }, { "epoch": 0.02129277566539924, "grad_norm": 7.265344101592903, "learning_rate": 9.513072037619573e-06, "loss": 1.2536, "step": 820 }, { "epoch": 0.02131874246499119, "grad_norm": 7.585652873457997, "learning_rate": 9.514800121270723e-06, "loss": 1.2911, "step": 821 }, { "epoch": 0.02134470926458314, "grad_norm": 9.816762982226818, "learning_rate": 9.51652610135026e-06, "loss": 1.3065, "step": 822 }, { "epoch": 0.02137067606417509, "grad_norm": 55.033406048054125, "learning_rate": 9.518249982973259e-06, "loss": 1.3019, "step": 823 }, { "epoch": 0.02139664286376704, "grad_norm": 7.799871874025926, "learning_rate": 9.519971771236152e-06, "loss": 1.3562, "step": 824 }, { "epoch": 0.02142260966335899, "grad_norm": 10.481865982034497, "learning_rate": 9.521691471216828e-06, "loss": 1.2744, "step": 825 }, { "epoch": 0.021448576462950942, "grad_norm": 5.641024673123932, "learning_rate": 9.523409087974731e-06, "loss": 1.2588, "step": 826 }, { "epoch": 0.021474543262542892, "grad_norm": 5.960853666063499, "learning_rate": 9.52512462655093e-06, "loss": 1.2903, "step": 827 }, { "epoch": 0.021500510062134843, "grad_norm": 5.084301382140283, "learning_rate": 9.526838091968224e-06, "loss": 1.3001, "step": 828 }, { "epoch": 0.021526476861726793, "grad_norm": 10.903577766846796, "learning_rate": 9.528549489231225e-06, "loss": 1.3012, "step": 829 }, { "epoch": 0.021552443661318744, "grad_norm": 6.290636024522288, "learning_rate": 9.530258823326437e-06, "loss": 1.2671, "step": 830 }, { "epoch": 0.021578410460910694, "grad_norm": 5.72639163232039, "learning_rate": 9.531966099222364e-06, "loss": 1.364, "step": 831 }, { "epoch": 0.021604377260502645, "grad_norm": 8.199991961219382, "learning_rate": 9.533671321869575e-06, "loss": 1.3145, "step": 832 }, { "epoch": 0.021630344060094592, "grad_norm": 8.670850868580098, "learning_rate": 9.5353744962008e-06, "loss": 1.2914, "step": 833 }, { "epoch": 0.021656310859686542, "grad_norm": 7.603732607920527, "learning_rate": 9.537075627131016e-06, "loss": 1.2973, "step": 834 }, { "epoch": 0.021682277659278493, "grad_norm": 9.550542798020247, "learning_rate": 9.538774719557533e-06, "loss": 1.3278, "step": 835 }, { "epoch": 0.021708244458870443, "grad_norm": 6.90376478678128, "learning_rate": 9.540471778360073e-06, "loss": 1.2987, "step": 836 }, { "epoch": 0.021734211258462394, "grad_norm": 6.150753142598995, "learning_rate": 9.54216680840086e-06, "loss": 1.3168, "step": 837 }, { "epoch": 0.021760178058054344, "grad_norm": 6.09144932280446, "learning_rate": 9.543859814524698e-06, "loss": 1.2922, "step": 838 }, { "epoch": 0.021786144857646295, "grad_norm": 12.822298796843992, "learning_rate": 9.54555080155906e-06, "loss": 1.3162, "step": 839 }, { "epoch": 0.021812111657238245, "grad_norm": 8.13494030237128, "learning_rate": 9.54723977431417e-06, "loss": 1.2857, "step": 840 }, { "epoch": 0.021838078456830196, "grad_norm": 7.570909096718917, "learning_rate": 9.548926737583076e-06, "loss": 1.302, "step": 841 }, { "epoch": 0.021864045256422146, "grad_norm": 6.438118321621105, "learning_rate": 9.55061169614174e-06, "loss": 1.3208, "step": 842 }, { "epoch": 0.021890012056014096, "grad_norm": 7.3194385442482774, "learning_rate": 9.552294654749126e-06, "loss": 1.3066, "step": 843 }, { "epoch": 0.021915978855606047, "grad_norm": 5.81334393024937, "learning_rate": 9.55397561814726e-06, "loss": 1.284, "step": 844 }, { "epoch": 0.021941945655197997, "grad_norm": 5.669002425557477, "learning_rate": 9.555654591061334e-06, "loss": 1.2842, "step": 845 }, { "epoch": 0.021967912454789948, "grad_norm": 9.83168981578391, "learning_rate": 9.557331578199763e-06, "loss": 1.2914, "step": 846 }, { "epoch": 0.0219938792543819, "grad_norm": 5.004403654654639, "learning_rate": 9.559006584254283e-06, "loss": 1.2675, "step": 847 }, { "epoch": 0.02201984605397385, "grad_norm": 5.598515531787687, "learning_rate": 9.560679613900021e-06, "loss": 1.2897, "step": 848 }, { "epoch": 0.0220458128535658, "grad_norm": 13.405993336886036, "learning_rate": 9.562350671795573e-06, "loss": 1.2598, "step": 849 }, { "epoch": 0.02207177965315775, "grad_norm": 8.810016688007392, "learning_rate": 9.564019762583084e-06, "loss": 1.2713, "step": 850 }, { "epoch": 0.0220977464527497, "grad_norm": 4.822900803651806, "learning_rate": 9.565686890888321e-06, "loss": 1.2995, "step": 851 }, { "epoch": 0.022123713252341647, "grad_norm": 8.380725782187989, "learning_rate": 9.567352061320757e-06, "loss": 1.3437, "step": 852 }, { "epoch": 0.022149680051933598, "grad_norm": 15.632187799862061, "learning_rate": 9.569015278473649e-06, "loss": 1.3065, "step": 853 }, { "epoch": 0.022175646851525548, "grad_norm": 6.554750935516936, "learning_rate": 9.570676546924098e-06, "loss": 1.271, "step": 854 }, { "epoch": 0.0222016136511175, "grad_norm": 5.467269958763876, "learning_rate": 9.572335871233142e-06, "loss": 1.2906, "step": 855 }, { "epoch": 0.02222758045070945, "grad_norm": 7.0094679426203985, "learning_rate": 9.573993255945826e-06, "loss": 1.3023, "step": 856 }, { "epoch": 0.0222535472503014, "grad_norm": 5.637045445693295, "learning_rate": 9.575648705591272e-06, "loss": 1.2869, "step": 857 }, { "epoch": 0.02227951404989335, "grad_norm": 7.239452713993271, "learning_rate": 9.577302224682756e-06, "loss": 1.3024, "step": 858 }, { "epoch": 0.0223054808494853, "grad_norm": 10.993978412764509, "learning_rate": 9.578953817717786e-06, "loss": 1.2899, "step": 859 }, { "epoch": 0.02233144764907725, "grad_norm": 5.587018955354914, "learning_rate": 9.580603489178172e-06, "loss": 1.3164, "step": 860 }, { "epoch": 0.0223574144486692, "grad_norm": 9.673838574784886, "learning_rate": 9.582251243530096e-06, "loss": 1.2987, "step": 861 }, { "epoch": 0.022383381248261152, "grad_norm": 12.867992679346035, "learning_rate": 9.583897085224188e-06, "loss": 1.2432, "step": 862 }, { "epoch": 0.022409348047853102, "grad_norm": 5.948304314571507, "learning_rate": 9.585541018695598e-06, "loss": 1.2896, "step": 863 }, { "epoch": 0.022435314847445053, "grad_norm": 6.3771274980945085, "learning_rate": 9.587183048364066e-06, "loss": 1.318, "step": 864 }, { "epoch": 0.022461281647037003, "grad_norm": 4.9450878315252424, "learning_rate": 9.588823178633999e-06, "loss": 1.2777, "step": 865 }, { "epoch": 0.022487248446628954, "grad_norm": 7.322631338393158, "learning_rate": 9.590461413894532e-06, "loss": 1.2895, "step": 866 }, { "epoch": 0.022513215246220904, "grad_norm": 5.996004999852148, "learning_rate": 9.592097758519608e-06, "loss": 1.2961, "step": 867 }, { "epoch": 0.022539182045812855, "grad_norm": 15.872299325058666, "learning_rate": 9.593732216868041e-06, "loss": 1.3118, "step": 868 }, { "epoch": 0.022565148845404805, "grad_norm": 10.305991228416378, "learning_rate": 9.595364793283588e-06, "loss": 1.3102, "step": 869 }, { "epoch": 0.022591115644996756, "grad_norm": 7.229728244512511, "learning_rate": 9.596995492095025e-06, "loss": 1.2981, "step": 870 }, { "epoch": 0.022617082444588703, "grad_norm": 25.113071931484956, "learning_rate": 9.598624317616204e-06, "loss": 1.3444, "step": 871 }, { "epoch": 0.022643049244180653, "grad_norm": 11.490754100996606, "learning_rate": 9.600251274146124e-06, "loss": 1.3223, "step": 872 }, { "epoch": 0.022669016043772604, "grad_norm": 4.9926604761761135, "learning_rate": 9.601876365969014e-06, "loss": 1.2855, "step": 873 }, { "epoch": 0.022694982843364554, "grad_norm": 4.931973418536181, "learning_rate": 9.603499597354378e-06, "loss": 1.3321, "step": 874 }, { "epoch": 0.022720949642956505, "grad_norm": 5.436327928522666, "learning_rate": 9.605120972557077e-06, "loss": 1.3053, "step": 875 }, { "epoch": 0.022746916442548455, "grad_norm": 8.78772255410856, "learning_rate": 9.606740495817399e-06, "loss": 1.3042, "step": 876 }, { "epoch": 0.022772883242140406, "grad_norm": 8.863838161492176, "learning_rate": 9.608358171361102e-06, "loss": 1.2963, "step": 877 }, { "epoch": 0.022798850041732356, "grad_norm": 5.743450249123793, "learning_rate": 9.609974003399514e-06, "loss": 1.3559, "step": 878 }, { "epoch": 0.022824816841324307, "grad_norm": 5.4974475486511825, "learning_rate": 9.611587996129575e-06, "loss": 1.2636, "step": 879 }, { "epoch": 0.022850783640916257, "grad_norm": 7.229712415280847, "learning_rate": 9.613200153733906e-06, "loss": 1.3024, "step": 880 }, { "epoch": 0.022876750440508208, "grad_norm": 9.665796394197534, "learning_rate": 9.614810480380884e-06, "loss": 1.3013, "step": 881 }, { "epoch": 0.022902717240100158, "grad_norm": 7.92096291131216, "learning_rate": 9.616418980224693e-06, "loss": 1.3147, "step": 882 }, { "epoch": 0.02292868403969211, "grad_norm": 6.818307903164356, "learning_rate": 9.6180256574054e-06, "loss": 1.29, "step": 883 }, { "epoch": 0.02295465083928406, "grad_norm": 5.7847685133399835, "learning_rate": 9.61963051604901e-06, "loss": 1.2674, "step": 884 }, { "epoch": 0.02298061763887601, "grad_norm": 6.394619837982118, "learning_rate": 9.621233560267536e-06, "loss": 1.3388, "step": 885 }, { "epoch": 0.02300658443846796, "grad_norm": 5.668685478676017, "learning_rate": 9.622834794159064e-06, "loss": 1.2962, "step": 886 }, { "epoch": 0.02303255123805991, "grad_norm": 9.472242258994749, "learning_rate": 9.624434221807802e-06, "loss": 1.3117, "step": 887 }, { "epoch": 0.02305851803765186, "grad_norm": 11.334118161421749, "learning_rate": 9.626031847284162e-06, "loss": 1.2593, "step": 888 }, { "epoch": 0.02308448483724381, "grad_norm": 6.080792074810655, "learning_rate": 9.627627674644803e-06, "loss": 1.3573, "step": 889 }, { "epoch": 0.02311045163683576, "grad_norm": 8.550900363920247, "learning_rate": 9.629221707932712e-06, "loss": 1.2709, "step": 890 }, { "epoch": 0.02313641843642771, "grad_norm": 8.165138049666838, "learning_rate": 9.630813951177247e-06, "loss": 1.246, "step": 891 }, { "epoch": 0.02316238523601966, "grad_norm": 5.552837423199632, "learning_rate": 9.632404408394215e-06, "loss": 1.2925, "step": 892 }, { "epoch": 0.02318835203561161, "grad_norm": 18.042003367088892, "learning_rate": 9.633993083585915e-06, "loss": 1.2752, "step": 893 }, { "epoch": 0.02321431883520356, "grad_norm": 6.71109239534936, "learning_rate": 9.635579980741213e-06, "loss": 1.3258, "step": 894 }, { "epoch": 0.02324028563479551, "grad_norm": 6.9044891460144715, "learning_rate": 9.637165103835603e-06, "loss": 1.2974, "step": 895 }, { "epoch": 0.02326625243438746, "grad_norm": 9.853750035925806, "learning_rate": 9.638748456831248e-06, "loss": 1.3332, "step": 896 }, { "epoch": 0.02329221923397941, "grad_norm": 9.558153947534306, "learning_rate": 9.640330043677061e-06, "loss": 1.3096, "step": 897 }, { "epoch": 0.023318186033571362, "grad_norm": 8.610965533376161, "learning_rate": 9.641909868308754e-06, "loss": 1.3152, "step": 898 }, { "epoch": 0.023344152833163313, "grad_norm": 7.142766385183039, "learning_rate": 9.643487934648895e-06, "loss": 1.2709, "step": 899 }, { "epoch": 0.023370119632755263, "grad_norm": 6.166920046712851, "learning_rate": 9.645064246606975e-06, "loss": 1.35, "step": 900 }, { "epoch": 0.023396086432347214, "grad_norm": 5.530068481462772, "learning_rate": 9.646638808079456e-06, "loss": 1.2657, "step": 901 }, { "epoch": 0.023422053231939164, "grad_norm": 10.349556851690181, "learning_rate": 9.648211622949834e-06, "loss": 1.3126, "step": 902 }, { "epoch": 0.023448020031531114, "grad_norm": 16.52874194555455, "learning_rate": 9.649782695088697e-06, "loss": 1.3455, "step": 903 }, { "epoch": 0.023473986831123065, "grad_norm": 17.179491209372024, "learning_rate": 9.651352028353778e-06, "loss": 1.2803, "step": 904 }, { "epoch": 0.023499953630715015, "grad_norm": 15.271347862967362, "learning_rate": 9.652919626590021e-06, "loss": 1.2948, "step": 905 }, { "epoch": 0.023525920430306966, "grad_norm": 6.534933557267464, "learning_rate": 9.654485493629629e-06, "loss": 1.3214, "step": 906 }, { "epoch": 0.023551887229898916, "grad_norm": 18.19862974269543, "learning_rate": 9.656049633292117e-06, "loss": 1.3455, "step": 907 }, { "epoch": 0.023577854029490867, "grad_norm": 8.118425262041375, "learning_rate": 9.657612049384382e-06, "loss": 1.2953, "step": 908 }, { "epoch": 0.023603820829082814, "grad_norm": 7.907715431677386, "learning_rate": 9.659172745700743e-06, "loss": 1.2986, "step": 909 }, { "epoch": 0.023629787628674764, "grad_norm": 6.806759716945963, "learning_rate": 9.660731726023005e-06, "loss": 1.2975, "step": 910 }, { "epoch": 0.023655754428266715, "grad_norm": 11.35526651544202, "learning_rate": 9.66228899412052e-06, "loss": 1.3155, "step": 911 }, { "epoch": 0.023681721227858665, "grad_norm": 8.15906040691162, "learning_rate": 9.66384455375022e-06, "loss": 1.3255, "step": 912 }, { "epoch": 0.023707688027450616, "grad_norm": 6.326080217960587, "learning_rate": 9.665398408656697e-06, "loss": 1.3095, "step": 913 }, { "epoch": 0.023733654827042566, "grad_norm": 6.377296781991996, "learning_rate": 9.666950562572242e-06, "loss": 1.3069, "step": 914 }, { "epoch": 0.023759621626634517, "grad_norm": 5.0605873921935505, "learning_rate": 9.668501019216903e-06, "loss": 1.2717, "step": 915 }, { "epoch": 0.023785588426226467, "grad_norm": 7.202311505039545, "learning_rate": 9.670049782298533e-06, "loss": 1.2766, "step": 916 }, { "epoch": 0.023811555225818418, "grad_norm": 5.985144025428875, "learning_rate": 9.671596855512855e-06, "loss": 1.2753, "step": 917 }, { "epoch": 0.023837522025410368, "grad_norm": 4.749374348444241, "learning_rate": 9.673142242543502e-06, "loss": 1.2848, "step": 918 }, { "epoch": 0.02386348882500232, "grad_norm": 6.7693010523288555, "learning_rate": 9.67468594706208e-06, "loss": 1.3632, "step": 919 }, { "epoch": 0.02388945562459427, "grad_norm": 7.2012825883135605, "learning_rate": 9.676227972728213e-06, "loss": 1.2742, "step": 920 }, { "epoch": 0.02391542242418622, "grad_norm": 6.82820991795022, "learning_rate": 9.677768323189597e-06, "loss": 1.3219, "step": 921 }, { "epoch": 0.02394138922377817, "grad_norm": 6.974731068229161, "learning_rate": 9.67930700208206e-06, "loss": 1.3178, "step": 922 }, { "epoch": 0.02396735602337012, "grad_norm": 15.13070596302278, "learning_rate": 9.680844013029594e-06, "loss": 1.284, "step": 923 }, { "epoch": 0.02399332282296207, "grad_norm": 6.438319181893629, "learning_rate": 9.68237935964443e-06, "loss": 1.2973, "step": 924 }, { "epoch": 0.02401928962255402, "grad_norm": 6.625264576361242, "learning_rate": 9.68391304552707e-06, "loss": 1.2586, "step": 925 }, { "epoch": 0.024045256422145972, "grad_norm": 15.059806614580237, "learning_rate": 9.68544507426635e-06, "loss": 1.3258, "step": 926 }, { "epoch": 0.024071223221737922, "grad_norm": 8.527297225594568, "learning_rate": 9.68697544943948e-06, "loss": 1.2771, "step": 927 }, { "epoch": 0.02409719002132987, "grad_norm": 5.311482410584564, "learning_rate": 9.688504174612103e-06, "loss": 1.2686, "step": 928 }, { "epoch": 0.02412315682092182, "grad_norm": 7.752117729025274, "learning_rate": 9.69003125333834e-06, "loss": 1.2727, "step": 929 }, { "epoch": 0.02414912362051377, "grad_norm": 4.8332415385682514, "learning_rate": 9.691556689160846e-06, "loss": 1.3241, "step": 930 }, { "epoch": 0.02417509042010572, "grad_norm": 6.002508274948092, "learning_rate": 9.693080485610847e-06, "loss": 1.3002, "step": 931 }, { "epoch": 0.02420105721969767, "grad_norm": 6.7418202889208985, "learning_rate": 9.694602646208196e-06, "loss": 1.2637, "step": 932 }, { "epoch": 0.024227024019289622, "grad_norm": 4.9300360941774635, "learning_rate": 9.696123174461428e-06, "loss": 1.3186, "step": 933 }, { "epoch": 0.024252990818881572, "grad_norm": 4.788797841063375, "learning_rate": 9.697642073867802e-06, "loss": 1.3139, "step": 934 }, { "epoch": 0.024278957618473523, "grad_norm": 25.32759725313847, "learning_rate": 9.699159347913341e-06, "loss": 1.3188, "step": 935 }, { "epoch": 0.024304924418065473, "grad_norm": 6.124397403891343, "learning_rate": 9.700675000072903e-06, "loss": 1.2059, "step": 936 }, { "epoch": 0.024330891217657424, "grad_norm": 6.15710195465208, "learning_rate": 9.702189033810203e-06, "loss": 1.2634, "step": 937 }, { "epoch": 0.024356858017249374, "grad_norm": 5.224370431159867, "learning_rate": 9.703701452577877e-06, "loss": 1.2649, "step": 938 }, { "epoch": 0.024382824816841325, "grad_norm": 9.860583247949691, "learning_rate": 9.705212259817522e-06, "loss": 1.3026, "step": 939 }, { "epoch": 0.024408791616433275, "grad_norm": 9.254174192775624, "learning_rate": 9.70672145895975e-06, "loss": 1.2627, "step": 940 }, { "epoch": 0.024434758416025226, "grad_norm": 10.661632621519558, "learning_rate": 9.708229053424222e-06, "loss": 1.2983, "step": 941 }, { "epoch": 0.024460725215617176, "grad_norm": 5.977909110957374, "learning_rate": 9.709735046619713e-06, "loss": 1.3384, "step": 942 }, { "epoch": 0.024486692015209126, "grad_norm": 5.354247017144938, "learning_rate": 9.711239441944137e-06, "loss": 1.2425, "step": 943 }, { "epoch": 0.024512658814801077, "grad_norm": 4.443145813892533, "learning_rate": 9.712742242784614e-06, "loss": 1.2567, "step": 944 }, { "epoch": 0.024538625614393027, "grad_norm": 16.416654996658025, "learning_rate": 9.7142434525175e-06, "loss": 1.3159, "step": 945 }, { "epoch": 0.024564592413984978, "grad_norm": 6.161107040111931, "learning_rate": 9.715743074508432e-06, "loss": 1.3107, "step": 946 }, { "epoch": 0.024590559213576925, "grad_norm": 5.047379506880239, "learning_rate": 9.717241112112394e-06, "loss": 1.2834, "step": 947 }, { "epoch": 0.024616526013168875, "grad_norm": 7.430957384821238, "learning_rate": 9.718737568673735e-06, "loss": 1.2714, "step": 948 }, { "epoch": 0.024642492812760826, "grad_norm": 13.865081816362068, "learning_rate": 9.720232447526234e-06, "loss": 1.2922, "step": 949 }, { "epoch": 0.024668459612352776, "grad_norm": 7.719543416253316, "learning_rate": 9.721725751993129e-06, "loss": 1.318, "step": 950 }, { "epoch": 0.024694426411944727, "grad_norm": 6.456343247810244, "learning_rate": 9.723217485387176e-06, "loss": 1.2868, "step": 951 }, { "epoch": 0.024720393211536677, "grad_norm": 6.514205375453248, "learning_rate": 9.724707651010681e-06, "loss": 1.2431, "step": 952 }, { "epoch": 0.024746360011128628, "grad_norm": 6.89294754296579, "learning_rate": 9.726196252155556e-06, "loss": 1.292, "step": 953 }, { "epoch": 0.024772326810720578, "grad_norm": 8.456403790955324, "learning_rate": 9.72768329210335e-06, "loss": 1.2675, "step": 954 }, { "epoch": 0.02479829361031253, "grad_norm": 6.805975072823477, "learning_rate": 9.729168774125295e-06, "loss": 1.2808, "step": 955 }, { "epoch": 0.02482426040990448, "grad_norm": 4.884198240863559, "learning_rate": 9.730652701482362e-06, "loss": 1.305, "step": 956 }, { "epoch": 0.02485022720949643, "grad_norm": 10.604731716447711, "learning_rate": 9.732135077425285e-06, "loss": 1.2619, "step": 957 }, { "epoch": 0.02487619400908838, "grad_norm": 5.267105573561905, "learning_rate": 9.733615905194618e-06, "loss": 1.2869, "step": 958 }, { "epoch": 0.02490216080868033, "grad_norm": 6.79009095930207, "learning_rate": 9.735095188020771e-06, "loss": 1.3256, "step": 959 }, { "epoch": 0.02492812760827228, "grad_norm": 5.721454585332429, "learning_rate": 9.736572929124054e-06, "loss": 1.3219, "step": 960 }, { "epoch": 0.02495409440786423, "grad_norm": 5.676630385875216, "learning_rate": 9.738049131714717e-06, "loss": 1.2697, "step": 961 }, { "epoch": 0.024980061207456182, "grad_norm": 19.56131748123225, "learning_rate": 9.739523798992998e-06, "loss": 1.2768, "step": 962 }, { "epoch": 0.025006028007048132, "grad_norm": 6.551889033587738, "learning_rate": 9.740996934149154e-06, "loss": 1.2909, "step": 963 }, { "epoch": 0.025031994806640083, "grad_norm": 6.254231355736228, "learning_rate": 9.742468540363514e-06, "loss": 1.2957, "step": 964 }, { "epoch": 0.025057961606232033, "grad_norm": 6.289857044141956, "learning_rate": 9.743938620806514e-06, "loss": 1.2626, "step": 965 }, { "epoch": 0.025083928405823984, "grad_norm": 7.581360808440901, "learning_rate": 9.745407178638734e-06, "loss": 1.316, "step": 966 }, { "epoch": 0.02510989520541593, "grad_norm": 9.473038813151181, "learning_rate": 9.746874217010951e-06, "loss": 1.2473, "step": 967 }, { "epoch": 0.02513586200500788, "grad_norm": 4.546115457863514, "learning_rate": 9.748339739064166e-06, "loss": 1.3249, "step": 968 }, { "epoch": 0.025161828804599832, "grad_norm": 8.662045298170554, "learning_rate": 9.749803747929655e-06, "loss": 1.3347, "step": 969 }, { "epoch": 0.025187795604191782, "grad_norm": 7.319306425578429, "learning_rate": 9.751266246729001e-06, "loss": 1.2844, "step": 970 }, { "epoch": 0.025213762403783733, "grad_norm": 5.191683817060053, "learning_rate": 9.752727238574139e-06, "loss": 1.3249, "step": 971 }, { "epoch": 0.025239729203375683, "grad_norm": 7.46033568826115, "learning_rate": 9.754186726567395e-06, "loss": 1.3356, "step": 972 }, { "epoch": 0.025265696002967634, "grad_norm": 4.9832783512072005, "learning_rate": 9.755644713801525e-06, "loss": 1.3125, "step": 973 }, { "epoch": 0.025291662802559584, "grad_norm": 6.483181542562136, "learning_rate": 9.757101203359754e-06, "loss": 1.2658, "step": 974 }, { "epoch": 0.025317629602151535, "grad_norm": 5.959667065429219, "learning_rate": 9.758556198315812e-06, "loss": 1.2394, "step": 975 }, { "epoch": 0.025343596401743485, "grad_norm": 9.80932254734702, "learning_rate": 9.76000970173398e-06, "loss": 1.3495, "step": 976 }, { "epoch": 0.025369563201335436, "grad_norm": 4.057787465402776, "learning_rate": 9.761461716669126e-06, "loss": 1.3405, "step": 977 }, { "epoch": 0.025395530000927386, "grad_norm": 10.611116561311272, "learning_rate": 9.76291224616674e-06, "loss": 1.2725, "step": 978 }, { "epoch": 0.025421496800519337, "grad_norm": 6.494670737292218, "learning_rate": 9.764361293262972e-06, "loss": 1.2891, "step": 979 }, { "epoch": 0.025447463600111287, "grad_norm": 9.537197171224495, "learning_rate": 9.765808860984678e-06, "loss": 1.3093, "step": 980 }, { "epoch": 0.025473430399703238, "grad_norm": 5.815006167014108, "learning_rate": 9.767254952349454e-06, "loss": 1.3046, "step": 981 }, { "epoch": 0.025499397199295188, "grad_norm": 7.86228521930036, "learning_rate": 9.768699570365663e-06, "loss": 1.3013, "step": 982 }, { "epoch": 0.02552536399888714, "grad_norm": 5.721077867335082, "learning_rate": 9.770142718032494e-06, "loss": 1.2978, "step": 983 }, { "epoch": 0.02555133079847909, "grad_norm": 5.856146245829168, "learning_rate": 9.771584398339979e-06, "loss": 1.2845, "step": 984 }, { "epoch": 0.02557729759807104, "grad_norm": 8.408391176395932, "learning_rate": 9.773024614269044e-06, "loss": 1.301, "step": 985 }, { "epoch": 0.025603264397662986, "grad_norm": 5.087963538645011, "learning_rate": 9.774463368791538e-06, "loss": 1.3114, "step": 986 }, { "epoch": 0.025629231197254937, "grad_norm": 9.0692643801184, "learning_rate": 9.775900664870271e-06, "loss": 1.2596, "step": 987 }, { "epoch": 0.025655197996846887, "grad_norm": 12.134828339860874, "learning_rate": 9.777336505459056e-06, "loss": 1.2722, "step": 988 }, { "epoch": 0.025681164796438838, "grad_norm": 6.718353401721282, "learning_rate": 9.778770893502738e-06, "loss": 1.2841, "step": 989 }, { "epoch": 0.02570713159603079, "grad_norm": 6.623555151813948, "learning_rate": 9.780203831937234e-06, "loss": 1.2548, "step": 990 }, { "epoch": 0.02573309839562274, "grad_norm": 10.501194840522352, "learning_rate": 9.781635323689574e-06, "loss": 1.3054, "step": 991 }, { "epoch": 0.02575906519521469, "grad_norm": 7.389525511172787, "learning_rate": 9.783065371677924e-06, "loss": 1.3585, "step": 992 }, { "epoch": 0.02578503199480664, "grad_norm": 7.097115008429206, "learning_rate": 9.784493978811635e-06, "loss": 1.2735, "step": 993 }, { "epoch": 0.02581099879439859, "grad_norm": 5.317230272863703, "learning_rate": 9.785921147991267e-06, "loss": 1.2838, "step": 994 }, { "epoch": 0.02583696559399054, "grad_norm": 6.095939047421082, "learning_rate": 9.787346882108637e-06, "loss": 1.2572, "step": 995 }, { "epoch": 0.02586293239358249, "grad_norm": 6.978365025575531, "learning_rate": 9.788771184046845e-06, "loss": 1.2687, "step": 996 }, { "epoch": 0.02588889919317444, "grad_norm": 13.259350626192834, "learning_rate": 9.790194056680308e-06, "loss": 1.2572, "step": 997 }, { "epoch": 0.025914865992766392, "grad_norm": 9.119014919686707, "learning_rate": 9.791615502874801e-06, "loss": 1.3067, "step": 998 }, { "epoch": 0.025940832792358343, "grad_norm": 7.927527463270512, "learning_rate": 9.79303552548749e-06, "loss": 1.2221, "step": 999 }, { "epoch": 0.025966799591950293, "grad_norm": 6.758283547979663, "learning_rate": 9.794454127366962e-06, "loss": 1.2497, "step": 1000 }, { "epoch": 0.025992766391542244, "grad_norm": 7.403770836629018, "learning_rate": 9.795871311353265e-06, "loss": 1.3469, "step": 1001 }, { "epoch": 0.026018733191134194, "grad_norm": 15.645674798249456, "learning_rate": 9.797287080277939e-06, "loss": 1.2602, "step": 1002 }, { "epoch": 0.026044699990726145, "grad_norm": 7.0923842405137565, "learning_rate": 9.79870143696405e-06, "loss": 1.3024, "step": 1003 }, { "epoch": 0.026070666790318095, "grad_norm": 15.947687939333136, "learning_rate": 9.800114384226223e-06, "loss": 1.2514, "step": 1004 }, { "epoch": 0.026096633589910042, "grad_norm": 8.867820737301034, "learning_rate": 9.801525924870682e-06, "loss": 1.2825, "step": 1005 }, { "epoch": 0.026122600389501992, "grad_norm": 21.461414577651404, "learning_rate": 9.802936061695272e-06, "loss": 1.2828, "step": 1006 }, { "epoch": 0.026148567189093943, "grad_norm": 8.469150435211214, "learning_rate": 9.804344797489503e-06, "loss": 1.2981, "step": 1007 }, { "epoch": 0.026174533988685893, "grad_norm": 10.420956595473665, "learning_rate": 9.805752135034576e-06, "loss": 1.3322, "step": 1008 }, { "epoch": 0.026200500788277844, "grad_norm": 6.847053462131083, "learning_rate": 9.807158077103422e-06, "loss": 1.3094, "step": 1009 }, { "epoch": 0.026226467587869794, "grad_norm": 10.458620186528178, "learning_rate": 9.808562626460728e-06, "loss": 1.363, "step": 1010 }, { "epoch": 0.026252434387461745, "grad_norm": 9.791115949578268, "learning_rate": 9.809965785862977e-06, "loss": 1.3119, "step": 1011 }, { "epoch": 0.026278401187053695, "grad_norm": 7.743233218433294, "learning_rate": 9.81136755805847e-06, "loss": 1.2529, "step": 1012 }, { "epoch": 0.026304367986645646, "grad_norm": 6.762522631482175, "learning_rate": 9.812767945787375e-06, "loss": 1.2703, "step": 1013 }, { "epoch": 0.026330334786237596, "grad_norm": 5.997442336129237, "learning_rate": 9.81416695178174e-06, "loss": 1.2476, "step": 1014 }, { "epoch": 0.026356301585829547, "grad_norm": 6.057686379305182, "learning_rate": 9.815564578765534e-06, "loss": 1.2868, "step": 1015 }, { "epoch": 0.026382268385421497, "grad_norm": 6.9888953274426155, "learning_rate": 9.816960829454688e-06, "loss": 1.2332, "step": 1016 }, { "epoch": 0.026408235185013448, "grad_norm": 5.994918260693183, "learning_rate": 9.818355706557106e-06, "loss": 1.3004, "step": 1017 }, { "epoch": 0.026434201984605398, "grad_norm": 20.47101982285911, "learning_rate": 9.819749212772719e-06, "loss": 1.2649, "step": 1018 }, { "epoch": 0.02646016878419735, "grad_norm": 7.834661925714113, "learning_rate": 9.821141350793496e-06, "loss": 1.269, "step": 1019 }, { "epoch": 0.0264861355837893, "grad_norm": 7.016448994187666, "learning_rate": 9.822532123303488e-06, "loss": 1.2671, "step": 1020 }, { "epoch": 0.02651210238338125, "grad_norm": 10.319168062018734, "learning_rate": 9.823921532978857e-06, "loss": 1.2819, "step": 1021 }, { "epoch": 0.0265380691829732, "grad_norm": 8.335203901793443, "learning_rate": 9.825309582487907e-06, "loss": 1.2271, "step": 1022 }, { "epoch": 0.02656403598256515, "grad_norm": 8.86659035709595, "learning_rate": 9.826696274491106e-06, "loss": 1.3004, "step": 1023 }, { "epoch": 0.026590002782157098, "grad_norm": 5.759697076461371, "learning_rate": 9.828081611641132e-06, "loss": 1.2976, "step": 1024 }, { "epoch": 0.026615969581749048, "grad_norm": 5.144906438008312, "learning_rate": 9.829465596582888e-06, "loss": 1.2771, "step": 1025 }, { "epoch": 0.026641936381341, "grad_norm": 8.762602966835512, "learning_rate": 9.830848231953547e-06, "loss": 1.2333, "step": 1026 }, { "epoch": 0.02666790318093295, "grad_norm": 9.294652849581757, "learning_rate": 9.832229520382572e-06, "loss": 1.2944, "step": 1027 }, { "epoch": 0.0266938699805249, "grad_norm": 9.372601418423425, "learning_rate": 9.833609464491743e-06, "loss": 1.3702, "step": 1028 }, { "epoch": 0.02671983678011685, "grad_norm": 12.490429075153717, "learning_rate": 9.834988066895202e-06, "loss": 1.3249, "step": 1029 }, { "epoch": 0.0267458035797088, "grad_norm": 7.940191338297178, "learning_rate": 9.836365330199466e-06, "loss": 1.2973, "step": 1030 }, { "epoch": 0.02677177037930075, "grad_norm": 15.315136989936061, "learning_rate": 9.837741257003469e-06, "loss": 1.3098, "step": 1031 }, { "epoch": 0.0267977371788927, "grad_norm": 6.843209859356881, "learning_rate": 9.83911584989858e-06, "loss": 1.2724, "step": 1032 }, { "epoch": 0.026823703978484652, "grad_norm": 6.574569212286721, "learning_rate": 9.840489111468645e-06, "loss": 1.3237, "step": 1033 }, { "epoch": 0.026849670778076602, "grad_norm": 6.66978318166563, "learning_rate": 9.841861044290009e-06, "loss": 1.248, "step": 1034 }, { "epoch": 0.026875637577668553, "grad_norm": 13.064317494670837, "learning_rate": 9.84323165093154e-06, "loss": 1.3519, "step": 1035 }, { "epoch": 0.026901604377260503, "grad_norm": 5.760771903132589, "learning_rate": 9.84460093395467e-06, "loss": 1.3055, "step": 1036 }, { "epoch": 0.026927571176852454, "grad_norm": 12.91904655599691, "learning_rate": 9.845968895913416e-06, "loss": 1.3158, "step": 1037 }, { "epoch": 0.026953537976444404, "grad_norm": 7.414051437771647, "learning_rate": 9.847335539354406e-06, "loss": 1.2901, "step": 1038 }, { "epoch": 0.026979504776036355, "grad_norm": 8.698376769423797, "learning_rate": 9.848700866816916e-06, "loss": 1.2545, "step": 1039 }, { "epoch": 0.027005471575628305, "grad_norm": 25.95310265840399, "learning_rate": 9.85006488083289e-06, "loss": 1.2896, "step": 1040 }, { "epoch": 0.027031438375220256, "grad_norm": 6.316976640622777, "learning_rate": 9.851427583926974e-06, "loss": 1.3023, "step": 1041 }, { "epoch": 0.027057405174812206, "grad_norm": 8.75734729194845, "learning_rate": 9.852788978616541e-06, "loss": 1.3546, "step": 1042 }, { "epoch": 0.027083371974404153, "grad_norm": 7.832613025851854, "learning_rate": 9.854149067411725e-06, "loss": 1.3085, "step": 1043 }, { "epoch": 0.027109338773996104, "grad_norm": 5.786629807901263, "learning_rate": 9.855507852815431e-06, "loss": 1.292, "step": 1044 }, { "epoch": 0.027135305573588054, "grad_norm": 6.641910204217509, "learning_rate": 9.856865337323388e-06, "loss": 1.3013, "step": 1045 }, { "epoch": 0.027161272373180004, "grad_norm": 8.89247624275867, "learning_rate": 9.858221523424158e-06, "loss": 1.2903, "step": 1046 }, { "epoch": 0.027187239172771955, "grad_norm": 7.59512655708994, "learning_rate": 9.859576413599167e-06, "loss": 1.2719, "step": 1047 }, { "epoch": 0.027213205972363905, "grad_norm": 13.562144472586645, "learning_rate": 9.860930010322737e-06, "loss": 1.2693, "step": 1048 }, { "epoch": 0.027239172771955856, "grad_norm": 8.082274324975623, "learning_rate": 9.862282316062114e-06, "loss": 1.2466, "step": 1049 }, { "epoch": 0.027265139571547806, "grad_norm": 6.916387927708562, "learning_rate": 9.863633333277485e-06, "loss": 1.317, "step": 1050 }, { "epoch": 0.027291106371139757, "grad_norm": 13.18405554746759, "learning_rate": 9.864983064422014e-06, "loss": 1.298, "step": 1051 }, { "epoch": 0.027317073170731707, "grad_norm": 11.17914017484476, "learning_rate": 9.866331511941864e-06, "loss": 1.2495, "step": 1052 }, { "epoch": 0.027343039970323658, "grad_norm": 6.638131999874204, "learning_rate": 9.86767867827623e-06, "loss": 1.2638, "step": 1053 }, { "epoch": 0.027369006769915608, "grad_norm": 5.252718993410029, "learning_rate": 9.869024565857359e-06, "loss": 1.3059, "step": 1054 }, { "epoch": 0.02739497356950756, "grad_norm": 6.303164084166287, "learning_rate": 9.870369177110576e-06, "loss": 1.3186, "step": 1055 }, { "epoch": 0.02742094036909951, "grad_norm": 8.735537538633052, "learning_rate": 9.871712514454312e-06, "loss": 1.3082, "step": 1056 }, { "epoch": 0.02744690716869146, "grad_norm": 5.4769325294594395, "learning_rate": 9.873054580300139e-06, "loss": 1.2345, "step": 1057 }, { "epoch": 0.02747287396828341, "grad_norm": 6.196905520580886, "learning_rate": 9.874395377052776e-06, "loss": 1.2992, "step": 1058 }, { "epoch": 0.02749884076787536, "grad_norm": 12.31199435827214, "learning_rate": 9.875734907110135e-06, "loss": 1.2388, "step": 1059 }, { "epoch": 0.02752480756746731, "grad_norm": 7.614061357704357, "learning_rate": 9.877073172863336e-06, "loss": 1.3431, "step": 1060 }, { "epoch": 0.02755077436705926, "grad_norm": 4.55036650638245, "learning_rate": 9.878410176696732e-06, "loss": 1.2793, "step": 1061 }, { "epoch": 0.02757674116665121, "grad_norm": 7.171198395371673, "learning_rate": 9.879745920987944e-06, "loss": 1.236, "step": 1062 }, { "epoch": 0.02760270796624316, "grad_norm": 12.267846012090597, "learning_rate": 9.88108040810787e-06, "loss": 1.2779, "step": 1063 }, { "epoch": 0.02762867476583511, "grad_norm": 6.188181521711713, "learning_rate": 9.88241364042073e-06, "loss": 1.3212, "step": 1064 }, { "epoch": 0.02765464156542706, "grad_norm": 5.5221304851121085, "learning_rate": 9.883745620284072e-06, "loss": 1.2987, "step": 1065 }, { "epoch": 0.02768060836501901, "grad_norm": 7.4426418450075, "learning_rate": 9.885076350048816e-06, "loss": 1.3036, "step": 1066 }, { "epoch": 0.02770657516461096, "grad_norm": 7.958376846061374, "learning_rate": 9.886405832059261e-06, "loss": 1.3301, "step": 1067 }, { "epoch": 0.02773254196420291, "grad_norm": 7.713984079872183, "learning_rate": 9.887734068653118e-06, "loss": 1.2916, "step": 1068 }, { "epoch": 0.027758508763794862, "grad_norm": 9.180510684500218, "learning_rate": 9.88906106216154e-06, "loss": 1.3304, "step": 1069 }, { "epoch": 0.027784475563386812, "grad_norm": 4.6742500723660125, "learning_rate": 9.890386814909139e-06, "loss": 1.2808, "step": 1070 }, { "epoch": 0.027810442362978763, "grad_norm": 6.290173924017107, "learning_rate": 9.891711329214011e-06, "loss": 1.3367, "step": 1071 }, { "epoch": 0.027836409162570713, "grad_norm": 4.803530173750026, "learning_rate": 9.89303460738776e-06, "loss": 1.3495, "step": 1072 }, { "epoch": 0.027862375962162664, "grad_norm": 4.918166152834282, "learning_rate": 9.894356651735526e-06, "loss": 1.2903, "step": 1073 }, { "epoch": 0.027888342761754614, "grad_norm": 7.036266927612349, "learning_rate": 9.895677464556009e-06, "loss": 1.2709, "step": 1074 }, { "epoch": 0.027914309561346565, "grad_norm": 8.30111993794338, "learning_rate": 9.896997048141487e-06, "loss": 1.346, "step": 1075 }, { "epoch": 0.027940276360938515, "grad_norm": 5.7953790343562055, "learning_rate": 9.89831540477785e-06, "loss": 1.2898, "step": 1076 }, { "epoch": 0.027966243160530466, "grad_norm": 8.365064506578461, "learning_rate": 9.899632536744611e-06, "loss": 1.2642, "step": 1077 }, { "epoch": 0.027992209960122416, "grad_norm": 5.868753562008038, "learning_rate": 9.900948446314938e-06, "loss": 1.2714, "step": 1078 }, { "epoch": 0.028018176759714367, "grad_norm": 30.79777673400813, "learning_rate": 9.90226313575568e-06, "loss": 1.2541, "step": 1079 }, { "epoch": 0.028044143559306317, "grad_norm": 5.131079626565473, "learning_rate": 9.903576607327382e-06, "loss": 1.3216, "step": 1080 }, { "epoch": 0.028070110358898264, "grad_norm": 7.652401077343971, "learning_rate": 9.904888863284312e-06, "loss": 1.356, "step": 1081 }, { "epoch": 0.028096077158490215, "grad_norm": 5.2639260650010895, "learning_rate": 9.906199905874489e-06, "loss": 1.3191, "step": 1082 }, { "epoch": 0.028122043958082165, "grad_norm": 10.588405645159488, "learning_rate": 9.907509737339701e-06, "loss": 1.2601, "step": 1083 }, { "epoch": 0.028148010757674116, "grad_norm": 6.939560867619823, "learning_rate": 9.908818359915526e-06, "loss": 1.2829, "step": 1084 }, { "epoch": 0.028173977557266066, "grad_norm": 8.117869373278994, "learning_rate": 9.910125775831356e-06, "loss": 1.2739, "step": 1085 }, { "epoch": 0.028199944356858016, "grad_norm": 8.365096884449724, "learning_rate": 9.911431987310429e-06, "loss": 1.2724, "step": 1086 }, { "epoch": 0.028225911156449967, "grad_norm": 16.42663354284076, "learning_rate": 9.912736996569839e-06, "loss": 1.301, "step": 1087 }, { "epoch": 0.028251877956041917, "grad_norm": 11.20570367267512, "learning_rate": 9.914040805820566e-06, "loss": 1.3151, "step": 1088 }, { "epoch": 0.028277844755633868, "grad_norm": 6.506893024311525, "learning_rate": 9.915343417267494e-06, "loss": 1.2834, "step": 1089 }, { "epoch": 0.02830381155522582, "grad_norm": 5.713663476035565, "learning_rate": 9.91664483310944e-06, "loss": 1.3185, "step": 1090 }, { "epoch": 0.02832977835481777, "grad_norm": 8.63059705381795, "learning_rate": 9.917945055539166e-06, "loss": 1.2822, "step": 1091 }, { "epoch": 0.02835574515440972, "grad_norm": 11.057373928219873, "learning_rate": 9.919244086743413e-06, "loss": 1.3149, "step": 1092 }, { "epoch": 0.02838171195400167, "grad_norm": 5.1049395332902625, "learning_rate": 9.920541928902911e-06, "loss": 1.3154, "step": 1093 }, { "epoch": 0.02840767875359362, "grad_norm": 14.220859647753146, "learning_rate": 9.921838584192414e-06, "loss": 1.2625, "step": 1094 }, { "epoch": 0.02843364555318557, "grad_norm": 6.253690316774968, "learning_rate": 9.923134054780712e-06, "loss": 1.2559, "step": 1095 }, { "epoch": 0.02845961235277752, "grad_norm": 5.732942906474517, "learning_rate": 9.924428342830654e-06, "loss": 1.2729, "step": 1096 }, { "epoch": 0.02848557915236947, "grad_norm": 12.446745689998023, "learning_rate": 9.925721450499174e-06, "loss": 1.3189, "step": 1097 }, { "epoch": 0.028511545951961422, "grad_norm": 11.539834070979625, "learning_rate": 9.927013379937308e-06, "loss": 1.3065, "step": 1098 }, { "epoch": 0.028537512751553373, "grad_norm": 6.907638203152674, "learning_rate": 9.928304133290221e-06, "loss": 1.2648, "step": 1099 }, { "epoch": 0.02856347955114532, "grad_norm": 6.9482031386156775, "learning_rate": 9.929593712697221e-06, "loss": 1.3044, "step": 1100 }, { "epoch": 0.02858944635073727, "grad_norm": 7.367737378405153, "learning_rate": 9.930882120291786e-06, "loss": 1.2944, "step": 1101 }, { "epoch": 0.02861541315032922, "grad_norm": 8.556297115979206, "learning_rate": 9.932169358201583e-06, "loss": 1.3347, "step": 1102 }, { "epoch": 0.02864137994992117, "grad_norm": 13.650762716929272, "learning_rate": 9.933455428548492e-06, "loss": 1.297, "step": 1103 }, { "epoch": 0.02866734674951312, "grad_norm": 8.254009312980543, "learning_rate": 9.934740333448617e-06, "loss": 1.2905, "step": 1104 }, { "epoch": 0.028693313549105072, "grad_norm": 7.556745625562825, "learning_rate": 9.936024075012323e-06, "loss": 1.3579, "step": 1105 }, { "epoch": 0.028719280348697022, "grad_norm": 5.325919223300342, "learning_rate": 9.937306655344245e-06, "loss": 1.2271, "step": 1106 }, { "epoch": 0.028745247148288973, "grad_norm": 5.61053471809934, "learning_rate": 9.938588076543308e-06, "loss": 1.2147, "step": 1107 }, { "epoch": 0.028771213947880923, "grad_norm": 7.589505740270034, "learning_rate": 9.939868340702757e-06, "loss": 1.2103, "step": 1108 }, { "epoch": 0.028797180747472874, "grad_norm": 6.291160243118104, "learning_rate": 9.941147449910168e-06, "loss": 1.2973, "step": 1109 }, { "epoch": 0.028823147547064824, "grad_norm": 6.017742626463795, "learning_rate": 9.942425406247476e-06, "loss": 1.2366, "step": 1110 }, { "epoch": 0.028849114346656775, "grad_norm": 7.604563356965189, "learning_rate": 9.943702211790988e-06, "loss": 1.3273, "step": 1111 }, { "epoch": 0.028875081146248725, "grad_norm": 5.610686337515092, "learning_rate": 9.944977868611409e-06, "loss": 1.3184, "step": 1112 }, { "epoch": 0.028901047945840676, "grad_norm": 6.508801590033893, "learning_rate": 9.94625237877386e-06, "loss": 1.2517, "step": 1113 }, { "epoch": 0.028927014745432626, "grad_norm": 9.339764513727046, "learning_rate": 9.947525744337894e-06, "loss": 1.3482, "step": 1114 }, { "epoch": 0.028952981545024577, "grad_norm": 5.796246188625047, "learning_rate": 9.94879796735753e-06, "loss": 1.2674, "step": 1115 }, { "epoch": 0.028978948344616527, "grad_norm": 6.687083578959679, "learning_rate": 9.950069049881252e-06, "loss": 1.2835, "step": 1116 }, { "epoch": 0.029004915144208478, "grad_norm": 8.015183345532403, "learning_rate": 9.951338993952048e-06, "loss": 1.264, "step": 1117 }, { "epoch": 0.029030881943800428, "grad_norm": 8.818356563191688, "learning_rate": 9.952607801607415e-06, "loss": 1.2409, "step": 1118 }, { "epoch": 0.02905684874339238, "grad_norm": 8.316739256191019, "learning_rate": 9.953875474879389e-06, "loss": 1.2932, "step": 1119 }, { "epoch": 0.029082815542984326, "grad_norm": 5.2969860864204135, "learning_rate": 9.955142015794563e-06, "loss": 1.3126, "step": 1120 }, { "epoch": 0.029108782342576276, "grad_norm": 8.004558694887704, "learning_rate": 9.956407426374096e-06, "loss": 1.3043, "step": 1121 }, { "epoch": 0.029134749142168227, "grad_norm": 5.215054054162986, "learning_rate": 9.957671708633747e-06, "loss": 1.3041, "step": 1122 }, { "epoch": 0.029160715941760177, "grad_norm": 4.984566137780617, "learning_rate": 9.958934864583887e-06, "loss": 1.2883, "step": 1123 }, { "epoch": 0.029186682741352128, "grad_norm": 5.8515963992676285, "learning_rate": 9.960196896229518e-06, "loss": 1.2986, "step": 1124 }, { "epoch": 0.029212649540944078, "grad_norm": 4.76953464917211, "learning_rate": 9.96145780557029e-06, "loss": 1.307, "step": 1125 }, { "epoch": 0.02923861634053603, "grad_norm": 10.915810665514108, "learning_rate": 9.962717594600529e-06, "loss": 1.2524, "step": 1126 }, { "epoch": 0.02926458314012798, "grad_norm": 8.335292687543234, "learning_rate": 9.963976265309245e-06, "loss": 1.318, "step": 1127 }, { "epoch": 0.02929054993971993, "grad_norm": 7.8148951406637766, "learning_rate": 9.965233819680155e-06, "loss": 1.2543, "step": 1128 }, { "epoch": 0.02931651673931188, "grad_norm": 62.18826357885485, "learning_rate": 9.966490259691707e-06, "loss": 1.2998, "step": 1129 }, { "epoch": 0.02934248353890383, "grad_norm": 5.6153817915432, "learning_rate": 9.967745587317093e-06, "loss": 1.2744, "step": 1130 }, { "epoch": 0.02936845033849578, "grad_norm": 5.683617912319774, "learning_rate": 9.968999804524268e-06, "loss": 1.3293, "step": 1131 }, { "epoch": 0.02939441713808773, "grad_norm": 5.148275416229453, "learning_rate": 9.970252913275966e-06, "loss": 1.3069, "step": 1132 }, { "epoch": 0.029420383937679682, "grad_norm": 5.18055112609583, "learning_rate": 9.971504915529725e-06, "loss": 1.2646, "step": 1133 }, { "epoch": 0.029446350737271632, "grad_norm": 5.603866039659951, "learning_rate": 9.972755813237904e-06, "loss": 1.2675, "step": 1134 }, { "epoch": 0.029472317536863583, "grad_norm": 13.454907306988348, "learning_rate": 9.974005608347697e-06, "loss": 1.2712, "step": 1135 }, { "epoch": 0.029498284336455533, "grad_norm": 7.329232658825036, "learning_rate": 9.975254302801151e-06, "loss": 1.3403, "step": 1136 }, { "epoch": 0.029524251136047484, "grad_norm": 4.739095720073457, "learning_rate": 9.976501898535191e-06, "loss": 1.2899, "step": 1137 }, { "epoch": 0.029550217935639434, "grad_norm": 14.136039808233456, "learning_rate": 9.977748397481631e-06, "loss": 1.2836, "step": 1138 }, { "epoch": 0.02957618473523138, "grad_norm": 5.581011369992656, "learning_rate": 9.978993801567193e-06, "loss": 1.2641, "step": 1139 }, { "epoch": 0.02960215153482333, "grad_norm": 4.833082696839791, "learning_rate": 9.980238112713535e-06, "loss": 1.2963, "step": 1140 }, { "epoch": 0.029628118334415282, "grad_norm": 5.94084250078561, "learning_rate": 9.981481332837249e-06, "loss": 1.2688, "step": 1141 }, { "epoch": 0.029654085134007233, "grad_norm": 6.349469092094617, "learning_rate": 9.982723463849895e-06, "loss": 1.3008, "step": 1142 }, { "epoch": 0.029680051933599183, "grad_norm": 7.966106380195767, "learning_rate": 9.983964507658016e-06, "loss": 1.2562, "step": 1143 }, { "epoch": 0.029706018733191134, "grad_norm": 19.006271130475756, "learning_rate": 9.98520446616315e-06, "loss": 1.3013, "step": 1144 }, { "epoch": 0.029731985532783084, "grad_norm": 5.407256633557118, "learning_rate": 9.986443341261849e-06, "loss": 1.3047, "step": 1145 }, { "epoch": 0.029757952332375034, "grad_norm": 6.161438590174236, "learning_rate": 9.987681134845703e-06, "loss": 1.2815, "step": 1146 }, { "epoch": 0.029783919131966985, "grad_norm": 11.131652900134036, "learning_rate": 9.988917848801347e-06, "loss": 1.3504, "step": 1147 }, { "epoch": 0.029809885931558935, "grad_norm": 8.078324276736927, "learning_rate": 9.990153485010489e-06, "loss": 1.2852, "step": 1148 }, { "epoch": 0.029835852731150886, "grad_norm": 5.826540488674399, "learning_rate": 9.991388045349918e-06, "loss": 1.2674, "step": 1149 }, { "epoch": 0.029861819530742836, "grad_norm": 6.355159627738942, "learning_rate": 9.992621531691526e-06, "loss": 1.3169, "step": 1150 }, { "epoch": 0.029887786330334787, "grad_norm": 6.400457973783588, "learning_rate": 9.993853945902325e-06, "loss": 1.2569, "step": 1151 }, { "epoch": 0.029913753129926737, "grad_norm": 9.598019522785188, "learning_rate": 9.99508528984446e-06, "loss": 1.2778, "step": 1152 }, { "epoch": 0.029939719929518688, "grad_norm": 5.07157141716493, "learning_rate": 9.996315565375231e-06, "loss": 1.2805, "step": 1153 }, { "epoch": 0.029965686729110638, "grad_norm": 8.507910693734308, "learning_rate": 9.997544774347108e-06, "loss": 1.2873, "step": 1154 }, { "epoch": 0.02999165352870259, "grad_norm": 4.759716936879111, "learning_rate": 9.998772918607745e-06, "loss": 1.2622, "step": 1155 }, { "epoch": 0.03001762032829454, "grad_norm": 15.711628993981854, "learning_rate": 1e-05, "loss": 1.2975, "step": 1156 }, { "epoch": 0.03004358712788649, "grad_norm": 5.704958769827323, "learning_rate": 1e-05, "loss": 1.2686, "step": 1157 }, { "epoch": 0.030069553927478437, "grad_norm": 11.405167703361354, "learning_rate": 1e-05, "loss": 1.2969, "step": 1158 }, { "epoch": 0.030095520727070387, "grad_norm": 5.8038110905721965, "learning_rate": 1e-05, "loss": 1.3172, "step": 1159 }, { "epoch": 0.030121487526662338, "grad_norm": 5.209212307829466, "learning_rate": 1e-05, "loss": 1.2935, "step": 1160 }, { "epoch": 0.030147454326254288, "grad_norm": 5.014955089668461, "learning_rate": 1e-05, "loss": 1.338, "step": 1161 }, { "epoch": 0.03017342112584624, "grad_norm": 8.799306963593825, "learning_rate": 1e-05, "loss": 1.3115, "step": 1162 }, { "epoch": 0.03019938792543819, "grad_norm": 4.537021092554795, "learning_rate": 1e-05, "loss": 1.2824, "step": 1163 }, { "epoch": 0.03022535472503014, "grad_norm": 6.4757861510718575, "learning_rate": 1e-05, "loss": 1.3057, "step": 1164 }, { "epoch": 0.03025132152462209, "grad_norm": 5.831442208510787, "learning_rate": 1e-05, "loss": 1.282, "step": 1165 }, { "epoch": 0.03027728832421404, "grad_norm": 8.033116461436856, "learning_rate": 1e-05, "loss": 1.303, "step": 1166 }, { "epoch": 0.03030325512380599, "grad_norm": 7.5971513830489705, "learning_rate": 1e-05, "loss": 1.2837, "step": 1167 }, { "epoch": 0.03032922192339794, "grad_norm": 7.725414287235593, "learning_rate": 1e-05, "loss": 1.26, "step": 1168 }, { "epoch": 0.030355188722989892, "grad_norm": 17.4647926392956, "learning_rate": 1e-05, "loss": 1.2975, "step": 1169 }, { "epoch": 0.030381155522581842, "grad_norm": 10.514839947943068, "learning_rate": 1e-05, "loss": 1.3081, "step": 1170 }, { "epoch": 0.030407122322173793, "grad_norm": 4.668977278931304, "learning_rate": 1e-05, "loss": 1.2813, "step": 1171 }, { "epoch": 0.030433089121765743, "grad_norm": 9.247054997297619, "learning_rate": 1e-05, "loss": 1.3499, "step": 1172 }, { "epoch": 0.030459055921357694, "grad_norm": 7.986505571289259, "learning_rate": 1e-05, "loss": 1.2754, "step": 1173 }, { "epoch": 0.030485022720949644, "grad_norm": 6.241017409791946, "learning_rate": 1e-05, "loss": 1.2616, "step": 1174 }, { "epoch": 0.030510989520541595, "grad_norm": 6.649446404151962, "learning_rate": 1e-05, "loss": 1.296, "step": 1175 }, { "epoch": 0.030536956320133545, "grad_norm": 8.567301719470455, "learning_rate": 1e-05, "loss": 1.301, "step": 1176 }, { "epoch": 0.030562923119725492, "grad_norm": 9.475958670478116, "learning_rate": 1e-05, "loss": 1.2707, "step": 1177 }, { "epoch": 0.030588889919317443, "grad_norm": 11.780495535979894, "learning_rate": 1e-05, "loss": 1.259, "step": 1178 }, { "epoch": 0.030614856718909393, "grad_norm": 7.469332748088021, "learning_rate": 1e-05, "loss": 1.2914, "step": 1179 }, { "epoch": 0.030640823518501344, "grad_norm": 10.573077982656883, "learning_rate": 1e-05, "loss": 1.3129, "step": 1180 }, { "epoch": 0.030666790318093294, "grad_norm": 8.475428801634878, "learning_rate": 1e-05, "loss": 1.3162, "step": 1181 }, { "epoch": 0.030692757117685245, "grad_norm": 5.347192202176903, "learning_rate": 1e-05, "loss": 1.2702, "step": 1182 }, { "epoch": 0.030718723917277195, "grad_norm": 5.672581147932818, "learning_rate": 1e-05, "loss": 1.3463, "step": 1183 }, { "epoch": 0.030744690716869146, "grad_norm": 6.454212605079143, "learning_rate": 1e-05, "loss": 1.3169, "step": 1184 }, { "epoch": 0.030770657516461096, "grad_norm": 12.429250463718388, "learning_rate": 1e-05, "loss": 1.2956, "step": 1185 }, { "epoch": 0.030796624316053046, "grad_norm": 7.954239383110013, "learning_rate": 1e-05, "loss": 1.2694, "step": 1186 }, { "epoch": 0.030822591115644997, "grad_norm": 7.826045873782921, "learning_rate": 1e-05, "loss": 1.2545, "step": 1187 }, { "epoch": 0.030848557915236947, "grad_norm": 5.073646988933162, "learning_rate": 1e-05, "loss": 1.2706, "step": 1188 }, { "epoch": 0.030874524714828898, "grad_norm": 9.106322102467423, "learning_rate": 1e-05, "loss": 1.3607, "step": 1189 }, { "epoch": 0.03090049151442085, "grad_norm": 13.130781371784154, "learning_rate": 1e-05, "loss": 1.3482, "step": 1190 }, { "epoch": 0.0309264583140128, "grad_norm": 11.873580687179455, "learning_rate": 1e-05, "loss": 1.2454, "step": 1191 }, { "epoch": 0.03095242511360475, "grad_norm": 5.982146403427537, "learning_rate": 1e-05, "loss": 1.3366, "step": 1192 }, { "epoch": 0.0309783919131967, "grad_norm": 8.986530397025325, "learning_rate": 1e-05, "loss": 1.2897, "step": 1193 }, { "epoch": 0.03100435871278865, "grad_norm": 8.614994160743723, "learning_rate": 1e-05, "loss": 1.2946, "step": 1194 }, { "epoch": 0.0310303255123806, "grad_norm": 6.8083867245669705, "learning_rate": 1e-05, "loss": 1.2818, "step": 1195 }, { "epoch": 0.031056292311972548, "grad_norm": 7.190492753292799, "learning_rate": 1e-05, "loss": 1.283, "step": 1196 }, { "epoch": 0.031082259111564498, "grad_norm": 7.464468395112331, "learning_rate": 1e-05, "loss": 1.2915, "step": 1197 }, { "epoch": 0.03110822591115645, "grad_norm": 6.7693475433094275, "learning_rate": 1e-05, "loss": 1.2938, "step": 1198 }, { "epoch": 0.0311341927107484, "grad_norm": 10.502012150836077, "learning_rate": 1e-05, "loss": 1.2797, "step": 1199 }, { "epoch": 0.03116015951034035, "grad_norm": 43.00080480487763, "learning_rate": 1e-05, "loss": 1.2769, "step": 1200 }, { "epoch": 0.0311861263099323, "grad_norm": 7.406563410183718, "learning_rate": 1e-05, "loss": 1.3208, "step": 1201 }, { "epoch": 0.03121209310952425, "grad_norm": 9.641912488735297, "learning_rate": 1e-05, "loss": 1.2822, "step": 1202 }, { "epoch": 0.0312380599091162, "grad_norm": 9.141513879056168, "learning_rate": 1e-05, "loss": 1.2793, "step": 1203 }, { "epoch": 0.031264026708708155, "grad_norm": 6.081075310231813, "learning_rate": 1e-05, "loss": 1.3334, "step": 1204 }, { "epoch": 0.031289993508300105, "grad_norm": 9.175164277895366, "learning_rate": 1e-05, "loss": 1.2577, "step": 1205 }, { "epoch": 0.03131596030789205, "grad_norm": 8.852813928527189, "learning_rate": 1e-05, "loss": 1.2783, "step": 1206 }, { "epoch": 0.031341927107484, "grad_norm": 8.815681478433714, "learning_rate": 1e-05, "loss": 1.3009, "step": 1207 }, { "epoch": 0.03136789390707595, "grad_norm": 13.404310392906268, "learning_rate": 1e-05, "loss": 1.294, "step": 1208 }, { "epoch": 0.0313938607066679, "grad_norm": 9.442077739102237, "learning_rate": 1e-05, "loss": 1.3175, "step": 1209 }, { "epoch": 0.03141982750625985, "grad_norm": 7.9859395444565475, "learning_rate": 1e-05, "loss": 1.2515, "step": 1210 }, { "epoch": 0.0314457943058518, "grad_norm": 8.829228533074552, "learning_rate": 1e-05, "loss": 1.347, "step": 1211 }, { "epoch": 0.03147176110544375, "grad_norm": 5.398500389693376, "learning_rate": 1e-05, "loss": 1.3454, "step": 1212 }, { "epoch": 0.0314977279050357, "grad_norm": 16.644715973071417, "learning_rate": 1e-05, "loss": 1.2729, "step": 1213 }, { "epoch": 0.03152369470462765, "grad_norm": 7.646571180062837, "learning_rate": 1e-05, "loss": 1.2489, "step": 1214 }, { "epoch": 0.0315496615042196, "grad_norm": 5.815216414146542, "learning_rate": 1e-05, "loss": 1.2655, "step": 1215 }, { "epoch": 0.031575628303811554, "grad_norm": 4.933497871198202, "learning_rate": 1e-05, "loss": 1.2444, "step": 1216 }, { "epoch": 0.031601595103403504, "grad_norm": 5.812606974612045, "learning_rate": 1e-05, "loss": 1.2935, "step": 1217 }, { "epoch": 0.031627561902995455, "grad_norm": 5.931935162056789, "learning_rate": 1e-05, "loss": 1.2773, "step": 1218 }, { "epoch": 0.031653528702587405, "grad_norm": 10.779004026133869, "learning_rate": 1e-05, "loss": 1.2677, "step": 1219 }, { "epoch": 0.031679495502179356, "grad_norm": 8.341190245517366, "learning_rate": 1e-05, "loss": 1.3194, "step": 1220 }, { "epoch": 0.031705462301771306, "grad_norm": 7.731110964907242, "learning_rate": 1e-05, "loss": 1.285, "step": 1221 }, { "epoch": 0.03173142910136326, "grad_norm": 7.713904956725646, "learning_rate": 1e-05, "loss": 1.2747, "step": 1222 }, { "epoch": 0.03175739590095521, "grad_norm": 8.559153550268292, "learning_rate": 1e-05, "loss": 1.2487, "step": 1223 }, { "epoch": 0.03178336270054716, "grad_norm": 12.238991910023996, "learning_rate": 1e-05, "loss": 1.34, "step": 1224 }, { "epoch": 0.03180932950013911, "grad_norm": 7.613851432920012, "learning_rate": 1e-05, "loss": 1.2953, "step": 1225 }, { "epoch": 0.03183529629973106, "grad_norm": 7.6095532155561445, "learning_rate": 1e-05, "loss": 1.3063, "step": 1226 }, { "epoch": 0.03186126309932301, "grad_norm": 9.693610251702616, "learning_rate": 1e-05, "loss": 1.2799, "step": 1227 }, { "epoch": 0.03188722989891496, "grad_norm": 8.430606993410603, "learning_rate": 1e-05, "loss": 1.2884, "step": 1228 }, { "epoch": 0.03191319669850691, "grad_norm": 19.016820589468452, "learning_rate": 1e-05, "loss": 1.2867, "step": 1229 }, { "epoch": 0.03193916349809886, "grad_norm": 8.350690034233791, "learning_rate": 1e-05, "loss": 1.3015, "step": 1230 }, { "epoch": 0.03196513029769081, "grad_norm": 4.953980958547024, "learning_rate": 1e-05, "loss": 1.2724, "step": 1231 }, { "epoch": 0.03199109709728276, "grad_norm": 6.962629838208837, "learning_rate": 1e-05, "loss": 1.2381, "step": 1232 }, { "epoch": 0.03201706389687471, "grad_norm": 8.829409561548632, "learning_rate": 1e-05, "loss": 1.3128, "step": 1233 }, { "epoch": 0.03204303069646666, "grad_norm": 11.243175555955322, "learning_rate": 1e-05, "loss": 1.2739, "step": 1234 }, { "epoch": 0.03206899749605861, "grad_norm": 7.7195048715993755, "learning_rate": 1e-05, "loss": 1.311, "step": 1235 }, { "epoch": 0.03209496429565056, "grad_norm": 5.294238489305202, "learning_rate": 1e-05, "loss": 1.2405, "step": 1236 }, { "epoch": 0.032120931095242514, "grad_norm": 7.307060370361224, "learning_rate": 1e-05, "loss": 1.2904, "step": 1237 }, { "epoch": 0.032146897894834464, "grad_norm": 6.055877214262894, "learning_rate": 1e-05, "loss": 1.2839, "step": 1238 }, { "epoch": 0.032172864694426415, "grad_norm": 7.8844218733124105, "learning_rate": 1e-05, "loss": 1.3695, "step": 1239 }, { "epoch": 0.032198831494018365, "grad_norm": 8.69614205597602, "learning_rate": 1e-05, "loss": 1.2393, "step": 1240 }, { "epoch": 0.032224798293610316, "grad_norm": 11.98923963666711, "learning_rate": 1e-05, "loss": 1.2746, "step": 1241 }, { "epoch": 0.032250765093202266, "grad_norm": 5.17861177024725, "learning_rate": 1e-05, "loss": 1.2874, "step": 1242 }, { "epoch": 0.032276731892794217, "grad_norm": 6.099947081789708, "learning_rate": 1e-05, "loss": 1.2706, "step": 1243 }, { "epoch": 0.03230269869238616, "grad_norm": 12.756239019049964, "learning_rate": 1e-05, "loss": 1.2524, "step": 1244 }, { "epoch": 0.03232866549197811, "grad_norm": 5.6104349394098305, "learning_rate": 1e-05, "loss": 1.2634, "step": 1245 }, { "epoch": 0.03235463229157006, "grad_norm": 4.573522552231114, "learning_rate": 1e-05, "loss": 1.3003, "step": 1246 }, { "epoch": 0.03238059909116201, "grad_norm": 5.285020373980933, "learning_rate": 1e-05, "loss": 1.2939, "step": 1247 }, { "epoch": 0.03240656589075396, "grad_norm": 6.6978491448004345, "learning_rate": 1e-05, "loss": 1.2356, "step": 1248 }, { "epoch": 0.03243253269034591, "grad_norm": 9.586709884454176, "learning_rate": 1e-05, "loss": 1.2865, "step": 1249 }, { "epoch": 0.03245849948993786, "grad_norm": 5.087433250795417, "learning_rate": 1e-05, "loss": 1.3387, "step": 1250 }, { "epoch": 0.03248446628952981, "grad_norm": 12.804145880666873, "learning_rate": 1e-05, "loss": 1.2915, "step": 1251 }, { "epoch": 0.032510433089121764, "grad_norm": 5.876258877550928, "learning_rate": 1e-05, "loss": 1.3534, "step": 1252 }, { "epoch": 0.032536399888713714, "grad_norm": 6.525342103605894, "learning_rate": 1e-05, "loss": 1.3328, "step": 1253 }, { "epoch": 0.032562366688305665, "grad_norm": 6.913313485647102, "learning_rate": 1e-05, "loss": 1.2654, "step": 1254 }, { "epoch": 0.032588333487897615, "grad_norm": 10.85038760850887, "learning_rate": 1e-05, "loss": 1.2485, "step": 1255 }, { "epoch": 0.032614300287489566, "grad_norm": 34.537960674293764, "learning_rate": 1e-05, "loss": 1.2917, "step": 1256 }, { "epoch": 0.032640267087081516, "grad_norm": 7.020000495149861, "learning_rate": 1e-05, "loss": 1.2845, "step": 1257 }, { "epoch": 0.03266623388667347, "grad_norm": 6.901514147052997, "learning_rate": 1e-05, "loss": 1.2777, "step": 1258 }, { "epoch": 0.03269220068626542, "grad_norm": 7.074747184846409, "learning_rate": 1e-05, "loss": 1.2814, "step": 1259 }, { "epoch": 0.03271816748585737, "grad_norm": 5.774886964955056, "learning_rate": 1e-05, "loss": 1.303, "step": 1260 }, { "epoch": 0.03274413428544932, "grad_norm": 5.152112005544501, "learning_rate": 1e-05, "loss": 1.2726, "step": 1261 }, { "epoch": 0.03277010108504127, "grad_norm": 10.593549788392021, "learning_rate": 1e-05, "loss": 1.2408, "step": 1262 }, { "epoch": 0.03279606788463322, "grad_norm": 5.882130293278905, "learning_rate": 1e-05, "loss": 1.2529, "step": 1263 }, { "epoch": 0.03282203468422517, "grad_norm": 6.598534172735621, "learning_rate": 1e-05, "loss": 1.2716, "step": 1264 }, { "epoch": 0.03284800148381712, "grad_norm": 9.357861531820744, "learning_rate": 1e-05, "loss": 1.325, "step": 1265 }, { "epoch": 0.03287396828340907, "grad_norm": 5.509338773455769, "learning_rate": 1e-05, "loss": 1.275, "step": 1266 }, { "epoch": 0.03289993508300102, "grad_norm": 6.328281428911475, "learning_rate": 1e-05, "loss": 1.2722, "step": 1267 }, { "epoch": 0.03292590188259297, "grad_norm": 9.011185794410562, "learning_rate": 1e-05, "loss": 1.2919, "step": 1268 }, { "epoch": 0.03295186868218492, "grad_norm": 5.396542516029994, "learning_rate": 1e-05, "loss": 1.2356, "step": 1269 }, { "epoch": 0.03297783548177687, "grad_norm": 8.003985365946221, "learning_rate": 1e-05, "loss": 1.2617, "step": 1270 }, { "epoch": 0.03300380228136882, "grad_norm": 7.115819537854547, "learning_rate": 1e-05, "loss": 1.3067, "step": 1271 }, { "epoch": 0.03302976908096077, "grad_norm": 46.00386362850407, "learning_rate": 1e-05, "loss": 1.2708, "step": 1272 }, { "epoch": 0.033055735880552724, "grad_norm": 7.253735073454857, "learning_rate": 1e-05, "loss": 1.3231, "step": 1273 }, { "epoch": 0.033081702680144674, "grad_norm": 6.602340814880723, "learning_rate": 1e-05, "loss": 1.2554, "step": 1274 }, { "epoch": 0.033107669479736625, "grad_norm": 36.37727689662961, "learning_rate": 1e-05, "loss": 1.3138, "step": 1275 }, { "epoch": 0.033133636279328575, "grad_norm": 9.262308128521518, "learning_rate": 1e-05, "loss": 1.2304, "step": 1276 }, { "epoch": 0.033159603078920526, "grad_norm": 19.235772758412416, "learning_rate": 1e-05, "loss": 1.2874, "step": 1277 }, { "epoch": 0.033185569878512476, "grad_norm": 6.468657119747601, "learning_rate": 1e-05, "loss": 1.2965, "step": 1278 }, { "epoch": 0.03321153667810443, "grad_norm": 6.397727097792017, "learning_rate": 1e-05, "loss": 1.2928, "step": 1279 }, { "epoch": 0.03323750347769638, "grad_norm": 5.877681363778691, "learning_rate": 1e-05, "loss": 1.3142, "step": 1280 }, { "epoch": 0.03326347027728833, "grad_norm": 19.19895925880082, "learning_rate": 1e-05, "loss": 1.295, "step": 1281 }, { "epoch": 0.03328943707688027, "grad_norm": 15.210322559154214, "learning_rate": 1e-05, "loss": 1.2607, "step": 1282 }, { "epoch": 0.03331540387647222, "grad_norm": 7.170934278836788, "learning_rate": 1e-05, "loss": 1.3358, "step": 1283 }, { "epoch": 0.03334137067606417, "grad_norm": 6.645033290943255, "learning_rate": 1e-05, "loss": 1.262, "step": 1284 }, { "epoch": 0.03336733747565612, "grad_norm": 10.013773016421636, "learning_rate": 1e-05, "loss": 1.2818, "step": 1285 }, { "epoch": 0.03339330427524807, "grad_norm": 6.534796668977759, "learning_rate": 1e-05, "loss": 1.3251, "step": 1286 }, { "epoch": 0.03341927107484002, "grad_norm": 6.92043149892748, "learning_rate": 1e-05, "loss": 1.3102, "step": 1287 }, { "epoch": 0.033445237874431974, "grad_norm": 27.43758786823914, "learning_rate": 1e-05, "loss": 1.2839, "step": 1288 }, { "epoch": 0.033471204674023924, "grad_norm": 7.785404674128468, "learning_rate": 1e-05, "loss": 1.2603, "step": 1289 }, { "epoch": 0.033497171473615875, "grad_norm": 10.682396813663603, "learning_rate": 1e-05, "loss": 1.3102, "step": 1290 }, { "epoch": 0.033523138273207825, "grad_norm": 8.896157418868501, "learning_rate": 1e-05, "loss": 1.2859, "step": 1291 }, { "epoch": 0.033549105072799776, "grad_norm": 25.194561519815764, "learning_rate": 1e-05, "loss": 1.2386, "step": 1292 }, { "epoch": 0.033575071872391726, "grad_norm": 7.877894520518539, "learning_rate": 1e-05, "loss": 1.278, "step": 1293 }, { "epoch": 0.03360103867198368, "grad_norm": 8.789514745309198, "learning_rate": 1e-05, "loss": 1.272, "step": 1294 }, { "epoch": 0.03362700547157563, "grad_norm": 7.074288311536793, "learning_rate": 1e-05, "loss": 1.2737, "step": 1295 }, { "epoch": 0.03365297227116758, "grad_norm": 7.906679164411286, "learning_rate": 1e-05, "loss": 1.3085, "step": 1296 }, { "epoch": 0.03367893907075953, "grad_norm": 8.254794952600134, "learning_rate": 1e-05, "loss": 1.2795, "step": 1297 }, { "epoch": 0.03370490587035148, "grad_norm": 6.320909092381442, "learning_rate": 1e-05, "loss": 1.3044, "step": 1298 }, { "epoch": 0.03373087266994343, "grad_norm": 6.757360895314919, "learning_rate": 1e-05, "loss": 1.2782, "step": 1299 }, { "epoch": 0.03375683946953538, "grad_norm": 13.089433418292561, "learning_rate": 1e-05, "loss": 1.2691, "step": 1300 }, { "epoch": 0.03378280626912733, "grad_norm": 6.344175728479478, "learning_rate": 1e-05, "loss": 1.2876, "step": 1301 }, { "epoch": 0.03380877306871928, "grad_norm": 6.69918145261119, "learning_rate": 1e-05, "loss": 1.3131, "step": 1302 }, { "epoch": 0.03383473986831123, "grad_norm": 8.060469497372123, "learning_rate": 1e-05, "loss": 1.2954, "step": 1303 }, { "epoch": 0.03386070666790318, "grad_norm": 6.988018135418475, "learning_rate": 1e-05, "loss": 1.2601, "step": 1304 }, { "epoch": 0.03388667346749513, "grad_norm": 9.599174432224546, "learning_rate": 1e-05, "loss": 1.3192, "step": 1305 }, { "epoch": 0.03391264026708708, "grad_norm": 7.41644281056642, "learning_rate": 1e-05, "loss": 1.2779, "step": 1306 }, { "epoch": 0.03393860706667903, "grad_norm": 6.4957437518461605, "learning_rate": 1e-05, "loss": 1.2446, "step": 1307 }, { "epoch": 0.03396457386627098, "grad_norm": 31.36272637174182, "learning_rate": 1e-05, "loss": 1.3113, "step": 1308 }, { "epoch": 0.033990540665862934, "grad_norm": 6.295854433715658, "learning_rate": 1e-05, "loss": 1.2611, "step": 1309 }, { "epoch": 0.034016507465454884, "grad_norm": 7.062798417800024, "learning_rate": 1e-05, "loss": 1.296, "step": 1310 }, { "epoch": 0.034042474265046835, "grad_norm": 8.709148411426119, "learning_rate": 1e-05, "loss": 1.2964, "step": 1311 }, { "epoch": 0.034068441064638785, "grad_norm": 9.029187240776652, "learning_rate": 1e-05, "loss": 1.3176, "step": 1312 }, { "epoch": 0.034094407864230736, "grad_norm": 6.8283831026018476, "learning_rate": 1e-05, "loss": 1.2983, "step": 1313 }, { "epoch": 0.034120374663822686, "grad_norm": 6.415306574690381, "learning_rate": 1e-05, "loss": 1.2572, "step": 1314 }, { "epoch": 0.03414634146341464, "grad_norm": 10.424017873191003, "learning_rate": 1e-05, "loss": 1.2812, "step": 1315 }, { "epoch": 0.03417230826300659, "grad_norm": 8.390611815486615, "learning_rate": 1e-05, "loss": 1.2989, "step": 1316 }, { "epoch": 0.03419827506259854, "grad_norm": 5.811672356860467, "learning_rate": 1e-05, "loss": 1.2801, "step": 1317 }, { "epoch": 0.03422424186219049, "grad_norm": 19.12354307920851, "learning_rate": 1e-05, "loss": 1.2185, "step": 1318 }, { "epoch": 0.03425020866178244, "grad_norm": 7.2307876689694535, "learning_rate": 1e-05, "loss": 1.2727, "step": 1319 }, { "epoch": 0.03427617546137438, "grad_norm": 10.61963257694615, "learning_rate": 1e-05, "loss": 1.2888, "step": 1320 }, { "epoch": 0.03430214226096633, "grad_norm": 44.449469862910064, "learning_rate": 1e-05, "loss": 1.279, "step": 1321 }, { "epoch": 0.03432810906055828, "grad_norm": 8.987658622095314, "learning_rate": 1e-05, "loss": 1.3048, "step": 1322 }, { "epoch": 0.034354075860150234, "grad_norm": 66.08941696789472, "learning_rate": 1e-05, "loss": 1.2854, "step": 1323 }, { "epoch": 0.034380042659742184, "grad_norm": 5.887929770299614, "learning_rate": 1e-05, "loss": 1.3075, "step": 1324 }, { "epoch": 0.034406009459334135, "grad_norm": 6.373643020118117, "learning_rate": 1e-05, "loss": 1.313, "step": 1325 }, { "epoch": 0.034431976258926085, "grad_norm": 8.354395786355715, "learning_rate": 1e-05, "loss": 1.3093, "step": 1326 }, { "epoch": 0.034457943058518035, "grad_norm": 7.288814168381656, "learning_rate": 1e-05, "loss": 1.2571, "step": 1327 }, { "epoch": 0.034483909858109986, "grad_norm": 9.323224133299922, "learning_rate": 1e-05, "loss": 1.2884, "step": 1328 }, { "epoch": 0.034509876657701936, "grad_norm": 13.656385198486937, "learning_rate": 1e-05, "loss": 1.3294, "step": 1329 }, { "epoch": 0.03453584345729389, "grad_norm": 7.513107290492215, "learning_rate": 1e-05, "loss": 1.3025, "step": 1330 }, { "epoch": 0.03456181025688584, "grad_norm": 6.667980191847242, "learning_rate": 1e-05, "loss": 1.2485, "step": 1331 }, { "epoch": 0.03458777705647779, "grad_norm": 7.939111260705651, "learning_rate": 1e-05, "loss": 1.2629, "step": 1332 }, { "epoch": 0.03461374385606974, "grad_norm": 6.163674462001961, "learning_rate": 1e-05, "loss": 1.3, "step": 1333 }, { "epoch": 0.03463971065566169, "grad_norm": 6.864736490658737, "learning_rate": 1e-05, "loss": 1.3206, "step": 1334 }, { "epoch": 0.03466567745525364, "grad_norm": 4.228491718321565, "learning_rate": 1e-05, "loss": 1.258, "step": 1335 }, { "epoch": 0.03469164425484559, "grad_norm": 5.717942811459134, "learning_rate": 1e-05, "loss": 1.264, "step": 1336 }, { "epoch": 0.03471761105443754, "grad_norm": 6.029926688537816, "learning_rate": 1e-05, "loss": 1.2607, "step": 1337 }, { "epoch": 0.03474357785402949, "grad_norm": 6.445534738119308, "learning_rate": 1e-05, "loss": 1.2639, "step": 1338 }, { "epoch": 0.03476954465362144, "grad_norm": 8.714780717499526, "learning_rate": 1e-05, "loss": 1.3112, "step": 1339 }, { "epoch": 0.03479551145321339, "grad_norm": 7.917055970374158, "learning_rate": 1e-05, "loss": 1.3202, "step": 1340 }, { "epoch": 0.03482147825280534, "grad_norm": 6.418837671321907, "learning_rate": 1e-05, "loss": 1.3004, "step": 1341 }, { "epoch": 0.03484744505239729, "grad_norm": 5.3144664042235785, "learning_rate": 1e-05, "loss": 1.2741, "step": 1342 }, { "epoch": 0.03487341185198924, "grad_norm": 7.45483839385641, "learning_rate": 1e-05, "loss": 1.2517, "step": 1343 }, { "epoch": 0.034899378651581194, "grad_norm": 5.368975079295825, "learning_rate": 1e-05, "loss": 1.2748, "step": 1344 }, { "epoch": 0.034925345451173144, "grad_norm": 5.83749495512195, "learning_rate": 1e-05, "loss": 1.2812, "step": 1345 }, { "epoch": 0.034951312250765094, "grad_norm": 5.948602836120789, "learning_rate": 1e-05, "loss": 1.2683, "step": 1346 }, { "epoch": 0.034977279050357045, "grad_norm": 5.544056006827126, "learning_rate": 1e-05, "loss": 1.2989, "step": 1347 }, { "epoch": 0.035003245849948995, "grad_norm": 5.2454461829246695, "learning_rate": 1e-05, "loss": 1.2333, "step": 1348 }, { "epoch": 0.035029212649540946, "grad_norm": 6.047484374576109, "learning_rate": 1e-05, "loss": 1.2236, "step": 1349 }, { "epoch": 0.035055179449132896, "grad_norm": 14.190101418646886, "learning_rate": 1e-05, "loss": 1.324, "step": 1350 }, { "epoch": 0.03508114624872485, "grad_norm": 5.513903643729548, "learning_rate": 1e-05, "loss": 1.2941, "step": 1351 }, { "epoch": 0.0351071130483168, "grad_norm": 8.444230171742086, "learning_rate": 1e-05, "loss": 1.2822, "step": 1352 }, { "epoch": 0.03513307984790875, "grad_norm": 7.544519408272912, "learning_rate": 1e-05, "loss": 1.3309, "step": 1353 }, { "epoch": 0.0351590466475007, "grad_norm": 23.154551167018845, "learning_rate": 1e-05, "loss": 1.311, "step": 1354 }, { "epoch": 0.03518501344709265, "grad_norm": 10.89652060875182, "learning_rate": 1e-05, "loss": 1.316, "step": 1355 }, { "epoch": 0.0352109802466846, "grad_norm": 8.023164590356185, "learning_rate": 1e-05, "loss": 1.2967, "step": 1356 }, { "epoch": 0.03523694704627655, "grad_norm": 15.070377300275407, "learning_rate": 1e-05, "loss": 1.2911, "step": 1357 }, { "epoch": 0.0352629138458685, "grad_norm": 5.098410130438144, "learning_rate": 1e-05, "loss": 1.267, "step": 1358 }, { "epoch": 0.035288880645460444, "grad_norm": 31.215906206499415, "learning_rate": 1e-05, "loss": 1.3349, "step": 1359 }, { "epoch": 0.035314847445052394, "grad_norm": 6.3400459186804445, "learning_rate": 1e-05, "loss": 1.2551, "step": 1360 }, { "epoch": 0.035340814244644345, "grad_norm": 10.067378975969902, "learning_rate": 1e-05, "loss": 1.2932, "step": 1361 }, { "epoch": 0.035366781044236295, "grad_norm": 6.16290512146121, "learning_rate": 1e-05, "loss": 1.2801, "step": 1362 }, { "epoch": 0.035392747843828246, "grad_norm": 7.364503276413202, "learning_rate": 1e-05, "loss": 1.2806, "step": 1363 }, { "epoch": 0.035418714643420196, "grad_norm": 6.001834271114117, "learning_rate": 1e-05, "loss": 1.2725, "step": 1364 }, { "epoch": 0.035444681443012147, "grad_norm": 6.286827230388621, "learning_rate": 1e-05, "loss": 1.306, "step": 1365 }, { "epoch": 0.0354706482426041, "grad_norm": 14.887464414052767, "learning_rate": 1e-05, "loss": 1.2591, "step": 1366 }, { "epoch": 0.03549661504219605, "grad_norm": 6.4016519560277, "learning_rate": 1e-05, "loss": 1.2636, "step": 1367 }, { "epoch": 0.035522581841788, "grad_norm": 6.615003132952744, "learning_rate": 1e-05, "loss": 1.3254, "step": 1368 }, { "epoch": 0.03554854864137995, "grad_norm": 6.365226883721428, "learning_rate": 1e-05, "loss": 1.2617, "step": 1369 }, { "epoch": 0.0355745154409719, "grad_norm": 5.387531310519334, "learning_rate": 1e-05, "loss": 1.2957, "step": 1370 }, { "epoch": 0.03560048224056385, "grad_norm": 8.066719784312165, "learning_rate": 1e-05, "loss": 1.243, "step": 1371 }, { "epoch": 0.0356264490401558, "grad_norm": 6.08541158771488, "learning_rate": 1e-05, "loss": 1.3317, "step": 1372 }, { "epoch": 0.03565241583974775, "grad_norm": 6.01246936814085, "learning_rate": 1e-05, "loss": 1.2753, "step": 1373 }, { "epoch": 0.0356783826393397, "grad_norm": 7.101684972128753, "learning_rate": 1e-05, "loss": 1.31, "step": 1374 }, { "epoch": 0.03570434943893165, "grad_norm": 6.467450144914387, "learning_rate": 1e-05, "loss": 1.2733, "step": 1375 }, { "epoch": 0.0357303162385236, "grad_norm": 6.2344759595913635, "learning_rate": 1e-05, "loss": 1.2791, "step": 1376 }, { "epoch": 0.03575628303811555, "grad_norm": 15.06135675040884, "learning_rate": 1e-05, "loss": 1.308, "step": 1377 }, { "epoch": 0.0357822498377075, "grad_norm": 6.847727834498194, "learning_rate": 1e-05, "loss": 1.2951, "step": 1378 }, { "epoch": 0.03580821663729945, "grad_norm": 16.621913996358654, "learning_rate": 1e-05, "loss": 1.2541, "step": 1379 }, { "epoch": 0.035834183436891404, "grad_norm": 5.994597864097815, "learning_rate": 1e-05, "loss": 1.2488, "step": 1380 }, { "epoch": 0.035860150236483354, "grad_norm": 6.220765673743623, "learning_rate": 1e-05, "loss": 1.2848, "step": 1381 }, { "epoch": 0.035886117036075305, "grad_norm": 9.416822404051864, "learning_rate": 1e-05, "loss": 1.2891, "step": 1382 }, { "epoch": 0.035912083835667255, "grad_norm": 22.42883099309451, "learning_rate": 1e-05, "loss": 1.2797, "step": 1383 }, { "epoch": 0.035938050635259206, "grad_norm": 7.98351042782689, "learning_rate": 1e-05, "loss": 1.2823, "step": 1384 }, { "epoch": 0.035964017434851156, "grad_norm": 6.81895742686551, "learning_rate": 1e-05, "loss": 1.2781, "step": 1385 }, { "epoch": 0.035989984234443106, "grad_norm": 7.810165422668291, "learning_rate": 1e-05, "loss": 1.1889, "step": 1386 }, { "epoch": 0.03601595103403506, "grad_norm": 5.895842571273773, "learning_rate": 1e-05, "loss": 1.2769, "step": 1387 }, { "epoch": 0.03604191783362701, "grad_norm": 6.526722776323092, "learning_rate": 1e-05, "loss": 1.2288, "step": 1388 }, { "epoch": 0.03606788463321896, "grad_norm": 12.14416625074695, "learning_rate": 1e-05, "loss": 1.3107, "step": 1389 }, { "epoch": 0.03609385143281091, "grad_norm": 7.416112329004734, "learning_rate": 1e-05, "loss": 1.2244, "step": 1390 }, { "epoch": 0.03611981823240286, "grad_norm": 4.8878074061509444, "learning_rate": 1e-05, "loss": 1.2709, "step": 1391 }, { "epoch": 0.03614578503199481, "grad_norm": 7.941081523119879, "learning_rate": 1e-05, "loss": 1.2672, "step": 1392 }, { "epoch": 0.03617175183158676, "grad_norm": 9.067869082360211, "learning_rate": 1e-05, "loss": 1.2382, "step": 1393 }, { "epoch": 0.03619771863117871, "grad_norm": 7.985763518651806, "learning_rate": 1e-05, "loss": 1.3221, "step": 1394 }, { "epoch": 0.03622368543077066, "grad_norm": 6.887612443341868, "learning_rate": 1e-05, "loss": 1.2917, "step": 1395 }, { "epoch": 0.03624965223036261, "grad_norm": 14.971962219761341, "learning_rate": 1e-05, "loss": 1.2713, "step": 1396 }, { "epoch": 0.036275619029954555, "grad_norm": 5.199144278321832, "learning_rate": 1e-05, "loss": 1.2148, "step": 1397 }, { "epoch": 0.036301585829546505, "grad_norm": 5.342294238157496, "learning_rate": 1e-05, "loss": 1.3212, "step": 1398 }, { "epoch": 0.036327552629138456, "grad_norm": 10.385855637088989, "learning_rate": 1e-05, "loss": 1.2486, "step": 1399 }, { "epoch": 0.036353519428730406, "grad_norm": 5.433797342968647, "learning_rate": 1e-05, "loss": 1.2774, "step": 1400 }, { "epoch": 0.03637948622832236, "grad_norm": 5.802861251137336, "learning_rate": 1e-05, "loss": 1.2568, "step": 1401 }, { "epoch": 0.03640545302791431, "grad_norm": 5.069784128081459, "learning_rate": 1e-05, "loss": 1.3167, "step": 1402 }, { "epoch": 0.03643141982750626, "grad_norm": 4.982687924712737, "learning_rate": 1e-05, "loss": 1.2826, "step": 1403 }, { "epoch": 0.03645738662709821, "grad_norm": 5.4853659057972965, "learning_rate": 1e-05, "loss": 1.231, "step": 1404 }, { "epoch": 0.03648335342669016, "grad_norm": 9.900111080278618, "learning_rate": 1e-05, "loss": 1.3256, "step": 1405 }, { "epoch": 0.03650932022628211, "grad_norm": 5.926785086674145, "learning_rate": 1e-05, "loss": 1.2859, "step": 1406 }, { "epoch": 0.03653528702587406, "grad_norm": 7.438520345500167, "learning_rate": 1e-05, "loss": 1.2069, "step": 1407 }, { "epoch": 0.03656125382546601, "grad_norm": 22.967430606237688, "learning_rate": 1e-05, "loss": 1.2574, "step": 1408 }, { "epoch": 0.03658722062505796, "grad_norm": 7.870748431863742, "learning_rate": 1e-05, "loss": 1.2345, "step": 1409 }, { "epoch": 0.03661318742464991, "grad_norm": 26.012742661275045, "learning_rate": 1e-05, "loss": 1.2422, "step": 1410 }, { "epoch": 0.03663915422424186, "grad_norm": 5.980456949592375, "learning_rate": 1e-05, "loss": 1.2297, "step": 1411 }, { "epoch": 0.03666512102383381, "grad_norm": 9.626887148853351, "learning_rate": 1e-05, "loss": 1.286, "step": 1412 }, { "epoch": 0.03669108782342576, "grad_norm": 4.69479230134919, "learning_rate": 1e-05, "loss": 1.3, "step": 1413 }, { "epoch": 0.03671705462301771, "grad_norm": 33.555599703840414, "learning_rate": 1e-05, "loss": 1.2996, "step": 1414 }, { "epoch": 0.03674302142260966, "grad_norm": 8.228200142673405, "learning_rate": 1e-05, "loss": 1.265, "step": 1415 }, { "epoch": 0.036768988222201614, "grad_norm": 7.377671356020894, "learning_rate": 1e-05, "loss": 1.2975, "step": 1416 }, { "epoch": 0.036794955021793564, "grad_norm": 5.1031794417234355, "learning_rate": 1e-05, "loss": 1.2498, "step": 1417 }, { "epoch": 0.036820921821385515, "grad_norm": 6.134105531484336, "learning_rate": 1e-05, "loss": 1.279, "step": 1418 }, { "epoch": 0.036846888620977465, "grad_norm": 24.025969124736935, "learning_rate": 1e-05, "loss": 1.2745, "step": 1419 }, { "epoch": 0.036872855420569416, "grad_norm": 6.122576798642336, "learning_rate": 1e-05, "loss": 1.2846, "step": 1420 }, { "epoch": 0.036898822220161366, "grad_norm": 6.7004810445241425, "learning_rate": 1e-05, "loss": 1.3152, "step": 1421 }, { "epoch": 0.03692478901975332, "grad_norm": 5.777547444537081, "learning_rate": 1e-05, "loss": 1.2586, "step": 1422 }, { "epoch": 0.03695075581934527, "grad_norm": 9.132151688293828, "learning_rate": 1e-05, "loss": 1.3147, "step": 1423 }, { "epoch": 0.03697672261893722, "grad_norm": 14.582955201105877, "learning_rate": 1e-05, "loss": 1.3026, "step": 1424 }, { "epoch": 0.03700268941852917, "grad_norm": 5.062321224117199, "learning_rate": 1e-05, "loss": 1.2416, "step": 1425 }, { "epoch": 0.03702865621812112, "grad_norm": 10.010121373251504, "learning_rate": 1e-05, "loss": 1.2497, "step": 1426 }, { "epoch": 0.03705462301771307, "grad_norm": 6.21851260244895, "learning_rate": 1e-05, "loss": 1.286, "step": 1427 }, { "epoch": 0.03708058981730502, "grad_norm": 5.777751792131986, "learning_rate": 1e-05, "loss": 1.2638, "step": 1428 }, { "epoch": 0.03710655661689697, "grad_norm": 7.849138696155418, "learning_rate": 1e-05, "loss": 1.2964, "step": 1429 }, { "epoch": 0.03713252341648892, "grad_norm": 21.727579421544664, "learning_rate": 1e-05, "loss": 1.3052, "step": 1430 }, { "epoch": 0.03715849021608087, "grad_norm": 7.168280604592141, "learning_rate": 1e-05, "loss": 1.2897, "step": 1431 }, { "epoch": 0.03718445701567282, "grad_norm": 6.860893620155457, "learning_rate": 1e-05, "loss": 1.2532, "step": 1432 }, { "epoch": 0.03721042381526477, "grad_norm": 5.91605366858022, "learning_rate": 1e-05, "loss": 1.2872, "step": 1433 }, { "epoch": 0.03723639061485672, "grad_norm": 13.10571499861797, "learning_rate": 1e-05, "loss": 1.2533, "step": 1434 }, { "epoch": 0.037262357414448666, "grad_norm": 5.526346738150228, "learning_rate": 1e-05, "loss": 1.259, "step": 1435 }, { "epoch": 0.037288324214040616, "grad_norm": 4.281932247404471, "learning_rate": 1e-05, "loss": 1.2192, "step": 1436 }, { "epoch": 0.03731429101363257, "grad_norm": 11.853483925697027, "learning_rate": 1e-05, "loss": 1.2821, "step": 1437 }, { "epoch": 0.03734025781322452, "grad_norm": 4.771696032132735, "learning_rate": 1e-05, "loss": 1.2678, "step": 1438 }, { "epoch": 0.03736622461281647, "grad_norm": 5.6152729278567834, "learning_rate": 1e-05, "loss": 1.3128, "step": 1439 }, { "epoch": 0.03739219141240842, "grad_norm": 5.658687508930889, "learning_rate": 1e-05, "loss": 1.2888, "step": 1440 }, { "epoch": 0.03741815821200037, "grad_norm": 10.231299484217153, "learning_rate": 1e-05, "loss": 1.3061, "step": 1441 }, { "epoch": 0.03744412501159232, "grad_norm": 7.139787286747414, "learning_rate": 1e-05, "loss": 1.3298, "step": 1442 }, { "epoch": 0.03747009181118427, "grad_norm": 7.868600333974291, "learning_rate": 1e-05, "loss": 1.3014, "step": 1443 }, { "epoch": 0.03749605861077622, "grad_norm": 4.475182019160143, "learning_rate": 1e-05, "loss": 1.2558, "step": 1444 }, { "epoch": 0.03752202541036817, "grad_norm": 7.16359949643762, "learning_rate": 1e-05, "loss": 1.2767, "step": 1445 }, { "epoch": 0.03754799220996012, "grad_norm": 4.6829685942504815, "learning_rate": 1e-05, "loss": 1.2677, "step": 1446 }, { "epoch": 0.03757395900955207, "grad_norm": 6.6403541419945, "learning_rate": 1e-05, "loss": 1.2828, "step": 1447 }, { "epoch": 0.03759992580914402, "grad_norm": 8.121197559010646, "learning_rate": 1e-05, "loss": 1.2488, "step": 1448 }, { "epoch": 0.03762589260873597, "grad_norm": 3.6836650756532414, "learning_rate": 1e-05, "loss": 1.3311, "step": 1449 }, { "epoch": 0.03765185940832792, "grad_norm": 6.9600940540141165, "learning_rate": 1e-05, "loss": 1.3126, "step": 1450 }, { "epoch": 0.03767782620791987, "grad_norm": 7.134632225261937, "learning_rate": 1e-05, "loss": 1.2938, "step": 1451 }, { "epoch": 0.037703793007511824, "grad_norm": 5.951107454217293, "learning_rate": 1e-05, "loss": 1.2828, "step": 1452 }, { "epoch": 0.037729759807103774, "grad_norm": 42.6941381071329, "learning_rate": 1e-05, "loss": 1.3023, "step": 1453 }, { "epoch": 0.037755726606695725, "grad_norm": 5.864211220282229, "learning_rate": 1e-05, "loss": 1.2938, "step": 1454 }, { "epoch": 0.037781693406287675, "grad_norm": 8.051684318285716, "learning_rate": 1e-05, "loss": 1.2994, "step": 1455 }, { "epoch": 0.037807660205879626, "grad_norm": 7.452089761616238, "learning_rate": 1e-05, "loss": 1.3164, "step": 1456 }, { "epoch": 0.037833627005471576, "grad_norm": 13.097605572337091, "learning_rate": 1e-05, "loss": 1.2758, "step": 1457 }, { "epoch": 0.03785959380506353, "grad_norm": 8.633665293002826, "learning_rate": 1e-05, "loss": 1.3111, "step": 1458 }, { "epoch": 0.03788556060465548, "grad_norm": 9.917684508161065, "learning_rate": 1e-05, "loss": 1.2751, "step": 1459 }, { "epoch": 0.03791152740424743, "grad_norm": 7.626841088325643, "learning_rate": 1e-05, "loss": 1.2691, "step": 1460 }, { "epoch": 0.03793749420383938, "grad_norm": 5.0988231273437545, "learning_rate": 1e-05, "loss": 1.2727, "step": 1461 }, { "epoch": 0.03796346100343133, "grad_norm": 6.954166497226147, "learning_rate": 1e-05, "loss": 1.2484, "step": 1462 }, { "epoch": 0.03798942780302328, "grad_norm": 6.744684740394798, "learning_rate": 1e-05, "loss": 1.2766, "step": 1463 }, { "epoch": 0.03801539460261523, "grad_norm": 46.278876120752976, "learning_rate": 1e-05, "loss": 1.2363, "step": 1464 }, { "epoch": 0.03804136140220718, "grad_norm": 8.444910031026316, "learning_rate": 1e-05, "loss": 1.3348, "step": 1465 }, { "epoch": 0.03806732820179913, "grad_norm": 6.044633474541921, "learning_rate": 1e-05, "loss": 1.3226, "step": 1466 }, { "epoch": 0.03809329500139108, "grad_norm": 19.769739258068913, "learning_rate": 1e-05, "loss": 1.3162, "step": 1467 }, { "epoch": 0.03811926180098303, "grad_norm": 6.8684058730582125, "learning_rate": 1e-05, "loss": 1.2503, "step": 1468 }, { "epoch": 0.03814522860057498, "grad_norm": 15.07984172950479, "learning_rate": 1e-05, "loss": 1.232, "step": 1469 }, { "epoch": 0.03817119540016693, "grad_norm": 6.185448508749752, "learning_rate": 1e-05, "loss": 1.2444, "step": 1470 }, { "epoch": 0.03819716219975888, "grad_norm": 5.5655541768352315, "learning_rate": 1e-05, "loss": 1.2807, "step": 1471 }, { "epoch": 0.03822312899935083, "grad_norm": 5.199508739154117, "learning_rate": 1e-05, "loss": 1.2854, "step": 1472 }, { "epoch": 0.03824909579894278, "grad_norm": 27.045202775428816, "learning_rate": 1e-05, "loss": 1.3158, "step": 1473 }, { "epoch": 0.03827506259853473, "grad_norm": 5.68296180174457, "learning_rate": 1e-05, "loss": 1.2989, "step": 1474 }, { "epoch": 0.03830102939812668, "grad_norm": 5.807861446146656, "learning_rate": 1e-05, "loss": 1.2668, "step": 1475 }, { "epoch": 0.03832699619771863, "grad_norm": 12.462547040446891, "learning_rate": 1e-05, "loss": 1.3103, "step": 1476 }, { "epoch": 0.03835296299731058, "grad_norm": 6.59137209279478, "learning_rate": 1e-05, "loss": 1.2854, "step": 1477 }, { "epoch": 0.03837892979690253, "grad_norm": 7.160386135408781, "learning_rate": 1e-05, "loss": 1.2857, "step": 1478 }, { "epoch": 0.03840489659649448, "grad_norm": 5.580823400582679, "learning_rate": 1e-05, "loss": 1.2698, "step": 1479 }, { "epoch": 0.03843086339608643, "grad_norm": 7.634270207688742, "learning_rate": 1e-05, "loss": 1.3161, "step": 1480 }, { "epoch": 0.03845683019567838, "grad_norm": 6.093433234956069, "learning_rate": 1e-05, "loss": 1.2266, "step": 1481 }, { "epoch": 0.03848279699527033, "grad_norm": 15.427247136920858, "learning_rate": 1e-05, "loss": 1.2883, "step": 1482 }, { "epoch": 0.03850876379486228, "grad_norm": 5.614115207763378, "learning_rate": 1e-05, "loss": 1.3175, "step": 1483 }, { "epoch": 0.03853473059445423, "grad_norm": 5.689081999031058, "learning_rate": 1e-05, "loss": 1.296, "step": 1484 }, { "epoch": 0.03856069739404618, "grad_norm": 5.808291316416336, "learning_rate": 1e-05, "loss": 1.3124, "step": 1485 }, { "epoch": 0.03858666419363813, "grad_norm": 10.792207760789763, "learning_rate": 1e-05, "loss": 1.2818, "step": 1486 }, { "epoch": 0.03861263099323008, "grad_norm": 6.335518443293532, "learning_rate": 1e-05, "loss": 1.2715, "step": 1487 }, { "epoch": 0.038638597792822034, "grad_norm": 5.887893164722799, "learning_rate": 1e-05, "loss": 1.2594, "step": 1488 }, { "epoch": 0.038664564592413984, "grad_norm": 3.993757980410344, "learning_rate": 1e-05, "loss": 1.3022, "step": 1489 }, { "epoch": 0.038690531392005935, "grad_norm": 5.856556288160858, "learning_rate": 1e-05, "loss": 1.253, "step": 1490 }, { "epoch": 0.038716498191597885, "grad_norm": 9.759506286265722, "learning_rate": 1e-05, "loss": 1.2971, "step": 1491 }, { "epoch": 0.038742464991189836, "grad_norm": 34.172771621753185, "learning_rate": 1e-05, "loss": 1.2788, "step": 1492 }, { "epoch": 0.038768431790781786, "grad_norm": 5.482639831624491, "learning_rate": 1e-05, "loss": 1.3126, "step": 1493 }, { "epoch": 0.03879439859037374, "grad_norm": 5.40406799383187, "learning_rate": 1e-05, "loss": 1.2687, "step": 1494 }, { "epoch": 0.03882036538996569, "grad_norm": 6.3338955412171725, "learning_rate": 1e-05, "loss": 1.2425, "step": 1495 }, { "epoch": 0.03884633218955764, "grad_norm": 8.60599535951128, "learning_rate": 1e-05, "loss": 1.2556, "step": 1496 }, { "epoch": 0.03887229898914959, "grad_norm": 6.157586122159124, "learning_rate": 1e-05, "loss": 1.3066, "step": 1497 }, { "epoch": 0.03889826578874154, "grad_norm": 6.147750021126885, "learning_rate": 1e-05, "loss": 1.2236, "step": 1498 }, { "epoch": 0.03892423258833349, "grad_norm": 6.727390712387615, "learning_rate": 1e-05, "loss": 1.2583, "step": 1499 }, { "epoch": 0.03895019938792544, "grad_norm": 6.72073519454351, "learning_rate": 1e-05, "loss": 1.2894, "step": 1500 }, { "epoch": 0.03897616618751739, "grad_norm": 8.904263673098331, "learning_rate": 1e-05, "loss": 1.3204, "step": 1501 }, { "epoch": 0.03900213298710934, "grad_norm": 6.464127957918814, "learning_rate": 1e-05, "loss": 1.2664, "step": 1502 }, { "epoch": 0.03902809978670129, "grad_norm": 8.704574382090135, "learning_rate": 1e-05, "loss": 1.2541, "step": 1503 }, { "epoch": 0.03905406658629324, "grad_norm": 230.90646780677235, "learning_rate": 1e-05, "loss": 1.2958, "step": 1504 }, { "epoch": 0.03908003338588519, "grad_norm": 8.7469568682809, "learning_rate": 1e-05, "loss": 1.266, "step": 1505 }, { "epoch": 0.03910600018547714, "grad_norm": 8.314097315545643, "learning_rate": 1e-05, "loss": 1.2709, "step": 1506 }, { "epoch": 0.03913196698506909, "grad_norm": 8.83228780448917, "learning_rate": 1e-05, "loss": 1.2937, "step": 1507 }, { "epoch": 0.03915793378466104, "grad_norm": 10.28455353045531, "learning_rate": 1e-05, "loss": 1.3232, "step": 1508 }, { "epoch": 0.039183900584252994, "grad_norm": 5.705399402007346, "learning_rate": 1e-05, "loss": 1.303, "step": 1509 }, { "epoch": 0.039209867383844944, "grad_norm": 6.962077277618459, "learning_rate": 1e-05, "loss": 1.2473, "step": 1510 }, { "epoch": 0.039235834183436895, "grad_norm": 6.118213241148117, "learning_rate": 1e-05, "loss": 1.2921, "step": 1511 }, { "epoch": 0.03926180098302884, "grad_norm": 19.792221329261764, "learning_rate": 1e-05, "loss": 1.2777, "step": 1512 }, { "epoch": 0.03928776778262079, "grad_norm": 6.994925430696589, "learning_rate": 1e-05, "loss": 1.2686, "step": 1513 }, { "epoch": 0.03931373458221274, "grad_norm": 4.883692303549341, "learning_rate": 1e-05, "loss": 1.2317, "step": 1514 }, { "epoch": 0.03933970138180469, "grad_norm": 6.984074050598138, "learning_rate": 1e-05, "loss": 1.2781, "step": 1515 }, { "epoch": 0.03936566818139664, "grad_norm": 6.60307108779902, "learning_rate": 1e-05, "loss": 1.3001, "step": 1516 }, { "epoch": 0.03939163498098859, "grad_norm": 6.929970168893084, "learning_rate": 1e-05, "loss": 1.3534, "step": 1517 }, { "epoch": 0.03941760178058054, "grad_norm": 11.377211806340256, "learning_rate": 1e-05, "loss": 1.2973, "step": 1518 }, { "epoch": 0.03944356858017249, "grad_norm": 5.198329972811036, "learning_rate": 1e-05, "loss": 1.2875, "step": 1519 }, { "epoch": 0.03946953537976444, "grad_norm": 6.934549890060232, "learning_rate": 1e-05, "loss": 1.2579, "step": 1520 }, { "epoch": 0.03949550217935639, "grad_norm": 6.806673410520037, "learning_rate": 1e-05, "loss": 1.2698, "step": 1521 }, { "epoch": 0.03952146897894834, "grad_norm": 6.016655375433953, "learning_rate": 1e-05, "loss": 1.2438, "step": 1522 }, { "epoch": 0.039547435778540294, "grad_norm": 6.917758380571187, "learning_rate": 1e-05, "loss": 1.2748, "step": 1523 }, { "epoch": 0.039573402578132244, "grad_norm": 6.925806586600582, "learning_rate": 1e-05, "loss": 1.249, "step": 1524 }, { "epoch": 0.039599369377724195, "grad_norm": 5.201284822046263, "learning_rate": 1e-05, "loss": 1.2219, "step": 1525 }, { "epoch": 0.039625336177316145, "grad_norm": 8.457430886920168, "learning_rate": 1e-05, "loss": 1.2354, "step": 1526 }, { "epoch": 0.039651302976908095, "grad_norm": 8.749286731531141, "learning_rate": 1e-05, "loss": 1.2678, "step": 1527 }, { "epoch": 0.039677269776500046, "grad_norm": 4.15235477420926, "learning_rate": 1e-05, "loss": 1.2707, "step": 1528 }, { "epoch": 0.039703236576091996, "grad_norm": 15.424367160273444, "learning_rate": 1e-05, "loss": 1.284, "step": 1529 }, { "epoch": 0.03972920337568395, "grad_norm": 7.951138051466379, "learning_rate": 1e-05, "loss": 1.3092, "step": 1530 }, { "epoch": 0.0397551701752759, "grad_norm": 11.134053179222466, "learning_rate": 1e-05, "loss": 1.2855, "step": 1531 }, { "epoch": 0.03978113697486785, "grad_norm": 11.399294219003483, "learning_rate": 1e-05, "loss": 1.298, "step": 1532 }, { "epoch": 0.0398071037744598, "grad_norm": 5.643768364590224, "learning_rate": 1e-05, "loss": 1.3036, "step": 1533 }, { "epoch": 0.03983307057405175, "grad_norm": 5.508199648405931, "learning_rate": 1e-05, "loss": 1.2639, "step": 1534 }, { "epoch": 0.0398590373736437, "grad_norm": 6.3387302024801455, "learning_rate": 1e-05, "loss": 1.3057, "step": 1535 }, { "epoch": 0.03988500417323565, "grad_norm": 9.329324769734393, "learning_rate": 1e-05, "loss": 1.2625, "step": 1536 }, { "epoch": 0.0399109709728276, "grad_norm": 4.171614620927037, "learning_rate": 1e-05, "loss": 1.31, "step": 1537 }, { "epoch": 0.03993693777241955, "grad_norm": 5.82453050808805, "learning_rate": 1e-05, "loss": 1.2761, "step": 1538 }, { "epoch": 0.0399629045720115, "grad_norm": 5.530596333473725, "learning_rate": 1e-05, "loss": 1.2582, "step": 1539 }, { "epoch": 0.03998887137160345, "grad_norm": 5.391156598844686, "learning_rate": 1e-05, "loss": 1.2772, "step": 1540 }, { "epoch": 0.0400148381711954, "grad_norm": 5.12743952305847, "learning_rate": 1e-05, "loss": 1.2514, "step": 1541 }, { "epoch": 0.04004080497078735, "grad_norm": 4.692946461497138, "learning_rate": 1e-05, "loss": 1.2877, "step": 1542 }, { "epoch": 0.0400667717703793, "grad_norm": 5.411136072994641, "learning_rate": 1e-05, "loss": 1.2847, "step": 1543 }, { "epoch": 0.040092738569971254, "grad_norm": 5.260227595289775, "learning_rate": 1e-05, "loss": 1.2832, "step": 1544 }, { "epoch": 0.040118705369563204, "grad_norm": 5.673134908168437, "learning_rate": 1e-05, "loss": 1.2988, "step": 1545 }, { "epoch": 0.040144672169155154, "grad_norm": 4.314192287837383, "learning_rate": 1e-05, "loss": 1.3112, "step": 1546 }, { "epoch": 0.040170638968747105, "grad_norm": 7.771581073474804, "learning_rate": 1e-05, "loss": 1.2592, "step": 1547 }, { "epoch": 0.040196605768339055, "grad_norm": 4.3489550683776415, "learning_rate": 1e-05, "loss": 1.3008, "step": 1548 }, { "epoch": 0.040222572567931006, "grad_norm": 9.45802792855695, "learning_rate": 1e-05, "loss": 1.3124, "step": 1549 }, { "epoch": 0.04024853936752295, "grad_norm": 7.97514344089384, "learning_rate": 1e-05, "loss": 1.2442, "step": 1550 }, { "epoch": 0.0402745061671149, "grad_norm": 4.9961943448507435, "learning_rate": 1e-05, "loss": 1.2687, "step": 1551 }, { "epoch": 0.04030047296670685, "grad_norm": 5.075194842727754, "learning_rate": 1e-05, "loss": 1.2991, "step": 1552 }, { "epoch": 0.0403264397662988, "grad_norm": 6.7905830813859485, "learning_rate": 1e-05, "loss": 1.277, "step": 1553 }, { "epoch": 0.04035240656589075, "grad_norm": 9.504327691796346, "learning_rate": 1e-05, "loss": 1.2381, "step": 1554 }, { "epoch": 0.0403783733654827, "grad_norm": 5.473604151491644, "learning_rate": 1e-05, "loss": 1.3012, "step": 1555 }, { "epoch": 0.04040434016507465, "grad_norm": 5.153298938006412, "learning_rate": 1e-05, "loss": 1.2892, "step": 1556 }, { "epoch": 0.0404303069646666, "grad_norm": 5.396366147083034, "learning_rate": 1e-05, "loss": 1.2443, "step": 1557 }, { "epoch": 0.04045627376425855, "grad_norm": 6.357134450107851, "learning_rate": 1e-05, "loss": 1.2963, "step": 1558 }, { "epoch": 0.040482240563850504, "grad_norm": 4.983854356318523, "learning_rate": 1e-05, "loss": 1.3173, "step": 1559 }, { "epoch": 0.040508207363442454, "grad_norm": 6.12650918857113, "learning_rate": 1e-05, "loss": 1.2828, "step": 1560 }, { "epoch": 0.040534174163034405, "grad_norm": 6.496121349723389, "learning_rate": 1e-05, "loss": 1.2582, "step": 1561 }, { "epoch": 0.040560140962626355, "grad_norm": 5.658301892799517, "learning_rate": 1e-05, "loss": 1.2937, "step": 1562 }, { "epoch": 0.040586107762218306, "grad_norm": 52.26523262184552, "learning_rate": 1e-05, "loss": 1.2989, "step": 1563 }, { "epoch": 0.040612074561810256, "grad_norm": 12.123319843722804, "learning_rate": 1e-05, "loss": 1.2935, "step": 1564 }, { "epoch": 0.040638041361402207, "grad_norm": 8.749335127501967, "learning_rate": 1e-05, "loss": 1.3102, "step": 1565 }, { "epoch": 0.04066400816099416, "grad_norm": 4.830442989403451, "learning_rate": 1e-05, "loss": 1.2955, "step": 1566 }, { "epoch": 0.04068997496058611, "grad_norm": 7.663544088051034, "learning_rate": 1e-05, "loss": 1.312, "step": 1567 }, { "epoch": 0.04071594176017806, "grad_norm": 41.410934286920224, "learning_rate": 1e-05, "loss": 1.274, "step": 1568 }, { "epoch": 0.04074190855977001, "grad_norm": 5.6033888304210775, "learning_rate": 1e-05, "loss": 1.2719, "step": 1569 }, { "epoch": 0.04076787535936196, "grad_norm": 5.278123821938238, "learning_rate": 1e-05, "loss": 1.2583, "step": 1570 }, { "epoch": 0.04079384215895391, "grad_norm": 6.682631387100397, "learning_rate": 1e-05, "loss": 1.2896, "step": 1571 }, { "epoch": 0.04081980895854586, "grad_norm": 4.811918743750767, "learning_rate": 1e-05, "loss": 1.2536, "step": 1572 }, { "epoch": 0.04084577575813781, "grad_norm": 6.676931426115066, "learning_rate": 1e-05, "loss": 1.2612, "step": 1573 }, { "epoch": 0.04087174255772976, "grad_norm": 5.849217758698327, "learning_rate": 1e-05, "loss": 1.3477, "step": 1574 }, { "epoch": 0.04089770935732171, "grad_norm": 5.183037366522977, "learning_rate": 1e-05, "loss": 1.2544, "step": 1575 }, { "epoch": 0.04092367615691366, "grad_norm": 4.2552906092234535, "learning_rate": 1e-05, "loss": 1.2379, "step": 1576 }, { "epoch": 0.04094964295650561, "grad_norm": 6.9336341878222205, "learning_rate": 1e-05, "loss": 1.3269, "step": 1577 }, { "epoch": 0.04097560975609756, "grad_norm": 7.0298327564732555, "learning_rate": 1e-05, "loss": 1.2885, "step": 1578 }, { "epoch": 0.04100157655568951, "grad_norm": 4.745809011198867, "learning_rate": 1e-05, "loss": 1.3265, "step": 1579 }, { "epoch": 0.041027543355281464, "grad_norm": 5.210202647128565, "learning_rate": 1e-05, "loss": 1.3042, "step": 1580 }, { "epoch": 0.041053510154873414, "grad_norm": 7.041838273284174, "learning_rate": 1e-05, "loss": 1.3106, "step": 1581 }, { "epoch": 0.041079476954465365, "grad_norm": 6.236160825641183, "learning_rate": 1e-05, "loss": 1.3374, "step": 1582 }, { "epoch": 0.041105443754057315, "grad_norm": 6.4219905350959, "learning_rate": 1e-05, "loss": 1.2489, "step": 1583 }, { "epoch": 0.041131410553649266, "grad_norm": 7.682454011570299, "learning_rate": 1e-05, "loss": 1.3087, "step": 1584 }, { "epoch": 0.041157377353241216, "grad_norm": 7.7177128268727495, "learning_rate": 1e-05, "loss": 1.286, "step": 1585 }, { "epoch": 0.041183344152833166, "grad_norm": 10.128778576735844, "learning_rate": 1e-05, "loss": 1.2398, "step": 1586 }, { "epoch": 0.04120931095242512, "grad_norm": 6.953697746775856, "learning_rate": 1e-05, "loss": 1.3156, "step": 1587 }, { "epoch": 0.04123527775201706, "grad_norm": 4.240081208009193, "learning_rate": 1e-05, "loss": 1.2764, "step": 1588 }, { "epoch": 0.04126124455160901, "grad_norm": 4.732164326333753, "learning_rate": 1e-05, "loss": 1.2564, "step": 1589 }, { "epoch": 0.04128721135120096, "grad_norm": 6.117078368478966, "learning_rate": 1e-05, "loss": 1.32, "step": 1590 }, { "epoch": 0.04131317815079291, "grad_norm": 6.6958135820187294, "learning_rate": 1e-05, "loss": 1.3126, "step": 1591 }, { "epoch": 0.04133914495038486, "grad_norm": 6.139776350787976, "learning_rate": 1e-05, "loss": 1.2992, "step": 1592 }, { "epoch": 0.04136511174997681, "grad_norm": 10.18764191511104, "learning_rate": 1e-05, "loss": 1.2713, "step": 1593 }, { "epoch": 0.04139107854956876, "grad_norm": 9.943031163250918, "learning_rate": 1e-05, "loss": 1.2872, "step": 1594 }, { "epoch": 0.041417045349160714, "grad_norm": 8.348140632881503, "learning_rate": 1e-05, "loss": 1.2591, "step": 1595 }, { "epoch": 0.041443012148752664, "grad_norm": 5.839804556724667, "learning_rate": 1e-05, "loss": 1.2527, "step": 1596 }, { "epoch": 0.041468978948344615, "grad_norm": 5.574327046786556, "learning_rate": 1e-05, "loss": 1.2672, "step": 1597 }, { "epoch": 0.041494945747936565, "grad_norm": 4.743062021981196, "learning_rate": 1e-05, "loss": 1.2872, "step": 1598 }, { "epoch": 0.041520912547528516, "grad_norm": 14.210874686330392, "learning_rate": 1e-05, "loss": 1.2772, "step": 1599 }, { "epoch": 0.041546879347120466, "grad_norm": 6.198361814097883, "learning_rate": 1e-05, "loss": 1.3522, "step": 1600 }, { "epoch": 0.04157284614671242, "grad_norm": 8.900735854685433, "learning_rate": 1e-05, "loss": 1.2899, "step": 1601 }, { "epoch": 0.04159881294630437, "grad_norm": 6.409308768627204, "learning_rate": 1e-05, "loss": 1.3042, "step": 1602 }, { "epoch": 0.04162477974589632, "grad_norm": 4.7726616611347, "learning_rate": 1e-05, "loss": 1.286, "step": 1603 }, { "epoch": 0.04165074654548827, "grad_norm": 6.194499978736966, "learning_rate": 1e-05, "loss": 1.2659, "step": 1604 }, { "epoch": 0.04167671334508022, "grad_norm": 6.592966910049068, "learning_rate": 1e-05, "loss": 1.2591, "step": 1605 }, { "epoch": 0.04170268014467217, "grad_norm": 5.141668799269536, "learning_rate": 1e-05, "loss": 1.2737, "step": 1606 }, { "epoch": 0.04172864694426412, "grad_norm": 5.3240494788641435, "learning_rate": 1e-05, "loss": 1.2803, "step": 1607 }, { "epoch": 0.04175461374385607, "grad_norm": 5.044879908724987, "learning_rate": 1e-05, "loss": 1.2347, "step": 1608 }, { "epoch": 0.04178058054344802, "grad_norm": 4.262929434972354, "learning_rate": 1e-05, "loss": 1.2958, "step": 1609 }, { "epoch": 0.04180654734303997, "grad_norm": 13.298172238608407, "learning_rate": 1e-05, "loss": 1.2524, "step": 1610 }, { "epoch": 0.04183251414263192, "grad_norm": 4.6723336597468, "learning_rate": 1e-05, "loss": 1.3352, "step": 1611 }, { "epoch": 0.04185848094222387, "grad_norm": 5.443286841887574, "learning_rate": 1e-05, "loss": 1.325, "step": 1612 }, { "epoch": 0.04188444774181582, "grad_norm": 6.777347690020469, "learning_rate": 1e-05, "loss": 1.2707, "step": 1613 }, { "epoch": 0.04191041454140777, "grad_norm": 5.986055700877157, "learning_rate": 1e-05, "loss": 1.2717, "step": 1614 }, { "epoch": 0.04193638134099972, "grad_norm": 4.1133304024247455, "learning_rate": 1e-05, "loss": 1.3131, "step": 1615 }, { "epoch": 0.041962348140591674, "grad_norm": 4.975403751111287, "learning_rate": 1e-05, "loss": 1.2994, "step": 1616 }, { "epoch": 0.041988314940183624, "grad_norm": 6.154702548812207, "learning_rate": 1e-05, "loss": 1.2708, "step": 1617 }, { "epoch": 0.042014281739775575, "grad_norm": 4.536578603464845, "learning_rate": 1e-05, "loss": 1.2784, "step": 1618 }, { "epoch": 0.042040248539367525, "grad_norm": 4.834343815550682, "learning_rate": 1e-05, "loss": 1.2852, "step": 1619 }, { "epoch": 0.042066215338959476, "grad_norm": 4.982393355158967, "learning_rate": 1e-05, "loss": 1.3138, "step": 1620 }, { "epoch": 0.042092182138551426, "grad_norm": 4.873503626515143, "learning_rate": 1e-05, "loss": 1.2717, "step": 1621 }, { "epoch": 0.04211814893814338, "grad_norm": 4.504399374362234, "learning_rate": 1e-05, "loss": 1.2838, "step": 1622 }, { "epoch": 0.04214411573773533, "grad_norm": 4.407870812961715, "learning_rate": 1e-05, "loss": 1.2536, "step": 1623 }, { "epoch": 0.04217008253732728, "grad_norm": 5.306418494614822, "learning_rate": 1e-05, "loss": 1.2655, "step": 1624 }, { "epoch": 0.04219604933691923, "grad_norm": 3.6244643243082963, "learning_rate": 1e-05, "loss": 1.2685, "step": 1625 }, { "epoch": 0.04222201613651118, "grad_norm": 3.928382125851428, "learning_rate": 1e-05, "loss": 1.304, "step": 1626 }, { "epoch": 0.04224798293610312, "grad_norm": 3.887207245919555, "learning_rate": 1e-05, "loss": 1.3044, "step": 1627 }, { "epoch": 0.04227394973569507, "grad_norm": 4.514326601895249, "learning_rate": 1e-05, "loss": 1.2872, "step": 1628 }, { "epoch": 0.04229991653528702, "grad_norm": 6.216234781571489, "learning_rate": 1e-05, "loss": 1.2877, "step": 1629 }, { "epoch": 0.04232588333487897, "grad_norm": 5.146570726860949, "learning_rate": 1e-05, "loss": 1.2476, "step": 1630 }, { "epoch": 0.042351850134470924, "grad_norm": 3.9483104230979045, "learning_rate": 1e-05, "loss": 1.2398, "step": 1631 }, { "epoch": 0.042377816934062874, "grad_norm": 15.242038837531044, "learning_rate": 1e-05, "loss": 1.237, "step": 1632 }, { "epoch": 0.042403783733654825, "grad_norm": 13.170678280081928, "learning_rate": 1e-05, "loss": 1.359, "step": 1633 }, { "epoch": 0.042429750533246775, "grad_norm": 5.698616649676542, "learning_rate": 1e-05, "loss": 1.2588, "step": 1634 }, { "epoch": 0.042455717332838726, "grad_norm": 5.083832242069257, "learning_rate": 1e-05, "loss": 1.2782, "step": 1635 }, { "epoch": 0.042481684132430676, "grad_norm": 6.053980238429501, "learning_rate": 1e-05, "loss": 1.284, "step": 1636 }, { "epoch": 0.04250765093202263, "grad_norm": 5.387022899204414, "learning_rate": 1e-05, "loss": 1.2932, "step": 1637 }, { "epoch": 0.04253361773161458, "grad_norm": 4.038909023770518, "learning_rate": 1e-05, "loss": 1.2293, "step": 1638 }, { "epoch": 0.04255958453120653, "grad_norm": 4.139760974489145, "learning_rate": 1e-05, "loss": 1.3216, "step": 1639 }, { "epoch": 0.04258555133079848, "grad_norm": 5.4035219575095725, "learning_rate": 1e-05, "loss": 1.2863, "step": 1640 }, { "epoch": 0.04261151813039043, "grad_norm": 6.2970761250448115, "learning_rate": 1e-05, "loss": 1.3261, "step": 1641 }, { "epoch": 0.04263748492998238, "grad_norm": 5.329534528779885, "learning_rate": 1e-05, "loss": 1.3137, "step": 1642 }, { "epoch": 0.04266345172957433, "grad_norm": 7.443329393881498, "learning_rate": 1e-05, "loss": 1.2892, "step": 1643 }, { "epoch": 0.04268941852916628, "grad_norm": 5.738344071409867, "learning_rate": 1e-05, "loss": 1.2635, "step": 1644 }, { "epoch": 0.04271538532875823, "grad_norm": 4.77996868841538, "learning_rate": 1e-05, "loss": 1.3205, "step": 1645 }, { "epoch": 0.04274135212835018, "grad_norm": 4.77479306716307, "learning_rate": 1e-05, "loss": 1.2774, "step": 1646 }, { "epoch": 0.04276731892794213, "grad_norm": 4.2632685709536355, "learning_rate": 1e-05, "loss": 1.2909, "step": 1647 }, { "epoch": 0.04279328572753408, "grad_norm": 7.572750994003929, "learning_rate": 1e-05, "loss": 1.2704, "step": 1648 }, { "epoch": 0.04281925252712603, "grad_norm": 5.083416901615475, "learning_rate": 1e-05, "loss": 1.2552, "step": 1649 }, { "epoch": 0.04284521932671798, "grad_norm": 5.0316762802842545, "learning_rate": 1e-05, "loss": 1.3323, "step": 1650 }, { "epoch": 0.04287118612630993, "grad_norm": 4.884612949308346, "learning_rate": 1e-05, "loss": 1.2424, "step": 1651 }, { "epoch": 0.042897152925901884, "grad_norm": 5.171474475852562, "learning_rate": 1e-05, "loss": 1.2861, "step": 1652 }, { "epoch": 0.042923119725493834, "grad_norm": 4.924495713804442, "learning_rate": 1e-05, "loss": 1.3101, "step": 1653 }, { "epoch": 0.042949086525085785, "grad_norm": 7.149377446314682, "learning_rate": 1e-05, "loss": 1.3096, "step": 1654 }, { "epoch": 0.042975053324677735, "grad_norm": 5.797001675355309, "learning_rate": 1e-05, "loss": 1.2918, "step": 1655 }, { "epoch": 0.043001020124269686, "grad_norm": 3.939321550735651, "learning_rate": 1e-05, "loss": 1.2736, "step": 1656 }, { "epoch": 0.043026986923861636, "grad_norm": 4.004895790449284, "learning_rate": 1e-05, "loss": 1.2812, "step": 1657 }, { "epoch": 0.04305295372345359, "grad_norm": 6.800438799444506, "learning_rate": 1e-05, "loss": 1.2928, "step": 1658 }, { "epoch": 0.04307892052304554, "grad_norm": 5.086438132402267, "learning_rate": 1e-05, "loss": 1.3017, "step": 1659 }, { "epoch": 0.04310488732263749, "grad_norm": 5.3678727913236735, "learning_rate": 1e-05, "loss": 1.2706, "step": 1660 }, { "epoch": 0.04313085412222944, "grad_norm": 4.224940752144444, "learning_rate": 1e-05, "loss": 1.2924, "step": 1661 }, { "epoch": 0.04315682092182139, "grad_norm": 4.507970321246173, "learning_rate": 1e-05, "loss": 1.2744, "step": 1662 }, { "epoch": 0.04318278772141334, "grad_norm": 9.29240717880775, "learning_rate": 1e-05, "loss": 1.2754, "step": 1663 }, { "epoch": 0.04320875452100529, "grad_norm": 5.16085960798301, "learning_rate": 1e-05, "loss": 1.2728, "step": 1664 }, { "epoch": 0.04323472132059723, "grad_norm": 4.185714796634459, "learning_rate": 1e-05, "loss": 1.2811, "step": 1665 }, { "epoch": 0.043260688120189184, "grad_norm": 5.423334194684249, "learning_rate": 1e-05, "loss": 1.3203, "step": 1666 }, { "epoch": 0.043286654919781134, "grad_norm": 4.431959534993998, "learning_rate": 1e-05, "loss": 1.3058, "step": 1667 }, { "epoch": 0.043312621719373084, "grad_norm": 4.223844941385418, "learning_rate": 1e-05, "loss": 1.262, "step": 1668 }, { "epoch": 0.043338588518965035, "grad_norm": 4.236600509195045, "learning_rate": 1e-05, "loss": 1.2575, "step": 1669 }, { "epoch": 0.043364555318556985, "grad_norm": 5.682279433514941, "learning_rate": 1e-05, "loss": 1.239, "step": 1670 }, { "epoch": 0.043390522118148936, "grad_norm": 5.437041690848293, "learning_rate": 1e-05, "loss": 1.2809, "step": 1671 }, { "epoch": 0.043416488917740886, "grad_norm": 6.959859471272626, "learning_rate": 1e-05, "loss": 1.2526, "step": 1672 }, { "epoch": 0.04344245571733284, "grad_norm": 7.1744513112454555, "learning_rate": 1e-05, "loss": 1.2431, "step": 1673 }, { "epoch": 0.04346842251692479, "grad_norm": 4.71719698581203, "learning_rate": 1e-05, "loss": 1.3192, "step": 1674 }, { "epoch": 0.04349438931651674, "grad_norm": 4.903291624312361, "learning_rate": 1e-05, "loss": 1.2535, "step": 1675 }, { "epoch": 0.04352035611610869, "grad_norm": 4.336343282321383, "learning_rate": 1e-05, "loss": 1.3066, "step": 1676 }, { "epoch": 0.04354632291570064, "grad_norm": 5.827510034265312, "learning_rate": 1e-05, "loss": 1.2613, "step": 1677 }, { "epoch": 0.04357228971529259, "grad_norm": 6.486585125516829, "learning_rate": 1e-05, "loss": 1.3346, "step": 1678 }, { "epoch": 0.04359825651488454, "grad_norm": 7.0430120804268865, "learning_rate": 1e-05, "loss": 1.2722, "step": 1679 }, { "epoch": 0.04362422331447649, "grad_norm": 5.577999166998078, "learning_rate": 1e-05, "loss": 1.2594, "step": 1680 }, { "epoch": 0.04365019011406844, "grad_norm": 7.111682425172254, "learning_rate": 1e-05, "loss": 1.2974, "step": 1681 }, { "epoch": 0.04367615691366039, "grad_norm": 5.432952382281181, "learning_rate": 1e-05, "loss": 1.311, "step": 1682 }, { "epoch": 0.04370212371325234, "grad_norm": 18.59346359417025, "learning_rate": 1e-05, "loss": 1.3342, "step": 1683 }, { "epoch": 0.04372809051284429, "grad_norm": 5.355081601273581, "learning_rate": 1e-05, "loss": 1.269, "step": 1684 }, { "epoch": 0.04375405731243624, "grad_norm": 5.601445706265851, "learning_rate": 1e-05, "loss": 1.2881, "step": 1685 }, { "epoch": 0.04378002411202819, "grad_norm": 4.83086152285268, "learning_rate": 1e-05, "loss": 1.2405, "step": 1686 }, { "epoch": 0.04380599091162014, "grad_norm": 9.684307138279701, "learning_rate": 1e-05, "loss": 1.2805, "step": 1687 }, { "epoch": 0.043831957711212094, "grad_norm": 5.304341772664668, "learning_rate": 1e-05, "loss": 1.2806, "step": 1688 }, { "epoch": 0.043857924510804044, "grad_norm": 7.197955678419139, "learning_rate": 1e-05, "loss": 1.2525, "step": 1689 }, { "epoch": 0.043883891310395995, "grad_norm": 3.992233485126849, "learning_rate": 1e-05, "loss": 1.2641, "step": 1690 }, { "epoch": 0.043909858109987945, "grad_norm": 6.153373859695, "learning_rate": 1e-05, "loss": 1.2583, "step": 1691 }, { "epoch": 0.043935824909579896, "grad_norm": 5.620913568277048, "learning_rate": 1e-05, "loss": 1.321, "step": 1692 }, { "epoch": 0.043961791709171846, "grad_norm": 4.874449283108671, "learning_rate": 1e-05, "loss": 1.2803, "step": 1693 }, { "epoch": 0.0439877585087638, "grad_norm": 6.551026405110059, "learning_rate": 1e-05, "loss": 1.2923, "step": 1694 }, { "epoch": 0.04401372530835575, "grad_norm": 7.427440090804518, "learning_rate": 1e-05, "loss": 1.2772, "step": 1695 }, { "epoch": 0.0440396921079477, "grad_norm": 6.21151020600519, "learning_rate": 1e-05, "loss": 1.2523, "step": 1696 }, { "epoch": 0.04406565890753965, "grad_norm": 5.006516977875523, "learning_rate": 1e-05, "loss": 1.246, "step": 1697 }, { "epoch": 0.0440916257071316, "grad_norm": 28.619785150325313, "learning_rate": 1e-05, "loss": 1.2812, "step": 1698 }, { "epoch": 0.04411759250672355, "grad_norm": 5.047571092910756, "learning_rate": 1e-05, "loss": 1.2984, "step": 1699 }, { "epoch": 0.0441435593063155, "grad_norm": 3.9835839776597997, "learning_rate": 1e-05, "loss": 1.335, "step": 1700 }, { "epoch": 0.04416952610590745, "grad_norm": 10.743939487812346, "learning_rate": 1e-05, "loss": 1.2598, "step": 1701 }, { "epoch": 0.0441954929054994, "grad_norm": 4.076693579604946, "learning_rate": 1e-05, "loss": 1.2902, "step": 1702 }, { "epoch": 0.044221459705091344, "grad_norm": 10.20258510966323, "learning_rate": 1e-05, "loss": 1.2388, "step": 1703 }, { "epoch": 0.044247426504683295, "grad_norm": 6.25949778824429, "learning_rate": 1e-05, "loss": 1.2606, "step": 1704 }, { "epoch": 0.044273393304275245, "grad_norm": 5.311398560279364, "learning_rate": 1e-05, "loss": 1.2303, "step": 1705 }, { "epoch": 0.044299360103867196, "grad_norm": 5.193179594775138, "learning_rate": 1e-05, "loss": 1.2751, "step": 1706 }, { "epoch": 0.044325326903459146, "grad_norm": 3.894491947323938, "learning_rate": 1e-05, "loss": 1.3187, "step": 1707 }, { "epoch": 0.044351293703051096, "grad_norm": 6.438040108846421, "learning_rate": 1e-05, "loss": 1.2699, "step": 1708 }, { "epoch": 0.04437726050264305, "grad_norm": 5.312434117525766, "learning_rate": 1e-05, "loss": 1.2379, "step": 1709 }, { "epoch": 0.044403227302235, "grad_norm": 4.215068878298872, "learning_rate": 1e-05, "loss": 1.3315, "step": 1710 }, { "epoch": 0.04442919410182695, "grad_norm": 5.926571717218644, "learning_rate": 1e-05, "loss": 1.252, "step": 1711 }, { "epoch": 0.0444551609014189, "grad_norm": 4.055077918041368, "learning_rate": 1e-05, "loss": 1.282, "step": 1712 }, { "epoch": 0.04448112770101085, "grad_norm": 5.243997139273066, "learning_rate": 1e-05, "loss": 1.2574, "step": 1713 }, { "epoch": 0.0445070945006028, "grad_norm": 7.081498510540348, "learning_rate": 1e-05, "loss": 1.2568, "step": 1714 }, { "epoch": 0.04453306130019475, "grad_norm": 12.43062658496835, "learning_rate": 1e-05, "loss": 1.2932, "step": 1715 }, { "epoch": 0.0445590280997867, "grad_norm": 5.874260064461875, "learning_rate": 1e-05, "loss": 1.3023, "step": 1716 }, { "epoch": 0.04458499489937865, "grad_norm": 4.308355800390538, "learning_rate": 1e-05, "loss": 1.2274, "step": 1717 }, { "epoch": 0.0446109616989706, "grad_norm": 7.035175228146251, "learning_rate": 1e-05, "loss": 1.2622, "step": 1718 }, { "epoch": 0.04463692849856255, "grad_norm": 6.220330579013227, "learning_rate": 1e-05, "loss": 1.2967, "step": 1719 }, { "epoch": 0.0446628952981545, "grad_norm": 8.537030954727332, "learning_rate": 1e-05, "loss": 1.3113, "step": 1720 }, { "epoch": 0.04468886209774645, "grad_norm": 4.851262130899062, "learning_rate": 1e-05, "loss": 1.2964, "step": 1721 }, { "epoch": 0.0447148288973384, "grad_norm": 6.260762827649841, "learning_rate": 1e-05, "loss": 1.2717, "step": 1722 }, { "epoch": 0.044740795696930354, "grad_norm": 7.610477062215085, "learning_rate": 1e-05, "loss": 1.3381, "step": 1723 }, { "epoch": 0.044766762496522304, "grad_norm": 7.12826061930869, "learning_rate": 1e-05, "loss": 1.3166, "step": 1724 }, { "epoch": 0.044792729296114255, "grad_norm": 5.59232094694219, "learning_rate": 1e-05, "loss": 1.283, "step": 1725 }, { "epoch": 0.044818696095706205, "grad_norm": 5.701685391934015, "learning_rate": 1e-05, "loss": 1.196, "step": 1726 }, { "epoch": 0.044844662895298155, "grad_norm": 4.278295458120855, "learning_rate": 1e-05, "loss": 1.3376, "step": 1727 }, { "epoch": 0.044870629694890106, "grad_norm": 7.83739612273229, "learning_rate": 1e-05, "loss": 1.2236, "step": 1728 }, { "epoch": 0.044896596494482056, "grad_norm": 4.977450448771007, "learning_rate": 1e-05, "loss": 1.3144, "step": 1729 }, { "epoch": 0.04492256329407401, "grad_norm": 12.213635713368365, "learning_rate": 1e-05, "loss": 1.3284, "step": 1730 }, { "epoch": 0.04494853009366596, "grad_norm": 3.9156131680272526, "learning_rate": 1e-05, "loss": 1.2689, "step": 1731 }, { "epoch": 0.04497449689325791, "grad_norm": 5.54414700330562, "learning_rate": 1e-05, "loss": 1.2966, "step": 1732 }, { "epoch": 0.04500046369284986, "grad_norm": 6.31626613558262, "learning_rate": 1e-05, "loss": 1.233, "step": 1733 }, { "epoch": 0.04502643049244181, "grad_norm": 97.40470102233259, "learning_rate": 1e-05, "loss": 1.2751, "step": 1734 }, { "epoch": 0.04505239729203376, "grad_norm": 4.071198526255572, "learning_rate": 1e-05, "loss": 1.279, "step": 1735 }, { "epoch": 0.04507836409162571, "grad_norm": 5.295535117856197, "learning_rate": 1e-05, "loss": 1.2556, "step": 1736 }, { "epoch": 0.04510433089121766, "grad_norm": 4.778193874623676, "learning_rate": 1e-05, "loss": 1.2667, "step": 1737 }, { "epoch": 0.04513029769080961, "grad_norm": 4.2079562531289545, "learning_rate": 1e-05, "loss": 1.2674, "step": 1738 }, { "epoch": 0.04515626449040156, "grad_norm": 5.085732920292595, "learning_rate": 1e-05, "loss": 1.2806, "step": 1739 }, { "epoch": 0.04518223128999351, "grad_norm": 4.883208772982395, "learning_rate": 1e-05, "loss": 1.2387, "step": 1740 }, { "epoch": 0.045208198089585455, "grad_norm": 7.106041028012495, "learning_rate": 1e-05, "loss": 1.2669, "step": 1741 }, { "epoch": 0.045234164889177406, "grad_norm": 7.31190005718685, "learning_rate": 1e-05, "loss": 1.2782, "step": 1742 }, { "epoch": 0.045260131688769356, "grad_norm": 4.661723561617678, "learning_rate": 1e-05, "loss": 1.2991, "step": 1743 }, { "epoch": 0.04528609848836131, "grad_norm": 5.102467011303496, "learning_rate": 1e-05, "loss": 1.2571, "step": 1744 }, { "epoch": 0.04531206528795326, "grad_norm": 4.392407472924515, "learning_rate": 1e-05, "loss": 1.2885, "step": 1745 }, { "epoch": 0.04533803208754521, "grad_norm": 5.921224029915704, "learning_rate": 1e-05, "loss": 1.2506, "step": 1746 }, { "epoch": 0.04536399888713716, "grad_norm": 4.802391434037606, "learning_rate": 1e-05, "loss": 1.3194, "step": 1747 }, { "epoch": 0.04538996568672911, "grad_norm": 6.968113344131262, "learning_rate": 1e-05, "loss": 1.2678, "step": 1748 }, { "epoch": 0.04541593248632106, "grad_norm": 5.2038778957435765, "learning_rate": 1e-05, "loss": 1.245, "step": 1749 }, { "epoch": 0.04544189928591301, "grad_norm": 8.2055646483792, "learning_rate": 1e-05, "loss": 1.3378, "step": 1750 }, { "epoch": 0.04546786608550496, "grad_norm": 3.964737433149686, "learning_rate": 1e-05, "loss": 1.224, "step": 1751 }, { "epoch": 0.04549383288509691, "grad_norm": 5.784562435235587, "learning_rate": 1e-05, "loss": 1.2872, "step": 1752 }, { "epoch": 0.04551979968468886, "grad_norm": 4.953625193469228, "learning_rate": 1e-05, "loss": 1.3006, "step": 1753 }, { "epoch": 0.04554576648428081, "grad_norm": 6.041612015400269, "learning_rate": 1e-05, "loss": 1.2465, "step": 1754 }, { "epoch": 0.04557173328387276, "grad_norm": 4.820537642255311, "learning_rate": 1e-05, "loss": 1.2843, "step": 1755 }, { "epoch": 0.04559770008346471, "grad_norm": 7.602786115627001, "learning_rate": 1e-05, "loss": 1.2397, "step": 1756 }, { "epoch": 0.04562366688305666, "grad_norm": 5.025139361674266, "learning_rate": 1e-05, "loss": 1.3117, "step": 1757 }, { "epoch": 0.04564963368264861, "grad_norm": 6.568494692703662, "learning_rate": 1e-05, "loss": 1.299, "step": 1758 }, { "epoch": 0.045675600482240564, "grad_norm": 6.397223836074355, "learning_rate": 1e-05, "loss": 1.2817, "step": 1759 }, { "epoch": 0.045701567281832514, "grad_norm": 4.921705261210047, "learning_rate": 1e-05, "loss": 1.3263, "step": 1760 }, { "epoch": 0.045727534081424465, "grad_norm": 49.52906393046536, "learning_rate": 1e-05, "loss": 1.289, "step": 1761 }, { "epoch": 0.045753500881016415, "grad_norm": 18.17337966685407, "learning_rate": 1e-05, "loss": 1.2657, "step": 1762 }, { "epoch": 0.045779467680608366, "grad_norm": 5.047517812318628, "learning_rate": 1e-05, "loss": 1.2537, "step": 1763 }, { "epoch": 0.045805434480200316, "grad_norm": 204.66814794014724, "learning_rate": 1e-05, "loss": 1.2836, "step": 1764 }, { "epoch": 0.045831401279792267, "grad_norm": 4.594737044189691, "learning_rate": 1e-05, "loss": 1.2608, "step": 1765 }, { "epoch": 0.04585736807938422, "grad_norm": 4.2036338980230425, "learning_rate": 1e-05, "loss": 1.27, "step": 1766 }, { "epoch": 0.04588333487897617, "grad_norm": 5.421232245875536, "learning_rate": 1e-05, "loss": 1.3085, "step": 1767 }, { "epoch": 0.04590930167856812, "grad_norm": 5.879850941539807, "learning_rate": 1e-05, "loss": 1.3296, "step": 1768 }, { "epoch": 0.04593526847816007, "grad_norm": 5.284169129569244, "learning_rate": 1e-05, "loss": 1.3006, "step": 1769 }, { "epoch": 0.04596123527775202, "grad_norm": 5.305541923376726, "learning_rate": 1e-05, "loss": 1.31, "step": 1770 }, { "epoch": 0.04598720207734397, "grad_norm": 5.61944441360337, "learning_rate": 1e-05, "loss": 1.239, "step": 1771 }, { "epoch": 0.04601316887693592, "grad_norm": 8.563844895377883, "learning_rate": 1e-05, "loss": 1.2828, "step": 1772 }, { "epoch": 0.04603913567652787, "grad_norm": 4.971019298638021, "learning_rate": 1e-05, "loss": 1.301, "step": 1773 }, { "epoch": 0.04606510247611982, "grad_norm": 4.260758862669352, "learning_rate": 1e-05, "loss": 1.2879, "step": 1774 }, { "epoch": 0.04609106927571177, "grad_norm": 5.546360561531825, "learning_rate": 1e-05, "loss": 1.2729, "step": 1775 }, { "epoch": 0.04611703607530372, "grad_norm": 11.194927115771742, "learning_rate": 1e-05, "loss": 1.3144, "step": 1776 }, { "epoch": 0.04614300287489567, "grad_norm": 5.836035874442788, "learning_rate": 1e-05, "loss": 1.3117, "step": 1777 }, { "epoch": 0.04616896967448762, "grad_norm": 4.776319966446448, "learning_rate": 1e-05, "loss": 1.2601, "step": 1778 }, { "epoch": 0.04619493647407957, "grad_norm": 13.615019616411628, "learning_rate": 1e-05, "loss": 1.257, "step": 1779 }, { "epoch": 0.04622090327367152, "grad_norm": 5.752946844870231, "learning_rate": 1e-05, "loss": 1.2526, "step": 1780 }, { "epoch": 0.04624687007326347, "grad_norm": 4.758484340393844, "learning_rate": 1e-05, "loss": 1.3009, "step": 1781 }, { "epoch": 0.04627283687285542, "grad_norm": 4.339254783955204, "learning_rate": 1e-05, "loss": 1.2821, "step": 1782 }, { "epoch": 0.04629880367244737, "grad_norm": 4.017213737197356, "learning_rate": 1e-05, "loss": 1.2881, "step": 1783 }, { "epoch": 0.04632477047203932, "grad_norm": 5.674581423207643, "learning_rate": 1e-05, "loss": 1.3225, "step": 1784 }, { "epoch": 0.04635073727163127, "grad_norm": 6.976242621971047, "learning_rate": 1e-05, "loss": 1.2717, "step": 1785 }, { "epoch": 0.04637670407122322, "grad_norm": 4.486924566496487, "learning_rate": 1e-05, "loss": 1.2615, "step": 1786 }, { "epoch": 0.04640267087081517, "grad_norm": 8.150372605494699, "learning_rate": 1e-05, "loss": 1.2535, "step": 1787 }, { "epoch": 0.04642863767040712, "grad_norm": 5.320427936752436, "learning_rate": 1e-05, "loss": 1.2731, "step": 1788 }, { "epoch": 0.04645460446999907, "grad_norm": 8.125498594978168, "learning_rate": 1e-05, "loss": 1.2726, "step": 1789 }, { "epoch": 0.04648057126959102, "grad_norm": 7.847599133767387, "learning_rate": 1e-05, "loss": 1.2425, "step": 1790 }, { "epoch": 0.04650653806918297, "grad_norm": 5.199338892946492, "learning_rate": 1e-05, "loss": 1.2908, "step": 1791 }, { "epoch": 0.04653250486877492, "grad_norm": 9.04706010645581, "learning_rate": 1e-05, "loss": 1.2479, "step": 1792 }, { "epoch": 0.04655847166836687, "grad_norm": 124.93831681198726, "learning_rate": 1e-05, "loss": 1.2702, "step": 1793 }, { "epoch": 0.04658443846795882, "grad_norm": 4.697535440860414, "learning_rate": 1e-05, "loss": 1.227, "step": 1794 }, { "epoch": 0.046610405267550774, "grad_norm": 5.690453066629724, "learning_rate": 1e-05, "loss": 1.2747, "step": 1795 }, { "epoch": 0.046636372067142724, "grad_norm": 4.914016699160566, "learning_rate": 1e-05, "loss": 1.2579, "step": 1796 }, { "epoch": 0.046662338866734675, "grad_norm": 5.16579630914425, "learning_rate": 1e-05, "loss": 1.2663, "step": 1797 }, { "epoch": 0.046688305666326625, "grad_norm": 4.638359158867013, "learning_rate": 1e-05, "loss": 1.2529, "step": 1798 }, { "epoch": 0.046714272465918576, "grad_norm": 5.034115372255264, "learning_rate": 1e-05, "loss": 1.2965, "step": 1799 }, { "epoch": 0.046740239265510526, "grad_norm": 5.7570823506688535, "learning_rate": 1e-05, "loss": 1.2264, "step": 1800 }, { "epoch": 0.04676620606510248, "grad_norm": 11.771405141439413, "learning_rate": 1e-05, "loss": 1.3194, "step": 1801 }, { "epoch": 0.04679217286469443, "grad_norm": 5.711536297801452, "learning_rate": 1e-05, "loss": 1.2243, "step": 1802 }, { "epoch": 0.04681813966428638, "grad_norm": 4.504145725894398, "learning_rate": 1e-05, "loss": 1.281, "step": 1803 }, { "epoch": 0.04684410646387833, "grad_norm": 17.22289058796688, "learning_rate": 1e-05, "loss": 1.2737, "step": 1804 }, { "epoch": 0.04687007326347028, "grad_norm": 5.557855617387043, "learning_rate": 1e-05, "loss": 1.3096, "step": 1805 }, { "epoch": 0.04689604006306223, "grad_norm": 9.79374311358046, "learning_rate": 1e-05, "loss": 1.3068, "step": 1806 }, { "epoch": 0.04692200686265418, "grad_norm": 4.7254048088285305, "learning_rate": 1e-05, "loss": 1.2739, "step": 1807 }, { "epoch": 0.04694797366224613, "grad_norm": 5.9310971721740895, "learning_rate": 1e-05, "loss": 1.2751, "step": 1808 }, { "epoch": 0.04697394046183808, "grad_norm": 6.024679290973257, "learning_rate": 1e-05, "loss": 1.3386, "step": 1809 }, { "epoch": 0.04699990726143003, "grad_norm": 11.9871566707384, "learning_rate": 1e-05, "loss": 1.3071, "step": 1810 }, { "epoch": 0.04702587406102198, "grad_norm": 6.128457475403744, "learning_rate": 1e-05, "loss": 1.2229, "step": 1811 }, { "epoch": 0.04705184086061393, "grad_norm": 11.25616184604657, "learning_rate": 1e-05, "loss": 1.213, "step": 1812 }, { "epoch": 0.04707780766020588, "grad_norm": 7.776655564260045, "learning_rate": 1e-05, "loss": 1.294, "step": 1813 }, { "epoch": 0.04710377445979783, "grad_norm": 6.407509879924348, "learning_rate": 1e-05, "loss": 1.2663, "step": 1814 }, { "epoch": 0.04712974125938978, "grad_norm": 8.986304140711853, "learning_rate": 1e-05, "loss": 1.2766, "step": 1815 }, { "epoch": 0.047155708058981734, "grad_norm": 5.143992888972231, "learning_rate": 1e-05, "loss": 1.3125, "step": 1816 }, { "epoch": 0.047181674858573684, "grad_norm": 10.601973411113656, "learning_rate": 1e-05, "loss": 1.2638, "step": 1817 }, { "epoch": 0.04720764165816563, "grad_norm": 10.412734935074694, "learning_rate": 1e-05, "loss": 1.2632, "step": 1818 }, { "epoch": 0.04723360845775758, "grad_norm": 5.695544579569308, "learning_rate": 1e-05, "loss": 1.3273, "step": 1819 }, { "epoch": 0.04725957525734953, "grad_norm": 7.505539946120632, "learning_rate": 1e-05, "loss": 1.3043, "step": 1820 }, { "epoch": 0.04728554205694148, "grad_norm": 16.054332987265948, "learning_rate": 1e-05, "loss": 1.2895, "step": 1821 }, { "epoch": 0.04731150885653343, "grad_norm": 7.460750621432476, "learning_rate": 1e-05, "loss": 1.272, "step": 1822 }, { "epoch": 0.04733747565612538, "grad_norm": 6.99519837409696, "learning_rate": 1e-05, "loss": 1.2453, "step": 1823 }, { "epoch": 0.04736344245571733, "grad_norm": 5.553610738913908, "learning_rate": 1e-05, "loss": 1.3091, "step": 1824 }, { "epoch": 0.04738940925530928, "grad_norm": 6.370998248536985, "learning_rate": 1e-05, "loss": 1.2821, "step": 1825 }, { "epoch": 0.04741537605490123, "grad_norm": 10.909058654621877, "learning_rate": 1e-05, "loss": 1.251, "step": 1826 }, { "epoch": 0.04744134285449318, "grad_norm": 8.838521861089944, "learning_rate": 1e-05, "loss": 1.2356, "step": 1827 }, { "epoch": 0.04746730965408513, "grad_norm": 5.671735777604333, "learning_rate": 1e-05, "loss": 1.2714, "step": 1828 }, { "epoch": 0.04749327645367708, "grad_norm": 9.667104601531584, "learning_rate": 1e-05, "loss": 1.2714, "step": 1829 }, { "epoch": 0.04751924325326903, "grad_norm": 6.224651354482697, "learning_rate": 1e-05, "loss": 1.241, "step": 1830 }, { "epoch": 0.047545210052860984, "grad_norm": 7.9197597952743335, "learning_rate": 1e-05, "loss": 1.2736, "step": 1831 }, { "epoch": 0.047571176852452934, "grad_norm": 9.063242072931661, "learning_rate": 1e-05, "loss": 1.2873, "step": 1832 }, { "epoch": 0.047597143652044885, "grad_norm": 26.38246779460924, "learning_rate": 1e-05, "loss": 1.3016, "step": 1833 }, { "epoch": 0.047623110451636835, "grad_norm": 6.020588518949119, "learning_rate": 1e-05, "loss": 1.2515, "step": 1834 }, { "epoch": 0.047649077251228786, "grad_norm": 9.365844732131038, "learning_rate": 1e-05, "loss": 1.2558, "step": 1835 }, { "epoch": 0.047675044050820736, "grad_norm": 6.30715595312507, "learning_rate": 1e-05, "loss": 1.3198, "step": 1836 }, { "epoch": 0.04770101085041269, "grad_norm": 6.036441762362213, "learning_rate": 1e-05, "loss": 1.2297, "step": 1837 }, { "epoch": 0.04772697765000464, "grad_norm": 25.37793673354007, "learning_rate": 1e-05, "loss": 1.2563, "step": 1838 }, { "epoch": 0.04775294444959659, "grad_norm": 6.673528319008572, "learning_rate": 1e-05, "loss": 1.2497, "step": 1839 }, { "epoch": 0.04777891124918854, "grad_norm": 12.81490307113777, "learning_rate": 1e-05, "loss": 1.259, "step": 1840 }, { "epoch": 0.04780487804878049, "grad_norm": 6.533374203664399, "learning_rate": 1e-05, "loss": 1.2667, "step": 1841 }, { "epoch": 0.04783084484837244, "grad_norm": 5.227597714299449, "learning_rate": 1e-05, "loss": 1.273, "step": 1842 }, { "epoch": 0.04785681164796439, "grad_norm": 5.442485759753275, "learning_rate": 1e-05, "loss": 1.2969, "step": 1843 }, { "epoch": 0.04788277844755634, "grad_norm": 4.487993754051065, "learning_rate": 1e-05, "loss": 1.3241, "step": 1844 }, { "epoch": 0.04790874524714829, "grad_norm": 29.77195019737013, "learning_rate": 1e-05, "loss": 1.2436, "step": 1845 }, { "epoch": 0.04793471204674024, "grad_norm": 5.450985901577286, "learning_rate": 1e-05, "loss": 1.2334, "step": 1846 }, { "epoch": 0.04796067884633219, "grad_norm": 5.4091210652232755, "learning_rate": 1e-05, "loss": 1.3121, "step": 1847 }, { "epoch": 0.04798664564592414, "grad_norm": 5.012766371044418, "learning_rate": 1e-05, "loss": 1.2827, "step": 1848 }, { "epoch": 0.04801261244551609, "grad_norm": 4.986513067895636, "learning_rate": 1e-05, "loss": 1.2933, "step": 1849 }, { "epoch": 0.04803857924510804, "grad_norm": 6.9523310861709025, "learning_rate": 1e-05, "loss": 1.2813, "step": 1850 }, { "epoch": 0.04806454604469999, "grad_norm": 5.75233611044583, "learning_rate": 1e-05, "loss": 1.2959, "step": 1851 }, { "epoch": 0.048090512844291944, "grad_norm": 6.400184998222677, "learning_rate": 1e-05, "loss": 1.2793, "step": 1852 }, { "epoch": 0.048116479643883894, "grad_norm": 6.288724333396645, "learning_rate": 1e-05, "loss": 1.2899, "step": 1853 }, { "epoch": 0.048142446443475845, "grad_norm": 8.156633506908081, "learning_rate": 1e-05, "loss": 1.2779, "step": 1854 }, { "epoch": 0.048168413243067795, "grad_norm": 21.957389487937768, "learning_rate": 1e-05, "loss": 1.2915, "step": 1855 }, { "epoch": 0.04819438004265974, "grad_norm": 7.501821423928569, "learning_rate": 1e-05, "loss": 1.2449, "step": 1856 }, { "epoch": 0.04822034684225169, "grad_norm": 10.132804970602445, "learning_rate": 1e-05, "loss": 1.2915, "step": 1857 }, { "epoch": 0.04824631364184364, "grad_norm": 8.386159720976142, "learning_rate": 1e-05, "loss": 1.3394, "step": 1858 }, { "epoch": 0.04827228044143559, "grad_norm": 5.933303474322564, "learning_rate": 1e-05, "loss": 1.2832, "step": 1859 }, { "epoch": 0.04829824724102754, "grad_norm": 6.02826581269407, "learning_rate": 1e-05, "loss": 1.2808, "step": 1860 }, { "epoch": 0.04832421404061949, "grad_norm": 5.5473966111480575, "learning_rate": 1e-05, "loss": 1.2992, "step": 1861 }, { "epoch": 0.04835018084021144, "grad_norm": 15.058370823576883, "learning_rate": 1e-05, "loss": 1.2718, "step": 1862 }, { "epoch": 0.04837614763980339, "grad_norm": 7.9987296049415715, "learning_rate": 1e-05, "loss": 1.2713, "step": 1863 }, { "epoch": 0.04840211443939534, "grad_norm": 5.040207370548831, "learning_rate": 1e-05, "loss": 1.2685, "step": 1864 }, { "epoch": 0.04842808123898729, "grad_norm": 8.871560303258693, "learning_rate": 1e-05, "loss": 1.3142, "step": 1865 }, { "epoch": 0.048454048038579244, "grad_norm": 5.335537494729656, "learning_rate": 1e-05, "loss": 1.3385, "step": 1866 }, { "epoch": 0.048480014838171194, "grad_norm": 5.930859194905484, "learning_rate": 1e-05, "loss": 1.3099, "step": 1867 }, { "epoch": 0.048505981637763144, "grad_norm": 19.378854675777976, "learning_rate": 1e-05, "loss": 1.2721, "step": 1868 }, { "epoch": 0.048531948437355095, "grad_norm": 7.510099858459487, "learning_rate": 1e-05, "loss": 1.2844, "step": 1869 }, { "epoch": 0.048557915236947045, "grad_norm": 8.108119860116501, "learning_rate": 1e-05, "loss": 1.2596, "step": 1870 }, { "epoch": 0.048583882036538996, "grad_norm": 11.019529172943797, "learning_rate": 1e-05, "loss": 1.268, "step": 1871 }, { "epoch": 0.048609848836130946, "grad_norm": 5.517152223606601, "learning_rate": 1e-05, "loss": 1.2922, "step": 1872 }, { "epoch": 0.0486358156357229, "grad_norm": 6.498840815545552, "learning_rate": 1e-05, "loss": 1.3098, "step": 1873 }, { "epoch": 0.04866178243531485, "grad_norm": 9.526329095980355, "learning_rate": 1e-05, "loss": 1.2861, "step": 1874 }, { "epoch": 0.0486877492349068, "grad_norm": 8.134909822026302, "learning_rate": 1e-05, "loss": 1.262, "step": 1875 }, { "epoch": 0.04871371603449875, "grad_norm": 16.934608803919765, "learning_rate": 1e-05, "loss": 1.2457, "step": 1876 }, { "epoch": 0.0487396828340907, "grad_norm": 7.840316689770792, "learning_rate": 1e-05, "loss": 1.333, "step": 1877 }, { "epoch": 0.04876564963368265, "grad_norm": 8.89527318353774, "learning_rate": 1e-05, "loss": 1.2377, "step": 1878 }, { "epoch": 0.0487916164332746, "grad_norm": 23.277465348839126, "learning_rate": 1e-05, "loss": 1.2792, "step": 1879 }, { "epoch": 0.04881758323286655, "grad_norm": 47.42240269475757, "learning_rate": 1e-05, "loss": 1.2559, "step": 1880 }, { "epoch": 0.0488435500324585, "grad_norm": 5.750630219793713, "learning_rate": 1e-05, "loss": 1.3052, "step": 1881 }, { "epoch": 0.04886951683205045, "grad_norm": 7.704687985211529, "learning_rate": 1e-05, "loss": 1.2874, "step": 1882 }, { "epoch": 0.0488954836316424, "grad_norm": 15.410966688923487, "learning_rate": 1e-05, "loss": 1.3248, "step": 1883 }, { "epoch": 0.04892145043123435, "grad_norm": 5.416523427781068, "learning_rate": 1e-05, "loss": 1.2823, "step": 1884 }, { "epoch": 0.0489474172308263, "grad_norm": 7.031083168593701, "learning_rate": 1e-05, "loss": 1.3193, "step": 1885 }, { "epoch": 0.04897338403041825, "grad_norm": 5.31595221156562, "learning_rate": 1e-05, "loss": 1.2714, "step": 1886 }, { "epoch": 0.048999350830010203, "grad_norm": 5.84494349077905, "learning_rate": 1e-05, "loss": 1.3204, "step": 1887 }, { "epoch": 0.049025317629602154, "grad_norm": 7.520574450770372, "learning_rate": 1e-05, "loss": 1.2886, "step": 1888 }, { "epoch": 0.049051284429194104, "grad_norm": 7.305471319076385, "learning_rate": 1e-05, "loss": 1.2647, "step": 1889 }, { "epoch": 0.049077251228786055, "grad_norm": 5.645390492018515, "learning_rate": 1e-05, "loss": 1.2263, "step": 1890 }, { "epoch": 0.049103218028378005, "grad_norm": 5.741811145723473, "learning_rate": 1e-05, "loss": 1.296, "step": 1891 }, { "epoch": 0.049129184827969956, "grad_norm": 5.444795646931189, "learning_rate": 1e-05, "loss": 1.2392, "step": 1892 }, { "epoch": 0.049155151627561906, "grad_norm": 9.381855810276782, "learning_rate": 1e-05, "loss": 1.2825, "step": 1893 }, { "epoch": 0.04918111842715385, "grad_norm": 5.961162117941997, "learning_rate": 1e-05, "loss": 1.2588, "step": 1894 }, { "epoch": 0.0492070852267458, "grad_norm": 10.247577660129764, "learning_rate": 1e-05, "loss": 1.3389, "step": 1895 }, { "epoch": 0.04923305202633775, "grad_norm": 7.182911345204422, "learning_rate": 1e-05, "loss": 1.3139, "step": 1896 }, { "epoch": 0.0492590188259297, "grad_norm": 5.600676611488091, "learning_rate": 1e-05, "loss": 1.2358, "step": 1897 }, { "epoch": 0.04928498562552165, "grad_norm": 10.935695651950535, "learning_rate": 1e-05, "loss": 1.2852, "step": 1898 }, { "epoch": 0.0493109524251136, "grad_norm": 16.173164765899156, "learning_rate": 1e-05, "loss": 1.2505, "step": 1899 }, { "epoch": 0.04933691922470555, "grad_norm": 6.5173668305576244, "learning_rate": 1e-05, "loss": 1.2505, "step": 1900 }, { "epoch": 0.0493628860242975, "grad_norm": 6.450296959953242, "learning_rate": 1e-05, "loss": 1.2519, "step": 1901 }, { "epoch": 0.049388852823889454, "grad_norm": 7.104233846784936, "learning_rate": 1e-05, "loss": 1.269, "step": 1902 }, { "epoch": 0.049414819623481404, "grad_norm": 8.57153333872391, "learning_rate": 1e-05, "loss": 1.3144, "step": 1903 }, { "epoch": 0.049440786423073355, "grad_norm": 6.667142024576866, "learning_rate": 1e-05, "loss": 1.3283, "step": 1904 }, { "epoch": 0.049466753222665305, "grad_norm": 7.236017621543435, "learning_rate": 1e-05, "loss": 1.2739, "step": 1905 }, { "epoch": 0.049492720022257256, "grad_norm": 7.5192151922214325, "learning_rate": 1e-05, "loss": 1.2498, "step": 1906 }, { "epoch": 0.049518686821849206, "grad_norm": 8.165164212446124, "learning_rate": 1e-05, "loss": 1.2889, "step": 1907 }, { "epoch": 0.049544653621441156, "grad_norm": 5.052787789326772, "learning_rate": 1e-05, "loss": 1.2799, "step": 1908 }, { "epoch": 0.04957062042103311, "grad_norm": 5.886304976230281, "learning_rate": 1e-05, "loss": 1.3188, "step": 1909 }, { "epoch": 0.04959658722062506, "grad_norm": 5.355468215775411, "learning_rate": 1e-05, "loss": 1.2952, "step": 1910 }, { "epoch": 0.04962255402021701, "grad_norm": 9.477256856389019, "learning_rate": 1e-05, "loss": 1.2978, "step": 1911 }, { "epoch": 0.04964852081980896, "grad_norm": 24.494144907826954, "learning_rate": 1e-05, "loss": 1.2791, "step": 1912 }, { "epoch": 0.04967448761940091, "grad_norm": 9.894700884353217, "learning_rate": 1e-05, "loss": 1.2989, "step": 1913 }, { "epoch": 0.04970045441899286, "grad_norm": 6.699718116410232, "learning_rate": 1e-05, "loss": 1.2645, "step": 1914 }, { "epoch": 0.04972642121858481, "grad_norm": 7.481172963752022, "learning_rate": 1e-05, "loss": 1.2843, "step": 1915 }, { "epoch": 0.04975238801817676, "grad_norm": 7.556064105527804, "learning_rate": 1e-05, "loss": 1.2682, "step": 1916 }, { "epoch": 0.04977835481776871, "grad_norm": 5.629437878802716, "learning_rate": 1e-05, "loss": 1.2963, "step": 1917 }, { "epoch": 0.04980432161736066, "grad_norm": 13.11422606999891, "learning_rate": 1e-05, "loss": 1.2631, "step": 1918 }, { "epoch": 0.04983028841695261, "grad_norm": 6.860488001868329, "learning_rate": 1e-05, "loss": 1.3416, "step": 1919 }, { "epoch": 0.04985625521654456, "grad_norm": 9.015231066052406, "learning_rate": 1e-05, "loss": 1.2571, "step": 1920 }, { "epoch": 0.04988222201613651, "grad_norm": 8.789088107601584, "learning_rate": 1e-05, "loss": 1.3522, "step": 1921 }, { "epoch": 0.04990818881572846, "grad_norm": 6.331732129251189, "learning_rate": 1e-05, "loss": 1.2859, "step": 1922 }, { "epoch": 0.049934155615320414, "grad_norm": 16.76210427268717, "learning_rate": 1e-05, "loss": 1.2846, "step": 1923 }, { "epoch": 0.049960122414912364, "grad_norm": 8.683869186336823, "learning_rate": 1e-05, "loss": 1.2787, "step": 1924 }, { "epoch": 0.049986089214504315, "grad_norm": 7.674416426052781, "learning_rate": 1e-05, "loss": 1.317, "step": 1925 }, { "epoch": 0.050012056014096265, "grad_norm": 35.80945599519595, "learning_rate": 1e-05, "loss": 1.3138, "step": 1926 }, { "epoch": 0.050038022813688215, "grad_norm": 5.267645835794885, "learning_rate": 1e-05, "loss": 1.3132, "step": 1927 }, { "epoch": 0.050063989613280166, "grad_norm": 7.4721245584010285, "learning_rate": 1e-05, "loss": 1.2943, "step": 1928 }, { "epoch": 0.050089956412872116, "grad_norm": 41.92557871265017, "learning_rate": 1e-05, "loss": 1.3037, "step": 1929 }, { "epoch": 0.05011592321246407, "grad_norm": 6.851926891767524, "learning_rate": 1e-05, "loss": 1.2423, "step": 1930 }, { "epoch": 0.05014189001205602, "grad_norm": 22.15501996541571, "learning_rate": 1e-05, "loss": 1.2962, "step": 1931 }, { "epoch": 0.05016785681164797, "grad_norm": 6.080687622538312, "learning_rate": 1e-05, "loss": 1.2635, "step": 1932 }, { "epoch": 0.05019382361123991, "grad_norm": 5.450659251805865, "learning_rate": 1e-05, "loss": 1.3056, "step": 1933 }, { "epoch": 0.05021979041083186, "grad_norm": 7.361712884197228, "learning_rate": 1e-05, "loss": 1.2682, "step": 1934 }, { "epoch": 0.05024575721042381, "grad_norm": 7.013675138365991, "learning_rate": 1e-05, "loss": 1.284, "step": 1935 }, { "epoch": 0.05027172401001576, "grad_norm": 5.417715098322971, "learning_rate": 1e-05, "loss": 1.278, "step": 1936 }, { "epoch": 0.05029769080960771, "grad_norm": 6.167276335125485, "learning_rate": 1e-05, "loss": 1.2989, "step": 1937 }, { "epoch": 0.050323657609199664, "grad_norm": 7.897196173165571, "learning_rate": 1e-05, "loss": 1.3121, "step": 1938 }, { "epoch": 0.050349624408791614, "grad_norm": 7.286270179164323, "learning_rate": 1e-05, "loss": 1.2444, "step": 1939 }, { "epoch": 0.050375591208383565, "grad_norm": 7.1853008057044505, "learning_rate": 1e-05, "loss": 1.2674, "step": 1940 }, { "epoch": 0.050401558007975515, "grad_norm": 6.271557669982008, "learning_rate": 1e-05, "loss": 1.3054, "step": 1941 }, { "epoch": 0.050427524807567466, "grad_norm": 5.917300588828061, "learning_rate": 1e-05, "loss": 1.2937, "step": 1942 }, { "epoch": 0.050453491607159416, "grad_norm": 7.138916078244428, "learning_rate": 1e-05, "loss": 1.2365, "step": 1943 }, { "epoch": 0.05047945840675137, "grad_norm": 7.104732934602221, "learning_rate": 1e-05, "loss": 1.3306, "step": 1944 }, { "epoch": 0.05050542520634332, "grad_norm": 5.576205225108238, "learning_rate": 1e-05, "loss": 1.3042, "step": 1945 }, { "epoch": 0.05053139200593527, "grad_norm": 8.66255833990095, "learning_rate": 1e-05, "loss": 1.2376, "step": 1946 }, { "epoch": 0.05055735880552722, "grad_norm": 18.0399261175364, "learning_rate": 1e-05, "loss": 1.2949, "step": 1947 }, { "epoch": 0.05058332560511917, "grad_norm": 7.159300840968631, "learning_rate": 1e-05, "loss": 1.2639, "step": 1948 }, { "epoch": 0.05060929240471112, "grad_norm": 17.2515010457084, "learning_rate": 1e-05, "loss": 1.2841, "step": 1949 }, { "epoch": 0.05063525920430307, "grad_norm": 8.019264391413625, "learning_rate": 1e-05, "loss": 1.3107, "step": 1950 }, { "epoch": 0.05066122600389502, "grad_norm": 7.399443126035868, "learning_rate": 1e-05, "loss": 1.2958, "step": 1951 }, { "epoch": 0.05068719280348697, "grad_norm": 7.1413207935928735, "learning_rate": 1e-05, "loss": 1.2988, "step": 1952 }, { "epoch": 0.05071315960307892, "grad_norm": 10.667791386753354, "learning_rate": 1e-05, "loss": 1.293, "step": 1953 }, { "epoch": 0.05073912640267087, "grad_norm": 18.724929528301136, "learning_rate": 1e-05, "loss": 1.2211, "step": 1954 }, { "epoch": 0.05076509320226282, "grad_norm": 7.579356677452445, "learning_rate": 1e-05, "loss": 1.3303, "step": 1955 }, { "epoch": 0.05079106000185477, "grad_norm": 10.801832990157996, "learning_rate": 1e-05, "loss": 1.2363, "step": 1956 }, { "epoch": 0.05081702680144672, "grad_norm": 8.379782606414185, "learning_rate": 1e-05, "loss": 1.2587, "step": 1957 }, { "epoch": 0.05084299360103867, "grad_norm": 5.610315270634359, "learning_rate": 1e-05, "loss": 1.3072, "step": 1958 }, { "epoch": 0.050868960400630624, "grad_norm": 7.233287886141325, "learning_rate": 1e-05, "loss": 1.2657, "step": 1959 }, { "epoch": 0.050894927200222574, "grad_norm": 8.832279166412116, "learning_rate": 1e-05, "loss": 1.2828, "step": 1960 }, { "epoch": 0.050920893999814525, "grad_norm": 6.818479940659199, "learning_rate": 1e-05, "loss": 1.2806, "step": 1961 }, { "epoch": 0.050946860799406475, "grad_norm": 5.51974342192168, "learning_rate": 1e-05, "loss": 1.296, "step": 1962 }, { "epoch": 0.050972827598998426, "grad_norm": 7.283832665035375, "learning_rate": 1e-05, "loss": 1.2683, "step": 1963 }, { "epoch": 0.050998794398590376, "grad_norm": 9.847875499592222, "learning_rate": 1e-05, "loss": 1.307, "step": 1964 }, { "epoch": 0.051024761198182327, "grad_norm": 7.043855075001682, "learning_rate": 1e-05, "loss": 1.2745, "step": 1965 }, { "epoch": 0.05105072799777428, "grad_norm": 7.726839467514404, "learning_rate": 1e-05, "loss": 1.2952, "step": 1966 }, { "epoch": 0.05107669479736623, "grad_norm": 8.155228474104435, "learning_rate": 1e-05, "loss": 1.2697, "step": 1967 }, { "epoch": 0.05110266159695818, "grad_norm": 5.484086510007219, "learning_rate": 1e-05, "loss": 1.2675, "step": 1968 }, { "epoch": 0.05112862839655013, "grad_norm": 7.460525133810408, "learning_rate": 1e-05, "loss": 1.2511, "step": 1969 }, { "epoch": 0.05115459519614208, "grad_norm": 6.515627049141615, "learning_rate": 1e-05, "loss": 1.2067, "step": 1970 }, { "epoch": 0.05118056199573402, "grad_norm": 8.996293894321584, "learning_rate": 1e-05, "loss": 1.2794, "step": 1971 }, { "epoch": 0.05120652879532597, "grad_norm": 5.22397137605732, "learning_rate": 1e-05, "loss": 1.2795, "step": 1972 }, { "epoch": 0.05123249559491792, "grad_norm": 7.822681841688889, "learning_rate": 1e-05, "loss": 1.2888, "step": 1973 }, { "epoch": 0.051258462394509874, "grad_norm": 8.473564327922082, "learning_rate": 1e-05, "loss": 1.2845, "step": 1974 }, { "epoch": 0.051284429194101824, "grad_norm": 5.960116708624079, "learning_rate": 1e-05, "loss": 1.271, "step": 1975 }, { "epoch": 0.051310395993693775, "grad_norm": 15.030822620938263, "learning_rate": 1e-05, "loss": 1.2761, "step": 1976 }, { "epoch": 0.051336362793285725, "grad_norm": 7.275435979304941, "learning_rate": 1e-05, "loss": 1.31, "step": 1977 }, { "epoch": 0.051362329592877676, "grad_norm": 6.196647278972719, "learning_rate": 1e-05, "loss": 1.2901, "step": 1978 }, { "epoch": 0.051388296392469626, "grad_norm": 12.669756729933813, "learning_rate": 1e-05, "loss": 1.268, "step": 1979 }, { "epoch": 0.05141426319206158, "grad_norm": 8.004492928565622, "learning_rate": 1e-05, "loss": 1.2793, "step": 1980 }, { "epoch": 0.05144022999165353, "grad_norm": 6.340285684785605, "learning_rate": 1e-05, "loss": 1.2903, "step": 1981 }, { "epoch": 0.05146619679124548, "grad_norm": 5.165443870475055, "learning_rate": 1e-05, "loss": 1.3074, "step": 1982 }, { "epoch": 0.05149216359083743, "grad_norm": 6.186666104930545, "learning_rate": 1e-05, "loss": 1.3231, "step": 1983 }, { "epoch": 0.05151813039042938, "grad_norm": 5.85508338927477, "learning_rate": 1e-05, "loss": 1.2739, "step": 1984 }, { "epoch": 0.05154409719002133, "grad_norm": 30.165300819570096, "learning_rate": 1e-05, "loss": 1.2715, "step": 1985 }, { "epoch": 0.05157006398961328, "grad_norm": 9.737936136960416, "learning_rate": 1e-05, "loss": 1.3253, "step": 1986 }, { "epoch": 0.05159603078920523, "grad_norm": 9.045412989730494, "learning_rate": 1e-05, "loss": 1.2976, "step": 1987 }, { "epoch": 0.05162199758879718, "grad_norm": 8.579616396138164, "learning_rate": 1e-05, "loss": 1.2561, "step": 1988 }, { "epoch": 0.05164796438838913, "grad_norm": 12.90841989079796, "learning_rate": 1e-05, "loss": 1.3233, "step": 1989 }, { "epoch": 0.05167393118798108, "grad_norm": 4.557910439763412, "learning_rate": 1e-05, "loss": 1.2706, "step": 1990 }, { "epoch": 0.05169989798757303, "grad_norm": 10.965495004261165, "learning_rate": 1e-05, "loss": 1.3033, "step": 1991 }, { "epoch": 0.05172586478716498, "grad_norm": 7.669509056722117, "learning_rate": 1e-05, "loss": 1.3076, "step": 1992 }, { "epoch": 0.05175183158675693, "grad_norm": 5.631973690903509, "learning_rate": 1e-05, "loss": 1.3025, "step": 1993 }, { "epoch": 0.05177779838634888, "grad_norm": 5.176555624559179, "learning_rate": 1e-05, "loss": 1.3025, "step": 1994 }, { "epoch": 0.051803765185940834, "grad_norm": 23.682636013230375, "learning_rate": 1e-05, "loss": 1.2575, "step": 1995 }, { "epoch": 0.051829731985532784, "grad_norm": 5.681461356315868, "learning_rate": 1e-05, "loss": 1.2854, "step": 1996 }, { "epoch": 0.051855698785124735, "grad_norm": 8.652945538468344, "learning_rate": 1e-05, "loss": 1.2497, "step": 1997 }, { "epoch": 0.051881665584716685, "grad_norm": 6.179629473890025, "learning_rate": 1e-05, "loss": 1.3119, "step": 1998 }, { "epoch": 0.051907632384308636, "grad_norm": 8.440798870444022, "learning_rate": 1e-05, "loss": 1.2387, "step": 1999 }, { "epoch": 0.051933599183900586, "grad_norm": 6.495810112005312, "learning_rate": 1e-05, "loss": 1.2576, "step": 2000 }, { "epoch": 0.05195956598349254, "grad_norm": 7.046141186588436, "learning_rate": 1e-05, "loss": 1.2812, "step": 2001 }, { "epoch": 0.05198553278308449, "grad_norm": 7.456441912935188, "learning_rate": 1e-05, "loss": 1.2388, "step": 2002 }, { "epoch": 0.05201149958267644, "grad_norm": 4.34248783792869, "learning_rate": 1e-05, "loss": 1.2659, "step": 2003 }, { "epoch": 0.05203746638226839, "grad_norm": 8.520546707020825, "learning_rate": 1e-05, "loss": 1.2979, "step": 2004 }, { "epoch": 0.05206343318186034, "grad_norm": 14.374311613102936, "learning_rate": 1e-05, "loss": 1.2371, "step": 2005 }, { "epoch": 0.05208939998145229, "grad_norm": 5.714403512285487, "learning_rate": 1e-05, "loss": 1.2518, "step": 2006 }, { "epoch": 0.05211536678104424, "grad_norm": 4.628496240412189, "learning_rate": 1e-05, "loss": 1.3026, "step": 2007 }, { "epoch": 0.05214133358063619, "grad_norm": 7.541762237224286, "learning_rate": 1e-05, "loss": 1.3003, "step": 2008 }, { "epoch": 0.052167300380228133, "grad_norm": 18.71537982234214, "learning_rate": 1e-05, "loss": 1.2937, "step": 2009 }, { "epoch": 0.052193267179820084, "grad_norm": 10.948203119703843, "learning_rate": 1e-05, "loss": 1.2741, "step": 2010 }, { "epoch": 0.052219233979412034, "grad_norm": 7.997555836193165, "learning_rate": 1e-05, "loss": 1.273, "step": 2011 }, { "epoch": 0.052245200779003985, "grad_norm": 10.929727637412899, "learning_rate": 1e-05, "loss": 1.2791, "step": 2012 }, { "epoch": 0.052271167578595935, "grad_norm": 4.554758061538012, "learning_rate": 1e-05, "loss": 1.2714, "step": 2013 }, { "epoch": 0.052297134378187886, "grad_norm": 12.979098610166343, "learning_rate": 1e-05, "loss": 1.2251, "step": 2014 }, { "epoch": 0.052323101177779836, "grad_norm": 9.99226194447933, "learning_rate": 1e-05, "loss": 1.2636, "step": 2015 }, { "epoch": 0.05234906797737179, "grad_norm": 4.777096808251354, "learning_rate": 1e-05, "loss": 1.2594, "step": 2016 }, { "epoch": 0.05237503477696374, "grad_norm": 10.711208170358281, "learning_rate": 1e-05, "loss": 1.2861, "step": 2017 }, { "epoch": 0.05240100157655569, "grad_norm": 7.09856933363848, "learning_rate": 1e-05, "loss": 1.2603, "step": 2018 }, { "epoch": 0.05242696837614764, "grad_norm": 5.477157669269089, "learning_rate": 1e-05, "loss": 1.2434, "step": 2019 }, { "epoch": 0.05245293517573959, "grad_norm": 6.302235936190437, "learning_rate": 1e-05, "loss": 1.3104, "step": 2020 }, { "epoch": 0.05247890197533154, "grad_norm": 8.077370820542944, "learning_rate": 1e-05, "loss": 1.2296, "step": 2021 }, { "epoch": 0.05250486877492349, "grad_norm": 7.6612368166339095, "learning_rate": 1e-05, "loss": 1.3311, "step": 2022 }, { "epoch": 0.05253083557451544, "grad_norm": 4.752078956603103, "learning_rate": 1e-05, "loss": 1.3173, "step": 2023 }, { "epoch": 0.05255680237410739, "grad_norm": 13.29327178797987, "learning_rate": 1e-05, "loss": 1.2686, "step": 2024 }, { "epoch": 0.05258276917369934, "grad_norm": 8.032413621017133, "learning_rate": 1e-05, "loss": 1.2665, "step": 2025 }, { "epoch": 0.05260873597329129, "grad_norm": 6.623783503712416, "learning_rate": 1e-05, "loss": 1.3128, "step": 2026 }, { "epoch": 0.05263470277288324, "grad_norm": 5.490555717539195, "learning_rate": 1e-05, "loss": 1.2268, "step": 2027 }, { "epoch": 0.05266066957247519, "grad_norm": 5.759088382636912, "learning_rate": 1e-05, "loss": 1.3001, "step": 2028 }, { "epoch": 0.05268663637206714, "grad_norm": 22.238250716319282, "learning_rate": 1e-05, "loss": 1.2768, "step": 2029 }, { "epoch": 0.05271260317165909, "grad_norm": 6.731680417268364, "learning_rate": 1e-05, "loss": 1.2657, "step": 2030 }, { "epoch": 0.052738569971251044, "grad_norm": 6.079375699767681, "learning_rate": 1e-05, "loss": 1.2514, "step": 2031 }, { "epoch": 0.052764536770842994, "grad_norm": 5.470313497203833, "learning_rate": 1e-05, "loss": 1.2676, "step": 2032 }, { "epoch": 0.052790503570434945, "grad_norm": 6.055924457819661, "learning_rate": 1e-05, "loss": 1.3347, "step": 2033 }, { "epoch": 0.052816470370026895, "grad_norm": 5.400938567082161, "learning_rate": 1e-05, "loss": 1.2915, "step": 2034 }, { "epoch": 0.052842437169618846, "grad_norm": 5.827353582387971, "learning_rate": 1e-05, "loss": 1.2935, "step": 2035 }, { "epoch": 0.052868403969210796, "grad_norm": 26.79458467799802, "learning_rate": 1e-05, "loss": 1.3218, "step": 2036 }, { "epoch": 0.05289437076880275, "grad_norm": 4.962234543779694, "learning_rate": 1e-05, "loss": 1.2613, "step": 2037 }, { "epoch": 0.0529203375683947, "grad_norm": 4.818645566211883, "learning_rate": 1e-05, "loss": 1.3221, "step": 2038 }, { "epoch": 0.05294630436798665, "grad_norm": 8.225762570171828, "learning_rate": 1e-05, "loss": 1.2549, "step": 2039 }, { "epoch": 0.0529722711675786, "grad_norm": 7.9022978905305505, "learning_rate": 1e-05, "loss": 1.3048, "step": 2040 }, { "epoch": 0.05299823796717055, "grad_norm": 4.242486932629864, "learning_rate": 1e-05, "loss": 1.2713, "step": 2041 }, { "epoch": 0.0530242047667625, "grad_norm": 4.380756405875824, "learning_rate": 1e-05, "loss": 1.29, "step": 2042 }, { "epoch": 0.05305017156635445, "grad_norm": 8.17119996763834, "learning_rate": 1e-05, "loss": 1.2984, "step": 2043 }, { "epoch": 0.0530761383659464, "grad_norm": 7.342557509979307, "learning_rate": 1e-05, "loss": 1.264, "step": 2044 }, { "epoch": 0.05310210516553835, "grad_norm": 9.825722502333917, "learning_rate": 1e-05, "loss": 1.3002, "step": 2045 }, { "epoch": 0.0531280719651303, "grad_norm": 13.629743677697192, "learning_rate": 1e-05, "loss": 1.3062, "step": 2046 }, { "epoch": 0.053154038764722245, "grad_norm": 5.649298173643902, "learning_rate": 1e-05, "loss": 1.2867, "step": 2047 }, { "epoch": 0.053180005564314195, "grad_norm": 5.946873062782954, "learning_rate": 1e-05, "loss": 1.2522, "step": 2048 }, { "epoch": 0.053205972363906145, "grad_norm": 7.704384226913788, "learning_rate": 1e-05, "loss": 1.2725, "step": 2049 }, { "epoch": 0.053231939163498096, "grad_norm": 22.801952024186026, "learning_rate": 1e-05, "loss": 1.2824, "step": 2050 }, { "epoch": 0.053257905963090046, "grad_norm": 22.67654801944346, "learning_rate": 1e-05, "loss": 1.2636, "step": 2051 }, { "epoch": 0.053283872762682, "grad_norm": 4.729420099175723, "learning_rate": 1e-05, "loss": 1.2399, "step": 2052 }, { "epoch": 0.05330983956227395, "grad_norm": 5.689237559746514, "learning_rate": 1e-05, "loss": 1.2896, "step": 2053 }, { "epoch": 0.0533358063618659, "grad_norm": 5.507604543669438, "learning_rate": 1e-05, "loss": 1.22, "step": 2054 }, { "epoch": 0.05336177316145785, "grad_norm": 6.31127917882317, "learning_rate": 1e-05, "loss": 1.2642, "step": 2055 }, { "epoch": 0.0533877399610498, "grad_norm": 5.66283829712148, "learning_rate": 1e-05, "loss": 1.279, "step": 2056 }, { "epoch": 0.05341370676064175, "grad_norm": 7.938103119772173, "learning_rate": 1e-05, "loss": 1.3028, "step": 2057 }, { "epoch": 0.0534396735602337, "grad_norm": 13.230785328918381, "learning_rate": 1e-05, "loss": 1.323, "step": 2058 }, { "epoch": 0.05346564035982565, "grad_norm": 5.974034391770983, "learning_rate": 1e-05, "loss": 1.2399, "step": 2059 }, { "epoch": 0.0534916071594176, "grad_norm": 9.088733665966833, "learning_rate": 1e-05, "loss": 1.2475, "step": 2060 }, { "epoch": 0.05351757395900955, "grad_norm": 13.003263210670488, "learning_rate": 1e-05, "loss": 1.3097, "step": 2061 }, { "epoch": 0.0535435407586015, "grad_norm": 10.033568307484483, "learning_rate": 1e-05, "loss": 1.2909, "step": 2062 }, { "epoch": 0.05356950755819345, "grad_norm": 5.878188221828239, "learning_rate": 1e-05, "loss": 1.3137, "step": 2063 }, { "epoch": 0.0535954743577854, "grad_norm": 4.478091632854469, "learning_rate": 1e-05, "loss": 1.286, "step": 2064 }, { "epoch": 0.05362144115737735, "grad_norm": 6.691927291082938, "learning_rate": 1e-05, "loss": 1.2652, "step": 2065 }, { "epoch": 0.053647407956969304, "grad_norm": 5.414715792170938, "learning_rate": 1e-05, "loss": 1.2798, "step": 2066 }, { "epoch": 0.053673374756561254, "grad_norm": 7.73023459844664, "learning_rate": 1e-05, "loss": 1.2454, "step": 2067 }, { "epoch": 0.053699341556153204, "grad_norm": 4.404931919641302, "learning_rate": 1e-05, "loss": 1.2073, "step": 2068 }, { "epoch": 0.053725308355745155, "grad_norm": 15.537032158238356, "learning_rate": 1e-05, "loss": 1.2803, "step": 2069 }, { "epoch": 0.053751275155337105, "grad_norm": 7.3226790048190145, "learning_rate": 1e-05, "loss": 1.3282, "step": 2070 }, { "epoch": 0.053777241954929056, "grad_norm": 5.24153144471979, "learning_rate": 1e-05, "loss": 1.2942, "step": 2071 }, { "epoch": 0.053803208754521006, "grad_norm": 4.78247453554177, "learning_rate": 1e-05, "loss": 1.3086, "step": 2072 }, { "epoch": 0.05382917555411296, "grad_norm": 6.1759103451482025, "learning_rate": 1e-05, "loss": 1.2815, "step": 2073 }, { "epoch": 0.05385514235370491, "grad_norm": 12.911274890076092, "learning_rate": 1e-05, "loss": 1.2536, "step": 2074 }, { "epoch": 0.05388110915329686, "grad_norm": 10.679697858558436, "learning_rate": 1e-05, "loss": 1.2737, "step": 2075 }, { "epoch": 0.05390707595288881, "grad_norm": 6.272036651368, "learning_rate": 1e-05, "loss": 1.2901, "step": 2076 }, { "epoch": 0.05393304275248076, "grad_norm": 7.375676528159764, "learning_rate": 1e-05, "loss": 1.3133, "step": 2077 }, { "epoch": 0.05395900955207271, "grad_norm": 5.128315853121752, "learning_rate": 1e-05, "loss": 1.2867, "step": 2078 }, { "epoch": 0.05398497635166466, "grad_norm": 7.685301427545087, "learning_rate": 1e-05, "loss": 1.3115, "step": 2079 }, { "epoch": 0.05401094315125661, "grad_norm": 4.5336594062296225, "learning_rate": 1e-05, "loss": 1.2903, "step": 2080 }, { "epoch": 0.05403690995084856, "grad_norm": 25.991695545013748, "learning_rate": 1e-05, "loss": 1.2875, "step": 2081 }, { "epoch": 0.05406287675044051, "grad_norm": 5.064863618631977, "learning_rate": 1e-05, "loss": 1.2662, "step": 2082 }, { "epoch": 0.05408884355003246, "grad_norm": 15.072720554273756, "learning_rate": 1e-05, "loss": 1.3109, "step": 2083 }, { "epoch": 0.05411481034962441, "grad_norm": 7.18625158995025, "learning_rate": 1e-05, "loss": 1.2089, "step": 2084 }, { "epoch": 0.05414077714921636, "grad_norm": 5.055108596737036, "learning_rate": 1e-05, "loss": 1.2243, "step": 2085 }, { "epoch": 0.054166743948808306, "grad_norm": 5.966859208451006, "learning_rate": 1e-05, "loss": 1.2946, "step": 2086 }, { "epoch": 0.05419271074840026, "grad_norm": 5.1205149308188025, "learning_rate": 1e-05, "loss": 1.2797, "step": 2087 }, { "epoch": 0.05421867754799221, "grad_norm": 11.006872977670907, "learning_rate": 1e-05, "loss": 1.2464, "step": 2088 }, { "epoch": 0.05424464434758416, "grad_norm": 5.368394385607203, "learning_rate": 1e-05, "loss": 1.2736, "step": 2089 }, { "epoch": 0.05427061114717611, "grad_norm": 9.175768777928944, "learning_rate": 1e-05, "loss": 1.2752, "step": 2090 }, { "epoch": 0.05429657794676806, "grad_norm": 6.976119861465484, "learning_rate": 1e-05, "loss": 1.2977, "step": 2091 }, { "epoch": 0.05432254474636001, "grad_norm": 6.2409437561385825, "learning_rate": 1e-05, "loss": 1.2971, "step": 2092 }, { "epoch": 0.05434851154595196, "grad_norm": 4.760066959326874, "learning_rate": 1e-05, "loss": 1.2807, "step": 2093 }, { "epoch": 0.05437447834554391, "grad_norm": 5.799261624921496, "learning_rate": 1e-05, "loss": 1.2916, "step": 2094 }, { "epoch": 0.05440044514513586, "grad_norm": 6.47698008775146, "learning_rate": 1e-05, "loss": 1.2256, "step": 2095 }, { "epoch": 0.05442641194472781, "grad_norm": 4.166591084112651, "learning_rate": 1e-05, "loss": 1.3049, "step": 2096 }, { "epoch": 0.05445237874431976, "grad_norm": 6.544880492632528, "learning_rate": 1e-05, "loss": 1.2374, "step": 2097 }, { "epoch": 0.05447834554391171, "grad_norm": 6.999009198503899, "learning_rate": 1e-05, "loss": 1.2905, "step": 2098 }, { "epoch": 0.05450431234350366, "grad_norm": 13.416292426218963, "learning_rate": 1e-05, "loss": 1.2269, "step": 2099 }, { "epoch": 0.05453027914309561, "grad_norm": 6.49773161994507, "learning_rate": 1e-05, "loss": 1.2274, "step": 2100 }, { "epoch": 0.05455624594268756, "grad_norm": 7.09213682141028, "learning_rate": 1e-05, "loss": 1.3246, "step": 2101 }, { "epoch": 0.054582212742279514, "grad_norm": 4.555797094956069, "learning_rate": 1e-05, "loss": 1.2481, "step": 2102 }, { "epoch": 0.054608179541871464, "grad_norm": 20.566335395770636, "learning_rate": 1e-05, "loss": 1.2446, "step": 2103 }, { "epoch": 0.054634146341463415, "grad_norm": 4.633022258570866, "learning_rate": 1e-05, "loss": 1.3527, "step": 2104 }, { "epoch": 0.054660113141055365, "grad_norm": 4.598408315853067, "learning_rate": 1e-05, "loss": 1.2975, "step": 2105 }, { "epoch": 0.054686079940647316, "grad_norm": 6.220679209137701, "learning_rate": 1e-05, "loss": 1.2744, "step": 2106 }, { "epoch": 0.054712046740239266, "grad_norm": 4.389019112459098, "learning_rate": 1e-05, "loss": 1.2431, "step": 2107 }, { "epoch": 0.054738013539831216, "grad_norm": 5.0000915518962, "learning_rate": 1e-05, "loss": 1.2645, "step": 2108 }, { "epoch": 0.05476398033942317, "grad_norm": 13.169168612829171, "learning_rate": 1e-05, "loss": 1.2784, "step": 2109 }, { "epoch": 0.05478994713901512, "grad_norm": 5.106315606768518, "learning_rate": 1e-05, "loss": 1.2647, "step": 2110 }, { "epoch": 0.05481591393860707, "grad_norm": 10.188652862658621, "learning_rate": 1e-05, "loss": 1.2566, "step": 2111 }, { "epoch": 0.05484188073819902, "grad_norm": 6.262976031761104, "learning_rate": 1e-05, "loss": 1.2669, "step": 2112 }, { "epoch": 0.05486784753779097, "grad_norm": 6.569535326267466, "learning_rate": 1e-05, "loss": 1.2529, "step": 2113 }, { "epoch": 0.05489381433738292, "grad_norm": 4.998011575136676, "learning_rate": 1e-05, "loss": 1.2523, "step": 2114 }, { "epoch": 0.05491978113697487, "grad_norm": 4.542393734097836, "learning_rate": 1e-05, "loss": 1.2817, "step": 2115 }, { "epoch": 0.05494574793656682, "grad_norm": 6.673422854976175, "learning_rate": 1e-05, "loss": 1.3104, "step": 2116 }, { "epoch": 0.05497171473615877, "grad_norm": 4.23693747598605, "learning_rate": 1e-05, "loss": 1.3252, "step": 2117 }, { "epoch": 0.05499768153575072, "grad_norm": 6.227500394976591, "learning_rate": 1e-05, "loss": 1.2458, "step": 2118 }, { "epoch": 0.05502364833534267, "grad_norm": 4.115792372883921, "learning_rate": 1e-05, "loss": 1.275, "step": 2119 }, { "epoch": 0.05504961513493462, "grad_norm": 6.5387740168406605, "learning_rate": 1e-05, "loss": 1.2348, "step": 2120 }, { "epoch": 0.05507558193452657, "grad_norm": 5.744558662215298, "learning_rate": 1e-05, "loss": 1.2587, "step": 2121 }, { "epoch": 0.05510154873411852, "grad_norm": 5.862946502665297, "learning_rate": 1e-05, "loss": 1.2482, "step": 2122 }, { "epoch": 0.055127515533710474, "grad_norm": 9.37712866458531, "learning_rate": 1e-05, "loss": 1.2675, "step": 2123 }, { "epoch": 0.05515348233330242, "grad_norm": 5.385608229436723, "learning_rate": 1e-05, "loss": 1.3648, "step": 2124 }, { "epoch": 0.05517944913289437, "grad_norm": 5.969059321991458, "learning_rate": 1e-05, "loss": 1.3041, "step": 2125 }, { "epoch": 0.05520541593248632, "grad_norm": 5.965984242944374, "learning_rate": 1e-05, "loss": 1.2962, "step": 2126 }, { "epoch": 0.05523138273207827, "grad_norm": 6.208389316110147, "learning_rate": 1e-05, "loss": 1.2368, "step": 2127 }, { "epoch": 0.05525734953167022, "grad_norm": 5.409067643467444, "learning_rate": 1e-05, "loss": 1.2486, "step": 2128 }, { "epoch": 0.05528331633126217, "grad_norm": 14.814548270335534, "learning_rate": 1e-05, "loss": 1.2862, "step": 2129 }, { "epoch": 0.05530928313085412, "grad_norm": 4.713593551884091, "learning_rate": 1e-05, "loss": 1.2756, "step": 2130 }, { "epoch": 0.05533524993044607, "grad_norm": 3.6869992708226667, "learning_rate": 1e-05, "loss": 1.2583, "step": 2131 }, { "epoch": 0.05536121673003802, "grad_norm": 4.121926289572622, "learning_rate": 1e-05, "loss": 1.2651, "step": 2132 }, { "epoch": 0.05538718352962997, "grad_norm": 5.174222016020315, "learning_rate": 1e-05, "loss": 1.3267, "step": 2133 }, { "epoch": 0.05541315032922192, "grad_norm": 6.186118231722982, "learning_rate": 1e-05, "loss": 1.2551, "step": 2134 }, { "epoch": 0.05543911712881387, "grad_norm": 4.754959127769943, "learning_rate": 1e-05, "loss": 1.3184, "step": 2135 }, { "epoch": 0.05546508392840582, "grad_norm": 4.800845612109245, "learning_rate": 1e-05, "loss": 1.2883, "step": 2136 }, { "epoch": 0.05549105072799777, "grad_norm": 4.335998644698378, "learning_rate": 1e-05, "loss": 1.2957, "step": 2137 }, { "epoch": 0.055517017527589724, "grad_norm": 4.035827639949442, "learning_rate": 1e-05, "loss": 1.3026, "step": 2138 }, { "epoch": 0.055542984327181674, "grad_norm": 35.0505790368763, "learning_rate": 1e-05, "loss": 1.2665, "step": 2139 }, { "epoch": 0.055568951126773625, "grad_norm": 6.336224983363116, "learning_rate": 1e-05, "loss": 1.2826, "step": 2140 }, { "epoch": 0.055594917926365575, "grad_norm": 5.337151988206336, "learning_rate": 1e-05, "loss": 1.3027, "step": 2141 }, { "epoch": 0.055620884725957526, "grad_norm": 5.174014107229659, "learning_rate": 1e-05, "loss": 1.3132, "step": 2142 }, { "epoch": 0.055646851525549476, "grad_norm": 5.529698903335321, "learning_rate": 1e-05, "loss": 1.2826, "step": 2143 }, { "epoch": 0.05567281832514143, "grad_norm": 5.778593205514163, "learning_rate": 1e-05, "loss": 1.2556, "step": 2144 }, { "epoch": 0.05569878512473338, "grad_norm": 4.094997303348161, "learning_rate": 1e-05, "loss": 1.2601, "step": 2145 }, { "epoch": 0.05572475192432533, "grad_norm": 5.944040330280242, "learning_rate": 1e-05, "loss": 1.2366, "step": 2146 }, { "epoch": 0.05575071872391728, "grad_norm": 4.008053302037369, "learning_rate": 1e-05, "loss": 1.278, "step": 2147 }, { "epoch": 0.05577668552350923, "grad_norm": 10.329469717323986, "learning_rate": 1e-05, "loss": 1.2944, "step": 2148 }, { "epoch": 0.05580265232310118, "grad_norm": 7.471299251134303, "learning_rate": 1e-05, "loss": 1.2926, "step": 2149 }, { "epoch": 0.05582861912269313, "grad_norm": 7.075287711288479, "learning_rate": 1e-05, "loss": 1.2304, "step": 2150 }, { "epoch": 0.05585458592228508, "grad_norm": 6.370971603672762, "learning_rate": 1e-05, "loss": 1.3261, "step": 2151 }, { "epoch": 0.05588055272187703, "grad_norm": 5.365230997735104, "learning_rate": 1e-05, "loss": 1.2352, "step": 2152 }, { "epoch": 0.05590651952146898, "grad_norm": 8.354356974450097, "learning_rate": 1e-05, "loss": 1.259, "step": 2153 }, { "epoch": 0.05593248632106093, "grad_norm": 7.047410086132291, "learning_rate": 1e-05, "loss": 1.3208, "step": 2154 }, { "epoch": 0.05595845312065288, "grad_norm": 3.594881360639084, "learning_rate": 1e-05, "loss": 1.2432, "step": 2155 }, { "epoch": 0.05598441992024483, "grad_norm": 5.987167941703173, "learning_rate": 1e-05, "loss": 1.264, "step": 2156 }, { "epoch": 0.05601038671983678, "grad_norm": 5.1282182218922445, "learning_rate": 1e-05, "loss": 1.2351, "step": 2157 }, { "epoch": 0.05603635351942873, "grad_norm": 4.871282089006325, "learning_rate": 1e-05, "loss": 1.2649, "step": 2158 }, { "epoch": 0.056062320319020684, "grad_norm": 4.268168937936228, "learning_rate": 1e-05, "loss": 1.2908, "step": 2159 }, { "epoch": 0.056088287118612634, "grad_norm": 8.865644655148726, "learning_rate": 1e-05, "loss": 1.2912, "step": 2160 }, { "epoch": 0.056114253918204585, "grad_norm": 3.755803766895284, "learning_rate": 1e-05, "loss": 1.2846, "step": 2161 }, { "epoch": 0.05614022071779653, "grad_norm": 5.348464938112044, "learning_rate": 1e-05, "loss": 1.2557, "step": 2162 }, { "epoch": 0.05616618751738848, "grad_norm": 4.541330844234099, "learning_rate": 1e-05, "loss": 1.305, "step": 2163 }, { "epoch": 0.05619215431698043, "grad_norm": 4.779272530715481, "learning_rate": 1e-05, "loss": 1.2729, "step": 2164 }, { "epoch": 0.05621812111657238, "grad_norm": 5.777653745803437, "learning_rate": 1e-05, "loss": 1.3027, "step": 2165 }, { "epoch": 0.05624408791616433, "grad_norm": 3.5768387660714707, "learning_rate": 1e-05, "loss": 1.2992, "step": 2166 }, { "epoch": 0.05627005471575628, "grad_norm": 18.361585638185975, "learning_rate": 1e-05, "loss": 1.289, "step": 2167 }, { "epoch": 0.05629602151534823, "grad_norm": 3.4707422807037895, "learning_rate": 1e-05, "loss": 1.2707, "step": 2168 }, { "epoch": 0.05632198831494018, "grad_norm": 7.198915919176787, "learning_rate": 1e-05, "loss": 1.2865, "step": 2169 }, { "epoch": 0.05634795511453213, "grad_norm": 4.832790848553686, "learning_rate": 1e-05, "loss": 1.2972, "step": 2170 }, { "epoch": 0.05637392191412408, "grad_norm": 5.55848222997319, "learning_rate": 1e-05, "loss": 1.2218, "step": 2171 }, { "epoch": 0.05639988871371603, "grad_norm": 4.259008397133711, "learning_rate": 1e-05, "loss": 1.2774, "step": 2172 }, { "epoch": 0.05642585551330798, "grad_norm": 8.800960696406245, "learning_rate": 1e-05, "loss": 1.2735, "step": 2173 }, { "epoch": 0.056451822312899934, "grad_norm": 5.55142549645806, "learning_rate": 1e-05, "loss": 1.2745, "step": 2174 }, { "epoch": 0.056477789112491884, "grad_norm": 5.666962484978381, "learning_rate": 1e-05, "loss": 1.2785, "step": 2175 }, { "epoch": 0.056503755912083835, "grad_norm": 6.940272421643194, "learning_rate": 1e-05, "loss": 1.2657, "step": 2176 }, { "epoch": 0.056529722711675785, "grad_norm": 5.993852645106484, "learning_rate": 1e-05, "loss": 1.3195, "step": 2177 }, { "epoch": 0.056555689511267736, "grad_norm": 11.020369656380916, "learning_rate": 1e-05, "loss": 1.316, "step": 2178 }, { "epoch": 0.056581656310859686, "grad_norm": 12.137440385289038, "learning_rate": 1e-05, "loss": 1.2974, "step": 2179 }, { "epoch": 0.05660762311045164, "grad_norm": 8.669186396983545, "learning_rate": 1e-05, "loss": 1.2825, "step": 2180 }, { "epoch": 0.05663358991004359, "grad_norm": 4.489921090976794, "learning_rate": 1e-05, "loss": 1.2374, "step": 2181 }, { "epoch": 0.05665955670963554, "grad_norm": 4.422105762943952, "learning_rate": 1e-05, "loss": 1.3015, "step": 2182 }, { "epoch": 0.05668552350922749, "grad_norm": 4.772096136398991, "learning_rate": 1e-05, "loss": 1.2127, "step": 2183 }, { "epoch": 0.05671149030881944, "grad_norm": 4.87855576171785, "learning_rate": 1e-05, "loss": 1.2883, "step": 2184 }, { "epoch": 0.05673745710841139, "grad_norm": 3.3675232474268757, "learning_rate": 1e-05, "loss": 1.2552, "step": 2185 }, { "epoch": 0.05676342390800334, "grad_norm": 3.7218685377498866, "learning_rate": 1e-05, "loss": 1.2536, "step": 2186 }, { "epoch": 0.05678939070759529, "grad_norm": 7.93535242539576, "learning_rate": 1e-05, "loss": 1.2875, "step": 2187 }, { "epoch": 0.05681535750718724, "grad_norm": 6.73312245910591, "learning_rate": 1e-05, "loss": 1.2614, "step": 2188 }, { "epoch": 0.05684132430677919, "grad_norm": 5.113366111010388, "learning_rate": 1e-05, "loss": 1.2775, "step": 2189 }, { "epoch": 0.05686729110637114, "grad_norm": 5.067644119509288, "learning_rate": 1e-05, "loss": 1.2602, "step": 2190 }, { "epoch": 0.05689325790596309, "grad_norm": 4.2003030485993795, "learning_rate": 1e-05, "loss": 1.3285, "step": 2191 }, { "epoch": 0.05691922470555504, "grad_norm": 6.851777128818337, "learning_rate": 1e-05, "loss": 1.3004, "step": 2192 }, { "epoch": 0.05694519150514699, "grad_norm": 6.649632849632535, "learning_rate": 1e-05, "loss": 1.2637, "step": 2193 }, { "epoch": 0.05697115830473894, "grad_norm": 4.188660190476685, "learning_rate": 1e-05, "loss": 1.3015, "step": 2194 }, { "epoch": 0.056997125104330894, "grad_norm": 6.152378472124242, "learning_rate": 1e-05, "loss": 1.3007, "step": 2195 }, { "epoch": 0.057023091903922844, "grad_norm": 5.104731605042667, "learning_rate": 1e-05, "loss": 1.277, "step": 2196 }, { "epoch": 0.057049058703514795, "grad_norm": 5.544410179421333, "learning_rate": 1e-05, "loss": 1.2874, "step": 2197 }, { "epoch": 0.057075025503106745, "grad_norm": 8.906708715571817, "learning_rate": 1e-05, "loss": 1.2889, "step": 2198 }, { "epoch": 0.057100992302698696, "grad_norm": 4.571346746291415, "learning_rate": 1e-05, "loss": 1.2735, "step": 2199 }, { "epoch": 0.05712695910229064, "grad_norm": 7.231198365234028, "learning_rate": 1e-05, "loss": 1.2703, "step": 2200 }, { "epoch": 0.05715292590188259, "grad_norm": 9.609493298306385, "learning_rate": 1e-05, "loss": 1.3214, "step": 2201 }, { "epoch": 0.05717889270147454, "grad_norm": 5.203608493278328, "learning_rate": 1e-05, "loss": 1.2238, "step": 2202 }, { "epoch": 0.05720485950106649, "grad_norm": 7.999355767059512, "learning_rate": 1e-05, "loss": 1.2675, "step": 2203 }, { "epoch": 0.05723082630065844, "grad_norm": 3.4900898276412433, "learning_rate": 1e-05, "loss": 1.2546, "step": 2204 }, { "epoch": 0.05725679310025039, "grad_norm": 10.636050233766072, "learning_rate": 1e-05, "loss": 1.2796, "step": 2205 }, { "epoch": 0.05728275989984234, "grad_norm": 4.9339260252195585, "learning_rate": 1e-05, "loss": 1.2555, "step": 2206 }, { "epoch": 0.05730872669943429, "grad_norm": 3.7491404501959393, "learning_rate": 1e-05, "loss": 1.2375, "step": 2207 }, { "epoch": 0.05733469349902624, "grad_norm": 4.585608501428043, "learning_rate": 1e-05, "loss": 1.3062, "step": 2208 }, { "epoch": 0.057360660298618193, "grad_norm": 9.671027372868787, "learning_rate": 1e-05, "loss": 1.2685, "step": 2209 }, { "epoch": 0.057386627098210144, "grad_norm": 4.139463326503779, "learning_rate": 1e-05, "loss": 1.2923, "step": 2210 }, { "epoch": 0.057412593897802094, "grad_norm": 6.2472048803555875, "learning_rate": 1e-05, "loss": 1.2659, "step": 2211 }, { "epoch": 0.057438560697394045, "grad_norm": 5.961385767569097, "learning_rate": 1e-05, "loss": 1.3522, "step": 2212 }, { "epoch": 0.057464527496985995, "grad_norm": 6.200505316046137, "learning_rate": 1e-05, "loss": 1.2501, "step": 2213 }, { "epoch": 0.057490494296577946, "grad_norm": 4.782215694066933, "learning_rate": 1e-05, "loss": 1.268, "step": 2214 }, { "epoch": 0.057516461096169896, "grad_norm": 15.263829198845336, "learning_rate": 1e-05, "loss": 1.2844, "step": 2215 }, { "epoch": 0.05754242789576185, "grad_norm": 4.446217959258784, "learning_rate": 1e-05, "loss": 1.3012, "step": 2216 }, { "epoch": 0.0575683946953538, "grad_norm": 8.893557636381553, "learning_rate": 1e-05, "loss": 1.2937, "step": 2217 }, { "epoch": 0.05759436149494575, "grad_norm": 4.722301031215497, "learning_rate": 1e-05, "loss": 1.3221, "step": 2218 }, { "epoch": 0.0576203282945377, "grad_norm": 69.40036839104314, "learning_rate": 1e-05, "loss": 1.2627, "step": 2219 }, { "epoch": 0.05764629509412965, "grad_norm": 11.842111761080202, "learning_rate": 1e-05, "loss": 1.3275, "step": 2220 }, { "epoch": 0.0576722618937216, "grad_norm": 5.6307592789945655, "learning_rate": 1e-05, "loss": 1.2236, "step": 2221 }, { "epoch": 0.05769822869331355, "grad_norm": 4.808233772891092, "learning_rate": 1e-05, "loss": 1.2675, "step": 2222 }, { "epoch": 0.0577241954929055, "grad_norm": 5.280035477432357, "learning_rate": 1e-05, "loss": 1.2851, "step": 2223 }, { "epoch": 0.05775016229249745, "grad_norm": 4.034780922464708, "learning_rate": 1e-05, "loss": 1.3052, "step": 2224 }, { "epoch": 0.0577761290920894, "grad_norm": 22.051805708598234, "learning_rate": 1e-05, "loss": 1.238, "step": 2225 }, { "epoch": 0.05780209589168135, "grad_norm": 5.450711741037744, "learning_rate": 1e-05, "loss": 1.305, "step": 2226 }, { "epoch": 0.0578280626912733, "grad_norm": 4.3696083633599825, "learning_rate": 1e-05, "loss": 1.2256, "step": 2227 }, { "epoch": 0.05785402949086525, "grad_norm": 5.9180728512914875, "learning_rate": 1e-05, "loss": 1.3062, "step": 2228 }, { "epoch": 0.0578799962904572, "grad_norm": 4.355681355166953, "learning_rate": 1e-05, "loss": 1.2549, "step": 2229 }, { "epoch": 0.05790596309004915, "grad_norm": 5.2809939350598265, "learning_rate": 1e-05, "loss": 1.3111, "step": 2230 }, { "epoch": 0.057931929889641104, "grad_norm": 4.68576851972901, "learning_rate": 1e-05, "loss": 1.2601, "step": 2231 }, { "epoch": 0.057957896689233054, "grad_norm": 5.849538944692865, "learning_rate": 1e-05, "loss": 1.2685, "step": 2232 }, { "epoch": 0.057983863488825005, "grad_norm": 4.7136321956589, "learning_rate": 1e-05, "loss": 1.236, "step": 2233 }, { "epoch": 0.058009830288416955, "grad_norm": 6.7464675137122105, "learning_rate": 1e-05, "loss": 1.2762, "step": 2234 }, { "epoch": 0.058035797088008906, "grad_norm": 7.565072623015849, "learning_rate": 1e-05, "loss": 1.2586, "step": 2235 }, { "epoch": 0.058061763887600856, "grad_norm": 4.461146361477116, "learning_rate": 1e-05, "loss": 1.2456, "step": 2236 }, { "epoch": 0.05808773068719281, "grad_norm": 7.212871056702403, "learning_rate": 1e-05, "loss": 1.3032, "step": 2237 }, { "epoch": 0.05811369748678476, "grad_norm": 5.0855455848181785, "learning_rate": 1e-05, "loss": 1.2937, "step": 2238 }, { "epoch": 0.0581396642863767, "grad_norm": 5.552641458233755, "learning_rate": 1e-05, "loss": 1.1913, "step": 2239 }, { "epoch": 0.05816563108596865, "grad_norm": 4.985394890696877, "learning_rate": 1e-05, "loss": 1.2628, "step": 2240 }, { "epoch": 0.0581915978855606, "grad_norm": 7.772797064541372, "learning_rate": 1e-05, "loss": 1.2624, "step": 2241 }, { "epoch": 0.05821756468515255, "grad_norm": 6.325479590285557, "learning_rate": 1e-05, "loss": 1.3003, "step": 2242 }, { "epoch": 0.0582435314847445, "grad_norm": 4.493274112805452, "learning_rate": 1e-05, "loss": 1.254, "step": 2243 }, { "epoch": 0.05826949828433645, "grad_norm": 7.631556959172469, "learning_rate": 1e-05, "loss": 1.257, "step": 2244 }, { "epoch": 0.058295465083928404, "grad_norm": 4.900515175031385, "learning_rate": 1e-05, "loss": 1.2488, "step": 2245 }, { "epoch": 0.058321431883520354, "grad_norm": 5.470481903495315, "learning_rate": 1e-05, "loss": 1.2562, "step": 2246 }, { "epoch": 0.058347398683112305, "grad_norm": 9.602108771456754, "learning_rate": 1e-05, "loss": 1.2884, "step": 2247 }, { "epoch": 0.058373365482704255, "grad_norm": 5.73692925965819, "learning_rate": 1e-05, "loss": 1.2438, "step": 2248 }, { "epoch": 0.058399332282296205, "grad_norm": 4.675415942614111, "learning_rate": 1e-05, "loss": 1.2639, "step": 2249 }, { "epoch": 0.058425299081888156, "grad_norm": 6.000421191372898, "learning_rate": 1e-05, "loss": 1.2668, "step": 2250 }, { "epoch": 0.058451265881480106, "grad_norm": 8.864863234754297, "learning_rate": 1e-05, "loss": 1.2745, "step": 2251 }, { "epoch": 0.05847723268107206, "grad_norm": 5.2239365074890936, "learning_rate": 1e-05, "loss": 1.2346, "step": 2252 }, { "epoch": 0.05850319948066401, "grad_norm": 15.63008511003299, "learning_rate": 1e-05, "loss": 1.3045, "step": 2253 }, { "epoch": 0.05852916628025596, "grad_norm": 5.586878301182305, "learning_rate": 1e-05, "loss": 1.3128, "step": 2254 }, { "epoch": 0.05855513307984791, "grad_norm": 4.378777970843502, "learning_rate": 1e-05, "loss": 1.2557, "step": 2255 }, { "epoch": 0.05858109987943986, "grad_norm": 4.814717364072209, "learning_rate": 1e-05, "loss": 1.2226, "step": 2256 }, { "epoch": 0.05860706667903181, "grad_norm": 5.569631569695033, "learning_rate": 1e-05, "loss": 1.3007, "step": 2257 }, { "epoch": 0.05863303347862376, "grad_norm": 4.385526146809592, "learning_rate": 1e-05, "loss": 1.2576, "step": 2258 }, { "epoch": 0.05865900027821571, "grad_norm": 6.980871768496637, "learning_rate": 1e-05, "loss": 1.3105, "step": 2259 }, { "epoch": 0.05868496707780766, "grad_norm": 9.49819848898373, "learning_rate": 1e-05, "loss": 1.2507, "step": 2260 }, { "epoch": 0.05871093387739961, "grad_norm": 4.672287530451429, "learning_rate": 1e-05, "loss": 1.301, "step": 2261 }, { "epoch": 0.05873690067699156, "grad_norm": 4.359510850327818, "learning_rate": 1e-05, "loss": 1.2886, "step": 2262 }, { "epoch": 0.05876286747658351, "grad_norm": 3.3740703573980735, "learning_rate": 1e-05, "loss": 1.23, "step": 2263 }, { "epoch": 0.05878883427617546, "grad_norm": 4.869521559569915, "learning_rate": 1e-05, "loss": 1.2585, "step": 2264 }, { "epoch": 0.05881480107576741, "grad_norm": 7.054705073777253, "learning_rate": 1e-05, "loss": 1.2642, "step": 2265 }, { "epoch": 0.058840767875359364, "grad_norm": 4.815428202158309, "learning_rate": 1e-05, "loss": 1.3374, "step": 2266 }, { "epoch": 0.058866734674951314, "grad_norm": 5.0615052552791795, "learning_rate": 1e-05, "loss": 1.3182, "step": 2267 }, { "epoch": 0.058892701474543264, "grad_norm": 9.374995117186229, "learning_rate": 1e-05, "loss": 1.271, "step": 2268 }, { "epoch": 0.058918668274135215, "grad_norm": 5.995192190975889, "learning_rate": 1e-05, "loss": 1.2276, "step": 2269 }, { "epoch": 0.058944635073727165, "grad_norm": 4.8726258977104004, "learning_rate": 1e-05, "loss": 1.2936, "step": 2270 }, { "epoch": 0.058970601873319116, "grad_norm": 7.894905968577811, "learning_rate": 1e-05, "loss": 1.3267, "step": 2271 }, { "epoch": 0.058996568672911066, "grad_norm": 6.047839499347125, "learning_rate": 1e-05, "loss": 1.312, "step": 2272 }, { "epoch": 0.05902253547250302, "grad_norm": 4.055405981246452, "learning_rate": 1e-05, "loss": 1.2886, "step": 2273 }, { "epoch": 0.05904850227209497, "grad_norm": 7.335206081665194, "learning_rate": 1e-05, "loss": 1.2233, "step": 2274 }, { "epoch": 0.05907446907168692, "grad_norm": 6.307132365081162, "learning_rate": 1e-05, "loss": 1.272, "step": 2275 }, { "epoch": 0.05910043587127887, "grad_norm": 5.223199831764427, "learning_rate": 1e-05, "loss": 1.2711, "step": 2276 }, { "epoch": 0.05912640267087081, "grad_norm": 6.519824387695325, "learning_rate": 1e-05, "loss": 1.2662, "step": 2277 }, { "epoch": 0.05915236947046276, "grad_norm": 4.013487725778614, "learning_rate": 1e-05, "loss": 1.2456, "step": 2278 }, { "epoch": 0.05917833627005471, "grad_norm": 5.140042257469226, "learning_rate": 1e-05, "loss": 1.2603, "step": 2279 }, { "epoch": 0.05920430306964666, "grad_norm": 5.604289605400596, "learning_rate": 1e-05, "loss": 1.3014, "step": 2280 }, { "epoch": 0.059230269869238614, "grad_norm": 5.87634424292663, "learning_rate": 1e-05, "loss": 1.2502, "step": 2281 }, { "epoch": 0.059256236668830564, "grad_norm": 7.078107484370574, "learning_rate": 1e-05, "loss": 1.2847, "step": 2282 }, { "epoch": 0.059282203468422515, "grad_norm": 5.8454509034503435, "learning_rate": 1e-05, "loss": 1.2851, "step": 2283 }, { "epoch": 0.059308170268014465, "grad_norm": 3.5700297504542475, "learning_rate": 1e-05, "loss": 1.2849, "step": 2284 }, { "epoch": 0.059334137067606416, "grad_norm": 3.975838406032593, "learning_rate": 1e-05, "loss": 1.2671, "step": 2285 }, { "epoch": 0.059360103867198366, "grad_norm": 5.211100013863885, "learning_rate": 1e-05, "loss": 1.2815, "step": 2286 }, { "epoch": 0.05938607066679032, "grad_norm": 6.059373501663849, "learning_rate": 1e-05, "loss": 1.274, "step": 2287 }, { "epoch": 0.05941203746638227, "grad_norm": 5.482858127459584, "learning_rate": 1e-05, "loss": 1.2821, "step": 2288 }, { "epoch": 0.05943800426597422, "grad_norm": 4.306768838440112, "learning_rate": 1e-05, "loss": 1.2563, "step": 2289 }, { "epoch": 0.05946397106556617, "grad_norm": 8.027105189581695, "learning_rate": 1e-05, "loss": 1.276, "step": 2290 }, { "epoch": 0.05948993786515812, "grad_norm": 4.690850649799628, "learning_rate": 1e-05, "loss": 1.2661, "step": 2291 }, { "epoch": 0.05951590466475007, "grad_norm": 5.879866512106828, "learning_rate": 1e-05, "loss": 1.2925, "step": 2292 }, { "epoch": 0.05954187146434202, "grad_norm": 4.18191565901341, "learning_rate": 1e-05, "loss": 1.2901, "step": 2293 }, { "epoch": 0.05956783826393397, "grad_norm": 4.967081617562996, "learning_rate": 1e-05, "loss": 1.2466, "step": 2294 }, { "epoch": 0.05959380506352592, "grad_norm": 3.703019667789273, "learning_rate": 1e-05, "loss": 1.2684, "step": 2295 }, { "epoch": 0.05961977186311787, "grad_norm": 5.5009622599061805, "learning_rate": 1e-05, "loss": 1.2646, "step": 2296 }, { "epoch": 0.05964573866270982, "grad_norm": 4.558686635461142, "learning_rate": 1e-05, "loss": 1.2973, "step": 2297 }, { "epoch": 0.05967170546230177, "grad_norm": 5.982609660271007, "learning_rate": 1e-05, "loss": 1.2571, "step": 2298 }, { "epoch": 0.05969767226189372, "grad_norm": 4.908390040088963, "learning_rate": 1e-05, "loss": 1.2843, "step": 2299 }, { "epoch": 0.05972363906148567, "grad_norm": 4.481130669147532, "learning_rate": 1e-05, "loss": 1.2808, "step": 2300 }, { "epoch": 0.05974960586107762, "grad_norm": 3.5730724393439486, "learning_rate": 1e-05, "loss": 1.2727, "step": 2301 }, { "epoch": 0.059775572660669574, "grad_norm": 4.033735351914181, "learning_rate": 1e-05, "loss": 1.2149, "step": 2302 }, { "epoch": 0.059801539460261524, "grad_norm": 5.6942849403103555, "learning_rate": 1e-05, "loss": 1.2904, "step": 2303 }, { "epoch": 0.059827506259853475, "grad_norm": 5.170881193166477, "learning_rate": 1e-05, "loss": 1.2998, "step": 2304 }, { "epoch": 0.059853473059445425, "grad_norm": 4.312084233241544, "learning_rate": 1e-05, "loss": 1.2832, "step": 2305 }, { "epoch": 0.059879439859037376, "grad_norm": 5.0482606668532926, "learning_rate": 1e-05, "loss": 1.2782, "step": 2306 }, { "epoch": 0.059905406658629326, "grad_norm": 5.90229786237126, "learning_rate": 1e-05, "loss": 1.2849, "step": 2307 }, { "epoch": 0.059931373458221276, "grad_norm": 5.5927958607078425, "learning_rate": 1e-05, "loss": 1.2635, "step": 2308 }, { "epoch": 0.05995734025781323, "grad_norm": 4.610571340709148, "learning_rate": 1e-05, "loss": 1.3089, "step": 2309 }, { "epoch": 0.05998330705740518, "grad_norm": 4.61690142111566, "learning_rate": 1e-05, "loss": 1.2388, "step": 2310 }, { "epoch": 0.06000927385699713, "grad_norm": 3.5911934425051095, "learning_rate": 1e-05, "loss": 1.2509, "step": 2311 }, { "epoch": 0.06003524065658908, "grad_norm": 3.837104076128251, "learning_rate": 1e-05, "loss": 1.2507, "step": 2312 }, { "epoch": 0.06006120745618103, "grad_norm": 7.398707416126371, "learning_rate": 1e-05, "loss": 1.3222, "step": 2313 }, { "epoch": 0.06008717425577298, "grad_norm": 8.422570251189981, "learning_rate": 1e-05, "loss": 1.2872, "step": 2314 }, { "epoch": 0.06011314105536492, "grad_norm": 5.58957520592272, "learning_rate": 1e-05, "loss": 1.2286, "step": 2315 }, { "epoch": 0.06013910785495687, "grad_norm": 8.196603620591027, "learning_rate": 1e-05, "loss": 1.2503, "step": 2316 }, { "epoch": 0.060165074654548824, "grad_norm": 3.792909662982741, "learning_rate": 1e-05, "loss": 1.2227, "step": 2317 }, { "epoch": 0.060191041454140774, "grad_norm": 10.296830172868091, "learning_rate": 1e-05, "loss": 1.318, "step": 2318 }, { "epoch": 0.060217008253732725, "grad_norm": 3.391007924760509, "learning_rate": 1e-05, "loss": 1.2827, "step": 2319 }, { "epoch": 0.060242975053324675, "grad_norm": 5.1377492356928345, "learning_rate": 1e-05, "loss": 1.2667, "step": 2320 }, { "epoch": 0.060268941852916626, "grad_norm": 5.820544838588359, "learning_rate": 1e-05, "loss": 1.3185, "step": 2321 }, { "epoch": 0.060294908652508576, "grad_norm": 4.2896003316450155, "learning_rate": 1e-05, "loss": 1.2375, "step": 2322 }, { "epoch": 0.06032087545210053, "grad_norm": 5.153583923965199, "learning_rate": 1e-05, "loss": 1.2615, "step": 2323 }, { "epoch": 0.06034684225169248, "grad_norm": 4.386900492951555, "learning_rate": 1e-05, "loss": 1.318, "step": 2324 }, { "epoch": 0.06037280905128443, "grad_norm": 4.069480415564269, "learning_rate": 1e-05, "loss": 1.2687, "step": 2325 }, { "epoch": 0.06039877585087638, "grad_norm": 17.40857115358337, "learning_rate": 1e-05, "loss": 1.2491, "step": 2326 }, { "epoch": 0.06042474265046833, "grad_norm": 5.829656077688534, "learning_rate": 1e-05, "loss": 1.2656, "step": 2327 }, { "epoch": 0.06045070945006028, "grad_norm": 4.522283060871091, "learning_rate": 1e-05, "loss": 1.3002, "step": 2328 }, { "epoch": 0.06047667624965223, "grad_norm": 5.980114089745043, "learning_rate": 1e-05, "loss": 1.2239, "step": 2329 }, { "epoch": 0.06050264304924418, "grad_norm": 4.810146796827644, "learning_rate": 1e-05, "loss": 1.2547, "step": 2330 }, { "epoch": 0.06052860984883613, "grad_norm": 11.007556054056828, "learning_rate": 1e-05, "loss": 1.3219, "step": 2331 }, { "epoch": 0.06055457664842808, "grad_norm": 6.996604095707453, "learning_rate": 1e-05, "loss": 1.2398, "step": 2332 }, { "epoch": 0.06058054344802003, "grad_norm": 5.058995670567847, "learning_rate": 1e-05, "loss": 1.2517, "step": 2333 }, { "epoch": 0.06060651024761198, "grad_norm": 3.9608783905614793, "learning_rate": 1e-05, "loss": 1.2239, "step": 2334 }, { "epoch": 0.06063247704720393, "grad_norm": 5.562423191183183, "learning_rate": 1e-05, "loss": 1.2859, "step": 2335 }, { "epoch": 0.06065844384679588, "grad_norm": 4.2421016227783275, "learning_rate": 1e-05, "loss": 1.2571, "step": 2336 }, { "epoch": 0.06068441064638783, "grad_norm": 5.233945914374233, "learning_rate": 1e-05, "loss": 1.3211, "step": 2337 }, { "epoch": 0.060710377445979784, "grad_norm": 4.192722764395131, "learning_rate": 1e-05, "loss": 1.257, "step": 2338 }, { "epoch": 0.060736344245571734, "grad_norm": 4.360898425078545, "learning_rate": 1e-05, "loss": 1.2642, "step": 2339 }, { "epoch": 0.060762311045163685, "grad_norm": 8.406717393779047, "learning_rate": 1e-05, "loss": 1.2874, "step": 2340 }, { "epoch": 0.060788277844755635, "grad_norm": 7.630034988743064, "learning_rate": 1e-05, "loss": 1.2741, "step": 2341 }, { "epoch": 0.060814244644347586, "grad_norm": 7.679103825257347, "learning_rate": 1e-05, "loss": 1.289, "step": 2342 }, { "epoch": 0.060840211443939536, "grad_norm": 5.099223186198182, "learning_rate": 1e-05, "loss": 1.291, "step": 2343 }, { "epoch": 0.06086617824353149, "grad_norm": 6.703955254343968, "learning_rate": 1e-05, "loss": 1.2082, "step": 2344 }, { "epoch": 0.06089214504312344, "grad_norm": 4.421470165255128, "learning_rate": 1e-05, "loss": 1.2677, "step": 2345 }, { "epoch": 0.06091811184271539, "grad_norm": 29.27368352912358, "learning_rate": 1e-05, "loss": 1.3272, "step": 2346 }, { "epoch": 0.06094407864230734, "grad_norm": 13.99062469461972, "learning_rate": 1e-05, "loss": 1.2286, "step": 2347 }, { "epoch": 0.06097004544189929, "grad_norm": 11.135336884848275, "learning_rate": 1e-05, "loss": 1.2831, "step": 2348 }, { "epoch": 0.06099601224149124, "grad_norm": 4.593455817735412, "learning_rate": 1e-05, "loss": 1.2783, "step": 2349 }, { "epoch": 0.06102197904108319, "grad_norm": 5.591833031231344, "learning_rate": 1e-05, "loss": 1.2858, "step": 2350 }, { "epoch": 0.06104794584067514, "grad_norm": 4.562972579937204, "learning_rate": 1e-05, "loss": 1.3029, "step": 2351 }, { "epoch": 0.06107391264026709, "grad_norm": 5.7692075934311395, "learning_rate": 1e-05, "loss": 1.2867, "step": 2352 }, { "epoch": 0.061099879439859034, "grad_norm": 4.496195774594501, "learning_rate": 1e-05, "loss": 1.2627, "step": 2353 }, { "epoch": 0.061125846239450984, "grad_norm": 4.125158364695042, "learning_rate": 1e-05, "loss": 1.232, "step": 2354 }, { "epoch": 0.061151813039042935, "grad_norm": 5.99980131456109, "learning_rate": 1e-05, "loss": 1.3245, "step": 2355 }, { "epoch": 0.061177779838634885, "grad_norm": 5.883929345449037, "learning_rate": 1e-05, "loss": 1.2908, "step": 2356 }, { "epoch": 0.061203746638226836, "grad_norm": 5.733753305770623, "learning_rate": 1e-05, "loss": 1.2483, "step": 2357 }, { "epoch": 0.061229713437818786, "grad_norm": 6.471845540376124, "learning_rate": 1e-05, "loss": 1.2507, "step": 2358 }, { "epoch": 0.06125568023741074, "grad_norm": 8.720831072339674, "learning_rate": 1e-05, "loss": 1.2356, "step": 2359 }, { "epoch": 0.06128164703700269, "grad_norm": 10.187484647587167, "learning_rate": 1e-05, "loss": 1.2384, "step": 2360 }, { "epoch": 0.06130761383659464, "grad_norm": 5.269673316573175, "learning_rate": 1e-05, "loss": 1.3271, "step": 2361 }, { "epoch": 0.06133358063618659, "grad_norm": 7.854279157269766, "learning_rate": 1e-05, "loss": 1.2268, "step": 2362 }, { "epoch": 0.06135954743577854, "grad_norm": 7.170798360091696, "learning_rate": 1e-05, "loss": 1.3302, "step": 2363 }, { "epoch": 0.06138551423537049, "grad_norm": 5.243731071065172, "learning_rate": 1e-05, "loss": 1.2672, "step": 2364 }, { "epoch": 0.06141148103496244, "grad_norm": 6.351582619798238, "learning_rate": 1e-05, "loss": 1.2832, "step": 2365 }, { "epoch": 0.06143744783455439, "grad_norm": 5.625596078442974, "learning_rate": 1e-05, "loss": 1.3066, "step": 2366 }, { "epoch": 0.06146341463414634, "grad_norm": 10.092257457253858, "learning_rate": 1e-05, "loss": 1.2362, "step": 2367 }, { "epoch": 0.06148938143373829, "grad_norm": 4.564177531327587, "learning_rate": 1e-05, "loss": 1.2659, "step": 2368 }, { "epoch": 0.06151534823333024, "grad_norm": 4.757610849200641, "learning_rate": 1e-05, "loss": 1.2618, "step": 2369 }, { "epoch": 0.06154131503292219, "grad_norm": 4.820485611192089, "learning_rate": 1e-05, "loss": 1.231, "step": 2370 }, { "epoch": 0.06156728183251414, "grad_norm": 13.483105861661022, "learning_rate": 1e-05, "loss": 1.2476, "step": 2371 }, { "epoch": 0.06159324863210609, "grad_norm": 6.318840845739752, "learning_rate": 1e-05, "loss": 1.267, "step": 2372 }, { "epoch": 0.06161921543169804, "grad_norm": 4.479383501822152, "learning_rate": 1e-05, "loss": 1.2457, "step": 2373 }, { "epoch": 0.061645182231289994, "grad_norm": 5.766277658796017, "learning_rate": 1e-05, "loss": 1.2576, "step": 2374 }, { "epoch": 0.061671149030881944, "grad_norm": 17.622583534755996, "learning_rate": 1e-05, "loss": 1.2938, "step": 2375 }, { "epoch": 0.061697115830473895, "grad_norm": 5.87116392173634, "learning_rate": 1e-05, "loss": 1.2926, "step": 2376 }, { "epoch": 0.061723082630065845, "grad_norm": 4.11004633188237, "learning_rate": 1e-05, "loss": 1.2301, "step": 2377 }, { "epoch": 0.061749049429657796, "grad_norm": 4.531143292453149, "learning_rate": 1e-05, "loss": 1.2525, "step": 2378 }, { "epoch": 0.061775016229249746, "grad_norm": 5.552112094978807, "learning_rate": 1e-05, "loss": 1.3019, "step": 2379 }, { "epoch": 0.0618009830288417, "grad_norm": 5.113914036773858, "learning_rate": 1e-05, "loss": 1.2293, "step": 2380 }, { "epoch": 0.06182694982843365, "grad_norm": 8.645997955103184, "learning_rate": 1e-05, "loss": 1.2842, "step": 2381 }, { "epoch": 0.0618529166280256, "grad_norm": 4.93911504286104, "learning_rate": 1e-05, "loss": 1.2578, "step": 2382 }, { "epoch": 0.06187888342761755, "grad_norm": 27.218570607904816, "learning_rate": 1e-05, "loss": 1.2295, "step": 2383 }, { "epoch": 0.0619048502272095, "grad_norm": 4.052791556449013, "learning_rate": 1e-05, "loss": 1.2902, "step": 2384 }, { "epoch": 0.06193081702680145, "grad_norm": 4.377808023941025, "learning_rate": 1e-05, "loss": 1.2656, "step": 2385 }, { "epoch": 0.0619567838263934, "grad_norm": 6.40507116869324, "learning_rate": 1e-05, "loss": 1.2725, "step": 2386 }, { "epoch": 0.06198275062598535, "grad_norm": 5.835721135756609, "learning_rate": 1e-05, "loss": 1.2517, "step": 2387 }, { "epoch": 0.0620087174255773, "grad_norm": 6.206867770424246, "learning_rate": 1e-05, "loss": 1.2628, "step": 2388 }, { "epoch": 0.06203468422516925, "grad_norm": 14.363630226757287, "learning_rate": 1e-05, "loss": 1.2454, "step": 2389 }, { "epoch": 0.0620606510247612, "grad_norm": 14.298023348690108, "learning_rate": 1e-05, "loss": 1.2761, "step": 2390 }, { "epoch": 0.06208661782435315, "grad_norm": 5.42564204060076, "learning_rate": 1e-05, "loss": 1.2832, "step": 2391 }, { "epoch": 0.062112584623945095, "grad_norm": 4.921363828242293, "learning_rate": 1e-05, "loss": 1.2937, "step": 2392 }, { "epoch": 0.062138551423537046, "grad_norm": 6.756107251885113, "learning_rate": 1e-05, "loss": 1.2753, "step": 2393 }, { "epoch": 0.062164518223128996, "grad_norm": 7.333335992061248, "learning_rate": 1e-05, "loss": 1.2617, "step": 2394 }, { "epoch": 0.06219048502272095, "grad_norm": 4.554392888534089, "learning_rate": 1e-05, "loss": 1.2748, "step": 2395 }, { "epoch": 0.0622164518223129, "grad_norm": 5.8122066300712785, "learning_rate": 1e-05, "loss": 1.2613, "step": 2396 }, { "epoch": 0.06224241862190485, "grad_norm": 5.566744781372458, "learning_rate": 1e-05, "loss": 1.2107, "step": 2397 }, { "epoch": 0.0622683854214968, "grad_norm": 6.209860424002757, "learning_rate": 1e-05, "loss": 1.2391, "step": 2398 }, { "epoch": 0.06229435222108875, "grad_norm": 4.790100084824099, "learning_rate": 1e-05, "loss": 1.296, "step": 2399 }, { "epoch": 0.0623203190206807, "grad_norm": 5.795231917894031, "learning_rate": 1e-05, "loss": 1.266, "step": 2400 }, { "epoch": 0.06234628582027265, "grad_norm": 4.638302822490153, "learning_rate": 1e-05, "loss": 1.2883, "step": 2401 }, { "epoch": 0.0623722526198646, "grad_norm": 8.631182679860565, "learning_rate": 1e-05, "loss": 1.3021, "step": 2402 }, { "epoch": 0.06239821941945655, "grad_norm": 8.760564040049287, "learning_rate": 1e-05, "loss": 1.261, "step": 2403 }, { "epoch": 0.0624241862190485, "grad_norm": 4.840639161396429, "learning_rate": 1e-05, "loss": 1.2414, "step": 2404 }, { "epoch": 0.06245015301864045, "grad_norm": 5.1279569329820385, "learning_rate": 1e-05, "loss": 1.2772, "step": 2405 }, { "epoch": 0.0624761198182324, "grad_norm": 5.678684655060632, "learning_rate": 1e-05, "loss": 1.2692, "step": 2406 }, { "epoch": 0.06250208661782436, "grad_norm": 4.825627245675839, "learning_rate": 1e-05, "loss": 1.2971, "step": 2407 }, { "epoch": 0.06252805341741631, "grad_norm": 5.78503524993734, "learning_rate": 1e-05, "loss": 1.2478, "step": 2408 }, { "epoch": 0.06255402021700826, "grad_norm": 16.04872722392067, "learning_rate": 1e-05, "loss": 1.3423, "step": 2409 }, { "epoch": 0.06257998701660021, "grad_norm": 6.0309754936721, "learning_rate": 1e-05, "loss": 1.2735, "step": 2410 }, { "epoch": 0.06260595381619215, "grad_norm": 7.2252120782157885, "learning_rate": 1e-05, "loss": 1.2821, "step": 2411 }, { "epoch": 0.0626319206157841, "grad_norm": 5.411781259789478, "learning_rate": 1e-05, "loss": 1.2899, "step": 2412 }, { "epoch": 0.06265788741537605, "grad_norm": 14.740615704331127, "learning_rate": 1e-05, "loss": 1.2833, "step": 2413 }, { "epoch": 0.062683854214968, "grad_norm": 4.519632323510584, "learning_rate": 1e-05, "loss": 1.235, "step": 2414 }, { "epoch": 0.06270982101455995, "grad_norm": 8.766193337468824, "learning_rate": 1e-05, "loss": 1.284, "step": 2415 }, { "epoch": 0.0627357878141519, "grad_norm": 18.094236588070746, "learning_rate": 1e-05, "loss": 1.3056, "step": 2416 }, { "epoch": 0.06276175461374385, "grad_norm": 5.119959888301128, "learning_rate": 1e-05, "loss": 1.2788, "step": 2417 }, { "epoch": 0.0627877214133358, "grad_norm": 9.485931419605611, "learning_rate": 1e-05, "loss": 1.2154, "step": 2418 }, { "epoch": 0.06281368821292775, "grad_norm": 6.65136768933066, "learning_rate": 1e-05, "loss": 1.2265, "step": 2419 }, { "epoch": 0.0628396550125197, "grad_norm": 11.583281446015942, "learning_rate": 1e-05, "loss": 1.2925, "step": 2420 }, { "epoch": 0.06286562181211165, "grad_norm": 5.037199495122474, "learning_rate": 1e-05, "loss": 1.3161, "step": 2421 }, { "epoch": 0.0628915886117036, "grad_norm": 5.7831724527428765, "learning_rate": 1e-05, "loss": 1.2741, "step": 2422 }, { "epoch": 0.06291755541129555, "grad_norm": 6.232944869450192, "learning_rate": 1e-05, "loss": 1.196, "step": 2423 }, { "epoch": 0.0629435222108875, "grad_norm": 14.144310099072028, "learning_rate": 1e-05, "loss": 1.3002, "step": 2424 }, { "epoch": 0.06296948901047945, "grad_norm": 9.419777910404664, "learning_rate": 1e-05, "loss": 1.2673, "step": 2425 }, { "epoch": 0.0629954558100714, "grad_norm": 5.816909921946451, "learning_rate": 1e-05, "loss": 1.2411, "step": 2426 }, { "epoch": 0.06302142260966336, "grad_norm": 6.003491974972124, "learning_rate": 1e-05, "loss": 1.2135, "step": 2427 }, { "epoch": 0.0630473894092553, "grad_norm": 5.325338848956108, "learning_rate": 1e-05, "loss": 1.2624, "step": 2428 }, { "epoch": 0.06307335620884726, "grad_norm": 5.85853639832243, "learning_rate": 1e-05, "loss": 1.2358, "step": 2429 }, { "epoch": 0.0630993230084392, "grad_norm": 7.052223763882506, "learning_rate": 1e-05, "loss": 1.2491, "step": 2430 }, { "epoch": 0.06312528980803116, "grad_norm": 4.519471744287391, "learning_rate": 1e-05, "loss": 1.2834, "step": 2431 }, { "epoch": 0.06315125660762311, "grad_norm": 13.222929742371003, "learning_rate": 1e-05, "loss": 1.2627, "step": 2432 }, { "epoch": 0.06317722340721506, "grad_norm": 4.700788050388757, "learning_rate": 1e-05, "loss": 1.2684, "step": 2433 }, { "epoch": 0.06320319020680701, "grad_norm": 70.05764925224084, "learning_rate": 1e-05, "loss": 1.249, "step": 2434 }, { "epoch": 0.06322915700639896, "grad_norm": 8.646876358787615, "learning_rate": 1e-05, "loss": 1.247, "step": 2435 }, { "epoch": 0.06325512380599091, "grad_norm": 7.776399502417003, "learning_rate": 1e-05, "loss": 1.3043, "step": 2436 }, { "epoch": 0.06328109060558286, "grad_norm": 5.187900573707142, "learning_rate": 1e-05, "loss": 1.2888, "step": 2437 }, { "epoch": 0.06330705740517481, "grad_norm": 37.29423528310458, "learning_rate": 1e-05, "loss": 1.2951, "step": 2438 }, { "epoch": 0.06333302420476676, "grad_norm": 5.269953637977809, "learning_rate": 1e-05, "loss": 1.2319, "step": 2439 }, { "epoch": 0.06335899100435871, "grad_norm": 5.206527193344818, "learning_rate": 1e-05, "loss": 1.2614, "step": 2440 }, { "epoch": 0.06338495780395066, "grad_norm": 14.83880828514346, "learning_rate": 1e-05, "loss": 1.2317, "step": 2441 }, { "epoch": 0.06341092460354261, "grad_norm": 5.899781252151827, "learning_rate": 1e-05, "loss": 1.2523, "step": 2442 }, { "epoch": 0.06343689140313456, "grad_norm": 5.204420140458789, "learning_rate": 1e-05, "loss": 1.2632, "step": 2443 }, { "epoch": 0.06346285820272651, "grad_norm": 25.611673602227388, "learning_rate": 1e-05, "loss": 1.2492, "step": 2444 }, { "epoch": 0.06348882500231846, "grad_norm": 4.868622349447836, "learning_rate": 1e-05, "loss": 1.267, "step": 2445 }, { "epoch": 0.06351479180191041, "grad_norm": 8.573095295846194, "learning_rate": 1e-05, "loss": 1.2564, "step": 2446 }, { "epoch": 0.06354075860150236, "grad_norm": 5.359331047389468, "learning_rate": 1e-05, "loss": 1.2214, "step": 2447 }, { "epoch": 0.06356672540109432, "grad_norm": 8.276726635354505, "learning_rate": 1e-05, "loss": 1.2655, "step": 2448 }, { "epoch": 0.06359269220068627, "grad_norm": 5.462679387339805, "learning_rate": 1e-05, "loss": 1.2378, "step": 2449 }, { "epoch": 0.06361865900027822, "grad_norm": 5.7255192283818515, "learning_rate": 1e-05, "loss": 1.285, "step": 2450 }, { "epoch": 0.06364462579987017, "grad_norm": 6.575737381696224, "learning_rate": 1e-05, "loss": 1.221, "step": 2451 }, { "epoch": 0.06367059259946212, "grad_norm": 8.947392464242288, "learning_rate": 1e-05, "loss": 1.2426, "step": 2452 }, { "epoch": 0.06369655939905407, "grad_norm": 5.775686855672246, "learning_rate": 1e-05, "loss": 1.256, "step": 2453 }, { "epoch": 0.06372252619864602, "grad_norm": 4.9214144052218165, "learning_rate": 1e-05, "loss": 1.2753, "step": 2454 }, { "epoch": 0.06374849299823797, "grad_norm": 4.92028085232657, "learning_rate": 1e-05, "loss": 1.2482, "step": 2455 }, { "epoch": 0.06377445979782992, "grad_norm": 6.694098808824433, "learning_rate": 1e-05, "loss": 1.2416, "step": 2456 }, { "epoch": 0.06380042659742187, "grad_norm": 6.041624643464664, "learning_rate": 1e-05, "loss": 1.2862, "step": 2457 }, { "epoch": 0.06382639339701382, "grad_norm": 4.5401030248976255, "learning_rate": 1e-05, "loss": 1.2577, "step": 2458 }, { "epoch": 0.06385236019660577, "grad_norm": 4.944557650600044, "learning_rate": 1e-05, "loss": 1.2577, "step": 2459 }, { "epoch": 0.06387832699619772, "grad_norm": 8.2366706210654, "learning_rate": 1e-05, "loss": 1.2814, "step": 2460 }, { "epoch": 0.06390429379578967, "grad_norm": 6.268752689684274, "learning_rate": 1e-05, "loss": 1.2272, "step": 2461 }, { "epoch": 0.06393026059538162, "grad_norm": 25.811878113725324, "learning_rate": 1e-05, "loss": 1.3314, "step": 2462 }, { "epoch": 0.06395622739497357, "grad_norm": 14.414316033508742, "learning_rate": 1e-05, "loss": 1.2915, "step": 2463 }, { "epoch": 0.06398219419456552, "grad_norm": 5.793412565180442, "learning_rate": 1e-05, "loss": 1.2305, "step": 2464 }, { "epoch": 0.06400816099415747, "grad_norm": 6.357787587887418, "learning_rate": 1e-05, "loss": 1.2503, "step": 2465 }, { "epoch": 0.06403412779374942, "grad_norm": 4.96144344122835, "learning_rate": 1e-05, "loss": 1.202, "step": 2466 }, { "epoch": 0.06406009459334137, "grad_norm": 4.879330594474093, "learning_rate": 1e-05, "loss": 1.2457, "step": 2467 }, { "epoch": 0.06408606139293332, "grad_norm": 44.75981910430381, "learning_rate": 1e-05, "loss": 1.3408, "step": 2468 }, { "epoch": 0.06411202819252527, "grad_norm": 5.090042543664392, "learning_rate": 1e-05, "loss": 1.2666, "step": 2469 }, { "epoch": 0.06413799499211723, "grad_norm": 9.107434231240278, "learning_rate": 1e-05, "loss": 1.309, "step": 2470 }, { "epoch": 0.06416396179170918, "grad_norm": 9.879893189134735, "learning_rate": 1e-05, "loss": 1.2608, "step": 2471 }, { "epoch": 0.06418992859130113, "grad_norm": 8.763963728912804, "learning_rate": 1e-05, "loss": 1.2445, "step": 2472 }, { "epoch": 0.06421589539089308, "grad_norm": 8.362937781217767, "learning_rate": 1e-05, "loss": 1.2714, "step": 2473 }, { "epoch": 0.06424186219048503, "grad_norm": 4.924716480440242, "learning_rate": 1e-05, "loss": 1.2809, "step": 2474 }, { "epoch": 0.06426782899007698, "grad_norm": 11.569242254054835, "learning_rate": 1e-05, "loss": 1.2651, "step": 2475 }, { "epoch": 0.06429379578966893, "grad_norm": 7.808826040426222, "learning_rate": 1e-05, "loss": 1.3415, "step": 2476 }, { "epoch": 0.06431976258926088, "grad_norm": 4.889807294547262, "learning_rate": 1e-05, "loss": 1.2746, "step": 2477 }, { "epoch": 0.06434572938885283, "grad_norm": 4.2868384203954, "learning_rate": 1e-05, "loss": 1.2245, "step": 2478 }, { "epoch": 0.06437169618844478, "grad_norm": 4.963206719875152, "learning_rate": 1e-05, "loss": 1.2316, "step": 2479 }, { "epoch": 0.06439766298803673, "grad_norm": 5.672394867919339, "learning_rate": 1e-05, "loss": 1.2817, "step": 2480 }, { "epoch": 0.06442362978762868, "grad_norm": 7.77861597297862, "learning_rate": 1e-05, "loss": 1.2879, "step": 2481 }, { "epoch": 0.06444959658722063, "grad_norm": 5.846854787574463, "learning_rate": 1e-05, "loss": 1.2617, "step": 2482 }, { "epoch": 0.06447556338681258, "grad_norm": 7.592776389167433, "learning_rate": 1e-05, "loss": 1.3314, "step": 2483 }, { "epoch": 0.06450153018640453, "grad_norm": 5.571760206168599, "learning_rate": 1e-05, "loss": 1.2591, "step": 2484 }, { "epoch": 0.06452749698599648, "grad_norm": 8.873515260103465, "learning_rate": 1e-05, "loss": 1.3137, "step": 2485 }, { "epoch": 0.06455346378558843, "grad_norm": 8.714019914276433, "learning_rate": 1e-05, "loss": 1.2995, "step": 2486 }, { "epoch": 0.06457943058518037, "grad_norm": 3.811346285996562, "learning_rate": 1e-05, "loss": 1.2365, "step": 2487 }, { "epoch": 0.06460539738477232, "grad_norm": 3.951087517057025, "learning_rate": 1e-05, "loss": 1.2565, "step": 2488 }, { "epoch": 0.06463136418436427, "grad_norm": 15.182586754297102, "learning_rate": 1e-05, "loss": 1.268, "step": 2489 }, { "epoch": 0.06465733098395622, "grad_norm": 5.869053204943787, "learning_rate": 1e-05, "loss": 1.2398, "step": 2490 }, { "epoch": 0.06468329778354817, "grad_norm": 5.190526975364103, "learning_rate": 1e-05, "loss": 1.2623, "step": 2491 }, { "epoch": 0.06470926458314012, "grad_norm": 6.0872154216320835, "learning_rate": 1e-05, "loss": 1.2451, "step": 2492 }, { "epoch": 0.06473523138273207, "grad_norm": 7.180258378504222, "learning_rate": 1e-05, "loss": 1.2374, "step": 2493 }, { "epoch": 0.06476119818232402, "grad_norm": 5.917391808717004, "learning_rate": 1e-05, "loss": 1.2522, "step": 2494 }, { "epoch": 0.06478716498191597, "grad_norm": 8.226501290726704, "learning_rate": 1e-05, "loss": 1.2987, "step": 2495 }, { "epoch": 0.06481313178150792, "grad_norm": 7.542543925763397, "learning_rate": 1e-05, "loss": 1.275, "step": 2496 }, { "epoch": 0.06483909858109987, "grad_norm": 4.957615883262002, "learning_rate": 1e-05, "loss": 1.2683, "step": 2497 }, { "epoch": 0.06486506538069182, "grad_norm": 5.438057004965819, "learning_rate": 1e-05, "loss": 1.2966, "step": 2498 }, { "epoch": 0.06489103218028378, "grad_norm": 5.068472459299618, "learning_rate": 1e-05, "loss": 1.2668, "step": 2499 }, { "epoch": 0.06491699897987573, "grad_norm": 5.6915989452408535, "learning_rate": 1e-05, "loss": 1.2619, "step": 2500 }, { "epoch": 0.06494296577946768, "grad_norm": 5.259070961426489, "learning_rate": 1e-05, "loss": 1.3037, "step": 2501 }, { "epoch": 0.06496893257905963, "grad_norm": 5.093719482330457, "learning_rate": 1e-05, "loss": 1.2622, "step": 2502 }, { "epoch": 0.06499489937865158, "grad_norm": 5.272519089586004, "learning_rate": 1e-05, "loss": 1.2424, "step": 2503 }, { "epoch": 0.06502086617824353, "grad_norm": 4.832780784505333, "learning_rate": 1e-05, "loss": 1.2574, "step": 2504 }, { "epoch": 0.06504683297783548, "grad_norm": 5.478216729535863, "learning_rate": 1e-05, "loss": 1.3022, "step": 2505 }, { "epoch": 0.06507279977742743, "grad_norm": 6.097127281768676, "learning_rate": 1e-05, "loss": 1.2472, "step": 2506 }, { "epoch": 0.06509876657701938, "grad_norm": 4.497172845091993, "learning_rate": 1e-05, "loss": 1.2913, "step": 2507 }, { "epoch": 0.06512473337661133, "grad_norm": 4.963889185563837, "learning_rate": 1e-05, "loss": 1.2734, "step": 2508 }, { "epoch": 0.06515070017620328, "grad_norm": 8.497146576290469, "learning_rate": 1e-05, "loss": 1.2363, "step": 2509 }, { "epoch": 0.06517666697579523, "grad_norm": 4.645616512057319, "learning_rate": 1e-05, "loss": 1.2416, "step": 2510 }, { "epoch": 0.06520263377538718, "grad_norm": 5.296234542251512, "learning_rate": 1e-05, "loss": 1.2677, "step": 2511 }, { "epoch": 0.06522860057497913, "grad_norm": 10.231099264051684, "learning_rate": 1e-05, "loss": 1.3021, "step": 2512 }, { "epoch": 0.06525456737457108, "grad_norm": 13.193992426753418, "learning_rate": 1e-05, "loss": 1.2731, "step": 2513 }, { "epoch": 0.06528053417416303, "grad_norm": 7.107852043264921, "learning_rate": 1e-05, "loss": 1.3051, "step": 2514 }, { "epoch": 0.06530650097375498, "grad_norm": 4.817642634863131, "learning_rate": 1e-05, "loss": 1.2578, "step": 2515 }, { "epoch": 0.06533246777334693, "grad_norm": 4.597555647197448, "learning_rate": 1e-05, "loss": 1.3073, "step": 2516 }, { "epoch": 0.06535843457293888, "grad_norm": 9.546560129320632, "learning_rate": 1e-05, "loss": 1.2573, "step": 2517 }, { "epoch": 0.06538440137253083, "grad_norm": 5.578513081191213, "learning_rate": 1e-05, "loss": 1.2897, "step": 2518 }, { "epoch": 0.06541036817212278, "grad_norm": 5.022972548163537, "learning_rate": 1e-05, "loss": 1.2889, "step": 2519 }, { "epoch": 0.06543633497171474, "grad_norm": 9.85036282718299, "learning_rate": 1e-05, "loss": 1.2764, "step": 2520 }, { "epoch": 0.06546230177130669, "grad_norm": 6.798775102583392, "learning_rate": 1e-05, "loss": 1.2685, "step": 2521 }, { "epoch": 0.06548826857089864, "grad_norm": 4.703684827366752, "learning_rate": 1e-05, "loss": 1.2742, "step": 2522 }, { "epoch": 0.06551423537049059, "grad_norm": 5.921979998863015, "learning_rate": 1e-05, "loss": 1.2474, "step": 2523 }, { "epoch": 0.06554020217008254, "grad_norm": 12.215350290304562, "learning_rate": 1e-05, "loss": 1.2387, "step": 2524 }, { "epoch": 0.06556616896967449, "grad_norm": 3.8385583815950453, "learning_rate": 1e-05, "loss": 1.256, "step": 2525 }, { "epoch": 0.06559213576926644, "grad_norm": 4.287449934864514, "learning_rate": 1e-05, "loss": 1.2187, "step": 2526 }, { "epoch": 0.06561810256885839, "grad_norm": 5.211221712931912, "learning_rate": 1e-05, "loss": 1.2322, "step": 2527 }, { "epoch": 0.06564406936845034, "grad_norm": 6.034185930289901, "learning_rate": 1e-05, "loss": 1.2968, "step": 2528 }, { "epoch": 0.06567003616804229, "grad_norm": 8.570049438333964, "learning_rate": 1e-05, "loss": 1.2391, "step": 2529 }, { "epoch": 0.06569600296763424, "grad_norm": 4.829377737142015, "learning_rate": 1e-05, "loss": 1.2622, "step": 2530 }, { "epoch": 0.06572196976722619, "grad_norm": 16.10721239879097, "learning_rate": 1e-05, "loss": 1.2105, "step": 2531 }, { "epoch": 0.06574793656681814, "grad_norm": 5.2746177426362015, "learning_rate": 1e-05, "loss": 1.2737, "step": 2532 }, { "epoch": 0.06577390336641009, "grad_norm": 4.920359157075746, "learning_rate": 1e-05, "loss": 1.2571, "step": 2533 }, { "epoch": 0.06579987016600204, "grad_norm": 5.603331984671651, "learning_rate": 1e-05, "loss": 1.2773, "step": 2534 }, { "epoch": 0.06582583696559399, "grad_norm": 6.038499971274006, "learning_rate": 1e-05, "loss": 1.2721, "step": 2535 }, { "epoch": 0.06585180376518594, "grad_norm": 5.400549380295378, "learning_rate": 1e-05, "loss": 1.3496, "step": 2536 }, { "epoch": 0.0658777705647779, "grad_norm": 6.300680747416058, "learning_rate": 1e-05, "loss": 1.3086, "step": 2537 }, { "epoch": 0.06590373736436984, "grad_norm": 9.80704225022784, "learning_rate": 1e-05, "loss": 1.2931, "step": 2538 }, { "epoch": 0.0659297041639618, "grad_norm": 10.670507295913744, "learning_rate": 1e-05, "loss": 1.2624, "step": 2539 }, { "epoch": 0.06595567096355374, "grad_norm": 5.603912328836188, "learning_rate": 1e-05, "loss": 1.2486, "step": 2540 }, { "epoch": 0.0659816377631457, "grad_norm": 5.217851572881418, "learning_rate": 1e-05, "loss": 1.298, "step": 2541 }, { "epoch": 0.06600760456273765, "grad_norm": 9.806244040415056, "learning_rate": 1e-05, "loss": 1.2602, "step": 2542 }, { "epoch": 0.0660335713623296, "grad_norm": 9.49969883491585, "learning_rate": 1e-05, "loss": 1.2182, "step": 2543 }, { "epoch": 0.06605953816192155, "grad_norm": 7.126954998372596, "learning_rate": 1e-05, "loss": 1.2469, "step": 2544 }, { "epoch": 0.0660855049615135, "grad_norm": 7.601666880064448, "learning_rate": 1e-05, "loss": 1.3032, "step": 2545 }, { "epoch": 0.06611147176110545, "grad_norm": 6.532993211780935, "learning_rate": 1e-05, "loss": 1.2437, "step": 2546 }, { "epoch": 0.0661374385606974, "grad_norm": 5.384807067645202, "learning_rate": 1e-05, "loss": 1.2852, "step": 2547 }, { "epoch": 0.06616340536028935, "grad_norm": 4.868269945406945, "learning_rate": 1e-05, "loss": 1.239, "step": 2548 }, { "epoch": 0.0661893721598813, "grad_norm": 7.0488029600028765, "learning_rate": 1e-05, "loss": 1.255, "step": 2549 }, { "epoch": 0.06621533895947325, "grad_norm": 5.763744347329016, "learning_rate": 1e-05, "loss": 1.253, "step": 2550 }, { "epoch": 0.0662413057590652, "grad_norm": 6.191904985334651, "learning_rate": 1e-05, "loss": 1.2482, "step": 2551 }, { "epoch": 0.06626727255865715, "grad_norm": 6.557217460941655, "learning_rate": 1e-05, "loss": 1.2784, "step": 2552 }, { "epoch": 0.0662932393582491, "grad_norm": 13.520668879107888, "learning_rate": 1e-05, "loss": 1.2545, "step": 2553 }, { "epoch": 0.06631920615784105, "grad_norm": 5.902085869821456, "learning_rate": 1e-05, "loss": 1.2696, "step": 2554 }, { "epoch": 0.066345172957433, "grad_norm": 5.197445814355746, "learning_rate": 1e-05, "loss": 1.3069, "step": 2555 }, { "epoch": 0.06637113975702495, "grad_norm": 8.648218277906972, "learning_rate": 1e-05, "loss": 1.2675, "step": 2556 }, { "epoch": 0.0663971065566169, "grad_norm": 5.420457814155467, "learning_rate": 1e-05, "loss": 1.3153, "step": 2557 }, { "epoch": 0.06642307335620885, "grad_norm": 6.679660658754551, "learning_rate": 1e-05, "loss": 1.2886, "step": 2558 }, { "epoch": 0.0664490401558008, "grad_norm": 72.77638924837677, "learning_rate": 1e-05, "loss": 1.3631, "step": 2559 }, { "epoch": 0.06647500695539275, "grad_norm": 7.582650569599296, "learning_rate": 1e-05, "loss": 1.2705, "step": 2560 }, { "epoch": 0.0665009737549847, "grad_norm": 7.364343994510469, "learning_rate": 1e-05, "loss": 1.2698, "step": 2561 }, { "epoch": 0.06652694055457666, "grad_norm": 9.957642684247535, "learning_rate": 1e-05, "loss": 1.2731, "step": 2562 }, { "epoch": 0.0665529073541686, "grad_norm": 4.611558920526792, "learning_rate": 1e-05, "loss": 1.228, "step": 2563 }, { "epoch": 0.06657887415376054, "grad_norm": 5.4795970481801675, "learning_rate": 1e-05, "loss": 1.262, "step": 2564 }, { "epoch": 0.06660484095335249, "grad_norm": 5.694567303249672, "learning_rate": 1e-05, "loss": 1.2901, "step": 2565 }, { "epoch": 0.06663080775294444, "grad_norm": 7.063433290104041, "learning_rate": 1e-05, "loss": 1.2724, "step": 2566 }, { "epoch": 0.0666567745525364, "grad_norm": 35.5400766163839, "learning_rate": 1e-05, "loss": 1.2519, "step": 2567 }, { "epoch": 0.06668274135212834, "grad_norm": 6.850158756393385, "learning_rate": 1e-05, "loss": 1.2637, "step": 2568 }, { "epoch": 0.0667087081517203, "grad_norm": 12.468182773164623, "learning_rate": 1e-05, "loss": 1.3001, "step": 2569 }, { "epoch": 0.06673467495131225, "grad_norm": 15.164095706107384, "learning_rate": 1e-05, "loss": 1.2508, "step": 2570 }, { "epoch": 0.0667606417509042, "grad_norm": 5.098429209836579, "learning_rate": 1e-05, "loss": 1.3053, "step": 2571 }, { "epoch": 0.06678660855049615, "grad_norm": 6.857357623506314, "learning_rate": 1e-05, "loss": 1.3059, "step": 2572 }, { "epoch": 0.0668125753500881, "grad_norm": 6.0452383423824525, "learning_rate": 1e-05, "loss": 1.2654, "step": 2573 }, { "epoch": 0.06683854214968005, "grad_norm": 6.100134375921311, "learning_rate": 1e-05, "loss": 1.263, "step": 2574 }, { "epoch": 0.066864508949272, "grad_norm": 5.454193622628383, "learning_rate": 1e-05, "loss": 1.2886, "step": 2575 }, { "epoch": 0.06689047574886395, "grad_norm": 4.937342580869212, "learning_rate": 1e-05, "loss": 1.2666, "step": 2576 }, { "epoch": 0.0669164425484559, "grad_norm": 6.476389331596276, "learning_rate": 1e-05, "loss": 1.3008, "step": 2577 }, { "epoch": 0.06694240934804785, "grad_norm": 5.074496337251829, "learning_rate": 1e-05, "loss": 1.3119, "step": 2578 }, { "epoch": 0.0669683761476398, "grad_norm": 5.092718101778712, "learning_rate": 1e-05, "loss": 1.2994, "step": 2579 }, { "epoch": 0.06699434294723175, "grad_norm": 15.30089842707876, "learning_rate": 1e-05, "loss": 1.2673, "step": 2580 }, { "epoch": 0.0670203097468237, "grad_norm": 13.01638582897834, "learning_rate": 1e-05, "loss": 1.2953, "step": 2581 }, { "epoch": 0.06704627654641565, "grad_norm": 6.026400657342034, "learning_rate": 1e-05, "loss": 1.3015, "step": 2582 }, { "epoch": 0.0670722433460076, "grad_norm": 6.320320949855291, "learning_rate": 1e-05, "loss": 1.2628, "step": 2583 }, { "epoch": 0.06709821014559955, "grad_norm": 8.936494784079432, "learning_rate": 1e-05, "loss": 1.2475, "step": 2584 }, { "epoch": 0.0671241769451915, "grad_norm": 5.174050786781035, "learning_rate": 1e-05, "loss": 1.2567, "step": 2585 }, { "epoch": 0.06715014374478345, "grad_norm": 4.762257671448749, "learning_rate": 1e-05, "loss": 1.2007, "step": 2586 }, { "epoch": 0.0671761105443754, "grad_norm": 4.622624379986683, "learning_rate": 1e-05, "loss": 1.3135, "step": 2587 }, { "epoch": 0.06720207734396735, "grad_norm": 5.479968961084827, "learning_rate": 1e-05, "loss": 1.2437, "step": 2588 }, { "epoch": 0.0672280441435593, "grad_norm": 5.319320250085873, "learning_rate": 1e-05, "loss": 1.241, "step": 2589 }, { "epoch": 0.06725401094315125, "grad_norm": 5.588375130548364, "learning_rate": 1e-05, "loss": 1.2915, "step": 2590 }, { "epoch": 0.0672799777427432, "grad_norm": 7.456199155955498, "learning_rate": 1e-05, "loss": 1.2839, "step": 2591 }, { "epoch": 0.06730594454233516, "grad_norm": 7.4276257531335865, "learning_rate": 1e-05, "loss": 1.3281, "step": 2592 }, { "epoch": 0.0673319113419271, "grad_norm": 5.738420852199072, "learning_rate": 1e-05, "loss": 1.2469, "step": 2593 }, { "epoch": 0.06735787814151906, "grad_norm": 4.71108345577432, "learning_rate": 1e-05, "loss": 1.2501, "step": 2594 }, { "epoch": 0.067383844941111, "grad_norm": 5.750282280668754, "learning_rate": 1e-05, "loss": 1.3136, "step": 2595 }, { "epoch": 0.06740981174070296, "grad_norm": 4.563776961042957, "learning_rate": 1e-05, "loss": 1.2596, "step": 2596 }, { "epoch": 0.06743577854029491, "grad_norm": 5.786980505547088, "learning_rate": 1e-05, "loss": 1.2559, "step": 2597 }, { "epoch": 0.06746174533988686, "grad_norm": 3.919825288032145, "learning_rate": 1e-05, "loss": 1.2527, "step": 2598 }, { "epoch": 0.06748771213947881, "grad_norm": 22.818955101276273, "learning_rate": 1e-05, "loss": 1.2527, "step": 2599 }, { "epoch": 0.06751367893907076, "grad_norm": 4.837043302993996, "learning_rate": 1e-05, "loss": 1.2533, "step": 2600 }, { "epoch": 0.06753964573866271, "grad_norm": 6.981204548604982, "learning_rate": 1e-05, "loss": 1.27, "step": 2601 }, { "epoch": 0.06756561253825466, "grad_norm": 5.624259900041762, "learning_rate": 1e-05, "loss": 1.2964, "step": 2602 }, { "epoch": 0.06759157933784661, "grad_norm": 4.506454924475349, "learning_rate": 1e-05, "loss": 1.284, "step": 2603 }, { "epoch": 0.06761754613743856, "grad_norm": 10.5882921591335, "learning_rate": 1e-05, "loss": 1.2885, "step": 2604 }, { "epoch": 0.06764351293703051, "grad_norm": 4.625276093361275, "learning_rate": 1e-05, "loss": 1.3171, "step": 2605 }, { "epoch": 0.06766947973662246, "grad_norm": 10.979402504595146, "learning_rate": 1e-05, "loss": 1.3131, "step": 2606 }, { "epoch": 0.06769544653621441, "grad_norm": 6.486503968558939, "learning_rate": 1e-05, "loss": 1.3188, "step": 2607 }, { "epoch": 0.06772141333580636, "grad_norm": 5.333187617854478, "learning_rate": 1e-05, "loss": 1.2822, "step": 2608 }, { "epoch": 0.06774738013539831, "grad_norm": 5.746746635559162, "learning_rate": 1e-05, "loss": 1.2665, "step": 2609 }, { "epoch": 0.06777334693499026, "grad_norm": 8.152707320744124, "learning_rate": 1e-05, "loss": 1.212, "step": 2610 }, { "epoch": 0.06779931373458221, "grad_norm": 8.288756455677955, "learning_rate": 1e-05, "loss": 1.2595, "step": 2611 }, { "epoch": 0.06782528053417416, "grad_norm": 12.370377727344767, "learning_rate": 1e-05, "loss": 1.2954, "step": 2612 }, { "epoch": 0.06785124733376612, "grad_norm": 6.118063599659018, "learning_rate": 1e-05, "loss": 1.2745, "step": 2613 }, { "epoch": 0.06787721413335807, "grad_norm": 6.504083084589267, "learning_rate": 1e-05, "loss": 1.285, "step": 2614 }, { "epoch": 0.06790318093295002, "grad_norm": 6.5999723607264835, "learning_rate": 1e-05, "loss": 1.2789, "step": 2615 }, { "epoch": 0.06792914773254197, "grad_norm": 5.04490618494705, "learning_rate": 1e-05, "loss": 1.2798, "step": 2616 }, { "epoch": 0.06795511453213392, "grad_norm": 7.488604980857661, "learning_rate": 1e-05, "loss": 1.2331, "step": 2617 }, { "epoch": 0.06798108133172587, "grad_norm": 7.725964591533468, "learning_rate": 1e-05, "loss": 1.3457, "step": 2618 }, { "epoch": 0.06800704813131782, "grad_norm": 6.397135880465924, "learning_rate": 1e-05, "loss": 1.2719, "step": 2619 }, { "epoch": 0.06803301493090977, "grad_norm": 5.864527357070318, "learning_rate": 1e-05, "loss": 1.2978, "step": 2620 }, { "epoch": 0.06805898173050172, "grad_norm": 6.24425303406013, "learning_rate": 1e-05, "loss": 1.2823, "step": 2621 }, { "epoch": 0.06808494853009367, "grad_norm": 6.498011651730504, "learning_rate": 1e-05, "loss": 1.2569, "step": 2622 }, { "epoch": 0.06811091532968562, "grad_norm": 19.458062877313868, "learning_rate": 1e-05, "loss": 1.2521, "step": 2623 }, { "epoch": 0.06813688212927757, "grad_norm": 8.352781515483938, "learning_rate": 1e-05, "loss": 1.2164, "step": 2624 }, { "epoch": 0.06816284892886952, "grad_norm": 5.715295634261776, "learning_rate": 1e-05, "loss": 1.2695, "step": 2625 }, { "epoch": 0.06818881572846147, "grad_norm": 14.824964469879957, "learning_rate": 1e-05, "loss": 1.233, "step": 2626 }, { "epoch": 0.06821478252805342, "grad_norm": 24.396665051293382, "learning_rate": 1e-05, "loss": 1.2633, "step": 2627 }, { "epoch": 0.06824074932764537, "grad_norm": 6.504992401798863, "learning_rate": 1e-05, "loss": 1.2673, "step": 2628 }, { "epoch": 0.06826671612723732, "grad_norm": 11.604934931765357, "learning_rate": 1e-05, "loss": 1.1967, "step": 2629 }, { "epoch": 0.06829268292682927, "grad_norm": 5.525578063827632, "learning_rate": 1e-05, "loss": 1.234, "step": 2630 }, { "epoch": 0.06831864972642122, "grad_norm": 11.561048261903524, "learning_rate": 1e-05, "loss": 1.2812, "step": 2631 }, { "epoch": 0.06834461652601317, "grad_norm": 10.102989101941331, "learning_rate": 1e-05, "loss": 1.274, "step": 2632 }, { "epoch": 0.06837058332560512, "grad_norm": 12.19903863339655, "learning_rate": 1e-05, "loss": 1.26, "step": 2633 }, { "epoch": 0.06839655012519708, "grad_norm": 5.105523106957671, "learning_rate": 1e-05, "loss": 1.2713, "step": 2634 }, { "epoch": 0.06842251692478903, "grad_norm": 6.967880673776236, "learning_rate": 1e-05, "loss": 1.2677, "step": 2635 }, { "epoch": 0.06844848372438098, "grad_norm": 21.96816794436967, "learning_rate": 1e-05, "loss": 1.2631, "step": 2636 }, { "epoch": 0.06847445052397293, "grad_norm": 6.950122244878143, "learning_rate": 1e-05, "loss": 1.269, "step": 2637 }, { "epoch": 0.06850041732356488, "grad_norm": 6.516154828015966, "learning_rate": 1e-05, "loss": 1.3109, "step": 2638 }, { "epoch": 0.06852638412315683, "grad_norm": 5.698071056368789, "learning_rate": 1e-05, "loss": 1.2174, "step": 2639 }, { "epoch": 0.06855235092274876, "grad_norm": 37.10807070076282, "learning_rate": 1e-05, "loss": 1.3047, "step": 2640 }, { "epoch": 0.06857831772234071, "grad_norm": 10.042384162901406, "learning_rate": 1e-05, "loss": 1.2987, "step": 2641 }, { "epoch": 0.06860428452193267, "grad_norm": 5.6708283943119255, "learning_rate": 1e-05, "loss": 1.271, "step": 2642 }, { "epoch": 0.06863025132152462, "grad_norm": 6.435780119540312, "learning_rate": 1e-05, "loss": 1.2651, "step": 2643 }, { "epoch": 0.06865621812111657, "grad_norm": 5.46919920030146, "learning_rate": 1e-05, "loss": 1.2681, "step": 2644 }, { "epoch": 0.06868218492070852, "grad_norm": 11.138117910727185, "learning_rate": 1e-05, "loss": 1.2273, "step": 2645 }, { "epoch": 0.06870815172030047, "grad_norm": 6.564315399477569, "learning_rate": 1e-05, "loss": 1.2793, "step": 2646 }, { "epoch": 0.06873411851989242, "grad_norm": 5.330705571522291, "learning_rate": 1e-05, "loss": 1.2011, "step": 2647 }, { "epoch": 0.06876008531948437, "grad_norm": 7.187088929736124, "learning_rate": 1e-05, "loss": 1.2689, "step": 2648 }, { "epoch": 0.06878605211907632, "grad_norm": 72.40215987045035, "learning_rate": 1e-05, "loss": 1.3284, "step": 2649 }, { "epoch": 0.06881201891866827, "grad_norm": 8.922208923397497, "learning_rate": 1e-05, "loss": 1.2733, "step": 2650 }, { "epoch": 0.06883798571826022, "grad_norm": 6.009293035224953, "learning_rate": 1e-05, "loss": 1.2875, "step": 2651 }, { "epoch": 0.06886395251785217, "grad_norm": 5.936998526828885, "learning_rate": 1e-05, "loss": 1.3207, "step": 2652 }, { "epoch": 0.06888991931744412, "grad_norm": 11.850013546191493, "learning_rate": 1e-05, "loss": 1.3048, "step": 2653 }, { "epoch": 0.06891588611703607, "grad_norm": 9.428428483680353, "learning_rate": 1e-05, "loss": 1.2953, "step": 2654 }, { "epoch": 0.06894185291662802, "grad_norm": 8.656458001787957, "learning_rate": 1e-05, "loss": 1.2534, "step": 2655 }, { "epoch": 0.06896781971621997, "grad_norm": 6.863145908330465, "learning_rate": 1e-05, "loss": 1.2613, "step": 2656 }, { "epoch": 0.06899378651581192, "grad_norm": 5.5814209504981935, "learning_rate": 1e-05, "loss": 1.2519, "step": 2657 }, { "epoch": 0.06901975331540387, "grad_norm": 6.96455181577601, "learning_rate": 1e-05, "loss": 1.2783, "step": 2658 }, { "epoch": 0.06904572011499582, "grad_norm": 6.905116475840352, "learning_rate": 1e-05, "loss": 1.2657, "step": 2659 }, { "epoch": 0.06907168691458777, "grad_norm": 7.055005984124688, "learning_rate": 1e-05, "loss": 1.3016, "step": 2660 }, { "epoch": 0.06909765371417972, "grad_norm": 24.063104953840252, "learning_rate": 1e-05, "loss": 1.2652, "step": 2661 }, { "epoch": 0.06912362051377167, "grad_norm": 7.302310463007207, "learning_rate": 1e-05, "loss": 1.2699, "step": 2662 }, { "epoch": 0.06914958731336363, "grad_norm": 19.09246811879458, "learning_rate": 1e-05, "loss": 1.238, "step": 2663 }, { "epoch": 0.06917555411295558, "grad_norm": 8.18424305672926, "learning_rate": 1e-05, "loss": 1.2584, "step": 2664 }, { "epoch": 0.06920152091254753, "grad_norm": 8.453074456872924, "learning_rate": 1e-05, "loss": 1.2449, "step": 2665 }, { "epoch": 0.06922748771213948, "grad_norm": 5.853116123337336, "learning_rate": 1e-05, "loss": 1.2452, "step": 2666 }, { "epoch": 0.06925345451173143, "grad_norm": 18.86646284840262, "learning_rate": 1e-05, "loss": 1.2828, "step": 2667 }, { "epoch": 0.06927942131132338, "grad_norm": 7.559240726510046, "learning_rate": 1e-05, "loss": 1.3108, "step": 2668 }, { "epoch": 0.06930538811091533, "grad_norm": 8.390028039023795, "learning_rate": 1e-05, "loss": 1.2445, "step": 2669 }, { "epoch": 0.06933135491050728, "grad_norm": 13.31825039023637, "learning_rate": 1e-05, "loss": 1.2486, "step": 2670 }, { "epoch": 0.06935732171009923, "grad_norm": 13.478119814911766, "learning_rate": 1e-05, "loss": 1.2778, "step": 2671 }, { "epoch": 0.06938328850969118, "grad_norm": 4.829952745858046, "learning_rate": 1e-05, "loss": 1.2068, "step": 2672 }, { "epoch": 0.06940925530928313, "grad_norm": 7.682609925655173, "learning_rate": 1e-05, "loss": 1.2464, "step": 2673 }, { "epoch": 0.06943522210887508, "grad_norm": 7.7213597383956145, "learning_rate": 1e-05, "loss": 1.2345, "step": 2674 }, { "epoch": 0.06946118890846703, "grad_norm": 7.793905361557331, "learning_rate": 1e-05, "loss": 1.2748, "step": 2675 }, { "epoch": 0.06948715570805898, "grad_norm": 9.118624615607942, "learning_rate": 1e-05, "loss": 1.2394, "step": 2676 }, { "epoch": 0.06951312250765093, "grad_norm": 9.151850662266819, "learning_rate": 1e-05, "loss": 1.2526, "step": 2677 }, { "epoch": 0.06953908930724288, "grad_norm": 5.405007809501885, "learning_rate": 1e-05, "loss": 1.2545, "step": 2678 }, { "epoch": 0.06956505610683483, "grad_norm": 14.500813428490508, "learning_rate": 1e-05, "loss": 1.2686, "step": 2679 }, { "epoch": 0.06959102290642678, "grad_norm": 6.0784535454062745, "learning_rate": 1e-05, "loss": 1.2349, "step": 2680 }, { "epoch": 0.06961698970601873, "grad_norm": 7.912010295834221, "learning_rate": 1e-05, "loss": 1.2407, "step": 2681 }, { "epoch": 0.06964295650561068, "grad_norm": 6.585397356538621, "learning_rate": 1e-05, "loss": 1.2756, "step": 2682 }, { "epoch": 0.06966892330520263, "grad_norm": 28.01388314614347, "learning_rate": 1e-05, "loss": 1.294, "step": 2683 }, { "epoch": 0.06969489010479459, "grad_norm": 7.312171080927199, "learning_rate": 1e-05, "loss": 1.2836, "step": 2684 }, { "epoch": 0.06972085690438654, "grad_norm": 7.771132421660966, "learning_rate": 1e-05, "loss": 1.2645, "step": 2685 }, { "epoch": 0.06974682370397849, "grad_norm": 5.542141466188859, "learning_rate": 1e-05, "loss": 1.3142, "step": 2686 }, { "epoch": 0.06977279050357044, "grad_norm": 17.017919632641554, "learning_rate": 1e-05, "loss": 1.287, "step": 2687 }, { "epoch": 0.06979875730316239, "grad_norm": 5.263379079286936, "learning_rate": 1e-05, "loss": 1.2858, "step": 2688 }, { "epoch": 0.06982472410275434, "grad_norm": 7.175379875308377, "learning_rate": 1e-05, "loss": 1.2748, "step": 2689 }, { "epoch": 0.06985069090234629, "grad_norm": 10.672483368228653, "learning_rate": 1e-05, "loss": 1.2561, "step": 2690 }, { "epoch": 0.06987665770193824, "grad_norm": 8.834762697514106, "learning_rate": 1e-05, "loss": 1.2822, "step": 2691 }, { "epoch": 0.06990262450153019, "grad_norm": 11.17436700517343, "learning_rate": 1e-05, "loss": 1.3203, "step": 2692 }, { "epoch": 0.06992859130112214, "grad_norm": 5.866800308150337, "learning_rate": 1e-05, "loss": 1.2759, "step": 2693 }, { "epoch": 0.06995455810071409, "grad_norm": 9.316386186430178, "learning_rate": 1e-05, "loss": 1.2579, "step": 2694 }, { "epoch": 0.06998052490030604, "grad_norm": 9.053985594828788, "learning_rate": 1e-05, "loss": 1.2573, "step": 2695 }, { "epoch": 0.07000649169989799, "grad_norm": 19.062101366221242, "learning_rate": 1e-05, "loss": 1.2667, "step": 2696 }, { "epoch": 0.07003245849948994, "grad_norm": 7.28323167021424, "learning_rate": 1e-05, "loss": 1.2508, "step": 2697 }, { "epoch": 0.07005842529908189, "grad_norm": 6.467527412206053, "learning_rate": 1e-05, "loss": 1.2879, "step": 2698 }, { "epoch": 0.07008439209867384, "grad_norm": 10.283552010641381, "learning_rate": 1e-05, "loss": 1.293, "step": 2699 }, { "epoch": 0.07011035889826579, "grad_norm": 6.987790222146105, "learning_rate": 1e-05, "loss": 1.2184, "step": 2700 }, { "epoch": 0.07013632569785774, "grad_norm": 6.649827894694173, "learning_rate": 1e-05, "loss": 1.2631, "step": 2701 }, { "epoch": 0.0701622924974497, "grad_norm": 9.295977586019943, "learning_rate": 1e-05, "loss": 1.2066, "step": 2702 }, { "epoch": 0.07018825929704164, "grad_norm": 7.549177556999794, "learning_rate": 1e-05, "loss": 1.2771, "step": 2703 }, { "epoch": 0.0702142260966336, "grad_norm": 14.703934455667385, "learning_rate": 1e-05, "loss": 1.2463, "step": 2704 }, { "epoch": 0.07024019289622555, "grad_norm": 6.959305868503462, "learning_rate": 1e-05, "loss": 1.2394, "step": 2705 }, { "epoch": 0.0702661596958175, "grad_norm": 13.804043075295345, "learning_rate": 1e-05, "loss": 1.2683, "step": 2706 }, { "epoch": 0.07029212649540945, "grad_norm": 13.841163410834715, "learning_rate": 1e-05, "loss": 1.2961, "step": 2707 }, { "epoch": 0.0703180932950014, "grad_norm": 11.2523946968445, "learning_rate": 1e-05, "loss": 1.2374, "step": 2708 }, { "epoch": 0.07034406009459335, "grad_norm": 5.629736876767487, "learning_rate": 1e-05, "loss": 1.2437, "step": 2709 }, { "epoch": 0.0703700268941853, "grad_norm": 12.144860430968127, "learning_rate": 1e-05, "loss": 1.2947, "step": 2710 }, { "epoch": 0.07039599369377725, "grad_norm": 4.283496350898315, "learning_rate": 1e-05, "loss": 1.2285, "step": 2711 }, { "epoch": 0.0704219604933692, "grad_norm": 6.992125578025394, "learning_rate": 1e-05, "loss": 1.2701, "step": 2712 }, { "epoch": 0.07044792729296115, "grad_norm": 7.4307084049020276, "learning_rate": 1e-05, "loss": 1.281, "step": 2713 }, { "epoch": 0.0704738940925531, "grad_norm": 19.64980891088702, "learning_rate": 1e-05, "loss": 1.2913, "step": 2714 }, { "epoch": 0.07049986089214505, "grad_norm": 6.650250377100474, "learning_rate": 1e-05, "loss": 1.2574, "step": 2715 }, { "epoch": 0.070525827691737, "grad_norm": 6.688222970832264, "learning_rate": 1e-05, "loss": 1.2699, "step": 2716 }, { "epoch": 0.07055179449132894, "grad_norm": 7.611832051655973, "learning_rate": 1e-05, "loss": 1.2712, "step": 2717 }, { "epoch": 0.07057776129092089, "grad_norm": 7.236134127797696, "learning_rate": 1e-05, "loss": 1.2999, "step": 2718 }, { "epoch": 0.07060372809051284, "grad_norm": 15.065085541349045, "learning_rate": 1e-05, "loss": 1.2737, "step": 2719 }, { "epoch": 0.07062969489010479, "grad_norm": 9.070811281774652, "learning_rate": 1e-05, "loss": 1.2705, "step": 2720 }, { "epoch": 0.07065566168969674, "grad_norm": 6.248107318400978, "learning_rate": 1e-05, "loss": 1.2589, "step": 2721 }, { "epoch": 0.07068162848928869, "grad_norm": 6.8717030335854234, "learning_rate": 1e-05, "loss": 1.2971, "step": 2722 }, { "epoch": 0.07070759528888064, "grad_norm": 6.349181307183184, "learning_rate": 1e-05, "loss": 1.2969, "step": 2723 }, { "epoch": 0.07073356208847259, "grad_norm": 13.359152844461347, "learning_rate": 1e-05, "loss": 1.2989, "step": 2724 }, { "epoch": 0.07075952888806454, "grad_norm": 16.310805894130908, "learning_rate": 1e-05, "loss": 1.2565, "step": 2725 }, { "epoch": 0.07078549568765649, "grad_norm": 7.202769372067751, "learning_rate": 1e-05, "loss": 1.2455, "step": 2726 }, { "epoch": 0.07081146248724844, "grad_norm": 9.449211079601062, "learning_rate": 1e-05, "loss": 1.2518, "step": 2727 }, { "epoch": 0.07083742928684039, "grad_norm": 10.852061315816737, "learning_rate": 1e-05, "loss": 1.2475, "step": 2728 }, { "epoch": 0.07086339608643234, "grad_norm": 5.844965354483907, "learning_rate": 1e-05, "loss": 1.2375, "step": 2729 }, { "epoch": 0.07088936288602429, "grad_norm": 9.948946136945352, "learning_rate": 1e-05, "loss": 1.2863, "step": 2730 }, { "epoch": 0.07091532968561624, "grad_norm": 4.9117979523547035, "learning_rate": 1e-05, "loss": 1.2346, "step": 2731 }, { "epoch": 0.0709412964852082, "grad_norm": 10.99985122580056, "learning_rate": 1e-05, "loss": 1.2779, "step": 2732 }, { "epoch": 0.07096726328480014, "grad_norm": 6.669435148785647, "learning_rate": 1e-05, "loss": 1.2456, "step": 2733 }, { "epoch": 0.0709932300843921, "grad_norm": 13.035512688975954, "learning_rate": 1e-05, "loss": 1.2672, "step": 2734 }, { "epoch": 0.07101919688398405, "grad_norm": 8.433267619765308, "learning_rate": 1e-05, "loss": 1.2549, "step": 2735 }, { "epoch": 0.071045163683576, "grad_norm": 22.472251905692286, "learning_rate": 1e-05, "loss": 1.2823, "step": 2736 }, { "epoch": 0.07107113048316795, "grad_norm": 9.040532288535813, "learning_rate": 1e-05, "loss": 1.2502, "step": 2737 }, { "epoch": 0.0710970972827599, "grad_norm": 8.496326325839672, "learning_rate": 1e-05, "loss": 1.2846, "step": 2738 }, { "epoch": 0.07112306408235185, "grad_norm": 8.010331158787226, "learning_rate": 1e-05, "loss": 1.2596, "step": 2739 }, { "epoch": 0.0711490308819438, "grad_norm": 11.173687981171925, "learning_rate": 1e-05, "loss": 1.2947, "step": 2740 }, { "epoch": 0.07117499768153575, "grad_norm": 9.331745103398651, "learning_rate": 1e-05, "loss": 1.2635, "step": 2741 }, { "epoch": 0.0712009644811277, "grad_norm": 7.452808941231618, "learning_rate": 1e-05, "loss": 1.2618, "step": 2742 }, { "epoch": 0.07122693128071965, "grad_norm": 5.517182819080339, "learning_rate": 1e-05, "loss": 1.2606, "step": 2743 }, { "epoch": 0.0712528980803116, "grad_norm": 10.83479119297241, "learning_rate": 1e-05, "loss": 1.2458, "step": 2744 }, { "epoch": 0.07127886487990355, "grad_norm": 7.713168578018988, "learning_rate": 1e-05, "loss": 1.3356, "step": 2745 }, { "epoch": 0.0713048316794955, "grad_norm": 10.528484400171205, "learning_rate": 1e-05, "loss": 1.25, "step": 2746 }, { "epoch": 0.07133079847908745, "grad_norm": 7.879556215831401, "learning_rate": 1e-05, "loss": 1.282, "step": 2747 }, { "epoch": 0.0713567652786794, "grad_norm": 5.951406796622687, "learning_rate": 1e-05, "loss": 1.3102, "step": 2748 }, { "epoch": 0.07138273207827135, "grad_norm": 15.628833025946937, "learning_rate": 1e-05, "loss": 1.2715, "step": 2749 }, { "epoch": 0.0714086988778633, "grad_norm": 6.890383623181613, "learning_rate": 1e-05, "loss": 1.2632, "step": 2750 }, { "epoch": 0.07143466567745525, "grad_norm": 12.515481745629902, "learning_rate": 1e-05, "loss": 1.2547, "step": 2751 }, { "epoch": 0.0714606324770472, "grad_norm": 6.968745347094159, "learning_rate": 1e-05, "loss": 1.2737, "step": 2752 }, { "epoch": 0.07148659927663915, "grad_norm": 6.809214508420166, "learning_rate": 1e-05, "loss": 1.2352, "step": 2753 }, { "epoch": 0.0715125660762311, "grad_norm": 7.8264655450111125, "learning_rate": 1e-05, "loss": 1.2278, "step": 2754 }, { "epoch": 0.07153853287582305, "grad_norm": 6.72867578501329, "learning_rate": 1e-05, "loss": 1.2907, "step": 2755 }, { "epoch": 0.071564499675415, "grad_norm": 8.428701376663707, "learning_rate": 1e-05, "loss": 1.2895, "step": 2756 }, { "epoch": 0.07159046647500696, "grad_norm": 7.536165145165998, "learning_rate": 1e-05, "loss": 1.2846, "step": 2757 }, { "epoch": 0.0716164332745989, "grad_norm": 7.9518929297225505, "learning_rate": 1e-05, "loss": 1.2567, "step": 2758 }, { "epoch": 0.07164240007419086, "grad_norm": 6.223200906503427, "learning_rate": 1e-05, "loss": 1.2932, "step": 2759 }, { "epoch": 0.07166836687378281, "grad_norm": 7.593418694949373, "learning_rate": 1e-05, "loss": 1.2685, "step": 2760 }, { "epoch": 0.07169433367337476, "grad_norm": 10.56630119229196, "learning_rate": 1e-05, "loss": 1.2346, "step": 2761 }, { "epoch": 0.07172030047296671, "grad_norm": 12.212624343660742, "learning_rate": 1e-05, "loss": 1.2822, "step": 2762 }, { "epoch": 0.07174626727255866, "grad_norm": 6.627118041990837, "learning_rate": 1e-05, "loss": 1.3062, "step": 2763 }, { "epoch": 0.07177223407215061, "grad_norm": 7.937883262692812, "learning_rate": 1e-05, "loss": 1.2469, "step": 2764 }, { "epoch": 0.07179820087174256, "grad_norm": 5.596444104517639, "learning_rate": 1e-05, "loss": 1.2523, "step": 2765 }, { "epoch": 0.07182416767133451, "grad_norm": 6.727538708108026, "learning_rate": 1e-05, "loss": 1.2264, "step": 2766 }, { "epoch": 0.07185013447092646, "grad_norm": 6.802321923955739, "learning_rate": 1e-05, "loss": 1.2659, "step": 2767 }, { "epoch": 0.07187610127051841, "grad_norm": 5.156344881051755, "learning_rate": 1e-05, "loss": 1.3262, "step": 2768 }, { "epoch": 0.07190206807011036, "grad_norm": 6.393864563692126, "learning_rate": 1e-05, "loss": 1.2514, "step": 2769 }, { "epoch": 0.07192803486970231, "grad_norm": 9.955128514501174, "learning_rate": 1e-05, "loss": 1.2133, "step": 2770 }, { "epoch": 0.07195400166929426, "grad_norm": 7.939214326052058, "learning_rate": 1e-05, "loss": 1.3213, "step": 2771 }, { "epoch": 0.07197996846888621, "grad_norm": 5.7947596060234945, "learning_rate": 1e-05, "loss": 1.2872, "step": 2772 }, { "epoch": 0.07200593526847816, "grad_norm": 9.669383971544422, "learning_rate": 1e-05, "loss": 1.2864, "step": 2773 }, { "epoch": 0.07203190206807011, "grad_norm": 9.11516385800075, "learning_rate": 1e-05, "loss": 1.2475, "step": 2774 }, { "epoch": 0.07205786886766206, "grad_norm": 5.34449629006162, "learning_rate": 1e-05, "loss": 1.2854, "step": 2775 }, { "epoch": 0.07208383566725401, "grad_norm": 22.606716461051803, "learning_rate": 1e-05, "loss": 1.2757, "step": 2776 }, { "epoch": 0.07210980246684597, "grad_norm": 7.712322817951218, "learning_rate": 1e-05, "loss": 1.2692, "step": 2777 }, { "epoch": 0.07213576926643792, "grad_norm": 5.438129958569312, "learning_rate": 1e-05, "loss": 1.2586, "step": 2778 }, { "epoch": 0.07216173606602987, "grad_norm": 9.735994501242422, "learning_rate": 1e-05, "loss": 1.2679, "step": 2779 }, { "epoch": 0.07218770286562182, "grad_norm": 8.746487593717308, "learning_rate": 1e-05, "loss": 1.2673, "step": 2780 }, { "epoch": 0.07221366966521377, "grad_norm": 7.353336391823007, "learning_rate": 1e-05, "loss": 1.2125, "step": 2781 }, { "epoch": 0.07223963646480572, "grad_norm": 8.775019597306063, "learning_rate": 1e-05, "loss": 1.2527, "step": 2782 }, { "epoch": 0.07226560326439767, "grad_norm": 8.747860892265589, "learning_rate": 1e-05, "loss": 1.3038, "step": 2783 }, { "epoch": 0.07229157006398962, "grad_norm": 8.785186645400954, "learning_rate": 1e-05, "loss": 1.2914, "step": 2784 }, { "epoch": 0.07231753686358157, "grad_norm": 6.284866430643167, "learning_rate": 1e-05, "loss": 1.2726, "step": 2785 }, { "epoch": 0.07234350366317352, "grad_norm": 21.87060363187665, "learning_rate": 1e-05, "loss": 1.253, "step": 2786 }, { "epoch": 0.07236947046276547, "grad_norm": 4.99028482728899, "learning_rate": 1e-05, "loss": 1.2889, "step": 2787 }, { "epoch": 0.07239543726235742, "grad_norm": 7.633013157466752, "learning_rate": 1e-05, "loss": 1.2878, "step": 2788 }, { "epoch": 0.07242140406194937, "grad_norm": 5.480619443424491, "learning_rate": 1e-05, "loss": 1.2781, "step": 2789 }, { "epoch": 0.07244737086154132, "grad_norm": 5.406939512173746, "learning_rate": 1e-05, "loss": 1.2305, "step": 2790 }, { "epoch": 0.07247333766113327, "grad_norm": 6.479874485354424, "learning_rate": 1e-05, "loss": 1.2771, "step": 2791 }, { "epoch": 0.07249930446072522, "grad_norm": 12.97142585816784, "learning_rate": 1e-05, "loss": 1.2589, "step": 2792 }, { "epoch": 0.07252527126031716, "grad_norm": 5.303075351606316, "learning_rate": 1e-05, "loss": 1.2993, "step": 2793 }, { "epoch": 0.07255123805990911, "grad_norm": 5.963164747117887, "learning_rate": 1e-05, "loss": 1.2398, "step": 2794 }, { "epoch": 0.07257720485950106, "grad_norm": 5.354036836744658, "learning_rate": 1e-05, "loss": 1.2828, "step": 2795 }, { "epoch": 0.07260317165909301, "grad_norm": 7.398509427036189, "learning_rate": 1e-05, "loss": 1.2291, "step": 2796 }, { "epoch": 0.07262913845868496, "grad_norm": 9.652891099540584, "learning_rate": 1e-05, "loss": 1.2281, "step": 2797 }, { "epoch": 0.07265510525827691, "grad_norm": 16.53639784273713, "learning_rate": 1e-05, "loss": 1.2864, "step": 2798 }, { "epoch": 0.07268107205786886, "grad_norm": 18.75592191638667, "learning_rate": 1e-05, "loss": 1.2624, "step": 2799 }, { "epoch": 0.07270703885746081, "grad_norm": 7.854241759517169, "learning_rate": 1e-05, "loss": 1.2867, "step": 2800 }, { "epoch": 0.07273300565705276, "grad_norm": 6.543222942823528, "learning_rate": 1e-05, "loss": 1.2549, "step": 2801 }, { "epoch": 0.07275897245664471, "grad_norm": 13.918586522389765, "learning_rate": 1e-05, "loss": 1.2753, "step": 2802 }, { "epoch": 0.07278493925623666, "grad_norm": 6.105739594820259, "learning_rate": 1e-05, "loss": 1.2244, "step": 2803 }, { "epoch": 0.07281090605582861, "grad_norm": 5.194803267453815, "learning_rate": 1e-05, "loss": 1.2561, "step": 2804 }, { "epoch": 0.07283687285542056, "grad_norm": 7.6164232389029936, "learning_rate": 1e-05, "loss": 1.2524, "step": 2805 }, { "epoch": 0.07286283965501252, "grad_norm": 6.4289995383961, "learning_rate": 1e-05, "loss": 1.2715, "step": 2806 }, { "epoch": 0.07288880645460447, "grad_norm": 8.599607378846578, "learning_rate": 1e-05, "loss": 1.2757, "step": 2807 }, { "epoch": 0.07291477325419642, "grad_norm": 11.624052275716664, "learning_rate": 1e-05, "loss": 1.3039, "step": 2808 }, { "epoch": 0.07294074005378837, "grad_norm": 6.791917417925078, "learning_rate": 1e-05, "loss": 1.2512, "step": 2809 }, { "epoch": 0.07296670685338032, "grad_norm": 18.576255556699593, "learning_rate": 1e-05, "loss": 1.2855, "step": 2810 }, { "epoch": 0.07299267365297227, "grad_norm": 6.991283984920974, "learning_rate": 1e-05, "loss": 1.2983, "step": 2811 }, { "epoch": 0.07301864045256422, "grad_norm": 6.457221478973415, "learning_rate": 1e-05, "loss": 1.2287, "step": 2812 }, { "epoch": 0.07304460725215617, "grad_norm": 6.373600525675316, "learning_rate": 1e-05, "loss": 1.2661, "step": 2813 }, { "epoch": 0.07307057405174812, "grad_norm": 13.906242319169227, "learning_rate": 1e-05, "loss": 1.2779, "step": 2814 }, { "epoch": 0.07309654085134007, "grad_norm": 5.255561789001969, "learning_rate": 1e-05, "loss": 1.3044, "step": 2815 }, { "epoch": 0.07312250765093202, "grad_norm": 6.800698233267003, "learning_rate": 1e-05, "loss": 1.2516, "step": 2816 }, { "epoch": 0.07314847445052397, "grad_norm": 5.559426304975551, "learning_rate": 1e-05, "loss": 1.2959, "step": 2817 }, { "epoch": 0.07317444125011592, "grad_norm": 5.258908660577114, "learning_rate": 1e-05, "loss": 1.3151, "step": 2818 }, { "epoch": 0.07320040804970787, "grad_norm": 8.06987951101289, "learning_rate": 1e-05, "loss": 1.2898, "step": 2819 }, { "epoch": 0.07322637484929982, "grad_norm": 6.691071884432754, "learning_rate": 1e-05, "loss": 1.282, "step": 2820 }, { "epoch": 0.07325234164889177, "grad_norm": 6.071114623343508, "learning_rate": 1e-05, "loss": 1.2245, "step": 2821 }, { "epoch": 0.07327830844848372, "grad_norm": 16.55110715176537, "learning_rate": 1e-05, "loss": 1.3083, "step": 2822 }, { "epoch": 0.07330427524807567, "grad_norm": 7.712809512266575, "learning_rate": 1e-05, "loss": 1.2785, "step": 2823 }, { "epoch": 0.07333024204766762, "grad_norm": 6.055188991328275, "learning_rate": 1e-05, "loss": 1.2154, "step": 2824 }, { "epoch": 0.07335620884725957, "grad_norm": 18.38436879169749, "learning_rate": 1e-05, "loss": 1.2091, "step": 2825 }, { "epoch": 0.07338217564685152, "grad_norm": 6.682180981055079, "learning_rate": 1e-05, "loss": 1.2057, "step": 2826 }, { "epoch": 0.07340814244644348, "grad_norm": 6.2008000011528255, "learning_rate": 1e-05, "loss": 1.2722, "step": 2827 }, { "epoch": 0.07343410924603543, "grad_norm": 9.521939292017114, "learning_rate": 1e-05, "loss": 1.2776, "step": 2828 }, { "epoch": 0.07346007604562738, "grad_norm": 7.754968096767695, "learning_rate": 1e-05, "loss": 1.2762, "step": 2829 }, { "epoch": 0.07348604284521933, "grad_norm": 6.7437466512677045, "learning_rate": 1e-05, "loss": 1.2638, "step": 2830 }, { "epoch": 0.07351200964481128, "grad_norm": 12.250617186899644, "learning_rate": 1e-05, "loss": 1.2298, "step": 2831 }, { "epoch": 0.07353797644440323, "grad_norm": 5.726098180997071, "learning_rate": 1e-05, "loss": 1.2618, "step": 2832 }, { "epoch": 0.07356394324399518, "grad_norm": 10.65848748217658, "learning_rate": 1e-05, "loss": 1.2476, "step": 2833 }, { "epoch": 0.07358991004358713, "grad_norm": 7.505585688560242, "learning_rate": 1e-05, "loss": 1.2535, "step": 2834 }, { "epoch": 0.07361587684317908, "grad_norm": 6.258453755377851, "learning_rate": 1e-05, "loss": 1.3277, "step": 2835 }, { "epoch": 0.07364184364277103, "grad_norm": 7.528873052679224, "learning_rate": 1e-05, "loss": 1.3025, "step": 2836 }, { "epoch": 0.07366781044236298, "grad_norm": 12.931321295001927, "learning_rate": 1e-05, "loss": 1.2961, "step": 2837 }, { "epoch": 0.07369377724195493, "grad_norm": 23.373374515257073, "learning_rate": 1e-05, "loss": 1.2486, "step": 2838 }, { "epoch": 0.07371974404154688, "grad_norm": 13.443276374734696, "learning_rate": 1e-05, "loss": 1.2755, "step": 2839 }, { "epoch": 0.07374571084113883, "grad_norm": 4.8922186744205955, "learning_rate": 1e-05, "loss": 1.2793, "step": 2840 }, { "epoch": 0.07377167764073078, "grad_norm": 4.623834617811927, "learning_rate": 1e-05, "loss": 1.3093, "step": 2841 }, { "epoch": 0.07379764444032273, "grad_norm": 5.172229766202825, "learning_rate": 1e-05, "loss": 1.2584, "step": 2842 }, { "epoch": 0.07382361123991468, "grad_norm": 8.215711604924614, "learning_rate": 1e-05, "loss": 1.2695, "step": 2843 }, { "epoch": 0.07384957803950663, "grad_norm": 8.868551141016232, "learning_rate": 1e-05, "loss": 1.2819, "step": 2844 }, { "epoch": 0.07387554483909858, "grad_norm": 6.288927538827511, "learning_rate": 1e-05, "loss": 1.2419, "step": 2845 }, { "epoch": 0.07390151163869053, "grad_norm": 6.123123504305185, "learning_rate": 1e-05, "loss": 1.258, "step": 2846 }, { "epoch": 0.07392747843828248, "grad_norm": 11.949817633733671, "learning_rate": 1e-05, "loss": 1.3015, "step": 2847 }, { "epoch": 0.07395344523787444, "grad_norm": 6.763401996582993, "learning_rate": 1e-05, "loss": 1.2894, "step": 2848 }, { "epoch": 0.07397941203746639, "grad_norm": 5.102075618158096, "learning_rate": 1e-05, "loss": 1.2329, "step": 2849 }, { "epoch": 0.07400537883705834, "grad_norm": 4.915311378713157, "learning_rate": 1e-05, "loss": 1.2681, "step": 2850 }, { "epoch": 0.07403134563665029, "grad_norm": 8.946086040517157, "learning_rate": 1e-05, "loss": 1.2795, "step": 2851 }, { "epoch": 0.07405731243624224, "grad_norm": 5.913334225923685, "learning_rate": 1e-05, "loss": 1.2094, "step": 2852 }, { "epoch": 0.07408327923583419, "grad_norm": 6.920940534146878, "learning_rate": 1e-05, "loss": 1.312, "step": 2853 }, { "epoch": 0.07410924603542614, "grad_norm": 6.423011252314858, "learning_rate": 1e-05, "loss": 1.3052, "step": 2854 }, { "epoch": 0.07413521283501809, "grad_norm": 8.234337012187043, "learning_rate": 1e-05, "loss": 1.2914, "step": 2855 }, { "epoch": 0.07416117963461004, "grad_norm": 5.666563631504775, "learning_rate": 1e-05, "loss": 1.2594, "step": 2856 }, { "epoch": 0.07418714643420199, "grad_norm": 7.384993942351236, "learning_rate": 1e-05, "loss": 1.2868, "step": 2857 }, { "epoch": 0.07421311323379394, "grad_norm": 8.619365676234217, "learning_rate": 1e-05, "loss": 1.3268, "step": 2858 }, { "epoch": 0.07423908003338589, "grad_norm": 5.805571662032779, "learning_rate": 1e-05, "loss": 1.2596, "step": 2859 }, { "epoch": 0.07426504683297784, "grad_norm": 8.222961507101052, "learning_rate": 1e-05, "loss": 1.2512, "step": 2860 }, { "epoch": 0.07429101363256979, "grad_norm": 6.617049142591337, "learning_rate": 1e-05, "loss": 1.2626, "step": 2861 }, { "epoch": 0.07431698043216174, "grad_norm": 8.661594324808718, "learning_rate": 1e-05, "loss": 1.2613, "step": 2862 }, { "epoch": 0.07434294723175369, "grad_norm": 15.400021421739616, "learning_rate": 1e-05, "loss": 1.2127, "step": 2863 }, { "epoch": 0.07436891403134564, "grad_norm": 6.603906700561176, "learning_rate": 1e-05, "loss": 1.2351, "step": 2864 }, { "epoch": 0.0743948808309376, "grad_norm": 5.330131980593163, "learning_rate": 1e-05, "loss": 1.2965, "step": 2865 }, { "epoch": 0.07442084763052954, "grad_norm": 6.6160102166278, "learning_rate": 1e-05, "loss": 1.3175, "step": 2866 }, { "epoch": 0.0744468144301215, "grad_norm": 6.764492442283381, "learning_rate": 1e-05, "loss": 1.2674, "step": 2867 }, { "epoch": 0.07447278122971344, "grad_norm": 6.924784787135592, "learning_rate": 1e-05, "loss": 1.2852, "step": 2868 }, { "epoch": 0.0744987480293054, "grad_norm": 6.632507389091807, "learning_rate": 1e-05, "loss": 1.2767, "step": 2869 }, { "epoch": 0.07452471482889733, "grad_norm": 6.452710006707574, "learning_rate": 1e-05, "loss": 1.2358, "step": 2870 }, { "epoch": 0.07455068162848928, "grad_norm": 10.378497166265248, "learning_rate": 1e-05, "loss": 1.2775, "step": 2871 }, { "epoch": 0.07457664842808123, "grad_norm": 7.177570948394608, "learning_rate": 1e-05, "loss": 1.2747, "step": 2872 }, { "epoch": 0.07460261522767318, "grad_norm": 6.306959988236241, "learning_rate": 1e-05, "loss": 1.2456, "step": 2873 }, { "epoch": 0.07462858202726513, "grad_norm": 7.335572189969099, "learning_rate": 1e-05, "loss": 1.2939, "step": 2874 }, { "epoch": 0.07465454882685708, "grad_norm": 12.922990854556584, "learning_rate": 1e-05, "loss": 1.2418, "step": 2875 }, { "epoch": 0.07468051562644903, "grad_norm": 6.1390158216760655, "learning_rate": 1e-05, "loss": 1.2592, "step": 2876 }, { "epoch": 0.07470648242604098, "grad_norm": 11.603835004704534, "learning_rate": 1e-05, "loss": 1.2405, "step": 2877 }, { "epoch": 0.07473244922563294, "grad_norm": 5.450205723790674, "learning_rate": 1e-05, "loss": 1.2575, "step": 2878 }, { "epoch": 0.07475841602522489, "grad_norm": 6.230157711671887, "learning_rate": 1e-05, "loss": 1.2809, "step": 2879 }, { "epoch": 0.07478438282481684, "grad_norm": 5.816498396369649, "learning_rate": 1e-05, "loss": 1.2659, "step": 2880 }, { "epoch": 0.07481034962440879, "grad_norm": 6.158007066039355, "learning_rate": 1e-05, "loss": 1.2553, "step": 2881 }, { "epoch": 0.07483631642400074, "grad_norm": 6.747765630445234, "learning_rate": 1e-05, "loss": 1.2879, "step": 2882 }, { "epoch": 0.07486228322359269, "grad_norm": 7.064815985462556, "learning_rate": 1e-05, "loss": 1.2425, "step": 2883 }, { "epoch": 0.07488825002318464, "grad_norm": 8.065605097926822, "learning_rate": 1e-05, "loss": 1.2296, "step": 2884 }, { "epoch": 0.07491421682277659, "grad_norm": 10.560145369598413, "learning_rate": 1e-05, "loss": 1.3026, "step": 2885 }, { "epoch": 0.07494018362236854, "grad_norm": 5.722675914535326, "learning_rate": 1e-05, "loss": 1.2291, "step": 2886 }, { "epoch": 0.07496615042196049, "grad_norm": 5.779811040832097, "learning_rate": 1e-05, "loss": 1.3019, "step": 2887 }, { "epoch": 0.07499211722155244, "grad_norm": 7.224889745168018, "learning_rate": 1e-05, "loss": 1.2895, "step": 2888 }, { "epoch": 0.07501808402114439, "grad_norm": 38.04916734624015, "learning_rate": 1e-05, "loss": 1.2643, "step": 2889 }, { "epoch": 0.07504405082073634, "grad_norm": 7.131413835520768, "learning_rate": 1e-05, "loss": 1.261, "step": 2890 }, { "epoch": 0.07507001762032829, "grad_norm": 12.835460800262611, "learning_rate": 1e-05, "loss": 1.3027, "step": 2891 }, { "epoch": 0.07509598441992024, "grad_norm": 8.73020352295673, "learning_rate": 1e-05, "loss": 1.2777, "step": 2892 }, { "epoch": 0.07512195121951219, "grad_norm": 9.151696019813077, "learning_rate": 1e-05, "loss": 1.2877, "step": 2893 }, { "epoch": 0.07514791801910414, "grad_norm": 30.85837913741228, "learning_rate": 1e-05, "loss": 1.2831, "step": 2894 }, { "epoch": 0.0751738848186961, "grad_norm": 13.678525168352397, "learning_rate": 1e-05, "loss": 1.2446, "step": 2895 }, { "epoch": 0.07519985161828804, "grad_norm": 10.039027161294806, "learning_rate": 1e-05, "loss": 1.2804, "step": 2896 }, { "epoch": 0.07522581841788, "grad_norm": 6.142615696924044, "learning_rate": 1e-05, "loss": 1.3226, "step": 2897 }, { "epoch": 0.07525178521747194, "grad_norm": 8.698107932418631, "learning_rate": 1e-05, "loss": 1.2314, "step": 2898 }, { "epoch": 0.0752777520170639, "grad_norm": 6.244344364931483, "learning_rate": 1e-05, "loss": 1.2163, "step": 2899 }, { "epoch": 0.07530371881665585, "grad_norm": 5.4200214658642905, "learning_rate": 1e-05, "loss": 1.2451, "step": 2900 }, { "epoch": 0.0753296856162478, "grad_norm": 7.924193274415571, "learning_rate": 1e-05, "loss": 1.239, "step": 2901 }, { "epoch": 0.07535565241583975, "grad_norm": 6.8783841213862065, "learning_rate": 1e-05, "loss": 1.2434, "step": 2902 }, { "epoch": 0.0753816192154317, "grad_norm": 8.597749868157956, "learning_rate": 1e-05, "loss": 1.2752, "step": 2903 }, { "epoch": 0.07540758601502365, "grad_norm": 8.821815274854282, "learning_rate": 1e-05, "loss": 1.2427, "step": 2904 }, { "epoch": 0.0754335528146156, "grad_norm": 9.888600315450638, "learning_rate": 1e-05, "loss": 1.2777, "step": 2905 }, { "epoch": 0.07545951961420755, "grad_norm": 7.109709813544607, "learning_rate": 1e-05, "loss": 1.2826, "step": 2906 }, { "epoch": 0.0754854864137995, "grad_norm": 8.228821823685152, "learning_rate": 1e-05, "loss": 1.2761, "step": 2907 }, { "epoch": 0.07551145321339145, "grad_norm": 8.005357855494735, "learning_rate": 1e-05, "loss": 1.2809, "step": 2908 }, { "epoch": 0.0755374200129834, "grad_norm": 7.444518552319558, "learning_rate": 1e-05, "loss": 1.2652, "step": 2909 }, { "epoch": 0.07556338681257535, "grad_norm": 13.362778828979643, "learning_rate": 1e-05, "loss": 1.2982, "step": 2910 }, { "epoch": 0.0755893536121673, "grad_norm": 45.3487785805748, "learning_rate": 1e-05, "loss": 1.273, "step": 2911 }, { "epoch": 0.07561532041175925, "grad_norm": 7.099275159774373, "learning_rate": 1e-05, "loss": 1.2432, "step": 2912 }, { "epoch": 0.0756412872113512, "grad_norm": 21.083686835701613, "learning_rate": 1e-05, "loss": 1.2453, "step": 2913 }, { "epoch": 0.07566725401094315, "grad_norm": 8.944801176024724, "learning_rate": 1e-05, "loss": 1.2782, "step": 2914 }, { "epoch": 0.0756932208105351, "grad_norm": 9.109879599151416, "learning_rate": 1e-05, "loss": 1.2283, "step": 2915 }, { "epoch": 0.07571918761012705, "grad_norm": 10.037052936027836, "learning_rate": 1e-05, "loss": 1.269, "step": 2916 }, { "epoch": 0.075745154409719, "grad_norm": 8.299652037163261, "learning_rate": 1e-05, "loss": 1.2354, "step": 2917 }, { "epoch": 0.07577112120931095, "grad_norm": 8.5627147341345, "learning_rate": 1e-05, "loss": 1.2436, "step": 2918 }, { "epoch": 0.0757970880089029, "grad_norm": 37.73112854823504, "learning_rate": 1e-05, "loss": 1.3093, "step": 2919 }, { "epoch": 0.07582305480849486, "grad_norm": 6.82168715563929, "learning_rate": 1e-05, "loss": 1.2332, "step": 2920 }, { "epoch": 0.0758490216080868, "grad_norm": 8.85199431454441, "learning_rate": 1e-05, "loss": 1.2314, "step": 2921 }, { "epoch": 0.07587498840767876, "grad_norm": 9.69565676954744, "learning_rate": 1e-05, "loss": 1.2885, "step": 2922 }, { "epoch": 0.0759009552072707, "grad_norm": 14.319718452115305, "learning_rate": 1e-05, "loss": 1.3014, "step": 2923 }, { "epoch": 0.07592692200686266, "grad_norm": 14.907148226180107, "learning_rate": 1e-05, "loss": 1.2406, "step": 2924 }, { "epoch": 0.07595288880645461, "grad_norm": 7.836740469747965, "learning_rate": 1e-05, "loss": 1.2774, "step": 2925 }, { "epoch": 0.07597885560604656, "grad_norm": 10.574481015287502, "learning_rate": 1e-05, "loss": 1.2802, "step": 2926 }, { "epoch": 0.07600482240563851, "grad_norm": 13.562620944701228, "learning_rate": 1e-05, "loss": 1.2815, "step": 2927 }, { "epoch": 0.07603078920523046, "grad_norm": 8.254922958559808, "learning_rate": 1e-05, "loss": 1.2989, "step": 2928 }, { "epoch": 0.07605675600482241, "grad_norm": 10.407373499434826, "learning_rate": 1e-05, "loss": 1.2761, "step": 2929 }, { "epoch": 0.07608272280441436, "grad_norm": 26.152378465716318, "learning_rate": 1e-05, "loss": 1.2908, "step": 2930 }, { "epoch": 0.07610868960400631, "grad_norm": 38.78489639636236, "learning_rate": 1e-05, "loss": 1.2945, "step": 2931 }, { "epoch": 0.07613465640359826, "grad_norm": 15.260154970114623, "learning_rate": 1e-05, "loss": 1.2342, "step": 2932 }, { "epoch": 0.07616062320319021, "grad_norm": 6.988407617900916, "learning_rate": 1e-05, "loss": 1.2246, "step": 2933 }, { "epoch": 0.07618659000278216, "grad_norm": 5.899541041739927, "learning_rate": 1e-05, "loss": 1.2922, "step": 2934 }, { "epoch": 0.07621255680237411, "grad_norm": 34.51293172210509, "learning_rate": 1e-05, "loss": 1.28, "step": 2935 }, { "epoch": 0.07623852360196606, "grad_norm": 8.867966564757241, "learning_rate": 1e-05, "loss": 1.2832, "step": 2936 }, { "epoch": 0.07626449040155801, "grad_norm": 8.27756772496645, "learning_rate": 1e-05, "loss": 1.288, "step": 2937 }, { "epoch": 0.07629045720114996, "grad_norm": 8.705081849828124, "learning_rate": 1e-05, "loss": 1.2926, "step": 2938 }, { "epoch": 0.07631642400074191, "grad_norm": 10.711148338552162, "learning_rate": 1e-05, "loss": 1.3032, "step": 2939 }, { "epoch": 0.07634239080033386, "grad_norm": 6.2628188853442825, "learning_rate": 1e-05, "loss": 1.3123, "step": 2940 }, { "epoch": 0.07636835759992582, "grad_norm": 5.1842999986922464, "learning_rate": 1e-05, "loss": 1.2877, "step": 2941 }, { "epoch": 0.07639432439951777, "grad_norm": 8.298085040057812, "learning_rate": 1e-05, "loss": 1.2489, "step": 2942 }, { "epoch": 0.07642029119910972, "grad_norm": 6.152320808488972, "learning_rate": 1e-05, "loss": 1.2716, "step": 2943 }, { "epoch": 0.07644625799870167, "grad_norm": 10.965025005738237, "learning_rate": 1e-05, "loss": 1.2546, "step": 2944 }, { "epoch": 0.07647222479829362, "grad_norm": 9.236989950054204, "learning_rate": 1e-05, "loss": 1.2608, "step": 2945 }, { "epoch": 0.07649819159788555, "grad_norm": 9.049999376159983, "learning_rate": 1e-05, "loss": 1.286, "step": 2946 }, { "epoch": 0.0765241583974775, "grad_norm": 8.921413215939012, "learning_rate": 1e-05, "loss": 1.2565, "step": 2947 }, { "epoch": 0.07655012519706945, "grad_norm": 13.528425454874116, "learning_rate": 1e-05, "loss": 1.2219, "step": 2948 }, { "epoch": 0.0765760919966614, "grad_norm": 6.471192714178969, "learning_rate": 1e-05, "loss": 1.1972, "step": 2949 }, { "epoch": 0.07660205879625336, "grad_norm": 7.6430829858962595, "learning_rate": 1e-05, "loss": 1.2498, "step": 2950 }, { "epoch": 0.0766280255958453, "grad_norm": 7.976855176437603, "learning_rate": 1e-05, "loss": 1.2889, "step": 2951 }, { "epoch": 0.07665399239543726, "grad_norm": 9.10515495424008, "learning_rate": 1e-05, "loss": 1.2631, "step": 2952 }, { "epoch": 0.07667995919502921, "grad_norm": 35.90384973482997, "learning_rate": 1e-05, "loss": 1.272, "step": 2953 }, { "epoch": 0.07670592599462116, "grad_norm": 7.4907573969936605, "learning_rate": 1e-05, "loss": 1.3077, "step": 2954 }, { "epoch": 0.07673189279421311, "grad_norm": 6.531633721719784, "learning_rate": 1e-05, "loss": 1.2806, "step": 2955 }, { "epoch": 0.07675785959380506, "grad_norm": 5.831561918594399, "learning_rate": 1e-05, "loss": 1.2882, "step": 2956 }, { "epoch": 0.07678382639339701, "grad_norm": 9.075821129506943, "learning_rate": 1e-05, "loss": 1.3052, "step": 2957 }, { "epoch": 0.07680979319298896, "grad_norm": 24.453687842963355, "learning_rate": 1e-05, "loss": 1.2477, "step": 2958 }, { "epoch": 0.07683575999258091, "grad_norm": 12.188183574579776, "learning_rate": 1e-05, "loss": 1.2498, "step": 2959 }, { "epoch": 0.07686172679217286, "grad_norm": 5.0290497891367325, "learning_rate": 1e-05, "loss": 1.2377, "step": 2960 }, { "epoch": 0.07688769359176481, "grad_norm": 5.356776194059007, "learning_rate": 1e-05, "loss": 1.293, "step": 2961 }, { "epoch": 0.07691366039135676, "grad_norm": 5.591319147369362, "learning_rate": 1e-05, "loss": 1.2823, "step": 2962 }, { "epoch": 0.07693962719094871, "grad_norm": 20.578618836614286, "learning_rate": 1e-05, "loss": 1.2785, "step": 2963 }, { "epoch": 0.07696559399054066, "grad_norm": 6.242676678755557, "learning_rate": 1e-05, "loss": 1.2885, "step": 2964 }, { "epoch": 0.07699156079013261, "grad_norm": 121.15022473410852, "learning_rate": 1e-05, "loss": 1.2736, "step": 2965 }, { "epoch": 0.07701752758972456, "grad_norm": 5.782789200718059, "learning_rate": 1e-05, "loss": 1.2856, "step": 2966 }, { "epoch": 0.07704349438931651, "grad_norm": 5.645898778707353, "learning_rate": 1e-05, "loss": 1.2901, "step": 2967 }, { "epoch": 0.07706946118890846, "grad_norm": 5.105292412792551, "learning_rate": 1e-05, "loss": 1.3347, "step": 2968 }, { "epoch": 0.07709542798850041, "grad_norm": 40.61714918131487, "learning_rate": 1e-05, "loss": 1.2803, "step": 2969 }, { "epoch": 0.07712139478809237, "grad_norm": 6.51956284631349, "learning_rate": 1e-05, "loss": 1.2568, "step": 2970 }, { "epoch": 0.07714736158768432, "grad_norm": 6.571693012295352, "learning_rate": 1e-05, "loss": 1.2461, "step": 2971 }, { "epoch": 0.07717332838727627, "grad_norm": 8.125304231083058, "learning_rate": 1e-05, "loss": 1.2443, "step": 2972 }, { "epoch": 0.07719929518686822, "grad_norm": 8.948037930881831, "learning_rate": 1e-05, "loss": 1.2754, "step": 2973 }, { "epoch": 0.07722526198646017, "grad_norm": 6.231599194567035, "learning_rate": 1e-05, "loss": 1.2825, "step": 2974 }, { "epoch": 0.07725122878605212, "grad_norm": 8.665823039540832, "learning_rate": 1e-05, "loss": 1.2556, "step": 2975 }, { "epoch": 0.07727719558564407, "grad_norm": 5.347243388520772, "learning_rate": 1e-05, "loss": 1.2763, "step": 2976 }, { "epoch": 0.07730316238523602, "grad_norm": 6.327827804556718, "learning_rate": 1e-05, "loss": 1.2389, "step": 2977 }, { "epoch": 0.07732912918482797, "grad_norm": 9.572148261433647, "learning_rate": 1e-05, "loss": 1.2771, "step": 2978 }, { "epoch": 0.07735509598441992, "grad_norm": 5.294963840724973, "learning_rate": 1e-05, "loss": 1.263, "step": 2979 }, { "epoch": 0.07738106278401187, "grad_norm": 5.310773860770588, "learning_rate": 1e-05, "loss": 1.2533, "step": 2980 }, { "epoch": 0.07740702958360382, "grad_norm": 6.566151211770853, "learning_rate": 1e-05, "loss": 1.2566, "step": 2981 }, { "epoch": 0.07743299638319577, "grad_norm": 6.017794289772098, "learning_rate": 1e-05, "loss": 1.2725, "step": 2982 }, { "epoch": 0.07745896318278772, "grad_norm": 6.413429665796999, "learning_rate": 1e-05, "loss": 1.26, "step": 2983 }, { "epoch": 0.07748492998237967, "grad_norm": 5.6680808079248495, "learning_rate": 1e-05, "loss": 1.313, "step": 2984 }, { "epoch": 0.07751089678197162, "grad_norm": 7.9203962943603266, "learning_rate": 1e-05, "loss": 1.3101, "step": 2985 }, { "epoch": 0.07753686358156357, "grad_norm": 7.1697683793328535, "learning_rate": 1e-05, "loss": 1.2567, "step": 2986 }, { "epoch": 0.07756283038115552, "grad_norm": 4.714962481352479, "learning_rate": 1e-05, "loss": 1.302, "step": 2987 }, { "epoch": 0.07758879718074747, "grad_norm": 4.9813866341684685, "learning_rate": 1e-05, "loss": 1.2418, "step": 2988 }, { "epoch": 0.07761476398033942, "grad_norm": 12.428408880063412, "learning_rate": 1e-05, "loss": 1.2637, "step": 2989 }, { "epoch": 0.07764073077993137, "grad_norm": 9.871974046867791, "learning_rate": 1e-05, "loss": 1.2833, "step": 2990 }, { "epoch": 0.07766669757952332, "grad_norm": 7.858978200796571, "learning_rate": 1e-05, "loss": 1.3028, "step": 2991 }, { "epoch": 0.07769266437911528, "grad_norm": 6.68428233092686, "learning_rate": 1e-05, "loss": 1.2763, "step": 2992 }, { "epoch": 0.07771863117870723, "grad_norm": 4.8345041939573905, "learning_rate": 1e-05, "loss": 1.2575, "step": 2993 }, { "epoch": 0.07774459797829918, "grad_norm": 9.35937866822236, "learning_rate": 1e-05, "loss": 1.2701, "step": 2994 }, { "epoch": 0.07777056477789113, "grad_norm": 8.598345717186861, "learning_rate": 1e-05, "loss": 1.2939, "step": 2995 }, { "epoch": 0.07779653157748308, "grad_norm": 6.871540100624618, "learning_rate": 1e-05, "loss": 1.213, "step": 2996 }, { "epoch": 0.07782249837707503, "grad_norm": 11.327631320061673, "learning_rate": 1e-05, "loss": 1.2512, "step": 2997 }, { "epoch": 0.07784846517666698, "grad_norm": 10.476363325963721, "learning_rate": 1e-05, "loss": 1.181, "step": 2998 }, { "epoch": 0.07787443197625893, "grad_norm": 5.525124126438141, "learning_rate": 1e-05, "loss": 1.2392, "step": 2999 }, { "epoch": 0.07790039877585088, "grad_norm": 8.017536017322435, "learning_rate": 1e-05, "loss": 1.2553, "step": 3000 } ], "logging_steps": 1.0, "max_steps": 38510, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.2253061281352253e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }