{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.999839871897518, "eval_steps": 500, "global_step": 3122, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0003202562049639712, "grad_norm": 0.5703125, "learning_rate": 1.597444089456869e-07, "loss": 2.5505, "step": 1 }, { "epoch": 0.0006405124099279424, "grad_norm": 0.73828125, "learning_rate": 3.194888178913738e-07, "loss": 2.302, "step": 2 }, { "epoch": 0.0009607686148919135, "grad_norm": 0.484375, "learning_rate": 4.792332268370607e-07, "loss": 2.0733, "step": 3 }, { "epoch": 0.0012810248198558848, "grad_norm": 0.52734375, "learning_rate": 6.389776357827476e-07, "loss": 2.5665, "step": 4 }, { "epoch": 0.0016012810248198558, "grad_norm": 0.578125, "learning_rate": 7.987220447284345e-07, "loss": 2.2262, "step": 5 }, { "epoch": 0.001921537229783827, "grad_norm": 0.48828125, "learning_rate": 9.584664536741213e-07, "loss": 2.7195, "step": 6 }, { "epoch": 0.0022417934347477983, "grad_norm": 1.1484375, "learning_rate": 1.1182108626198083e-06, "loss": 2.3025, "step": 7 }, { "epoch": 0.0025620496397117695, "grad_norm": 0.267578125, "learning_rate": 1.2779552715654952e-06, "loss": 1.7873, "step": 8 }, { "epoch": 0.0028823058446757407, "grad_norm": 0.38671875, "learning_rate": 1.4376996805111822e-06, "loss": 2.1716, "step": 9 }, { "epoch": 0.0032025620496397116, "grad_norm": 0.6328125, "learning_rate": 1.597444089456869e-06, "loss": 2.1507, "step": 10 }, { "epoch": 0.003522818254603683, "grad_norm": 0.77734375, "learning_rate": 1.7571884984025559e-06, "loss": 2.2224, "step": 11 }, { "epoch": 0.003843074459567654, "grad_norm": 0.5, "learning_rate": 1.9169329073482426e-06, "loss": 2.3008, "step": 12 }, { "epoch": 0.004163330664531626, "grad_norm": 0.4609375, "learning_rate": 2.0766773162939296e-06, "loss": 2.3001, "step": 13 }, { "epoch": 0.0044835868694955965, "grad_norm": 0.52734375, "learning_rate": 2.2364217252396165e-06, "loss": 2.1583, "step": 14 }, { "epoch": 0.004803843074459567, "grad_norm": 0.60546875, "learning_rate": 2.3961661341853035e-06, "loss": 2.2752, "step": 15 }, { "epoch": 0.005124099279423539, "grad_norm": 0.443359375, "learning_rate": 2.5559105431309904e-06, "loss": 1.6753, "step": 16 }, { "epoch": 0.00544435548438751, "grad_norm": 1.4609375, "learning_rate": 2.7156549520766774e-06, "loss": 2.4427, "step": 17 }, { "epoch": 0.0057646116893514815, "grad_norm": 1.0546875, "learning_rate": 2.8753993610223644e-06, "loss": 2.3853, "step": 18 }, { "epoch": 0.006084867894315452, "grad_norm": 0.51953125, "learning_rate": 3.0351437699680513e-06, "loss": 2.4479, "step": 19 }, { "epoch": 0.006405124099279423, "grad_norm": 0.59375, "learning_rate": 3.194888178913738e-06, "loss": 1.9668, "step": 20 }, { "epoch": 0.006725380304243395, "grad_norm": 1.2890625, "learning_rate": 3.354632587859425e-06, "loss": 2.2987, "step": 21 }, { "epoch": 0.007045636509207366, "grad_norm": 0.48046875, "learning_rate": 3.5143769968051118e-06, "loss": 1.7833, "step": 22 }, { "epoch": 0.007365892714171337, "grad_norm": 0.494140625, "learning_rate": 3.6741214057507987e-06, "loss": 2.3454, "step": 23 }, { "epoch": 0.007686148919135308, "grad_norm": 0.609375, "learning_rate": 3.833865814696485e-06, "loss": 2.1669, "step": 24 }, { "epoch": 0.008006405124099279, "grad_norm": 0.45703125, "learning_rate": 3.993610223642173e-06, "loss": 2.2589, "step": 25 }, { "epoch": 0.008326661329063251, "grad_norm": 0.458984375, "learning_rate": 4.153354632587859e-06, "loss": 2.3922, "step": 26 }, { "epoch": 0.008646917534027222, "grad_norm": 0.5, "learning_rate": 4.3130990415335465e-06, "loss": 2.8444, "step": 27 }, { "epoch": 0.008967173738991193, "grad_norm": 0.45703125, "learning_rate": 4.472843450479233e-06, "loss": 1.6251, "step": 28 }, { "epoch": 0.009287429943955164, "grad_norm": 0.57421875, "learning_rate": 4.6325878594249205e-06, "loss": 1.4672, "step": 29 }, { "epoch": 0.009607686148919135, "grad_norm": 0.5625, "learning_rate": 4.792332268370607e-06, "loss": 2.2324, "step": 30 }, { "epoch": 0.009927942353883107, "grad_norm": 0.455078125, "learning_rate": 4.952076677316294e-06, "loss": 1.8752, "step": 31 }, { "epoch": 0.010248198558847078, "grad_norm": 1.921875, "learning_rate": 5.111821086261981e-06, "loss": 2.4251, "step": 32 }, { "epoch": 0.010568454763811049, "grad_norm": 0.412109375, "learning_rate": 5.2715654952076674e-06, "loss": 2.2702, "step": 33 }, { "epoch": 0.01088871096877502, "grad_norm": 5.3125, "learning_rate": 5.431309904153355e-06, "loss": 2.2467, "step": 34 }, { "epoch": 0.01120896717373899, "grad_norm": 0.357421875, "learning_rate": 5.591054313099041e-06, "loss": 1.8738, "step": 35 }, { "epoch": 0.011529223378702963, "grad_norm": 0.5234375, "learning_rate": 5.750798722044729e-06, "loss": 2.0557, "step": 36 }, { "epoch": 0.011849479583666934, "grad_norm": 0.5390625, "learning_rate": 5.910543130990415e-06, "loss": 2.236, "step": 37 }, { "epoch": 0.012169735788630905, "grad_norm": 0.984375, "learning_rate": 6.070287539936103e-06, "loss": 1.9037, "step": 38 }, { "epoch": 0.012489991993594875, "grad_norm": 0.46484375, "learning_rate": 6.230031948881789e-06, "loss": 1.9012, "step": 39 }, { "epoch": 0.012810248198558846, "grad_norm": 2.1875, "learning_rate": 6.389776357827476e-06, "loss": 1.4725, "step": 40 }, { "epoch": 0.013130504403522819, "grad_norm": 0.6796875, "learning_rate": 6.549520766773164e-06, "loss": 2.4334, "step": 41 }, { "epoch": 0.01345076060848679, "grad_norm": 0.48046875, "learning_rate": 6.70926517571885e-06, "loss": 2.2842, "step": 42 }, { "epoch": 0.01377101681345076, "grad_norm": 0.40625, "learning_rate": 6.869009584664538e-06, "loss": 1.993, "step": 43 }, { "epoch": 0.014091273018414731, "grad_norm": 0.50390625, "learning_rate": 7.0287539936102235e-06, "loss": 1.7918, "step": 44 }, { "epoch": 0.014411529223378704, "grad_norm": 0.4375, "learning_rate": 7.188498402555911e-06, "loss": 2.573, "step": 45 }, { "epoch": 0.014731785428342675, "grad_norm": 1.578125, "learning_rate": 7.3482428115015974e-06, "loss": 2.0233, "step": 46 }, { "epoch": 0.015052041633306645, "grad_norm": 0.47265625, "learning_rate": 7.507987220447285e-06, "loss": 2.1536, "step": 47 }, { "epoch": 0.015372297838270616, "grad_norm": 0.5703125, "learning_rate": 7.66773162939297e-06, "loss": 2.4289, "step": 48 }, { "epoch": 0.015692554043234587, "grad_norm": 0.412109375, "learning_rate": 7.82747603833866e-06, "loss": 1.7267, "step": 49 }, { "epoch": 0.016012810248198558, "grad_norm": 0.384765625, "learning_rate": 7.987220447284345e-06, "loss": 1.8542, "step": 50 }, { "epoch": 0.01633306645316253, "grad_norm": 0.494140625, "learning_rate": 8.146964856230033e-06, "loss": 1.6736, "step": 51 }, { "epoch": 0.016653322658126503, "grad_norm": 0.59375, "learning_rate": 8.306709265175718e-06, "loss": 2.1386, "step": 52 }, { "epoch": 0.016973578863090474, "grad_norm": 0.45703125, "learning_rate": 8.466453674121406e-06, "loss": 2.3592, "step": 53 }, { "epoch": 0.017293835068054444, "grad_norm": 0.546875, "learning_rate": 8.626198083067093e-06, "loss": 1.732, "step": 54 }, { "epoch": 0.017614091273018415, "grad_norm": 0.77734375, "learning_rate": 8.78594249201278e-06, "loss": 1.5725, "step": 55 }, { "epoch": 0.017934347477982386, "grad_norm": 0.3984375, "learning_rate": 8.945686900958466e-06, "loss": 2.4561, "step": 56 }, { "epoch": 0.018254603682946357, "grad_norm": 0.734375, "learning_rate": 9.105431309904154e-06, "loss": 1.8126, "step": 57 }, { "epoch": 0.018574859887910328, "grad_norm": 0.474609375, "learning_rate": 9.265175718849841e-06, "loss": 2.1243, "step": 58 }, { "epoch": 0.0188951160928743, "grad_norm": 0.54296875, "learning_rate": 9.424920127795528e-06, "loss": 2.3236, "step": 59 }, { "epoch": 0.01921537229783827, "grad_norm": 0.5234375, "learning_rate": 9.584664536741214e-06, "loss": 2.3305, "step": 60 }, { "epoch": 0.01953562850280224, "grad_norm": 0.49609375, "learning_rate": 9.744408945686901e-06, "loss": 2.1348, "step": 61 }, { "epoch": 0.019855884707766214, "grad_norm": 0.314453125, "learning_rate": 9.904153354632589e-06, "loss": 1.91, "step": 62 }, { "epoch": 0.020176140912730185, "grad_norm": 0.53515625, "learning_rate": 1.0063897763578276e-05, "loss": 2.4788, "step": 63 }, { "epoch": 0.020496397117694156, "grad_norm": 0.84765625, "learning_rate": 1.0223642172523962e-05, "loss": 2.4574, "step": 64 }, { "epoch": 0.020816653322658127, "grad_norm": 0.8203125, "learning_rate": 1.038338658146965e-05, "loss": 2.3022, "step": 65 }, { "epoch": 0.021136909527622098, "grad_norm": 0.470703125, "learning_rate": 1.0543130990415335e-05, "loss": 1.4106, "step": 66 }, { "epoch": 0.02145716573258607, "grad_norm": 0.52734375, "learning_rate": 1.0702875399361024e-05, "loss": 2.4533, "step": 67 }, { "epoch": 0.02177742193755004, "grad_norm": 0.66015625, "learning_rate": 1.086261980830671e-05, "loss": 2.6732, "step": 68 }, { "epoch": 0.02209767814251401, "grad_norm": 0.609375, "learning_rate": 1.1022364217252397e-05, "loss": 2.09, "step": 69 }, { "epoch": 0.02241793434747798, "grad_norm": 0.451171875, "learning_rate": 1.1182108626198083e-05, "loss": 2.4639, "step": 70 }, { "epoch": 0.022738190552441955, "grad_norm": 0.5859375, "learning_rate": 1.134185303514377e-05, "loss": 2.195, "step": 71 }, { "epoch": 0.023058446757405926, "grad_norm": 0.55078125, "learning_rate": 1.1501597444089457e-05, "loss": 2.0553, "step": 72 }, { "epoch": 0.023378702962369897, "grad_norm": 0.46484375, "learning_rate": 1.1661341853035145e-05, "loss": 1.6604, "step": 73 }, { "epoch": 0.023698959167333868, "grad_norm": 0.48046875, "learning_rate": 1.182108626198083e-05, "loss": 2.1982, "step": 74 }, { "epoch": 0.02401921537229784, "grad_norm": 0.45703125, "learning_rate": 1.1980830670926518e-05, "loss": 2.2353, "step": 75 }, { "epoch": 0.02433947157726181, "grad_norm": 0.4609375, "learning_rate": 1.2140575079872205e-05, "loss": 2.0923, "step": 76 }, { "epoch": 0.02465972778222578, "grad_norm": 0.5078125, "learning_rate": 1.2300319488817893e-05, "loss": 2.379, "step": 77 }, { "epoch": 0.02497998398718975, "grad_norm": 1.0625, "learning_rate": 1.2460063897763578e-05, "loss": 1.7649, "step": 78 }, { "epoch": 0.02530024019215372, "grad_norm": 0.64453125, "learning_rate": 1.2619808306709266e-05, "loss": 2.475, "step": 79 }, { "epoch": 0.025620496397117692, "grad_norm": 0.453125, "learning_rate": 1.2779552715654951e-05, "loss": 2.4901, "step": 80 }, { "epoch": 0.025940752602081667, "grad_norm": 0.458984375, "learning_rate": 1.2939297124600639e-05, "loss": 2.4741, "step": 81 }, { "epoch": 0.026261008807045638, "grad_norm": 0.37890625, "learning_rate": 1.3099041533546328e-05, "loss": 2.1992, "step": 82 }, { "epoch": 0.02658126501200961, "grad_norm": 0.65234375, "learning_rate": 1.3258785942492014e-05, "loss": 2.6322, "step": 83 }, { "epoch": 0.02690152121697358, "grad_norm": 0.78515625, "learning_rate": 1.34185303514377e-05, "loss": 2.2932, "step": 84 }, { "epoch": 0.02722177742193755, "grad_norm": 0.462890625, "learning_rate": 1.3578274760383387e-05, "loss": 1.828, "step": 85 }, { "epoch": 0.02754203362690152, "grad_norm": 0.490234375, "learning_rate": 1.3738019169329076e-05, "loss": 2.0083, "step": 86 }, { "epoch": 0.02786228983186549, "grad_norm": 0.37109375, "learning_rate": 1.3897763578274761e-05, "loss": 2.0055, "step": 87 }, { "epoch": 0.028182546036829462, "grad_norm": 2.890625, "learning_rate": 1.4057507987220447e-05, "loss": 2.3489, "step": 88 }, { "epoch": 0.028502802241793433, "grad_norm": 0.5390625, "learning_rate": 1.4217252396166134e-05, "loss": 2.0319, "step": 89 }, { "epoch": 0.028823058446757407, "grad_norm": 0.6015625, "learning_rate": 1.4376996805111822e-05, "loss": 1.8418, "step": 90 }, { "epoch": 0.02914331465172138, "grad_norm": 0.6640625, "learning_rate": 1.453674121405751e-05, "loss": 2.1933, "step": 91 }, { "epoch": 0.02946357085668535, "grad_norm": 0.51953125, "learning_rate": 1.4696485623003195e-05, "loss": 2.238, "step": 92 }, { "epoch": 0.02978382706164932, "grad_norm": 0.494140625, "learning_rate": 1.485623003194888e-05, "loss": 1.8154, "step": 93 }, { "epoch": 0.03010408326661329, "grad_norm": 0.375, "learning_rate": 1.501597444089457e-05, "loss": 2.0664, "step": 94 }, { "epoch": 0.03042433947157726, "grad_norm": 0.56640625, "learning_rate": 1.5175718849840257e-05, "loss": 2.4674, "step": 95 }, { "epoch": 0.030744595676541232, "grad_norm": 0.404296875, "learning_rate": 1.533546325878594e-05, "loss": 2.5225, "step": 96 }, { "epoch": 0.031064851881505203, "grad_norm": 0.796875, "learning_rate": 1.549520766773163e-05, "loss": 2.5732, "step": 97 }, { "epoch": 0.031385108086469174, "grad_norm": 0.484375, "learning_rate": 1.565495207667732e-05, "loss": 2.4263, "step": 98 }, { "epoch": 0.03170536429143315, "grad_norm": 0.45703125, "learning_rate": 1.5814696485623005e-05, "loss": 1.8861, "step": 99 }, { "epoch": 0.032025620496397116, "grad_norm": 0.44921875, "learning_rate": 1.597444089456869e-05, "loss": 2.187, "step": 100 }, { "epoch": 0.03234587670136109, "grad_norm": 0.58203125, "learning_rate": 1.6134185303514376e-05, "loss": 2.4841, "step": 101 }, { "epoch": 0.03266613290632506, "grad_norm": 0.88671875, "learning_rate": 1.6293929712460065e-05, "loss": 2.1184, "step": 102 }, { "epoch": 0.03298638911128903, "grad_norm": 0.47265625, "learning_rate": 1.645367412140575e-05, "loss": 2.5642, "step": 103 }, { "epoch": 0.033306645316253006, "grad_norm": 0.318359375, "learning_rate": 1.6613418530351437e-05, "loss": 2.2907, "step": 104 }, { "epoch": 0.03362690152121697, "grad_norm": 0.3828125, "learning_rate": 1.6773162939297126e-05, "loss": 2.4367, "step": 105 }, { "epoch": 0.03394715772618095, "grad_norm": 0.4140625, "learning_rate": 1.693290734824281e-05, "loss": 2.5302, "step": 106 }, { "epoch": 0.034267413931144915, "grad_norm": 0.5703125, "learning_rate": 1.70926517571885e-05, "loss": 2.1865, "step": 107 }, { "epoch": 0.03458767013610889, "grad_norm": 0.55078125, "learning_rate": 1.7252396166134186e-05, "loss": 2.1695, "step": 108 }, { "epoch": 0.034907926341072856, "grad_norm": 0.52734375, "learning_rate": 1.7412140575079875e-05, "loss": 2.3653, "step": 109 }, { "epoch": 0.03522818254603683, "grad_norm": 0.396484375, "learning_rate": 1.757188498402556e-05, "loss": 1.9492, "step": 110 }, { "epoch": 0.0355484387510008, "grad_norm": 0.388671875, "learning_rate": 1.7731629392971247e-05, "loss": 2.7231, "step": 111 }, { "epoch": 0.03586869495596477, "grad_norm": 0.40234375, "learning_rate": 1.7891373801916932e-05, "loss": 2.2126, "step": 112 }, { "epoch": 0.03618895116092874, "grad_norm": 0.416015625, "learning_rate": 1.805111821086262e-05, "loss": 2.4667, "step": 113 }, { "epoch": 0.036509207365892714, "grad_norm": 0.64453125, "learning_rate": 1.8210862619808307e-05, "loss": 2.3994, "step": 114 }, { "epoch": 0.03682946357085669, "grad_norm": 0.5859375, "learning_rate": 1.8370607028753993e-05, "loss": 2.2501, "step": 115 }, { "epoch": 0.037149719775820655, "grad_norm": 1.703125, "learning_rate": 1.8530351437699682e-05, "loss": 2.5277, "step": 116 }, { "epoch": 0.03746997598078463, "grad_norm": 0.39453125, "learning_rate": 1.869009584664537e-05, "loss": 2.1283, "step": 117 }, { "epoch": 0.0377902321857486, "grad_norm": 1.6328125, "learning_rate": 1.8849840255591057e-05, "loss": 2.5671, "step": 118 }, { "epoch": 0.03811048839071257, "grad_norm": 1.796875, "learning_rate": 1.9009584664536742e-05, "loss": 1.7323, "step": 119 }, { "epoch": 0.03843074459567654, "grad_norm": 0.486328125, "learning_rate": 1.9169329073482428e-05, "loss": 2.4496, "step": 120 }, { "epoch": 0.03875100080064051, "grad_norm": 0.6171875, "learning_rate": 1.9329073482428117e-05, "loss": 2.3247, "step": 121 }, { "epoch": 0.03907125700560448, "grad_norm": 0.318359375, "learning_rate": 1.9488817891373803e-05, "loss": 2.044, "step": 122 }, { "epoch": 0.039391513210568455, "grad_norm": 0.55859375, "learning_rate": 1.964856230031949e-05, "loss": 2.2774, "step": 123 }, { "epoch": 0.03971176941553243, "grad_norm": 0.765625, "learning_rate": 1.9808306709265177e-05, "loss": 2.6735, "step": 124 }, { "epoch": 0.040032025620496396, "grad_norm": 0.498046875, "learning_rate": 1.9968051118210863e-05, "loss": 2.5676, "step": 125 }, { "epoch": 0.04035228182546037, "grad_norm": 0.41015625, "learning_rate": 2.0127795527156552e-05, "loss": 2.132, "step": 126 }, { "epoch": 0.04067253803042434, "grad_norm": 0.4609375, "learning_rate": 2.0287539936102238e-05, "loss": 2.6514, "step": 127 }, { "epoch": 0.04099279423538831, "grad_norm": 0.65234375, "learning_rate": 2.0447284345047924e-05, "loss": 2.6641, "step": 128 }, { "epoch": 0.04131305044035228, "grad_norm": 0.494140625, "learning_rate": 2.0607028753993613e-05, "loss": 2.442, "step": 129 }, { "epoch": 0.041633306645316254, "grad_norm": 0.3359375, "learning_rate": 2.07667731629393e-05, "loss": 1.4356, "step": 130 }, { "epoch": 0.04195356285028022, "grad_norm": 0.34765625, "learning_rate": 2.0926517571884984e-05, "loss": 1.8376, "step": 131 }, { "epoch": 0.042273819055244195, "grad_norm": 0.3671875, "learning_rate": 2.108626198083067e-05, "loss": 2.4422, "step": 132 }, { "epoch": 0.04259407526020817, "grad_norm": 0.578125, "learning_rate": 2.124600638977636e-05, "loss": 2.1663, "step": 133 }, { "epoch": 0.04291433146517214, "grad_norm": 0.6171875, "learning_rate": 2.1405750798722048e-05, "loss": 2.5161, "step": 134 }, { "epoch": 0.04323458767013611, "grad_norm": 0.6328125, "learning_rate": 2.1565495207667734e-05, "loss": 2.295, "step": 135 }, { "epoch": 0.04355484387510008, "grad_norm": 0.48046875, "learning_rate": 2.172523961661342e-05, "loss": 2.6238, "step": 136 }, { "epoch": 0.04387510008006405, "grad_norm": 1.9609375, "learning_rate": 2.188498402555911e-05, "loss": 2.0364, "step": 137 }, { "epoch": 0.04419535628502802, "grad_norm": 0.77734375, "learning_rate": 2.2044728434504794e-05, "loss": 2.7412, "step": 138 }, { "epoch": 0.044515612489991994, "grad_norm": 1.0703125, "learning_rate": 2.220447284345048e-05, "loss": 1.8156, "step": 139 }, { "epoch": 0.04483586869495596, "grad_norm": 0.91796875, "learning_rate": 2.2364217252396165e-05, "loss": 2.3815, "step": 140 }, { "epoch": 0.045156124899919936, "grad_norm": 0.439453125, "learning_rate": 2.2523961661341854e-05, "loss": 2.6271, "step": 141 }, { "epoch": 0.04547638110488391, "grad_norm": 1.7421875, "learning_rate": 2.268370607028754e-05, "loss": 1.6873, "step": 142 }, { "epoch": 0.04579663730984788, "grad_norm": 0.431640625, "learning_rate": 2.284345047923323e-05, "loss": 2.3417, "step": 143 }, { "epoch": 0.04611689351481185, "grad_norm": 0.4609375, "learning_rate": 2.3003194888178915e-05, "loss": 2.1776, "step": 144 }, { "epoch": 0.04643714971977582, "grad_norm": 0.498046875, "learning_rate": 2.3162939297124604e-05, "loss": 2.2616, "step": 145 }, { "epoch": 0.046757405924739794, "grad_norm": 0.60546875, "learning_rate": 2.332268370607029e-05, "loss": 2.5723, "step": 146 }, { "epoch": 0.04707766212970376, "grad_norm": 0.470703125, "learning_rate": 2.3482428115015975e-05, "loss": 2.4548, "step": 147 }, { "epoch": 0.047397918334667735, "grad_norm": 0.9375, "learning_rate": 2.364217252396166e-05, "loss": 2.4736, "step": 148 }, { "epoch": 0.0477181745396317, "grad_norm": 0.482421875, "learning_rate": 2.380191693290735e-05, "loss": 2.386, "step": 149 }, { "epoch": 0.04803843074459568, "grad_norm": 0.486328125, "learning_rate": 2.3961661341853036e-05, "loss": 2.644, "step": 150 }, { "epoch": 0.04835868694955965, "grad_norm": 0.58203125, "learning_rate": 2.412140575079872e-05, "loss": 2.1945, "step": 151 }, { "epoch": 0.04867894315452362, "grad_norm": 0.66796875, "learning_rate": 2.428115015974441e-05, "loss": 2.2383, "step": 152 }, { "epoch": 0.04899919935948759, "grad_norm": 0.4296875, "learning_rate": 2.44408945686901e-05, "loss": 2.5643, "step": 153 }, { "epoch": 0.04931945556445156, "grad_norm": 0.431640625, "learning_rate": 2.4600638977635785e-05, "loss": 1.6623, "step": 154 }, { "epoch": 0.049639711769415534, "grad_norm": 0.7890625, "learning_rate": 2.476038338658147e-05, "loss": 2.1212, "step": 155 }, { "epoch": 0.0499599679743795, "grad_norm": 0.453125, "learning_rate": 2.4920127795527157e-05, "loss": 2.0499, "step": 156 }, { "epoch": 0.050280224179343476, "grad_norm": 0.58984375, "learning_rate": 2.5079872204472842e-05, "loss": 2.004, "step": 157 }, { "epoch": 0.05060048038430744, "grad_norm": 0.50390625, "learning_rate": 2.523961661341853e-05, "loss": 2.4388, "step": 158 }, { "epoch": 0.05092073658927142, "grad_norm": 0.431640625, "learning_rate": 2.539936102236422e-05, "loss": 2.2752, "step": 159 }, { "epoch": 0.051240992794235385, "grad_norm": 0.470703125, "learning_rate": 2.5559105431309903e-05, "loss": 2.1717, "step": 160 }, { "epoch": 0.05156124899919936, "grad_norm": 0.46875, "learning_rate": 2.5718849840255592e-05, "loss": 2.3856, "step": 161 }, { "epoch": 0.05188150520416333, "grad_norm": 0.96484375, "learning_rate": 2.5878594249201278e-05, "loss": 1.7901, "step": 162 }, { "epoch": 0.0522017614091273, "grad_norm": 0.62890625, "learning_rate": 2.6038338658146967e-05, "loss": 2.1485, "step": 163 }, { "epoch": 0.052522017614091275, "grad_norm": 1.046875, "learning_rate": 2.6198083067092656e-05, "loss": 2.4288, "step": 164 }, { "epoch": 0.05284227381905524, "grad_norm": 0.333984375, "learning_rate": 2.6357827476038338e-05, "loss": 2.4144, "step": 165 }, { "epoch": 0.05316253002401922, "grad_norm": 0.390625, "learning_rate": 2.6517571884984027e-05, "loss": 2.4021, "step": 166 }, { "epoch": 0.053482786228983184, "grad_norm": 0.439453125, "learning_rate": 2.6677316293929716e-05, "loss": 2.2037, "step": 167 }, { "epoch": 0.05380304243394716, "grad_norm": 0.63671875, "learning_rate": 2.68370607028754e-05, "loss": 2.33, "step": 168 }, { "epoch": 0.054123298638911126, "grad_norm": 0.7421875, "learning_rate": 2.6996805111821088e-05, "loss": 2.3497, "step": 169 }, { "epoch": 0.0544435548438751, "grad_norm": 0.4765625, "learning_rate": 2.7156549520766773e-05, "loss": 2.0505, "step": 170 }, { "epoch": 0.054763811048839074, "grad_norm": 0.478515625, "learning_rate": 2.7316293929712462e-05, "loss": 2.5688, "step": 171 }, { "epoch": 0.05508406725380304, "grad_norm": 0.57421875, "learning_rate": 2.747603833865815e-05, "loss": 2.5961, "step": 172 }, { "epoch": 0.055404323458767016, "grad_norm": 0.490234375, "learning_rate": 2.7635782747603834e-05, "loss": 1.9069, "step": 173 }, { "epoch": 0.05572457966373098, "grad_norm": 0.498046875, "learning_rate": 2.7795527156549523e-05, "loss": 2.3172, "step": 174 }, { "epoch": 0.05604483586869496, "grad_norm": 0.67578125, "learning_rate": 2.7955271565495212e-05, "loss": 2.5545, "step": 175 }, { "epoch": 0.056365092073658925, "grad_norm": 0.75390625, "learning_rate": 2.8115015974440894e-05, "loss": 2.248, "step": 176 }, { "epoch": 0.0566853482786229, "grad_norm": 0.486328125, "learning_rate": 2.8274760383386583e-05, "loss": 2.4285, "step": 177 }, { "epoch": 0.057005604483586866, "grad_norm": 0.373046875, "learning_rate": 2.843450479233227e-05, "loss": 1.4118, "step": 178 }, { "epoch": 0.05732586068855084, "grad_norm": 0.6328125, "learning_rate": 2.8594249201277955e-05, "loss": 1.8937, "step": 179 }, { "epoch": 0.057646116893514815, "grad_norm": 0.482421875, "learning_rate": 2.8753993610223644e-05, "loss": 1.7478, "step": 180 }, { "epoch": 0.05796637309847878, "grad_norm": 0.55859375, "learning_rate": 2.891373801916933e-05, "loss": 1.8904, "step": 181 }, { "epoch": 0.05828662930344276, "grad_norm": 0.455078125, "learning_rate": 2.907348242811502e-05, "loss": 2.6784, "step": 182 }, { "epoch": 0.058606885508406724, "grad_norm": 0.515625, "learning_rate": 2.9233226837060707e-05, "loss": 2.4389, "step": 183 }, { "epoch": 0.0589271417133707, "grad_norm": 0.5703125, "learning_rate": 2.939297124600639e-05, "loss": 2.5659, "step": 184 }, { "epoch": 0.059247397918334666, "grad_norm": 0.498046875, "learning_rate": 2.955271565495208e-05, "loss": 2.0976, "step": 185 }, { "epoch": 0.05956765412329864, "grad_norm": 1.546875, "learning_rate": 2.971246006389776e-05, "loss": 2.5468, "step": 186 }, { "epoch": 0.05988791032826261, "grad_norm": 0.455078125, "learning_rate": 2.987220447284345e-05, "loss": 2.3513, "step": 187 }, { "epoch": 0.06020816653322658, "grad_norm": 0.5234375, "learning_rate": 3.003194888178914e-05, "loss": 1.9096, "step": 188 }, { "epoch": 0.060528422738190556, "grad_norm": 0.44921875, "learning_rate": 3.0191693290734825e-05, "loss": 1.768, "step": 189 }, { "epoch": 0.06084867894315452, "grad_norm": 0.404296875, "learning_rate": 3.0351437699680514e-05, "loss": 1.7802, "step": 190 }, { "epoch": 0.0611689351481185, "grad_norm": 0.46484375, "learning_rate": 3.0511182108626203e-05, "loss": 1.9335, "step": 191 }, { "epoch": 0.061489191353082465, "grad_norm": 0.439453125, "learning_rate": 3.067092651757188e-05, "loss": 2.3403, "step": 192 }, { "epoch": 0.06180944755804644, "grad_norm": 0.51171875, "learning_rate": 3.083067092651757e-05, "loss": 2.1596, "step": 193 }, { "epoch": 0.062129703763010406, "grad_norm": 0.390625, "learning_rate": 3.099041533546326e-05, "loss": 2.6126, "step": 194 }, { "epoch": 0.06244995996797438, "grad_norm": 0.322265625, "learning_rate": 3.115015974440895e-05, "loss": 1.772, "step": 195 }, { "epoch": 0.06277021617293835, "grad_norm": 0.63671875, "learning_rate": 3.130990415335464e-05, "loss": 2.694, "step": 196 }, { "epoch": 0.06309047237790232, "grad_norm": 0.484375, "learning_rate": 3.146964856230032e-05, "loss": 2.045, "step": 197 }, { "epoch": 0.0634107285828663, "grad_norm": 1.828125, "learning_rate": 3.162939297124601e-05, "loss": 1.7398, "step": 198 }, { "epoch": 0.06373098478783026, "grad_norm": 0.59375, "learning_rate": 3.17891373801917e-05, "loss": 2.344, "step": 199 }, { "epoch": 0.06405124099279423, "grad_norm": 0.5546875, "learning_rate": 3.194888178913738e-05, "loss": 2.1837, "step": 200 }, { "epoch": 0.06437149719775821, "grad_norm": 1.171875, "learning_rate": 3.210862619808307e-05, "loss": 2.0782, "step": 201 }, { "epoch": 0.06469175340272218, "grad_norm": 0.44921875, "learning_rate": 3.226837060702875e-05, "loss": 2.2655, "step": 202 }, { "epoch": 0.06501200960768615, "grad_norm": 0.462890625, "learning_rate": 3.242811501597444e-05, "loss": 2.55, "step": 203 }, { "epoch": 0.06533226581265011, "grad_norm": 0.59765625, "learning_rate": 3.258785942492013e-05, "loss": 2.2424, "step": 204 }, { "epoch": 0.0656525220176141, "grad_norm": 0.486328125, "learning_rate": 3.274760383386581e-05, "loss": 2.5929, "step": 205 }, { "epoch": 0.06597277822257806, "grad_norm": 0.400390625, "learning_rate": 3.29073482428115e-05, "loss": 2.2174, "step": 206 }, { "epoch": 0.06629303442754203, "grad_norm": 0.5078125, "learning_rate": 3.306709265175719e-05, "loss": 2.685, "step": 207 }, { "epoch": 0.06661329063250601, "grad_norm": 0.486328125, "learning_rate": 3.322683706070287e-05, "loss": 2.242, "step": 208 }, { "epoch": 0.06693354683746998, "grad_norm": 0.4375, "learning_rate": 3.338658146964856e-05, "loss": 1.9813, "step": 209 }, { "epoch": 0.06725380304243395, "grad_norm": 0.5390625, "learning_rate": 3.354632587859425e-05, "loss": 2.0788, "step": 210 }, { "epoch": 0.06757405924739791, "grad_norm": 0.478515625, "learning_rate": 3.3706070287539934e-05, "loss": 2.2187, "step": 211 }, { "epoch": 0.0678943154523619, "grad_norm": 0.4765625, "learning_rate": 3.386581469648562e-05, "loss": 2.2772, "step": 212 }, { "epoch": 0.06821457165732586, "grad_norm": 0.373046875, "learning_rate": 3.402555910543131e-05, "loss": 1.4853, "step": 213 }, { "epoch": 0.06853482786228983, "grad_norm": 0.53515625, "learning_rate": 3.4185303514377e-05, "loss": 2.0633, "step": 214 }, { "epoch": 0.0688550840672538, "grad_norm": 0.6953125, "learning_rate": 3.434504792332269e-05, "loss": 2.3552, "step": 215 }, { "epoch": 0.06917534027221778, "grad_norm": 0.498046875, "learning_rate": 3.450479233226837e-05, "loss": 2.7011, "step": 216 }, { "epoch": 0.06949559647718175, "grad_norm": 0.439453125, "learning_rate": 3.466453674121406e-05, "loss": 2.0144, "step": 217 }, { "epoch": 0.06981585268214571, "grad_norm": 0.640625, "learning_rate": 3.482428115015975e-05, "loss": 2.293, "step": 218 }, { "epoch": 0.0701361088871097, "grad_norm": 0.376953125, "learning_rate": 3.498402555910543e-05, "loss": 2.2046, "step": 219 }, { "epoch": 0.07045636509207366, "grad_norm": 1.0390625, "learning_rate": 3.514376996805112e-05, "loss": 2.0654, "step": 220 }, { "epoch": 0.07077662129703763, "grad_norm": 1.1484375, "learning_rate": 3.5303514376996804e-05, "loss": 2.3501, "step": 221 }, { "epoch": 0.0710968775020016, "grad_norm": 0.51171875, "learning_rate": 3.546325878594249e-05, "loss": 2.3258, "step": 222 }, { "epoch": 0.07141713370696558, "grad_norm": 0.41796875, "learning_rate": 3.562300319488818e-05, "loss": 2.3782, "step": 223 }, { "epoch": 0.07173738991192954, "grad_norm": 0.5078125, "learning_rate": 3.5782747603833865e-05, "loss": 2.2688, "step": 224 }, { "epoch": 0.07205764611689351, "grad_norm": 0.67578125, "learning_rate": 3.5942492012779554e-05, "loss": 2.2505, "step": 225 }, { "epoch": 0.07237790232185748, "grad_norm": 1.0859375, "learning_rate": 3.610223642172524e-05, "loss": 2.4121, "step": 226 }, { "epoch": 0.07269815852682146, "grad_norm": 0.55078125, "learning_rate": 3.6261980830670925e-05, "loss": 2.4804, "step": 227 }, { "epoch": 0.07301841473178543, "grad_norm": 0.6640625, "learning_rate": 3.6421725239616614e-05, "loss": 2.4248, "step": 228 }, { "epoch": 0.0733386709367494, "grad_norm": 0.91015625, "learning_rate": 3.65814696485623e-05, "loss": 2.2826, "step": 229 }, { "epoch": 0.07365892714171338, "grad_norm": 0.392578125, "learning_rate": 3.6741214057507985e-05, "loss": 1.8805, "step": 230 }, { "epoch": 0.07397918334667734, "grad_norm": 1.265625, "learning_rate": 3.6900958466453675e-05, "loss": 1.9896, "step": 231 }, { "epoch": 0.07429943955164131, "grad_norm": 1.515625, "learning_rate": 3.7060702875399364e-05, "loss": 2.6494, "step": 232 }, { "epoch": 0.07461969575660528, "grad_norm": 0.5390625, "learning_rate": 3.722044728434505e-05, "loss": 1.5742, "step": 233 }, { "epoch": 0.07493995196156926, "grad_norm": 0.353515625, "learning_rate": 3.738019169329074e-05, "loss": 1.9401, "step": 234 }, { "epoch": 0.07526020816653323, "grad_norm": 0.46875, "learning_rate": 3.7539936102236424e-05, "loss": 2.5195, "step": 235 }, { "epoch": 0.0755804643714972, "grad_norm": 0.4375, "learning_rate": 3.769968051118211e-05, "loss": 2.1442, "step": 236 }, { "epoch": 0.07590072057646118, "grad_norm": 2.015625, "learning_rate": 3.7859424920127795e-05, "loss": 2.2005, "step": 237 }, { "epoch": 0.07622097678142514, "grad_norm": 0.4609375, "learning_rate": 3.8019169329073485e-05, "loss": 2.5538, "step": 238 }, { "epoch": 0.07654123298638911, "grad_norm": 0.6875, "learning_rate": 3.8178913738019174e-05, "loss": 2.22, "step": 239 }, { "epoch": 0.07686148919135308, "grad_norm": 0.54296875, "learning_rate": 3.8338658146964856e-05, "loss": 2.4174, "step": 240 }, { "epoch": 0.07718174539631706, "grad_norm": 0.478515625, "learning_rate": 3.8498402555910545e-05, "loss": 2.3724, "step": 241 }, { "epoch": 0.07750200160128103, "grad_norm": 0.59375, "learning_rate": 3.8658146964856234e-05, "loss": 2.1111, "step": 242 }, { "epoch": 0.077822257806245, "grad_norm": 0.6171875, "learning_rate": 3.8817891373801916e-05, "loss": 2.2114, "step": 243 }, { "epoch": 0.07814251401120896, "grad_norm": 0.5, "learning_rate": 3.8977635782747605e-05, "loss": 2.718, "step": 244 }, { "epoch": 0.07846277021617294, "grad_norm": 0.5859375, "learning_rate": 3.913738019169329e-05, "loss": 2.1236, "step": 245 }, { "epoch": 0.07878302642113691, "grad_norm": 0.6171875, "learning_rate": 3.929712460063898e-05, "loss": 2.3918, "step": 246 }, { "epoch": 0.07910328262610088, "grad_norm": 0.474609375, "learning_rate": 3.9456869009584666e-05, "loss": 1.8762, "step": 247 }, { "epoch": 0.07942353883106486, "grad_norm": 0.61328125, "learning_rate": 3.9616613418530355e-05, "loss": 2.2969, "step": 248 }, { "epoch": 0.07974379503602883, "grad_norm": 0.4375, "learning_rate": 3.9776357827476044e-05, "loss": 2.1578, "step": 249 }, { "epoch": 0.08006405124099279, "grad_norm": 0.46484375, "learning_rate": 3.9936102236421726e-05, "loss": 2.5525, "step": 250 }, { "epoch": 0.08038430744595676, "grad_norm": 0.4921875, "learning_rate": 4.0095846645367415e-05, "loss": 2.4644, "step": 251 }, { "epoch": 0.08070456365092074, "grad_norm": 0.400390625, "learning_rate": 4.0255591054313104e-05, "loss": 1.2111, "step": 252 }, { "epoch": 0.08102481985588471, "grad_norm": 0.451171875, "learning_rate": 4.041533546325879e-05, "loss": 2.522, "step": 253 }, { "epoch": 0.08134507606084868, "grad_norm": 0.41796875, "learning_rate": 4.0575079872204476e-05, "loss": 1.5268, "step": 254 }, { "epoch": 0.08166533226581266, "grad_norm": 0.408203125, "learning_rate": 4.0734824281150165e-05, "loss": 2.331, "step": 255 }, { "epoch": 0.08198558847077662, "grad_norm": 0.8828125, "learning_rate": 4.089456869009585e-05, "loss": 2.2693, "step": 256 }, { "epoch": 0.08230584467574059, "grad_norm": 0.58203125, "learning_rate": 4.1054313099041536e-05, "loss": 2.0041, "step": 257 }, { "epoch": 0.08262610088070456, "grad_norm": 0.30078125, "learning_rate": 4.1214057507987225e-05, "loss": 1.814, "step": 258 }, { "epoch": 0.08294635708566854, "grad_norm": 0.59375, "learning_rate": 4.137380191693291e-05, "loss": 2.1807, "step": 259 }, { "epoch": 0.08326661329063251, "grad_norm": 0.431640625, "learning_rate": 4.15335463258786e-05, "loss": 1.8303, "step": 260 }, { "epoch": 0.08358686949559647, "grad_norm": 0.46484375, "learning_rate": 4.169329073482428e-05, "loss": 2.0817, "step": 261 }, { "epoch": 0.08390712570056044, "grad_norm": 0.384765625, "learning_rate": 4.185303514376997e-05, "loss": 2.572, "step": 262 }, { "epoch": 0.08422738190552442, "grad_norm": 0.5234375, "learning_rate": 4.201277955271566e-05, "loss": 2.4427, "step": 263 }, { "epoch": 0.08454763811048839, "grad_norm": 0.470703125, "learning_rate": 4.217252396166134e-05, "loss": 1.6031, "step": 264 }, { "epoch": 0.08486789431545236, "grad_norm": 0.388671875, "learning_rate": 4.233226837060703e-05, "loss": 2.3756, "step": 265 }, { "epoch": 0.08518815052041634, "grad_norm": 0.75, "learning_rate": 4.249201277955272e-05, "loss": 2.5675, "step": 266 }, { "epoch": 0.0855084067253803, "grad_norm": 0.6171875, "learning_rate": 4.265175718849841e-05, "loss": 2.3992, "step": 267 }, { "epoch": 0.08582866293034427, "grad_norm": 2.296875, "learning_rate": 4.2811501597444096e-05, "loss": 1.8396, "step": 268 }, { "epoch": 0.08614891913530824, "grad_norm": 0.625, "learning_rate": 4.297124600638978e-05, "loss": 1.8471, "step": 269 }, { "epoch": 0.08646917534027222, "grad_norm": 0.6171875, "learning_rate": 4.313099041533547e-05, "loss": 2.4466, "step": 270 }, { "epoch": 0.08678943154523619, "grad_norm": 0.384765625, "learning_rate": 4.3290734824281156e-05, "loss": 2.4973, "step": 271 }, { "epoch": 0.08710968775020016, "grad_norm": 0.5234375, "learning_rate": 4.345047923322684e-05, "loss": 2.3791, "step": 272 }, { "epoch": 0.08742994395516412, "grad_norm": 0.455078125, "learning_rate": 4.361022364217253e-05, "loss": 2.2905, "step": 273 }, { "epoch": 0.0877502001601281, "grad_norm": 0.44140625, "learning_rate": 4.376996805111822e-05, "loss": 2.1515, "step": 274 }, { "epoch": 0.08807045636509207, "grad_norm": 0.4921875, "learning_rate": 4.39297124600639e-05, "loss": 2.2641, "step": 275 }, { "epoch": 0.08839071257005604, "grad_norm": 0.7265625, "learning_rate": 4.408945686900959e-05, "loss": 2.4807, "step": 276 }, { "epoch": 0.08871096877502002, "grad_norm": 0.419921875, "learning_rate": 4.424920127795527e-05, "loss": 2.3933, "step": 277 }, { "epoch": 0.08903122497998399, "grad_norm": 0.53125, "learning_rate": 4.440894568690096e-05, "loss": 1.7232, "step": 278 }, { "epoch": 0.08935148118494796, "grad_norm": 3.390625, "learning_rate": 4.456869009584665e-05, "loss": 2.3984, "step": 279 }, { "epoch": 0.08967173738991192, "grad_norm": 0.4921875, "learning_rate": 4.472843450479233e-05, "loss": 2.2911, "step": 280 }, { "epoch": 0.0899919935948759, "grad_norm": 0.5234375, "learning_rate": 4.488817891373802e-05, "loss": 2.3337, "step": 281 }, { "epoch": 0.09031224979983987, "grad_norm": 0.859375, "learning_rate": 4.504792332268371e-05, "loss": 2.0545, "step": 282 }, { "epoch": 0.09063250600480384, "grad_norm": 0.52734375, "learning_rate": 4.520766773162939e-05, "loss": 2.1376, "step": 283 }, { "epoch": 0.09095276220976782, "grad_norm": 0.55859375, "learning_rate": 4.536741214057508e-05, "loss": 2.1938, "step": 284 }, { "epoch": 0.09127301841473179, "grad_norm": 0.53515625, "learning_rate": 4.552715654952077e-05, "loss": 2.4041, "step": 285 }, { "epoch": 0.09159327461969576, "grad_norm": 0.58203125, "learning_rate": 4.568690095846646e-05, "loss": 2.2543, "step": 286 }, { "epoch": 0.09191353082465972, "grad_norm": 0.55859375, "learning_rate": 4.584664536741215e-05, "loss": 2.0115, "step": 287 }, { "epoch": 0.0922337870296237, "grad_norm": 0.40234375, "learning_rate": 4.600638977635783e-05, "loss": 2.3996, "step": 288 }, { "epoch": 0.09255404323458767, "grad_norm": 2.546875, "learning_rate": 4.616613418530352e-05, "loss": 2.7077, "step": 289 }, { "epoch": 0.09287429943955164, "grad_norm": 0.63671875, "learning_rate": 4.632587859424921e-05, "loss": 2.0203, "step": 290 }, { "epoch": 0.0931945556445156, "grad_norm": 0.455078125, "learning_rate": 4.648562300319489e-05, "loss": 2.2094, "step": 291 }, { "epoch": 0.09351481184947959, "grad_norm": 0.40234375, "learning_rate": 4.664536741214058e-05, "loss": 2.2682, "step": 292 }, { "epoch": 0.09383506805444355, "grad_norm": 0.5234375, "learning_rate": 4.680511182108626e-05, "loss": 2.1374, "step": 293 }, { "epoch": 0.09415532425940752, "grad_norm": 0.9609375, "learning_rate": 4.696485623003195e-05, "loss": 2.272, "step": 294 }, { "epoch": 0.0944755804643715, "grad_norm": 0.5078125, "learning_rate": 4.712460063897764e-05, "loss": 2.384, "step": 295 }, { "epoch": 0.09479583666933547, "grad_norm": 0.671875, "learning_rate": 4.728434504792332e-05, "loss": 2.4198, "step": 296 }, { "epoch": 0.09511609287429944, "grad_norm": 0.3125, "learning_rate": 4.744408945686901e-05, "loss": 2.3007, "step": 297 }, { "epoch": 0.0954363490792634, "grad_norm": 0.357421875, "learning_rate": 4.76038338658147e-05, "loss": 2.6156, "step": 298 }, { "epoch": 0.09575660528422739, "grad_norm": 0.462890625, "learning_rate": 4.776357827476038e-05, "loss": 2.2731, "step": 299 }, { "epoch": 0.09607686148919135, "grad_norm": 0.349609375, "learning_rate": 4.792332268370607e-05, "loss": 2.5316, "step": 300 }, { "epoch": 0.09639711769415532, "grad_norm": 3.75, "learning_rate": 4.8083067092651754e-05, "loss": 1.9556, "step": 301 }, { "epoch": 0.0967173738991193, "grad_norm": 0.408203125, "learning_rate": 4.824281150159744e-05, "loss": 2.4008, "step": 302 }, { "epoch": 0.09703763010408327, "grad_norm": 0.5703125, "learning_rate": 4.840255591054313e-05, "loss": 2.2126, "step": 303 }, { "epoch": 0.09735788630904724, "grad_norm": 0.51171875, "learning_rate": 4.856230031948882e-05, "loss": 1.9912, "step": 304 }, { "epoch": 0.0976781425140112, "grad_norm": 0.65625, "learning_rate": 4.872204472843451e-05, "loss": 2.3782, "step": 305 }, { "epoch": 0.09799839871897519, "grad_norm": 0.94140625, "learning_rate": 4.88817891373802e-05, "loss": 2.2945, "step": 306 }, { "epoch": 0.09831865492393915, "grad_norm": 0.416015625, "learning_rate": 4.904153354632588e-05, "loss": 2.7067, "step": 307 }, { "epoch": 0.09863891112890312, "grad_norm": 0.423828125, "learning_rate": 4.920127795527157e-05, "loss": 2.0701, "step": 308 }, { "epoch": 0.09895916733386709, "grad_norm": 0.609375, "learning_rate": 4.936102236421725e-05, "loss": 2.5141, "step": 309 }, { "epoch": 0.09927942353883107, "grad_norm": 0.4609375, "learning_rate": 4.952076677316294e-05, "loss": 2.1579, "step": 310 }, { "epoch": 0.09959967974379504, "grad_norm": 0.498046875, "learning_rate": 4.968051118210863e-05, "loss": 2.4784, "step": 311 }, { "epoch": 0.099919935948759, "grad_norm": 0.498046875, "learning_rate": 4.984025559105431e-05, "loss": 2.285, "step": 312 }, { "epoch": 0.10024019215372298, "grad_norm": 0.74609375, "learning_rate": 5e-05, "loss": 2.4196, "step": 313 }, { "epoch": 0.10056044835868695, "grad_norm": 0.60546875, "learning_rate": 4.9999984364699426e-05, "loss": 1.4867, "step": 314 }, { "epoch": 0.10088070456365092, "grad_norm": 0.50390625, "learning_rate": 4.999993745881724e-05, "loss": 2.2844, "step": 315 }, { "epoch": 0.10120096076861489, "grad_norm": 0.703125, "learning_rate": 4.9999859282412125e-05, "loss": 1.8481, "step": 316 }, { "epoch": 0.10152121697357887, "grad_norm": 0.59375, "learning_rate": 4.9999749835581864e-05, "loss": 2.4302, "step": 317 }, { "epoch": 0.10184147317854284, "grad_norm": 0.494140625, "learning_rate": 4.999960911846336e-05, "loss": 2.3678, "step": 318 }, { "epoch": 0.1021617293835068, "grad_norm": 0.5546875, "learning_rate": 4.999943713123261e-05, "loss": 2.1427, "step": 319 }, { "epoch": 0.10248198558847077, "grad_norm": 0.47265625, "learning_rate": 4.9999233874104755e-05, "loss": 1.591, "step": 320 }, { "epoch": 0.10280224179343475, "grad_norm": 0.4765625, "learning_rate": 4.9998999347334036e-05, "loss": 2.353, "step": 321 }, { "epoch": 0.10312249799839872, "grad_norm": 0.3984375, "learning_rate": 4.9998733551213795e-05, "loss": 2.0951, "step": 322 }, { "epoch": 0.10344275420336269, "grad_norm": 0.5234375, "learning_rate": 4.9998436486076495e-05, "loss": 2.4225, "step": 323 }, { "epoch": 0.10376301040832667, "grad_norm": 0.419921875, "learning_rate": 4.999810815229372e-05, "loss": 1.8435, "step": 324 }, { "epoch": 0.10408326661329063, "grad_norm": 0.546875, "learning_rate": 4.999774855027615e-05, "loss": 2.0037, "step": 325 }, { "epoch": 0.1044035228182546, "grad_norm": 0.4453125, "learning_rate": 4.999735768047359e-05, "loss": 2.6103, "step": 326 }, { "epoch": 0.10472377902321857, "grad_norm": 0.4765625, "learning_rate": 4.999693554337495e-05, "loss": 2.0573, "step": 327 }, { "epoch": 0.10504403522818255, "grad_norm": 0.39453125, "learning_rate": 4.9996482139508236e-05, "loss": 2.1116, "step": 328 }, { "epoch": 0.10536429143314652, "grad_norm": 0.37890625, "learning_rate": 4.999599746944059e-05, "loss": 2.677, "step": 329 }, { "epoch": 0.10568454763811048, "grad_norm": 0.48046875, "learning_rate": 4.9995481533778256e-05, "loss": 2.2031, "step": 330 }, { "epoch": 0.10600480384307447, "grad_norm": 0.6328125, "learning_rate": 4.999493433316656e-05, "loss": 2.1728, "step": 331 }, { "epoch": 0.10632506004803843, "grad_norm": 0.494140625, "learning_rate": 4.9994355868289965e-05, "loss": 2.239, "step": 332 }, { "epoch": 0.1066453162530024, "grad_norm": 0.431640625, "learning_rate": 4.999374613987202e-05, "loss": 2.5327, "step": 333 }, { "epoch": 0.10696557245796637, "grad_norm": 0.9765625, "learning_rate": 4.9993105148675405e-05, "loss": 2.2862, "step": 334 }, { "epoch": 0.10728582866293035, "grad_norm": 0.36328125, "learning_rate": 4.9992432895501874e-05, "loss": 2.0678, "step": 335 }, { "epoch": 0.10760608486789432, "grad_norm": 0.671875, "learning_rate": 4.99917293811923e-05, "loss": 2.2222, "step": 336 }, { "epoch": 0.10792634107285828, "grad_norm": 0.474609375, "learning_rate": 4.9990994606626647e-05, "loss": 2.195, "step": 337 }, { "epoch": 0.10824659727782225, "grad_norm": 2.109375, "learning_rate": 4.9990228572724e-05, "loss": 2.0068, "step": 338 }, { "epoch": 0.10856685348278623, "grad_norm": 0.88671875, "learning_rate": 4.998943128044254e-05, "loss": 2.0167, "step": 339 }, { "epoch": 0.1088871096877502, "grad_norm": 0.443359375, "learning_rate": 4.9988602730779515e-05, "loss": 2.3055, "step": 340 }, { "epoch": 0.10920736589271417, "grad_norm": 0.439453125, "learning_rate": 4.998774292477131e-05, "loss": 2.2542, "step": 341 }, { "epoch": 0.10952762209767815, "grad_norm": 0.41015625, "learning_rate": 4.99868518634934e-05, "loss": 2.564, "step": 342 }, { "epoch": 0.10984787830264212, "grad_norm": 0.4375, "learning_rate": 4.998592954806033e-05, "loss": 2.2191, "step": 343 }, { "epoch": 0.11016813450760608, "grad_norm": 0.546875, "learning_rate": 4.998497597962576e-05, "loss": 2.5834, "step": 344 }, { "epoch": 0.11048839071257005, "grad_norm": 0.51171875, "learning_rate": 4.998399115938244e-05, "loss": 2.6079, "step": 345 }, { "epoch": 0.11080864691753403, "grad_norm": 0.6171875, "learning_rate": 4.99829750885622e-05, "loss": 2.0724, "step": 346 }, { "epoch": 0.111128903122498, "grad_norm": 0.51171875, "learning_rate": 4.9981927768435964e-05, "loss": 2.0157, "step": 347 }, { "epoch": 0.11144915932746197, "grad_norm": 0.58203125, "learning_rate": 4.998084920031376e-05, "loss": 2.4644, "step": 348 }, { "epoch": 0.11176941553242595, "grad_norm": 0.388671875, "learning_rate": 4.997973938554466e-05, "loss": 2.6107, "step": 349 }, { "epoch": 0.11208967173738991, "grad_norm": 0.4921875, "learning_rate": 4.997859832551689e-05, "loss": 2.3877, "step": 350 }, { "epoch": 0.11240992794235388, "grad_norm": 0.466796875, "learning_rate": 4.997742602165767e-05, "loss": 2.179, "step": 351 }, { "epoch": 0.11273018414731785, "grad_norm": 1.2890625, "learning_rate": 4.997622247543338e-05, "loss": 2.5496, "step": 352 }, { "epoch": 0.11305044035228183, "grad_norm": 0.515625, "learning_rate": 4.9974987688349415e-05, "loss": 1.8469, "step": 353 }, { "epoch": 0.1133706965572458, "grad_norm": 0.5703125, "learning_rate": 4.9973721661950295e-05, "loss": 2.3798, "step": 354 }, { "epoch": 0.11369095276220977, "grad_norm": 0.875, "learning_rate": 4.9972424397819596e-05, "loss": 2.4496, "step": 355 }, { "epoch": 0.11401120896717373, "grad_norm": 0.47265625, "learning_rate": 4.997109589757997e-05, "loss": 2.0667, "step": 356 }, { "epoch": 0.11433146517213771, "grad_norm": 3.4375, "learning_rate": 4.996973616289312e-05, "loss": 2.1528, "step": 357 }, { "epoch": 0.11465172137710168, "grad_norm": 0.4609375, "learning_rate": 4.996834519545985e-05, "loss": 1.8263, "step": 358 }, { "epoch": 0.11497197758206565, "grad_norm": 0.466796875, "learning_rate": 4.996692299702e-05, "loss": 2.629, "step": 359 }, { "epoch": 0.11529223378702963, "grad_norm": 0.55859375, "learning_rate": 4.996546956935252e-05, "loss": 2.4788, "step": 360 }, { "epoch": 0.1156124899919936, "grad_norm": 0.41015625, "learning_rate": 4.9963984914275363e-05, "loss": 2.2397, "step": 361 }, { "epoch": 0.11593274619695756, "grad_norm": 0.5078125, "learning_rate": 4.996246903364559e-05, "loss": 2.3035, "step": 362 }, { "epoch": 0.11625300240192153, "grad_norm": 0.64453125, "learning_rate": 4.996092192935929e-05, "loss": 2.6009, "step": 363 }, { "epoch": 0.11657325860688551, "grad_norm": 0.52734375, "learning_rate": 4.99593436033516e-05, "loss": 2.3057, "step": 364 }, { "epoch": 0.11689351481184948, "grad_norm": 0.42578125, "learning_rate": 4.9957734057596774e-05, "loss": 2.5767, "step": 365 }, { "epoch": 0.11721377101681345, "grad_norm": 0.466796875, "learning_rate": 4.995609329410804e-05, "loss": 1.9374, "step": 366 }, { "epoch": 0.11753402722177742, "grad_norm": 0.640625, "learning_rate": 4.995442131493771e-05, "loss": 1.9711, "step": 367 }, { "epoch": 0.1178542834267414, "grad_norm": 0.70703125, "learning_rate": 4.9952718122177126e-05, "loss": 2.3285, "step": 368 }, { "epoch": 0.11817453963170536, "grad_norm": 0.546875, "learning_rate": 4.9950983717956703e-05, "loss": 2.0349, "step": 369 }, { "epoch": 0.11849479583666933, "grad_norm": 0.578125, "learning_rate": 4.994921810444586e-05, "loss": 2.1336, "step": 370 }, { "epoch": 0.11881505204163331, "grad_norm": 0.447265625, "learning_rate": 4.994742128385307e-05, "loss": 2.319, "step": 371 }, { "epoch": 0.11913530824659728, "grad_norm": 0.408203125, "learning_rate": 4.994559325842585e-05, "loss": 2.0951, "step": 372 }, { "epoch": 0.11945556445156125, "grad_norm": 0.9765625, "learning_rate": 4.994373403045073e-05, "loss": 2.4311, "step": 373 }, { "epoch": 0.11977582065652521, "grad_norm": 0.6484375, "learning_rate": 4.994184360225328e-05, "loss": 1.8299, "step": 374 }, { "epoch": 0.1200960768614892, "grad_norm": 0.435546875, "learning_rate": 4.993992197619809e-05, "loss": 2.4937, "step": 375 }, { "epoch": 0.12041633306645316, "grad_norm": 0.486328125, "learning_rate": 4.9937969154688766e-05, "loss": 2.3098, "step": 376 }, { "epoch": 0.12073658927141713, "grad_norm": 0.55078125, "learning_rate": 4.993598514016797e-05, "loss": 2.2064, "step": 377 }, { "epoch": 0.12105684547638111, "grad_norm": 0.48828125, "learning_rate": 4.9933969935117335e-05, "loss": 2.0565, "step": 378 }, { "epoch": 0.12137710168134508, "grad_norm": 0.453125, "learning_rate": 4.9931923542057534e-05, "loss": 2.6077, "step": 379 }, { "epoch": 0.12169735788630905, "grad_norm": 0.375, "learning_rate": 4.9929845963548246e-05, "loss": 2.3129, "step": 380 }, { "epoch": 0.12201761409127301, "grad_norm": 0.48046875, "learning_rate": 4.9927737202188155e-05, "loss": 2.5353, "step": 381 }, { "epoch": 0.122337870296237, "grad_norm": 0.439453125, "learning_rate": 4.992559726061494e-05, "loss": 2.0995, "step": 382 }, { "epoch": 0.12265812650120096, "grad_norm": 0.41796875, "learning_rate": 4.9923426141505305e-05, "loss": 1.9616, "step": 383 }, { "epoch": 0.12297838270616493, "grad_norm": 0.40625, "learning_rate": 4.9921223847574935e-05, "loss": 2.1245, "step": 384 }, { "epoch": 0.1232986389111289, "grad_norm": 0.7734375, "learning_rate": 4.991899038157851e-05, "loss": 2.2955, "step": 385 }, { "epoch": 0.12361889511609288, "grad_norm": 0.59765625, "learning_rate": 4.991672574630971e-05, "loss": 2.2364, "step": 386 }, { "epoch": 0.12393915132105685, "grad_norm": 0.4609375, "learning_rate": 4.9914429944601185e-05, "loss": 2.2836, "step": 387 }, { "epoch": 0.12425940752602081, "grad_norm": 0.486328125, "learning_rate": 4.991210297932459e-05, "loss": 1.9992, "step": 388 }, { "epoch": 0.1245796637309848, "grad_norm": 0.33984375, "learning_rate": 4.9909744853390534e-05, "loss": 2.2879, "step": 389 }, { "epoch": 0.12489991993594876, "grad_norm": 0.87109375, "learning_rate": 4.990735556974863e-05, "loss": 2.2375, "step": 390 }, { "epoch": 0.12522017614091274, "grad_norm": 0.59375, "learning_rate": 4.990493513138744e-05, "loss": 2.5697, "step": 391 }, { "epoch": 0.1255404323458767, "grad_norm": 0.40234375, "learning_rate": 4.990248354133452e-05, "loss": 1.9927, "step": 392 }, { "epoch": 0.12586068855084068, "grad_norm": 0.5, "learning_rate": 4.9900000802656376e-05, "loss": 2.2876, "step": 393 }, { "epoch": 0.12618094475580463, "grad_norm": 0.53515625, "learning_rate": 4.989748691845847e-05, "loss": 2.3259, "step": 394 }, { "epoch": 0.1265012009607686, "grad_norm": 0.53515625, "learning_rate": 4.989494189188523e-05, "loss": 2.6403, "step": 395 }, { "epoch": 0.1268214571657326, "grad_norm": 0.84765625, "learning_rate": 4.989236572612004e-05, "loss": 2.4835, "step": 396 }, { "epoch": 0.12714171337069655, "grad_norm": 0.45703125, "learning_rate": 4.988975842438523e-05, "loss": 2.2297, "step": 397 }, { "epoch": 0.12746196957566053, "grad_norm": 0.65234375, "learning_rate": 4.9887119989942073e-05, "loss": 2.4837, "step": 398 }, { "epoch": 0.1277822257806245, "grad_norm": 0.478515625, "learning_rate": 4.9884450426090786e-05, "loss": 2.9118, "step": 399 }, { "epoch": 0.12810248198558846, "grad_norm": 0.51171875, "learning_rate": 4.9881749736170525e-05, "loss": 2.4227, "step": 400 }, { "epoch": 0.12842273819055244, "grad_norm": 0.5234375, "learning_rate": 4.987901792355938e-05, "loss": 2.6651, "step": 401 }, { "epoch": 0.12874299439551642, "grad_norm": 0.55859375, "learning_rate": 4.987625499167436e-05, "loss": 2.61, "step": 402 }, { "epoch": 0.12906325060048038, "grad_norm": 0.341796875, "learning_rate": 4.9873460943971416e-05, "loss": 2.1197, "step": 403 }, { "epoch": 0.12938350680544436, "grad_norm": 0.58984375, "learning_rate": 4.987063578394541e-05, "loss": 2.6976, "step": 404 }, { "epoch": 0.1297037630104083, "grad_norm": 0.65234375, "learning_rate": 4.986777951513011e-05, "loss": 2.242, "step": 405 }, { "epoch": 0.1300240192153723, "grad_norm": 0.46484375, "learning_rate": 4.9864892141098215e-05, "loss": 2.3003, "step": 406 }, { "epoch": 0.13034427542033628, "grad_norm": 0.396484375, "learning_rate": 4.986197366546133e-05, "loss": 1.8946, "step": 407 }, { "epoch": 0.13066453162530023, "grad_norm": 0.349609375, "learning_rate": 4.985902409186993e-05, "loss": 2.1781, "step": 408 }, { "epoch": 0.1309847878302642, "grad_norm": 0.515625, "learning_rate": 4.985604342401343e-05, "loss": 1.9627, "step": 409 }, { "epoch": 0.1313050440352282, "grad_norm": 0.4140625, "learning_rate": 4.9853031665620116e-05, "loss": 2.161, "step": 410 }, { "epoch": 0.13162530024019214, "grad_norm": 1.7734375, "learning_rate": 4.984998882045717e-05, "loss": 2.3552, "step": 411 }, { "epoch": 0.13194555644515613, "grad_norm": 0.439453125, "learning_rate": 4.9846914892330654e-05, "loss": 2.2441, "step": 412 }, { "epoch": 0.1322658126501201, "grad_norm": 0.396484375, "learning_rate": 4.984380988508551e-05, "loss": 1.9077, "step": 413 }, { "epoch": 0.13258606885508406, "grad_norm": 0.48828125, "learning_rate": 4.9840673802605566e-05, "loss": 2.5129, "step": 414 }, { "epoch": 0.13290632506004804, "grad_norm": 0.46875, "learning_rate": 4.98375066488135e-05, "loss": 2.2193, "step": 415 }, { "epoch": 0.13322658126501202, "grad_norm": 0.48828125, "learning_rate": 4.983430842767086e-05, "loss": 2.5459, "step": 416 }, { "epoch": 0.13354683746997598, "grad_norm": 0.490234375, "learning_rate": 4.9831079143178066e-05, "loss": 2.6023, "step": 417 }, { "epoch": 0.13386709367493996, "grad_norm": 9.5, "learning_rate": 4.982781879937438e-05, "loss": 2.1457, "step": 418 }, { "epoch": 0.1341873498799039, "grad_norm": 0.404296875, "learning_rate": 4.982452740033793e-05, "loss": 2.0819, "step": 419 }, { "epoch": 0.1345076060848679, "grad_norm": 0.703125, "learning_rate": 4.982120495018566e-05, "loss": 2.6695, "step": 420 }, { "epoch": 0.13482786228983187, "grad_norm": 0.59765625, "learning_rate": 4.981785145307337e-05, "loss": 2.4198, "step": 421 }, { "epoch": 0.13514811849479583, "grad_norm": 0.6171875, "learning_rate": 4.9814466913195717e-05, "loss": 2.0445, "step": 422 }, { "epoch": 0.1354683746997598, "grad_norm": 0.40625, "learning_rate": 4.981105133478614e-05, "loss": 2.3541, "step": 423 }, { "epoch": 0.1357886309047238, "grad_norm": 0.34375, "learning_rate": 4.9807604722116945e-05, "loss": 1.7954, "step": 424 }, { "epoch": 0.13610888710968774, "grad_norm": 0.640625, "learning_rate": 4.980412707949923e-05, "loss": 2.3201, "step": 425 }, { "epoch": 0.13642914331465172, "grad_norm": 0.67578125, "learning_rate": 4.9800618411282914e-05, "loss": 2.3104, "step": 426 }, { "epoch": 0.1367493995196157, "grad_norm": 0.41796875, "learning_rate": 4.979707872185672e-05, "loss": 2.4099, "step": 427 }, { "epoch": 0.13706965572457966, "grad_norm": 0.3671875, "learning_rate": 4.979350801564818e-05, "loss": 2.0, "step": 428 }, { "epoch": 0.13738991192954364, "grad_norm": 0.466796875, "learning_rate": 4.9789906297123626e-05, "loss": 2.3145, "step": 429 }, { "epoch": 0.1377101681345076, "grad_norm": 0.451171875, "learning_rate": 4.978627357078817e-05, "loss": 2.1489, "step": 430 }, { "epoch": 0.13803042433947157, "grad_norm": 0.474609375, "learning_rate": 4.97826098411857e-05, "loss": 2.0548, "step": 431 }, { "epoch": 0.13835068054443556, "grad_norm": 0.7890625, "learning_rate": 4.9778915112898914e-05, "loss": 2.3199, "step": 432 }, { "epoch": 0.1386709367493995, "grad_norm": 12.4375, "learning_rate": 4.977518939054927e-05, "loss": 2.1056, "step": 433 }, { "epoch": 0.1389911929543635, "grad_norm": 0.443359375, "learning_rate": 4.977143267879697e-05, "loss": 2.2003, "step": 434 }, { "epoch": 0.13931144915932747, "grad_norm": 0.51953125, "learning_rate": 4.9767644982341014e-05, "loss": 2.1945, "step": 435 }, { "epoch": 0.13963170536429143, "grad_norm": 0.47265625, "learning_rate": 4.9763826305919146e-05, "loss": 2.4707, "step": 436 }, { "epoch": 0.1399519615692554, "grad_norm": 0.455078125, "learning_rate": 4.975997665430785e-05, "loss": 2.5856, "step": 437 }, { "epoch": 0.1402722177742194, "grad_norm": 0.4921875, "learning_rate": 4.9756096032322376e-05, "loss": 2.2698, "step": 438 }, { "epoch": 0.14059247397918334, "grad_norm": 0.34765625, "learning_rate": 4.975218444481668e-05, "loss": 2.5266, "step": 439 }, { "epoch": 0.14091273018414732, "grad_norm": 0.5859375, "learning_rate": 4.974824189668348e-05, "loss": 2.2826, "step": 440 }, { "epoch": 0.14123298638911128, "grad_norm": 1.28125, "learning_rate": 4.974426839285422e-05, "loss": 1.8341, "step": 441 }, { "epoch": 0.14155324259407526, "grad_norm": 0.439453125, "learning_rate": 4.9740263938299034e-05, "loss": 2.6753, "step": 442 }, { "epoch": 0.14187349879903924, "grad_norm": 0.458984375, "learning_rate": 4.973622853802681e-05, "loss": 2.436, "step": 443 }, { "epoch": 0.1421937550040032, "grad_norm": 0.439453125, "learning_rate": 4.97321621970851e-05, "loss": 2.1716, "step": 444 }, { "epoch": 0.14251401120896717, "grad_norm": 0.62109375, "learning_rate": 4.97280649205602e-05, "loss": 2.0777, "step": 445 }, { "epoch": 0.14283426741393115, "grad_norm": 0.41015625, "learning_rate": 4.9723936713577084e-05, "loss": 1.7713, "step": 446 }, { "epoch": 0.1431545236188951, "grad_norm": 0.52734375, "learning_rate": 4.97197775812994e-05, "loss": 2.645, "step": 447 }, { "epoch": 0.1434747798238591, "grad_norm": 0.53515625, "learning_rate": 4.97155875289295e-05, "loss": 2.1758, "step": 448 }, { "epoch": 0.14379503602882307, "grad_norm": 0.447265625, "learning_rate": 4.9711366561708395e-05, "loss": 2.2733, "step": 449 }, { "epoch": 0.14411529223378702, "grad_norm": 0.42578125, "learning_rate": 4.9707114684915776e-05, "loss": 1.9709, "step": 450 }, { "epoch": 0.144435548438751, "grad_norm": 0.279296875, "learning_rate": 4.9702831903869996e-05, "loss": 1.9222, "step": 451 }, { "epoch": 0.14475580464371496, "grad_norm": 0.55078125, "learning_rate": 4.969851822392805e-05, "loss": 2.0273, "step": 452 }, { "epoch": 0.14507606084867894, "grad_norm": 0.470703125, "learning_rate": 4.9694173650485595e-05, "loss": 2.286, "step": 453 }, { "epoch": 0.14539631705364292, "grad_norm": 0.640625, "learning_rate": 4.968979818897694e-05, "loss": 2.6001, "step": 454 }, { "epoch": 0.14571657325860687, "grad_norm": 0.486328125, "learning_rate": 4.968539184487502e-05, "loss": 2.4443, "step": 455 }, { "epoch": 0.14603682946357086, "grad_norm": 0.62890625, "learning_rate": 4.9680954623691374e-05, "loss": 2.2168, "step": 456 }, { "epoch": 0.14635708566853484, "grad_norm": 0.5, "learning_rate": 4.9676486530976196e-05, "loss": 2.4374, "step": 457 }, { "epoch": 0.1466773418734988, "grad_norm": 0.46875, "learning_rate": 4.967198757231829e-05, "loss": 2.1691, "step": 458 }, { "epoch": 0.14699759807846277, "grad_norm": 0.83203125, "learning_rate": 4.966745775334505e-05, "loss": 2.7222, "step": 459 }, { "epoch": 0.14731785428342675, "grad_norm": 0.53125, "learning_rate": 4.966289707972249e-05, "loss": 1.9953, "step": 460 }, { "epoch": 0.1476381104883907, "grad_norm": 0.255859375, "learning_rate": 4.965830555715522e-05, "loss": 2.0458, "step": 461 }, { "epoch": 0.1479583666933547, "grad_norm": 0.56640625, "learning_rate": 4.9653683191386405e-05, "loss": 2.2034, "step": 462 }, { "epoch": 0.14827862289831867, "grad_norm": 0.470703125, "learning_rate": 4.964902998819782e-05, "loss": 2.0623, "step": 463 }, { "epoch": 0.14859887910328262, "grad_norm": 0.490234375, "learning_rate": 4.964434595340982e-05, "loss": 2.0543, "step": 464 }, { "epoch": 0.1489191353082466, "grad_norm": 0.6171875, "learning_rate": 4.963963109288129e-05, "loss": 1.7578, "step": 465 }, { "epoch": 0.14923939151321056, "grad_norm": 0.3515625, "learning_rate": 4.963488541250969e-05, "loss": 1.9211, "step": 466 }, { "epoch": 0.14955964771817454, "grad_norm": 0.423828125, "learning_rate": 4.963010891823103e-05, "loss": 2.1433, "step": 467 }, { "epoch": 0.14987990392313852, "grad_norm": 0.86328125, "learning_rate": 4.962530161601988e-05, "loss": 2.7066, "step": 468 }, { "epoch": 0.15020016012810247, "grad_norm": 0.40234375, "learning_rate": 4.9620463511889314e-05, "loss": 1.7484, "step": 469 }, { "epoch": 0.15052041633306645, "grad_norm": 0.380859375, "learning_rate": 4.961559461189096e-05, "loss": 2.2189, "step": 470 }, { "epoch": 0.15084067253803043, "grad_norm": 0.33203125, "learning_rate": 4.961069492211494e-05, "loss": 1.9058, "step": 471 }, { "epoch": 0.1511609287429944, "grad_norm": 0.490234375, "learning_rate": 4.960576444868992e-05, "loss": 2.1691, "step": 472 }, { "epoch": 0.15148118494795837, "grad_norm": 0.35546875, "learning_rate": 4.960080319778305e-05, "loss": 2.2587, "step": 473 }, { "epoch": 0.15180144115292235, "grad_norm": 0.57421875, "learning_rate": 4.959581117559998e-05, "loss": 1.6996, "step": 474 }, { "epoch": 0.1521216973578863, "grad_norm": 0.419921875, "learning_rate": 4.959078838838485e-05, "loss": 2.3875, "step": 475 }, { "epoch": 0.15244195356285029, "grad_norm": 0.515625, "learning_rate": 4.958573484242027e-05, "loss": 2.4589, "step": 476 }, { "epoch": 0.15276220976781424, "grad_norm": 2.578125, "learning_rate": 4.9580650544027376e-05, "loss": 2.1532, "step": 477 }, { "epoch": 0.15308246597277822, "grad_norm": 0.478515625, "learning_rate": 4.957553549956568e-05, "loss": 2.2743, "step": 478 }, { "epoch": 0.1534027221777422, "grad_norm": 0.435546875, "learning_rate": 4.957038971543325e-05, "loss": 1.9105, "step": 479 }, { "epoch": 0.15372297838270615, "grad_norm": 0.376953125, "learning_rate": 4.956521319806653e-05, "loss": 2.2667, "step": 480 }, { "epoch": 0.15404323458767014, "grad_norm": 0.5625, "learning_rate": 4.956000595394043e-05, "loss": 1.9691, "step": 481 }, { "epoch": 0.15436349079263412, "grad_norm": 1.03125, "learning_rate": 4.955476798956831e-05, "loss": 1.549, "step": 482 }, { "epoch": 0.15468374699759807, "grad_norm": 0.50390625, "learning_rate": 4.954949931150193e-05, "loss": 1.8682, "step": 483 }, { "epoch": 0.15500400320256205, "grad_norm": 0.51171875, "learning_rate": 4.9544199926331484e-05, "loss": 1.846, "step": 484 }, { "epoch": 0.15532425940752603, "grad_norm": 0.66796875, "learning_rate": 4.953886984068558e-05, "loss": 2.1578, "step": 485 }, { "epoch": 0.15564451561249, "grad_norm": 0.546875, "learning_rate": 4.95335090612312e-05, "loss": 2.2433, "step": 486 }, { "epoch": 0.15596477181745397, "grad_norm": 0.58984375, "learning_rate": 4.952811759467374e-05, "loss": 2.2284, "step": 487 }, { "epoch": 0.15628502802241792, "grad_norm": 0.478515625, "learning_rate": 4.9522695447757e-05, "loss": 2.5033, "step": 488 }, { "epoch": 0.1566052842273819, "grad_norm": 0.39453125, "learning_rate": 4.9517242627263096e-05, "loss": 1.7995, "step": 489 }, { "epoch": 0.15692554043234588, "grad_norm": 0.4375, "learning_rate": 4.951175914001256e-05, "loss": 2.5606, "step": 490 }, { "epoch": 0.15724579663730984, "grad_norm": 1.7265625, "learning_rate": 4.950624499286428e-05, "loss": 2.3475, "step": 491 }, { "epoch": 0.15756605284227382, "grad_norm": 0.474609375, "learning_rate": 4.950070019271547e-05, "loss": 2.3772, "step": 492 }, { "epoch": 0.1578863090472378, "grad_norm": 0.40234375, "learning_rate": 4.949512474650171e-05, "loss": 1.9745, "step": 493 }, { "epoch": 0.15820656525220175, "grad_norm": 0.58984375, "learning_rate": 4.94895186611969e-05, "loss": 2.5581, "step": 494 }, { "epoch": 0.15852682145716573, "grad_norm": 0.82421875, "learning_rate": 4.948388194381326e-05, "loss": 2.3336, "step": 495 }, { "epoch": 0.15884707766212972, "grad_norm": 0.443359375, "learning_rate": 4.947821460140134e-05, "loss": 2.1051, "step": 496 }, { "epoch": 0.15916733386709367, "grad_norm": 0.55078125, "learning_rate": 4.947251664104998e-05, "loss": 2.4866, "step": 497 }, { "epoch": 0.15948759007205765, "grad_norm": 0.419921875, "learning_rate": 4.946678806988633e-05, "loss": 2.2679, "step": 498 }, { "epoch": 0.1598078462770216, "grad_norm": 0.419921875, "learning_rate": 4.9461028895075825e-05, "loss": 2.4515, "step": 499 }, { "epoch": 0.16012810248198558, "grad_norm": 0.466796875, "learning_rate": 4.9455239123822184e-05, "loss": 1.6854, "step": 500 }, { "epoch": 0.16044835868694957, "grad_norm": 0.97265625, "learning_rate": 4.944941876336738e-05, "loss": 2.0233, "step": 501 }, { "epoch": 0.16076861489191352, "grad_norm": 0.5390625, "learning_rate": 4.944356782099167e-05, "loss": 2.467, "step": 502 }, { "epoch": 0.1610888710968775, "grad_norm": 0.4921875, "learning_rate": 4.943768630401355e-05, "loss": 2.4588, "step": 503 }, { "epoch": 0.16140912730184148, "grad_norm": 0.5625, "learning_rate": 4.943177421978976e-05, "loss": 2.4571, "step": 504 }, { "epoch": 0.16172938350680544, "grad_norm": 0.408203125, "learning_rate": 4.9425831575715286e-05, "loss": 2.6707, "step": 505 }, { "epoch": 0.16204963971176942, "grad_norm": 0.87890625, "learning_rate": 4.941985837922332e-05, "loss": 2.1903, "step": 506 }, { "epoch": 0.1623698959167334, "grad_norm": 0.302734375, "learning_rate": 4.941385463778528e-05, "loss": 1.7641, "step": 507 }, { "epoch": 0.16269015212169735, "grad_norm": 0.89453125, "learning_rate": 4.9407820358910804e-05, "loss": 2.4916, "step": 508 }, { "epoch": 0.16301040832666133, "grad_norm": 1.59375, "learning_rate": 4.94017555501477e-05, "loss": 2.0425, "step": 509 }, { "epoch": 0.1633306645316253, "grad_norm": 0.48046875, "learning_rate": 4.939566021908197e-05, "loss": 2.0939, "step": 510 }, { "epoch": 0.16365092073658927, "grad_norm": 0.609375, "learning_rate": 4.938953437333783e-05, "loss": 2.3065, "step": 511 }, { "epoch": 0.16397117694155325, "grad_norm": 0.443359375, "learning_rate": 4.93833780205776e-05, "loss": 2.7856, "step": 512 }, { "epoch": 0.1642914331465172, "grad_norm": 0.51953125, "learning_rate": 4.937719116850181e-05, "loss": 2.3894, "step": 513 }, { "epoch": 0.16461168935148118, "grad_norm": 0.419921875, "learning_rate": 4.937097382484913e-05, "loss": 2.5719, "step": 514 }, { "epoch": 0.16493194555644516, "grad_norm": 0.392578125, "learning_rate": 4.936472599739635e-05, "loss": 1.8058, "step": 515 }, { "epoch": 0.16525220176140912, "grad_norm": 0.43359375, "learning_rate": 4.935844769395842e-05, "loss": 2.2764, "step": 516 }, { "epoch": 0.1655724579663731, "grad_norm": 0.46875, "learning_rate": 4.9352138922388375e-05, "loss": 1.6205, "step": 517 }, { "epoch": 0.16589271417133708, "grad_norm": 0.59765625, "learning_rate": 4.9345799690577385e-05, "loss": 1.8822, "step": 518 }, { "epoch": 0.16621297037630103, "grad_norm": 0.6953125, "learning_rate": 4.933943000645471e-05, "loss": 2.3701, "step": 519 }, { "epoch": 0.16653322658126501, "grad_norm": 1.2265625, "learning_rate": 4.9333029877987715e-05, "loss": 2.474, "step": 520 }, { "epoch": 0.166853482786229, "grad_norm": 0.6015625, "learning_rate": 4.932659931318182e-05, "loss": 2.0294, "step": 521 }, { "epoch": 0.16717373899119295, "grad_norm": 0.46484375, "learning_rate": 4.932013832008054e-05, "loss": 2.218, "step": 522 }, { "epoch": 0.16749399519615693, "grad_norm": 0.828125, "learning_rate": 4.931364690676544e-05, "loss": 2.5899, "step": 523 }, { "epoch": 0.16781425140112088, "grad_norm": 0.55859375, "learning_rate": 4.930712508135613e-05, "loss": 1.828, "step": 524 }, { "epoch": 0.16813450760608487, "grad_norm": 5.3125, "learning_rate": 4.930057285201027e-05, "loss": 2.627, "step": 525 }, { "epoch": 0.16845476381104885, "grad_norm": 1.25, "learning_rate": 4.929399022692355e-05, "loss": 2.1664, "step": 526 }, { "epoch": 0.1687750200160128, "grad_norm": 0.43359375, "learning_rate": 4.928737721432967e-05, "loss": 1.8216, "step": 527 }, { "epoch": 0.16909527622097678, "grad_norm": 0.52734375, "learning_rate": 4.9280733822500346e-05, "loss": 1.9037, "step": 528 }, { "epoch": 0.16941553242594076, "grad_norm": 0.48046875, "learning_rate": 4.927406005974529e-05, "loss": 2.0505, "step": 529 }, { "epoch": 0.16973578863090472, "grad_norm": 0.64453125, "learning_rate": 4.9267355934412214e-05, "loss": 2.3799, "step": 530 }, { "epoch": 0.1700560448358687, "grad_norm": 0.52734375, "learning_rate": 4.926062145488679e-05, "loss": 2.2806, "step": 531 }, { "epoch": 0.17037630104083268, "grad_norm": 0.314453125, "learning_rate": 4.9253856629592663e-05, "loss": 2.4759, "step": 532 }, { "epoch": 0.17069655724579663, "grad_norm": 0.486328125, "learning_rate": 4.924706146699146e-05, "loss": 2.6853, "step": 533 }, { "epoch": 0.1710168134507606, "grad_norm": 0.5546875, "learning_rate": 4.924023597558271e-05, "loss": 1.9226, "step": 534 }, { "epoch": 0.17133706965572457, "grad_norm": 0.8515625, "learning_rate": 4.923338016390392e-05, "loss": 1.8459, "step": 535 }, { "epoch": 0.17165732586068855, "grad_norm": 0.408203125, "learning_rate": 4.922649404053048e-05, "loss": 2.0748, "step": 536 }, { "epoch": 0.17197758206565253, "grad_norm": 0.72265625, "learning_rate": 4.921957761407575e-05, "loss": 2.0169, "step": 537 }, { "epoch": 0.17229783827061648, "grad_norm": 0.75390625, "learning_rate": 4.921263089319095e-05, "loss": 1.7651, "step": 538 }, { "epoch": 0.17261809447558046, "grad_norm": 0.53125, "learning_rate": 4.920565388656519e-05, "loss": 2.498, "step": 539 }, { "epoch": 0.17293835068054444, "grad_norm": 0.52734375, "learning_rate": 4.919864660292549e-05, "loss": 2.1956, "step": 540 }, { "epoch": 0.1732586068855084, "grad_norm": 0.671875, "learning_rate": 4.919160905103673e-05, "loss": 2.4898, "step": 541 }, { "epoch": 0.17357886309047238, "grad_norm": 0.494140625, "learning_rate": 4.9184541239701666e-05, "loss": 2.1794, "step": 542 }, { "epoch": 0.17389911929543636, "grad_norm": 0.5, "learning_rate": 4.9177443177760854e-05, "loss": 2.1124, "step": 543 }, { "epoch": 0.17421937550040031, "grad_norm": 1.0078125, "learning_rate": 4.9170314874092734e-05, "loss": 2.161, "step": 544 }, { "epoch": 0.1745396317053643, "grad_norm": 0.66796875, "learning_rate": 4.916315633761356e-05, "loss": 2.2314, "step": 545 }, { "epoch": 0.17485988791032825, "grad_norm": 0.400390625, "learning_rate": 4.915596757727741e-05, "loss": 1.9608, "step": 546 }, { "epoch": 0.17518014411529223, "grad_norm": 0.46484375, "learning_rate": 4.9148748602076146e-05, "loss": 1.8623, "step": 547 }, { "epoch": 0.1755004003202562, "grad_norm": 0.5078125, "learning_rate": 4.914149942103944e-05, "loss": 2.0954, "step": 548 }, { "epoch": 0.17582065652522016, "grad_norm": 0.36328125, "learning_rate": 4.913422004323474e-05, "loss": 2.0595, "step": 549 }, { "epoch": 0.17614091273018415, "grad_norm": 0.369140625, "learning_rate": 4.912691047776727e-05, "loss": 2.1955, "step": 550 }, { "epoch": 0.17646116893514813, "grad_norm": 0.470703125, "learning_rate": 4.9119570733780015e-05, "loss": 2.0215, "step": 551 }, { "epoch": 0.17678142514011208, "grad_norm": 0.498046875, "learning_rate": 4.91122008204537e-05, "loss": 2.29, "step": 552 }, { "epoch": 0.17710168134507606, "grad_norm": 0.494140625, "learning_rate": 4.910480074700678e-05, "loss": 2.1564, "step": 553 }, { "epoch": 0.17742193755004004, "grad_norm": 0.47265625, "learning_rate": 4.909737052269545e-05, "loss": 2.0404, "step": 554 }, { "epoch": 0.177742193755004, "grad_norm": 0.4296875, "learning_rate": 4.908991015681362e-05, "loss": 2.3606, "step": 555 }, { "epoch": 0.17806244995996798, "grad_norm": 0.6328125, "learning_rate": 4.908241965869289e-05, "loss": 2.4787, "step": 556 }, { "epoch": 0.17838270616493196, "grad_norm": 0.48828125, "learning_rate": 4.907489903770256e-05, "loss": 2.3174, "step": 557 }, { "epoch": 0.1787029623698959, "grad_norm": 0.484375, "learning_rate": 4.9067348303249594e-05, "loss": 2.5595, "step": 558 }, { "epoch": 0.1790232185748599, "grad_norm": 0.5625, "learning_rate": 4.905976746477864e-05, "loss": 2.6765, "step": 559 }, { "epoch": 0.17934347477982385, "grad_norm": 0.490234375, "learning_rate": 4.905215653177199e-05, "loss": 2.6524, "step": 560 }, { "epoch": 0.17966373098478783, "grad_norm": 0.5, "learning_rate": 4.904451551374959e-05, "loss": 2.497, "step": 561 }, { "epoch": 0.1799839871897518, "grad_norm": 0.52734375, "learning_rate": 4.903684442026899e-05, "loss": 2.0149, "step": 562 }, { "epoch": 0.18030424339471576, "grad_norm": 0.486328125, "learning_rate": 4.9029143260925395e-05, "loss": 2.1776, "step": 563 }, { "epoch": 0.18062449959967974, "grad_norm": 0.470703125, "learning_rate": 4.902141204535159e-05, "loss": 1.8415, "step": 564 }, { "epoch": 0.18094475580464373, "grad_norm": 0.53515625, "learning_rate": 4.901365078321798e-05, "loss": 2.2762, "step": 565 }, { "epoch": 0.18126501200960768, "grad_norm": 0.40234375, "learning_rate": 4.9005859484232526e-05, "loss": 2.6546, "step": 566 }, { "epoch": 0.18158526821457166, "grad_norm": 0.45703125, "learning_rate": 4.899803815814077e-05, "loss": 2.1712, "step": 567 }, { "epoch": 0.18190552441953564, "grad_norm": 0.400390625, "learning_rate": 4.899018681472582e-05, "loss": 2.5383, "step": 568 }, { "epoch": 0.1822257806244996, "grad_norm": 0.76953125, "learning_rate": 4.8982305463808334e-05, "loss": 2.3668, "step": 569 }, { "epoch": 0.18254603682946358, "grad_norm": 0.431640625, "learning_rate": 4.897439411524648e-05, "loss": 2.4888, "step": 570 }, { "epoch": 0.18286629303442753, "grad_norm": 0.376953125, "learning_rate": 4.896645277893597e-05, "loss": 1.5993, "step": 571 }, { "epoch": 0.1831865492393915, "grad_norm": 0.50390625, "learning_rate": 4.895848146481002e-05, "loss": 2.1428, "step": 572 }, { "epoch": 0.1835068054443555, "grad_norm": 0.921875, "learning_rate": 4.895048018283933e-05, "loss": 2.6575, "step": 573 }, { "epoch": 0.18382706164931945, "grad_norm": 1.9296875, "learning_rate": 4.8942448943032116e-05, "loss": 2.0452, "step": 574 }, { "epoch": 0.18414731785428343, "grad_norm": 0.57421875, "learning_rate": 4.893438775543403e-05, "loss": 1.9505, "step": 575 }, { "epoch": 0.1844675740592474, "grad_norm": 0.439453125, "learning_rate": 4.89262966301282e-05, "loss": 1.9657, "step": 576 }, { "epoch": 0.18478783026421136, "grad_norm": 0.46875, "learning_rate": 4.8918175577235214e-05, "loss": 2.5377, "step": 577 }, { "epoch": 0.18510808646917534, "grad_norm": 0.40625, "learning_rate": 4.891002460691306e-05, "loss": 2.1608, "step": 578 }, { "epoch": 0.18542834267413932, "grad_norm": 0.421875, "learning_rate": 4.890184372935719e-05, "loss": 2.0481, "step": 579 }, { "epoch": 0.18574859887910328, "grad_norm": 0.404296875, "learning_rate": 4.889363295480042e-05, "loss": 1.9898, "step": 580 }, { "epoch": 0.18606885508406726, "grad_norm": 0.5859375, "learning_rate": 4.8885392293513e-05, "loss": 2.3255, "step": 581 }, { "epoch": 0.1863891112890312, "grad_norm": 0.6171875, "learning_rate": 4.8877121755802536e-05, "loss": 2.2675, "step": 582 }, { "epoch": 0.1867093674939952, "grad_norm": 0.60546875, "learning_rate": 4.8868821352014024e-05, "loss": 2.2714, "step": 583 }, { "epoch": 0.18702962369895917, "grad_norm": 0.5703125, "learning_rate": 4.886049109252981e-05, "loss": 2.0052, "step": 584 }, { "epoch": 0.18734987990392313, "grad_norm": 0.8125, "learning_rate": 4.8852130987769574e-05, "loss": 2.4493, "step": 585 }, { "epoch": 0.1876701361088871, "grad_norm": 0.6953125, "learning_rate": 4.884374104819035e-05, "loss": 2.2796, "step": 586 }, { "epoch": 0.1879903923138511, "grad_norm": 0.265625, "learning_rate": 4.883532128428645e-05, "loss": 1.4809, "step": 587 }, { "epoch": 0.18831064851881504, "grad_norm": 0.359375, "learning_rate": 4.882687170658955e-05, "loss": 1.5241, "step": 588 }, { "epoch": 0.18863090472377902, "grad_norm": 0.337890625, "learning_rate": 4.881839232566856e-05, "loss": 2.1891, "step": 589 }, { "epoch": 0.188951160928743, "grad_norm": 0.640625, "learning_rate": 4.880988315212971e-05, "loss": 2.2043, "step": 590 }, { "epoch": 0.18927141713370696, "grad_norm": 0.7265625, "learning_rate": 4.880134419661647e-05, "loss": 2.3487, "step": 591 }, { "epoch": 0.18959167333867094, "grad_norm": 0.498046875, "learning_rate": 4.879277546980957e-05, "loss": 2.6678, "step": 592 }, { "epoch": 0.1899119295436349, "grad_norm": 0.4375, "learning_rate": 4.8784176982426984e-05, "loss": 2.1424, "step": 593 }, { "epoch": 0.19023218574859888, "grad_norm": 0.455078125, "learning_rate": 4.87755487452239e-05, "loss": 2.3376, "step": 594 }, { "epoch": 0.19055244195356286, "grad_norm": 0.46875, "learning_rate": 4.8766890768992734e-05, "loss": 2.4265, "step": 595 }, { "epoch": 0.1908726981585268, "grad_norm": 0.65234375, "learning_rate": 4.875820306456309e-05, "loss": 2.366, "step": 596 }, { "epoch": 0.1911929543634908, "grad_norm": 0.4765625, "learning_rate": 4.8749485642801754e-05, "loss": 2.4942, "step": 597 }, { "epoch": 0.19151321056845477, "grad_norm": 0.431640625, "learning_rate": 4.874073851461268e-05, "loss": 1.9429, "step": 598 }, { "epoch": 0.19183346677341873, "grad_norm": 0.5625, "learning_rate": 4.873196169093699e-05, "loss": 2.1505, "step": 599 }, { "epoch": 0.1921537229783827, "grad_norm": 0.8046875, "learning_rate": 4.872315518275296e-05, "loss": 2.373, "step": 600 }, { "epoch": 0.1924739791833467, "grad_norm": 0.5703125, "learning_rate": 4.871431900107597e-05, "loss": 2.2588, "step": 601 }, { "epoch": 0.19279423538831064, "grad_norm": 0.71484375, "learning_rate": 4.870545315695853e-05, "loss": 2.6125, "step": 602 }, { "epoch": 0.19311449159327462, "grad_norm": 0.515625, "learning_rate": 4.869655766149024e-05, "loss": 2.489, "step": 603 }, { "epoch": 0.1934347477982386, "grad_norm": 0.39453125, "learning_rate": 4.868763252579782e-05, "loss": 2.29, "step": 604 }, { "epoch": 0.19375500400320256, "grad_norm": 0.51953125, "learning_rate": 4.867867776104503e-05, "loss": 2.3659, "step": 605 }, { "epoch": 0.19407526020816654, "grad_norm": 0.427734375, "learning_rate": 4.866969337843271e-05, "loss": 1.8565, "step": 606 }, { "epoch": 0.1943955164131305, "grad_norm": 0.439453125, "learning_rate": 4.8660679389198746e-05, "loss": 2.5902, "step": 607 }, { "epoch": 0.19471577261809447, "grad_norm": 0.365234375, "learning_rate": 4.8651635804618046e-05, "loss": 2.3034, "step": 608 }, { "epoch": 0.19503602882305845, "grad_norm": 0.35546875, "learning_rate": 4.8642562636002546e-05, "loss": 2.4728, "step": 609 }, { "epoch": 0.1953562850280224, "grad_norm": 0.455078125, "learning_rate": 4.8633459894701186e-05, "loss": 2.1875, "step": 610 }, { "epoch": 0.1956765412329864, "grad_norm": 0.466796875, "learning_rate": 4.8624327592099884e-05, "loss": 2.2461, "step": 611 }, { "epoch": 0.19599679743795037, "grad_norm": 0.390625, "learning_rate": 4.861516573962156e-05, "loss": 1.8182, "step": 612 }, { "epoch": 0.19631705364291432, "grad_norm": 0.99609375, "learning_rate": 4.860597434872606e-05, "loss": 1.7308, "step": 613 }, { "epoch": 0.1966373098478783, "grad_norm": 0.412109375, "learning_rate": 4.859675343091021e-05, "loss": 2.5587, "step": 614 }, { "epoch": 0.1969575660528423, "grad_norm": 3.15625, "learning_rate": 4.858750299770774e-05, "loss": 2.2981, "step": 615 }, { "epoch": 0.19727782225780624, "grad_norm": 0.984375, "learning_rate": 4.8578223060689346e-05, "loss": 2.356, "step": 616 }, { "epoch": 0.19759807846277022, "grad_norm": 0.47265625, "learning_rate": 4.8568913631462565e-05, "loss": 1.9542, "step": 617 }, { "epoch": 0.19791833466773417, "grad_norm": 0.369140625, "learning_rate": 4.855957472167187e-05, "loss": 2.3769, "step": 618 }, { "epoch": 0.19823859087269816, "grad_norm": 0.56640625, "learning_rate": 4.85502063429986e-05, "loss": 2.4044, "step": 619 }, { "epoch": 0.19855884707766214, "grad_norm": 0.53125, "learning_rate": 4.8540808507160926e-05, "loss": 2.4242, "step": 620 }, { "epoch": 0.1988791032826261, "grad_norm": 0.65625, "learning_rate": 4.85313812259139e-05, "loss": 2.3783, "step": 621 }, { "epoch": 0.19919935948759007, "grad_norm": 0.5234375, "learning_rate": 4.85219245110494e-05, "loss": 2.5425, "step": 622 }, { "epoch": 0.19951961569255405, "grad_norm": 0.373046875, "learning_rate": 4.8512438374396095e-05, "loss": 2.3095, "step": 623 }, { "epoch": 0.199839871897518, "grad_norm": 0.87109375, "learning_rate": 4.850292282781949e-05, "loss": 2.5806, "step": 624 }, { "epoch": 0.200160128102482, "grad_norm": 0.609375, "learning_rate": 4.849337788322185e-05, "loss": 2.4287, "step": 625 }, { "epoch": 0.20048038430744597, "grad_norm": 0.55078125, "learning_rate": 4.848380355254222e-05, "loss": 2.6023, "step": 626 }, { "epoch": 0.20080064051240992, "grad_norm": 0.625, "learning_rate": 4.847419984775641e-05, "loss": 2.0865, "step": 627 }, { "epoch": 0.2011208967173739, "grad_norm": 1.6484375, "learning_rate": 4.8464566780876954e-05, "loss": 1.8258, "step": 628 }, { "epoch": 0.20144115292233786, "grad_norm": 0.515625, "learning_rate": 4.845490436395314e-05, "loss": 2.2667, "step": 629 }, { "epoch": 0.20176140912730184, "grad_norm": 0.5234375, "learning_rate": 4.8445212609070934e-05, "loss": 2.394, "step": 630 }, { "epoch": 0.20208166533226582, "grad_norm": 0.58984375, "learning_rate": 4.8435491528353026e-05, "loss": 2.6123, "step": 631 }, { "epoch": 0.20240192153722977, "grad_norm": 1.25, "learning_rate": 4.8425741133958774e-05, "loss": 2.2056, "step": 632 }, { "epoch": 0.20272217774219375, "grad_norm": 0.4453125, "learning_rate": 4.841596143808421e-05, "loss": 2.6455, "step": 633 }, { "epoch": 0.20304243394715774, "grad_norm": 0.37890625, "learning_rate": 4.840615245296201e-05, "loss": 2.1008, "step": 634 }, { "epoch": 0.2033626901521217, "grad_norm": 0.453125, "learning_rate": 4.8396314190861495e-05, "loss": 1.9474, "step": 635 }, { "epoch": 0.20368294635708567, "grad_norm": 0.578125, "learning_rate": 4.838644666408858e-05, "loss": 2.1314, "step": 636 }, { "epoch": 0.20400320256204965, "grad_norm": 0.4375, "learning_rate": 4.8376549884985825e-05, "loss": 2.4542, "step": 637 }, { "epoch": 0.2043234587670136, "grad_norm": 0.40234375, "learning_rate": 4.836662386593235e-05, "loss": 2.2233, "step": 638 }, { "epoch": 0.20464371497197759, "grad_norm": 0.4296875, "learning_rate": 4.835666861934387e-05, "loss": 1.8406, "step": 639 }, { "epoch": 0.20496397117694154, "grad_norm": 0.458984375, "learning_rate": 4.834668415767263e-05, "loss": 2.3927, "step": 640 }, { "epoch": 0.20528422738190552, "grad_norm": 0.458984375, "learning_rate": 4.833667049340744e-05, "loss": 2.4086, "step": 641 }, { "epoch": 0.2056044835868695, "grad_norm": 0.53125, "learning_rate": 4.832662763907363e-05, "loss": 1.7985, "step": 642 }, { "epoch": 0.20592473979183346, "grad_norm": 0.423828125, "learning_rate": 4.831655560723306e-05, "loss": 2.2216, "step": 643 }, { "epoch": 0.20624499599679744, "grad_norm": 1.3515625, "learning_rate": 4.8306454410484055e-05, "loss": 2.4505, "step": 644 }, { "epoch": 0.20656525220176142, "grad_norm": 0.60546875, "learning_rate": 4.829632406146143e-05, "loss": 2.1029, "step": 645 }, { "epoch": 0.20688550840672537, "grad_norm": 0.5, "learning_rate": 4.828616457283648e-05, "loss": 2.4966, "step": 646 }, { "epoch": 0.20720576461168935, "grad_norm": 0.384765625, "learning_rate": 4.8275975957316943e-05, "loss": 1.7865, "step": 647 }, { "epoch": 0.20752602081665333, "grad_norm": 0.4140625, "learning_rate": 4.826575822764696e-05, "loss": 1.674, "step": 648 }, { "epoch": 0.2078462770216173, "grad_norm": 0.404296875, "learning_rate": 4.825551139660715e-05, "loss": 1.9932, "step": 649 }, { "epoch": 0.20816653322658127, "grad_norm": 0.84765625, "learning_rate": 4.824523547701445e-05, "loss": 2.18, "step": 650 }, { "epoch": 0.20848678943154525, "grad_norm": 1.7421875, "learning_rate": 4.8234930481722264e-05, "loss": 1.9729, "step": 651 }, { "epoch": 0.2088070456365092, "grad_norm": 0.5625, "learning_rate": 4.8224596423620315e-05, "loss": 2.0876, "step": 652 }, { "epoch": 0.20912730184147318, "grad_norm": 0.8359375, "learning_rate": 4.821423331563469e-05, "loss": 2.0135, "step": 653 }, { "epoch": 0.20944755804643714, "grad_norm": 0.416015625, "learning_rate": 4.820384117072781e-05, "loss": 2.3225, "step": 654 }, { "epoch": 0.20976781425140112, "grad_norm": 1.65625, "learning_rate": 4.819342000189844e-05, "loss": 1.6923, "step": 655 }, { "epoch": 0.2100880704563651, "grad_norm": 0.435546875, "learning_rate": 4.818296982218161e-05, "loss": 2.1741, "step": 656 }, { "epoch": 0.21040832666132905, "grad_norm": 0.671875, "learning_rate": 4.817249064464865e-05, "loss": 2.285, "step": 657 }, { "epoch": 0.21072858286629303, "grad_norm": 0.6484375, "learning_rate": 4.816198248240718e-05, "loss": 2.1315, "step": 658 }, { "epoch": 0.21104883907125702, "grad_norm": 0.498046875, "learning_rate": 4.815144534860106e-05, "loss": 1.9737, "step": 659 }, { "epoch": 0.21136909527622097, "grad_norm": 0.453125, "learning_rate": 4.81408792564104e-05, "loss": 1.6167, "step": 660 }, { "epoch": 0.21168935148118495, "grad_norm": 0.421875, "learning_rate": 4.8130284219051494e-05, "loss": 2.4866, "step": 661 }, { "epoch": 0.21200960768614893, "grad_norm": 0.7578125, "learning_rate": 4.81196602497769e-05, "loss": 2.2202, "step": 662 }, { "epoch": 0.21232986389111289, "grad_norm": 0.51953125, "learning_rate": 4.810900736187531e-05, "loss": 2.1335, "step": 663 }, { "epoch": 0.21265012009607687, "grad_norm": 0.58203125, "learning_rate": 4.809832556867164e-05, "loss": 2.0997, "step": 664 }, { "epoch": 0.21297037630104082, "grad_norm": 1.9765625, "learning_rate": 4.80876148835269e-05, "loss": 2.5337, "step": 665 }, { "epoch": 0.2132906325060048, "grad_norm": 0.5390625, "learning_rate": 4.80768753198383e-05, "loss": 2.4361, "step": 666 }, { "epoch": 0.21361088871096878, "grad_norm": 0.408203125, "learning_rate": 4.8066106891039135e-05, "loss": 2.7403, "step": 667 }, { "epoch": 0.21393114491593274, "grad_norm": 0.44921875, "learning_rate": 4.805530961059881e-05, "loss": 1.8824, "step": 668 }, { "epoch": 0.21425140112089672, "grad_norm": 0.51953125, "learning_rate": 4.804448349202283e-05, "loss": 1.9837, "step": 669 }, { "epoch": 0.2145716573258607, "grad_norm": 0.52734375, "learning_rate": 4.803362854885276e-05, "loss": 2.0329, "step": 670 }, { "epoch": 0.21489191353082465, "grad_norm": 0.91015625, "learning_rate": 4.802274479466623e-05, "loss": 2.314, "step": 671 }, { "epoch": 0.21521216973578863, "grad_norm": 0.44140625, "learning_rate": 4.8011832243076895e-05, "loss": 2.324, "step": 672 }, { "epoch": 0.21553242594075261, "grad_norm": 0.53515625, "learning_rate": 4.8000890907734434e-05, "loss": 1.7424, "step": 673 }, { "epoch": 0.21585268214571657, "grad_norm": 0.37109375, "learning_rate": 4.7989920802324537e-05, "loss": 1.981, "step": 674 }, { "epoch": 0.21617293835068055, "grad_norm": 0.412109375, "learning_rate": 4.797892194056889e-05, "loss": 1.6027, "step": 675 }, { "epoch": 0.2164931945556445, "grad_norm": 0.48828125, "learning_rate": 4.79678943362251e-05, "loss": 2.5835, "step": 676 }, { "epoch": 0.21681345076060848, "grad_norm": 0.7734375, "learning_rate": 4.79568380030868e-05, "loss": 2.4851, "step": 677 }, { "epoch": 0.21713370696557246, "grad_norm": 0.4296875, "learning_rate": 4.794575295498348e-05, "loss": 2.6476, "step": 678 }, { "epoch": 0.21745396317053642, "grad_norm": 0.427734375, "learning_rate": 4.793463920578061e-05, "loss": 2.6871, "step": 679 }, { "epoch": 0.2177742193755004, "grad_norm": 0.279296875, "learning_rate": 4.792349676937953e-05, "loss": 2.124, "step": 680 }, { "epoch": 0.21809447558046438, "grad_norm": 0.546875, "learning_rate": 4.791232565971745e-05, "loss": 2.3615, "step": 681 }, { "epoch": 0.21841473178542833, "grad_norm": 0.462890625, "learning_rate": 4.7901125890767484e-05, "loss": 1.8175, "step": 682 }, { "epoch": 0.21873498799039232, "grad_norm": 0.75390625, "learning_rate": 4.788989747653857e-05, "loss": 2.2353, "step": 683 }, { "epoch": 0.2190552441953563, "grad_norm": 0.52734375, "learning_rate": 4.787864043107546e-05, "loss": 1.9429, "step": 684 }, { "epoch": 0.21937550040032025, "grad_norm": 0.48828125, "learning_rate": 4.786735476845876e-05, "loss": 2.1238, "step": 685 }, { "epoch": 0.21969575660528423, "grad_norm": 0.361328125, "learning_rate": 4.7856040502804836e-05, "loss": 1.7298, "step": 686 }, { "epoch": 0.22001601281024818, "grad_norm": 0.51953125, "learning_rate": 4.784469764826585e-05, "loss": 1.9109, "step": 687 }, { "epoch": 0.22033626901521217, "grad_norm": 0.474609375, "learning_rate": 4.783332621902971e-05, "loss": 2.7153, "step": 688 }, { "epoch": 0.22065652522017615, "grad_norm": 0.6171875, "learning_rate": 4.782192622932008e-05, "loss": 2.3001, "step": 689 }, { "epoch": 0.2209767814251401, "grad_norm": 0.6640625, "learning_rate": 4.7810497693396327e-05, "loss": 2.0766, "step": 690 }, { "epoch": 0.22129703763010408, "grad_norm": 0.640625, "learning_rate": 4.779904062555356e-05, "loss": 2.2382, "step": 691 }, { "epoch": 0.22161729383506806, "grad_norm": 0.51171875, "learning_rate": 4.7787555040122534e-05, "loss": 2.6912, "step": 692 }, { "epoch": 0.22193755004003202, "grad_norm": 0.5703125, "learning_rate": 4.777604095146971e-05, "loss": 2.3497, "step": 693 }, { "epoch": 0.222257806244996, "grad_norm": 0.458984375, "learning_rate": 4.7764498373997194e-05, "loss": 2.4798, "step": 694 }, { "epoch": 0.22257806244995998, "grad_norm": 0.89453125, "learning_rate": 4.775292732214269e-05, "loss": 2.6314, "step": 695 }, { "epoch": 0.22289831865492393, "grad_norm": 0.51953125, "learning_rate": 4.774132781037958e-05, "loss": 2.2279, "step": 696 }, { "epoch": 0.2232185748598879, "grad_norm": 0.53515625, "learning_rate": 4.7729699853216784e-05, "loss": 2.6009, "step": 697 }, { "epoch": 0.2235388310648519, "grad_norm": 0.546875, "learning_rate": 4.771804346519886e-05, "loss": 2.1287, "step": 698 }, { "epoch": 0.22385908726981585, "grad_norm": 0.392578125, "learning_rate": 4.770635866090587e-05, "loss": 2.0747, "step": 699 }, { "epoch": 0.22417934347477983, "grad_norm": 1.2578125, "learning_rate": 4.769464545495347e-05, "loss": 2.1692, "step": 700 }, { "epoch": 0.22449959967974378, "grad_norm": 0.5625, "learning_rate": 4.768290386199281e-05, "loss": 2.1745, "step": 701 }, { "epoch": 0.22481985588470776, "grad_norm": 1.0234375, "learning_rate": 4.767113389671055e-05, "loss": 2.4571, "step": 702 }, { "epoch": 0.22514011208967175, "grad_norm": 0.486328125, "learning_rate": 4.7659335573828855e-05, "loss": 2.1443, "step": 703 }, { "epoch": 0.2254603682946357, "grad_norm": 0.365234375, "learning_rate": 4.7647508908105354e-05, "loss": 2.645, "step": 704 }, { "epoch": 0.22578062449959968, "grad_norm": 0.50390625, "learning_rate": 4.7635653914333123e-05, "loss": 2.5967, "step": 705 }, { "epoch": 0.22610088070456366, "grad_norm": 0.3828125, "learning_rate": 4.7623770607340656e-05, "loss": 1.8238, "step": 706 }, { "epoch": 0.22642113690952761, "grad_norm": 0.4765625, "learning_rate": 4.761185900199191e-05, "loss": 2.1985, "step": 707 }, { "epoch": 0.2267413931144916, "grad_norm": 1.15625, "learning_rate": 4.7599919113186184e-05, "loss": 1.9056, "step": 708 }, { "epoch": 0.22706164931945558, "grad_norm": 0.53515625, "learning_rate": 4.7587950955858184e-05, "loss": 2.0725, "step": 709 }, { "epoch": 0.22738190552441953, "grad_norm": 0.62109375, "learning_rate": 4.757595454497798e-05, "loss": 2.1442, "step": 710 }, { "epoch": 0.2277021617293835, "grad_norm": 0.34375, "learning_rate": 4.756392989555094e-05, "loss": 1.8137, "step": 711 }, { "epoch": 0.22802241793434747, "grad_norm": 0.515625, "learning_rate": 4.755187702261782e-05, "loss": 2.4148, "step": 712 }, { "epoch": 0.22834267413931145, "grad_norm": 0.54296875, "learning_rate": 4.753979594125463e-05, "loss": 2.579, "step": 713 }, { "epoch": 0.22866293034427543, "grad_norm": 0.55078125, "learning_rate": 4.752768666657267e-05, "loss": 2.4936, "step": 714 }, { "epoch": 0.22898318654923938, "grad_norm": 0.498046875, "learning_rate": 4.751554921371852e-05, "loss": 2.104, "step": 715 }, { "epoch": 0.22930344275420336, "grad_norm": 0.3984375, "learning_rate": 4.7503383597873994e-05, "loss": 2.0582, "step": 716 }, { "epoch": 0.22962369895916734, "grad_norm": 0.435546875, "learning_rate": 4.7491189834256147e-05, "loss": 2.2863, "step": 717 }, { "epoch": 0.2299439551641313, "grad_norm": 0.5703125, "learning_rate": 4.747896793811721e-05, "loss": 2.4744, "step": 718 }, { "epoch": 0.23026421136909528, "grad_norm": 0.4921875, "learning_rate": 4.746671792474465e-05, "loss": 2.4798, "step": 719 }, { "epoch": 0.23058446757405926, "grad_norm": 0.60546875, "learning_rate": 4.7454439809461064e-05, "loss": 2.4989, "step": 720 }, { "epoch": 0.2309047237790232, "grad_norm": 0.482421875, "learning_rate": 4.7442133607624216e-05, "loss": 2.4011, "step": 721 }, { "epoch": 0.2312249799839872, "grad_norm": 0.498046875, "learning_rate": 4.7429799334627e-05, "loss": 1.8769, "step": 722 }, { "epoch": 0.23154523618895115, "grad_norm": 0.4296875, "learning_rate": 4.7417437005897415e-05, "loss": 2.1301, "step": 723 }, { "epoch": 0.23186549239391513, "grad_norm": 0.484375, "learning_rate": 4.740504663689857e-05, "loss": 1.8229, "step": 724 }, { "epoch": 0.2321857485988791, "grad_norm": 0.48828125, "learning_rate": 4.739262824312863e-05, "loss": 2.2332, "step": 725 }, { "epoch": 0.23250600480384306, "grad_norm": 0.462890625, "learning_rate": 4.7380181840120825e-05, "loss": 2.4415, "step": 726 }, { "epoch": 0.23282626100880705, "grad_norm": 0.80078125, "learning_rate": 4.736770744344341e-05, "loss": 2.4059, "step": 727 }, { "epoch": 0.23314651721377103, "grad_norm": 0.58203125, "learning_rate": 4.735520506869966e-05, "loss": 2.3278, "step": 728 }, { "epoch": 0.23346677341873498, "grad_norm": 0.439453125, "learning_rate": 4.734267473152785e-05, "loss": 2.3155, "step": 729 }, { "epoch": 0.23378702962369896, "grad_norm": 0.40234375, "learning_rate": 4.7330116447601225e-05, "loss": 1.8021, "step": 730 }, { "epoch": 0.23410728582866294, "grad_norm": 0.4765625, "learning_rate": 4.731753023262798e-05, "loss": 2.0426, "step": 731 }, { "epoch": 0.2344275420336269, "grad_norm": 0.59765625, "learning_rate": 4.730491610235128e-05, "loss": 1.5206, "step": 732 }, { "epoch": 0.23474779823859088, "grad_norm": 0.5859375, "learning_rate": 4.729227407254916e-05, "loss": 2.1711, "step": 733 }, { "epoch": 0.23506805444355483, "grad_norm": 0.578125, "learning_rate": 4.727960415903458e-05, "loss": 2.246, "step": 734 }, { "epoch": 0.2353883106485188, "grad_norm": 0.44140625, "learning_rate": 4.7266906377655375e-05, "loss": 1.8435, "step": 735 }, { "epoch": 0.2357085668534828, "grad_norm": 0.423828125, "learning_rate": 4.725418074429423e-05, "loss": 1.9849, "step": 736 }, { "epoch": 0.23602882305844675, "grad_norm": 0.41015625, "learning_rate": 4.724142727486869e-05, "loss": 2.6527, "step": 737 }, { "epoch": 0.23634907926341073, "grad_norm": 0.412109375, "learning_rate": 4.722864598533108e-05, "loss": 2.5422, "step": 738 }, { "epoch": 0.2366693354683747, "grad_norm": 0.60546875, "learning_rate": 4.721583689166855e-05, "loss": 2.0435, "step": 739 }, { "epoch": 0.23698959167333866, "grad_norm": 0.8515625, "learning_rate": 4.720300000990302e-05, "loss": 1.571, "step": 740 }, { "epoch": 0.23730984787830264, "grad_norm": 0.51953125, "learning_rate": 4.719013535609119e-05, "loss": 1.9472, "step": 741 }, { "epoch": 0.23763010408326662, "grad_norm": 0.5078125, "learning_rate": 4.717724294632446e-05, "loss": 2.4293, "step": 742 }, { "epoch": 0.23795036028823058, "grad_norm": 0.5, "learning_rate": 4.716432279672897e-05, "loss": 2.4161, "step": 743 }, { "epoch": 0.23827061649319456, "grad_norm": 0.486328125, "learning_rate": 4.7151374923465554e-05, "loss": 2.4954, "step": 744 }, { "epoch": 0.23859087269815854, "grad_norm": 1.1953125, "learning_rate": 4.7138399342729723e-05, "loss": 2.3457, "step": 745 }, { "epoch": 0.2389111289031225, "grad_norm": 1.6015625, "learning_rate": 4.712539607075164e-05, "loss": 2.2736, "step": 746 }, { "epoch": 0.23923138510808648, "grad_norm": 0.384765625, "learning_rate": 4.711236512379612e-05, "loss": 2.1532, "step": 747 }, { "epoch": 0.23955164131305043, "grad_norm": 0.70703125, "learning_rate": 4.709930651816258e-05, "loss": 2.0565, "step": 748 }, { "epoch": 0.2398718975180144, "grad_norm": 0.671875, "learning_rate": 4.708622027018504e-05, "loss": 2.5243, "step": 749 }, { "epoch": 0.2401921537229784, "grad_norm": 0.59765625, "learning_rate": 4.70731063962321e-05, "loss": 2.1374, "step": 750 }, { "epoch": 0.24051240992794234, "grad_norm": 0.52734375, "learning_rate": 4.705996491270689e-05, "loss": 2.0636, "step": 751 }, { "epoch": 0.24083266613290633, "grad_norm": 0.8515625, "learning_rate": 4.704679583604711e-05, "loss": 2.4744, "step": 752 }, { "epoch": 0.2411529223378703, "grad_norm": 0.6171875, "learning_rate": 4.703359918272495e-05, "loss": 2.2096, "step": 753 }, { "epoch": 0.24147317854283426, "grad_norm": 0.58203125, "learning_rate": 4.7020374969247106e-05, "loss": 2.576, "step": 754 }, { "epoch": 0.24179343474779824, "grad_norm": 0.6953125, "learning_rate": 4.700712321215474e-05, "loss": 2.3694, "step": 755 }, { "epoch": 0.24211369095276222, "grad_norm": 0.451171875, "learning_rate": 4.699384392802346e-05, "loss": 2.4635, "step": 756 }, { "epoch": 0.24243394715772618, "grad_norm": 0.5625, "learning_rate": 4.698053713346334e-05, "loss": 2.7555, "step": 757 }, { "epoch": 0.24275420336269016, "grad_norm": 0.6640625, "learning_rate": 4.696720284511881e-05, "loss": 2.4657, "step": 758 }, { "epoch": 0.2430744595676541, "grad_norm": 1.2734375, "learning_rate": 4.6953841079668724e-05, "loss": 1.6963, "step": 759 }, { "epoch": 0.2433947157726181, "grad_norm": 0.439453125, "learning_rate": 4.6940451853826316e-05, "loss": 2.1481, "step": 760 }, { "epoch": 0.24371497197758207, "grad_norm": 1.0546875, "learning_rate": 4.692703518433913e-05, "loss": 2.1694, "step": 761 }, { "epoch": 0.24403522818254603, "grad_norm": 0.47265625, "learning_rate": 4.6913591087989084e-05, "loss": 2.0632, "step": 762 }, { "epoch": 0.24435548438751, "grad_norm": 0.453125, "learning_rate": 4.690011958159235e-05, "loss": 2.0139, "step": 763 }, { "epoch": 0.244675740592474, "grad_norm": 0.3671875, "learning_rate": 4.6886620681999434e-05, "loss": 2.0248, "step": 764 }, { "epoch": 0.24499599679743794, "grad_norm": 0.4921875, "learning_rate": 4.6873094406095066e-05, "loss": 2.3785, "step": 765 }, { "epoch": 0.24531625300240192, "grad_norm": 0.515625, "learning_rate": 4.685954077079825e-05, "loss": 2.6487, "step": 766 }, { "epoch": 0.2456365092073659, "grad_norm": 0.953125, "learning_rate": 4.68459597930622e-05, "loss": 2.0873, "step": 767 }, { "epoch": 0.24595676541232986, "grad_norm": 0.53515625, "learning_rate": 4.683235148987432e-05, "loss": 2.0824, "step": 768 }, { "epoch": 0.24627702161729384, "grad_norm": 0.41796875, "learning_rate": 4.68187158782562e-05, "loss": 2.6039, "step": 769 }, { "epoch": 0.2465972778222578, "grad_norm": 0.71484375, "learning_rate": 4.680505297526361e-05, "loss": 2.1172, "step": 770 }, { "epoch": 0.24691753402722177, "grad_norm": 0.5546875, "learning_rate": 4.679136279798642e-05, "loss": 2.3141, "step": 771 }, { "epoch": 0.24723779023218576, "grad_norm": 0.474609375, "learning_rate": 4.677764536354864e-05, "loss": 2.0481, "step": 772 }, { "epoch": 0.2475580464371497, "grad_norm": 0.376953125, "learning_rate": 4.6763900689108365e-05, "loss": 1.7297, "step": 773 }, { "epoch": 0.2478783026421137, "grad_norm": 1.484375, "learning_rate": 4.675012879185777e-05, "loss": 2.4405, "step": 774 }, { "epoch": 0.24819855884707767, "grad_norm": 0.53125, "learning_rate": 4.673632968902306e-05, "loss": 2.1231, "step": 775 }, { "epoch": 0.24851881505204163, "grad_norm": 0.53125, "learning_rate": 4.67225033978645e-05, "loss": 1.7979, "step": 776 }, { "epoch": 0.2488390712570056, "grad_norm": 0.5, "learning_rate": 4.670864993567635e-05, "loss": 1.7725, "step": 777 }, { "epoch": 0.2491593274619696, "grad_norm": 0.396484375, "learning_rate": 4.6694769319786834e-05, "loss": 2.0591, "step": 778 }, { "epoch": 0.24947958366693354, "grad_norm": 0.47265625, "learning_rate": 4.668086156755818e-05, "loss": 2.1371, "step": 779 }, { "epoch": 0.24979983987189752, "grad_norm": 0.37109375, "learning_rate": 4.666692669638653e-05, "loss": 2.1823, "step": 780 }, { "epoch": 0.2501200960768615, "grad_norm": 0.734375, "learning_rate": 4.665296472370195e-05, "loss": 1.8962, "step": 781 }, { "epoch": 0.2504403522818255, "grad_norm": 0.40625, "learning_rate": 4.663897566696843e-05, "loss": 2.6859, "step": 782 }, { "epoch": 0.25076060848678944, "grad_norm": 0.671875, "learning_rate": 4.66249595436838e-05, "loss": 2.4368, "step": 783 }, { "epoch": 0.2510808646917534, "grad_norm": 0.419921875, "learning_rate": 4.661091637137977e-05, "loss": 1.895, "step": 784 }, { "epoch": 0.2514011208967174, "grad_norm": 0.369140625, "learning_rate": 4.659684616762188e-05, "loss": 2.1725, "step": 785 }, { "epoch": 0.25172137710168135, "grad_norm": 0.54296875, "learning_rate": 4.6582748950009475e-05, "loss": 2.7823, "step": 786 }, { "epoch": 0.2520416333066453, "grad_norm": 0.92578125, "learning_rate": 4.65686247361757e-05, "loss": 2.3225, "step": 787 }, { "epoch": 0.25236188951160926, "grad_norm": 0.48828125, "learning_rate": 4.655447354378745e-05, "loss": 2.1723, "step": 788 }, { "epoch": 0.25268214571657327, "grad_norm": 0.412109375, "learning_rate": 4.654029539054539e-05, "loss": 1.9019, "step": 789 }, { "epoch": 0.2530024019215372, "grad_norm": 0.451171875, "learning_rate": 4.652609029418389e-05, "loss": 2.178, "step": 790 }, { "epoch": 0.2533226581265012, "grad_norm": 1.03125, "learning_rate": 4.651185827247103e-05, "loss": 1.4949, "step": 791 }, { "epoch": 0.2536429143314652, "grad_norm": 0.58984375, "learning_rate": 4.649759934320855e-05, "loss": 2.6352, "step": 792 }, { "epoch": 0.25396317053642914, "grad_norm": 0.431640625, "learning_rate": 4.6483313524231874e-05, "loss": 2.1695, "step": 793 }, { "epoch": 0.2542834267413931, "grad_norm": 0.50390625, "learning_rate": 4.646900083341005e-05, "loss": 2.4847, "step": 794 }, { "epoch": 0.2546036829463571, "grad_norm": 0.5703125, "learning_rate": 4.645466128864573e-05, "loss": 2.0651, "step": 795 }, { "epoch": 0.25492393915132106, "grad_norm": 0.4765625, "learning_rate": 4.644029490787517e-05, "loss": 2.7276, "step": 796 }, { "epoch": 0.255244195356285, "grad_norm": 0.5625, "learning_rate": 4.642590170906816e-05, "loss": 2.5108, "step": 797 }, { "epoch": 0.255564451561249, "grad_norm": 0.44921875, "learning_rate": 4.6411481710228096e-05, "loss": 2.3693, "step": 798 }, { "epoch": 0.25588470776621297, "grad_norm": 0.4765625, "learning_rate": 4.6397034929391834e-05, "loss": 2.333, "step": 799 }, { "epoch": 0.2562049639711769, "grad_norm": 0.5078125, "learning_rate": 4.638256138462975e-05, "loss": 2.4582, "step": 800 }, { "epoch": 0.25652522017614093, "grad_norm": 0.365234375, "learning_rate": 4.636806109404573e-05, "loss": 2.5472, "step": 801 }, { "epoch": 0.2568454763811049, "grad_norm": 0.5078125, "learning_rate": 4.635353407577706e-05, "loss": 2.4438, "step": 802 }, { "epoch": 0.25716573258606884, "grad_norm": 0.44140625, "learning_rate": 4.633898034799449e-05, "loss": 2.3358, "step": 803 }, { "epoch": 0.25748598879103285, "grad_norm": 0.921875, "learning_rate": 4.632439992890218e-05, "loss": 2.4705, "step": 804 }, { "epoch": 0.2578062449959968, "grad_norm": 0.3984375, "learning_rate": 4.630979283673766e-05, "loss": 2.163, "step": 805 }, { "epoch": 0.25812650120096076, "grad_norm": 0.51953125, "learning_rate": 4.629515908977184e-05, "loss": 1.7262, "step": 806 }, { "epoch": 0.25844675740592477, "grad_norm": 0.4296875, "learning_rate": 4.628049870630896e-05, "loss": 2.2349, "step": 807 }, { "epoch": 0.2587670136108887, "grad_norm": 0.431640625, "learning_rate": 4.626581170468658e-05, "loss": 2.3561, "step": 808 }, { "epoch": 0.25908726981585267, "grad_norm": 0.458984375, "learning_rate": 4.625109810327556e-05, "loss": 2.5208, "step": 809 }, { "epoch": 0.2594075260208166, "grad_norm": 1.953125, "learning_rate": 4.6236357920480013e-05, "loss": 2.0921, "step": 810 }, { "epoch": 0.25972778222578063, "grad_norm": 0.828125, "learning_rate": 4.6221591174737314e-05, "loss": 2.2984, "step": 811 }, { "epoch": 0.2600480384307446, "grad_norm": 0.3359375, "learning_rate": 4.620679788451808e-05, "loss": 1.6466, "step": 812 }, { "epoch": 0.26036829463570854, "grad_norm": 0.427734375, "learning_rate": 4.61919780683261e-05, "loss": 2.3331, "step": 813 }, { "epoch": 0.26068855084067255, "grad_norm": 0.404296875, "learning_rate": 4.6177131744698364e-05, "loss": 2.313, "step": 814 }, { "epoch": 0.2610088070456365, "grad_norm": 0.5703125, "learning_rate": 4.616225893220502e-05, "loss": 2.4781, "step": 815 }, { "epoch": 0.26132906325060046, "grad_norm": 0.447265625, "learning_rate": 4.6147359649449305e-05, "loss": 2.2835, "step": 816 }, { "epoch": 0.26164931945556447, "grad_norm": 0.443359375, "learning_rate": 4.613243391506764e-05, "loss": 2.2348, "step": 817 }, { "epoch": 0.2619695756605284, "grad_norm": 0.345703125, "learning_rate": 4.6117481747729476e-05, "loss": 2.4842, "step": 818 }, { "epoch": 0.2622898318654924, "grad_norm": 0.51953125, "learning_rate": 4.6102503166137336e-05, "loss": 2.0446, "step": 819 }, { "epoch": 0.2626100880704564, "grad_norm": 0.46875, "learning_rate": 4.608749818902679e-05, "loss": 1.8785, "step": 820 }, { "epoch": 0.26293034427542034, "grad_norm": 0.76953125, "learning_rate": 4.607246683516644e-05, "loss": 2.5239, "step": 821 }, { "epoch": 0.2632506004803843, "grad_norm": 0.49609375, "learning_rate": 4.605740912335786e-05, "loss": 2.61, "step": 822 }, { "epoch": 0.2635708566853483, "grad_norm": 0.48046875, "learning_rate": 4.604232507243559e-05, "loss": 1.8585, "step": 823 }, { "epoch": 0.26389111289031225, "grad_norm": 0.73828125, "learning_rate": 4.6027214701267126e-05, "loss": 2.3605, "step": 824 }, { "epoch": 0.2642113690952762, "grad_norm": 0.3828125, "learning_rate": 4.6012078028752885e-05, "loss": 2.2571, "step": 825 }, { "epoch": 0.2645316253002402, "grad_norm": 0.40625, "learning_rate": 4.5996915073826187e-05, "loss": 1.8091, "step": 826 }, { "epoch": 0.26485188150520417, "grad_norm": 0.45703125, "learning_rate": 4.598172585545321e-05, "loss": 2.3873, "step": 827 }, { "epoch": 0.2651721377101681, "grad_norm": 0.5703125, "learning_rate": 4.5966510392633e-05, "loss": 2.171, "step": 828 }, { "epoch": 0.26549239391513213, "grad_norm": 2.75, "learning_rate": 4.595126870439742e-05, "loss": 2.4189, "step": 829 }, { "epoch": 0.2658126501200961, "grad_norm": 0.482421875, "learning_rate": 4.593600080981114e-05, "loss": 2.3643, "step": 830 }, { "epoch": 0.26613290632506004, "grad_norm": 0.52734375, "learning_rate": 4.592070672797161e-05, "loss": 2.1314, "step": 831 }, { "epoch": 0.26645316253002405, "grad_norm": 0.6328125, "learning_rate": 4.590538647800904e-05, "loss": 2.5095, "step": 832 }, { "epoch": 0.266773418734988, "grad_norm": 0.6015625, "learning_rate": 4.5890040079086374e-05, "loss": 2.1051, "step": 833 }, { "epoch": 0.26709367493995195, "grad_norm": 0.54296875, "learning_rate": 4.587466755039924e-05, "loss": 2.6451, "step": 834 }, { "epoch": 0.2674139311449159, "grad_norm": 0.486328125, "learning_rate": 4.585926891117597e-05, "loss": 2.3557, "step": 835 }, { "epoch": 0.2677341873498799, "grad_norm": 0.435546875, "learning_rate": 4.584384418067756e-05, "loss": 2.2055, "step": 836 }, { "epoch": 0.26805444355484387, "grad_norm": 0.55859375, "learning_rate": 4.582839337819763e-05, "loss": 1.711, "step": 837 }, { "epoch": 0.2683746997598078, "grad_norm": 1.6640625, "learning_rate": 4.581291652306241e-05, "loss": 2.7919, "step": 838 }, { "epoch": 0.26869495596477183, "grad_norm": 0.51171875, "learning_rate": 4.579741363463074e-05, "loss": 2.3713, "step": 839 }, { "epoch": 0.2690152121697358, "grad_norm": 0.625, "learning_rate": 4.578188473229398e-05, "loss": 2.4531, "step": 840 }, { "epoch": 0.26933546837469974, "grad_norm": 0.48828125, "learning_rate": 4.576632983547607e-05, "loss": 2.2186, "step": 841 }, { "epoch": 0.26965572457966375, "grad_norm": 0.51953125, "learning_rate": 4.575074896363345e-05, "loss": 2.3781, "step": 842 }, { "epoch": 0.2699759807846277, "grad_norm": 0.482421875, "learning_rate": 4.573514213625505e-05, "loss": 2.1847, "step": 843 }, { "epoch": 0.27029623698959165, "grad_norm": 0.421875, "learning_rate": 4.571950937286225e-05, "loss": 2.1921, "step": 844 }, { "epoch": 0.27061649319455566, "grad_norm": 0.3984375, "learning_rate": 4.570385069300891e-05, "loss": 2.6008, "step": 845 }, { "epoch": 0.2709367493995196, "grad_norm": 0.671875, "learning_rate": 4.568816611628126e-05, "loss": 2.5111, "step": 846 }, { "epoch": 0.27125700560448357, "grad_norm": 0.412109375, "learning_rate": 4.5672455662297966e-05, "loss": 2.4546, "step": 847 }, { "epoch": 0.2715772618094476, "grad_norm": 0.490234375, "learning_rate": 4.565671935071002e-05, "loss": 2.2381, "step": 848 }, { "epoch": 0.27189751801441153, "grad_norm": 0.66015625, "learning_rate": 4.56409572012008e-05, "loss": 2.5955, "step": 849 }, { "epoch": 0.2722177742193755, "grad_norm": 0.51171875, "learning_rate": 4.562516923348597e-05, "loss": 2.2343, "step": 850 }, { "epoch": 0.2725380304243395, "grad_norm": 0.45703125, "learning_rate": 4.56093554673135e-05, "loss": 2.4699, "step": 851 }, { "epoch": 0.27285828662930345, "grad_norm": 0.5703125, "learning_rate": 4.559351592246363e-05, "loss": 2.4249, "step": 852 }, { "epoch": 0.2731785428342674, "grad_norm": 0.9765625, "learning_rate": 4.5577650618748846e-05, "loss": 1.7957, "step": 853 }, { "epoch": 0.2734987990392314, "grad_norm": 0.61328125, "learning_rate": 4.556175957601385e-05, "loss": 2.3653, "step": 854 }, { "epoch": 0.27381905524419536, "grad_norm": 0.4765625, "learning_rate": 4.5545842814135545e-05, "loss": 2.7702, "step": 855 }, { "epoch": 0.2741393114491593, "grad_norm": 0.453125, "learning_rate": 4.552990035302299e-05, "loss": 2.3475, "step": 856 }, { "epoch": 0.27445956765412327, "grad_norm": 0.61328125, "learning_rate": 4.55139322126174e-05, "loss": 2.1684, "step": 857 }, { "epoch": 0.2747798238590873, "grad_norm": 0.396484375, "learning_rate": 4.549793841289211e-05, "loss": 1.8475, "step": 858 }, { "epoch": 0.27510008006405123, "grad_norm": 1.1953125, "learning_rate": 4.548191897385257e-05, "loss": 2.3642, "step": 859 }, { "epoch": 0.2754203362690152, "grad_norm": 0.458984375, "learning_rate": 4.546587391553624e-05, "loss": 1.9096, "step": 860 }, { "epoch": 0.2757405924739792, "grad_norm": 0.54296875, "learning_rate": 4.54498032580127e-05, "loss": 2.4123, "step": 861 }, { "epoch": 0.27606084867894315, "grad_norm": 0.5703125, "learning_rate": 4.54337070213835e-05, "loss": 2.431, "step": 862 }, { "epoch": 0.2763811048839071, "grad_norm": 0.49609375, "learning_rate": 4.54175852257822e-05, "loss": 2.5574, "step": 863 }, { "epoch": 0.2767013610888711, "grad_norm": 0.46875, "learning_rate": 4.5401437891374326e-05, "loss": 2.4494, "step": 864 }, { "epoch": 0.27702161729383507, "grad_norm": 0.416015625, "learning_rate": 4.5385265038357363e-05, "loss": 2.2868, "step": 865 }, { "epoch": 0.277341873498799, "grad_norm": 1.3359375, "learning_rate": 4.5369066686960694e-05, "loss": 2.2944, "step": 866 }, { "epoch": 0.277662129703763, "grad_norm": 0.7734375, "learning_rate": 4.535284285744561e-05, "loss": 1.8372, "step": 867 }, { "epoch": 0.277982385908727, "grad_norm": 0.5, "learning_rate": 4.5336593570105266e-05, "loss": 2.3709, "step": 868 }, { "epoch": 0.27830264211369093, "grad_norm": 0.443359375, "learning_rate": 4.5320318845264664e-05, "loss": 2.1215, "step": 869 }, { "epoch": 0.27862289831865494, "grad_norm": 0.61328125, "learning_rate": 4.530401870328063e-05, "loss": 2.8229, "step": 870 }, { "epoch": 0.2789431545236189, "grad_norm": 0.9765625, "learning_rate": 4.528769316454176e-05, "loss": 2.272, "step": 871 }, { "epoch": 0.27926341072858285, "grad_norm": 0.546875, "learning_rate": 4.5271342249468424e-05, "loss": 2.1168, "step": 872 }, { "epoch": 0.27958366693354686, "grad_norm": 0.33203125, "learning_rate": 4.525496597851275e-05, "loss": 1.5178, "step": 873 }, { "epoch": 0.2799039231385108, "grad_norm": 0.44140625, "learning_rate": 4.523856437215858e-05, "loss": 1.6794, "step": 874 }, { "epoch": 0.28022417934347477, "grad_norm": 0.375, "learning_rate": 4.522213745092143e-05, "loss": 2.3376, "step": 875 }, { "epoch": 0.2805444355484388, "grad_norm": 0.62890625, "learning_rate": 4.520568523534848e-05, "loss": 2.4695, "step": 876 }, { "epoch": 0.28086469175340273, "grad_norm": 0.4609375, "learning_rate": 4.5189207746018566e-05, "loss": 1.6355, "step": 877 }, { "epoch": 0.2811849479583667, "grad_norm": 0.49609375, "learning_rate": 4.5172705003542125e-05, "loss": 2.0943, "step": 878 }, { "epoch": 0.2815052041633307, "grad_norm": 0.62109375, "learning_rate": 4.515617702856119e-05, "loss": 1.9592, "step": 879 }, { "epoch": 0.28182546036829464, "grad_norm": 0.484375, "learning_rate": 4.5139623841749346e-05, "loss": 2.0776, "step": 880 }, { "epoch": 0.2821457165732586, "grad_norm": 0.5859375, "learning_rate": 4.5123045463811706e-05, "loss": 2.2985, "step": 881 }, { "epoch": 0.28246597277822255, "grad_norm": 0.431640625, "learning_rate": 4.510644191548492e-05, "loss": 1.9178, "step": 882 }, { "epoch": 0.28278622898318656, "grad_norm": 0.3984375, "learning_rate": 4.50898132175371e-05, "loss": 2.3869, "step": 883 }, { "epoch": 0.2831064851881505, "grad_norm": 0.625, "learning_rate": 4.507315939076782e-05, "loss": 1.5066, "step": 884 }, { "epoch": 0.28342674139311447, "grad_norm": 0.435546875, "learning_rate": 4.5056480456008085e-05, "loss": 2.5622, "step": 885 }, { "epoch": 0.2837469975980785, "grad_norm": 1.0546875, "learning_rate": 4.5039776434120306e-05, "loss": 2.3004, "step": 886 }, { "epoch": 0.28406725380304243, "grad_norm": 0.5078125, "learning_rate": 4.502304734599828e-05, "loss": 2.3096, "step": 887 }, { "epoch": 0.2843875100080064, "grad_norm": 0.6015625, "learning_rate": 4.5006293212567164e-05, "loss": 2.1645, "step": 888 }, { "epoch": 0.2847077662129704, "grad_norm": 0.5625, "learning_rate": 4.4989514054783406e-05, "loss": 2.2932, "step": 889 }, { "epoch": 0.28502802241793435, "grad_norm": 0.462890625, "learning_rate": 4.49727098936348e-05, "loss": 2.1796, "step": 890 }, { "epoch": 0.2853482786228983, "grad_norm": 0.404296875, "learning_rate": 4.495588075014038e-05, "loss": 1.7093, "step": 891 }, { "epoch": 0.2856685348278623, "grad_norm": 0.4921875, "learning_rate": 4.4939026645350454e-05, "loss": 2.5258, "step": 892 }, { "epoch": 0.28598879103282626, "grad_norm": 2.9375, "learning_rate": 4.4922147600346545e-05, "loss": 1.979, "step": 893 }, { "epoch": 0.2863090472377902, "grad_norm": 0.361328125, "learning_rate": 4.4905243636241354e-05, "loss": 2.0999, "step": 894 }, { "epoch": 0.2866293034427542, "grad_norm": 0.53515625, "learning_rate": 4.4888314774178776e-05, "loss": 2.4738, "step": 895 }, { "epoch": 0.2869495596477182, "grad_norm": 0.8046875, "learning_rate": 4.4871361035333836e-05, "loss": 2.3017, "step": 896 }, { "epoch": 0.28726981585268213, "grad_norm": 0.384765625, "learning_rate": 4.485438244091268e-05, "loss": 2.2876, "step": 897 }, { "epoch": 0.28759007205764614, "grad_norm": 0.458984375, "learning_rate": 4.483737901215254e-05, "loss": 2.1632, "step": 898 }, { "epoch": 0.2879103282626101, "grad_norm": 0.54296875, "learning_rate": 4.482035077032171e-05, "loss": 2.4111, "step": 899 }, { "epoch": 0.28823058446757405, "grad_norm": 0.36328125, "learning_rate": 4.4803297736719534e-05, "loss": 2.6698, "step": 900 }, { "epoch": 0.28855084067253806, "grad_norm": 1.5390625, "learning_rate": 4.478621993267635e-05, "loss": 2.4385, "step": 901 }, { "epoch": 0.288871096877502, "grad_norm": 1.1953125, "learning_rate": 4.476911737955349e-05, "loss": 1.9053, "step": 902 }, { "epoch": 0.28919135308246596, "grad_norm": 0.56640625, "learning_rate": 4.475199009874323e-05, "loss": 1.8812, "step": 903 }, { "epoch": 0.2895116092874299, "grad_norm": 0.40234375, "learning_rate": 4.47348381116688e-05, "loss": 1.812, "step": 904 }, { "epoch": 0.2898318654923939, "grad_norm": 0.50390625, "learning_rate": 4.47176614397843e-05, "loss": 2.3822, "step": 905 }, { "epoch": 0.2901521216973579, "grad_norm": 0.8828125, "learning_rate": 4.470046010457473e-05, "loss": 1.8144, "step": 906 }, { "epoch": 0.29047237790232183, "grad_norm": 0.6171875, "learning_rate": 4.4683234127555944e-05, "loss": 2.3208, "step": 907 }, { "epoch": 0.29079263410728584, "grad_norm": 0.42578125, "learning_rate": 4.46659835302746e-05, "loss": 2.0736, "step": 908 }, { "epoch": 0.2911128903122498, "grad_norm": 0.515625, "learning_rate": 4.4648708334308156e-05, "loss": 2.6272, "step": 909 }, { "epoch": 0.29143314651721375, "grad_norm": 0.466796875, "learning_rate": 4.463140856126485e-05, "loss": 2.2224, "step": 910 }, { "epoch": 0.29175340272217776, "grad_norm": 0.45703125, "learning_rate": 4.461408423278365e-05, "loss": 2.1467, "step": 911 }, { "epoch": 0.2920736589271417, "grad_norm": 0.74609375, "learning_rate": 4.4596735370534246e-05, "loss": 2.5051, "step": 912 }, { "epoch": 0.29239391513210566, "grad_norm": 0.466796875, "learning_rate": 4.457936199621701e-05, "loss": 2.0413, "step": 913 }, { "epoch": 0.2927141713370697, "grad_norm": 0.5546875, "learning_rate": 4.456196413156297e-05, "loss": 2.1757, "step": 914 }, { "epoch": 0.2930344275420336, "grad_norm": 0.68359375, "learning_rate": 4.45445417983338e-05, "loss": 2.51, "step": 915 }, { "epoch": 0.2933546837469976, "grad_norm": 0.64453125, "learning_rate": 4.4527095018321786e-05, "loss": 2.3372, "step": 916 }, { "epoch": 0.2936749399519616, "grad_norm": 0.82421875, "learning_rate": 4.4509623813349755e-05, "loss": 2.4337, "step": 917 }, { "epoch": 0.29399519615692554, "grad_norm": 1.109375, "learning_rate": 4.4492128205271134e-05, "loss": 2.0431, "step": 918 }, { "epoch": 0.2943154523618895, "grad_norm": 0.51953125, "learning_rate": 4.4474608215969825e-05, "loss": 2.0454, "step": 919 }, { "epoch": 0.2946357085668535, "grad_norm": 0.60546875, "learning_rate": 4.445706386736027e-05, "loss": 1.6327, "step": 920 }, { "epoch": 0.29495596477181746, "grad_norm": 0.57421875, "learning_rate": 4.4439495181387355e-05, "loss": 2.1912, "step": 921 }, { "epoch": 0.2952762209767814, "grad_norm": 1.4609375, "learning_rate": 4.442190218002643e-05, "loss": 2.5112, "step": 922 }, { "epoch": 0.2955964771817454, "grad_norm": 0.64453125, "learning_rate": 4.440428488528322e-05, "loss": 2.2843, "step": 923 }, { "epoch": 0.2959167333867094, "grad_norm": 0.408203125, "learning_rate": 4.4386643319193874e-05, "loss": 2.0363, "step": 924 }, { "epoch": 0.2962369895916733, "grad_norm": 0.515625, "learning_rate": 4.436897750382489e-05, "loss": 2.4265, "step": 925 }, { "epoch": 0.29655724579663734, "grad_norm": 0.58984375, "learning_rate": 4.435128746127309e-05, "loss": 2.0594, "step": 926 }, { "epoch": 0.2968775020016013, "grad_norm": 0.482421875, "learning_rate": 4.4333573213665604e-05, "loss": 2.3537, "step": 927 }, { "epoch": 0.29719775820656524, "grad_norm": 0.52734375, "learning_rate": 4.431583478315984e-05, "loss": 2.7564, "step": 928 }, { "epoch": 0.2975180144115292, "grad_norm": 0.49609375, "learning_rate": 4.429807219194345e-05, "loss": 2.0438, "step": 929 }, { "epoch": 0.2978382706164932, "grad_norm": 0.6015625, "learning_rate": 4.428028546223432e-05, "loss": 2.0035, "step": 930 }, { "epoch": 0.29815852682145716, "grad_norm": 0.87109375, "learning_rate": 4.4262474616280506e-05, "loss": 1.9675, "step": 931 }, { "epoch": 0.2984787830264211, "grad_norm": 0.74609375, "learning_rate": 4.4244639676360253e-05, "loss": 2.2379, "step": 932 }, { "epoch": 0.2987990392313851, "grad_norm": 1.8359375, "learning_rate": 4.422678066478192e-05, "loss": 1.8438, "step": 933 }, { "epoch": 0.2991192954363491, "grad_norm": 0.4765625, "learning_rate": 4.4208897603884016e-05, "loss": 2.166, "step": 934 }, { "epoch": 0.29943955164131303, "grad_norm": 0.4609375, "learning_rate": 4.419099051603507e-05, "loss": 2.4782, "step": 935 }, { "epoch": 0.29975980784627704, "grad_norm": 0.37890625, "learning_rate": 4.417305942363371e-05, "loss": 1.9755, "step": 936 }, { "epoch": 0.300080064051241, "grad_norm": 0.5625, "learning_rate": 4.415510434910859e-05, "loss": 2.6677, "step": 937 }, { "epoch": 0.30040032025620494, "grad_norm": 0.494140625, "learning_rate": 4.413712531491833e-05, "loss": 2.0538, "step": 938 }, { "epoch": 0.30072057646116895, "grad_norm": 0.6328125, "learning_rate": 4.411912234355156e-05, "loss": 2.4744, "step": 939 }, { "epoch": 0.3010408326661329, "grad_norm": 0.8203125, "learning_rate": 4.41010954575268e-05, "loss": 2.4799, "step": 940 }, { "epoch": 0.30136108887109686, "grad_norm": 0.4609375, "learning_rate": 4.408304467939254e-05, "loss": 2.5785, "step": 941 }, { "epoch": 0.30168134507606087, "grad_norm": 0.65234375, "learning_rate": 4.4064970031727114e-05, "loss": 2.2337, "step": 942 }, { "epoch": 0.3020016012810248, "grad_norm": 0.5703125, "learning_rate": 4.4046871537138734e-05, "loss": 2.5195, "step": 943 }, { "epoch": 0.3023218574859888, "grad_norm": 0.62109375, "learning_rate": 4.402874921826542e-05, "loss": 2.3741, "step": 944 }, { "epoch": 0.3026421136909528, "grad_norm": 0.5234375, "learning_rate": 4.4010603097775015e-05, "loss": 2.4184, "step": 945 }, { "epoch": 0.30296236989591674, "grad_norm": 0.267578125, "learning_rate": 4.399243319836512e-05, "loss": 2.1592, "step": 946 }, { "epoch": 0.3032826261008807, "grad_norm": 0.419921875, "learning_rate": 4.397423954276308e-05, "loss": 1.8628, "step": 947 }, { "epoch": 0.3036028823058447, "grad_norm": 0.439453125, "learning_rate": 4.395602215372596e-05, "loss": 1.8973, "step": 948 }, { "epoch": 0.30392313851080865, "grad_norm": 0.5546875, "learning_rate": 4.3937781054040505e-05, "loss": 2.1797, "step": 949 }, { "epoch": 0.3042433947157726, "grad_norm": 2.40625, "learning_rate": 4.391951626652312e-05, "loss": 2.0715, "step": 950 }, { "epoch": 0.30456365092073656, "grad_norm": 0.40234375, "learning_rate": 4.3901227814019845e-05, "loss": 1.859, "step": 951 }, { "epoch": 0.30488390712570057, "grad_norm": 0.40234375, "learning_rate": 4.388291571940631e-05, "loss": 2.4701, "step": 952 }, { "epoch": 0.3052041633306645, "grad_norm": 0.53125, "learning_rate": 4.386458000558773e-05, "loss": 2.0972, "step": 953 }, { "epoch": 0.3055244195356285, "grad_norm": 8.5, "learning_rate": 4.384622069549886e-05, "loss": 1.9298, "step": 954 }, { "epoch": 0.3058446757405925, "grad_norm": 0.5859375, "learning_rate": 4.382783781210395e-05, "loss": 2.1858, "step": 955 }, { "epoch": 0.30616493194555644, "grad_norm": 0.453125, "learning_rate": 4.3809431378396776e-05, "loss": 2.1796, "step": 956 }, { "epoch": 0.3064851881505204, "grad_norm": 0.50390625, "learning_rate": 4.379100141740053e-05, "loss": 2.071, "step": 957 }, { "epoch": 0.3068054443554844, "grad_norm": 1.8359375, "learning_rate": 4.377254795216785e-05, "loss": 2.5639, "step": 958 }, { "epoch": 0.30712570056044836, "grad_norm": 0.859375, "learning_rate": 4.3754071005780785e-05, "loss": 2.4597, "step": 959 }, { "epoch": 0.3074459567654123, "grad_norm": 0.8984375, "learning_rate": 4.3735570601350736e-05, "loss": 2.6462, "step": 960 }, { "epoch": 0.3077662129703763, "grad_norm": 0.5390625, "learning_rate": 4.3717046762018454e-05, "loss": 2.2894, "step": 961 }, { "epoch": 0.30808646917534027, "grad_norm": 0.486328125, "learning_rate": 4.369849951095401e-05, "loss": 1.6288, "step": 962 }, { "epoch": 0.3084067253803042, "grad_norm": 0.298828125, "learning_rate": 4.3679928871356743e-05, "loss": 1.4223, "step": 963 }, { "epoch": 0.30872698158526823, "grad_norm": 0.5390625, "learning_rate": 4.366133486645526e-05, "loss": 2.0063, "step": 964 }, { "epoch": 0.3090472377902322, "grad_norm": 0.4921875, "learning_rate": 4.364271751950739e-05, "loss": 2.1687, "step": 965 }, { "epoch": 0.30936749399519614, "grad_norm": 0.6015625, "learning_rate": 4.362407685380015e-05, "loss": 2.4734, "step": 966 }, { "epoch": 0.30968775020016015, "grad_norm": 0.51171875, "learning_rate": 4.3605412892649744e-05, "loss": 2.215, "step": 967 }, { "epoch": 0.3100080064051241, "grad_norm": 1.5078125, "learning_rate": 4.3586725659401504e-05, "loss": 2.1618, "step": 968 }, { "epoch": 0.31032826261008806, "grad_norm": 0.4296875, "learning_rate": 4.356801517742986e-05, "loss": 2.6403, "step": 969 }, { "epoch": 0.31064851881505207, "grad_norm": 0.51171875, "learning_rate": 4.3549281470138334e-05, "loss": 2.0227, "step": 970 }, { "epoch": 0.310968775020016, "grad_norm": 0.7578125, "learning_rate": 4.353052456095952e-05, "loss": 2.4126, "step": 971 }, { "epoch": 0.31128903122498, "grad_norm": 0.609375, "learning_rate": 4.351174447335498e-05, "loss": 2.1611, "step": 972 }, { "epoch": 0.311609287429944, "grad_norm": 0.59375, "learning_rate": 4.349294123081531e-05, "loss": 2.2805, "step": 973 }, { "epoch": 0.31192954363490794, "grad_norm": 0.56640625, "learning_rate": 4.347411485686006e-05, "loss": 2.2679, "step": 974 }, { "epoch": 0.3122497998398719, "grad_norm": 0.7421875, "learning_rate": 4.345526537503771e-05, "loss": 2.2008, "step": 975 }, { "epoch": 0.31257005604483584, "grad_norm": 2.0, "learning_rate": 4.3436392808925654e-05, "loss": 1.8056, "step": 976 }, { "epoch": 0.31289031224979985, "grad_norm": 0.5234375, "learning_rate": 4.341749718213014e-05, "loss": 2.3376, "step": 977 }, { "epoch": 0.3132105684547638, "grad_norm": 0.91015625, "learning_rate": 4.339857851828628e-05, "loss": 2.4653, "step": 978 }, { "epoch": 0.31353082465972776, "grad_norm": 0.828125, "learning_rate": 4.337963684105799e-05, "loss": 2.2227, "step": 979 }, { "epoch": 0.31385108086469177, "grad_norm": 0.375, "learning_rate": 4.336067217413797e-05, "loss": 2.2355, "step": 980 }, { "epoch": 0.3141713370696557, "grad_norm": 0.490234375, "learning_rate": 4.334168454124769e-05, "loss": 2.4895, "step": 981 }, { "epoch": 0.3144915932746197, "grad_norm": 0.43359375, "learning_rate": 4.332267396613734e-05, "loss": 1.9417, "step": 982 }, { "epoch": 0.3148118494795837, "grad_norm": 0.53125, "learning_rate": 4.330364047258579e-05, "loss": 2.6751, "step": 983 }, { "epoch": 0.31513210568454764, "grad_norm": 0.369140625, "learning_rate": 4.3284584084400604e-05, "loss": 1.9346, "step": 984 }, { "epoch": 0.3154523618895116, "grad_norm": 0.46875, "learning_rate": 4.3265504825417966e-05, "loss": 2.1739, "step": 985 }, { "epoch": 0.3157726180944756, "grad_norm": 0.388671875, "learning_rate": 4.3246402719502676e-05, "loss": 1.7595, "step": 986 }, { "epoch": 0.31609287429943955, "grad_norm": 0.5078125, "learning_rate": 4.3227277790548104e-05, "loss": 2.2438, "step": 987 }, { "epoch": 0.3164131305044035, "grad_norm": 0.546875, "learning_rate": 4.3208130062476174e-05, "loss": 2.4872, "step": 988 }, { "epoch": 0.3167333867093675, "grad_norm": 0.46875, "learning_rate": 4.318895955923731e-05, "loss": 2.4345, "step": 989 }, { "epoch": 0.31705364291433147, "grad_norm": 0.447265625, "learning_rate": 4.316976630481046e-05, "loss": 2.2613, "step": 990 }, { "epoch": 0.3173738991192954, "grad_norm": 0.46484375, "learning_rate": 4.3150550323203e-05, "loss": 2.1071, "step": 991 }, { "epoch": 0.31769415532425943, "grad_norm": 0.5234375, "learning_rate": 4.313131163845073e-05, "loss": 2.1463, "step": 992 }, { "epoch": 0.3180144115292234, "grad_norm": 0.98828125, "learning_rate": 4.311205027461789e-05, "loss": 1.4839, "step": 993 }, { "epoch": 0.31833466773418734, "grad_norm": 0.640625, "learning_rate": 4.309276625579703e-05, "loss": 2.4197, "step": 994 }, { "epoch": 0.31865492393915135, "grad_norm": 0.35546875, "learning_rate": 4.307345960610906e-05, "loss": 2.0932, "step": 995 }, { "epoch": 0.3189751801441153, "grad_norm": 0.5703125, "learning_rate": 4.305413034970322e-05, "loss": 2.1792, "step": 996 }, { "epoch": 0.31929543634907925, "grad_norm": 0.421875, "learning_rate": 4.3034778510757e-05, "loss": 2.5835, "step": 997 }, { "epoch": 0.3196156925540432, "grad_norm": 0.5078125, "learning_rate": 4.3015404113476156e-05, "loss": 1.9542, "step": 998 }, { "epoch": 0.3199359487590072, "grad_norm": 0.384765625, "learning_rate": 4.2996007182094634e-05, "loss": 1.7533, "step": 999 }, { "epoch": 0.32025620496397117, "grad_norm": 0.66796875, "learning_rate": 4.297658774087459e-05, "loss": 2.4536, "step": 1000 }, { "epoch": 0.3205764611689351, "grad_norm": 0.796875, "learning_rate": 4.295714581410633e-05, "loss": 2.0973, "step": 1001 }, { "epoch": 0.32089671737389913, "grad_norm": 0.419921875, "learning_rate": 4.293768142610828e-05, "loss": 2.5078, "step": 1002 }, { "epoch": 0.3212169735788631, "grad_norm": 0.7734375, "learning_rate": 4.291819460122697e-05, "loss": 1.7711, "step": 1003 }, { "epoch": 0.32153722978382704, "grad_norm": 0.3828125, "learning_rate": 4.2898685363836974e-05, "loss": 2.1171, "step": 1004 }, { "epoch": 0.32185748598879105, "grad_norm": 0.5234375, "learning_rate": 4.287915373834094e-05, "loss": 2.2499, "step": 1005 }, { "epoch": 0.322177742193755, "grad_norm": 0.51171875, "learning_rate": 4.2859599749169474e-05, "loss": 1.9514, "step": 1006 }, { "epoch": 0.32249799839871895, "grad_norm": 0.44140625, "learning_rate": 4.284002342078119e-05, "loss": 2.2806, "step": 1007 }, { "epoch": 0.32281825460368296, "grad_norm": 0.484375, "learning_rate": 4.2820424777662616e-05, "loss": 1.9288, "step": 1008 }, { "epoch": 0.3231385108086469, "grad_norm": 0.55859375, "learning_rate": 4.280080384432821e-05, "loss": 2.4497, "step": 1009 }, { "epoch": 0.32345876701361087, "grad_norm": 0.53125, "learning_rate": 4.278116064532032e-05, "loss": 2.273, "step": 1010 }, { "epoch": 0.3237790232185749, "grad_norm": 0.64453125, "learning_rate": 4.2761495205209114e-05, "loss": 2.2382, "step": 1011 }, { "epoch": 0.32409927942353883, "grad_norm": 0.4921875, "learning_rate": 4.274180754859261e-05, "loss": 2.0357, "step": 1012 }, { "epoch": 0.3244195356285028, "grad_norm": 0.6171875, "learning_rate": 4.27220977000966e-05, "loss": 1.655, "step": 1013 }, { "epoch": 0.3247397918334668, "grad_norm": 0.49609375, "learning_rate": 4.2702365684374626e-05, "loss": 2.2382, "step": 1014 }, { "epoch": 0.32506004803843075, "grad_norm": 0.322265625, "learning_rate": 4.2682611526107986e-05, "loss": 1.7932, "step": 1015 }, { "epoch": 0.3253803042433947, "grad_norm": 0.59375, "learning_rate": 4.266283525000564e-05, "loss": 2.5412, "step": 1016 }, { "epoch": 0.3257005604483587, "grad_norm": 0.6953125, "learning_rate": 4.2643036880804236e-05, "loss": 2.1839, "step": 1017 }, { "epoch": 0.32602081665332266, "grad_norm": 0.5078125, "learning_rate": 4.2623216443268055e-05, "loss": 2.3921, "step": 1018 }, { "epoch": 0.3263410728582866, "grad_norm": 0.53515625, "learning_rate": 4.2603373962188966e-05, "loss": 2.1971, "step": 1019 }, { "epoch": 0.3266613290632506, "grad_norm": 0.7890625, "learning_rate": 4.258350946238643e-05, "loss": 2.4754, "step": 1020 }, { "epoch": 0.3269815852682146, "grad_norm": 0.60546875, "learning_rate": 4.256362296870744e-05, "loss": 2.1899, "step": 1021 }, { "epoch": 0.32730184147317853, "grad_norm": 0.47265625, "learning_rate": 4.25437145060265e-05, "loss": 2.5574, "step": 1022 }, { "epoch": 0.3276220976781425, "grad_norm": 0.55859375, "learning_rate": 4.252378409924559e-05, "loss": 2.1373, "step": 1023 }, { "epoch": 0.3279423538831065, "grad_norm": 0.439453125, "learning_rate": 4.2503831773294144e-05, "loss": 2.4083, "step": 1024 }, { "epoch": 0.32826261008807045, "grad_norm": 0.466796875, "learning_rate": 4.248385755312901e-05, "loss": 1.7979, "step": 1025 }, { "epoch": 0.3285828662930344, "grad_norm": 0.546875, "learning_rate": 4.246386146373444e-05, "loss": 2.0333, "step": 1026 }, { "epoch": 0.3289031224979984, "grad_norm": 0.404296875, "learning_rate": 4.244384353012199e-05, "loss": 1.9149, "step": 1027 }, { "epoch": 0.32922337870296237, "grad_norm": 0.796875, "learning_rate": 4.2423803777330606e-05, "loss": 2.0417, "step": 1028 }, { "epoch": 0.3295436349079263, "grad_norm": 0.431640625, "learning_rate": 4.240374223042647e-05, "loss": 1.9744, "step": 1029 }, { "epoch": 0.32986389111289033, "grad_norm": 0.484375, "learning_rate": 4.2383658914503056e-05, "loss": 2.2739, "step": 1030 }, { "epoch": 0.3301841473178543, "grad_norm": 0.30078125, "learning_rate": 4.236355385468106e-05, "loss": 1.5005, "step": 1031 }, { "epoch": 0.33050440352281824, "grad_norm": 0.451171875, "learning_rate": 4.234342707610837e-05, "loss": 2.1598, "step": 1032 }, { "epoch": 0.33082465972778224, "grad_norm": 0.99609375, "learning_rate": 4.2323278603960046e-05, "loss": 2.4382, "step": 1033 }, { "epoch": 0.3311449159327462, "grad_norm": 0.380859375, "learning_rate": 4.230310846343828e-05, "loss": 1.963, "step": 1034 }, { "epoch": 0.33146517213771015, "grad_norm": 0.390625, "learning_rate": 4.228291667977238e-05, "loss": 1.9068, "step": 1035 }, { "epoch": 0.33178542834267416, "grad_norm": 0.42578125, "learning_rate": 4.22627032782187e-05, "loss": 1.4926, "step": 1036 }, { "epoch": 0.3321056845476381, "grad_norm": 0.5, "learning_rate": 4.2242468284060644e-05, "loss": 2.3178, "step": 1037 }, { "epoch": 0.33242594075260207, "grad_norm": 0.41796875, "learning_rate": 4.222221172260865e-05, "loss": 2.6006, "step": 1038 }, { "epoch": 0.3327461969575661, "grad_norm": 0.5390625, "learning_rate": 4.2201933619200095e-05, "loss": 2.1655, "step": 1039 }, { "epoch": 0.33306645316253003, "grad_norm": 0.875, "learning_rate": 4.218163399919933e-05, "loss": 2.2719, "step": 1040 }, { "epoch": 0.333386709367494, "grad_norm": 0.8203125, "learning_rate": 4.216131288799761e-05, "loss": 2.1145, "step": 1041 }, { "epoch": 0.333706965572458, "grad_norm": 0.498046875, "learning_rate": 4.214097031101305e-05, "loss": 1.9372, "step": 1042 }, { "epoch": 0.33402722177742195, "grad_norm": 0.59765625, "learning_rate": 4.212060629369065e-05, "loss": 2.4681, "step": 1043 }, { "epoch": 0.3343474779823859, "grad_norm": 0.37890625, "learning_rate": 4.210022086150221e-05, "loss": 1.957, "step": 1044 }, { "epoch": 0.33466773418734985, "grad_norm": 0.36328125, "learning_rate": 4.207981403994632e-05, "loss": 2.0795, "step": 1045 }, { "epoch": 0.33498799039231386, "grad_norm": 0.52734375, "learning_rate": 4.2059385854548316e-05, "loss": 2.4652, "step": 1046 }, { "epoch": 0.3353082465972778, "grad_norm": 0.51171875, "learning_rate": 4.203893633086027e-05, "loss": 2.3613, "step": 1047 }, { "epoch": 0.33562850280224177, "grad_norm": 0.67578125, "learning_rate": 4.201846549446094e-05, "loss": 2.4627, "step": 1048 }, { "epoch": 0.3359487590072058, "grad_norm": 0.408203125, "learning_rate": 4.1997973370955734e-05, "loss": 2.232, "step": 1049 }, { "epoch": 0.33626901521216973, "grad_norm": 0.421875, "learning_rate": 4.19774599859767e-05, "loss": 2.6032, "step": 1050 }, { "epoch": 0.3365892714171337, "grad_norm": 0.455078125, "learning_rate": 4.1956925365182466e-05, "loss": 2.4148, "step": 1051 }, { "epoch": 0.3369095276220977, "grad_norm": 0.36328125, "learning_rate": 4.193636953425823e-05, "loss": 1.6132, "step": 1052 }, { "epoch": 0.33722978382706165, "grad_norm": 0.48046875, "learning_rate": 4.191579251891572e-05, "loss": 2.2119, "step": 1053 }, { "epoch": 0.3375500400320256, "grad_norm": 0.71484375, "learning_rate": 4.1895194344893177e-05, "loss": 2.0276, "step": 1054 }, { "epoch": 0.3378702962369896, "grad_norm": 0.416015625, "learning_rate": 4.187457503795527e-05, "loss": 2.0666, "step": 1055 }, { "epoch": 0.33819055244195356, "grad_norm": 0.490234375, "learning_rate": 4.185393462389313e-05, "loss": 2.1673, "step": 1056 }, { "epoch": 0.3385108086469175, "grad_norm": 0.369140625, "learning_rate": 4.1833273128524284e-05, "loss": 2.0761, "step": 1057 }, { "epoch": 0.3388310648518815, "grad_norm": 0.40625, "learning_rate": 4.181259057769264e-05, "loss": 2.4819, "step": 1058 }, { "epoch": 0.3391513210568455, "grad_norm": 0.50390625, "learning_rate": 4.1791886997268414e-05, "loss": 2.1933, "step": 1059 }, { "epoch": 0.33947157726180943, "grad_norm": 0.419921875, "learning_rate": 4.177116241314815e-05, "loss": 2.2095, "step": 1060 }, { "epoch": 0.33979183346677344, "grad_norm": 0.58984375, "learning_rate": 4.175041685125465e-05, "loss": 2.0342, "step": 1061 }, { "epoch": 0.3401120896717374, "grad_norm": 0.55078125, "learning_rate": 4.172965033753697e-05, "loss": 2.3994, "step": 1062 }, { "epoch": 0.34043234587670135, "grad_norm": 0.310546875, "learning_rate": 4.170886289797036e-05, "loss": 2.0665, "step": 1063 }, { "epoch": 0.34075260208166536, "grad_norm": 0.51953125, "learning_rate": 4.1688054558556245e-05, "loss": 1.9771, "step": 1064 }, { "epoch": 0.3410728582866293, "grad_norm": 0.5625, "learning_rate": 4.16672253453222e-05, "loss": 2.5785, "step": 1065 }, { "epoch": 0.34139311449159326, "grad_norm": 0.53515625, "learning_rate": 4.1646375284321916e-05, "loss": 2.0384, "step": 1066 }, { "epoch": 0.3417133706965573, "grad_norm": 0.443359375, "learning_rate": 4.1625504401635126e-05, "loss": 2.1372, "step": 1067 }, { "epoch": 0.3420336269015212, "grad_norm": 0.416015625, "learning_rate": 4.1604612723367666e-05, "loss": 1.9049, "step": 1068 }, { "epoch": 0.3423538831064852, "grad_norm": 0.33203125, "learning_rate": 4.1583700275651314e-05, "loss": 2.012, "step": 1069 }, { "epoch": 0.34267413931144913, "grad_norm": 0.609375, "learning_rate": 4.1562767084643885e-05, "loss": 2.1901, "step": 1070 }, { "epoch": 0.34299439551641314, "grad_norm": 0.59765625, "learning_rate": 4.154181317652911e-05, "loss": 2.4813, "step": 1071 }, { "epoch": 0.3433146517213771, "grad_norm": 0.5, "learning_rate": 4.152083857751665e-05, "loss": 2.4138, "step": 1072 }, { "epoch": 0.34363490792634105, "grad_norm": 0.470703125, "learning_rate": 4.149984331384203e-05, "loss": 2.4764, "step": 1073 }, { "epoch": 0.34395516413130506, "grad_norm": 0.462890625, "learning_rate": 4.147882741176662e-05, "loss": 2.5071, "step": 1074 }, { "epoch": 0.344275420336269, "grad_norm": 0.431640625, "learning_rate": 4.145779089757764e-05, "loss": 2.285, "step": 1075 }, { "epoch": 0.34459567654123296, "grad_norm": 0.3984375, "learning_rate": 4.1436733797588056e-05, "loss": 2.2156, "step": 1076 }, { "epoch": 0.344915932746197, "grad_norm": 3.4375, "learning_rate": 4.141565613813658e-05, "loss": 2.0952, "step": 1077 }, { "epoch": 0.3452361889511609, "grad_norm": 0.515625, "learning_rate": 4.139455794558768e-05, "loss": 2.3122, "step": 1078 }, { "epoch": 0.3455564451561249, "grad_norm": 0.361328125, "learning_rate": 4.137343924633147e-05, "loss": 1.9691, "step": 1079 }, { "epoch": 0.3458767013610889, "grad_norm": 0.345703125, "learning_rate": 4.135230006678373e-05, "loss": 2.2278, "step": 1080 }, { "epoch": 0.34619695756605284, "grad_norm": 0.470703125, "learning_rate": 4.133114043338585e-05, "loss": 1.9843, "step": 1081 }, { "epoch": 0.3465172137710168, "grad_norm": 0.5546875, "learning_rate": 4.130996037260482e-05, "loss": 1.5984, "step": 1082 }, { "epoch": 0.3468374699759808, "grad_norm": 0.302734375, "learning_rate": 4.128875991093315e-05, "loss": 2.2697, "step": 1083 }, { "epoch": 0.34715772618094476, "grad_norm": 0.5234375, "learning_rate": 4.12675390748889e-05, "loss": 2.314, "step": 1084 }, { "epoch": 0.3474779823859087, "grad_norm": 0.5078125, "learning_rate": 4.12462978910156e-05, "loss": 2.4867, "step": 1085 }, { "epoch": 0.3477982385908727, "grad_norm": 0.48828125, "learning_rate": 4.1225036385882235e-05, "loss": 1.9026, "step": 1086 }, { "epoch": 0.3481184947958367, "grad_norm": 0.435546875, "learning_rate": 4.12037545860832e-05, "loss": 2.6461, "step": 1087 }, { "epoch": 0.34843875100080063, "grad_norm": 0.6875, "learning_rate": 4.118245251823829e-05, "loss": 2.2835, "step": 1088 }, { "epoch": 0.34875900720576464, "grad_norm": 0.392578125, "learning_rate": 4.116113020899264e-05, "loss": 2.3112, "step": 1089 }, { "epoch": 0.3490792634107286, "grad_norm": 0.400390625, "learning_rate": 4.1139787685016707e-05, "loss": 2.0939, "step": 1090 }, { "epoch": 0.34939951961569254, "grad_norm": 0.52734375, "learning_rate": 4.1118424973006237e-05, "loss": 1.9811, "step": 1091 }, { "epoch": 0.3497197758206565, "grad_norm": 0.458984375, "learning_rate": 4.1097042099682216e-05, "loss": 2.5208, "step": 1092 }, { "epoch": 0.3500400320256205, "grad_norm": 0.390625, "learning_rate": 4.107563909179087e-05, "loss": 1.6477, "step": 1093 }, { "epoch": 0.35036028823058446, "grad_norm": 0.55859375, "learning_rate": 4.105421597610358e-05, "loss": 2.1296, "step": 1094 }, { "epoch": 0.3506805444355484, "grad_norm": 0.419921875, "learning_rate": 4.1032772779416904e-05, "loss": 2.5322, "step": 1095 }, { "epoch": 0.3510008006405124, "grad_norm": 0.443359375, "learning_rate": 4.10113095285525e-05, "loss": 1.8264, "step": 1096 }, { "epoch": 0.3513210568454764, "grad_norm": 0.53515625, "learning_rate": 4.098982625035713e-05, "loss": 2.2612, "step": 1097 }, { "epoch": 0.35164131305044033, "grad_norm": 0.439453125, "learning_rate": 4.096832297170259e-05, "loss": 1.9801, "step": 1098 }, { "epoch": 0.35196156925540434, "grad_norm": 0.5546875, "learning_rate": 4.094679971948569e-05, "loss": 2.2899, "step": 1099 }, { "epoch": 0.3522818254603683, "grad_norm": 0.50390625, "learning_rate": 4.0925256520628244e-05, "loss": 2.1428, "step": 1100 }, { "epoch": 0.35260208166533225, "grad_norm": 0.40625, "learning_rate": 4.090369340207699e-05, "loss": 1.8895, "step": 1101 }, { "epoch": 0.35292233787029625, "grad_norm": 0.44140625, "learning_rate": 4.088211039080361e-05, "loss": 2.3364, "step": 1102 }, { "epoch": 0.3532425940752602, "grad_norm": 0.5546875, "learning_rate": 4.086050751380464e-05, "loss": 1.9986, "step": 1103 }, { "epoch": 0.35356285028022416, "grad_norm": 0.384765625, "learning_rate": 4.083888479810149e-05, "loss": 2.2253, "step": 1104 }, { "epoch": 0.35388310648518817, "grad_norm": 0.384765625, "learning_rate": 4.081724227074036e-05, "loss": 2.1126, "step": 1105 }, { "epoch": 0.3542033626901521, "grad_norm": 1.5078125, "learning_rate": 4.079557995879225e-05, "loss": 2.427, "step": 1106 }, { "epoch": 0.3545236188951161, "grad_norm": 0.5859375, "learning_rate": 4.0773897889352913e-05, "loss": 2.4259, "step": 1107 }, { "epoch": 0.3548438751000801, "grad_norm": 0.3984375, "learning_rate": 4.075219608954278e-05, "loss": 2.0609, "step": 1108 }, { "epoch": 0.35516413130504404, "grad_norm": 0.86328125, "learning_rate": 4.073047458650699e-05, "loss": 2.5113, "step": 1109 }, { "epoch": 0.355484387510008, "grad_norm": 0.6796875, "learning_rate": 4.070873340741534e-05, "loss": 2.1542, "step": 1110 }, { "epoch": 0.355804643714972, "grad_norm": 0.373046875, "learning_rate": 4.06869725794622e-05, "loss": 2.0773, "step": 1111 }, { "epoch": 0.35612489991993596, "grad_norm": 0.44140625, "learning_rate": 4.066519212986654e-05, "loss": 1.5918, "step": 1112 }, { "epoch": 0.3564451561248999, "grad_norm": 0.337890625, "learning_rate": 4.064339208587187e-05, "loss": 1.8691, "step": 1113 }, { "epoch": 0.3567654123298639, "grad_norm": 0.56640625, "learning_rate": 4.0621572474746227e-05, "loss": 2.0796, "step": 1114 }, { "epoch": 0.35708566853482787, "grad_norm": 0.439453125, "learning_rate": 4.059973332378207e-05, "loss": 2.437, "step": 1115 }, { "epoch": 0.3574059247397918, "grad_norm": 0.494140625, "learning_rate": 4.057787466029637e-05, "loss": 2.6097, "step": 1116 }, { "epoch": 0.3577261809447558, "grad_norm": 0.5234375, "learning_rate": 4.055599651163044e-05, "loss": 2.567, "step": 1117 }, { "epoch": 0.3580464371497198, "grad_norm": 0.47265625, "learning_rate": 4.053409890515002e-05, "loss": 2.0911, "step": 1118 }, { "epoch": 0.35836669335468374, "grad_norm": 0.51953125, "learning_rate": 4.051218186824515e-05, "loss": 2.3121, "step": 1119 }, { "epoch": 0.3586869495596477, "grad_norm": 0.55859375, "learning_rate": 4.049024542833018e-05, "loss": 2.4189, "step": 1120 }, { "epoch": 0.3590072057646117, "grad_norm": 0.5, "learning_rate": 4.046828961284375e-05, "loss": 2.6126, "step": 1121 }, { "epoch": 0.35932746196957566, "grad_norm": 0.4609375, "learning_rate": 4.0446314449248716e-05, "loss": 2.208, "step": 1122 }, { "epoch": 0.3596477181745396, "grad_norm": 0.53125, "learning_rate": 4.042431996503213e-05, "loss": 1.7206, "step": 1123 }, { "epoch": 0.3599679743795036, "grad_norm": 0.427734375, "learning_rate": 4.040230618770524e-05, "loss": 2.3154, "step": 1124 }, { "epoch": 0.3602882305844676, "grad_norm": 0.53515625, "learning_rate": 4.03802731448034e-05, "loss": 2.3005, "step": 1125 }, { "epoch": 0.3606084867894315, "grad_norm": 1.3046875, "learning_rate": 4.0358220863886074e-05, "loss": 2.2069, "step": 1126 }, { "epoch": 0.36092874299439553, "grad_norm": 0.376953125, "learning_rate": 4.033614937253678e-05, "loss": 2.4196, "step": 1127 }, { "epoch": 0.3612489991993595, "grad_norm": 0.427734375, "learning_rate": 4.031405869836307e-05, "loss": 2.4691, "step": 1128 }, { "epoch": 0.36156925540432344, "grad_norm": 0.494140625, "learning_rate": 4.0291948868996496e-05, "loss": 2.7772, "step": 1129 }, { "epoch": 0.36188951160928745, "grad_norm": 0.52734375, "learning_rate": 4.026981991209256e-05, "loss": 2.2456, "step": 1130 }, { "epoch": 0.3622097678142514, "grad_norm": 0.5546875, "learning_rate": 4.024767185533069e-05, "loss": 2.1406, "step": 1131 }, { "epoch": 0.36253002401921536, "grad_norm": 0.5078125, "learning_rate": 4.022550472641422e-05, "loss": 2.3268, "step": 1132 }, { "epoch": 0.36285028022417937, "grad_norm": 0.53125, "learning_rate": 4.020331855307031e-05, "loss": 2.2335, "step": 1133 }, { "epoch": 0.3631705364291433, "grad_norm": 0.43359375, "learning_rate": 4.018111336304997e-05, "loss": 2.384, "step": 1134 }, { "epoch": 0.3634907926341073, "grad_norm": 0.4296875, "learning_rate": 4.015888918412799e-05, "loss": 2.0399, "step": 1135 }, { "epoch": 0.3638110488390713, "grad_norm": 0.458984375, "learning_rate": 4.01366460441029e-05, "loss": 2.2721, "step": 1136 }, { "epoch": 0.36413130504403524, "grad_norm": 0.38671875, "learning_rate": 4.011438397079696e-05, "loss": 1.8505, "step": 1137 }, { "epoch": 0.3644515612489992, "grad_norm": 0.39453125, "learning_rate": 4.0092102992056095e-05, "loss": 1.8735, "step": 1138 }, { "epoch": 0.36477181745396314, "grad_norm": 0.44140625, "learning_rate": 4.006980313574991e-05, "loss": 2.2612, "step": 1139 }, { "epoch": 0.36509207365892715, "grad_norm": 0.53515625, "learning_rate": 4.004748442977158e-05, "loss": 2.2203, "step": 1140 }, { "epoch": 0.3654123298638911, "grad_norm": 0.609375, "learning_rate": 4.002514690203788e-05, "loss": 2.1066, "step": 1141 }, { "epoch": 0.36573258606885506, "grad_norm": 0.462890625, "learning_rate": 4.000279058048915e-05, "loss": 2.2341, "step": 1142 }, { "epoch": 0.36605284227381907, "grad_norm": 0.72265625, "learning_rate": 3.998041549308919e-05, "loss": 2.4673, "step": 1143 }, { "epoch": 0.366373098478783, "grad_norm": 0.427734375, "learning_rate": 3.995802166782531e-05, "loss": 2.1514, "step": 1144 }, { "epoch": 0.366693354683747, "grad_norm": 0.54296875, "learning_rate": 3.993560913270823e-05, "loss": 1.8712, "step": 1145 }, { "epoch": 0.367013610888711, "grad_norm": 0.46484375, "learning_rate": 3.991317791577212e-05, "loss": 2.6296, "step": 1146 }, { "epoch": 0.36733386709367494, "grad_norm": 0.48046875, "learning_rate": 3.989072804507444e-05, "loss": 2.4734, "step": 1147 }, { "epoch": 0.3676541232986389, "grad_norm": 0.5546875, "learning_rate": 3.986825954869607e-05, "loss": 2.5021, "step": 1148 }, { "epoch": 0.3679743795036029, "grad_norm": 0.48046875, "learning_rate": 3.9845772454741124e-05, "loss": 2.0079, "step": 1149 }, { "epoch": 0.36829463570856685, "grad_norm": 0.65625, "learning_rate": 3.982326679133701e-05, "loss": 2.6182, "step": 1150 }, { "epoch": 0.3686148919135308, "grad_norm": 0.43359375, "learning_rate": 3.9800742586634346e-05, "loss": 1.8996, "step": 1151 }, { "epoch": 0.3689351481184948, "grad_norm": 0.65625, "learning_rate": 3.977819986880694e-05, "loss": 2.4005, "step": 1152 }, { "epoch": 0.36925540432345877, "grad_norm": 0.95703125, "learning_rate": 3.9755638666051795e-05, "loss": 2.2208, "step": 1153 }, { "epoch": 0.3695756605284227, "grad_norm": 1.2421875, "learning_rate": 3.973305900658897e-05, "loss": 1.9195, "step": 1154 }, { "epoch": 0.36989591673338673, "grad_norm": 0.462890625, "learning_rate": 3.971046091866166e-05, "loss": 1.4788, "step": 1155 }, { "epoch": 0.3702161729383507, "grad_norm": 0.455078125, "learning_rate": 3.968784443053612e-05, "loss": 2.1755, "step": 1156 }, { "epoch": 0.37053642914331464, "grad_norm": 0.53515625, "learning_rate": 3.966520957050156e-05, "loss": 2.3551, "step": 1157 }, { "epoch": 0.37085668534827865, "grad_norm": 0.53125, "learning_rate": 3.964255636687023e-05, "loss": 2.3383, "step": 1158 }, { "epoch": 0.3711769415532426, "grad_norm": 0.451171875, "learning_rate": 3.961988484797729e-05, "loss": 1.8462, "step": 1159 }, { "epoch": 0.37149719775820655, "grad_norm": 0.52734375, "learning_rate": 3.959719504218083e-05, "loss": 2.2438, "step": 1160 }, { "epoch": 0.37181745396317056, "grad_norm": 0.5078125, "learning_rate": 3.95744869778618e-05, "loss": 2.2278, "step": 1161 }, { "epoch": 0.3721377101681345, "grad_norm": 0.455078125, "learning_rate": 3.9551760683424e-05, "loss": 2.2491, "step": 1162 }, { "epoch": 0.37245796637309847, "grad_norm": 0.80078125, "learning_rate": 3.952901618729402e-05, "loss": 1.8948, "step": 1163 }, { "epoch": 0.3727782225780624, "grad_norm": 0.55078125, "learning_rate": 3.950625351792122e-05, "loss": 1.8903, "step": 1164 }, { "epoch": 0.37309847878302643, "grad_norm": 0.359375, "learning_rate": 3.948347270377769e-05, "loss": 1.2149, "step": 1165 }, { "epoch": 0.3734187349879904, "grad_norm": 0.38671875, "learning_rate": 3.946067377335824e-05, "loss": 2.2527, "step": 1166 }, { "epoch": 0.37373899119295434, "grad_norm": 0.62890625, "learning_rate": 3.94378567551803e-05, "loss": 1.5419, "step": 1167 }, { "epoch": 0.37405924739791835, "grad_norm": 0.453125, "learning_rate": 3.941502167778395e-05, "loss": 1.9236, "step": 1168 }, { "epoch": 0.3743795036028823, "grad_norm": 0.443359375, "learning_rate": 3.9392168569731854e-05, "loss": 2.3081, "step": 1169 }, { "epoch": 0.37469975980784626, "grad_norm": 0.515625, "learning_rate": 3.9369297459609247e-05, "loss": 2.1903, "step": 1170 }, { "epoch": 0.37502001601281026, "grad_norm": 0.859375, "learning_rate": 3.934640837602383e-05, "loss": 2.5142, "step": 1171 }, { "epoch": 0.3753402722177742, "grad_norm": 0.58203125, "learning_rate": 3.932350134760585e-05, "loss": 2.1751, "step": 1172 }, { "epoch": 0.37566052842273817, "grad_norm": 0.6484375, "learning_rate": 3.930057640300795e-05, "loss": 2.3225, "step": 1173 }, { "epoch": 0.3759807846277022, "grad_norm": 0.5078125, "learning_rate": 3.92776335709052e-05, "loss": 2.1874, "step": 1174 }, { "epoch": 0.37630104083266613, "grad_norm": 0.51953125, "learning_rate": 3.925467287999507e-05, "loss": 2.1818, "step": 1175 }, { "epoch": 0.3766212970376301, "grad_norm": 0.5234375, "learning_rate": 3.9231694358997326e-05, "loss": 2.558, "step": 1176 }, { "epoch": 0.3769415532425941, "grad_norm": 0.38671875, "learning_rate": 3.920869803665405e-05, "loss": 2.0283, "step": 1177 }, { "epoch": 0.37726180944755805, "grad_norm": 0.5, "learning_rate": 3.91856839417296e-05, "loss": 2.6307, "step": 1178 }, { "epoch": 0.377582065652522, "grad_norm": 0.578125, "learning_rate": 3.916265210301057e-05, "loss": 1.9584, "step": 1179 }, { "epoch": 0.377902321857486, "grad_norm": 0.5, "learning_rate": 3.913960254930572e-05, "loss": 1.7026, "step": 1180 }, { "epoch": 0.37822257806244997, "grad_norm": 0.50390625, "learning_rate": 3.9116535309446e-05, "loss": 2.0822, "step": 1181 }, { "epoch": 0.3785428342674139, "grad_norm": 0.60546875, "learning_rate": 3.909345041228447e-05, "loss": 2.5519, "step": 1182 }, { "epoch": 0.37886309047237793, "grad_norm": 0.447265625, "learning_rate": 3.9070347886696254e-05, "loss": 2.2184, "step": 1183 }, { "epoch": 0.3791833466773419, "grad_norm": 0.921875, "learning_rate": 3.9047227761578566e-05, "loss": 1.765, "step": 1184 }, { "epoch": 0.37950360288230583, "grad_norm": 0.65625, "learning_rate": 3.902409006585061e-05, "loss": 2.3076, "step": 1185 }, { "epoch": 0.3798238590872698, "grad_norm": 0.490234375, "learning_rate": 3.900093482845357e-05, "loss": 2.0427, "step": 1186 }, { "epoch": 0.3801441152922338, "grad_norm": 1.640625, "learning_rate": 3.8977762078350576e-05, "loss": 1.6712, "step": 1187 }, { "epoch": 0.38046437149719775, "grad_norm": 3.296875, "learning_rate": 3.895457184452665e-05, "loss": 2.4834, "step": 1188 }, { "epoch": 0.3807846277021617, "grad_norm": 0.349609375, "learning_rate": 3.8931364155988726e-05, "loss": 2.2172, "step": 1189 }, { "epoch": 0.3811048839071257, "grad_norm": 0.58984375, "learning_rate": 3.89081390417655e-05, "loss": 2.2178, "step": 1190 }, { "epoch": 0.38142514011208967, "grad_norm": 0.6171875, "learning_rate": 3.888489653090752e-05, "loss": 2.2411, "step": 1191 }, { "epoch": 0.3817453963170536, "grad_norm": 0.400390625, "learning_rate": 3.886163665248709e-05, "loss": 2.7614, "step": 1192 }, { "epoch": 0.38206565252201763, "grad_norm": 0.56640625, "learning_rate": 3.8838359435598205e-05, "loss": 2.2614, "step": 1193 }, { "epoch": 0.3823859087269816, "grad_norm": 0.51171875, "learning_rate": 3.881506490935657e-05, "loss": 2.0497, "step": 1194 }, { "epoch": 0.38270616493194554, "grad_norm": 0.61328125, "learning_rate": 3.879175310289955e-05, "loss": 2.3673, "step": 1195 }, { "epoch": 0.38302642113690955, "grad_norm": 4.6875, "learning_rate": 3.876842404538611e-05, "loss": 2.7132, "step": 1196 }, { "epoch": 0.3833466773418735, "grad_norm": 0.498046875, "learning_rate": 3.874507776599678e-05, "loss": 2.1813, "step": 1197 }, { "epoch": 0.38366693354683745, "grad_norm": 1.3359375, "learning_rate": 3.872171429393368e-05, "loss": 2.3327, "step": 1198 }, { "epoch": 0.38398718975180146, "grad_norm": 0.640625, "learning_rate": 3.8698333658420366e-05, "loss": 1.5882, "step": 1199 }, { "epoch": 0.3843074459567654, "grad_norm": 0.53515625, "learning_rate": 3.867493588870193e-05, "loss": 2.0014, "step": 1200 }, { "epoch": 0.38462770216172937, "grad_norm": 0.494140625, "learning_rate": 3.865152101404485e-05, "loss": 2.2738, "step": 1201 }, { "epoch": 0.3849479583666934, "grad_norm": 0.44921875, "learning_rate": 3.862808906373701e-05, "loss": 2.3236, "step": 1202 }, { "epoch": 0.38526821457165733, "grad_norm": 0.43359375, "learning_rate": 3.860464006708766e-05, "loss": 2.6712, "step": 1203 }, { "epoch": 0.3855884707766213, "grad_norm": 0.470703125, "learning_rate": 3.8581174053427374e-05, "loss": 1.9265, "step": 1204 }, { "epoch": 0.3859087269815853, "grad_norm": 0.46484375, "learning_rate": 3.855769105210801e-05, "loss": 2.5721, "step": 1205 }, { "epoch": 0.38622898318654925, "grad_norm": 0.462890625, "learning_rate": 3.8534191092502655e-05, "loss": 2.0682, "step": 1206 }, { "epoch": 0.3865492393915132, "grad_norm": 0.671875, "learning_rate": 3.851067420400564e-05, "loss": 2.3802, "step": 1207 }, { "epoch": 0.3868694955964772, "grad_norm": 0.5234375, "learning_rate": 3.848714041603245e-05, "loss": 2.3584, "step": 1208 }, { "epoch": 0.38718975180144116, "grad_norm": 0.6328125, "learning_rate": 3.846358975801971e-05, "loss": 2.3271, "step": 1209 }, { "epoch": 0.3875100080064051, "grad_norm": 0.60546875, "learning_rate": 3.844002225942514e-05, "loss": 2.2613, "step": 1210 }, { "epoch": 0.38783026421136907, "grad_norm": 0.435546875, "learning_rate": 3.8416437949727555e-05, "loss": 2.1979, "step": 1211 }, { "epoch": 0.3881505204163331, "grad_norm": 0.578125, "learning_rate": 3.839283685842676e-05, "loss": 2.7155, "step": 1212 }, { "epoch": 0.38847077662129703, "grad_norm": 0.4296875, "learning_rate": 3.836921901504358e-05, "loss": 1.9949, "step": 1213 }, { "epoch": 0.388791032826261, "grad_norm": 0.53125, "learning_rate": 3.8345584449119776e-05, "loss": 2.1843, "step": 1214 }, { "epoch": 0.389111289031225, "grad_norm": 0.5390625, "learning_rate": 3.832193319021802e-05, "loss": 2.2712, "step": 1215 }, { "epoch": 0.38943154523618895, "grad_norm": 0.609375, "learning_rate": 3.8298265267921884e-05, "loss": 2.1854, "step": 1216 }, { "epoch": 0.3897518014411529, "grad_norm": 1.0625, "learning_rate": 3.827458071183578e-05, "loss": 2.5162, "step": 1217 }, { "epoch": 0.3900720576461169, "grad_norm": 0.369140625, "learning_rate": 3.825087955158492e-05, "loss": 2.4592, "step": 1218 }, { "epoch": 0.39039231385108086, "grad_norm": 0.71875, "learning_rate": 3.822716181681528e-05, "loss": 1.9011, "step": 1219 }, { "epoch": 0.3907125700560448, "grad_norm": 0.416015625, "learning_rate": 3.820342753719357e-05, "loss": 2.2409, "step": 1220 }, { "epoch": 0.3910328262610088, "grad_norm": 0.4375, "learning_rate": 3.817967674240721e-05, "loss": 1.9098, "step": 1221 }, { "epoch": 0.3913530824659728, "grad_norm": 0.484375, "learning_rate": 3.815590946216425e-05, "loss": 2.0723, "step": 1222 }, { "epoch": 0.39167333867093673, "grad_norm": 0.373046875, "learning_rate": 3.813212572619338e-05, "loss": 2.3035, "step": 1223 }, { "epoch": 0.39199359487590074, "grad_norm": 0.439453125, "learning_rate": 3.810832556424388e-05, "loss": 1.9079, "step": 1224 }, { "epoch": 0.3923138510808647, "grad_norm": 1.546875, "learning_rate": 3.808450900608555e-05, "loss": 1.7984, "step": 1225 }, { "epoch": 0.39263410728582865, "grad_norm": 0.486328125, "learning_rate": 3.806067608150872e-05, "loss": 2.1234, "step": 1226 }, { "epoch": 0.39295436349079266, "grad_norm": 0.5, "learning_rate": 3.803682682032419e-05, "loss": 2.2432, "step": 1227 }, { "epoch": 0.3932746196957566, "grad_norm": 1.0703125, "learning_rate": 3.8012961252363184e-05, "loss": 1.9167, "step": 1228 }, { "epoch": 0.39359487590072056, "grad_norm": 0.42578125, "learning_rate": 3.798907940747732e-05, "loss": 2.6162, "step": 1229 }, { "epoch": 0.3939151321056846, "grad_norm": 0.45703125, "learning_rate": 3.79651813155386e-05, "loss": 2.239, "step": 1230 }, { "epoch": 0.3942353883106485, "grad_norm": 0.65625, "learning_rate": 3.794126700643933e-05, "loss": 1.7196, "step": 1231 }, { "epoch": 0.3945556445156125, "grad_norm": 0.408203125, "learning_rate": 3.791733651009209e-05, "loss": 2.1263, "step": 1232 }, { "epoch": 0.39487590072057643, "grad_norm": 0.427734375, "learning_rate": 3.789338985642973e-05, "loss": 2.0495, "step": 1233 }, { "epoch": 0.39519615692554044, "grad_norm": 0.45703125, "learning_rate": 3.786942707540529e-05, "loss": 2.2755, "step": 1234 }, { "epoch": 0.3955164131305044, "grad_norm": 0.494140625, "learning_rate": 3.784544819699202e-05, "loss": 2.5356, "step": 1235 }, { "epoch": 0.39583666933546835, "grad_norm": 0.39453125, "learning_rate": 3.7821453251183245e-05, "loss": 2.3301, "step": 1236 }, { "epoch": 0.39615692554043236, "grad_norm": 0.48046875, "learning_rate": 3.7797442267992436e-05, "loss": 2.3236, "step": 1237 }, { "epoch": 0.3964771817453963, "grad_norm": 0.5625, "learning_rate": 3.7773415277453103e-05, "loss": 2.3126, "step": 1238 }, { "epoch": 0.39679743795036027, "grad_norm": 0.349609375, "learning_rate": 3.77493723096188e-05, "loss": 2.2927, "step": 1239 }, { "epoch": 0.3971176941553243, "grad_norm": 0.5078125, "learning_rate": 3.772531339456302e-05, "loss": 1.9049, "step": 1240 }, { "epoch": 0.39743795036028823, "grad_norm": 0.6796875, "learning_rate": 3.770123856237925e-05, "loss": 2.5726, "step": 1241 }, { "epoch": 0.3977582065652522, "grad_norm": 0.490234375, "learning_rate": 3.767714784318088e-05, "loss": 2.2026, "step": 1242 }, { "epoch": 0.3980784627702162, "grad_norm": 0.62890625, "learning_rate": 3.765304126710113e-05, "loss": 2.08, "step": 1243 }, { "epoch": 0.39839871897518014, "grad_norm": 0.453125, "learning_rate": 3.762891886429312e-05, "loss": 1.8371, "step": 1244 }, { "epoch": 0.3987189751801441, "grad_norm": 0.69140625, "learning_rate": 3.760478066492971e-05, "loss": 2.2806, "step": 1245 }, { "epoch": 0.3990392313851081, "grad_norm": 0.54296875, "learning_rate": 3.758062669920354e-05, "loss": 2.312, "step": 1246 }, { "epoch": 0.39935948759007206, "grad_norm": 0.33203125, "learning_rate": 3.7556456997326975e-05, "loss": 2.1427, "step": 1247 }, { "epoch": 0.399679743795036, "grad_norm": 0.53515625, "learning_rate": 3.7532271589532065e-05, "loss": 2.2062, "step": 1248 }, { "epoch": 0.4, "grad_norm": 0.470703125, "learning_rate": 3.75080705060705e-05, "loss": 2.4721, "step": 1249 }, { "epoch": 0.400320256204964, "grad_norm": 0.49609375, "learning_rate": 3.748385377721356e-05, "loss": 2.5967, "step": 1250 }, { "epoch": 0.40064051240992793, "grad_norm": 0.5859375, "learning_rate": 3.7459621433252135e-05, "loss": 2.1675, "step": 1251 }, { "epoch": 0.40096076861489194, "grad_norm": 0.42578125, "learning_rate": 3.74353735044966e-05, "loss": 1.9072, "step": 1252 }, { "epoch": 0.4012810248198559, "grad_norm": 0.388671875, "learning_rate": 3.741111002127688e-05, "loss": 1.6543, "step": 1253 }, { "epoch": 0.40160128102481985, "grad_norm": 0.75390625, "learning_rate": 3.738683101394228e-05, "loss": 1.8961, "step": 1254 }, { "epoch": 0.40192153722978385, "grad_norm": 0.6328125, "learning_rate": 3.73625365128616e-05, "loss": 2.2844, "step": 1255 }, { "epoch": 0.4022417934347478, "grad_norm": 1.0703125, "learning_rate": 3.733822654842298e-05, "loss": 2.2592, "step": 1256 }, { "epoch": 0.40256204963971176, "grad_norm": 0.57421875, "learning_rate": 3.73139011510339e-05, "loss": 2.8168, "step": 1257 }, { "epoch": 0.4028823058446757, "grad_norm": 0.71875, "learning_rate": 3.728956035112115e-05, "loss": 1.9843, "step": 1258 }, { "epoch": 0.4032025620496397, "grad_norm": 0.56640625, "learning_rate": 3.72652041791308e-05, "loss": 2.2555, "step": 1259 }, { "epoch": 0.4035228182546037, "grad_norm": 0.416015625, "learning_rate": 3.7240832665528124e-05, "loss": 2.3078, "step": 1260 }, { "epoch": 0.40384307445956763, "grad_norm": 0.53125, "learning_rate": 3.721644584079761e-05, "loss": 1.7157, "step": 1261 }, { "epoch": 0.40416333066453164, "grad_norm": 0.7421875, "learning_rate": 3.7192043735442865e-05, "loss": 2.4214, "step": 1262 }, { "epoch": 0.4044835868694956, "grad_norm": 0.5859375, "learning_rate": 3.7167626379986655e-05, "loss": 2.5919, "step": 1263 }, { "epoch": 0.40480384307445955, "grad_norm": 0.33984375, "learning_rate": 3.714319380497077e-05, "loss": 2.6869, "step": 1264 }, { "epoch": 0.40512409927942356, "grad_norm": 0.77734375, "learning_rate": 3.7118746040956076e-05, "loss": 2.0477, "step": 1265 }, { "epoch": 0.4054443554843875, "grad_norm": 0.5078125, "learning_rate": 3.709428311852242e-05, "loss": 1.8267, "step": 1266 }, { "epoch": 0.40576461168935146, "grad_norm": 0.52734375, "learning_rate": 3.706980506826863e-05, "loss": 2.1828, "step": 1267 }, { "epoch": 0.40608486789431547, "grad_norm": 0.515625, "learning_rate": 3.704531192081241e-05, "loss": 2.2251, "step": 1268 }, { "epoch": 0.4064051240992794, "grad_norm": 0.369140625, "learning_rate": 3.70208037067904e-05, "loss": 1.7606, "step": 1269 }, { "epoch": 0.4067253803042434, "grad_norm": 0.494140625, "learning_rate": 3.699628045685805e-05, "loss": 2.4729, "step": 1270 }, { "epoch": 0.4070456365092074, "grad_norm": 0.40625, "learning_rate": 3.697174220168965e-05, "loss": 1.5516, "step": 1271 }, { "epoch": 0.40736589271417134, "grad_norm": 0.474609375, "learning_rate": 3.694718897197822e-05, "loss": 2.0469, "step": 1272 }, { "epoch": 0.4076861489191353, "grad_norm": 0.486328125, "learning_rate": 3.692262079843553e-05, "loss": 1.997, "step": 1273 }, { "epoch": 0.4080064051240993, "grad_norm": 0.48828125, "learning_rate": 3.689803771179207e-05, "loss": 2.3182, "step": 1274 }, { "epoch": 0.40832666132906326, "grad_norm": 0.474609375, "learning_rate": 3.687343974279691e-05, "loss": 2.4617, "step": 1275 }, { "epoch": 0.4086469175340272, "grad_norm": 0.609375, "learning_rate": 3.684882692221782e-05, "loss": 2.3608, "step": 1276 }, { "epoch": 0.4089671737389912, "grad_norm": 0.458984375, "learning_rate": 3.682419928084109e-05, "loss": 2.1872, "step": 1277 }, { "epoch": 0.40928742994395517, "grad_norm": 0.5703125, "learning_rate": 3.679955684947158e-05, "loss": 2.0972, "step": 1278 }, { "epoch": 0.4096076861489191, "grad_norm": 0.40625, "learning_rate": 3.6774899658932615e-05, "loss": 2.086, "step": 1279 }, { "epoch": 0.4099279423538831, "grad_norm": 0.4765625, "learning_rate": 3.6750227740066015e-05, "loss": 2.3214, "step": 1280 }, { "epoch": 0.4102481985588471, "grad_norm": 1.6640625, "learning_rate": 3.6725541123732e-05, "loss": 2.2449, "step": 1281 }, { "epoch": 0.41056845476381104, "grad_norm": 3.421875, "learning_rate": 3.6700839840809206e-05, "loss": 1.619, "step": 1282 }, { "epoch": 0.410888710968775, "grad_norm": 0.56640625, "learning_rate": 3.6676123922194574e-05, "loss": 2.1188, "step": 1283 }, { "epoch": 0.411208967173739, "grad_norm": 0.5078125, "learning_rate": 3.6651393398803366e-05, "loss": 2.6604, "step": 1284 }, { "epoch": 0.41152922337870296, "grad_norm": 1.15625, "learning_rate": 3.6626648301569124e-05, "loss": 1.9312, "step": 1285 }, { "epoch": 0.4118494795836669, "grad_norm": 0.474609375, "learning_rate": 3.6601888661443603e-05, "loss": 2.0071, "step": 1286 }, { "epoch": 0.4121697357886309, "grad_norm": 0.66015625, "learning_rate": 3.657711450939676e-05, "loss": 2.1451, "step": 1287 }, { "epoch": 0.4124899919935949, "grad_norm": 0.61328125, "learning_rate": 3.6552325876416704e-05, "loss": 2.6817, "step": 1288 }, { "epoch": 0.4128102481985588, "grad_norm": 0.37890625, "learning_rate": 3.652752279350965e-05, "loss": 2.2703, "step": 1289 }, { "epoch": 0.41313050440352284, "grad_norm": 0.5078125, "learning_rate": 3.65027052916999e-05, "loss": 1.6399, "step": 1290 }, { "epoch": 0.4134507606084868, "grad_norm": 0.478515625, "learning_rate": 3.647787340202975e-05, "loss": 2.139, "step": 1291 }, { "epoch": 0.41377101681345074, "grad_norm": 0.67578125, "learning_rate": 3.6453027155559565e-05, "loss": 2.0271, "step": 1292 }, { "epoch": 0.41409127301841475, "grad_norm": 0.48828125, "learning_rate": 3.6428166583367615e-05, "loss": 2.1767, "step": 1293 }, { "epoch": 0.4144115292233787, "grad_norm": 1.1796875, "learning_rate": 3.640329171655008e-05, "loss": 1.9477, "step": 1294 }, { "epoch": 0.41473178542834266, "grad_norm": 0.49609375, "learning_rate": 3.637840258622106e-05, "loss": 3.0508, "step": 1295 }, { "epoch": 0.41505204163330667, "grad_norm": 0.455078125, "learning_rate": 3.635349922351249e-05, "loss": 2.6344, "step": 1296 }, { "epoch": 0.4153722978382706, "grad_norm": 4.125, "learning_rate": 3.632858165957407e-05, "loss": 2.1484, "step": 1297 }, { "epoch": 0.4156925540432346, "grad_norm": 0.3984375, "learning_rate": 3.63036499255733e-05, "loss": 2.0567, "step": 1298 }, { "epoch": 0.4160128102481986, "grad_norm": 0.890625, "learning_rate": 3.627870405269539e-05, "loss": 2.6683, "step": 1299 }, { "epoch": 0.41633306645316254, "grad_norm": 0.55859375, "learning_rate": 3.625374407214325e-05, "loss": 2.0889, "step": 1300 }, { "epoch": 0.4166533226581265, "grad_norm": 0.5625, "learning_rate": 3.62287700151374e-05, "loss": 2.3849, "step": 1301 }, { "epoch": 0.4169735788630905, "grad_norm": 0.46875, "learning_rate": 3.6203781912916e-05, "loss": 2.5107, "step": 1302 }, { "epoch": 0.41729383506805445, "grad_norm": 0.55078125, "learning_rate": 3.6178779796734785e-05, "loss": 2.0597, "step": 1303 }, { "epoch": 0.4176140912730184, "grad_norm": 0.70703125, "learning_rate": 3.615376369786699e-05, "loss": 2.3948, "step": 1304 }, { "epoch": 0.41793434747798236, "grad_norm": 0.53125, "learning_rate": 3.6128733647603345e-05, "loss": 2.1924, "step": 1305 }, { "epoch": 0.41825460368294637, "grad_norm": 0.640625, "learning_rate": 3.610368967725205e-05, "loss": 2.1514, "step": 1306 }, { "epoch": 0.4185748598879103, "grad_norm": 0.59375, "learning_rate": 3.607863181813871e-05, "loss": 2.1708, "step": 1307 }, { "epoch": 0.4188951160928743, "grad_norm": 0.5234375, "learning_rate": 3.6053560101606285e-05, "loss": 2.4148, "step": 1308 }, { "epoch": 0.4192153722978383, "grad_norm": 0.486328125, "learning_rate": 3.602847455901509e-05, "loss": 2.3323, "step": 1309 }, { "epoch": 0.41953562850280224, "grad_norm": 0.68359375, "learning_rate": 3.600337522174272e-05, "loss": 2.2687, "step": 1310 }, { "epoch": 0.4198558847077662, "grad_norm": 0.45703125, "learning_rate": 3.597826212118404e-05, "loss": 2.5078, "step": 1311 }, { "epoch": 0.4201761409127302, "grad_norm": 0.609375, "learning_rate": 3.5953135288751094e-05, "loss": 2.1704, "step": 1312 }, { "epoch": 0.42049639711769415, "grad_norm": 0.5078125, "learning_rate": 3.5927994755873157e-05, "loss": 2.3606, "step": 1313 }, { "epoch": 0.4208166533226581, "grad_norm": 0.70703125, "learning_rate": 3.5902840553996594e-05, "loss": 1.9514, "step": 1314 }, { "epoch": 0.4211369095276221, "grad_norm": 0.455078125, "learning_rate": 3.58776727145849e-05, "loss": 2.6369, "step": 1315 }, { "epoch": 0.42145716573258607, "grad_norm": 0.421875, "learning_rate": 3.58524912691186e-05, "loss": 2.2219, "step": 1316 }, { "epoch": 0.42177742193755, "grad_norm": 0.4921875, "learning_rate": 3.5827296249095264e-05, "loss": 1.7456, "step": 1317 }, { "epoch": 0.42209767814251403, "grad_norm": 0.5078125, "learning_rate": 3.580208768602941e-05, "loss": 2.2512, "step": 1318 }, { "epoch": 0.422417934347478, "grad_norm": 0.6484375, "learning_rate": 3.577686561145254e-05, "loss": 2.1483, "step": 1319 }, { "epoch": 0.42273819055244194, "grad_norm": 0.3671875, "learning_rate": 3.575163005691302e-05, "loss": 1.9452, "step": 1320 }, { "epoch": 0.42305844675740595, "grad_norm": 0.65625, "learning_rate": 3.572638105397607e-05, "loss": 2.1301, "step": 1321 }, { "epoch": 0.4233787029623699, "grad_norm": 0.431640625, "learning_rate": 3.5701118634223785e-05, "loss": 2.099, "step": 1322 }, { "epoch": 0.42369895916733386, "grad_norm": 0.45703125, "learning_rate": 3.567584282925498e-05, "loss": 2.2, "step": 1323 }, { "epoch": 0.42401921537229786, "grad_norm": 0.466796875, "learning_rate": 3.565055367068525e-05, "loss": 2.4591, "step": 1324 }, { "epoch": 0.4243394715772618, "grad_norm": 0.52734375, "learning_rate": 3.562525119014688e-05, "loss": 2.1612, "step": 1325 }, { "epoch": 0.42465972778222577, "grad_norm": 0.5, "learning_rate": 3.5599935419288824e-05, "loss": 1.9581, "step": 1326 }, { "epoch": 0.4249799839871897, "grad_norm": 0.54296875, "learning_rate": 3.557460638977666e-05, "loss": 2.4488, "step": 1327 }, { "epoch": 0.42530024019215373, "grad_norm": 0.5234375, "learning_rate": 3.554926413329254e-05, "loss": 2.6207, "step": 1328 }, { "epoch": 0.4256204963971177, "grad_norm": 0.494140625, "learning_rate": 3.552390868153516e-05, "loss": 2.0293, "step": 1329 }, { "epoch": 0.42594075260208164, "grad_norm": 0.390625, "learning_rate": 3.549854006621976e-05, "loss": 2.2195, "step": 1330 }, { "epoch": 0.42626100880704565, "grad_norm": 0.99609375, "learning_rate": 3.547315831907798e-05, "loss": 2.9019, "step": 1331 }, { "epoch": 0.4265812650120096, "grad_norm": 0.515625, "learning_rate": 3.544776347185793e-05, "loss": 2.5745, "step": 1332 }, { "epoch": 0.42690152121697356, "grad_norm": 0.54296875, "learning_rate": 3.54223555563241e-05, "loss": 2.6656, "step": 1333 }, { "epoch": 0.42722177742193757, "grad_norm": 0.44921875, "learning_rate": 3.5396934604257326e-05, "loss": 2.2501, "step": 1334 }, { "epoch": 0.4275420336269015, "grad_norm": 0.4609375, "learning_rate": 3.537150064745474e-05, "loss": 2.2139, "step": 1335 }, { "epoch": 0.42786228983186547, "grad_norm": 0.44921875, "learning_rate": 3.534605371772974e-05, "loss": 1.9124, "step": 1336 }, { "epoch": 0.4281825460368295, "grad_norm": 0.56640625, "learning_rate": 3.5320593846911973e-05, "loss": 2.6009, "step": 1337 }, { "epoch": 0.42850280224179343, "grad_norm": 0.41796875, "learning_rate": 3.529512106684724e-05, "loss": 1.5908, "step": 1338 }, { "epoch": 0.4288230584467574, "grad_norm": 0.41015625, "learning_rate": 3.526963540939752e-05, "loss": 2.011, "step": 1339 }, { "epoch": 0.4291433146517214, "grad_norm": 0.54296875, "learning_rate": 3.5244136906440886e-05, "loss": 2.0686, "step": 1340 }, { "epoch": 0.42946357085668535, "grad_norm": 0.9453125, "learning_rate": 3.521862558987148e-05, "loss": 1.9993, "step": 1341 }, { "epoch": 0.4297838270616493, "grad_norm": 0.419921875, "learning_rate": 3.5193101491599453e-05, "loss": 2.6751, "step": 1342 }, { "epoch": 0.4301040832666133, "grad_norm": 0.453125, "learning_rate": 3.5167564643550975e-05, "loss": 2.1176, "step": 1343 }, { "epoch": 0.43042433947157727, "grad_norm": 0.44140625, "learning_rate": 3.5142015077668155e-05, "loss": 2.3972, "step": 1344 }, { "epoch": 0.4307445956765412, "grad_norm": 0.291015625, "learning_rate": 3.5116452825909e-05, "loss": 1.8073, "step": 1345 }, { "epoch": 0.43106485188150523, "grad_norm": 0.62890625, "learning_rate": 3.5090877920247375e-05, "loss": 2.6247, "step": 1346 }, { "epoch": 0.4313851080864692, "grad_norm": 0.41796875, "learning_rate": 3.5065290392673e-05, "loss": 1.8295, "step": 1347 }, { "epoch": 0.43170536429143314, "grad_norm": 0.50390625, "learning_rate": 3.503969027519137e-05, "loss": 2.4511, "step": 1348 }, { "epoch": 0.43202562049639714, "grad_norm": 0.474609375, "learning_rate": 3.501407759982372e-05, "loss": 2.578, "step": 1349 }, { "epoch": 0.4323458767013611, "grad_norm": 0.64453125, "learning_rate": 3.498845239860702e-05, "loss": 2.1768, "step": 1350 }, { "epoch": 0.43266613290632505, "grad_norm": 0.3984375, "learning_rate": 3.496281470359386e-05, "loss": 2.084, "step": 1351 }, { "epoch": 0.432986389111289, "grad_norm": 0.423828125, "learning_rate": 3.493716454685251e-05, "loss": 2.215, "step": 1352 }, { "epoch": 0.433306645316253, "grad_norm": 0.64453125, "learning_rate": 3.491150196046679e-05, "loss": 2.1564, "step": 1353 }, { "epoch": 0.43362690152121697, "grad_norm": 0.60546875, "learning_rate": 3.4885826976536074e-05, "loss": 2.0137, "step": 1354 }, { "epoch": 0.4339471577261809, "grad_norm": 0.486328125, "learning_rate": 3.486013962717526e-05, "loss": 1.9845, "step": 1355 }, { "epoch": 0.43426741393114493, "grad_norm": 0.357421875, "learning_rate": 3.483443994451471e-05, "loss": 1.7616, "step": 1356 }, { "epoch": 0.4345876701361089, "grad_norm": 0.439453125, "learning_rate": 3.480872796070018e-05, "loss": 2.5005, "step": 1357 }, { "epoch": 0.43490792634107284, "grad_norm": 0.46484375, "learning_rate": 3.478300370789286e-05, "loss": 2.4385, "step": 1358 }, { "epoch": 0.43522818254603685, "grad_norm": 0.52734375, "learning_rate": 3.475726721826926e-05, "loss": 2.3023, "step": 1359 }, { "epoch": 0.4355484387510008, "grad_norm": 0.48046875, "learning_rate": 3.473151852402119e-05, "loss": 2.0413, "step": 1360 }, { "epoch": 0.43586869495596475, "grad_norm": 0.62109375, "learning_rate": 3.470575765735574e-05, "loss": 2.4551, "step": 1361 }, { "epoch": 0.43618895116092876, "grad_norm": 0.50390625, "learning_rate": 3.4679984650495225e-05, "loss": 2.5011, "step": 1362 }, { "epoch": 0.4365092073658927, "grad_norm": 0.52734375, "learning_rate": 3.465419953567714e-05, "loss": 1.651, "step": 1363 }, { "epoch": 0.43682946357085667, "grad_norm": 0.54296875, "learning_rate": 3.4628402345154136e-05, "loss": 2.2163, "step": 1364 }, { "epoch": 0.4371497197758207, "grad_norm": 0.42578125, "learning_rate": 3.4602593111193946e-05, "loss": 2.1722, "step": 1365 }, { "epoch": 0.43746997598078463, "grad_norm": 0.455078125, "learning_rate": 3.457677186607938e-05, "loss": 2.5205, "step": 1366 }, { "epoch": 0.4377902321857486, "grad_norm": 0.66796875, "learning_rate": 3.455093864210828e-05, "loss": 2.0087, "step": 1367 }, { "epoch": 0.4381104883907126, "grad_norm": 0.65625, "learning_rate": 3.452509347159346e-05, "loss": 2.0194, "step": 1368 }, { "epoch": 0.43843074459567655, "grad_norm": 0.48046875, "learning_rate": 3.449923638686269e-05, "loss": 2.6472, "step": 1369 }, { "epoch": 0.4387510008006405, "grad_norm": 0.419921875, "learning_rate": 3.447336742025862e-05, "loss": 2.1153, "step": 1370 }, { "epoch": 0.4390712570056045, "grad_norm": 0.5078125, "learning_rate": 3.4447486604138776e-05, "loss": 2.0821, "step": 1371 }, { "epoch": 0.43939151321056846, "grad_norm": 0.51953125, "learning_rate": 3.442159397087551e-05, "loss": 2.3243, "step": 1372 }, { "epoch": 0.4397117694155324, "grad_norm": 0.65625, "learning_rate": 3.4395689552855955e-05, "loss": 2.3805, "step": 1373 }, { "epoch": 0.44003202562049637, "grad_norm": 0.70703125, "learning_rate": 3.436977338248197e-05, "loss": 2.2077, "step": 1374 }, { "epoch": 0.4403522818254604, "grad_norm": 0.5703125, "learning_rate": 3.434384549217013e-05, "loss": 2.3821, "step": 1375 }, { "epoch": 0.44067253803042433, "grad_norm": 0.9453125, "learning_rate": 3.431790591435167e-05, "loss": 2.0472, "step": 1376 }, { "epoch": 0.4409927942353883, "grad_norm": 0.26171875, "learning_rate": 3.4291954681472427e-05, "loss": 1.6111, "step": 1377 }, { "epoch": 0.4413130504403523, "grad_norm": 0.423828125, "learning_rate": 3.426599182599283e-05, "loss": 2.5023, "step": 1378 }, { "epoch": 0.44163330664531625, "grad_norm": 0.42578125, "learning_rate": 3.424001738038784e-05, "loss": 2.4245, "step": 1379 }, { "epoch": 0.4419535628502802, "grad_norm": 0.375, "learning_rate": 3.4214031377146915e-05, "loss": 2.2083, "step": 1380 }, { "epoch": 0.4422738190552442, "grad_norm": 0.80859375, "learning_rate": 3.4188033848773995e-05, "loss": 2.5568, "step": 1381 }, { "epoch": 0.44259407526020816, "grad_norm": 0.60546875, "learning_rate": 3.4162024827787384e-05, "loss": 2.4897, "step": 1382 }, { "epoch": 0.4429143314651721, "grad_norm": 0.48046875, "learning_rate": 3.4136004346719815e-05, "loss": 2.2791, "step": 1383 }, { "epoch": 0.4432345876701361, "grad_norm": 0.4296875, "learning_rate": 3.410997243811832e-05, "loss": 1.5351, "step": 1384 }, { "epoch": 0.4435548438751001, "grad_norm": 0.54296875, "learning_rate": 3.408392913454423e-05, "loss": 2.3915, "step": 1385 }, { "epoch": 0.44387510008006403, "grad_norm": 0.412109375, "learning_rate": 3.4057874468573145e-05, "loss": 1.9231, "step": 1386 }, { "epoch": 0.44419535628502804, "grad_norm": 0.546875, "learning_rate": 3.4031808472794875e-05, "loss": 2.1587, "step": 1387 }, { "epoch": 0.444515612489992, "grad_norm": 0.53515625, "learning_rate": 3.400573117981338e-05, "loss": 1.8927, "step": 1388 }, { "epoch": 0.44483586869495595, "grad_norm": 0.46484375, "learning_rate": 3.3979642622246765e-05, "loss": 1.9999, "step": 1389 }, { "epoch": 0.44515612489991996, "grad_norm": 0.515625, "learning_rate": 3.395354283272724e-05, "loss": 1.7148, "step": 1390 }, { "epoch": 0.4454763811048839, "grad_norm": 0.54296875, "learning_rate": 3.3927431843901034e-05, "loss": 2.6938, "step": 1391 }, { "epoch": 0.44579663730984787, "grad_norm": 0.58203125, "learning_rate": 3.39013096884284e-05, "loss": 2.2425, "step": 1392 }, { "epoch": 0.4461168935148119, "grad_norm": 0.431640625, "learning_rate": 3.3875176398983565e-05, "loss": 2.3499, "step": 1393 }, { "epoch": 0.4464371497197758, "grad_norm": 0.455078125, "learning_rate": 3.3849032008254676e-05, "loss": 2.2984, "step": 1394 }, { "epoch": 0.4467574059247398, "grad_norm": 6.21875, "learning_rate": 3.382287654894377e-05, "loss": 1.9116, "step": 1395 }, { "epoch": 0.4470776621297038, "grad_norm": 0.67578125, "learning_rate": 3.379671005376671e-05, "loss": 2.2836, "step": 1396 }, { "epoch": 0.44739791833466774, "grad_norm": 0.328125, "learning_rate": 3.377053255545319e-05, "loss": 1.6366, "step": 1397 }, { "epoch": 0.4477181745396317, "grad_norm": 0.470703125, "learning_rate": 3.374434408674665e-05, "loss": 2.4968, "step": 1398 }, { "epoch": 0.44803843074459565, "grad_norm": 0.37109375, "learning_rate": 3.371814468040426e-05, "loss": 2.4331, "step": 1399 }, { "epoch": 0.44835868694955966, "grad_norm": 0.50390625, "learning_rate": 3.3691934369196855e-05, "loss": 2.381, "step": 1400 }, { "epoch": 0.4486789431545236, "grad_norm": 0.87109375, "learning_rate": 3.366571318590894e-05, "loss": 1.7545, "step": 1401 }, { "epoch": 0.44899919935948757, "grad_norm": 0.380859375, "learning_rate": 3.3639481163338584e-05, "loss": 2.4128, "step": 1402 }, { "epoch": 0.4493194555644516, "grad_norm": 0.37890625, "learning_rate": 3.3613238334297446e-05, "loss": 1.8779, "step": 1403 }, { "epoch": 0.44963971176941553, "grad_norm": 0.447265625, "learning_rate": 3.358698473161068e-05, "loss": 2.3488, "step": 1404 }, { "epoch": 0.4499599679743795, "grad_norm": 0.54296875, "learning_rate": 3.3560720388116934e-05, "loss": 1.6194, "step": 1405 }, { "epoch": 0.4502802241793435, "grad_norm": 1.2421875, "learning_rate": 3.3534445336668266e-05, "loss": 2.0731, "step": 1406 }, { "epoch": 0.45060048038430744, "grad_norm": 0.349609375, "learning_rate": 3.350815961013014e-05, "loss": 2.0653, "step": 1407 }, { "epoch": 0.4509207365892714, "grad_norm": 0.56640625, "learning_rate": 3.348186324138139e-05, "loss": 2.5686, "step": 1408 }, { "epoch": 0.4512409927942354, "grad_norm": 0.59375, "learning_rate": 3.345555626331415e-05, "loss": 2.0038, "step": 1409 }, { "epoch": 0.45156124899919936, "grad_norm": 0.56640625, "learning_rate": 3.34292387088338e-05, "loss": 2.3573, "step": 1410 }, { "epoch": 0.4518815052041633, "grad_norm": 0.56640625, "learning_rate": 3.340291061085898e-05, "loss": 2.2184, "step": 1411 }, { "epoch": 0.4522017614091273, "grad_norm": 0.5703125, "learning_rate": 3.3376572002321506e-05, "loss": 2.8585, "step": 1412 }, { "epoch": 0.4525220176140913, "grad_norm": 0.400390625, "learning_rate": 3.335022291616635e-05, "loss": 2.5615, "step": 1413 }, { "epoch": 0.45284227381905523, "grad_norm": 0.734375, "learning_rate": 3.332386338535158e-05, "loss": 2.1242, "step": 1414 }, { "epoch": 0.45316253002401924, "grad_norm": 0.88671875, "learning_rate": 3.329749344284832e-05, "loss": 2.2279, "step": 1415 }, { "epoch": 0.4534827862289832, "grad_norm": 0.64453125, "learning_rate": 3.327111312164075e-05, "loss": 2.3805, "step": 1416 }, { "epoch": 0.45380304243394715, "grad_norm": 0.41796875, "learning_rate": 3.324472245472598e-05, "loss": 2.2084, "step": 1417 }, { "epoch": 0.45412329863891115, "grad_norm": 0.5390625, "learning_rate": 3.321832147511412e-05, "loss": 2.6584, "step": 1418 }, { "epoch": 0.4544435548438751, "grad_norm": 0.5078125, "learning_rate": 3.319191021582814e-05, "loss": 1.9955, "step": 1419 }, { "epoch": 0.45476381104883906, "grad_norm": 0.455078125, "learning_rate": 3.3165488709903876e-05, "loss": 2.1134, "step": 1420 }, { "epoch": 0.455084067253803, "grad_norm": 0.474609375, "learning_rate": 3.313905699038997e-05, "loss": 2.4377, "step": 1421 }, { "epoch": 0.455404323458767, "grad_norm": 0.498046875, "learning_rate": 3.3112615090347875e-05, "loss": 2.3482, "step": 1422 }, { "epoch": 0.455724579663731, "grad_norm": 0.416015625, "learning_rate": 3.3086163042851757e-05, "loss": 2.2396, "step": 1423 }, { "epoch": 0.45604483586869493, "grad_norm": 0.63671875, "learning_rate": 3.305970088098844e-05, "loss": 1.9499, "step": 1424 }, { "epoch": 0.45636509207365894, "grad_norm": 0.40234375, "learning_rate": 3.303322863785747e-05, "loss": 2.4442, "step": 1425 }, { "epoch": 0.4566853482786229, "grad_norm": 0.5078125, "learning_rate": 3.300674634657094e-05, "loss": 2.0177, "step": 1426 }, { "epoch": 0.45700560448358685, "grad_norm": 0.3203125, "learning_rate": 3.2980254040253554e-05, "loss": 2.3229, "step": 1427 }, { "epoch": 0.45732586068855086, "grad_norm": 0.478515625, "learning_rate": 3.2953751752042514e-05, "loss": 1.5711, "step": 1428 }, { "epoch": 0.4576461168935148, "grad_norm": 0.486328125, "learning_rate": 3.2927239515087524e-05, "loss": 2.3031, "step": 1429 }, { "epoch": 0.45796637309847876, "grad_norm": 1.03125, "learning_rate": 3.290071736255073e-05, "loss": 2.4136, "step": 1430 }, { "epoch": 0.45828662930344277, "grad_norm": 0.412109375, "learning_rate": 3.287418532760667e-05, "loss": 2.4615, "step": 1431 }, { "epoch": 0.4586068855084067, "grad_norm": 0.5078125, "learning_rate": 3.284764344344226e-05, "loss": 2.3343, "step": 1432 }, { "epoch": 0.4589271417133707, "grad_norm": 0.76171875, "learning_rate": 3.282109174325672e-05, "loss": 2.6001, "step": 1433 }, { "epoch": 0.4592473979183347, "grad_norm": 0.75, "learning_rate": 3.279453026026157e-05, "loss": 2.282, "step": 1434 }, { "epoch": 0.45956765412329864, "grad_norm": 0.68359375, "learning_rate": 3.2767959027680525e-05, "loss": 2.6124, "step": 1435 }, { "epoch": 0.4598879103282626, "grad_norm": 0.47265625, "learning_rate": 3.2741378078749545e-05, "loss": 1.9807, "step": 1436 }, { "epoch": 0.4602081665332266, "grad_norm": 0.6328125, "learning_rate": 3.271478744671672e-05, "loss": 2.7701, "step": 1437 }, { "epoch": 0.46052842273819056, "grad_norm": 2.09375, "learning_rate": 3.2688187164842235e-05, "loss": 2.2759, "step": 1438 }, { "epoch": 0.4608486789431545, "grad_norm": 0.474609375, "learning_rate": 3.266157726639836e-05, "loss": 2.1833, "step": 1439 }, { "epoch": 0.4611689351481185, "grad_norm": 0.64453125, "learning_rate": 3.263495778466942e-05, "loss": 2.2356, "step": 1440 }, { "epoch": 0.4614891913530825, "grad_norm": 0.94140625, "learning_rate": 3.260832875295169e-05, "loss": 2.2155, "step": 1441 }, { "epoch": 0.4618094475580464, "grad_norm": 0.349609375, "learning_rate": 3.25816902045534e-05, "loss": 1.4778, "step": 1442 }, { "epoch": 0.46212970376301044, "grad_norm": 0.447265625, "learning_rate": 3.255504217279469e-05, "loss": 1.7559, "step": 1443 }, { "epoch": 0.4624499599679744, "grad_norm": 0.515625, "learning_rate": 3.2528384691007555e-05, "loss": 2.1804, "step": 1444 }, { "epoch": 0.46277021617293834, "grad_norm": 0.69140625, "learning_rate": 3.250171779253583e-05, "loss": 2.2981, "step": 1445 }, { "epoch": 0.4630904723779023, "grad_norm": 0.48046875, "learning_rate": 3.2475041510735086e-05, "loss": 2.4689, "step": 1446 }, { "epoch": 0.4634107285828663, "grad_norm": 0.43359375, "learning_rate": 3.244835587897268e-05, "loss": 2.397, "step": 1447 }, { "epoch": 0.46373098478783026, "grad_norm": 0.482421875, "learning_rate": 3.242166093062764e-05, "loss": 2.2169, "step": 1448 }, { "epoch": 0.4640512409927942, "grad_norm": 0.57421875, "learning_rate": 3.239495669909064e-05, "loss": 2.2407, "step": 1449 }, { "epoch": 0.4643714971977582, "grad_norm": 0.57421875, "learning_rate": 3.236824321776397e-05, "loss": 2.1264, "step": 1450 }, { "epoch": 0.4646917534027222, "grad_norm": 0.46875, "learning_rate": 3.2341520520061516e-05, "loss": 2.2125, "step": 1451 }, { "epoch": 0.4650120096076861, "grad_norm": 0.435546875, "learning_rate": 3.2314788639408656e-05, "loss": 2.1495, "step": 1452 }, { "epoch": 0.46533226581265014, "grad_norm": 0.466796875, "learning_rate": 3.2288047609242266e-05, "loss": 2.029, "step": 1453 }, { "epoch": 0.4656525220176141, "grad_norm": 0.625, "learning_rate": 3.226129746301067e-05, "loss": 1.9174, "step": 1454 }, { "epoch": 0.46597277822257804, "grad_norm": 0.494140625, "learning_rate": 3.223453823417362e-05, "loss": 2.5737, "step": 1455 }, { "epoch": 0.46629303442754205, "grad_norm": 0.62109375, "learning_rate": 3.220776995620217e-05, "loss": 2.2023, "step": 1456 }, { "epoch": 0.466613290632506, "grad_norm": 0.4375, "learning_rate": 3.218099266257874e-05, "loss": 2.6245, "step": 1457 }, { "epoch": 0.46693354683746996, "grad_norm": 0.455078125, "learning_rate": 3.2154206386797e-05, "loss": 1.9389, "step": 1458 }, { "epoch": 0.46725380304243397, "grad_norm": 0.53125, "learning_rate": 3.212741116236189e-05, "loss": 2.0921, "step": 1459 }, { "epoch": 0.4675740592473979, "grad_norm": 0.5078125, "learning_rate": 3.210060702278951e-05, "loss": 2.5264, "step": 1460 }, { "epoch": 0.4678943154523619, "grad_norm": 0.87890625, "learning_rate": 3.207379400160711e-05, "loss": 1.9612, "step": 1461 }, { "epoch": 0.4682145716573259, "grad_norm": 0.482421875, "learning_rate": 3.204697213235309e-05, "loss": 2.329, "step": 1462 }, { "epoch": 0.46853482786228984, "grad_norm": 0.58203125, "learning_rate": 3.2020141448576866e-05, "loss": 2.4217, "step": 1463 }, { "epoch": 0.4688550840672538, "grad_norm": 0.453125, "learning_rate": 3.1993301983838906e-05, "loss": 1.7458, "step": 1464 }, { "epoch": 0.4691753402722178, "grad_norm": 0.3828125, "learning_rate": 3.1966453771710655e-05, "loss": 2.112, "step": 1465 }, { "epoch": 0.46949559647718175, "grad_norm": 0.546875, "learning_rate": 3.1939596845774525e-05, "loss": 2.4703, "step": 1466 }, { "epoch": 0.4698158526821457, "grad_norm": 0.71875, "learning_rate": 3.191273123962378e-05, "loss": 2.4275, "step": 1467 }, { "epoch": 0.47013610888710966, "grad_norm": 0.53125, "learning_rate": 3.188585698686257e-05, "loss": 2.4174, "step": 1468 }, { "epoch": 0.47045636509207367, "grad_norm": 0.6328125, "learning_rate": 3.185897412110587e-05, "loss": 2.2741, "step": 1469 }, { "epoch": 0.4707766212970376, "grad_norm": 0.80078125, "learning_rate": 3.1832082675979394e-05, "loss": 2.2359, "step": 1470 }, { "epoch": 0.4710968775020016, "grad_norm": 0.6953125, "learning_rate": 3.180518268511963e-05, "loss": 2.174, "step": 1471 }, { "epoch": 0.4714171337069656, "grad_norm": 0.447265625, "learning_rate": 3.1778274182173716e-05, "loss": 2.2126, "step": 1472 }, { "epoch": 0.47173738991192954, "grad_norm": 0.392578125, "learning_rate": 3.175135720079947e-05, "loss": 2.0206, "step": 1473 }, { "epoch": 0.4720576461168935, "grad_norm": 0.74609375, "learning_rate": 3.172443177466529e-05, "loss": 2.4025, "step": 1474 }, { "epoch": 0.4723779023218575, "grad_norm": 0.73828125, "learning_rate": 3.169749793745014e-05, "loss": 2.6801, "step": 1475 }, { "epoch": 0.47269815852682145, "grad_norm": 0.55859375, "learning_rate": 3.167055572284352e-05, "loss": 2.2362, "step": 1476 }, { "epoch": 0.4730184147317854, "grad_norm": 0.43359375, "learning_rate": 3.164360516454541e-05, "loss": 2.0306, "step": 1477 }, { "epoch": 0.4733386709367494, "grad_norm": 0.431640625, "learning_rate": 3.1616646296266196e-05, "loss": 2.1496, "step": 1478 }, { "epoch": 0.47365892714171337, "grad_norm": 1.171875, "learning_rate": 3.158967915172669e-05, "loss": 2.6622, "step": 1479 }, { "epoch": 0.4739791833466773, "grad_norm": 0.80078125, "learning_rate": 3.1562703764658055e-05, "loss": 2.2467, "step": 1480 }, { "epoch": 0.47429943955164133, "grad_norm": 0.349609375, "learning_rate": 3.1535720168801725e-05, "loss": 1.7394, "step": 1481 }, { "epoch": 0.4746196957566053, "grad_norm": 0.462890625, "learning_rate": 3.150872839790946e-05, "loss": 1.9964, "step": 1482 }, { "epoch": 0.47493995196156924, "grad_norm": 0.5703125, "learning_rate": 3.148172848574319e-05, "loss": 2.3196, "step": 1483 }, { "epoch": 0.47526020816653325, "grad_norm": 0.40234375, "learning_rate": 3.145472046607508e-05, "loss": 2.052, "step": 1484 }, { "epoch": 0.4755804643714972, "grad_norm": 0.4296875, "learning_rate": 3.142770437268739e-05, "loss": 1.8644, "step": 1485 }, { "epoch": 0.47590072057646116, "grad_norm": 0.83203125, "learning_rate": 3.140068023937251e-05, "loss": 2.1707, "step": 1486 }, { "epoch": 0.47622097678142516, "grad_norm": 0.59375, "learning_rate": 3.137364809993288e-05, "loss": 1.9486, "step": 1487 }, { "epoch": 0.4765412329863891, "grad_norm": 0.431640625, "learning_rate": 3.1346607988180933e-05, "loss": 1.5815, "step": 1488 }, { "epoch": 0.47686148919135307, "grad_norm": 0.59375, "learning_rate": 3.131955993793911e-05, "loss": 2.0043, "step": 1489 }, { "epoch": 0.4771817453963171, "grad_norm": 0.546875, "learning_rate": 3.129250398303975e-05, "loss": 2.2674, "step": 1490 }, { "epoch": 0.47750200160128103, "grad_norm": 0.3828125, "learning_rate": 3.12654401573251e-05, "loss": 2.037, "step": 1491 }, { "epoch": 0.477822257806245, "grad_norm": 0.427734375, "learning_rate": 3.123836849464723e-05, "loss": 1.7428, "step": 1492 }, { "epoch": 0.47814251401120894, "grad_norm": 0.921875, "learning_rate": 3.121128902886804e-05, "loss": 2.6863, "step": 1493 }, { "epoch": 0.47846277021617295, "grad_norm": 0.515625, "learning_rate": 3.118420179385919e-05, "loss": 2.0104, "step": 1494 }, { "epoch": 0.4787830264211369, "grad_norm": 0.5234375, "learning_rate": 3.115710682350201e-05, "loss": 2.1841, "step": 1495 }, { "epoch": 0.47910328262610086, "grad_norm": 0.61328125, "learning_rate": 3.113000415168756e-05, "loss": 2.503, "step": 1496 }, { "epoch": 0.47942353883106487, "grad_norm": 0.32421875, "learning_rate": 3.110289381231651e-05, "loss": 1.9073, "step": 1497 }, { "epoch": 0.4797437950360288, "grad_norm": 1.421875, "learning_rate": 3.107577583929914e-05, "loss": 1.9962, "step": 1498 }, { "epoch": 0.4800640512409928, "grad_norm": 0.48046875, "learning_rate": 3.104865026655525e-05, "loss": 2.0423, "step": 1499 }, { "epoch": 0.4803843074459568, "grad_norm": 0.41796875, "learning_rate": 3.102151712801415e-05, "loss": 2.1058, "step": 1500 }, { "epoch": 0.48070456365092074, "grad_norm": 0.60546875, "learning_rate": 3.0994376457614636e-05, "loss": 2.2794, "step": 1501 }, { "epoch": 0.4810248198558847, "grad_norm": 0.98046875, "learning_rate": 3.09672282893049e-05, "loss": 2.1654, "step": 1502 }, { "epoch": 0.4813450760608487, "grad_norm": 0.40625, "learning_rate": 3.094007265704253e-05, "loss": 1.8414, "step": 1503 }, { "epoch": 0.48166533226581265, "grad_norm": 0.470703125, "learning_rate": 3.091290959479444e-05, "loss": 1.9365, "step": 1504 }, { "epoch": 0.4819855884707766, "grad_norm": 1.1796875, "learning_rate": 3.0885739136536854e-05, "loss": 2.2331, "step": 1505 }, { "epoch": 0.4823058446757406, "grad_norm": 0.435546875, "learning_rate": 3.0858561316255224e-05, "loss": 1.8893, "step": 1506 }, { "epoch": 0.48262610088070457, "grad_norm": 0.34375, "learning_rate": 3.083137616794422e-05, "loss": 1.8019, "step": 1507 }, { "epoch": 0.4829463570856685, "grad_norm": 0.439453125, "learning_rate": 3.080418372560768e-05, "loss": 1.971, "step": 1508 }, { "epoch": 0.48326661329063253, "grad_norm": 1.4765625, "learning_rate": 3.077698402325857e-05, "loss": 2.0729, "step": 1509 }, { "epoch": 0.4835868694955965, "grad_norm": 0.609375, "learning_rate": 3.0749777094918924e-05, "loss": 2.444, "step": 1510 }, { "epoch": 0.48390712570056044, "grad_norm": 0.44140625, "learning_rate": 3.072256297461983e-05, "loss": 2.6036, "step": 1511 }, { "epoch": 0.48422738190552445, "grad_norm": 0.4921875, "learning_rate": 3.069534169640136e-05, "loss": 2.5759, "step": 1512 }, { "epoch": 0.4845476381104884, "grad_norm": 0.484375, "learning_rate": 3.066811329431254e-05, "loss": 2.5265, "step": 1513 }, { "epoch": 0.48486789431545235, "grad_norm": 0.6875, "learning_rate": 3.0640877802411316e-05, "loss": 2.5761, "step": 1514 }, { "epoch": 0.4851881505204163, "grad_norm": 0.478515625, "learning_rate": 3.0613635254764495e-05, "loss": 1.9635, "step": 1515 }, { "epoch": 0.4855084067253803, "grad_norm": 0.435546875, "learning_rate": 3.058638568544772e-05, "loss": 1.9067, "step": 1516 }, { "epoch": 0.48582866293034427, "grad_norm": 0.59375, "learning_rate": 3.055912912854538e-05, "loss": 2.7006, "step": 1517 }, { "epoch": 0.4861489191353082, "grad_norm": 1.8671875, "learning_rate": 3.0531865618150654e-05, "loss": 2.7706, "step": 1518 }, { "epoch": 0.48646917534027223, "grad_norm": 2.703125, "learning_rate": 3.050459518836539e-05, "loss": 2.2856, "step": 1519 }, { "epoch": 0.4867894315452362, "grad_norm": 0.498046875, "learning_rate": 3.0477317873300103e-05, "loss": 2.2043, "step": 1520 }, { "epoch": 0.48710968775020014, "grad_norm": 1.3828125, "learning_rate": 3.045003370707391e-05, "loss": 1.7819, "step": 1521 }, { "epoch": 0.48742994395516415, "grad_norm": 0.57421875, "learning_rate": 3.0422742723814503e-05, "loss": 2.2534, "step": 1522 }, { "epoch": 0.4877502001601281, "grad_norm": 0.37109375, "learning_rate": 3.0395444957658097e-05, "loss": 1.9324, "step": 1523 }, { "epoch": 0.48807045636509205, "grad_norm": 0.498046875, "learning_rate": 3.0368140442749398e-05, "loss": 2.167, "step": 1524 }, { "epoch": 0.48839071257005606, "grad_norm": 0.546875, "learning_rate": 3.034082921324155e-05, "loss": 2.6054, "step": 1525 }, { "epoch": 0.48871096877502, "grad_norm": 0.546875, "learning_rate": 3.03135113032961e-05, "loss": 2.4174, "step": 1526 }, { "epoch": 0.48903122497998397, "grad_norm": 0.62890625, "learning_rate": 3.0286186747082934e-05, "loss": 2.3323, "step": 1527 }, { "epoch": 0.489351481184948, "grad_norm": 1.953125, "learning_rate": 3.0258855578780265e-05, "loss": 2.1546, "step": 1528 }, { "epoch": 0.48967173738991193, "grad_norm": 0.55078125, "learning_rate": 3.023151783257459e-05, "loss": 2.4691, "step": 1529 }, { "epoch": 0.4899919935948759, "grad_norm": 0.396484375, "learning_rate": 3.0204173542660613e-05, "loss": 2.0822, "step": 1530 }, { "epoch": 0.4903122497998399, "grad_norm": 0.89453125, "learning_rate": 3.0176822743241222e-05, "loss": 2.6899, "step": 1531 }, { "epoch": 0.49063250600480385, "grad_norm": 0.51171875, "learning_rate": 3.014946546852746e-05, "loss": 2.5449, "step": 1532 }, { "epoch": 0.4909527622097678, "grad_norm": 0.71484375, "learning_rate": 3.0122101752738468e-05, "loss": 2.0894, "step": 1533 }, { "epoch": 0.4912730184147318, "grad_norm": 0.46484375, "learning_rate": 3.0094731630101436e-05, "loss": 2.2903, "step": 1534 }, { "epoch": 0.49159327461969576, "grad_norm": 0.427734375, "learning_rate": 3.0067355134851565e-05, "loss": 2.2904, "step": 1535 }, { "epoch": 0.4919135308246597, "grad_norm": 0.365234375, "learning_rate": 3.0039972301232045e-05, "loss": 2.2971, "step": 1536 }, { "epoch": 0.4922337870296237, "grad_norm": 0.46875, "learning_rate": 3.001258316349398e-05, "loss": 1.7656, "step": 1537 }, { "epoch": 0.4925540432345877, "grad_norm": 0.48828125, "learning_rate": 2.998518775589635e-05, "loss": 2.0205, "step": 1538 }, { "epoch": 0.49287429943955163, "grad_norm": 0.59765625, "learning_rate": 2.9957786112706006e-05, "loss": 2.3737, "step": 1539 }, { "epoch": 0.4931945556445156, "grad_norm": 0.40625, "learning_rate": 2.9930378268197577e-05, "loss": 2.1592, "step": 1540 }, { "epoch": 0.4935148118494796, "grad_norm": 0.5390625, "learning_rate": 2.990296425665345e-05, "loss": 2.3567, "step": 1541 }, { "epoch": 0.49383506805444355, "grad_norm": 0.53515625, "learning_rate": 2.987554411236373e-05, "loss": 2.3767, "step": 1542 }, { "epoch": 0.4941553242594075, "grad_norm": 1.625, "learning_rate": 2.9848117869626192e-05, "loss": 2.5746, "step": 1543 }, { "epoch": 0.4944755804643715, "grad_norm": 0.349609375, "learning_rate": 2.9820685562746254e-05, "loss": 1.4638, "step": 1544 }, { "epoch": 0.49479583666933546, "grad_norm": 0.97265625, "learning_rate": 2.9793247226036885e-05, "loss": 1.9633, "step": 1545 }, { "epoch": 0.4951160928742994, "grad_norm": 0.4765625, "learning_rate": 2.9765802893818627e-05, "loss": 1.7408, "step": 1546 }, { "epoch": 0.4954363490792634, "grad_norm": 0.474609375, "learning_rate": 2.9738352600419506e-05, "loss": 2.0108, "step": 1547 }, { "epoch": 0.4957566052842274, "grad_norm": 0.388671875, "learning_rate": 2.9710896380175024e-05, "loss": 2.3077, "step": 1548 }, { "epoch": 0.49607686148919133, "grad_norm": 0.86328125, "learning_rate": 2.9683434267428055e-05, "loss": 2.0489, "step": 1549 }, { "epoch": 0.49639711769415534, "grad_norm": 1.3125, "learning_rate": 2.9655966296528893e-05, "loss": 2.8007, "step": 1550 }, { "epoch": 0.4967173738991193, "grad_norm": 0.412109375, "learning_rate": 2.962849250183513e-05, "loss": 2.0005, "step": 1551 }, { "epoch": 0.49703763010408325, "grad_norm": 0.6953125, "learning_rate": 2.9601012917711646e-05, "loss": 1.5309, "step": 1552 }, { "epoch": 0.49735788630904726, "grad_norm": 0.46484375, "learning_rate": 2.957352757853056e-05, "loss": 2.4253, "step": 1553 }, { "epoch": 0.4976781425140112, "grad_norm": 0.416015625, "learning_rate": 2.9546036518671206e-05, "loss": 1.8466, "step": 1554 }, { "epoch": 0.49799839871897517, "grad_norm": 0.609375, "learning_rate": 2.9518539772520064e-05, "loss": 2.2555, "step": 1555 }, { "epoch": 0.4983186549239392, "grad_norm": 0.419921875, "learning_rate": 2.9491037374470715e-05, "loss": 2.0387, "step": 1556 }, { "epoch": 0.49863891112890313, "grad_norm": 0.7578125, "learning_rate": 2.9463529358923824e-05, "loss": 2.4217, "step": 1557 }, { "epoch": 0.4989591673338671, "grad_norm": 0.640625, "learning_rate": 2.9436015760287082e-05, "loss": 2.1835, "step": 1558 }, { "epoch": 0.4992794235388311, "grad_norm": 0.51953125, "learning_rate": 2.9408496612975156e-05, "loss": 2.1866, "step": 1559 }, { "epoch": 0.49959967974379504, "grad_norm": 0.4296875, "learning_rate": 2.9380971951409652e-05, "loss": 2.0304, "step": 1560 }, { "epoch": 0.499919935948759, "grad_norm": 0.640625, "learning_rate": 2.9353441810019093e-05, "loss": 1.4819, "step": 1561 }, { "epoch": 0.500240192153723, "grad_norm": 0.65234375, "learning_rate": 2.9325906223238836e-05, "loss": 1.8902, "step": 1562 }, { "epoch": 0.500560448358687, "grad_norm": 0.419921875, "learning_rate": 2.929836522551104e-05, "loss": 2.3322, "step": 1563 }, { "epoch": 0.500880704563651, "grad_norm": 0.470703125, "learning_rate": 2.927081885128467e-05, "loss": 1.9023, "step": 1564 }, { "epoch": 0.5012009607686149, "grad_norm": 0.4296875, "learning_rate": 2.9243267135015383e-05, "loss": 2.1684, "step": 1565 }, { "epoch": 0.5015212169735789, "grad_norm": 0.54296875, "learning_rate": 2.921571011116552e-05, "loss": 2.0673, "step": 1566 }, { "epoch": 0.5018414731785429, "grad_norm": 0.50390625, "learning_rate": 2.9188147814204086e-05, "loss": 2.5705, "step": 1567 }, { "epoch": 0.5021617293835068, "grad_norm": 0.515625, "learning_rate": 2.916058027860665e-05, "loss": 2.6609, "step": 1568 }, { "epoch": 0.5024819855884708, "grad_norm": 0.484375, "learning_rate": 2.913300753885536e-05, "loss": 2.2706, "step": 1569 }, { "epoch": 0.5028022417934348, "grad_norm": 0.50390625, "learning_rate": 2.910542962943885e-05, "loss": 2.5645, "step": 1570 }, { "epoch": 0.5031224979983987, "grad_norm": 0.578125, "learning_rate": 2.907784658485225e-05, "loss": 2.7824, "step": 1571 }, { "epoch": 0.5034427542033627, "grad_norm": 0.58203125, "learning_rate": 2.905025843959708e-05, "loss": 1.5037, "step": 1572 }, { "epoch": 0.5037630104083266, "grad_norm": 1.1015625, "learning_rate": 2.902266522818125e-05, "loss": 2.2413, "step": 1573 }, { "epoch": 0.5040832666132906, "grad_norm": 0.439453125, "learning_rate": 2.899506698511903e-05, "loss": 2.5246, "step": 1574 }, { "epoch": 0.5044035228182546, "grad_norm": 0.4375, "learning_rate": 2.8967463744930958e-05, "loss": 1.7726, "step": 1575 }, { "epoch": 0.5047237790232185, "grad_norm": 0.51953125, "learning_rate": 2.8939855542143828e-05, "loss": 1.8297, "step": 1576 }, { "epoch": 0.5050440352281825, "grad_norm": 0.4453125, "learning_rate": 2.8912242411290646e-05, "loss": 1.8675, "step": 1577 }, { "epoch": 0.5053642914331465, "grad_norm": 1.0546875, "learning_rate": 2.8884624386910586e-05, "loss": 2.3941, "step": 1578 }, { "epoch": 0.5056845476381104, "grad_norm": 0.51171875, "learning_rate": 2.8857001503548925e-05, "loss": 2.5356, "step": 1579 }, { "epoch": 0.5060048038430744, "grad_norm": 0.6640625, "learning_rate": 2.882937379575704e-05, "loss": 2.5708, "step": 1580 }, { "epoch": 0.5063250600480385, "grad_norm": 0.50390625, "learning_rate": 2.8801741298092332e-05, "loss": 2.1658, "step": 1581 }, { "epoch": 0.5066453162530024, "grad_norm": 0.50390625, "learning_rate": 2.8774104045118183e-05, "loss": 1.7119, "step": 1582 }, { "epoch": 0.5069655724579664, "grad_norm": 0.71875, "learning_rate": 2.8746462071403952e-05, "loss": 1.936, "step": 1583 }, { "epoch": 0.5072858286629304, "grad_norm": 0.462890625, "learning_rate": 2.8718815411524863e-05, "loss": 2.1219, "step": 1584 }, { "epoch": 0.5076060848678943, "grad_norm": 0.52734375, "learning_rate": 2.8691164100062034e-05, "loss": 2.5705, "step": 1585 }, { "epoch": 0.5079263410728583, "grad_norm": 0.48046875, "learning_rate": 2.8663508171602386e-05, "loss": 2.2458, "step": 1586 }, { "epoch": 0.5082465972778223, "grad_norm": 1.390625, "learning_rate": 2.8635847660738635e-05, "loss": 1.8396, "step": 1587 }, { "epoch": 0.5085668534827862, "grad_norm": 0.515625, "learning_rate": 2.8608182602069195e-05, "loss": 2.5934, "step": 1588 }, { "epoch": 0.5088871096877502, "grad_norm": 0.625, "learning_rate": 2.8580513030198186e-05, "loss": 2.4622, "step": 1589 }, { "epoch": 0.5092073658927142, "grad_norm": 0.490234375, "learning_rate": 2.8552838979735393e-05, "loss": 2.2395, "step": 1590 }, { "epoch": 0.5095276220976781, "grad_norm": 0.51953125, "learning_rate": 2.8525160485296167e-05, "loss": 2.172, "step": 1591 }, { "epoch": 0.5098478783026421, "grad_norm": 0.5078125, "learning_rate": 2.849747758150143e-05, "loss": 2.4499, "step": 1592 }, { "epoch": 0.5101681345076061, "grad_norm": 0.5390625, "learning_rate": 2.846979030297764e-05, "loss": 2.2869, "step": 1593 }, { "epoch": 0.51048839071257, "grad_norm": 0.37109375, "learning_rate": 2.8442098684356707e-05, "loss": 1.8278, "step": 1594 }, { "epoch": 0.510808646917534, "grad_norm": 0.458984375, "learning_rate": 2.841440276027596e-05, "loss": 2.2451, "step": 1595 }, { "epoch": 0.511128903122498, "grad_norm": 0.859375, "learning_rate": 2.8386702565378137e-05, "loss": 2.3676, "step": 1596 }, { "epoch": 0.5114491593274619, "grad_norm": 0.921875, "learning_rate": 2.8358998134311316e-05, "loss": 2.4343, "step": 1597 }, { "epoch": 0.5117694155324259, "grad_norm": 0.62109375, "learning_rate": 2.8331289501728854e-05, "loss": 2.516, "step": 1598 }, { "epoch": 0.51208967173739, "grad_norm": 0.5234375, "learning_rate": 2.830357670228937e-05, "loss": 2.5648, "step": 1599 }, { "epoch": 0.5124099279423538, "grad_norm": 0.5234375, "learning_rate": 2.8275859770656714e-05, "loss": 2.4266, "step": 1600 }, { "epoch": 0.5127301841473179, "grad_norm": 0.43359375, "learning_rate": 2.8248138741499885e-05, "loss": 2.4382, "step": 1601 }, { "epoch": 0.5130504403522819, "grad_norm": 0.578125, "learning_rate": 2.822041364949301e-05, "loss": 2.4932, "step": 1602 }, { "epoch": 0.5133706965572458, "grad_norm": 0.51171875, "learning_rate": 2.8192684529315305e-05, "loss": 2.121, "step": 1603 }, { "epoch": 0.5136909527622098, "grad_norm": 0.5859375, "learning_rate": 2.816495141565102e-05, "loss": 2.0563, "step": 1604 }, { "epoch": 0.5140112089671738, "grad_norm": 0.396484375, "learning_rate": 2.8137214343189395e-05, "loss": 2.4438, "step": 1605 }, { "epoch": 0.5143314651721377, "grad_norm": 0.369140625, "learning_rate": 2.8109473346624627e-05, "loss": 2.3331, "step": 1606 }, { "epoch": 0.5146517213771017, "grad_norm": 0.43359375, "learning_rate": 2.808172846065583e-05, "loss": 2.2871, "step": 1607 }, { "epoch": 0.5149719775820657, "grad_norm": 0.53515625, "learning_rate": 2.8053979719986972e-05, "loss": 2.4588, "step": 1608 }, { "epoch": 0.5152922337870296, "grad_norm": 0.5703125, "learning_rate": 2.802622715932684e-05, "loss": 2.5462, "step": 1609 }, { "epoch": 0.5156124899919936, "grad_norm": 0.4765625, "learning_rate": 2.799847081338901e-05, "loss": 1.7209, "step": 1610 }, { "epoch": 0.5159327461969576, "grad_norm": 1.4609375, "learning_rate": 2.7970710716891783e-05, "loss": 2.7945, "step": 1611 }, { "epoch": 0.5162530024019215, "grad_norm": 0.5390625, "learning_rate": 2.7942946904558155e-05, "loss": 2.2602, "step": 1612 }, { "epoch": 0.5165732586068855, "grad_norm": 0.466796875, "learning_rate": 2.791517941111577e-05, "loss": 2.0592, "step": 1613 }, { "epoch": 0.5168935148118495, "grad_norm": 0.5234375, "learning_rate": 2.7887408271296883e-05, "loss": 2.1679, "step": 1614 }, { "epoch": 0.5172137710168134, "grad_norm": 0.52734375, "learning_rate": 2.7859633519838296e-05, "loss": 2.5147, "step": 1615 }, { "epoch": 0.5175340272217774, "grad_norm": 0.62109375, "learning_rate": 2.7831855191481336e-05, "loss": 2.471, "step": 1616 }, { "epoch": 0.5178542834267414, "grad_norm": 0.69921875, "learning_rate": 2.7804073320971812e-05, "loss": 2.2137, "step": 1617 }, { "epoch": 0.5181745396317053, "grad_norm": 1.640625, "learning_rate": 2.777628794305995e-05, "loss": 2.0848, "step": 1618 }, { "epoch": 0.5184947958366694, "grad_norm": 0.53515625, "learning_rate": 2.774849909250037e-05, "loss": 2.0605, "step": 1619 }, { "epoch": 0.5188150520416333, "grad_norm": 0.75, "learning_rate": 2.7720706804052026e-05, "loss": 2.6878, "step": 1620 }, { "epoch": 0.5191353082465973, "grad_norm": 0.490234375, "learning_rate": 2.7692911112478193e-05, "loss": 2.4631, "step": 1621 }, { "epoch": 0.5194555644515613, "grad_norm": 0.62109375, "learning_rate": 2.7665112052546387e-05, "loss": 2.5574, "step": 1622 }, { "epoch": 0.5197758206565252, "grad_norm": 0.421875, "learning_rate": 2.763730965902834e-05, "loss": 2.0565, "step": 1623 }, { "epoch": 0.5200960768614892, "grad_norm": 0.462890625, "learning_rate": 2.7609503966699945e-05, "loss": 1.9217, "step": 1624 }, { "epoch": 0.5204163330664532, "grad_norm": 0.46875, "learning_rate": 2.7581695010341246e-05, "loss": 2.049, "step": 1625 }, { "epoch": 0.5207365892714171, "grad_norm": 0.439453125, "learning_rate": 2.7553882824736348e-05, "loss": 2.1618, "step": 1626 }, { "epoch": 0.5210568454763811, "grad_norm": 0.494140625, "learning_rate": 2.7526067444673402e-05, "loss": 1.9671, "step": 1627 }, { "epoch": 0.5213771016813451, "grad_norm": 0.60546875, "learning_rate": 2.749824890494455e-05, "loss": 2.1159, "step": 1628 }, { "epoch": 0.521697357886309, "grad_norm": 0.423828125, "learning_rate": 2.74704272403459e-05, "loss": 1.5528, "step": 1629 }, { "epoch": 0.522017614091273, "grad_norm": 0.98828125, "learning_rate": 2.744260248567745e-05, "loss": 2.162, "step": 1630 }, { "epoch": 0.522337870296237, "grad_norm": 0.578125, "learning_rate": 2.741477467574307e-05, "loss": 1.8618, "step": 1631 }, { "epoch": 0.5226581265012009, "grad_norm": 0.49609375, "learning_rate": 2.738694384535046e-05, "loss": 1.8943, "step": 1632 }, { "epoch": 0.5229783827061649, "grad_norm": 0.73046875, "learning_rate": 2.73591100293111e-05, "loss": 2.0352, "step": 1633 }, { "epoch": 0.5232986389111289, "grad_norm": 0.51953125, "learning_rate": 2.7331273262440175e-05, "loss": 2.1791, "step": 1634 }, { "epoch": 0.5236188951160928, "grad_norm": 0.9375, "learning_rate": 2.7303433579556602e-05, "loss": 2.4775, "step": 1635 }, { "epoch": 0.5239391513210568, "grad_norm": 0.412109375, "learning_rate": 2.727559101548292e-05, "loss": 2.1351, "step": 1636 }, { "epoch": 0.5242594075260208, "grad_norm": 0.609375, "learning_rate": 2.7247745605045277e-05, "loss": 2.3471, "step": 1637 }, { "epoch": 0.5245796637309847, "grad_norm": 1.0390625, "learning_rate": 2.7219897383073373e-05, "loss": 1.6935, "step": 1638 }, { "epoch": 0.5248999199359488, "grad_norm": 0.451171875, "learning_rate": 2.7192046384400444e-05, "loss": 2.4096, "step": 1639 }, { "epoch": 0.5252201761409128, "grad_norm": 0.462890625, "learning_rate": 2.7164192643863196e-05, "loss": 2.2081, "step": 1640 }, { "epoch": 0.5255404323458767, "grad_norm": 0.5546875, "learning_rate": 2.7136336196301737e-05, "loss": 2.1331, "step": 1641 }, { "epoch": 0.5258606885508407, "grad_norm": 0.54296875, "learning_rate": 2.7108477076559595e-05, "loss": 2.1021, "step": 1642 }, { "epoch": 0.5261809447558047, "grad_norm": 0.546875, "learning_rate": 2.7080615319483634e-05, "loss": 2.491, "step": 1643 }, { "epoch": 0.5265012009607686, "grad_norm": 0.68359375, "learning_rate": 2.705275095992399e-05, "loss": 2.7152, "step": 1644 }, { "epoch": 0.5268214571657326, "grad_norm": 0.5234375, "learning_rate": 2.7024884032734082e-05, "loss": 2.138, "step": 1645 }, { "epoch": 0.5271417133706966, "grad_norm": 0.484375, "learning_rate": 2.6997014572770534e-05, "loss": 2.4889, "step": 1646 }, { "epoch": 0.5274619695756605, "grad_norm": 0.9453125, "learning_rate": 2.696914261489315e-05, "loss": 2.4738, "step": 1647 }, { "epoch": 0.5277822257806245, "grad_norm": 0.427734375, "learning_rate": 2.694126819396482e-05, "loss": 2.4388, "step": 1648 }, { "epoch": 0.5281024819855885, "grad_norm": 0.5078125, "learning_rate": 2.6913391344851556e-05, "loss": 2.0395, "step": 1649 }, { "epoch": 0.5284227381905524, "grad_norm": 1.125, "learning_rate": 2.688551210242239e-05, "loss": 2.4296, "step": 1650 }, { "epoch": 0.5287429943955164, "grad_norm": 0.453125, "learning_rate": 2.6857630501549347e-05, "loss": 2.1873, "step": 1651 }, { "epoch": 0.5290632506004804, "grad_norm": 0.46484375, "learning_rate": 2.68297465771074e-05, "loss": 2.5596, "step": 1652 }, { "epoch": 0.5293835068054443, "grad_norm": 0.48046875, "learning_rate": 2.6801860363974434e-05, "loss": 1.793, "step": 1653 }, { "epoch": 0.5297037630104083, "grad_norm": 0.58203125, "learning_rate": 2.6773971897031207e-05, "loss": 2.0554, "step": 1654 }, { "epoch": 0.5300240192153723, "grad_norm": 0.37109375, "learning_rate": 2.6746081211161268e-05, "loss": 1.7286, "step": 1655 }, { "epoch": 0.5303442754203362, "grad_norm": 0.396484375, "learning_rate": 2.6718188341250955e-05, "loss": 2.0696, "step": 1656 }, { "epoch": 0.5306645316253003, "grad_norm": 0.51171875, "learning_rate": 2.6690293322189353e-05, "loss": 1.9679, "step": 1657 }, { "epoch": 0.5309847878302643, "grad_norm": 0.5859375, "learning_rate": 2.6662396188868228e-05, "loss": 2.3103, "step": 1658 }, { "epoch": 0.5313050440352282, "grad_norm": 0.51953125, "learning_rate": 2.6634496976181968e-05, "loss": 2.3372, "step": 1659 }, { "epoch": 0.5316253002401922, "grad_norm": 0.578125, "learning_rate": 2.6606595719027583e-05, "loss": 2.3772, "step": 1660 }, { "epoch": 0.5319455564451562, "grad_norm": 0.734375, "learning_rate": 2.657869245230464e-05, "loss": 2.5822, "step": 1661 }, { "epoch": 0.5322658126501201, "grad_norm": 0.5390625, "learning_rate": 2.6550787210915223e-05, "loss": 1.8052, "step": 1662 }, { "epoch": 0.5325860688550841, "grad_norm": 0.69140625, "learning_rate": 2.6522880029763862e-05, "loss": 2.583, "step": 1663 }, { "epoch": 0.5329063250600481, "grad_norm": 0.50390625, "learning_rate": 2.6494970943757548e-05, "loss": 2.0871, "step": 1664 }, { "epoch": 0.533226581265012, "grad_norm": 0.5625, "learning_rate": 2.6467059987805633e-05, "loss": 2.2921, "step": 1665 }, { "epoch": 0.533546837469976, "grad_norm": 0.56640625, "learning_rate": 2.6439147196819795e-05, "loss": 1.3349, "step": 1666 }, { "epoch": 0.5338670936749399, "grad_norm": 1.0390625, "learning_rate": 2.6411232605714043e-05, "loss": 2.4875, "step": 1667 }, { "epoch": 0.5341873498799039, "grad_norm": 0.423828125, "learning_rate": 2.6383316249404615e-05, "loss": 1.9192, "step": 1668 }, { "epoch": 0.5345076060848679, "grad_norm": 0.427734375, "learning_rate": 2.6355398162809957e-05, "loss": 2.4075, "step": 1669 }, { "epoch": 0.5348278622898318, "grad_norm": 0.77734375, "learning_rate": 2.632747838085068e-05, "loss": 2.5198, "step": 1670 }, { "epoch": 0.5351481184947958, "grad_norm": 0.423828125, "learning_rate": 2.6299556938449528e-05, "loss": 2.4702, "step": 1671 }, { "epoch": 0.5354683746997598, "grad_norm": 0.56640625, "learning_rate": 2.627163387053131e-05, "loss": 2.4223, "step": 1672 }, { "epoch": 0.5357886309047237, "grad_norm": 0.546875, "learning_rate": 2.624370921202286e-05, "loss": 2.0182, "step": 1673 }, { "epoch": 0.5361088871096877, "grad_norm": 0.6328125, "learning_rate": 2.6215782997853026e-05, "loss": 2.2894, "step": 1674 }, { "epoch": 0.5364291433146517, "grad_norm": 0.58203125, "learning_rate": 2.6187855262952587e-05, "loss": 2.3239, "step": 1675 }, { "epoch": 0.5367493995196156, "grad_norm": 0.78125, "learning_rate": 2.615992604225422e-05, "loss": 2.5619, "step": 1676 }, { "epoch": 0.5370696557245797, "grad_norm": 0.6328125, "learning_rate": 2.6131995370692463e-05, "loss": 1.886, "step": 1677 }, { "epoch": 0.5373899119295437, "grad_norm": 0.8125, "learning_rate": 2.6104063283203672e-05, "loss": 2.2598, "step": 1678 }, { "epoch": 0.5377101681345076, "grad_norm": 0.408203125, "learning_rate": 2.607612981472599e-05, "loss": 1.7341, "step": 1679 }, { "epoch": 0.5380304243394716, "grad_norm": 1.328125, "learning_rate": 2.6048195000199248e-05, "loss": 1.8696, "step": 1680 }, { "epoch": 0.5383506805444356, "grad_norm": 0.51953125, "learning_rate": 2.6020258874565002e-05, "loss": 2.6784, "step": 1681 }, { "epoch": 0.5386709367493995, "grad_norm": 0.765625, "learning_rate": 2.5992321472766418e-05, "loss": 1.9824, "step": 1682 }, { "epoch": 0.5389911929543635, "grad_norm": 0.56640625, "learning_rate": 2.596438282974828e-05, "loss": 2.1182, "step": 1683 }, { "epoch": 0.5393114491593275, "grad_norm": 0.51953125, "learning_rate": 2.59364429804569e-05, "loss": 1.9854, "step": 1684 }, { "epoch": 0.5396317053642914, "grad_norm": 0.46484375, "learning_rate": 2.5908501959840125e-05, "loss": 2.27, "step": 1685 }, { "epoch": 0.5399519615692554, "grad_norm": 0.6796875, "learning_rate": 2.588055980284725e-05, "loss": 2.0739, "step": 1686 }, { "epoch": 0.5402722177742194, "grad_norm": 0.50390625, "learning_rate": 2.5852616544429008e-05, "loss": 2.2536, "step": 1687 }, { "epoch": 0.5405924739791833, "grad_norm": 0.5234375, "learning_rate": 2.5824672219537483e-05, "loss": 2.2221, "step": 1688 }, { "epoch": 0.5409127301841473, "grad_norm": 0.72265625, "learning_rate": 2.5796726863126113e-05, "loss": 2.7278, "step": 1689 }, { "epoch": 0.5412329863891113, "grad_norm": 0.5703125, "learning_rate": 2.5768780510149633e-05, "loss": 2.2353, "step": 1690 }, { "epoch": 0.5415532425940752, "grad_norm": 0.4140625, "learning_rate": 2.5740833195563996e-05, "loss": 1.9356, "step": 1691 }, { "epoch": 0.5418734987990392, "grad_norm": 0.52734375, "learning_rate": 2.571288495432639e-05, "loss": 1.7192, "step": 1692 }, { "epoch": 0.5421937550040032, "grad_norm": 0.50390625, "learning_rate": 2.5684935821395135e-05, "loss": 2.4505, "step": 1693 }, { "epoch": 0.5425140112089671, "grad_norm": 0.427734375, "learning_rate": 2.5656985831729692e-05, "loss": 1.8894, "step": 1694 }, { "epoch": 0.5428342674139311, "grad_norm": 0.4296875, "learning_rate": 2.5629035020290566e-05, "loss": 1.8297, "step": 1695 }, { "epoch": 0.5431545236188952, "grad_norm": 0.43359375, "learning_rate": 2.5601083422039314e-05, "loss": 2.2753, "step": 1696 }, { "epoch": 0.543474779823859, "grad_norm": 0.51171875, "learning_rate": 2.5573131071938456e-05, "loss": 2.8383, "step": 1697 }, { "epoch": 0.5437950360288231, "grad_norm": 0.44921875, "learning_rate": 2.554517800495148e-05, "loss": 2.4237, "step": 1698 }, { "epoch": 0.5441152922337871, "grad_norm": 0.40625, "learning_rate": 2.5517224256042745e-05, "loss": 2.0334, "step": 1699 }, { "epoch": 0.544435548438751, "grad_norm": 0.5078125, "learning_rate": 2.5489269860177474e-05, "loss": 2.4893, "step": 1700 }, { "epoch": 0.544755804643715, "grad_norm": 0.60546875, "learning_rate": 2.5461314852321694e-05, "loss": 2.0231, "step": 1701 }, { "epoch": 0.545076060848679, "grad_norm": 0.5390625, "learning_rate": 2.5433359267442204e-05, "loss": 2.486, "step": 1702 }, { "epoch": 0.5453963170536429, "grad_norm": 0.79296875, "learning_rate": 2.5405403140506517e-05, "loss": 2.3233, "step": 1703 }, { "epoch": 0.5457165732586069, "grad_norm": 0.53515625, "learning_rate": 2.5377446506482837e-05, "loss": 2.0646, "step": 1704 }, { "epoch": 0.5460368294635709, "grad_norm": 0.439453125, "learning_rate": 2.5349489400339997e-05, "loss": 2.1051, "step": 1705 }, { "epoch": 0.5463570856685348, "grad_norm": 0.5078125, "learning_rate": 2.5321531857047404e-05, "loss": 2.5043, "step": 1706 }, { "epoch": 0.5466773418734988, "grad_norm": 0.494140625, "learning_rate": 2.5293573911575036e-05, "loss": 1.8195, "step": 1707 }, { "epoch": 0.5469975980784628, "grad_norm": 0.50390625, "learning_rate": 2.5265615598893356e-05, "loss": 2.6448, "step": 1708 }, { "epoch": 0.5473178542834267, "grad_norm": 0.53515625, "learning_rate": 2.5237656953973303e-05, "loss": 2.3232, "step": 1709 }, { "epoch": 0.5476381104883907, "grad_norm": 0.63671875, "learning_rate": 2.5209698011786214e-05, "loss": 1.9998, "step": 1710 }, { "epoch": 0.5479583666933547, "grad_norm": 0.5859375, "learning_rate": 2.5181738807303806e-05, "loss": 2.0763, "step": 1711 }, { "epoch": 0.5482786228983186, "grad_norm": 0.58203125, "learning_rate": 2.5153779375498137e-05, "loss": 2.3951, "step": 1712 }, { "epoch": 0.5485988791032826, "grad_norm": 0.71484375, "learning_rate": 2.5125819751341522e-05, "loss": 1.9252, "step": 1713 }, { "epoch": 0.5489191353082465, "grad_norm": 0.546875, "learning_rate": 2.509785996980653e-05, "loss": 2.0553, "step": 1714 }, { "epoch": 0.5492393915132106, "grad_norm": 0.490234375, "learning_rate": 2.506990006586594e-05, "loss": 2.1069, "step": 1715 }, { "epoch": 0.5495596477181746, "grad_norm": 0.6171875, "learning_rate": 2.5041940074492676e-05, "loss": 2.3812, "step": 1716 }, { "epoch": 0.5498799039231385, "grad_norm": 0.5078125, "learning_rate": 2.5013980030659743e-05, "loss": 2.1561, "step": 1717 }, { "epoch": 0.5502001601281025, "grad_norm": 0.51953125, "learning_rate": 2.498601996934026e-05, "loss": 1.9662, "step": 1718 }, { "epoch": 0.5505204163330665, "grad_norm": 0.421875, "learning_rate": 2.4958059925507333e-05, "loss": 1.932, "step": 1719 }, { "epoch": 0.5508406725380304, "grad_norm": 0.85546875, "learning_rate": 2.4930099934134065e-05, "loss": 2.3476, "step": 1720 }, { "epoch": 0.5511609287429944, "grad_norm": 0.5078125, "learning_rate": 2.490214003019347e-05, "loss": 2.2002, "step": 1721 }, { "epoch": 0.5514811849479584, "grad_norm": 0.51171875, "learning_rate": 2.4874180248658484e-05, "loss": 2.3532, "step": 1722 }, { "epoch": 0.5518014411529223, "grad_norm": 0.53125, "learning_rate": 2.4846220624501872e-05, "loss": 2.258, "step": 1723 }, { "epoch": 0.5521216973578863, "grad_norm": 0.61328125, "learning_rate": 2.48182611926962e-05, "loss": 2.4362, "step": 1724 }, { "epoch": 0.5524419535628503, "grad_norm": 0.6171875, "learning_rate": 2.479030198821379e-05, "loss": 1.7804, "step": 1725 }, { "epoch": 0.5527622097678142, "grad_norm": 0.515625, "learning_rate": 2.4762343046026702e-05, "loss": 2.7527, "step": 1726 }, { "epoch": 0.5530824659727782, "grad_norm": 0.625, "learning_rate": 2.473438440110665e-05, "loss": 2.4608, "step": 1727 }, { "epoch": 0.5534027221777422, "grad_norm": 0.380859375, "learning_rate": 2.4706426088424973e-05, "loss": 1.8141, "step": 1728 }, { "epoch": 0.5537229783827061, "grad_norm": 0.53515625, "learning_rate": 2.46784681429526e-05, "loss": 2.2575, "step": 1729 }, { "epoch": 0.5540432345876701, "grad_norm": 0.33203125, "learning_rate": 2.4650510599660005e-05, "loss": 2.6203, "step": 1730 }, { "epoch": 0.5543634907926341, "grad_norm": 0.5234375, "learning_rate": 2.4622553493517165e-05, "loss": 2.1001, "step": 1731 }, { "epoch": 0.554683746997598, "grad_norm": 0.6796875, "learning_rate": 2.4594596859493485e-05, "loss": 2.3472, "step": 1732 }, { "epoch": 0.555004003202562, "grad_norm": 0.462890625, "learning_rate": 2.4566640732557802e-05, "loss": 2.0663, "step": 1733 }, { "epoch": 0.555324259407526, "grad_norm": 0.53515625, "learning_rate": 2.4538685147678316e-05, "loss": 2.3392, "step": 1734 }, { "epoch": 0.55564451561249, "grad_norm": 0.70703125, "learning_rate": 2.4510730139822535e-05, "loss": 2.1949, "step": 1735 }, { "epoch": 0.555964771817454, "grad_norm": 0.478515625, "learning_rate": 2.4482775743957258e-05, "loss": 2.517, "step": 1736 }, { "epoch": 0.556285028022418, "grad_norm": 0.43359375, "learning_rate": 2.4454821995048522e-05, "loss": 2.4129, "step": 1737 }, { "epoch": 0.5566052842273819, "grad_norm": 0.7109375, "learning_rate": 2.442686892806155e-05, "loss": 2.5147, "step": 1738 }, { "epoch": 0.5569255404323459, "grad_norm": 0.578125, "learning_rate": 2.4398916577960695e-05, "loss": 2.8204, "step": 1739 }, { "epoch": 0.5572457966373099, "grad_norm": 0.5859375, "learning_rate": 2.437096497970944e-05, "loss": 2.6784, "step": 1740 }, { "epoch": 0.5575660528422738, "grad_norm": 0.453125, "learning_rate": 2.4343014168270314e-05, "loss": 2.1789, "step": 1741 }, { "epoch": 0.5578863090472378, "grad_norm": 0.466796875, "learning_rate": 2.431506417860487e-05, "loss": 2.8155, "step": 1742 }, { "epoch": 0.5582065652522018, "grad_norm": 0.4765625, "learning_rate": 2.4287115045673617e-05, "loss": 2.0444, "step": 1743 }, { "epoch": 0.5585268214571657, "grad_norm": 0.59375, "learning_rate": 2.4259166804436006e-05, "loss": 2.646, "step": 1744 }, { "epoch": 0.5588470776621297, "grad_norm": 0.5859375, "learning_rate": 2.4231219489850376e-05, "loss": 2.2975, "step": 1745 }, { "epoch": 0.5591673338670937, "grad_norm": 0.765625, "learning_rate": 2.4203273136873892e-05, "loss": 2.1118, "step": 1746 }, { "epoch": 0.5594875900720576, "grad_norm": 0.462890625, "learning_rate": 2.4175327780462523e-05, "loss": 2.1737, "step": 1747 }, { "epoch": 0.5598078462770216, "grad_norm": 0.412109375, "learning_rate": 2.4147383455571004e-05, "loss": 2.2012, "step": 1748 }, { "epoch": 0.5601281024819856, "grad_norm": 0.5625, "learning_rate": 2.4119440197152754e-05, "loss": 1.4749, "step": 1749 }, { "epoch": 0.5604483586869495, "grad_norm": 0.640625, "learning_rate": 2.409149804015988e-05, "loss": 2.1635, "step": 1750 }, { "epoch": 0.5607686148919135, "grad_norm": 0.39453125, "learning_rate": 2.4063557019543107e-05, "loss": 1.4912, "step": 1751 }, { "epoch": 0.5610888710968776, "grad_norm": 0.5078125, "learning_rate": 2.403561717025173e-05, "loss": 2.4878, "step": 1752 }, { "epoch": 0.5614091273018414, "grad_norm": 0.55078125, "learning_rate": 2.400767852723359e-05, "loss": 2.3552, "step": 1753 }, { "epoch": 0.5617293835068055, "grad_norm": 0.4765625, "learning_rate": 2.3979741125435004e-05, "loss": 1.6363, "step": 1754 }, { "epoch": 0.5620496397117695, "grad_norm": 1.078125, "learning_rate": 2.3951804999800754e-05, "loss": 2.0071, "step": 1755 }, { "epoch": 0.5623698959167334, "grad_norm": 0.5625, "learning_rate": 2.3923870185274018e-05, "loss": 2.3637, "step": 1756 }, { "epoch": 0.5626901521216974, "grad_norm": 0.494140625, "learning_rate": 2.389593671679633e-05, "loss": 2.4539, "step": 1757 }, { "epoch": 0.5630104083266614, "grad_norm": 0.271484375, "learning_rate": 2.3868004629307543e-05, "loss": 2.087, "step": 1758 }, { "epoch": 0.5633306645316253, "grad_norm": 0.83984375, "learning_rate": 2.3840073957745793e-05, "loss": 2.3264, "step": 1759 }, { "epoch": 0.5636509207365893, "grad_norm": 0.55859375, "learning_rate": 2.381214473704742e-05, "loss": 2.3831, "step": 1760 }, { "epoch": 0.5639711769415532, "grad_norm": 0.52734375, "learning_rate": 2.3784217002146976e-05, "loss": 2.5073, "step": 1761 }, { "epoch": 0.5642914331465172, "grad_norm": 0.494140625, "learning_rate": 2.3756290787977147e-05, "loss": 2.3088, "step": 1762 }, { "epoch": 0.5646116893514812, "grad_norm": 0.69921875, "learning_rate": 2.3728366129468696e-05, "loss": 2.3283, "step": 1763 }, { "epoch": 0.5649319455564451, "grad_norm": 0.46484375, "learning_rate": 2.3700443061550478e-05, "loss": 2.4001, "step": 1764 }, { "epoch": 0.5652522017614091, "grad_norm": 0.50390625, "learning_rate": 2.3672521619149322e-05, "loss": 2.3446, "step": 1765 }, { "epoch": 0.5655724579663731, "grad_norm": 0.58203125, "learning_rate": 2.364460183719005e-05, "loss": 2.1358, "step": 1766 }, { "epoch": 0.565892714171337, "grad_norm": 0.57421875, "learning_rate": 2.3616683750595394e-05, "loss": 2.2924, "step": 1767 }, { "epoch": 0.566212970376301, "grad_norm": 0.494140625, "learning_rate": 2.3588767394285963e-05, "loss": 1.6567, "step": 1768 }, { "epoch": 0.566533226581265, "grad_norm": 0.462890625, "learning_rate": 2.356085280318021e-05, "loss": 2.1492, "step": 1769 }, { "epoch": 0.5668534827862289, "grad_norm": 0.42578125, "learning_rate": 2.3532940012194383e-05, "loss": 2.6153, "step": 1770 }, { "epoch": 0.5671737389911929, "grad_norm": 0.490234375, "learning_rate": 2.3505029056242454e-05, "loss": 2.1282, "step": 1771 }, { "epoch": 0.567493995196157, "grad_norm": 0.5546875, "learning_rate": 2.347711997023614e-05, "loss": 2.281, "step": 1772 }, { "epoch": 0.5678142514011209, "grad_norm": 0.5859375, "learning_rate": 2.344921278908479e-05, "loss": 2.5034, "step": 1773 }, { "epoch": 0.5681345076060849, "grad_norm": 0.6640625, "learning_rate": 2.342130754769536e-05, "loss": 2.2516, "step": 1774 }, { "epoch": 0.5684547638110489, "grad_norm": 0.484375, "learning_rate": 2.3393404280972426e-05, "loss": 2.1612, "step": 1775 }, { "epoch": 0.5687750200160128, "grad_norm": 0.470703125, "learning_rate": 2.336550302381804e-05, "loss": 2.4927, "step": 1776 }, { "epoch": 0.5690952762209768, "grad_norm": 0.47265625, "learning_rate": 2.333760381113178e-05, "loss": 2.2154, "step": 1777 }, { "epoch": 0.5694155324259408, "grad_norm": 0.48046875, "learning_rate": 2.330970667781065e-05, "loss": 1.8648, "step": 1778 }, { "epoch": 0.5697357886309047, "grad_norm": 0.5859375, "learning_rate": 2.328181165874905e-05, "loss": 2.3315, "step": 1779 }, { "epoch": 0.5700560448358687, "grad_norm": 0.46484375, "learning_rate": 2.325391878883874e-05, "loss": 2.1458, "step": 1780 }, { "epoch": 0.5703763010408327, "grad_norm": 0.439453125, "learning_rate": 2.3226028102968806e-05, "loss": 1.9153, "step": 1781 }, { "epoch": 0.5706965572457966, "grad_norm": 0.59375, "learning_rate": 2.3198139636025565e-05, "loss": 1.8971, "step": 1782 }, { "epoch": 0.5710168134507606, "grad_norm": 0.53125, "learning_rate": 2.3170253422892606e-05, "loss": 2.1762, "step": 1783 }, { "epoch": 0.5713370696557246, "grad_norm": 0.43359375, "learning_rate": 2.3142369498450665e-05, "loss": 1.8368, "step": 1784 }, { "epoch": 0.5716573258606885, "grad_norm": 0.423828125, "learning_rate": 2.3114487897577614e-05, "loss": 2.1639, "step": 1785 }, { "epoch": 0.5719775820656525, "grad_norm": 0.58203125, "learning_rate": 2.3086608655148447e-05, "loss": 2.3979, "step": 1786 }, { "epoch": 0.5722978382706165, "grad_norm": 0.427734375, "learning_rate": 2.305873180603519e-05, "loss": 1.7269, "step": 1787 }, { "epoch": 0.5726180944755804, "grad_norm": 0.484375, "learning_rate": 2.303085738510686e-05, "loss": 2.2162, "step": 1788 }, { "epoch": 0.5729383506805444, "grad_norm": 0.392578125, "learning_rate": 2.3002985427229472e-05, "loss": 2.1375, "step": 1789 }, { "epoch": 0.5732586068855084, "grad_norm": 0.58984375, "learning_rate": 2.2975115967265924e-05, "loss": 1.917, "step": 1790 }, { "epoch": 0.5735788630904723, "grad_norm": 0.4140625, "learning_rate": 2.294724904007602e-05, "loss": 2.2711, "step": 1791 }, { "epoch": 0.5738991192954364, "grad_norm": 0.490234375, "learning_rate": 2.2919384680516382e-05, "loss": 2.324, "step": 1792 }, { "epoch": 0.5742193755004004, "grad_norm": 0.8671875, "learning_rate": 2.2891522923440408e-05, "loss": 2.2216, "step": 1793 }, { "epoch": 0.5745396317053643, "grad_norm": 0.58203125, "learning_rate": 2.286366380369827e-05, "loss": 1.962, "step": 1794 }, { "epoch": 0.5748598879103283, "grad_norm": 0.53125, "learning_rate": 2.2835807356136817e-05, "loss": 2.2255, "step": 1795 }, { "epoch": 0.5751801441152923, "grad_norm": 0.53125, "learning_rate": 2.2807953615599552e-05, "loss": 1.8936, "step": 1796 }, { "epoch": 0.5755004003202562, "grad_norm": 0.50390625, "learning_rate": 2.2780102616926633e-05, "loss": 1.6788, "step": 1797 }, { "epoch": 0.5758206565252202, "grad_norm": 0.5390625, "learning_rate": 2.2752254394954736e-05, "loss": 2.1119, "step": 1798 }, { "epoch": 0.5761409127301842, "grad_norm": 0.58203125, "learning_rate": 2.272440898451708e-05, "loss": 2.2388, "step": 1799 }, { "epoch": 0.5764611689351481, "grad_norm": 0.5078125, "learning_rate": 2.2696566420443403e-05, "loss": 2.1596, "step": 1800 }, { "epoch": 0.5767814251401121, "grad_norm": 0.4765625, "learning_rate": 2.266872673755983e-05, "loss": 2.5837, "step": 1801 }, { "epoch": 0.5771016813450761, "grad_norm": 0.5859375, "learning_rate": 2.2640889970688904e-05, "loss": 1.9443, "step": 1802 }, { "epoch": 0.57742193755004, "grad_norm": 0.5546875, "learning_rate": 2.2613056154649542e-05, "loss": 2.2111, "step": 1803 }, { "epoch": 0.577742193755004, "grad_norm": 0.5703125, "learning_rate": 2.2585225324256928e-05, "loss": 2.4131, "step": 1804 }, { "epoch": 0.578062449959968, "grad_norm": 0.5546875, "learning_rate": 2.255739751432256e-05, "loss": 2.4771, "step": 1805 }, { "epoch": 0.5783827061649319, "grad_norm": 0.458984375, "learning_rate": 2.252957275965411e-05, "loss": 2.0832, "step": 1806 }, { "epoch": 0.5787029623698959, "grad_norm": 0.6171875, "learning_rate": 2.2501751095055448e-05, "loss": 2.2442, "step": 1807 }, { "epoch": 0.5790232185748598, "grad_norm": 0.41015625, "learning_rate": 2.2473932555326604e-05, "loss": 2.3033, "step": 1808 }, { "epoch": 0.5793434747798238, "grad_norm": 0.423828125, "learning_rate": 2.2446117175263658e-05, "loss": 2.1901, "step": 1809 }, { "epoch": 0.5796637309847879, "grad_norm": 3.03125, "learning_rate": 2.2418304989658757e-05, "loss": 2.1636, "step": 1810 }, { "epoch": 0.5799839871897517, "grad_norm": 0.416015625, "learning_rate": 2.239049603330006e-05, "loss": 2.5267, "step": 1811 }, { "epoch": 0.5803042433947158, "grad_norm": 0.443359375, "learning_rate": 2.2362690340971678e-05, "loss": 1.9348, "step": 1812 }, { "epoch": 0.5806244995996798, "grad_norm": 1.0234375, "learning_rate": 2.233488794745362e-05, "loss": 2.3273, "step": 1813 }, { "epoch": 0.5809447558046437, "grad_norm": 0.69921875, "learning_rate": 2.2307088887521816e-05, "loss": 1.9374, "step": 1814 }, { "epoch": 0.5812650120096077, "grad_norm": 0.5234375, "learning_rate": 2.2279293195947977e-05, "loss": 2.5428, "step": 1815 }, { "epoch": 0.5815852682145717, "grad_norm": 0.43359375, "learning_rate": 2.225150090749964e-05, "loss": 1.9329, "step": 1816 }, { "epoch": 0.5819055244195356, "grad_norm": 0.484375, "learning_rate": 2.222371205694006e-05, "loss": 2.0871, "step": 1817 }, { "epoch": 0.5822257806244996, "grad_norm": 0.78125, "learning_rate": 2.219592667902819e-05, "loss": 2.1848, "step": 1818 }, { "epoch": 0.5825460368294636, "grad_norm": 0.51171875, "learning_rate": 2.2168144808518667e-05, "loss": 1.9535, "step": 1819 }, { "epoch": 0.5828662930344275, "grad_norm": 0.447265625, "learning_rate": 2.2140366480161713e-05, "loss": 2.2323, "step": 1820 }, { "epoch": 0.5831865492393915, "grad_norm": 0.63671875, "learning_rate": 2.211259172870312e-05, "loss": 2.3016, "step": 1821 }, { "epoch": 0.5835068054443555, "grad_norm": 0.392578125, "learning_rate": 2.2084820588884234e-05, "loss": 1.6852, "step": 1822 }, { "epoch": 0.5838270616493194, "grad_norm": 2.328125, "learning_rate": 2.2057053095441854e-05, "loss": 2.3596, "step": 1823 }, { "epoch": 0.5841473178542834, "grad_norm": 0.392578125, "learning_rate": 2.202928928310822e-05, "loss": 1.8076, "step": 1824 }, { "epoch": 0.5844675740592474, "grad_norm": 0.50390625, "learning_rate": 2.2001529186610996e-05, "loss": 2.6958, "step": 1825 }, { "epoch": 0.5847878302642113, "grad_norm": 0.77734375, "learning_rate": 2.197377284067317e-05, "loss": 2.3731, "step": 1826 }, { "epoch": 0.5851080864691753, "grad_norm": 0.53125, "learning_rate": 2.194602028001303e-05, "loss": 2.4798, "step": 1827 }, { "epoch": 0.5854283426741393, "grad_norm": 0.5625, "learning_rate": 2.1918271539344177e-05, "loss": 2.7211, "step": 1828 }, { "epoch": 0.5857485988791032, "grad_norm": 0.482421875, "learning_rate": 2.1890526653375372e-05, "loss": 2.6734, "step": 1829 }, { "epoch": 0.5860688550840673, "grad_norm": 0.62890625, "learning_rate": 2.186278565681061e-05, "loss": 2.3266, "step": 1830 }, { "epoch": 0.5863891112890313, "grad_norm": 1.1015625, "learning_rate": 2.183504858434899e-05, "loss": 2.572, "step": 1831 }, { "epoch": 0.5867093674939952, "grad_norm": 0.59765625, "learning_rate": 2.1807315470684697e-05, "loss": 2.4207, "step": 1832 }, { "epoch": 0.5870296236989592, "grad_norm": 0.74609375, "learning_rate": 2.1779586350506996e-05, "loss": 2.1545, "step": 1833 }, { "epoch": 0.5873498799039232, "grad_norm": 0.447265625, "learning_rate": 2.175186125850012e-05, "loss": 2.287, "step": 1834 }, { "epoch": 0.5876701361088871, "grad_norm": 0.333984375, "learning_rate": 2.1724140229343288e-05, "loss": 2.3484, "step": 1835 }, { "epoch": 0.5879903923138511, "grad_norm": 0.478515625, "learning_rate": 2.169642329771063e-05, "loss": 2.0906, "step": 1836 }, { "epoch": 0.5883106485188151, "grad_norm": 0.51953125, "learning_rate": 2.1668710498271162e-05, "loss": 2.4591, "step": 1837 }, { "epoch": 0.588630904723779, "grad_norm": 0.52734375, "learning_rate": 2.1641001865688686e-05, "loss": 1.9441, "step": 1838 }, { "epoch": 0.588951160928743, "grad_norm": 0.51171875, "learning_rate": 2.1613297434621865e-05, "loss": 2.4205, "step": 1839 }, { "epoch": 0.589271417133707, "grad_norm": 0.6484375, "learning_rate": 2.158559723972404e-05, "loss": 2.1306, "step": 1840 }, { "epoch": 0.5895916733386709, "grad_norm": 0.400390625, "learning_rate": 2.15579013156433e-05, "loss": 1.831, "step": 1841 }, { "epoch": 0.5899119295436349, "grad_norm": 0.5546875, "learning_rate": 2.1530209697022366e-05, "loss": 1.9654, "step": 1842 }, { "epoch": 0.5902321857485989, "grad_norm": 0.5078125, "learning_rate": 2.150252241849857e-05, "loss": 2.4949, "step": 1843 }, { "epoch": 0.5905524419535628, "grad_norm": 0.64453125, "learning_rate": 2.1474839514703843e-05, "loss": 1.785, "step": 1844 }, { "epoch": 0.5908726981585268, "grad_norm": 0.5078125, "learning_rate": 2.1447161020264616e-05, "loss": 1.4549, "step": 1845 }, { "epoch": 0.5911929543634908, "grad_norm": 0.54296875, "learning_rate": 2.141948696980181e-05, "loss": 2.0587, "step": 1846 }, { "epoch": 0.5915132105684547, "grad_norm": 0.54296875, "learning_rate": 2.1391817397930814e-05, "loss": 1.8139, "step": 1847 }, { "epoch": 0.5918334667734187, "grad_norm": 0.458984375, "learning_rate": 2.1364152339261374e-05, "loss": 2.3478, "step": 1848 }, { "epoch": 0.5921537229783828, "grad_norm": 0.90625, "learning_rate": 2.133649182839761e-05, "loss": 1.9641, "step": 1849 }, { "epoch": 0.5924739791833467, "grad_norm": 0.52734375, "learning_rate": 2.1308835899937972e-05, "loss": 2.3559, "step": 1850 }, { "epoch": 0.5927942353883107, "grad_norm": 0.546875, "learning_rate": 2.128118458847515e-05, "loss": 2.3271, "step": 1851 }, { "epoch": 0.5931144915932747, "grad_norm": 0.478515625, "learning_rate": 2.1253537928596057e-05, "loss": 2.0447, "step": 1852 }, { "epoch": 0.5934347477982386, "grad_norm": 0.50390625, "learning_rate": 2.1225895954881823e-05, "loss": 2.2835, "step": 1853 }, { "epoch": 0.5937550040032026, "grad_norm": 0.53515625, "learning_rate": 2.119825870190767e-05, "loss": 1.9083, "step": 1854 }, { "epoch": 0.5940752602081665, "grad_norm": 0.5234375, "learning_rate": 2.1170626204242962e-05, "loss": 2.0059, "step": 1855 }, { "epoch": 0.5943955164131305, "grad_norm": 0.6015625, "learning_rate": 2.114299849645108e-05, "loss": 2.408, "step": 1856 }, { "epoch": 0.5947157726180945, "grad_norm": 0.55859375, "learning_rate": 2.1115375613089416e-05, "loss": 2.1296, "step": 1857 }, { "epoch": 0.5950360288230584, "grad_norm": 0.5703125, "learning_rate": 2.1087757588709356e-05, "loss": 2.2477, "step": 1858 }, { "epoch": 0.5953562850280224, "grad_norm": 0.578125, "learning_rate": 2.106014445785618e-05, "loss": 2.5451, "step": 1859 }, { "epoch": 0.5956765412329864, "grad_norm": 0.5, "learning_rate": 2.1032536255069045e-05, "loss": 1.9078, "step": 1860 }, { "epoch": 0.5959967974379503, "grad_norm": 0.5859375, "learning_rate": 2.1004933014880976e-05, "loss": 2.3821, "step": 1861 }, { "epoch": 0.5963170536429143, "grad_norm": 0.310546875, "learning_rate": 2.097733477181876e-05, "loss": 1.7438, "step": 1862 }, { "epoch": 0.5966373098478783, "grad_norm": 0.61328125, "learning_rate": 2.094974156040293e-05, "loss": 1.9844, "step": 1863 }, { "epoch": 0.5969575660528422, "grad_norm": 0.57421875, "learning_rate": 2.092215341514776e-05, "loss": 2.3406, "step": 1864 }, { "epoch": 0.5972778222578062, "grad_norm": 0.474609375, "learning_rate": 2.0894570370561156e-05, "loss": 2.2888, "step": 1865 }, { "epoch": 0.5975980784627702, "grad_norm": 0.5234375, "learning_rate": 2.0866992461144645e-05, "loss": 2.4423, "step": 1866 }, { "epoch": 0.5979183346677341, "grad_norm": 0.490234375, "learning_rate": 2.0839419721393354e-05, "loss": 2.4976, "step": 1867 }, { "epoch": 0.5982385908726982, "grad_norm": 0.431640625, "learning_rate": 2.0811852185795916e-05, "loss": 2.1009, "step": 1868 }, { "epoch": 0.5985588470776622, "grad_norm": 0.41015625, "learning_rate": 2.078428988883448e-05, "loss": 2.1664, "step": 1869 }, { "epoch": 0.5988791032826261, "grad_norm": 0.59765625, "learning_rate": 2.075673286498463e-05, "loss": 2.3979, "step": 1870 }, { "epoch": 0.5991993594875901, "grad_norm": 0.451171875, "learning_rate": 2.0729181148715332e-05, "loss": 1.8152, "step": 1871 }, { "epoch": 0.5995196156925541, "grad_norm": 0.494140625, "learning_rate": 2.070163477448896e-05, "loss": 2.2508, "step": 1872 }, { "epoch": 0.599839871897518, "grad_norm": 0.4921875, "learning_rate": 2.0674093776761176e-05, "loss": 2.1103, "step": 1873 }, { "epoch": 0.600160128102482, "grad_norm": 0.5078125, "learning_rate": 2.064655818998091e-05, "loss": 2.4928, "step": 1874 }, { "epoch": 0.600480384307446, "grad_norm": 0.58203125, "learning_rate": 2.061902804859035e-05, "loss": 2.1205, "step": 1875 }, { "epoch": 0.6008006405124099, "grad_norm": 0.59375, "learning_rate": 2.0591503387024853e-05, "loss": 2.2769, "step": 1876 }, { "epoch": 0.6011208967173739, "grad_norm": 0.57421875, "learning_rate": 2.056398423971292e-05, "loss": 2.2979, "step": 1877 }, { "epoch": 0.6014411529223379, "grad_norm": 0.318359375, "learning_rate": 2.053647064107618e-05, "loss": 1.9518, "step": 1878 }, { "epoch": 0.6017614091273018, "grad_norm": 0.46875, "learning_rate": 2.0508962625529284e-05, "loss": 2.1684, "step": 1879 }, { "epoch": 0.6020816653322658, "grad_norm": 0.51171875, "learning_rate": 2.0481460227479942e-05, "loss": 2.4742, "step": 1880 }, { "epoch": 0.6024019215372298, "grad_norm": 0.439453125, "learning_rate": 2.0453963481328796e-05, "loss": 2.1223, "step": 1881 }, { "epoch": 0.6027221777421937, "grad_norm": 0.55078125, "learning_rate": 2.042647242146944e-05, "loss": 1.8289, "step": 1882 }, { "epoch": 0.6030424339471577, "grad_norm": 0.3828125, "learning_rate": 2.0398987082288363e-05, "loss": 2.206, "step": 1883 }, { "epoch": 0.6033626901521217, "grad_norm": 0.431640625, "learning_rate": 2.037150749816488e-05, "loss": 1.9582, "step": 1884 }, { "epoch": 0.6036829463570856, "grad_norm": 0.56640625, "learning_rate": 2.0344033703471106e-05, "loss": 2.1088, "step": 1885 }, { "epoch": 0.6040032025620496, "grad_norm": 0.359375, "learning_rate": 2.0316565732571947e-05, "loss": 1.6995, "step": 1886 }, { "epoch": 0.6043234587670137, "grad_norm": 0.63671875, "learning_rate": 2.028910361982499e-05, "loss": 2.3343, "step": 1887 }, { "epoch": 0.6046437149719776, "grad_norm": 0.44140625, "learning_rate": 2.0261647399580493e-05, "loss": 2.5429, "step": 1888 }, { "epoch": 0.6049639711769416, "grad_norm": 0.48046875, "learning_rate": 2.023419710618138e-05, "loss": 2.6612, "step": 1889 }, { "epoch": 0.6052842273819056, "grad_norm": 1.140625, "learning_rate": 2.0206752773963128e-05, "loss": 2.0385, "step": 1890 }, { "epoch": 0.6056044835868695, "grad_norm": 0.462890625, "learning_rate": 2.0179314437253756e-05, "loss": 1.9074, "step": 1891 }, { "epoch": 0.6059247397918335, "grad_norm": 0.54296875, "learning_rate": 2.015188213037381e-05, "loss": 2.1805, "step": 1892 }, { "epoch": 0.6062449959967975, "grad_norm": 0.4765625, "learning_rate": 2.012445588763627e-05, "loss": 2.714, "step": 1893 }, { "epoch": 0.6065652522017614, "grad_norm": 0.498046875, "learning_rate": 2.009703574334656e-05, "loss": 2.5523, "step": 1894 }, { "epoch": 0.6068855084067254, "grad_norm": 0.451171875, "learning_rate": 2.0069621731802435e-05, "loss": 2.4568, "step": 1895 }, { "epoch": 0.6072057646116894, "grad_norm": 0.388671875, "learning_rate": 2.0042213887293994e-05, "loss": 2.021, "step": 1896 }, { "epoch": 0.6075260208166533, "grad_norm": 1.1484375, "learning_rate": 2.0014812244103653e-05, "loss": 2.636, "step": 1897 }, { "epoch": 0.6078462770216173, "grad_norm": 0.498046875, "learning_rate": 1.9987416836506034e-05, "loss": 1.8559, "step": 1898 }, { "epoch": 0.6081665332265813, "grad_norm": 0.53515625, "learning_rate": 1.9960027698767957e-05, "loss": 1.9964, "step": 1899 }, { "epoch": 0.6084867894315452, "grad_norm": 0.4453125, "learning_rate": 1.993264486514844e-05, "loss": 2.045, "step": 1900 }, { "epoch": 0.6088070456365092, "grad_norm": 1.140625, "learning_rate": 1.9905268369898576e-05, "loss": 2.0631, "step": 1901 }, { "epoch": 0.6091273018414731, "grad_norm": 0.470703125, "learning_rate": 1.9877898247261534e-05, "loss": 1.8805, "step": 1902 }, { "epoch": 0.6094475580464371, "grad_norm": 0.5234375, "learning_rate": 1.9850534531472546e-05, "loss": 2.0472, "step": 1903 }, { "epoch": 0.6097678142514011, "grad_norm": 0.474609375, "learning_rate": 1.9823177256758787e-05, "loss": 2.5672, "step": 1904 }, { "epoch": 0.610088070456365, "grad_norm": 0.431640625, "learning_rate": 1.9795826457339393e-05, "loss": 1.8286, "step": 1905 }, { "epoch": 0.610408326661329, "grad_norm": 0.55859375, "learning_rate": 1.9768482167425413e-05, "loss": 1.6787, "step": 1906 }, { "epoch": 0.6107285828662931, "grad_norm": 0.5625, "learning_rate": 1.974114442121973e-05, "loss": 2.3295, "step": 1907 }, { "epoch": 0.611048839071257, "grad_norm": 0.609375, "learning_rate": 1.9713813252917075e-05, "loss": 2.6286, "step": 1908 }, { "epoch": 0.611369095276221, "grad_norm": 0.6796875, "learning_rate": 1.9686488696703914e-05, "loss": 2.2819, "step": 1909 }, { "epoch": 0.611689351481185, "grad_norm": 0.375, "learning_rate": 1.965917078675845e-05, "loss": 2.2437, "step": 1910 }, { "epoch": 0.6120096076861489, "grad_norm": 0.50390625, "learning_rate": 1.9631859557250605e-05, "loss": 2.1633, "step": 1911 }, { "epoch": 0.6123298638911129, "grad_norm": 0.51953125, "learning_rate": 1.960455504234191e-05, "loss": 2.1793, "step": 1912 }, { "epoch": 0.6126501200960769, "grad_norm": 0.515625, "learning_rate": 1.95772572761855e-05, "loss": 2.4874, "step": 1913 }, { "epoch": 0.6129703763010408, "grad_norm": 0.53515625, "learning_rate": 1.9549966292926095e-05, "loss": 2.1151, "step": 1914 }, { "epoch": 0.6132906325060048, "grad_norm": 0.482421875, "learning_rate": 1.9522682126699903e-05, "loss": 2.0204, "step": 1915 }, { "epoch": 0.6136108887109688, "grad_norm": 0.8671875, "learning_rate": 1.949540481163461e-05, "loss": 2.5082, "step": 1916 }, { "epoch": 0.6139311449159327, "grad_norm": 0.41796875, "learning_rate": 1.9468134381849355e-05, "loss": 2.2812, "step": 1917 }, { "epoch": 0.6142514011208967, "grad_norm": 0.515625, "learning_rate": 1.944087087145462e-05, "loss": 2.2605, "step": 1918 }, { "epoch": 0.6145716573258607, "grad_norm": 1.9921875, "learning_rate": 1.9413614314552293e-05, "loss": 1.7432, "step": 1919 }, { "epoch": 0.6148919135308246, "grad_norm": 0.74609375, "learning_rate": 1.938636474523551e-05, "loss": 2.4423, "step": 1920 }, { "epoch": 0.6152121697357886, "grad_norm": 0.69140625, "learning_rate": 1.935912219758868e-05, "loss": 2.3166, "step": 1921 }, { "epoch": 0.6155324259407526, "grad_norm": 0.255859375, "learning_rate": 1.9331886705687464e-05, "loss": 1.7659, "step": 1922 }, { "epoch": 0.6158526821457165, "grad_norm": 0.59375, "learning_rate": 1.9304658303598648e-05, "loss": 2.425, "step": 1923 }, { "epoch": 0.6161729383506805, "grad_norm": 0.54296875, "learning_rate": 1.927743702538017e-05, "loss": 1.8974, "step": 1924 }, { "epoch": 0.6164931945556446, "grad_norm": 0.36328125, "learning_rate": 1.925022290508108e-05, "loss": 1.9298, "step": 1925 }, { "epoch": 0.6168134507606085, "grad_norm": 4.03125, "learning_rate": 1.922301597674144e-05, "loss": 2.4168, "step": 1926 }, { "epoch": 0.6171337069655725, "grad_norm": 0.4296875, "learning_rate": 1.919581627439232e-05, "loss": 2.331, "step": 1927 }, { "epoch": 0.6174539631705365, "grad_norm": 0.54296875, "learning_rate": 1.9168623832055786e-05, "loss": 1.7454, "step": 1928 }, { "epoch": 0.6177742193755004, "grad_norm": 0.79296875, "learning_rate": 1.9141438683744785e-05, "loss": 2.4372, "step": 1929 }, { "epoch": 0.6180944755804644, "grad_norm": 0.5390625, "learning_rate": 1.9114260863463145e-05, "loss": 2.159, "step": 1930 }, { "epoch": 0.6184147317854284, "grad_norm": 0.48828125, "learning_rate": 1.908709040520556e-05, "loss": 2.2052, "step": 1931 }, { "epoch": 0.6187349879903923, "grad_norm": 1.7734375, "learning_rate": 1.9059927342957468e-05, "loss": 2.6305, "step": 1932 }, { "epoch": 0.6190552441953563, "grad_norm": 0.625, "learning_rate": 1.9032771710695106e-05, "loss": 2.5403, "step": 1933 }, { "epoch": 0.6193755004003203, "grad_norm": 0.578125, "learning_rate": 1.9005623542385377e-05, "loss": 1.5766, "step": 1934 }, { "epoch": 0.6196957566052842, "grad_norm": 0.46875, "learning_rate": 1.8978482871985852e-05, "loss": 2.1659, "step": 1935 }, { "epoch": 0.6200160128102482, "grad_norm": 0.6015625, "learning_rate": 1.895134973344476e-05, "loss": 1.9693, "step": 1936 }, { "epoch": 0.6203362690152122, "grad_norm": 1.2265625, "learning_rate": 1.8924224160700866e-05, "loss": 2.5216, "step": 1937 }, { "epoch": 0.6206565252201761, "grad_norm": 0.404296875, "learning_rate": 1.8897106187683482e-05, "loss": 1.6666, "step": 1938 }, { "epoch": 0.6209767814251401, "grad_norm": 0.59375, "learning_rate": 1.8869995848312445e-05, "loss": 2.3252, "step": 1939 }, { "epoch": 0.6212970376301041, "grad_norm": 0.5, "learning_rate": 1.8842893176498003e-05, "loss": 2.1655, "step": 1940 }, { "epoch": 0.621617293835068, "grad_norm": 0.67578125, "learning_rate": 1.881579820614082e-05, "loss": 2.7821, "step": 1941 }, { "epoch": 0.621937550040032, "grad_norm": 0.56640625, "learning_rate": 1.878871097113196e-05, "loss": 2.2319, "step": 1942 }, { "epoch": 0.622257806244996, "grad_norm": 0.9296875, "learning_rate": 1.876163150535277e-05, "loss": 2.2277, "step": 1943 }, { "epoch": 0.62257806244996, "grad_norm": 0.51953125, "learning_rate": 1.8734559842674908e-05, "loss": 1.8544, "step": 1944 }, { "epoch": 0.622898318654924, "grad_norm": 0.50390625, "learning_rate": 1.8707496016960262e-05, "loss": 2.4545, "step": 1945 }, { "epoch": 0.623218574859888, "grad_norm": 0.73046875, "learning_rate": 1.8680440062060895e-05, "loss": 2.2949, "step": 1946 }, { "epoch": 0.6235388310648519, "grad_norm": 0.474609375, "learning_rate": 1.8653392011819072e-05, "loss": 2.4566, "step": 1947 }, { "epoch": 0.6238590872698159, "grad_norm": 0.294921875, "learning_rate": 1.8626351900067135e-05, "loss": 1.9222, "step": 1948 }, { "epoch": 0.6241793434747798, "grad_norm": 0.462890625, "learning_rate": 1.8599319760627494e-05, "loss": 2.7594, "step": 1949 }, { "epoch": 0.6244995996797438, "grad_norm": 0.6875, "learning_rate": 1.8572295627312616e-05, "loss": 2.2509, "step": 1950 }, { "epoch": 0.6248198558847078, "grad_norm": 0.51953125, "learning_rate": 1.8545279533924934e-05, "loss": 2.6185, "step": 1951 }, { "epoch": 0.6251401120896717, "grad_norm": 0.5390625, "learning_rate": 1.8518271514256812e-05, "loss": 2.1968, "step": 1952 }, { "epoch": 0.6254603682946357, "grad_norm": 1.640625, "learning_rate": 1.8491271602090552e-05, "loss": 2.1392, "step": 1953 }, { "epoch": 0.6257806244995997, "grad_norm": 0.54296875, "learning_rate": 1.8464279831198288e-05, "loss": 2.3392, "step": 1954 }, { "epoch": 0.6261008807045636, "grad_norm": 0.89453125, "learning_rate": 1.8437296235341954e-05, "loss": 2.3781, "step": 1955 }, { "epoch": 0.6264211369095276, "grad_norm": 0.6328125, "learning_rate": 1.8410320848273315e-05, "loss": 2.1818, "step": 1956 }, { "epoch": 0.6267413931144916, "grad_norm": 0.36328125, "learning_rate": 1.8383353703733803e-05, "loss": 2.3265, "step": 1957 }, { "epoch": 0.6270616493194555, "grad_norm": 0.67578125, "learning_rate": 1.8356394835454596e-05, "loss": 2.1143, "step": 1958 }, { "epoch": 0.6273819055244195, "grad_norm": 0.427734375, "learning_rate": 1.8329444277156483e-05, "loss": 1.686, "step": 1959 }, { "epoch": 0.6277021617293835, "grad_norm": 0.474609375, "learning_rate": 1.8302502062549858e-05, "loss": 2.2293, "step": 1960 }, { "epoch": 0.6280224179343474, "grad_norm": 0.462890625, "learning_rate": 1.8275568225334722e-05, "loss": 2.3718, "step": 1961 }, { "epoch": 0.6283426741393114, "grad_norm": 0.62890625, "learning_rate": 1.824864279920054e-05, "loss": 2.0935, "step": 1962 }, { "epoch": 0.6286629303442755, "grad_norm": 0.48828125, "learning_rate": 1.8221725817826286e-05, "loss": 2.1848, "step": 1963 }, { "epoch": 0.6289831865492393, "grad_norm": 0.5625, "learning_rate": 1.819481731488038e-05, "loss": 2.5737, "step": 1964 }, { "epoch": 0.6293034427542034, "grad_norm": 0.59765625, "learning_rate": 1.8167917324020615e-05, "loss": 2.2572, "step": 1965 }, { "epoch": 0.6296236989591674, "grad_norm": 0.466796875, "learning_rate": 1.814102587889414e-05, "loss": 2.2202, "step": 1966 }, { "epoch": 0.6299439551641313, "grad_norm": 1.921875, "learning_rate": 1.8114143013137434e-05, "loss": 2.5762, "step": 1967 }, { "epoch": 0.6302642113690953, "grad_norm": 0.50390625, "learning_rate": 1.8087268760376237e-05, "loss": 2.2381, "step": 1968 }, { "epoch": 0.6305844675740593, "grad_norm": 0.5390625, "learning_rate": 1.806040315422548e-05, "loss": 2.179, "step": 1969 }, { "epoch": 0.6309047237790232, "grad_norm": 0.5625, "learning_rate": 1.8033546228289347e-05, "loss": 2.1213, "step": 1970 }, { "epoch": 0.6312249799839872, "grad_norm": 0.462890625, "learning_rate": 1.8006698016161097e-05, "loss": 1.9795, "step": 1971 }, { "epoch": 0.6315452361889512, "grad_norm": 0.5234375, "learning_rate": 1.7979858551423147e-05, "loss": 2.1043, "step": 1972 }, { "epoch": 0.6318654923939151, "grad_norm": 0.546875, "learning_rate": 1.795302786764692e-05, "loss": 1.9355, "step": 1973 }, { "epoch": 0.6321857485988791, "grad_norm": 0.55859375, "learning_rate": 1.7926205998392888e-05, "loss": 2.4791, "step": 1974 }, { "epoch": 0.6325060048038431, "grad_norm": 0.482421875, "learning_rate": 1.7899392977210498e-05, "loss": 2.5029, "step": 1975 }, { "epoch": 0.632826261008807, "grad_norm": 0.79296875, "learning_rate": 1.787258883763812e-05, "loss": 2.1131, "step": 1976 }, { "epoch": 0.633146517213771, "grad_norm": 0.451171875, "learning_rate": 1.7845793613203e-05, "loss": 2.3818, "step": 1977 }, { "epoch": 0.633466773418735, "grad_norm": 0.41015625, "learning_rate": 1.7819007337421272e-05, "loss": 2.3837, "step": 1978 }, { "epoch": 0.6337870296236989, "grad_norm": 0.9140625, "learning_rate": 1.7792230043797845e-05, "loss": 2.3031, "step": 1979 }, { "epoch": 0.6341072858286629, "grad_norm": 0.404296875, "learning_rate": 1.776546176582639e-05, "loss": 2.1035, "step": 1980 }, { "epoch": 0.634427542033627, "grad_norm": 0.5390625, "learning_rate": 1.773870253698933e-05, "loss": 2.4333, "step": 1981 }, { "epoch": 0.6347477982385908, "grad_norm": 0.421875, "learning_rate": 1.7711952390757736e-05, "loss": 2.0962, "step": 1982 }, { "epoch": 0.6350680544435549, "grad_norm": 0.8046875, "learning_rate": 1.7685211360591353e-05, "loss": 2.0404, "step": 1983 }, { "epoch": 0.6353883106485189, "grad_norm": 0.46875, "learning_rate": 1.7658479479938493e-05, "loss": 1.8208, "step": 1984 }, { "epoch": 0.6357085668534828, "grad_norm": 0.58203125, "learning_rate": 1.7631756782236024e-05, "loss": 1.9733, "step": 1985 }, { "epoch": 0.6360288230584468, "grad_norm": 0.484375, "learning_rate": 1.760504330090937e-05, "loss": 2.1959, "step": 1986 }, { "epoch": 0.6363490792634108, "grad_norm": 0.7421875, "learning_rate": 1.757833906937237e-05, "loss": 2.3309, "step": 1987 }, { "epoch": 0.6366693354683747, "grad_norm": 0.486328125, "learning_rate": 1.7551644121027317e-05, "loss": 2.45, "step": 1988 }, { "epoch": 0.6369895916733387, "grad_norm": 0.625, "learning_rate": 1.7524958489264913e-05, "loss": 2.2078, "step": 1989 }, { "epoch": 0.6373098478783027, "grad_norm": 0.56640625, "learning_rate": 1.7498282207464185e-05, "loss": 2.45, "step": 1990 }, { "epoch": 0.6376301040832666, "grad_norm": 0.44921875, "learning_rate": 1.7471615308992444e-05, "loss": 2.5175, "step": 1991 }, { "epoch": 0.6379503602882306, "grad_norm": 0.84765625, "learning_rate": 1.7444957827205318e-05, "loss": 1.9297, "step": 1992 }, { "epoch": 0.6382706164931946, "grad_norm": 0.546875, "learning_rate": 1.741830979544661e-05, "loss": 2.3512, "step": 1993 }, { "epoch": 0.6385908726981585, "grad_norm": 0.546875, "learning_rate": 1.7391671247048315e-05, "loss": 2.4583, "step": 1994 }, { "epoch": 0.6389111289031225, "grad_norm": 0.48828125, "learning_rate": 1.7365042215330584e-05, "loss": 2.1487, "step": 1995 }, { "epoch": 0.6392313851080864, "grad_norm": 0.455078125, "learning_rate": 1.7338422733601636e-05, "loss": 1.959, "step": 1996 }, { "epoch": 0.6395516413130504, "grad_norm": 0.484375, "learning_rate": 1.7311812835157774e-05, "loss": 2.6023, "step": 1997 }, { "epoch": 0.6398718975180144, "grad_norm": 0.546875, "learning_rate": 1.7285212553283295e-05, "loss": 2.2752, "step": 1998 }, { "epoch": 0.6401921537229783, "grad_norm": 0.515625, "learning_rate": 1.7258621921250454e-05, "loss": 1.882, "step": 1999 }, { "epoch": 0.6405124099279423, "grad_norm": 2.875, "learning_rate": 1.7232040972319474e-05, "loss": 2.0141, "step": 2000 }, { "epoch": 0.6408326661329063, "grad_norm": 0.375, "learning_rate": 1.7205469739738442e-05, "loss": 1.99, "step": 2001 }, { "epoch": 0.6411529223378702, "grad_norm": 0.31640625, "learning_rate": 1.7178908256743277e-05, "loss": 1.6528, "step": 2002 }, { "epoch": 0.6414731785428343, "grad_norm": 0.91796875, "learning_rate": 1.7152356556557745e-05, "loss": 2.0187, "step": 2003 }, { "epoch": 0.6417934347477983, "grad_norm": 0.31640625, "learning_rate": 1.7125814672393337e-05, "loss": 2.4054, "step": 2004 }, { "epoch": 0.6421136909527622, "grad_norm": 0.65625, "learning_rate": 1.7099282637449273e-05, "loss": 1.9533, "step": 2005 }, { "epoch": 0.6424339471577262, "grad_norm": 0.515625, "learning_rate": 1.7072760484912482e-05, "loss": 2.1352, "step": 2006 }, { "epoch": 0.6427542033626902, "grad_norm": 0.640625, "learning_rate": 1.70462482479575e-05, "loss": 2.5819, "step": 2007 }, { "epoch": 0.6430744595676541, "grad_norm": 0.48828125, "learning_rate": 1.701974595974645e-05, "loss": 2.4939, "step": 2008 }, { "epoch": 0.6433947157726181, "grad_norm": 0.578125, "learning_rate": 1.6993253653429063e-05, "loss": 2.0688, "step": 2009 }, { "epoch": 0.6437149719775821, "grad_norm": 0.5390625, "learning_rate": 1.6966771362142532e-05, "loss": 2.3022, "step": 2010 }, { "epoch": 0.644035228182546, "grad_norm": 0.466796875, "learning_rate": 1.694029911901156e-05, "loss": 2.4128, "step": 2011 }, { "epoch": 0.64435548438751, "grad_norm": 0.453125, "learning_rate": 1.691383695714826e-05, "loss": 1.9259, "step": 2012 }, { "epoch": 0.644675740592474, "grad_norm": 0.63671875, "learning_rate": 1.688738490965212e-05, "loss": 2.2305, "step": 2013 }, { "epoch": 0.6449959967974379, "grad_norm": 0.45703125, "learning_rate": 1.686094300961003e-05, "loss": 1.9657, "step": 2014 }, { "epoch": 0.6453162530024019, "grad_norm": 0.71875, "learning_rate": 1.683451129009614e-05, "loss": 2.2175, "step": 2015 }, { "epoch": 0.6456365092073659, "grad_norm": 0.71484375, "learning_rate": 1.6808089784171863e-05, "loss": 2.3302, "step": 2016 }, { "epoch": 0.6459567654123298, "grad_norm": 0.52734375, "learning_rate": 1.6781678524885882e-05, "loss": 1.7276, "step": 2017 }, { "epoch": 0.6462770216172938, "grad_norm": 0.3984375, "learning_rate": 1.6755277545274023e-05, "loss": 1.9889, "step": 2018 }, { "epoch": 0.6465972778222578, "grad_norm": 0.408203125, "learning_rate": 1.6728886878359258e-05, "loss": 2.0543, "step": 2019 }, { "epoch": 0.6469175340272217, "grad_norm": 0.427734375, "learning_rate": 1.670250655715168e-05, "loss": 2.6313, "step": 2020 }, { "epoch": 0.6472377902321858, "grad_norm": 0.609375, "learning_rate": 1.667613661464842e-05, "loss": 2.2489, "step": 2021 }, { "epoch": 0.6475580464371498, "grad_norm": 1.515625, "learning_rate": 1.664977708383365e-05, "loss": 2.7663, "step": 2022 }, { "epoch": 0.6478783026421137, "grad_norm": 0.55859375, "learning_rate": 1.6623427997678497e-05, "loss": 2.5957, "step": 2023 }, { "epoch": 0.6481985588470777, "grad_norm": 0.65234375, "learning_rate": 1.659708938914102e-05, "loss": 2.0438, "step": 2024 }, { "epoch": 0.6485188150520417, "grad_norm": 0.4765625, "learning_rate": 1.6570761291166207e-05, "loss": 2.1817, "step": 2025 }, { "epoch": 0.6488390712570056, "grad_norm": 0.478515625, "learning_rate": 1.654444373668586e-05, "loss": 2.3496, "step": 2026 }, { "epoch": 0.6491593274619696, "grad_norm": 0.546875, "learning_rate": 1.6518136758618607e-05, "loss": 2.2006, "step": 2027 }, { "epoch": 0.6494795836669336, "grad_norm": 0.5703125, "learning_rate": 1.649184038986986e-05, "loss": 1.8331, "step": 2028 }, { "epoch": 0.6497998398718975, "grad_norm": 0.51171875, "learning_rate": 1.6465554663331746e-05, "loss": 2.3697, "step": 2029 }, { "epoch": 0.6501200960768615, "grad_norm": 0.498046875, "learning_rate": 1.6439279611883075e-05, "loss": 2.2204, "step": 2030 }, { "epoch": 0.6504403522818255, "grad_norm": 1.3671875, "learning_rate": 1.6413015268389323e-05, "loss": 2.3094, "step": 2031 }, { "epoch": 0.6507606084867894, "grad_norm": 0.97265625, "learning_rate": 1.6386761665702564e-05, "loss": 2.6854, "step": 2032 }, { "epoch": 0.6510808646917534, "grad_norm": 0.498046875, "learning_rate": 1.6360518836661418e-05, "loss": 2.6873, "step": 2033 }, { "epoch": 0.6514011208967174, "grad_norm": 0.419921875, "learning_rate": 1.6334286814091072e-05, "loss": 2.2687, "step": 2034 }, { "epoch": 0.6517213771016813, "grad_norm": 0.337890625, "learning_rate": 1.6308065630803144e-05, "loss": 2.0965, "step": 2035 }, { "epoch": 0.6520416333066453, "grad_norm": 0.59375, "learning_rate": 1.6281855319595747e-05, "loss": 2.6411, "step": 2036 }, { "epoch": 0.6523618895116093, "grad_norm": 0.64453125, "learning_rate": 1.6255655913253357e-05, "loss": 2.0638, "step": 2037 }, { "epoch": 0.6526821457165732, "grad_norm": 0.546875, "learning_rate": 1.622946744454681e-05, "loss": 2.3375, "step": 2038 }, { "epoch": 0.6530024019215372, "grad_norm": 0.71875, "learning_rate": 1.620328994623329e-05, "loss": 2.0346, "step": 2039 }, { "epoch": 0.6533226581265013, "grad_norm": 0.390625, "learning_rate": 1.617712345105624e-05, "loss": 2.1022, "step": 2040 }, { "epoch": 0.6536429143314652, "grad_norm": 0.7265625, "learning_rate": 1.6150967991745323e-05, "loss": 2.2443, "step": 2041 }, { "epoch": 0.6539631705364292, "grad_norm": 0.515625, "learning_rate": 1.6124823601016437e-05, "loss": 1.8261, "step": 2042 }, { "epoch": 0.6542834267413931, "grad_norm": 0.58203125, "learning_rate": 1.6098690311571608e-05, "loss": 1.3938, "step": 2043 }, { "epoch": 0.6546036829463571, "grad_norm": 0.46484375, "learning_rate": 1.6072568156098972e-05, "loss": 2.2079, "step": 2044 }, { "epoch": 0.6549239391513211, "grad_norm": 0.265625, "learning_rate": 1.6046457167272773e-05, "loss": 1.7936, "step": 2045 }, { "epoch": 0.655244195356285, "grad_norm": 0.56640625, "learning_rate": 1.6020357377753237e-05, "loss": 2.3265, "step": 2046 }, { "epoch": 0.655564451561249, "grad_norm": 0.51953125, "learning_rate": 1.5994268820186624e-05, "loss": 2.1049, "step": 2047 }, { "epoch": 0.655884707766213, "grad_norm": 0.48046875, "learning_rate": 1.5968191527205134e-05, "loss": 2.2532, "step": 2048 }, { "epoch": 0.6562049639711769, "grad_norm": 0.56640625, "learning_rate": 1.5942125531426848e-05, "loss": 2.5682, "step": 2049 }, { "epoch": 0.6565252201761409, "grad_norm": 0.9375, "learning_rate": 1.591607086545577e-05, "loss": 2.6043, "step": 2050 }, { "epoch": 0.6568454763811049, "grad_norm": 0.82421875, "learning_rate": 1.589002756188169e-05, "loss": 1.8284, "step": 2051 }, { "epoch": 0.6571657325860688, "grad_norm": 0.6875, "learning_rate": 1.5863995653280187e-05, "loss": 2.0559, "step": 2052 }, { "epoch": 0.6574859887910328, "grad_norm": 0.578125, "learning_rate": 1.5837975172212615e-05, "loss": 2.0543, "step": 2053 }, { "epoch": 0.6578062449959968, "grad_norm": 0.57421875, "learning_rate": 1.5811966151226014e-05, "loss": 2.0767, "step": 2054 }, { "epoch": 0.6581265012009607, "grad_norm": 0.94921875, "learning_rate": 1.578596862285308e-05, "loss": 2.3316, "step": 2055 }, { "epoch": 0.6584467574059247, "grad_norm": 0.51953125, "learning_rate": 1.575998261961217e-05, "loss": 2.4229, "step": 2056 }, { "epoch": 0.6587670136108887, "grad_norm": 0.4765625, "learning_rate": 1.5734008174007182e-05, "loss": 2.3858, "step": 2057 }, { "epoch": 0.6590872698158526, "grad_norm": 0.451171875, "learning_rate": 1.5708045318527583e-05, "loss": 2.5705, "step": 2058 }, { "epoch": 0.6594075260208166, "grad_norm": 0.4140625, "learning_rate": 1.568209408564834e-05, "loss": 2.8622, "step": 2059 }, { "epoch": 0.6597277822257807, "grad_norm": 0.48046875, "learning_rate": 1.5656154507829867e-05, "loss": 2.4677, "step": 2060 }, { "epoch": 0.6600480384307446, "grad_norm": 0.478515625, "learning_rate": 1.5630226617518033e-05, "loss": 2.4751, "step": 2061 }, { "epoch": 0.6603682946357086, "grad_norm": 0.53515625, "learning_rate": 1.560431044714405e-05, "loss": 2.2698, "step": 2062 }, { "epoch": 0.6606885508406726, "grad_norm": 0.431640625, "learning_rate": 1.5578406029124485e-05, "loss": 2.0726, "step": 2063 }, { "epoch": 0.6610088070456365, "grad_norm": 0.49609375, "learning_rate": 1.555251339586123e-05, "loss": 1.9642, "step": 2064 }, { "epoch": 0.6613290632506005, "grad_norm": 1.421875, "learning_rate": 1.5526632579741386e-05, "loss": 1.9934, "step": 2065 }, { "epoch": 0.6616493194555645, "grad_norm": 1.65625, "learning_rate": 1.5500763613137316e-05, "loss": 2.7078, "step": 2066 }, { "epoch": 0.6619695756605284, "grad_norm": 0.66796875, "learning_rate": 1.5474906528406542e-05, "loss": 2.3785, "step": 2067 }, { "epoch": 0.6622898318654924, "grad_norm": 0.6875, "learning_rate": 1.544906135789172e-05, "loss": 2.1679, "step": 2068 }, { "epoch": 0.6626100880704564, "grad_norm": 0.515625, "learning_rate": 1.542322813392062e-05, "loss": 2.3335, "step": 2069 }, { "epoch": 0.6629303442754203, "grad_norm": 0.63671875, "learning_rate": 1.5397406888806063e-05, "loss": 2.1106, "step": 2070 }, { "epoch": 0.6632506004803843, "grad_norm": 0.5546875, "learning_rate": 1.537159765484587e-05, "loss": 1.8888, "step": 2071 }, { "epoch": 0.6635708566853483, "grad_norm": 0.48046875, "learning_rate": 1.5345800464322856e-05, "loss": 2.4113, "step": 2072 }, { "epoch": 0.6638911128903122, "grad_norm": 0.431640625, "learning_rate": 1.5320015349504778e-05, "loss": 1.6792, "step": 2073 }, { "epoch": 0.6642113690952762, "grad_norm": 0.70703125, "learning_rate": 1.529424234264426e-05, "loss": 2.4077, "step": 2074 }, { "epoch": 0.6645316253002402, "grad_norm": 2.109375, "learning_rate": 1.5268481475978817e-05, "loss": 2.5355, "step": 2075 }, { "epoch": 0.6648518815052041, "grad_norm": 0.82421875, "learning_rate": 1.5242732781730749e-05, "loss": 2.0814, "step": 2076 }, { "epoch": 0.6651721377101681, "grad_norm": 0.5703125, "learning_rate": 1.5216996292107144e-05, "loss": 2.5232, "step": 2077 }, { "epoch": 0.6654923939151322, "grad_norm": 0.54296875, "learning_rate": 1.5191272039299825e-05, "loss": 2.3116, "step": 2078 }, { "epoch": 0.665812650120096, "grad_norm": 0.51171875, "learning_rate": 1.5165560055485301e-05, "loss": 1.9626, "step": 2079 }, { "epoch": 0.6661329063250601, "grad_norm": 0.349609375, "learning_rate": 1.5139860372824741e-05, "loss": 1.8281, "step": 2080 }, { "epoch": 0.6664531625300241, "grad_norm": 0.470703125, "learning_rate": 1.5114173023463932e-05, "loss": 2.2899, "step": 2081 }, { "epoch": 0.666773418734988, "grad_norm": 0.46484375, "learning_rate": 1.5088498039533219e-05, "loss": 1.9186, "step": 2082 }, { "epoch": 0.667093674939952, "grad_norm": 0.48828125, "learning_rate": 1.5062835453147495e-05, "loss": 2.5786, "step": 2083 }, { "epoch": 0.667413931144916, "grad_norm": 0.63671875, "learning_rate": 1.5037185296406142e-05, "loss": 2.3842, "step": 2084 }, { "epoch": 0.6677341873498799, "grad_norm": 0.515625, "learning_rate": 1.5011547601392983e-05, "loss": 2.313, "step": 2085 }, { "epoch": 0.6680544435548439, "grad_norm": 0.578125, "learning_rate": 1.4985922400176278e-05, "loss": 2.6968, "step": 2086 }, { "epoch": 0.6683746997598079, "grad_norm": 0.455078125, "learning_rate": 1.4960309724808633e-05, "loss": 1.9038, "step": 2087 }, { "epoch": 0.6686949559647718, "grad_norm": 0.8359375, "learning_rate": 1.4934709607327001e-05, "loss": 1.8091, "step": 2088 }, { "epoch": 0.6690152121697358, "grad_norm": 0.5078125, "learning_rate": 1.4909122079752632e-05, "loss": 2.426, "step": 2089 }, { "epoch": 0.6693354683746997, "grad_norm": 0.4140625, "learning_rate": 1.4883547174091011e-05, "loss": 1.9204, "step": 2090 }, { "epoch": 0.6696557245796637, "grad_norm": 0.55078125, "learning_rate": 1.4857984922331847e-05, "loss": 2.1271, "step": 2091 }, { "epoch": 0.6699759807846277, "grad_norm": 0.53515625, "learning_rate": 1.4832435356449026e-05, "loss": 2.5165, "step": 2092 }, { "epoch": 0.6702962369895916, "grad_norm": 0.388671875, "learning_rate": 1.4806898508400552e-05, "loss": 1.994, "step": 2093 }, { "epoch": 0.6706164931945556, "grad_norm": 0.53515625, "learning_rate": 1.478137441012853e-05, "loss": 2.0746, "step": 2094 }, { "epoch": 0.6709367493995196, "grad_norm": 0.373046875, "learning_rate": 1.4755863093559118e-05, "loss": 1.7466, "step": 2095 }, { "epoch": 0.6712570056044835, "grad_norm": 0.51953125, "learning_rate": 1.473036459060248e-05, "loss": 2.247, "step": 2096 }, { "epoch": 0.6715772618094475, "grad_norm": 0.71484375, "learning_rate": 1.4704878933152761e-05, "loss": 2.4396, "step": 2097 }, { "epoch": 0.6718975180144116, "grad_norm": 0.5078125, "learning_rate": 1.4679406153088032e-05, "loss": 2.3092, "step": 2098 }, { "epoch": 0.6722177742193755, "grad_norm": 0.5625, "learning_rate": 1.4653946282270258e-05, "loss": 2.0661, "step": 2099 }, { "epoch": 0.6725380304243395, "grad_norm": 0.8671875, "learning_rate": 1.4628499352545266e-05, "loss": 2.3798, "step": 2100 }, { "epoch": 0.6728582866293035, "grad_norm": 0.98046875, "learning_rate": 1.4603065395742677e-05, "loss": 2.1861, "step": 2101 }, { "epoch": 0.6731785428342674, "grad_norm": 0.45703125, "learning_rate": 1.4577644443675895e-05, "loss": 1.6774, "step": 2102 }, { "epoch": 0.6734987990392314, "grad_norm": 0.494140625, "learning_rate": 1.4552236528142078e-05, "loss": 2.7617, "step": 2103 }, { "epoch": 0.6738190552441954, "grad_norm": 0.515625, "learning_rate": 1.452684168092203e-05, "loss": 2.9759, "step": 2104 }, { "epoch": 0.6741393114491593, "grad_norm": 0.53125, "learning_rate": 1.4501459933780249e-05, "loss": 2.4151, "step": 2105 }, { "epoch": 0.6744595676541233, "grad_norm": 0.318359375, "learning_rate": 1.4476091318464836e-05, "loss": 2.0188, "step": 2106 }, { "epoch": 0.6747798238590873, "grad_norm": 0.474609375, "learning_rate": 1.4450735866707474e-05, "loss": 2.5587, "step": 2107 }, { "epoch": 0.6751000800640512, "grad_norm": 0.451171875, "learning_rate": 1.4425393610223342e-05, "loss": 2.2319, "step": 2108 }, { "epoch": 0.6754203362690152, "grad_norm": 0.470703125, "learning_rate": 1.4400064580711182e-05, "loss": 2.3308, "step": 2109 }, { "epoch": 0.6757405924739792, "grad_norm": 1.46875, "learning_rate": 1.4374748809853125e-05, "loss": 1.9952, "step": 2110 }, { "epoch": 0.6760608486789431, "grad_norm": 0.482421875, "learning_rate": 1.4349446329314756e-05, "loss": 2.4647, "step": 2111 }, { "epoch": 0.6763811048839071, "grad_norm": 0.419921875, "learning_rate": 1.4324157170745025e-05, "loss": 2.0062, "step": 2112 }, { "epoch": 0.6767013610888711, "grad_norm": 0.53515625, "learning_rate": 1.429888136577622e-05, "loss": 2.5416, "step": 2113 }, { "epoch": 0.677021617293835, "grad_norm": 0.53515625, "learning_rate": 1.4273618946023925e-05, "loss": 2.4491, "step": 2114 }, { "epoch": 0.677341873498799, "grad_norm": 0.40234375, "learning_rate": 1.4248369943086998e-05, "loss": 2.3076, "step": 2115 }, { "epoch": 0.677662129703763, "grad_norm": 0.59765625, "learning_rate": 1.422313438854746e-05, "loss": 2.4278, "step": 2116 }, { "epoch": 0.677982385908727, "grad_norm": 0.609375, "learning_rate": 1.419791231397059e-05, "loss": 2.1202, "step": 2117 }, { "epoch": 0.678302642113691, "grad_norm": 0.4296875, "learning_rate": 1.4172703750904746e-05, "loss": 1.8748, "step": 2118 }, { "epoch": 0.678622898318655, "grad_norm": 0.3125, "learning_rate": 1.4147508730881403e-05, "loss": 2.3358, "step": 2119 }, { "epoch": 0.6789431545236189, "grad_norm": 0.7890625, "learning_rate": 1.4122327285415105e-05, "loss": 2.2464, "step": 2120 }, { "epoch": 0.6792634107285829, "grad_norm": 0.5078125, "learning_rate": 1.4097159446003417e-05, "loss": 2.0882, "step": 2121 }, { "epoch": 0.6795836669335469, "grad_norm": 0.4453125, "learning_rate": 1.4072005244126846e-05, "loss": 1.9614, "step": 2122 }, { "epoch": 0.6799039231385108, "grad_norm": 1.28125, "learning_rate": 1.4046864711248913e-05, "loss": 2.0258, "step": 2123 }, { "epoch": 0.6802241793434748, "grad_norm": 0.515625, "learning_rate": 1.4021737878815966e-05, "loss": 1.9489, "step": 2124 }, { "epoch": 0.6805444355484388, "grad_norm": 0.58984375, "learning_rate": 1.3996624778257287e-05, "loss": 2.5073, "step": 2125 }, { "epoch": 0.6808646917534027, "grad_norm": 0.45703125, "learning_rate": 1.3971525440984914e-05, "loss": 1.9678, "step": 2126 }, { "epoch": 0.6811849479583667, "grad_norm": 0.3984375, "learning_rate": 1.3946439898393718e-05, "loss": 2.1684, "step": 2127 }, { "epoch": 0.6815052041633307, "grad_norm": 0.51171875, "learning_rate": 1.3921368181861293e-05, "loss": 2.5053, "step": 2128 }, { "epoch": 0.6818254603682946, "grad_norm": 0.41796875, "learning_rate": 1.3896310322747957e-05, "loss": 2.5186, "step": 2129 }, { "epoch": 0.6821457165732586, "grad_norm": 0.41796875, "learning_rate": 1.3871266352396656e-05, "loss": 2.1431, "step": 2130 }, { "epoch": 0.6824659727782226, "grad_norm": 0.75390625, "learning_rate": 1.3846236302133023e-05, "loss": 2.2754, "step": 2131 }, { "epoch": 0.6827862289831865, "grad_norm": 0.34375, "learning_rate": 1.3821220203265223e-05, "loss": 2.122, "step": 2132 }, { "epoch": 0.6831064851881505, "grad_norm": 0.58984375, "learning_rate": 1.3796218087084e-05, "loss": 2.2143, "step": 2133 }, { "epoch": 0.6834267413931145, "grad_norm": 0.51953125, "learning_rate": 1.3771229984862608e-05, "loss": 2.6696, "step": 2134 }, { "epoch": 0.6837469975980784, "grad_norm": 0.50390625, "learning_rate": 1.374625592785677e-05, "loss": 2.481, "step": 2135 }, { "epoch": 0.6840672538030425, "grad_norm": 0.376953125, "learning_rate": 1.3721295947304607e-05, "loss": 2.4655, "step": 2136 }, { "epoch": 0.6843875100080064, "grad_norm": 0.443359375, "learning_rate": 1.3696350074426708e-05, "loss": 2.3388, "step": 2137 }, { "epoch": 0.6847077662129704, "grad_norm": 0.58984375, "learning_rate": 1.3671418340425928e-05, "loss": 2.2813, "step": 2138 }, { "epoch": 0.6850280224179344, "grad_norm": 0.466796875, "learning_rate": 1.3646500776487519e-05, "loss": 2.7145, "step": 2139 }, { "epoch": 0.6853482786228983, "grad_norm": 0.515625, "learning_rate": 1.3621597413778939e-05, "loss": 2.4922, "step": 2140 }, { "epoch": 0.6856685348278623, "grad_norm": 0.86328125, "learning_rate": 1.3596708283449921e-05, "loss": 2.0813, "step": 2141 }, { "epoch": 0.6859887910328263, "grad_norm": 0.5078125, "learning_rate": 1.3571833416632396e-05, "loss": 1.8429, "step": 2142 }, { "epoch": 0.6863090472377902, "grad_norm": 1.0546875, "learning_rate": 1.3546972844440441e-05, "loss": 2.3635, "step": 2143 }, { "epoch": 0.6866293034427542, "grad_norm": 0.494140625, "learning_rate": 1.3522126597970244e-05, "loss": 2.161, "step": 2144 }, { "epoch": 0.6869495596477182, "grad_norm": 0.546875, "learning_rate": 1.3497294708300118e-05, "loss": 2.2284, "step": 2145 }, { "epoch": 0.6872698158526821, "grad_norm": 0.51171875, "learning_rate": 1.3472477206490355e-05, "loss": 2.4539, "step": 2146 }, { "epoch": 0.6875900720576461, "grad_norm": 0.5703125, "learning_rate": 1.34476741235833e-05, "loss": 2.5645, "step": 2147 }, { "epoch": 0.6879103282626101, "grad_norm": 0.44921875, "learning_rate": 1.3422885490603244e-05, "loss": 2.0052, "step": 2148 }, { "epoch": 0.688230584467574, "grad_norm": 0.37890625, "learning_rate": 1.3398111338556412e-05, "loss": 1.8642, "step": 2149 }, { "epoch": 0.688550840672538, "grad_norm": 0.55859375, "learning_rate": 1.3373351698430884e-05, "loss": 2.543, "step": 2150 }, { "epoch": 0.688871096877502, "grad_norm": 0.462890625, "learning_rate": 1.3348606601196648e-05, "loss": 2.08, "step": 2151 }, { "epoch": 0.6891913530824659, "grad_norm": 0.4453125, "learning_rate": 1.332387607780543e-05, "loss": 1.7781, "step": 2152 }, { "epoch": 0.6895116092874299, "grad_norm": 0.5859375, "learning_rate": 1.3299160159190801e-05, "loss": 2.3629, "step": 2153 }, { "epoch": 0.689831865492394, "grad_norm": 0.59375, "learning_rate": 1.3274458876267998e-05, "loss": 2.1964, "step": 2154 }, { "epoch": 0.6901521216973578, "grad_norm": 0.416015625, "learning_rate": 1.324977225993399e-05, "loss": 1.7751, "step": 2155 }, { "epoch": 0.6904723779023219, "grad_norm": 0.51953125, "learning_rate": 1.3225100341067393e-05, "loss": 1.352, "step": 2156 }, { "epoch": 0.6907926341072859, "grad_norm": 0.56640625, "learning_rate": 1.3200443150528438e-05, "loss": 1.7647, "step": 2157 }, { "epoch": 0.6911128903122498, "grad_norm": 0.88671875, "learning_rate": 1.317580071915891e-05, "loss": 1.7487, "step": 2158 }, { "epoch": 0.6914331465172138, "grad_norm": 0.83203125, "learning_rate": 1.3151173077782192e-05, "loss": 2.8391, "step": 2159 }, { "epoch": 0.6917534027221778, "grad_norm": 0.60546875, "learning_rate": 1.3126560257203097e-05, "loss": 2.3178, "step": 2160 }, { "epoch": 0.6920736589271417, "grad_norm": 0.39453125, "learning_rate": 1.3101962288207948e-05, "loss": 1.708, "step": 2161 }, { "epoch": 0.6923939151321057, "grad_norm": 1.078125, "learning_rate": 1.3077379201564472e-05, "loss": 1.5599, "step": 2162 }, { "epoch": 0.6927141713370697, "grad_norm": 0.4140625, "learning_rate": 1.3052811028021788e-05, "loss": 2.1572, "step": 2163 }, { "epoch": 0.6930344275420336, "grad_norm": 0.59765625, "learning_rate": 1.3028257798310356e-05, "loss": 2.1006, "step": 2164 }, { "epoch": 0.6933546837469976, "grad_norm": 0.341796875, "learning_rate": 1.3003719543141957e-05, "loss": 2.0096, "step": 2165 }, { "epoch": 0.6936749399519616, "grad_norm": 3.328125, "learning_rate": 1.29791962932096e-05, "loss": 2.0032, "step": 2166 }, { "epoch": 0.6939951961569255, "grad_norm": 0.46875, "learning_rate": 1.29546880791876e-05, "loss": 2.5821, "step": 2167 }, { "epoch": 0.6943154523618895, "grad_norm": 1.1796875, "learning_rate": 1.2930194931731382e-05, "loss": 2.3261, "step": 2168 }, { "epoch": 0.6946357085668535, "grad_norm": 0.462890625, "learning_rate": 1.290571688147758e-05, "loss": 2.4759, "step": 2169 }, { "epoch": 0.6949559647718174, "grad_norm": 0.84765625, "learning_rate": 1.2881253959043926e-05, "loss": 2.61, "step": 2170 }, { "epoch": 0.6952762209767814, "grad_norm": 0.5234375, "learning_rate": 1.2856806195029241e-05, "loss": 2.443, "step": 2171 }, { "epoch": 0.6955964771817454, "grad_norm": 0.44140625, "learning_rate": 1.2832373620013352e-05, "loss": 2.3406, "step": 2172 }, { "epoch": 0.6959167333867093, "grad_norm": 0.54296875, "learning_rate": 1.280795626455714e-05, "loss": 2.389, "step": 2173 }, { "epoch": 0.6962369895916733, "grad_norm": 0.4140625, "learning_rate": 1.2783554159202404e-05, "loss": 2.4067, "step": 2174 }, { "epoch": 0.6965572457966374, "grad_norm": 0.443359375, "learning_rate": 1.2759167334471884e-05, "loss": 2.178, "step": 2175 }, { "epoch": 0.6968775020016013, "grad_norm": 0.373046875, "learning_rate": 1.2734795820869208e-05, "loss": 2.0938, "step": 2176 }, { "epoch": 0.6971977582065653, "grad_norm": 0.4296875, "learning_rate": 1.2710439648878855e-05, "loss": 1.9665, "step": 2177 }, { "epoch": 0.6975180144115293, "grad_norm": 0.5390625, "learning_rate": 1.2686098848966107e-05, "loss": 2.3108, "step": 2178 }, { "epoch": 0.6978382706164932, "grad_norm": 0.4453125, "learning_rate": 1.2661773451577036e-05, "loss": 1.9203, "step": 2179 }, { "epoch": 0.6981585268214572, "grad_norm": 0.91015625, "learning_rate": 1.2637463487138398e-05, "loss": 2.6032, "step": 2180 }, { "epoch": 0.6984787830264212, "grad_norm": 0.65625, "learning_rate": 1.2613168986057728e-05, "loss": 2.3923, "step": 2181 }, { "epoch": 0.6987990392313851, "grad_norm": 0.71484375, "learning_rate": 1.2588889978723134e-05, "loss": 1.8835, "step": 2182 }, { "epoch": 0.6991192954363491, "grad_norm": 0.42578125, "learning_rate": 1.2564626495503401e-05, "loss": 1.7873, "step": 2183 }, { "epoch": 0.699439551641313, "grad_norm": 0.4921875, "learning_rate": 1.2540378566747874e-05, "loss": 1.8807, "step": 2184 }, { "epoch": 0.699759807846277, "grad_norm": 0.5625, "learning_rate": 1.2516146222786451e-05, "loss": 2.3201, "step": 2185 }, { "epoch": 0.700080064051241, "grad_norm": 0.396484375, "learning_rate": 1.2491929493929508e-05, "loss": 1.9649, "step": 2186 }, { "epoch": 0.7004003202562049, "grad_norm": 0.55078125, "learning_rate": 1.2467728410467944e-05, "loss": 1.8792, "step": 2187 }, { "epoch": 0.7007205764611689, "grad_norm": 0.6171875, "learning_rate": 1.2443543002673022e-05, "loss": 2.3325, "step": 2188 }, { "epoch": 0.7010408326661329, "grad_norm": 0.73828125, "learning_rate": 1.241937330079647e-05, "loss": 2.6154, "step": 2189 }, { "epoch": 0.7013610888710968, "grad_norm": 0.44921875, "learning_rate": 1.2395219335070301e-05, "loss": 2.442, "step": 2190 }, { "epoch": 0.7016813450760608, "grad_norm": 0.46484375, "learning_rate": 1.2371081135706888e-05, "loss": 2.3294, "step": 2191 }, { "epoch": 0.7020016012810248, "grad_norm": 0.5625, "learning_rate": 1.234695873289887e-05, "loss": 2.4601, "step": 2192 }, { "epoch": 0.7023218574859887, "grad_norm": 0.490234375, "learning_rate": 1.2322852156819137e-05, "loss": 2.0798, "step": 2193 }, { "epoch": 0.7026421136909528, "grad_norm": 0.4453125, "learning_rate": 1.2298761437620748e-05, "loss": 2.283, "step": 2194 }, { "epoch": 0.7029623698959168, "grad_norm": 0.498046875, "learning_rate": 1.2274686605436989e-05, "loss": 2.3894, "step": 2195 }, { "epoch": 0.7032826261008807, "grad_norm": 0.53125, "learning_rate": 1.2250627690381213e-05, "loss": 2.4578, "step": 2196 }, { "epoch": 0.7036028823058447, "grad_norm": 0.52734375, "learning_rate": 1.2226584722546899e-05, "loss": 2.2, "step": 2197 }, { "epoch": 0.7039231385108087, "grad_norm": 0.412109375, "learning_rate": 1.2202557732007569e-05, "loss": 1.6341, "step": 2198 }, { "epoch": 0.7042433947157726, "grad_norm": 0.58984375, "learning_rate": 1.217854674881677e-05, "loss": 2.0122, "step": 2199 }, { "epoch": 0.7045636509207366, "grad_norm": 0.5703125, "learning_rate": 1.2154551803007986e-05, "loss": 2.0305, "step": 2200 }, { "epoch": 0.7048839071257006, "grad_norm": 0.93359375, "learning_rate": 1.2130572924594713e-05, "loss": 1.7452, "step": 2201 }, { "epoch": 0.7052041633306645, "grad_norm": 0.5625, "learning_rate": 1.2106610143570271e-05, "loss": 1.9185, "step": 2202 }, { "epoch": 0.7055244195356285, "grad_norm": 0.494140625, "learning_rate": 1.208266348990792e-05, "loss": 2.1781, "step": 2203 }, { "epoch": 0.7058446757405925, "grad_norm": 0.59765625, "learning_rate": 1.2058732993560678e-05, "loss": 2.6887, "step": 2204 }, { "epoch": 0.7061649319455564, "grad_norm": 0.53515625, "learning_rate": 1.2034818684461402e-05, "loss": 2.3117, "step": 2205 }, { "epoch": 0.7064851881505204, "grad_norm": 0.46484375, "learning_rate": 1.2010920592522681e-05, "loss": 2.0872, "step": 2206 }, { "epoch": 0.7068054443554844, "grad_norm": 1.296875, "learning_rate": 1.198703874763683e-05, "loss": 2.1383, "step": 2207 }, { "epoch": 0.7071257005604483, "grad_norm": 1.625, "learning_rate": 1.1963173179675812e-05, "loss": 2.1324, "step": 2208 }, { "epoch": 0.7074459567654123, "grad_norm": 0.515625, "learning_rate": 1.1939323918491288e-05, "loss": 2.2598, "step": 2209 }, { "epoch": 0.7077662129703763, "grad_norm": 0.84765625, "learning_rate": 1.1915490993914458e-05, "loss": 2.0589, "step": 2210 }, { "epoch": 0.7080864691753402, "grad_norm": 0.7109375, "learning_rate": 1.1891674435756129e-05, "loss": 2.522, "step": 2211 }, { "epoch": 0.7084067253803042, "grad_norm": 0.53515625, "learning_rate": 1.1867874273806623e-05, "loss": 2.4149, "step": 2212 }, { "epoch": 0.7087269815852683, "grad_norm": 1.140625, "learning_rate": 1.1844090537835764e-05, "loss": 2.5084, "step": 2213 }, { "epoch": 0.7090472377902322, "grad_norm": 0.6328125, "learning_rate": 1.1820323257592796e-05, "loss": 2.5254, "step": 2214 }, { "epoch": 0.7093674939951962, "grad_norm": 0.73828125, "learning_rate": 1.1796572462806437e-05, "loss": 2.2132, "step": 2215 }, { "epoch": 0.7096877502001602, "grad_norm": 0.5078125, "learning_rate": 1.177283818318472e-05, "loss": 2.0243, "step": 2216 }, { "epoch": 0.7100080064051241, "grad_norm": 0.640625, "learning_rate": 1.1749120448415085e-05, "loss": 2.1482, "step": 2217 }, { "epoch": 0.7103282626100881, "grad_norm": 0.466796875, "learning_rate": 1.1725419288164222e-05, "loss": 1.9561, "step": 2218 }, { "epoch": 0.7106485188150521, "grad_norm": 0.5078125, "learning_rate": 1.1701734732078118e-05, "loss": 2.4484, "step": 2219 }, { "epoch": 0.710968775020016, "grad_norm": 0.7109375, "learning_rate": 1.1678066809781988e-05, "loss": 2.4673, "step": 2220 }, { "epoch": 0.71128903122498, "grad_norm": 0.3515625, "learning_rate": 1.1654415550880243e-05, "loss": 2.1008, "step": 2221 }, { "epoch": 0.711609287429944, "grad_norm": 0.6875, "learning_rate": 1.163078098495642e-05, "loss": 2.556, "step": 2222 }, { "epoch": 0.7119295436349079, "grad_norm": 0.478515625, "learning_rate": 1.1607163141573246e-05, "loss": 2.6396, "step": 2223 }, { "epoch": 0.7122497998398719, "grad_norm": 0.55859375, "learning_rate": 1.158356205027245e-05, "loss": 2.5558, "step": 2224 }, { "epoch": 0.7125700560448359, "grad_norm": 0.45703125, "learning_rate": 1.1559977740574861e-05, "loss": 2.4041, "step": 2225 }, { "epoch": 0.7128903122497998, "grad_norm": 0.609375, "learning_rate": 1.1536410241980297e-05, "loss": 2.3616, "step": 2226 }, { "epoch": 0.7132105684547638, "grad_norm": 0.56640625, "learning_rate": 1.1512859583967552e-05, "loss": 2.4075, "step": 2227 }, { "epoch": 0.7135308246597278, "grad_norm": 0.431640625, "learning_rate": 1.148932579599436e-05, "loss": 2.4303, "step": 2228 }, { "epoch": 0.7138510808646917, "grad_norm": 0.6484375, "learning_rate": 1.1465808907497352e-05, "loss": 2.5852, "step": 2229 }, { "epoch": 0.7141713370696557, "grad_norm": 1.5859375, "learning_rate": 1.1442308947891992e-05, "loss": 2.5451, "step": 2230 }, { "epoch": 0.7144915932746196, "grad_norm": 0.61328125, "learning_rate": 1.141882594657263e-05, "loss": 1.6112, "step": 2231 }, { "epoch": 0.7148118494795836, "grad_norm": 0.5234375, "learning_rate": 1.1395359932912348e-05, "loss": 2.3431, "step": 2232 }, { "epoch": 0.7151321056845477, "grad_norm": 8.1875, "learning_rate": 1.1371910936263e-05, "loss": 2.2054, "step": 2233 }, { "epoch": 0.7154523618895116, "grad_norm": 0.365234375, "learning_rate": 1.1348478985955161e-05, "loss": 1.7845, "step": 2234 }, { "epoch": 0.7157726180944756, "grad_norm": 0.46875, "learning_rate": 1.1325064111298087e-05, "loss": 2.5106, "step": 2235 }, { "epoch": 0.7160928742994396, "grad_norm": 0.376953125, "learning_rate": 1.1301666341579633e-05, "loss": 2.3032, "step": 2236 }, { "epoch": 0.7164131305044035, "grad_norm": 0.498046875, "learning_rate": 1.1278285706066335e-05, "loss": 2.0377, "step": 2237 }, { "epoch": 0.7167333867093675, "grad_norm": 0.51953125, "learning_rate": 1.1254922234003224e-05, "loss": 2.2911, "step": 2238 }, { "epoch": 0.7170536429143315, "grad_norm": 0.5, "learning_rate": 1.1231575954613902e-05, "loss": 1.9858, "step": 2239 }, { "epoch": 0.7173738991192954, "grad_norm": 0.7578125, "learning_rate": 1.1208246897100458e-05, "loss": 2.2628, "step": 2240 }, { "epoch": 0.7176941553242594, "grad_norm": 0.474609375, "learning_rate": 1.1184935090643433e-05, "loss": 1.7735, "step": 2241 }, { "epoch": 0.7180144115292234, "grad_norm": 0.322265625, "learning_rate": 1.1161640564401805e-05, "loss": 1.6539, "step": 2242 }, { "epoch": 0.7183346677341873, "grad_norm": 1.078125, "learning_rate": 1.1138363347512926e-05, "loss": 2.1595, "step": 2243 }, { "epoch": 0.7186549239391513, "grad_norm": 0.310546875, "learning_rate": 1.1115103469092475e-05, "loss": 1.9965, "step": 2244 }, { "epoch": 0.7189751801441153, "grad_norm": 0.68359375, "learning_rate": 1.1091860958234508e-05, "loss": 2.3327, "step": 2245 }, { "epoch": 0.7192954363490792, "grad_norm": 0.4375, "learning_rate": 1.1068635844011286e-05, "loss": 2.1586, "step": 2246 }, { "epoch": 0.7196156925540432, "grad_norm": 0.5859375, "learning_rate": 1.1045428155473347e-05, "loss": 1.9614, "step": 2247 }, { "epoch": 0.7199359487590072, "grad_norm": 0.4375, "learning_rate": 1.1022237921649429e-05, "loss": 1.8978, "step": 2248 }, { "epoch": 0.7202562049639711, "grad_norm": 0.52734375, "learning_rate": 1.0999065171546432e-05, "loss": 2.1097, "step": 2249 }, { "epoch": 0.7205764611689351, "grad_norm": 0.46875, "learning_rate": 1.0975909934149392e-05, "loss": 2.1752, "step": 2250 }, { "epoch": 0.7208967173738992, "grad_norm": 0.65234375, "learning_rate": 1.0952772238421444e-05, "loss": 2.2443, "step": 2251 }, { "epoch": 0.721216973578863, "grad_norm": 0.578125, "learning_rate": 1.0929652113303754e-05, "loss": 2.4343, "step": 2252 }, { "epoch": 0.7215372297838271, "grad_norm": 0.47265625, "learning_rate": 1.0906549587715546e-05, "loss": 2.1622, "step": 2253 }, { "epoch": 0.7218574859887911, "grad_norm": 0.515625, "learning_rate": 1.0883464690554005e-05, "loss": 2.4342, "step": 2254 }, { "epoch": 0.722177742193755, "grad_norm": 0.498046875, "learning_rate": 1.0860397450694285e-05, "loss": 2.6058, "step": 2255 }, { "epoch": 0.722497998398719, "grad_norm": 0.5703125, "learning_rate": 1.0837347896989437e-05, "loss": 2.2438, "step": 2256 }, { "epoch": 0.722818254603683, "grad_norm": 0.58984375, "learning_rate": 1.081431605827041e-05, "loss": 2.1936, "step": 2257 }, { "epoch": 0.7231385108086469, "grad_norm": 0.52734375, "learning_rate": 1.0791301963345954e-05, "loss": 2.4717, "step": 2258 }, { "epoch": 0.7234587670136109, "grad_norm": 1.0078125, "learning_rate": 1.0768305641002688e-05, "loss": 2.1485, "step": 2259 }, { "epoch": 0.7237790232185749, "grad_norm": 0.62109375, "learning_rate": 1.0745327120004935e-05, "loss": 2.4316, "step": 2260 }, { "epoch": 0.7240992794235388, "grad_norm": 0.58984375, "learning_rate": 1.0722366429094798e-05, "loss": 1.7809, "step": 2261 }, { "epoch": 0.7244195356285028, "grad_norm": 0.6484375, "learning_rate": 1.0699423596992056e-05, "loss": 2.3482, "step": 2262 }, { "epoch": 0.7247397918334668, "grad_norm": 0.453125, "learning_rate": 1.0676498652394157e-05, "loss": 2.037, "step": 2263 }, { "epoch": 0.7250600480384307, "grad_norm": 0.419921875, "learning_rate": 1.065359162397617e-05, "loss": 2.3973, "step": 2264 }, { "epoch": 0.7253803042433947, "grad_norm": 0.404296875, "learning_rate": 1.0630702540390766e-05, "loss": 2.2461, "step": 2265 }, { "epoch": 0.7257005604483587, "grad_norm": 0.87109375, "learning_rate": 1.060783143026814e-05, "loss": 2.4446, "step": 2266 }, { "epoch": 0.7260208166533226, "grad_norm": 0.57421875, "learning_rate": 1.0584978322216057e-05, "loss": 1.8983, "step": 2267 }, { "epoch": 0.7263410728582866, "grad_norm": 0.72265625, "learning_rate": 1.0562143244819707e-05, "loss": 1.9454, "step": 2268 }, { "epoch": 0.7266613290632506, "grad_norm": 0.478515625, "learning_rate": 1.0539326226641768e-05, "loss": 2.2344, "step": 2269 }, { "epoch": 0.7269815852682145, "grad_norm": 0.46484375, "learning_rate": 1.0516527296222308e-05, "loss": 2.0809, "step": 2270 }, { "epoch": 0.7273018414731786, "grad_norm": 0.4765625, "learning_rate": 1.0493746482078783e-05, "loss": 1.7608, "step": 2271 }, { "epoch": 0.7276220976781426, "grad_norm": 0.455078125, "learning_rate": 1.0470983812705984e-05, "loss": 2.4922, "step": 2272 }, { "epoch": 0.7279423538831065, "grad_norm": 0.33984375, "learning_rate": 1.0448239316576009e-05, "loss": 1.822, "step": 2273 }, { "epoch": 0.7282626100880705, "grad_norm": 0.376953125, "learning_rate": 1.0425513022138203e-05, "loss": 2.2417, "step": 2274 }, { "epoch": 0.7285828662930345, "grad_norm": 0.5234375, "learning_rate": 1.0402804957819174e-05, "loss": 1.8278, "step": 2275 }, { "epoch": 0.7289031224979984, "grad_norm": 0.62890625, "learning_rate": 1.0380115152022716e-05, "loss": 2.0429, "step": 2276 }, { "epoch": 0.7292233787029624, "grad_norm": 0.3671875, "learning_rate": 1.0357443633129777e-05, "loss": 2.1444, "step": 2277 }, { "epoch": 0.7295436349079263, "grad_norm": 0.53515625, "learning_rate": 1.0334790429498445e-05, "loss": 2.5259, "step": 2278 }, { "epoch": 0.7298638911128903, "grad_norm": 0.51171875, "learning_rate": 1.0312155569463897e-05, "loss": 2.4972, "step": 2279 }, { "epoch": 0.7301841473178543, "grad_norm": 0.5234375, "learning_rate": 1.0289539081338335e-05, "loss": 1.7737, "step": 2280 }, { "epoch": 0.7305044035228182, "grad_norm": 0.40234375, "learning_rate": 1.026694099341104e-05, "loss": 2.4937, "step": 2281 }, { "epoch": 0.7308246597277822, "grad_norm": 0.62890625, "learning_rate": 1.024436133394822e-05, "loss": 1.7839, "step": 2282 }, { "epoch": 0.7311449159327462, "grad_norm": 0.470703125, "learning_rate": 1.022180013119306e-05, "loss": 2.5823, "step": 2283 }, { "epoch": 0.7314651721377101, "grad_norm": 0.640625, "learning_rate": 1.0199257413365663e-05, "loss": 2.5663, "step": 2284 }, { "epoch": 0.7317854283426741, "grad_norm": 0.49609375, "learning_rate": 1.0176733208662993e-05, "loss": 2.738, "step": 2285 }, { "epoch": 0.7321056845476381, "grad_norm": 0.5625, "learning_rate": 1.0154227545258873e-05, "loss": 2.3974, "step": 2286 }, { "epoch": 0.732425940752602, "grad_norm": 0.93359375, "learning_rate": 1.0131740451303936e-05, "loss": 2.6672, "step": 2287 }, { "epoch": 0.732746196957566, "grad_norm": 0.412109375, "learning_rate": 1.0109271954925562e-05, "loss": 1.7195, "step": 2288 }, { "epoch": 0.73306645316253, "grad_norm": 0.51171875, "learning_rate": 1.0086822084227896e-05, "loss": 2.2517, "step": 2289 }, { "epoch": 0.733386709367494, "grad_norm": 0.3828125, "learning_rate": 1.0064390867291774e-05, "loss": 2.5776, "step": 2290 }, { "epoch": 0.733706965572458, "grad_norm": 0.609375, "learning_rate": 1.00419783321747e-05, "loss": 2.0832, "step": 2291 }, { "epoch": 0.734027221777422, "grad_norm": 0.5546875, "learning_rate": 1.0019584506910817e-05, "loss": 1.78, "step": 2292 }, { "epoch": 0.7343474779823859, "grad_norm": 0.52734375, "learning_rate": 9.99720941951086e-06, "loss": 2.3546, "step": 2293 }, { "epoch": 0.7346677341873499, "grad_norm": 0.6328125, "learning_rate": 9.974853097962116e-06, "loss": 2.3111, "step": 2294 }, { "epoch": 0.7349879903923139, "grad_norm": 0.625, "learning_rate": 9.952515570228432e-06, "loss": 1.9775, "step": 2295 }, { "epoch": 0.7353082465972778, "grad_norm": 0.494140625, "learning_rate": 9.930196864250102e-06, "loss": 2.029, "step": 2296 }, { "epoch": 0.7356285028022418, "grad_norm": 0.49609375, "learning_rate": 9.90789700794391e-06, "loss": 2.0413, "step": 2297 }, { "epoch": 0.7359487590072058, "grad_norm": 0.546875, "learning_rate": 9.885616029203046e-06, "loss": 1.8798, "step": 2298 }, { "epoch": 0.7362690152121697, "grad_norm": 0.33984375, "learning_rate": 9.8633539558971e-06, "loss": 1.7159, "step": 2299 }, { "epoch": 0.7365892714171337, "grad_norm": 0.466796875, "learning_rate": 9.84111081587201e-06, "loss": 2.5267, "step": 2300 }, { "epoch": 0.7369095276220977, "grad_norm": 1.5234375, "learning_rate": 9.818886636950037e-06, "loss": 2.7178, "step": 2301 }, { "epoch": 0.7372297838270616, "grad_norm": 0.4765625, "learning_rate": 9.796681446929698e-06, "loss": 2.4225, "step": 2302 }, { "epoch": 0.7375500400320256, "grad_norm": 0.4296875, "learning_rate": 9.77449527358579e-06, "loss": 2.337, "step": 2303 }, { "epoch": 0.7378702962369896, "grad_norm": 0.81640625, "learning_rate": 9.752328144669316e-06, "loss": 2.5408, "step": 2304 }, { "epoch": 0.7381905524419535, "grad_norm": 0.53515625, "learning_rate": 9.730180087907445e-06, "loss": 1.9689, "step": 2305 }, { "epoch": 0.7385108086469175, "grad_norm": 0.396484375, "learning_rate": 9.708051131003507e-06, "loss": 2.0099, "step": 2306 }, { "epoch": 0.7388310648518815, "grad_norm": 0.59765625, "learning_rate": 9.685941301636927e-06, "loss": 2.1473, "step": 2307 }, { "epoch": 0.7391513210568454, "grad_norm": 0.5625, "learning_rate": 9.663850627463219e-06, "loss": 1.9139, "step": 2308 }, { "epoch": 0.7394715772618095, "grad_norm": 0.59765625, "learning_rate": 9.641779136113931e-06, "loss": 2.5305, "step": 2309 }, { "epoch": 0.7397918334667735, "grad_norm": 0.51171875, "learning_rate": 9.619726855196603e-06, "loss": 2.4336, "step": 2310 }, { "epoch": 0.7401120896717374, "grad_norm": 0.60546875, "learning_rate": 9.597693812294761e-06, "loss": 2.3245, "step": 2311 }, { "epoch": 0.7404323458767014, "grad_norm": 0.5859375, "learning_rate": 9.57568003496787e-06, "loss": 2.3801, "step": 2312 }, { "epoch": 0.7407526020816654, "grad_norm": 0.47265625, "learning_rate": 9.553685550751294e-06, "loss": 2.4653, "step": 2313 }, { "epoch": 0.7410728582866293, "grad_norm": 0.494140625, "learning_rate": 9.531710387156254e-06, "loss": 2.639, "step": 2314 }, { "epoch": 0.7413931144915933, "grad_norm": 0.404296875, "learning_rate": 9.509754571669821e-06, "loss": 2.0894, "step": 2315 }, { "epoch": 0.7417133706965573, "grad_norm": 0.53515625, "learning_rate": 9.48781813175486e-06, "loss": 1.8972, "step": 2316 }, { "epoch": 0.7420336269015212, "grad_norm": 0.87109375, "learning_rate": 9.465901094849985e-06, "loss": 1.8963, "step": 2317 }, { "epoch": 0.7423538831064852, "grad_norm": 0.30078125, "learning_rate": 9.44400348836956e-06, "loss": 2.1828, "step": 2318 }, { "epoch": 0.7426741393114492, "grad_norm": 0.578125, "learning_rate": 9.422125339703635e-06, "loss": 1.7612, "step": 2319 }, { "epoch": 0.7429943955164131, "grad_norm": 0.58203125, "learning_rate": 9.400266676217927e-06, "loss": 2.154, "step": 2320 }, { "epoch": 0.7433146517213771, "grad_norm": 0.60546875, "learning_rate": 9.378427525253781e-06, "loss": 2.6834, "step": 2321 }, { "epoch": 0.7436349079263411, "grad_norm": 0.5, "learning_rate": 9.356607914128126e-06, "loss": 1.8765, "step": 2322 }, { "epoch": 0.743955164131305, "grad_norm": 0.82421875, "learning_rate": 9.33480787013346e-06, "loss": 2.3637, "step": 2323 }, { "epoch": 0.744275420336269, "grad_norm": 1.203125, "learning_rate": 9.313027420537809e-06, "loss": 2.5115, "step": 2324 }, { "epoch": 0.7445956765412329, "grad_norm": 0.51953125, "learning_rate": 9.291266592584666e-06, "loss": 1.9597, "step": 2325 }, { "epoch": 0.7449159327461969, "grad_norm": 0.466796875, "learning_rate": 9.269525413493008e-06, "loss": 2.3502, "step": 2326 }, { "epoch": 0.745236188951161, "grad_norm": 0.44921875, "learning_rate": 9.247803910457226e-06, "loss": 2.5435, "step": 2327 }, { "epoch": 0.7455564451561248, "grad_norm": 0.451171875, "learning_rate": 9.226102110647092e-06, "loss": 2.3085, "step": 2328 }, { "epoch": 0.7458767013610889, "grad_norm": 0.48828125, "learning_rate": 9.204420041207746e-06, "loss": 2.0914, "step": 2329 }, { "epoch": 0.7461969575660529, "grad_norm": 1.6328125, "learning_rate": 9.182757729259638e-06, "loss": 1.7195, "step": 2330 }, { "epoch": 0.7465172137710168, "grad_norm": 0.43359375, "learning_rate": 9.161115201898518e-06, "loss": 2.1269, "step": 2331 }, { "epoch": 0.7468374699759808, "grad_norm": 0.609375, "learning_rate": 9.139492486195364e-06, "loss": 2.4199, "step": 2332 }, { "epoch": 0.7471577261809448, "grad_norm": 0.546875, "learning_rate": 9.117889609196395e-06, "loss": 2.0749, "step": 2333 }, { "epoch": 0.7474779823859087, "grad_norm": 0.4921875, "learning_rate": 9.09630659792301e-06, "loss": 2.6966, "step": 2334 }, { "epoch": 0.7477982385908727, "grad_norm": 0.67578125, "learning_rate": 9.074743479371761e-06, "loss": 1.7836, "step": 2335 }, { "epoch": 0.7481184947958367, "grad_norm": 0.66796875, "learning_rate": 9.05320028051431e-06, "loss": 2.27, "step": 2336 }, { "epoch": 0.7484387510008006, "grad_norm": 0.57421875, "learning_rate": 9.031677028297413e-06, "loss": 1.792, "step": 2337 }, { "epoch": 0.7487590072057646, "grad_norm": 0.51171875, "learning_rate": 9.010173749642878e-06, "loss": 2.2518, "step": 2338 }, { "epoch": 0.7490792634107286, "grad_norm": 0.412109375, "learning_rate": 8.988690471447503e-06, "loss": 1.1478, "step": 2339 }, { "epoch": 0.7493995196156925, "grad_norm": 0.49609375, "learning_rate": 8.967227220583102e-06, "loss": 2.3766, "step": 2340 }, { "epoch": 0.7497197758206565, "grad_norm": 0.6015625, "learning_rate": 8.945784023896425e-06, "loss": 1.8921, "step": 2341 }, { "epoch": 0.7500400320256205, "grad_norm": 0.486328125, "learning_rate": 8.924360908209137e-06, "loss": 1.6513, "step": 2342 }, { "epoch": 0.7503602882305844, "grad_norm": 0.63671875, "learning_rate": 8.902957900317783e-06, "loss": 1.5818, "step": 2343 }, { "epoch": 0.7506805444355484, "grad_norm": 0.408203125, "learning_rate": 8.881575026993766e-06, "loss": 2.2554, "step": 2344 }, { "epoch": 0.7510008006405124, "grad_norm": 0.56640625, "learning_rate": 8.860212314983292e-06, "loss": 2.2229, "step": 2345 }, { "epoch": 0.7513210568454763, "grad_norm": 1.6484375, "learning_rate": 8.838869791007367e-06, "loss": 2.2565, "step": 2346 }, { "epoch": 0.7516413130504404, "grad_norm": 0.5078125, "learning_rate": 8.817547481761717e-06, "loss": 2.5122, "step": 2347 }, { "epoch": 0.7519615692554044, "grad_norm": 0.47265625, "learning_rate": 8.796245413916807e-06, "loss": 2.6242, "step": 2348 }, { "epoch": 0.7522818254603683, "grad_norm": 0.60546875, "learning_rate": 8.774963614117773e-06, "loss": 2.2074, "step": 2349 }, { "epoch": 0.7526020816653323, "grad_norm": 0.56640625, "learning_rate": 8.753702108984405e-06, "loss": 2.5946, "step": 2350 }, { "epoch": 0.7529223378702963, "grad_norm": 0.431640625, "learning_rate": 8.732460925111105e-06, "loss": 2.4789, "step": 2351 }, { "epoch": 0.7532425940752602, "grad_norm": 0.58203125, "learning_rate": 8.711240089066863e-06, "loss": 2.1403, "step": 2352 }, { "epoch": 0.7535628502802242, "grad_norm": 0.80078125, "learning_rate": 8.690039627395195e-06, "loss": 2.0411, "step": 2353 }, { "epoch": 0.7538831064851882, "grad_norm": 0.494140625, "learning_rate": 8.668859566614154e-06, "loss": 2.4965, "step": 2354 }, { "epoch": 0.7542033626901521, "grad_norm": 0.54296875, "learning_rate": 8.647699933216274e-06, "loss": 2.275, "step": 2355 }, { "epoch": 0.7545236188951161, "grad_norm": 0.55078125, "learning_rate": 8.626560753668531e-06, "loss": 2.3405, "step": 2356 }, { "epoch": 0.7548438751000801, "grad_norm": 0.498046875, "learning_rate": 8.605442054412318e-06, "loss": 2.0917, "step": 2357 }, { "epoch": 0.755164131305044, "grad_norm": 0.515625, "learning_rate": 8.584343861863418e-06, "loss": 2.1474, "step": 2358 }, { "epoch": 0.755484387510008, "grad_norm": 0.4765625, "learning_rate": 8.563266202411948e-06, "loss": 2.2758, "step": 2359 }, { "epoch": 0.755804643714972, "grad_norm": 0.59765625, "learning_rate": 8.542209102422366e-06, "loss": 2.5798, "step": 2360 }, { "epoch": 0.7561248999199359, "grad_norm": 0.5546875, "learning_rate": 8.521172588233378e-06, "loss": 2.485, "step": 2361 }, { "epoch": 0.7564451561248999, "grad_norm": 0.5, "learning_rate": 8.500156686157975e-06, "loss": 2.354, "step": 2362 }, { "epoch": 0.7567654123298639, "grad_norm": 0.47265625, "learning_rate": 8.479161422483353e-06, "loss": 1.9641, "step": 2363 }, { "epoch": 0.7570856685348278, "grad_norm": 0.43359375, "learning_rate": 8.458186823470888e-06, "loss": 2.5159, "step": 2364 }, { "epoch": 0.7574059247397918, "grad_norm": 0.50390625, "learning_rate": 8.437232915356114e-06, "loss": 2.1078, "step": 2365 }, { "epoch": 0.7577261809447559, "grad_norm": 0.59765625, "learning_rate": 8.416299724348695e-06, "loss": 2.3243, "step": 2366 }, { "epoch": 0.7580464371497198, "grad_norm": 0.515625, "learning_rate": 8.395387276632338e-06, "loss": 1.7567, "step": 2367 }, { "epoch": 0.7583666933546838, "grad_norm": 0.44921875, "learning_rate": 8.374495598364876e-06, "loss": 2.1981, "step": 2368 }, { "epoch": 0.7586869495596478, "grad_norm": 0.609375, "learning_rate": 8.353624715678093e-06, "loss": 2.1722, "step": 2369 }, { "epoch": 0.7590072057646117, "grad_norm": 0.53515625, "learning_rate": 8.332774654677802e-06, "loss": 2.2055, "step": 2370 }, { "epoch": 0.7593274619695757, "grad_norm": 0.55078125, "learning_rate": 8.311945441443759e-06, "loss": 2.5394, "step": 2371 }, { "epoch": 0.7596477181745396, "grad_norm": 0.5234375, "learning_rate": 8.291137102029644e-06, "loss": 2.7369, "step": 2372 }, { "epoch": 0.7599679743795036, "grad_norm": 0.4453125, "learning_rate": 8.270349662463032e-06, "loss": 2.2094, "step": 2373 }, { "epoch": 0.7602882305844676, "grad_norm": 0.373046875, "learning_rate": 8.249583148745357e-06, "loss": 2.7189, "step": 2374 }, { "epoch": 0.7606084867894315, "grad_norm": 0.51953125, "learning_rate": 8.228837586851859e-06, "loss": 2.4148, "step": 2375 }, { "epoch": 0.7609287429943955, "grad_norm": 0.58203125, "learning_rate": 8.20811300273159e-06, "loss": 2.1258, "step": 2376 }, { "epoch": 0.7612489991993595, "grad_norm": 0.52734375, "learning_rate": 8.187409422307365e-06, "loss": 1.9111, "step": 2377 }, { "epoch": 0.7615692554043234, "grad_norm": 0.427734375, "learning_rate": 8.166726871475713e-06, "loss": 2.2352, "step": 2378 }, { "epoch": 0.7618895116092874, "grad_norm": 0.546875, "learning_rate": 8.146065376106873e-06, "loss": 2.1225, "step": 2379 }, { "epoch": 0.7622097678142514, "grad_norm": 1.0078125, "learning_rate": 8.125424962044742e-06, "loss": 2.1848, "step": 2380 }, { "epoch": 0.7625300240192153, "grad_norm": 0.6015625, "learning_rate": 8.10480565510683e-06, "loss": 2.2108, "step": 2381 }, { "epoch": 0.7628502802241793, "grad_norm": 0.484375, "learning_rate": 8.084207481084282e-06, "loss": 2.359, "step": 2382 }, { "epoch": 0.7631705364291433, "grad_norm": 0.49609375, "learning_rate": 8.063630465741775e-06, "loss": 1.9673, "step": 2383 }, { "epoch": 0.7634907926341072, "grad_norm": 0.609375, "learning_rate": 8.043074634817541e-06, "loss": 2.294, "step": 2384 }, { "epoch": 0.7638110488390712, "grad_norm": 0.404296875, "learning_rate": 8.022540014023306e-06, "loss": 2.1543, "step": 2385 }, { "epoch": 0.7641313050440353, "grad_norm": 0.474609375, "learning_rate": 8.002026629044268e-06, "loss": 2.2717, "step": 2386 }, { "epoch": 0.7644515612489992, "grad_norm": 0.4609375, "learning_rate": 7.981534505539063e-06, "loss": 2.7924, "step": 2387 }, { "epoch": 0.7647718174539632, "grad_norm": 0.466796875, "learning_rate": 7.961063669139738e-06, "loss": 2.4514, "step": 2388 }, { "epoch": 0.7650920736589272, "grad_norm": 0.4609375, "learning_rate": 7.940614145451683e-06, "loss": 2.3244, "step": 2389 }, { "epoch": 0.7654123298638911, "grad_norm": 1.03125, "learning_rate": 7.920185960053688e-06, "loss": 1.8777, "step": 2390 }, { "epoch": 0.7657325860688551, "grad_norm": 0.5859375, "learning_rate": 7.899779138497796e-06, "loss": 2.4442, "step": 2391 }, { "epoch": 0.7660528422738191, "grad_norm": 0.9609375, "learning_rate": 7.879393706309354e-06, "loss": 2.174, "step": 2392 }, { "epoch": 0.766373098478783, "grad_norm": 0.6015625, "learning_rate": 7.859029688986952e-06, "loss": 1.5029, "step": 2393 }, { "epoch": 0.766693354683747, "grad_norm": 0.54296875, "learning_rate": 7.838687112002397e-06, "loss": 2.0569, "step": 2394 }, { "epoch": 0.767013610888711, "grad_norm": 0.5078125, "learning_rate": 7.818366000800666e-06, "loss": 2.0224, "step": 2395 }, { "epoch": 0.7673338670936749, "grad_norm": 0.474609375, "learning_rate": 7.798066380799907e-06, "loss": 2.3784, "step": 2396 }, { "epoch": 0.7676541232986389, "grad_norm": 0.478515625, "learning_rate": 7.777788277391348e-06, "loss": 2.0502, "step": 2397 }, { "epoch": 0.7679743795036029, "grad_norm": 0.455078125, "learning_rate": 7.757531715939358e-06, "loss": 1.6753, "step": 2398 }, { "epoch": 0.7682946357085668, "grad_norm": 0.515625, "learning_rate": 7.73729672178131e-06, "loss": 2.408, "step": 2399 }, { "epoch": 0.7686148919135308, "grad_norm": 0.48046875, "learning_rate": 7.717083320227627e-06, "loss": 2.0858, "step": 2400 }, { "epoch": 0.7689351481184948, "grad_norm": 0.59765625, "learning_rate": 7.69689153656172e-06, "loss": 2.4706, "step": 2401 }, { "epoch": 0.7692554043234587, "grad_norm": 1.125, "learning_rate": 7.676721396039963e-06, "loss": 2.2135, "step": 2402 }, { "epoch": 0.7695756605284227, "grad_norm": 0.482421875, "learning_rate": 7.65657292389163e-06, "loss": 2.5483, "step": 2403 }, { "epoch": 0.7698959167333868, "grad_norm": 0.515625, "learning_rate": 7.636446145318945e-06, "loss": 2.4565, "step": 2404 }, { "epoch": 0.7702161729383507, "grad_norm": 0.51171875, "learning_rate": 7.6163410854969476e-06, "loss": 2.0011, "step": 2405 }, { "epoch": 0.7705364291433147, "grad_norm": 0.55078125, "learning_rate": 7.596257769573534e-06, "loss": 2.4179, "step": 2406 }, { "epoch": 0.7708566853482787, "grad_norm": 1.046875, "learning_rate": 7.576196222669396e-06, "loss": 2.2249, "step": 2407 }, { "epoch": 0.7711769415532426, "grad_norm": 0.55859375, "learning_rate": 7.5561564698780065e-06, "loss": 2.3321, "step": 2408 }, { "epoch": 0.7714971977582066, "grad_norm": 0.79296875, "learning_rate": 7.536138536265563e-06, "loss": 2.2056, "step": 2409 }, { "epoch": 0.7718174539631706, "grad_norm": 0.5703125, "learning_rate": 7.516142446870989e-06, "loss": 2.3361, "step": 2410 }, { "epoch": 0.7721377101681345, "grad_norm": 0.4765625, "learning_rate": 7.496168226705852e-06, "loss": 1.8415, "step": 2411 }, { "epoch": 0.7724579663730985, "grad_norm": 0.5546875, "learning_rate": 7.476215900754413e-06, "loss": 2.214, "step": 2412 }, { "epoch": 0.7727782225780625, "grad_norm": 0.435546875, "learning_rate": 7.456285493973506e-06, "loss": 2.1685, "step": 2413 }, { "epoch": 0.7730984787830264, "grad_norm": 0.5078125, "learning_rate": 7.4363770312925625e-06, "loss": 1.6529, "step": 2414 }, { "epoch": 0.7734187349879904, "grad_norm": 0.5625, "learning_rate": 7.416490537613571e-06, "loss": 2.5159, "step": 2415 }, { "epoch": 0.7737389911929544, "grad_norm": 0.5390625, "learning_rate": 7.396626037811042e-06, "loss": 1.9631, "step": 2416 }, { "epoch": 0.7740592473979183, "grad_norm": 0.67578125, "learning_rate": 7.376783556731948e-06, "loss": 2.3239, "step": 2417 }, { "epoch": 0.7743795036028823, "grad_norm": 0.478515625, "learning_rate": 7.35696311919577e-06, "loss": 1.8956, "step": 2418 }, { "epoch": 0.7746997598078462, "grad_norm": 0.91015625, "learning_rate": 7.337164749994368e-06, "loss": 2.2789, "step": 2419 }, { "epoch": 0.7750200160128102, "grad_norm": 0.4921875, "learning_rate": 7.317388473892023e-06, "loss": 2.1517, "step": 2420 }, { "epoch": 0.7753402722177742, "grad_norm": 0.40234375, "learning_rate": 7.297634315625377e-06, "loss": 1.8434, "step": 2421 }, { "epoch": 0.7756605284227381, "grad_norm": 0.462890625, "learning_rate": 7.277902299903405e-06, "loss": 2.4362, "step": 2422 }, { "epoch": 0.7759807846277021, "grad_norm": 0.478515625, "learning_rate": 7.2581924514073916e-06, "loss": 2.2693, "step": 2423 }, { "epoch": 0.7763010408326662, "grad_norm": 0.64453125, "learning_rate": 7.238504794790893e-06, "loss": 2.4455, "step": 2424 }, { "epoch": 0.77662129703763, "grad_norm": 0.515625, "learning_rate": 7.218839354679683e-06, "loss": 2.5956, "step": 2425 }, { "epoch": 0.7769415532425941, "grad_norm": 0.86328125, "learning_rate": 7.199196155671795e-06, "loss": 2.187, "step": 2426 }, { "epoch": 0.7772618094475581, "grad_norm": 0.4921875, "learning_rate": 7.179575222337393e-06, "loss": 2.5182, "step": 2427 }, { "epoch": 0.777582065652522, "grad_norm": 0.84765625, "learning_rate": 7.159976579218819e-06, "loss": 1.9639, "step": 2428 }, { "epoch": 0.777902321857486, "grad_norm": 0.53515625, "learning_rate": 7.140400250830529e-06, "loss": 2.2367, "step": 2429 }, { "epoch": 0.77822257806245, "grad_norm": 0.58984375, "learning_rate": 7.12084626165907e-06, "loss": 2.4143, "step": 2430 }, { "epoch": 0.7785428342674139, "grad_norm": 0.55859375, "learning_rate": 7.10131463616302e-06, "loss": 1.9487, "step": 2431 }, { "epoch": 0.7788630904723779, "grad_norm": 0.64453125, "learning_rate": 7.0818053987730395e-06, "loss": 2.0972, "step": 2432 }, { "epoch": 0.7791833466773419, "grad_norm": 0.5234375, "learning_rate": 7.062318573891716e-06, "loss": 2.4615, "step": 2433 }, { "epoch": 0.7795036028823058, "grad_norm": 0.55859375, "learning_rate": 7.042854185893674e-06, "loss": 2.0897, "step": 2434 }, { "epoch": 0.7798238590872698, "grad_norm": 0.83984375, "learning_rate": 7.023412259125414e-06, "loss": 2.435, "step": 2435 }, { "epoch": 0.7801441152922338, "grad_norm": 0.451171875, "learning_rate": 7.00399281790537e-06, "loss": 2.3402, "step": 2436 }, { "epoch": 0.7804643714971977, "grad_norm": 0.5234375, "learning_rate": 6.984595886523848e-06, "loss": 2.1758, "step": 2437 }, { "epoch": 0.7807846277021617, "grad_norm": 0.416015625, "learning_rate": 6.965221489243007e-06, "loss": 2.3811, "step": 2438 }, { "epoch": 0.7811048839071257, "grad_norm": 0.337890625, "learning_rate": 6.94586965029678e-06, "loss": 1.8303, "step": 2439 }, { "epoch": 0.7814251401120896, "grad_norm": 0.53515625, "learning_rate": 6.926540393890948e-06, "loss": 2.0528, "step": 2440 }, { "epoch": 0.7817453963170536, "grad_norm": 0.3828125, "learning_rate": 6.907233744202985e-06, "loss": 2.1735, "step": 2441 }, { "epoch": 0.7820656525220177, "grad_norm": 0.51953125, "learning_rate": 6.887949725382117e-06, "loss": 2.369, "step": 2442 }, { "epoch": 0.7823859087269815, "grad_norm": 0.5703125, "learning_rate": 6.8686883615492666e-06, "loss": 2.4711, "step": 2443 }, { "epoch": 0.7827061649319456, "grad_norm": 0.478515625, "learning_rate": 6.84944967679701e-06, "loss": 2.5019, "step": 2444 }, { "epoch": 0.7830264211369096, "grad_norm": 0.55859375, "learning_rate": 6.830233695189539e-06, "loss": 2.4643, "step": 2445 }, { "epoch": 0.7833466773418735, "grad_norm": 0.67578125, "learning_rate": 6.8110404407626946e-06, "loss": 1.7722, "step": 2446 }, { "epoch": 0.7836669335468375, "grad_norm": 0.365234375, "learning_rate": 6.791869937523831e-06, "loss": 1.7393, "step": 2447 }, { "epoch": 0.7839871897518015, "grad_norm": 0.396484375, "learning_rate": 6.772722209451904e-06, "loss": 1.9059, "step": 2448 }, { "epoch": 0.7843074459567654, "grad_norm": 0.57421875, "learning_rate": 6.753597280497329e-06, "loss": 1.9944, "step": 2449 }, { "epoch": 0.7846277021617294, "grad_norm": 0.44921875, "learning_rate": 6.734495174582037e-06, "loss": 2.2997, "step": 2450 }, { "epoch": 0.7849479583666934, "grad_norm": 0.392578125, "learning_rate": 6.715415915599399e-06, "loss": 2.5738, "step": 2451 }, { "epoch": 0.7852682145716573, "grad_norm": 0.474609375, "learning_rate": 6.69635952741422e-06, "loss": 2.5298, "step": 2452 }, { "epoch": 0.7855884707766213, "grad_norm": 0.341796875, "learning_rate": 6.677326033862666e-06, "loss": 2.172, "step": 2453 }, { "epoch": 0.7859087269815853, "grad_norm": 0.37109375, "learning_rate": 6.658315458752315e-06, "loss": 1.9617, "step": 2454 }, { "epoch": 0.7862289831865492, "grad_norm": 0.5234375, "learning_rate": 6.639327825862035e-06, "loss": 2.0634, "step": 2455 }, { "epoch": 0.7865492393915132, "grad_norm": 0.5078125, "learning_rate": 6.620363158942016e-06, "loss": 2.2057, "step": 2456 }, { "epoch": 0.7868694955964772, "grad_norm": 0.5234375, "learning_rate": 6.601421481713724e-06, "loss": 2.4316, "step": 2457 }, { "epoch": 0.7871897518014411, "grad_norm": 0.5390625, "learning_rate": 6.582502817869868e-06, "loss": 2.2507, "step": 2458 }, { "epoch": 0.7875100080064051, "grad_norm": 0.53125, "learning_rate": 6.563607191074345e-06, "loss": 2.264, "step": 2459 }, { "epoch": 0.7878302642113691, "grad_norm": 0.408203125, "learning_rate": 6.5447346249622926e-06, "loss": 2.4022, "step": 2460 }, { "epoch": 0.788150520416333, "grad_norm": 0.515625, "learning_rate": 6.52588514313994e-06, "loss": 2.6929, "step": 2461 }, { "epoch": 0.788470776621297, "grad_norm": 0.5, "learning_rate": 6.507058769184699e-06, "loss": 2.1782, "step": 2462 }, { "epoch": 0.7887910328262611, "grad_norm": 0.8359375, "learning_rate": 6.488255526645029e-06, "loss": 1.8834, "step": 2463 }, { "epoch": 0.789111289031225, "grad_norm": 0.5234375, "learning_rate": 6.46947543904049e-06, "loss": 2.4088, "step": 2464 }, { "epoch": 0.789431545236189, "grad_norm": 0.486328125, "learning_rate": 6.450718529861663e-06, "loss": 2.7434, "step": 2465 }, { "epoch": 0.7897518014411529, "grad_norm": 0.365234375, "learning_rate": 6.431984822570147e-06, "loss": 2.0478, "step": 2466 }, { "epoch": 0.7900720576461169, "grad_norm": 0.9765625, "learning_rate": 6.413274340598496e-06, "loss": 2.5495, "step": 2467 }, { "epoch": 0.7903923138510809, "grad_norm": 0.68359375, "learning_rate": 6.394587107350258e-06, "loss": 1.9307, "step": 2468 }, { "epoch": 0.7907125700560448, "grad_norm": 0.37890625, "learning_rate": 6.3759231461998525e-06, "loss": 1.8534, "step": 2469 }, { "epoch": 0.7910328262610088, "grad_norm": 0.357421875, "learning_rate": 6.357282480492616e-06, "loss": 1.9793, "step": 2470 }, { "epoch": 0.7913530824659728, "grad_norm": 0.5234375, "learning_rate": 6.338665133544744e-06, "loss": 2.3047, "step": 2471 }, { "epoch": 0.7916733386709367, "grad_norm": 0.75, "learning_rate": 6.32007112864326e-06, "loss": 2.4341, "step": 2472 }, { "epoch": 0.7919935948759007, "grad_norm": 0.453125, "learning_rate": 6.3015004890459915e-06, "loss": 1.6601, "step": 2473 }, { "epoch": 0.7923138510808647, "grad_norm": 0.455078125, "learning_rate": 6.282953237981551e-06, "loss": 2.1922, "step": 2474 }, { "epoch": 0.7926341072858286, "grad_norm": 0.41796875, "learning_rate": 6.264429398649263e-06, "loss": 1.6936, "step": 2475 }, { "epoch": 0.7929543634907926, "grad_norm": 0.48828125, "learning_rate": 6.245928994219222e-06, "loss": 1.8489, "step": 2476 }, { "epoch": 0.7932746196957566, "grad_norm": 1.03125, "learning_rate": 6.227452047832155e-06, "loss": 2.7892, "step": 2477 }, { "epoch": 0.7935948759007205, "grad_norm": 0.478515625, "learning_rate": 6.208998582599476e-06, "loss": 2.4221, "step": 2478 }, { "epoch": 0.7939151321056845, "grad_norm": 0.50390625, "learning_rate": 6.1905686216032264e-06, "loss": 2.2199, "step": 2479 }, { "epoch": 0.7942353883106485, "grad_norm": 0.703125, "learning_rate": 6.1721621878960526e-06, "loss": 2.4266, "step": 2480 }, { "epoch": 0.7945556445156124, "grad_norm": 0.43359375, "learning_rate": 6.153779304501139e-06, "loss": 1.9723, "step": 2481 }, { "epoch": 0.7948759007205765, "grad_norm": 0.515625, "learning_rate": 6.135419994412272e-06, "loss": 2.4696, "step": 2482 }, { "epoch": 0.7951961569255405, "grad_norm": 0.48046875, "learning_rate": 6.1170842805936905e-06, "loss": 2.081, "step": 2483 }, { "epoch": 0.7955164131305044, "grad_norm": 0.9609375, "learning_rate": 6.098772185980159e-06, "loss": 2.4716, "step": 2484 }, { "epoch": 0.7958366693354684, "grad_norm": 0.65234375, "learning_rate": 6.080483733476883e-06, "loss": 1.9128, "step": 2485 }, { "epoch": 0.7961569255404324, "grad_norm": 0.5625, "learning_rate": 6.062218945959497e-06, "loss": 2.4175, "step": 2486 }, { "epoch": 0.7964771817453963, "grad_norm": 0.76171875, "learning_rate": 6.04397784627404e-06, "loss": 2.0246, "step": 2487 }, { "epoch": 0.7967974379503603, "grad_norm": 0.447265625, "learning_rate": 6.025760457236926e-06, "loss": 2.3537, "step": 2488 }, { "epoch": 0.7971176941553243, "grad_norm": 0.8515625, "learning_rate": 6.007566801634876e-06, "loss": 2.7204, "step": 2489 }, { "epoch": 0.7974379503602882, "grad_norm": 0.48828125, "learning_rate": 5.989396902224986e-06, "loss": 2.1869, "step": 2490 }, { "epoch": 0.7977582065652522, "grad_norm": 0.5859375, "learning_rate": 5.97125078173458e-06, "loss": 2.5756, "step": 2491 }, { "epoch": 0.7980784627702162, "grad_norm": 0.51171875, "learning_rate": 5.953128462861268e-06, "loss": 2.2436, "step": 2492 }, { "epoch": 0.7983987189751801, "grad_norm": 0.3984375, "learning_rate": 5.935029968272885e-06, "loss": 1.6643, "step": 2493 }, { "epoch": 0.7987189751801441, "grad_norm": 0.478515625, "learning_rate": 5.916955320607465e-06, "loss": 2.1245, "step": 2494 }, { "epoch": 0.7990392313851081, "grad_norm": 0.419921875, "learning_rate": 5.898904542473197e-06, "loss": 2.1199, "step": 2495 }, { "epoch": 0.799359487590072, "grad_norm": 0.4765625, "learning_rate": 5.8808776564484524e-06, "loss": 2.4544, "step": 2496 }, { "epoch": 0.799679743795036, "grad_norm": 0.58203125, "learning_rate": 5.862874685081673e-06, "loss": 2.386, "step": 2497 }, { "epoch": 0.8, "grad_norm": 0.60546875, "learning_rate": 5.844895650891416e-06, "loss": 2.2664, "step": 2498 }, { "epoch": 0.8003202562049639, "grad_norm": 0.453125, "learning_rate": 5.826940576366291e-06, "loss": 2.2293, "step": 2499 }, { "epoch": 0.800640512409928, "grad_norm": 0.474609375, "learning_rate": 5.809009483964936e-06, "loss": 1.8299, "step": 2500 }, { "epoch": 0.800960768614892, "grad_norm": 0.4453125, "learning_rate": 5.791102396115991e-06, "loss": 2.4019, "step": 2501 }, { "epoch": 0.8012810248198559, "grad_norm": 0.494140625, "learning_rate": 5.773219335218083e-06, "loss": 2.5534, "step": 2502 }, { "epoch": 0.8016012810248199, "grad_norm": 0.478515625, "learning_rate": 5.75536032363975e-06, "loss": 2.1044, "step": 2503 }, { "epoch": 0.8019215372297839, "grad_norm": 0.359375, "learning_rate": 5.737525383719499e-06, "loss": 1.7572, "step": 2504 }, { "epoch": 0.8022417934347478, "grad_norm": 0.46484375, "learning_rate": 5.719714537765688e-06, "loss": 2.5762, "step": 2505 }, { "epoch": 0.8025620496397118, "grad_norm": 0.81640625, "learning_rate": 5.7019278080565535e-06, "loss": 2.3808, "step": 2506 }, { "epoch": 0.8028823058446758, "grad_norm": 0.4765625, "learning_rate": 5.684165216840165e-06, "loss": 2.4309, "step": 2507 }, { "epoch": 0.8032025620496397, "grad_norm": 2.203125, "learning_rate": 5.666426786334405e-06, "loss": 2.4435, "step": 2508 }, { "epoch": 0.8035228182546037, "grad_norm": 0.48046875, "learning_rate": 5.648712538726911e-06, "loss": 2.1253, "step": 2509 }, { "epoch": 0.8038430744595677, "grad_norm": 0.62109375, "learning_rate": 5.631022496175115e-06, "loss": 2.5646, "step": 2510 }, { "epoch": 0.8041633306645316, "grad_norm": 0.490234375, "learning_rate": 5.61335668080612e-06, "loss": 2.106, "step": 2511 }, { "epoch": 0.8044835868694956, "grad_norm": 0.5078125, "learning_rate": 5.595715114716782e-06, "loss": 2.2236, "step": 2512 }, { "epoch": 0.8048038430744595, "grad_norm": 0.53125, "learning_rate": 5.578097819973577e-06, "loss": 2.451, "step": 2513 }, { "epoch": 0.8051240992794235, "grad_norm": 0.84765625, "learning_rate": 5.560504818612641e-06, "loss": 2.1605, "step": 2514 }, { "epoch": 0.8054443554843875, "grad_norm": 0.4765625, "learning_rate": 5.5429361326397285e-06, "loss": 2.5384, "step": 2515 }, { "epoch": 0.8057646116893514, "grad_norm": 0.36328125, "learning_rate": 5.525391784030182e-06, "loss": 2.0079, "step": 2516 }, { "epoch": 0.8060848678943154, "grad_norm": 0.375, "learning_rate": 5.507871794728872e-06, "loss": 1.5529, "step": 2517 }, { "epoch": 0.8064051240992794, "grad_norm": 0.388671875, "learning_rate": 5.4903761866502475e-06, "loss": 2.1881, "step": 2518 }, { "epoch": 0.8067253803042433, "grad_norm": 0.6015625, "learning_rate": 5.47290498167822e-06, "loss": 2.1702, "step": 2519 }, { "epoch": 0.8070456365092074, "grad_norm": 0.40234375, "learning_rate": 5.455458201666197e-06, "loss": 2.2788, "step": 2520 }, { "epoch": 0.8073658927141714, "grad_norm": 0.57421875, "learning_rate": 5.43803586843703e-06, "loss": 2.2958, "step": 2521 }, { "epoch": 0.8076861489191353, "grad_norm": 0.57421875, "learning_rate": 5.420638003783002e-06, "loss": 2.4007, "step": 2522 }, { "epoch": 0.8080064051240993, "grad_norm": 0.416015625, "learning_rate": 5.403264629465757e-06, "loss": 2.1257, "step": 2523 }, { "epoch": 0.8083266613290633, "grad_norm": 0.40234375, "learning_rate": 5.3859157672163564e-06, "loss": 2.2356, "step": 2524 }, { "epoch": 0.8086469175340272, "grad_norm": 0.58984375, "learning_rate": 5.368591438735149e-06, "loss": 2.5254, "step": 2525 }, { "epoch": 0.8089671737389912, "grad_norm": 0.5, "learning_rate": 5.3512916656918485e-06, "loss": 2.1586, "step": 2526 }, { "epoch": 0.8092874299439552, "grad_norm": 1.1875, "learning_rate": 5.334016469725409e-06, "loss": 2.3058, "step": 2527 }, { "epoch": 0.8096076861489191, "grad_norm": 0.99609375, "learning_rate": 5.31676587244406e-06, "loss": 2.061, "step": 2528 }, { "epoch": 0.8099279423538831, "grad_norm": 0.470703125, "learning_rate": 5.299539895425271e-06, "loss": 2.6985, "step": 2529 }, { "epoch": 0.8102481985588471, "grad_norm": 0.5234375, "learning_rate": 5.282338560215711e-06, "loss": 1.9756, "step": 2530 }, { "epoch": 0.810568454763811, "grad_norm": 0.4375, "learning_rate": 5.265161888331205e-06, "loss": 2.1155, "step": 2531 }, { "epoch": 0.810888710968775, "grad_norm": 1.1953125, "learning_rate": 5.248009901256776e-06, "loss": 2.2378, "step": 2532 }, { "epoch": 0.811208967173739, "grad_norm": 0.55078125, "learning_rate": 5.230882620446517e-06, "loss": 2.1956, "step": 2533 }, { "epoch": 0.8115292233787029, "grad_norm": 0.416015625, "learning_rate": 5.213780067323654e-06, "loss": 2.4922, "step": 2534 }, { "epoch": 0.8118494795836669, "grad_norm": 0.85546875, "learning_rate": 5.19670226328047e-06, "loss": 2.0935, "step": 2535 }, { "epoch": 0.8121697357886309, "grad_norm": 0.470703125, "learning_rate": 5.1796492296782905e-06, "loss": 1.7568, "step": 2536 }, { "epoch": 0.8124899919935948, "grad_norm": 0.498046875, "learning_rate": 5.162620987847466e-06, "loss": 2.527, "step": 2537 }, { "epoch": 0.8128102481985588, "grad_norm": 0.5234375, "learning_rate": 5.145617559087332e-06, "loss": 2.5302, "step": 2538 }, { "epoch": 0.8131305044035229, "grad_norm": 0.56640625, "learning_rate": 5.128638964666166e-06, "loss": 2.2782, "step": 2539 }, { "epoch": 0.8134507606084868, "grad_norm": 0.56640625, "learning_rate": 5.111685225821233e-06, "loss": 2.4849, "step": 2540 }, { "epoch": 0.8137710168134508, "grad_norm": 0.41796875, "learning_rate": 5.0947563637586545e-06, "loss": 2.1828, "step": 2541 }, { "epoch": 0.8140912730184148, "grad_norm": 0.453125, "learning_rate": 5.077852399653463e-06, "loss": 2.3564, "step": 2542 }, { "epoch": 0.8144115292233787, "grad_norm": 0.48828125, "learning_rate": 5.060973354649548e-06, "loss": 1.8303, "step": 2543 }, { "epoch": 0.8147317854283427, "grad_norm": 0.59375, "learning_rate": 5.044119249859627e-06, "loss": 2.3438, "step": 2544 }, { "epoch": 0.8150520416333067, "grad_norm": 0.4375, "learning_rate": 5.027290106365204e-06, "loss": 1.9135, "step": 2545 }, { "epoch": 0.8153722978382706, "grad_norm": 0.3515625, "learning_rate": 5.010485945216601e-06, "loss": 2.1691, "step": 2546 }, { "epoch": 0.8156925540432346, "grad_norm": 0.86328125, "learning_rate": 4.993706787432844e-06, "loss": 2.5787, "step": 2547 }, { "epoch": 0.8160128102481986, "grad_norm": 0.482421875, "learning_rate": 4.976952654001718e-06, "loss": 2.1237, "step": 2548 }, { "epoch": 0.8163330664531625, "grad_norm": 0.44921875, "learning_rate": 4.960223565879696e-06, "loss": 2.3624, "step": 2549 }, { "epoch": 0.8166533226581265, "grad_norm": 0.53125, "learning_rate": 4.94351954399192e-06, "loss": 1.776, "step": 2550 }, { "epoch": 0.8169735788630905, "grad_norm": 0.53125, "learning_rate": 4.926840609232183e-06, "loss": 2.3213, "step": 2551 }, { "epoch": 0.8172938350680544, "grad_norm": 1.0234375, "learning_rate": 4.910186782462908e-06, "loss": 2.7406, "step": 2552 }, { "epoch": 0.8176140912730184, "grad_norm": 0.431640625, "learning_rate": 4.8935580845150765e-06, "loss": 2.2117, "step": 2553 }, { "epoch": 0.8179343474779824, "grad_norm": 0.38671875, "learning_rate": 4.876954536188297e-06, "loss": 1.6675, "step": 2554 }, { "epoch": 0.8182546036829463, "grad_norm": 0.67578125, "learning_rate": 4.860376158250659e-06, "loss": 2.0331, "step": 2555 }, { "epoch": 0.8185748598879103, "grad_norm": 0.4375, "learning_rate": 4.84382297143881e-06, "loss": 2.0956, "step": 2556 }, { "epoch": 0.8188951160928744, "grad_norm": 0.59375, "learning_rate": 4.827294996457873e-06, "loss": 2.4513, "step": 2557 }, { "epoch": 0.8192153722978383, "grad_norm": 1.46875, "learning_rate": 4.810792253981439e-06, "loss": 2.3092, "step": 2558 }, { "epoch": 0.8195356285028023, "grad_norm": 0.34765625, "learning_rate": 4.79431476465152e-06, "loss": 1.9795, "step": 2559 }, { "epoch": 0.8198558847077662, "grad_norm": 0.43359375, "learning_rate": 4.777862549078579e-06, "loss": 2.2858, "step": 2560 }, { "epoch": 0.8201761409127302, "grad_norm": 0.421875, "learning_rate": 4.7614356278414235e-06, "loss": 2.7803, "step": 2561 }, { "epoch": 0.8204963971176942, "grad_norm": 1.0625, "learning_rate": 4.745034021487252e-06, "loss": 2.0644, "step": 2562 }, { "epoch": 0.8208166533226581, "grad_norm": 1.234375, "learning_rate": 4.728657750531581e-06, "loss": 1.9754, "step": 2563 }, { "epoch": 0.8211369095276221, "grad_norm": 1.390625, "learning_rate": 4.712306835458249e-06, "loss": 2.4305, "step": 2564 }, { "epoch": 0.8214571657325861, "grad_norm": 0.83203125, "learning_rate": 4.695981296719376e-06, "loss": 2.0684, "step": 2565 }, { "epoch": 0.82177742193755, "grad_norm": 0.40234375, "learning_rate": 4.679681154735341e-06, "loss": 2.2563, "step": 2566 }, { "epoch": 0.822097678142514, "grad_norm": 0.5078125, "learning_rate": 4.6634064298947315e-06, "loss": 2.162, "step": 2567 }, { "epoch": 0.822417934347478, "grad_norm": 0.44140625, "learning_rate": 4.647157142554396e-06, "loss": 2.3221, "step": 2568 }, { "epoch": 0.8227381905524419, "grad_norm": 0.453125, "learning_rate": 4.6309333130393125e-06, "loss": 2.307, "step": 2569 }, { "epoch": 0.8230584467574059, "grad_norm": 0.5703125, "learning_rate": 4.614734961642644e-06, "loss": 2.0686, "step": 2570 }, { "epoch": 0.8233787029623699, "grad_norm": 0.6328125, "learning_rate": 4.598562108625676e-06, "loss": 2.1714, "step": 2571 }, { "epoch": 0.8236989591673338, "grad_norm": 1.6796875, "learning_rate": 4.582414774217811e-06, "loss": 2.3787, "step": 2572 }, { "epoch": 0.8240192153722978, "grad_norm": 0.4609375, "learning_rate": 4.5662929786165014e-06, "loss": 2.2512, "step": 2573 }, { "epoch": 0.8243394715772618, "grad_norm": 0.376953125, "learning_rate": 4.550196741987306e-06, "loss": 2.0636, "step": 2574 }, { "epoch": 0.8246597277822257, "grad_norm": 0.50390625, "learning_rate": 4.534126084463755e-06, "loss": 1.9996, "step": 2575 }, { "epoch": 0.8249799839871897, "grad_norm": 0.45703125, "learning_rate": 4.518081026147439e-06, "loss": 1.8507, "step": 2576 }, { "epoch": 0.8253002401921538, "grad_norm": 0.58984375, "learning_rate": 4.502061587107889e-06, "loss": 2.1717, "step": 2577 }, { "epoch": 0.8256204963971177, "grad_norm": 2.109375, "learning_rate": 4.4860677873826025e-06, "loss": 2.3655, "step": 2578 }, { "epoch": 0.8259407526020817, "grad_norm": 0.443359375, "learning_rate": 4.4700996469770165e-06, "loss": 2.014, "step": 2579 }, { "epoch": 0.8262610088070457, "grad_norm": 0.515625, "learning_rate": 4.454157185864469e-06, "loss": 2.4824, "step": 2580 }, { "epoch": 0.8265812650120096, "grad_norm": 0.578125, "learning_rate": 4.438240423986154e-06, "loss": 2.2044, "step": 2581 }, { "epoch": 0.8269015212169736, "grad_norm": 0.455078125, "learning_rate": 4.422349381251162e-06, "loss": 2.2686, "step": 2582 }, { "epoch": 0.8272217774219376, "grad_norm": 0.400390625, "learning_rate": 4.406484077536377e-06, "loss": 2.5848, "step": 2583 }, { "epoch": 0.8275420336269015, "grad_norm": 0.56640625, "learning_rate": 4.3906445326865085e-06, "loss": 1.95, "step": 2584 }, { "epoch": 0.8278622898318655, "grad_norm": 0.50390625, "learning_rate": 4.374830766514037e-06, "loss": 1.8195, "step": 2585 }, { "epoch": 0.8281825460368295, "grad_norm": 0.453125, "learning_rate": 4.3590427987992126e-06, "loss": 1.9135, "step": 2586 }, { "epoch": 0.8285028022417934, "grad_norm": 0.474609375, "learning_rate": 4.3432806492899826e-06, "loss": 2.2409, "step": 2587 }, { "epoch": 0.8288230584467574, "grad_norm": 0.466796875, "learning_rate": 4.327544337702047e-06, "loss": 1.7218, "step": 2588 }, { "epoch": 0.8291433146517214, "grad_norm": 0.5078125, "learning_rate": 4.3118338837187415e-06, "loss": 2.4292, "step": 2589 }, { "epoch": 0.8294635708566853, "grad_norm": 0.490234375, "learning_rate": 4.296149306991098e-06, "loss": 2.1419, "step": 2590 }, { "epoch": 0.8297838270616493, "grad_norm": 0.640625, "learning_rate": 4.280490627137751e-06, "loss": 2.5364, "step": 2591 }, { "epoch": 0.8301040832666133, "grad_norm": 0.4296875, "learning_rate": 4.264857863744956e-06, "loss": 2.121, "step": 2592 }, { "epoch": 0.8304243394715772, "grad_norm": 1.4140625, "learning_rate": 4.24925103636655e-06, "loss": 2.5428, "step": 2593 }, { "epoch": 0.8307445956765412, "grad_norm": 0.5703125, "learning_rate": 4.233670164523937e-06, "loss": 2.0423, "step": 2594 }, { "epoch": 0.8310648518815053, "grad_norm": 0.419921875, "learning_rate": 4.218115267706021e-06, "loss": 2.4526, "step": 2595 }, { "epoch": 0.8313851080864691, "grad_norm": 0.396484375, "learning_rate": 4.2025863653692706e-06, "loss": 2.3808, "step": 2596 }, { "epoch": 0.8317053642914332, "grad_norm": 0.53515625, "learning_rate": 4.1870834769375925e-06, "loss": 2.3936, "step": 2597 }, { "epoch": 0.8320256204963972, "grad_norm": 0.52734375, "learning_rate": 4.171606621802377e-06, "loss": 2.443, "step": 2598 }, { "epoch": 0.8323458767013611, "grad_norm": 1.296875, "learning_rate": 4.1561558193224435e-06, "loss": 2.0169, "step": 2599 }, { "epoch": 0.8326661329063251, "grad_norm": 0.37890625, "learning_rate": 4.14073108882404e-06, "loss": 2.1607, "step": 2600 }, { "epoch": 0.8329863891112891, "grad_norm": 0.5234375, "learning_rate": 4.125332449600766e-06, "loss": 2.3544, "step": 2601 }, { "epoch": 0.833306645316253, "grad_norm": 0.52734375, "learning_rate": 4.109959920913637e-06, "loss": 2.0641, "step": 2602 }, { "epoch": 0.833626901521217, "grad_norm": 0.373046875, "learning_rate": 4.094613521990956e-06, "loss": 2.1128, "step": 2603 }, { "epoch": 0.833947157726181, "grad_norm": 0.52734375, "learning_rate": 4.079293272028395e-06, "loss": 1.7678, "step": 2604 }, { "epoch": 0.8342674139311449, "grad_norm": 0.76953125, "learning_rate": 4.063999190188869e-06, "loss": 2.2154, "step": 2605 }, { "epoch": 0.8345876701361089, "grad_norm": 0.482421875, "learning_rate": 4.04873129560259e-06, "loss": 2.4884, "step": 2606 }, { "epoch": 0.8349079263410728, "grad_norm": 0.5546875, "learning_rate": 4.033489607367011e-06, "loss": 2.6735, "step": 2607 }, { "epoch": 0.8352281825460368, "grad_norm": 0.69140625, "learning_rate": 4.018274144546802e-06, "loss": 2.579, "step": 2608 }, { "epoch": 0.8355484387510008, "grad_norm": 0.609375, "learning_rate": 4.003084926173819e-06, "loss": 2.5984, "step": 2609 }, { "epoch": 0.8358686949559647, "grad_norm": 0.55078125, "learning_rate": 3.987921971247122e-06, "loss": 2.7333, "step": 2610 }, { "epoch": 0.8361889511609287, "grad_norm": 0.439453125, "learning_rate": 3.972785298732881e-06, "loss": 2.0657, "step": 2611 }, { "epoch": 0.8365092073658927, "grad_norm": 0.546875, "learning_rate": 3.9576749275644175e-06, "loss": 2.1598, "step": 2612 }, { "epoch": 0.8368294635708566, "grad_norm": 0.45703125, "learning_rate": 3.942590876642146e-06, "loss": 2.1292, "step": 2613 }, { "epoch": 0.8371497197758206, "grad_norm": 1.203125, "learning_rate": 3.927533164833561e-06, "loss": 2.3755, "step": 2614 }, { "epoch": 0.8374699759807847, "grad_norm": 0.359375, "learning_rate": 3.91250181097321e-06, "loss": 2.1559, "step": 2615 }, { "epoch": 0.8377902321857486, "grad_norm": 0.462890625, "learning_rate": 3.897496833862677e-06, "loss": 2.4086, "step": 2616 }, { "epoch": 0.8381104883907126, "grad_norm": 2.890625, "learning_rate": 3.882518252270531e-06, "loss": 2.1332, "step": 2617 }, { "epoch": 0.8384307445956766, "grad_norm": 0.5546875, "learning_rate": 3.867566084932367e-06, "loss": 1.9732, "step": 2618 }, { "epoch": 0.8387510008006405, "grad_norm": 0.87890625, "learning_rate": 3.852640350550696e-06, "loss": 2.4946, "step": 2619 }, { "epoch": 0.8390712570056045, "grad_norm": 0.5546875, "learning_rate": 3.837741067794992e-06, "loss": 2.6402, "step": 2620 }, { "epoch": 0.8393915132105685, "grad_norm": 1.625, "learning_rate": 3.822868255301637e-06, "loss": 2.6491, "step": 2621 }, { "epoch": 0.8397117694155324, "grad_norm": 0.64453125, "learning_rate": 3.808021931673905e-06, "loss": 2.4719, "step": 2622 }, { "epoch": 0.8400320256204964, "grad_norm": 0.474609375, "learning_rate": 3.7932021154819225e-06, "loss": 1.9543, "step": 2623 }, { "epoch": 0.8403522818254604, "grad_norm": 0.4453125, "learning_rate": 3.7784088252626914e-06, "loss": 1.8275, "step": 2624 }, { "epoch": 0.8406725380304243, "grad_norm": 0.796875, "learning_rate": 3.7636420795199996e-06, "loss": 1.8942, "step": 2625 }, { "epoch": 0.8409927942353883, "grad_norm": 0.48828125, "learning_rate": 3.7489018967244494e-06, "loss": 1.9354, "step": 2626 }, { "epoch": 0.8413130504403523, "grad_norm": 0.3515625, "learning_rate": 3.7341882953134206e-06, "loss": 1.8982, "step": 2627 }, { "epoch": 0.8416333066453162, "grad_norm": 1.0625, "learning_rate": 3.7195012936910383e-06, "loss": 2.8696, "step": 2628 }, { "epoch": 0.8419535628502802, "grad_norm": 0.6796875, "learning_rate": 3.7048409102281577e-06, "loss": 2.3575, "step": 2629 }, { "epoch": 0.8422738190552442, "grad_norm": 0.82421875, "learning_rate": 3.690207163262341e-06, "loss": 2.4408, "step": 2630 }, { "epoch": 0.8425940752602081, "grad_norm": 0.41015625, "learning_rate": 3.675600071097818e-06, "loss": 2.1077, "step": 2631 }, { "epoch": 0.8429143314651721, "grad_norm": 0.45703125, "learning_rate": 3.6610196520055124e-06, "loss": 2.1918, "step": 2632 }, { "epoch": 0.8432345876701361, "grad_norm": 0.57421875, "learning_rate": 3.646465924222947e-06, "loss": 2.4826, "step": 2633 }, { "epoch": 0.8435548438751, "grad_norm": 0.494140625, "learning_rate": 3.631938905954277e-06, "loss": 2.3311, "step": 2634 }, { "epoch": 0.8438751000800641, "grad_norm": 0.447265625, "learning_rate": 3.6174386153702473e-06, "loss": 2.4787, "step": 2635 }, { "epoch": 0.8441953562850281, "grad_norm": 0.44140625, "learning_rate": 3.6029650706081775e-06, "loss": 1.9944, "step": 2636 }, { "epoch": 0.844515612489992, "grad_norm": 0.478515625, "learning_rate": 3.588518289771908e-06, "loss": 2.4228, "step": 2637 }, { "epoch": 0.844835868694956, "grad_norm": 0.42578125, "learning_rate": 3.5740982909318406e-06, "loss": 2.4106, "step": 2638 }, { "epoch": 0.84515612489992, "grad_norm": 0.515625, "learning_rate": 3.5597050921248347e-06, "loss": 2.4529, "step": 2639 }, { "epoch": 0.8454763811048839, "grad_norm": 0.478515625, "learning_rate": 3.5453387113542737e-06, "loss": 2.5682, "step": 2640 }, { "epoch": 0.8457966373098479, "grad_norm": 1.609375, "learning_rate": 3.5309991665899545e-06, "loss": 2.0301, "step": 2641 }, { "epoch": 0.8461168935148119, "grad_norm": 0.42578125, "learning_rate": 3.516686475768127e-06, "loss": 2.1756, "step": 2642 }, { "epoch": 0.8464371497197758, "grad_norm": 0.8046875, "learning_rate": 3.502400656791455e-06, "loss": 1.9134, "step": 2643 }, { "epoch": 0.8467574059247398, "grad_norm": 0.890625, "learning_rate": 3.4881417275289847e-06, "loss": 2.3704, "step": 2644 }, { "epoch": 0.8470776621297038, "grad_norm": 0.57421875, "learning_rate": 3.4739097058161114e-06, "loss": 2.5802, "step": 2645 }, { "epoch": 0.8473979183346677, "grad_norm": 0.45703125, "learning_rate": 3.459704609454614e-06, "loss": 1.5905, "step": 2646 }, { "epoch": 0.8477181745396317, "grad_norm": 0.55859375, "learning_rate": 3.445526456212553e-06, "loss": 2.2604, "step": 2647 }, { "epoch": 0.8480384307445957, "grad_norm": 0.482421875, "learning_rate": 3.4313752638243075e-06, "loss": 2.3032, "step": 2648 }, { "epoch": 0.8483586869495596, "grad_norm": 0.447265625, "learning_rate": 3.41725104999053e-06, "loss": 2.215, "step": 2649 }, { "epoch": 0.8486789431545236, "grad_norm": 0.56640625, "learning_rate": 3.4031538323781254e-06, "loss": 2.3237, "step": 2650 }, { "epoch": 0.8489991993594876, "grad_norm": 0.6328125, "learning_rate": 3.389083628620235e-06, "loss": 1.5601, "step": 2651 }, { "epoch": 0.8493194555644515, "grad_norm": 0.283203125, "learning_rate": 3.3750404563162085e-06, "loss": 2.3487, "step": 2652 }, { "epoch": 0.8496397117694156, "grad_norm": 0.55859375, "learning_rate": 3.3610243330315733e-06, "loss": 1.9077, "step": 2653 }, { "epoch": 0.8499599679743794, "grad_norm": 0.427734375, "learning_rate": 3.347035276298052e-06, "loss": 1.879, "step": 2654 }, { "epoch": 0.8502802241793435, "grad_norm": 0.546875, "learning_rate": 3.3330733036134765e-06, "loss": 2.0285, "step": 2655 }, { "epoch": 0.8506004803843075, "grad_norm": 1.21875, "learning_rate": 3.3191384324418236e-06, "loss": 2.5377, "step": 2656 }, { "epoch": 0.8509207365892714, "grad_norm": 0.498046875, "learning_rate": 3.3052306802131644e-06, "loss": 1.8052, "step": 2657 }, { "epoch": 0.8512409927942354, "grad_norm": 0.5390625, "learning_rate": 3.291350064323659e-06, "loss": 1.8514, "step": 2658 }, { "epoch": 0.8515612489991994, "grad_norm": 0.490234375, "learning_rate": 3.277496602135496e-06, "loss": 2.2399, "step": 2659 }, { "epoch": 0.8518815052041633, "grad_norm": 0.419921875, "learning_rate": 3.2636703109769406e-06, "loss": 2.274, "step": 2660 }, { "epoch": 0.8522017614091273, "grad_norm": 0.578125, "learning_rate": 3.249871208142238e-06, "loss": 2.2057, "step": 2661 }, { "epoch": 0.8525220176140913, "grad_norm": 0.55078125, "learning_rate": 3.236099310891638e-06, "loss": 2.5783, "step": 2662 }, { "epoch": 0.8528422738190552, "grad_norm": 0.65625, "learning_rate": 3.2223546364513636e-06, "loss": 2.2508, "step": 2663 }, { "epoch": 0.8531625300240192, "grad_norm": 0.46875, "learning_rate": 3.2086372020135815e-06, "loss": 2.5538, "step": 2664 }, { "epoch": 0.8534827862289832, "grad_norm": 0.625, "learning_rate": 3.194947024736392e-06, "loss": 2.0804, "step": 2665 }, { "epoch": 0.8538030424339471, "grad_norm": 0.640625, "learning_rate": 3.1812841217438023e-06, "loss": 2.5366, "step": 2666 }, { "epoch": 0.8541232986389111, "grad_norm": 0.66015625, "learning_rate": 3.1676485101256844e-06, "loss": 2.1118, "step": 2667 }, { "epoch": 0.8544435548438751, "grad_norm": 0.578125, "learning_rate": 3.1540402069378084e-06, "loss": 2.3653, "step": 2668 }, { "epoch": 0.854763811048839, "grad_norm": 0.5625, "learning_rate": 3.140459229201753e-06, "loss": 2.4334, "step": 2669 }, { "epoch": 0.855084067253803, "grad_norm": 0.5, "learning_rate": 3.1269055939049356e-06, "loss": 2.5745, "step": 2670 }, { "epoch": 0.855404323458767, "grad_norm": 0.376953125, "learning_rate": 3.11337931800057e-06, "loss": 1.7917, "step": 2671 }, { "epoch": 0.8557245796637309, "grad_norm": 0.61328125, "learning_rate": 3.0998804184076495e-06, "loss": 2.3825, "step": 2672 }, { "epoch": 0.856044835868695, "grad_norm": 0.59765625, "learning_rate": 3.0864089120109186e-06, "loss": 2.1879, "step": 2673 }, { "epoch": 0.856365092073659, "grad_norm": 0.408203125, "learning_rate": 3.07296481566087e-06, "loss": 2.2279, "step": 2674 }, { "epoch": 0.8566853482786229, "grad_norm": 0.62890625, "learning_rate": 3.059548146173691e-06, "loss": 2.1929, "step": 2675 }, { "epoch": 0.8570056044835869, "grad_norm": 0.625, "learning_rate": 3.046158920331277e-06, "loss": 2.2886, "step": 2676 }, { "epoch": 0.8573258606885509, "grad_norm": 0.55078125, "learning_rate": 3.032797154881198e-06, "loss": 2.8405, "step": 2677 }, { "epoch": 0.8576461168935148, "grad_norm": 0.48046875, "learning_rate": 3.019462866536668e-06, "loss": 1.712, "step": 2678 }, { "epoch": 0.8579663730984788, "grad_norm": 0.47265625, "learning_rate": 3.0061560719765376e-06, "loss": 2.4549, "step": 2679 }, { "epoch": 0.8582866293034428, "grad_norm": 0.61328125, "learning_rate": 2.9928767878452698e-06, "loss": 2.2048, "step": 2680 }, { "epoch": 0.8586068855084067, "grad_norm": 0.828125, "learning_rate": 2.9796250307528973e-06, "loss": 2.1138, "step": 2681 }, { "epoch": 0.8589271417133707, "grad_norm": 3.390625, "learning_rate": 2.9664008172750557e-06, "loss": 2.0606, "step": 2682 }, { "epoch": 0.8592473979183347, "grad_norm": 0.5859375, "learning_rate": 2.9532041639528967e-06, "loss": 2.1709, "step": 2683 }, { "epoch": 0.8595676541232986, "grad_norm": 0.61328125, "learning_rate": 2.9400350872931138e-06, "loss": 2.5132, "step": 2684 }, { "epoch": 0.8598879103282626, "grad_norm": 0.5625, "learning_rate": 2.926893603767908e-06, "loss": 2.2391, "step": 2685 }, { "epoch": 0.8602081665332266, "grad_norm": 0.62890625, "learning_rate": 2.9137797298149584e-06, "loss": 2.4567, "step": 2686 }, { "epoch": 0.8605284227381905, "grad_norm": 0.61328125, "learning_rate": 2.9006934818374187e-06, "loss": 2.511, "step": 2687 }, { "epoch": 0.8608486789431545, "grad_norm": 0.48046875, "learning_rate": 2.8876348762038846e-06, "loss": 2.4784, "step": 2688 }, { "epoch": 0.8611689351481185, "grad_norm": 0.625, "learning_rate": 2.874603929248365e-06, "loss": 1.7693, "step": 2689 }, { "epoch": 0.8614891913530824, "grad_norm": 0.421875, "learning_rate": 2.8616006572702857e-06, "loss": 2.4669, "step": 2690 }, { "epoch": 0.8618094475580464, "grad_norm": 0.453125, "learning_rate": 2.848625076534453e-06, "loss": 1.9782, "step": 2691 }, { "epoch": 0.8621297037630105, "grad_norm": 0.51953125, "learning_rate": 2.835677203271037e-06, "loss": 2.2473, "step": 2692 }, { "epoch": 0.8624499599679744, "grad_norm": 0.6328125, "learning_rate": 2.8227570536755425e-06, "loss": 2.3449, "step": 2693 }, { "epoch": 0.8627702161729384, "grad_norm": 0.50390625, "learning_rate": 2.8098646439088106e-06, "loss": 2.292, "step": 2694 }, { "epoch": 0.8630904723779024, "grad_norm": 0.498046875, "learning_rate": 2.796999990096974e-06, "loss": 2.395, "step": 2695 }, { "epoch": 0.8634107285828663, "grad_norm": 0.43359375, "learning_rate": 2.7841631083314584e-06, "loss": 2.19, "step": 2696 }, { "epoch": 0.8637309847878303, "grad_norm": 0.357421875, "learning_rate": 2.771354014668931e-06, "loss": 1.5913, "step": 2697 }, { "epoch": 0.8640512409927943, "grad_norm": 0.470703125, "learning_rate": 2.75857272513132e-06, "loss": 2.1674, "step": 2698 }, { "epoch": 0.8643714971977582, "grad_norm": 0.77734375, "learning_rate": 2.74581925570577e-06, "loss": 2.3245, "step": 2699 }, { "epoch": 0.8646917534027222, "grad_norm": 0.51953125, "learning_rate": 2.733093622344629e-06, "loss": 1.8666, "step": 2700 }, { "epoch": 0.8650120096076861, "grad_norm": 0.51171875, "learning_rate": 2.7203958409654227e-06, "loss": 2.2844, "step": 2701 }, { "epoch": 0.8653322658126501, "grad_norm": 0.94140625, "learning_rate": 2.707725927450844e-06, "loss": 1.9818, "step": 2702 }, { "epoch": 0.8656525220176141, "grad_norm": 0.578125, "learning_rate": 2.6950838976487287e-06, "loss": 1.9651, "step": 2703 }, { "epoch": 0.865972778222578, "grad_norm": 1.3203125, "learning_rate": 2.68246976737202e-06, "loss": 2.1429, "step": 2704 }, { "epoch": 0.866293034427542, "grad_norm": 0.7109375, "learning_rate": 2.6698835523987832e-06, "loss": 2.1287, "step": 2705 }, { "epoch": 0.866613290632506, "grad_norm": 0.5234375, "learning_rate": 2.657325268472155e-06, "loss": 2.2263, "step": 2706 }, { "epoch": 0.8669335468374699, "grad_norm": 0.51953125, "learning_rate": 2.6447949313003456e-06, "loss": 2.1328, "step": 2707 }, { "epoch": 0.8672538030424339, "grad_norm": 0.640625, "learning_rate": 2.6322925565565954e-06, "loss": 2.1588, "step": 2708 }, { "epoch": 0.8675740592473979, "grad_norm": 0.55078125, "learning_rate": 2.619818159879178e-06, "loss": 2.1747, "step": 2709 }, { "epoch": 0.8678943154523618, "grad_norm": 0.54296875, "learning_rate": 2.607371756871374e-06, "loss": 1.9288, "step": 2710 }, { "epoch": 0.8682145716573259, "grad_norm": 0.59375, "learning_rate": 2.594953363101435e-06, "loss": 2.3231, "step": 2711 }, { "epoch": 0.8685348278622899, "grad_norm": 0.55859375, "learning_rate": 2.5825629941025887e-06, "loss": 2.2347, "step": 2712 }, { "epoch": 0.8688550840672538, "grad_norm": 0.50390625, "learning_rate": 2.5702006653730072e-06, "loss": 2.5465, "step": 2713 }, { "epoch": 0.8691753402722178, "grad_norm": 0.984375, "learning_rate": 2.557866392375788e-06, "loss": 2.163, "step": 2714 }, { "epoch": 0.8694955964771818, "grad_norm": 0.56640625, "learning_rate": 2.545560190538937e-06, "loss": 2.5304, "step": 2715 }, { "epoch": 0.8698158526821457, "grad_norm": 0.6015625, "learning_rate": 2.5332820752553498e-06, "loss": 2.4132, "step": 2716 }, { "epoch": 0.8701361088871097, "grad_norm": 0.419921875, "learning_rate": 2.521032061882786e-06, "loss": 2.2646, "step": 2717 }, { "epoch": 0.8704563650920737, "grad_norm": 0.83984375, "learning_rate": 2.5088101657438605e-06, "loss": 2.1072, "step": 2718 }, { "epoch": 0.8707766212970376, "grad_norm": 0.42578125, "learning_rate": 2.4966164021260092e-06, "loss": 2.3744, "step": 2719 }, { "epoch": 0.8710968775020016, "grad_norm": 0.53515625, "learning_rate": 2.4844507862814835e-06, "loss": 1.145, "step": 2720 }, { "epoch": 0.8714171337069656, "grad_norm": 0.396484375, "learning_rate": 2.472313333427334e-06, "loss": 1.9044, "step": 2721 }, { "epoch": 0.8717373899119295, "grad_norm": 0.515625, "learning_rate": 2.4602040587453774e-06, "loss": 1.9285, "step": 2722 }, { "epoch": 0.8720576461168935, "grad_norm": 0.87109375, "learning_rate": 2.4481229773821835e-06, "loss": 2.2034, "step": 2723 }, { "epoch": 0.8723779023218575, "grad_norm": 0.6171875, "learning_rate": 2.436070104449059e-06, "loss": 2.475, "step": 2724 }, { "epoch": 0.8726981585268214, "grad_norm": 0.671875, "learning_rate": 2.4240454550220367e-06, "loss": 2.528, "step": 2725 }, { "epoch": 0.8730184147317854, "grad_norm": 0.55859375, "learning_rate": 2.412049044141823e-06, "loss": 2.067, "step": 2726 }, { "epoch": 0.8733386709367494, "grad_norm": 0.51171875, "learning_rate": 2.4000808868138212e-06, "loss": 1.8833, "step": 2727 }, { "epoch": 0.8736589271417133, "grad_norm": 0.59765625, "learning_rate": 2.3881409980080954e-06, "loss": 2.4556, "step": 2728 }, { "epoch": 0.8739791833466773, "grad_norm": 0.50390625, "learning_rate": 2.3762293926593433e-06, "loss": 1.9098, "step": 2729 }, { "epoch": 0.8742994395516414, "grad_norm": 0.5859375, "learning_rate": 2.3643460856668823e-06, "loss": 2.4019, "step": 2730 }, { "epoch": 0.8746196957566053, "grad_norm": 0.5078125, "learning_rate": 2.3524910918946484e-06, "loss": 2.2306, "step": 2731 }, { "epoch": 0.8749399519615693, "grad_norm": 0.484375, "learning_rate": 2.340664426171149e-06, "loss": 2.2344, "step": 2732 }, { "epoch": 0.8752602081665333, "grad_norm": 0.453125, "learning_rate": 2.3288661032894564e-06, "loss": 2.1087, "step": 2733 }, { "epoch": 0.8755804643714972, "grad_norm": 0.345703125, "learning_rate": 2.3170961380071993e-06, "loss": 2.0231, "step": 2734 }, { "epoch": 0.8759007205764612, "grad_norm": 0.640625, "learning_rate": 2.305354545046537e-06, "loss": 2.5689, "step": 2735 }, { "epoch": 0.8762209767814252, "grad_norm": 0.609375, "learning_rate": 2.2936413390941325e-06, "loss": 2.1854, "step": 2736 }, { "epoch": 0.8765412329863891, "grad_norm": 0.49609375, "learning_rate": 2.2819565348011466e-06, "loss": 2.2984, "step": 2737 }, { "epoch": 0.8768614891913531, "grad_norm": 0.5234375, "learning_rate": 2.2703001467832137e-06, "loss": 2.3775, "step": 2738 }, { "epoch": 0.8771817453963171, "grad_norm": 0.69140625, "learning_rate": 2.258672189620431e-06, "loss": 2.5423, "step": 2739 }, { "epoch": 0.877502001601281, "grad_norm": 1.5, "learning_rate": 2.2470726778573146e-06, "loss": 1.8138, "step": 2740 }, { "epoch": 0.877822257806245, "grad_norm": 0.4375, "learning_rate": 2.235501626002817e-06, "loss": 1.9844, "step": 2741 }, { "epoch": 0.878142514011209, "grad_norm": 0.5703125, "learning_rate": 2.2239590485302887e-06, "loss": 1.8902, "step": 2742 }, { "epoch": 0.8784627702161729, "grad_norm": 0.59765625, "learning_rate": 2.212444959877466e-06, "loss": 1.7303, "step": 2743 }, { "epoch": 0.8787830264211369, "grad_norm": 0.53125, "learning_rate": 2.2009593744464456e-06, "loss": 2.1603, "step": 2744 }, { "epoch": 0.8791032826261009, "grad_norm": 0.494140625, "learning_rate": 2.1895023066036745e-06, "loss": 1.5858, "step": 2745 }, { "epoch": 0.8794235388310648, "grad_norm": 0.6640625, "learning_rate": 2.1780737706799255e-06, "loss": 2.3251, "step": 2746 }, { "epoch": 0.8797437950360288, "grad_norm": 0.6171875, "learning_rate": 2.1666737809702963e-06, "loss": 2.1973, "step": 2747 }, { "epoch": 0.8800640512409927, "grad_norm": 0.5546875, "learning_rate": 2.1553023517341566e-06, "loss": 2.2133, "step": 2748 }, { "epoch": 0.8803843074459567, "grad_norm": 0.44921875, "learning_rate": 2.1439594971951666e-06, "loss": 2.7235, "step": 2749 }, { "epoch": 0.8807045636509208, "grad_norm": 0.470703125, "learning_rate": 2.1326452315412413e-06, "loss": 1.9469, "step": 2750 }, { "epoch": 0.8810248198558847, "grad_norm": 2.046875, "learning_rate": 2.1213595689245386e-06, "loss": 2.0951, "step": 2751 }, { "epoch": 0.8813450760608487, "grad_norm": 1.6796875, "learning_rate": 2.110102523461435e-06, "loss": 2.1507, "step": 2752 }, { "epoch": 0.8816653322658127, "grad_norm": 0.494140625, "learning_rate": 2.098874109232521e-06, "loss": 2.0099, "step": 2753 }, { "epoch": 0.8819855884707766, "grad_norm": 0.98828125, "learning_rate": 2.08767434028255e-06, "loss": 1.7282, "step": 2754 }, { "epoch": 0.8823058446757406, "grad_norm": 6.3125, "learning_rate": 2.07650323062048e-06, "loss": 2.3165, "step": 2755 }, { "epoch": 0.8826261008807046, "grad_norm": 0.419921875, "learning_rate": 2.065360794219395e-06, "loss": 2.093, "step": 2756 }, { "epoch": 0.8829463570856685, "grad_norm": 0.890625, "learning_rate": 2.054247045016522e-06, "loss": 1.8592, "step": 2757 }, { "epoch": 0.8832666132906325, "grad_norm": 0.7265625, "learning_rate": 2.0431619969132065e-06, "loss": 2.0496, "step": 2758 }, { "epoch": 0.8835868694955965, "grad_norm": 0.6328125, "learning_rate": 2.032105663774897e-06, "loss": 2.394, "step": 2759 }, { "epoch": 0.8839071257005604, "grad_norm": 0.3671875, "learning_rate": 2.021078059431114e-06, "loss": 1.8352, "step": 2760 }, { "epoch": 0.8842273819055244, "grad_norm": 0.96484375, "learning_rate": 2.010079197675463e-06, "loss": 1.8796, "step": 2761 }, { "epoch": 0.8845476381104884, "grad_norm": 0.71875, "learning_rate": 1.9991090922655694e-06, "loss": 1.7734, "step": 2762 }, { "epoch": 0.8848678943154523, "grad_norm": 0.99609375, "learning_rate": 1.9881677569231094e-06, "loss": 2.5961, "step": 2763 }, { "epoch": 0.8851881505204163, "grad_norm": 0.48046875, "learning_rate": 1.977255205333775e-06, "loss": 1.9614, "step": 2764 }, { "epoch": 0.8855084067253803, "grad_norm": 0.60546875, "learning_rate": 1.966371451147239e-06, "loss": 2.529, "step": 2765 }, { "epoch": 0.8858286629303442, "grad_norm": 0.640625, "learning_rate": 1.9555165079771703e-06, "loss": 2.4828, "step": 2766 }, { "epoch": 0.8861489191353082, "grad_norm": 0.75390625, "learning_rate": 1.9446903894011937e-06, "loss": 2.0968, "step": 2767 }, { "epoch": 0.8864691753402723, "grad_norm": 0.47265625, "learning_rate": 1.933893108960866e-06, "loss": 1.7931, "step": 2768 }, { "epoch": 0.8867894315452362, "grad_norm": 0.6484375, "learning_rate": 1.9231246801617033e-06, "loss": 2.7387, "step": 2769 }, { "epoch": 0.8871096877502002, "grad_norm": 0.486328125, "learning_rate": 1.9123851164731004e-06, "loss": 2.1077, "step": 2770 }, { "epoch": 0.8874299439551642, "grad_norm": 0.458984375, "learning_rate": 1.901674431328368e-06, "loss": 2.2398, "step": 2771 }, { "epoch": 0.8877502001601281, "grad_norm": 0.458984375, "learning_rate": 1.8909926381246862e-06, "loss": 2.5055, "step": 2772 }, { "epoch": 0.8880704563650921, "grad_norm": 0.41015625, "learning_rate": 1.8803397502231034e-06, "loss": 2.0305, "step": 2773 }, { "epoch": 0.8883907125700561, "grad_norm": 0.47265625, "learning_rate": 1.8697157809485055e-06, "loss": 2.3018, "step": 2774 }, { "epoch": 0.88871096877502, "grad_norm": 0.56640625, "learning_rate": 1.8591207435896096e-06, "loss": 2.3575, "step": 2775 }, { "epoch": 0.889031224979984, "grad_norm": 0.486328125, "learning_rate": 1.8485546513989372e-06, "loss": 2.1141, "step": 2776 }, { "epoch": 0.889351481184948, "grad_norm": 0.6640625, "learning_rate": 1.8380175175928216e-06, "loss": 1.9234, "step": 2777 }, { "epoch": 0.8896717373899119, "grad_norm": 0.455078125, "learning_rate": 1.8275093553513533e-06, "loss": 2.3495, "step": 2778 }, { "epoch": 0.8899919935948759, "grad_norm": 0.4921875, "learning_rate": 1.817030177818399e-06, "loss": 2.2586, "step": 2779 }, { "epoch": 0.8903122497998399, "grad_norm": 0.50390625, "learning_rate": 1.8065799981015625e-06, "loss": 2.1218, "step": 2780 }, { "epoch": 0.8906325060048038, "grad_norm": 0.5234375, "learning_rate": 1.7961588292721848e-06, "loss": 2.3084, "step": 2781 }, { "epoch": 0.8909527622097678, "grad_norm": 0.59375, "learning_rate": 1.7857666843653088e-06, "loss": 1.9461, "step": 2782 }, { "epoch": 0.8912730184147318, "grad_norm": 0.490234375, "learning_rate": 1.775403576379689e-06, "loss": 2.3538, "step": 2783 }, { "epoch": 0.8915932746196957, "grad_norm": 0.470703125, "learning_rate": 1.76506951827774e-06, "loss": 2.2233, "step": 2784 }, { "epoch": 0.8919135308246597, "grad_norm": 0.435546875, "learning_rate": 1.7547645229855497e-06, "loss": 2.5818, "step": 2785 }, { "epoch": 0.8922337870296237, "grad_norm": 0.68359375, "learning_rate": 1.7444886033928603e-06, "loss": 2.1235, "step": 2786 }, { "epoch": 0.8925540432345876, "grad_norm": 0.44140625, "learning_rate": 1.7342417723530374e-06, "loss": 2.3278, "step": 2787 }, { "epoch": 0.8928742994395517, "grad_norm": 0.439453125, "learning_rate": 1.7240240426830617e-06, "loss": 2.1486, "step": 2788 }, { "epoch": 0.8931945556445157, "grad_norm": 1.515625, "learning_rate": 1.713835427163521e-06, "loss": 2.7315, "step": 2789 }, { "epoch": 0.8935148118494796, "grad_norm": 0.408203125, "learning_rate": 1.7036759385385682e-06, "loss": 1.8048, "step": 2790 }, { "epoch": 0.8938350680544436, "grad_norm": 0.890625, "learning_rate": 1.693545589515952e-06, "loss": 2.6576, "step": 2791 }, { "epoch": 0.8941553242594076, "grad_norm": 0.412109375, "learning_rate": 1.6834443927669446e-06, "loss": 1.56, "step": 2792 }, { "epoch": 0.8944755804643715, "grad_norm": 0.4296875, "learning_rate": 1.6733723609263674e-06, "loss": 2.1219, "step": 2793 }, { "epoch": 0.8947958366693355, "grad_norm": 0.83203125, "learning_rate": 1.6633295065925647e-06, "loss": 1.3158, "step": 2794 }, { "epoch": 0.8951160928742994, "grad_norm": 0.494140625, "learning_rate": 1.653315842327377e-06, "loss": 2.4756, "step": 2795 }, { "epoch": 0.8954363490792634, "grad_norm": 0.3828125, "learning_rate": 1.6433313806561379e-06, "loss": 2.2544, "step": 2796 }, { "epoch": 0.8957566052842274, "grad_norm": 0.419921875, "learning_rate": 1.6333761340676518e-06, "loss": 2.0936, "step": 2797 }, { "epoch": 0.8960768614891913, "grad_norm": 0.423828125, "learning_rate": 1.6234501150141745e-06, "loss": 2.3256, "step": 2798 }, { "epoch": 0.8963971176941553, "grad_norm": 0.546875, "learning_rate": 1.6135533359114247e-06, "loss": 2.2462, "step": 2799 }, { "epoch": 0.8967173738991193, "grad_norm": 0.4609375, "learning_rate": 1.6036858091385137e-06, "loss": 2.377, "step": 2800 }, { "epoch": 0.8970376301040832, "grad_norm": 0.70703125, "learning_rate": 1.593847547037991e-06, "loss": 2.2872, "step": 2801 }, { "epoch": 0.8973578863090472, "grad_norm": 0.5, "learning_rate": 1.5840385619157904e-06, "loss": 1.8975, "step": 2802 }, { "epoch": 0.8976781425140112, "grad_norm": 0.5625, "learning_rate": 1.574258866041231e-06, "loss": 2.7625, "step": 2803 }, { "epoch": 0.8979983987189751, "grad_norm": 0.498046875, "learning_rate": 1.5645084716469777e-06, "loss": 2.2334, "step": 2804 }, { "epoch": 0.8983186549239391, "grad_norm": 2.90625, "learning_rate": 1.5547873909290716e-06, "loss": 2.3039, "step": 2805 }, { "epoch": 0.8986389111289032, "grad_norm": 0.44140625, "learning_rate": 1.5450956360468698e-06, "loss": 1.9124, "step": 2806 }, { "epoch": 0.898959167333867, "grad_norm": 0.455078125, "learning_rate": 1.5354332191230502e-06, "loss": 1.9301, "step": 2807 }, { "epoch": 0.8992794235388311, "grad_norm": 0.498046875, "learning_rate": 1.5258001522435978e-06, "loss": 2.5829, "step": 2808 }, { "epoch": 0.8995996797437951, "grad_norm": 0.5390625, "learning_rate": 1.516196447457785e-06, "loss": 2.4198, "step": 2809 }, { "epoch": 0.899919935948759, "grad_norm": 0.7578125, "learning_rate": 1.5066221167781557e-06, "loss": 2.4602, "step": 2810 }, { "epoch": 0.900240192153723, "grad_norm": 0.515625, "learning_rate": 1.4970771721805165e-06, "loss": 2.4408, "step": 2811 }, { "epoch": 0.900560448358687, "grad_norm": 0.55859375, "learning_rate": 1.4875616256039031e-06, "loss": 2.3519, "step": 2812 }, { "epoch": 0.9008807045636509, "grad_norm": 0.31640625, "learning_rate": 1.478075488950606e-06, "loss": 1.9272, "step": 2813 }, { "epoch": 0.9012009607686149, "grad_norm": 0.42578125, "learning_rate": 1.468618774086103e-06, "loss": 2.1546, "step": 2814 }, { "epoch": 0.9015212169735789, "grad_norm": 0.5390625, "learning_rate": 1.4591914928390793e-06, "loss": 2.2757, "step": 2815 }, { "epoch": 0.9018414731785428, "grad_norm": 0.5234375, "learning_rate": 1.449793657001408e-06, "loss": 2.3683, "step": 2816 }, { "epoch": 0.9021617293835068, "grad_norm": 0.50390625, "learning_rate": 1.4404252783281308e-06, "loss": 1.7778, "step": 2817 }, { "epoch": 0.9024819855884708, "grad_norm": 0.326171875, "learning_rate": 1.431086368537432e-06, "loss": 2.2228, "step": 2818 }, { "epoch": 0.9028022417934347, "grad_norm": 0.515625, "learning_rate": 1.4217769393106567e-06, "loss": 2.0747, "step": 2819 }, { "epoch": 0.9031224979983987, "grad_norm": 0.4453125, "learning_rate": 1.4124970022922513e-06, "loss": 1.7582, "step": 2820 }, { "epoch": 0.9034427542033627, "grad_norm": 0.6484375, "learning_rate": 1.4032465690897944e-06, "loss": 2.561, "step": 2821 }, { "epoch": 0.9037630104083266, "grad_norm": 0.515625, "learning_rate": 1.3940256512739447e-06, "loss": 1.9619, "step": 2822 }, { "epoch": 0.9040832666132906, "grad_norm": 0.478515625, "learning_rate": 1.3848342603784453e-06, "loss": 2.1804, "step": 2823 }, { "epoch": 0.9044035228182546, "grad_norm": 0.5546875, "learning_rate": 1.3756724079001165e-06, "loss": 2.6895, "step": 2824 }, { "epoch": 0.9047237790232185, "grad_norm": 0.421875, "learning_rate": 1.366540105298819e-06, "loss": 1.9138, "step": 2825 }, { "epoch": 0.9050440352281826, "grad_norm": 0.396484375, "learning_rate": 1.3574373639974518e-06, "loss": 2.2131, "step": 2826 }, { "epoch": 0.9053642914331466, "grad_norm": 0.412109375, "learning_rate": 1.3483641953819542e-06, "loss": 2.1942, "step": 2827 }, { "epoch": 0.9056845476381105, "grad_norm": 0.75, "learning_rate": 1.3393206108012562e-06, "loss": 1.9268, "step": 2828 }, { "epoch": 0.9060048038430745, "grad_norm": 0.435546875, "learning_rate": 1.33030662156729e-06, "loss": 2.5872, "step": 2829 }, { "epoch": 0.9063250600480385, "grad_norm": 0.5703125, "learning_rate": 1.3213222389549724e-06, "loss": 2.0611, "step": 2830 }, { "epoch": 0.9066453162530024, "grad_norm": 0.58203125, "learning_rate": 1.3123674742021862e-06, "loss": 2.4414, "step": 2831 }, { "epoch": 0.9069655724579664, "grad_norm": 0.61328125, "learning_rate": 1.303442338509761e-06, "loss": 2.0825, "step": 2832 }, { "epoch": 0.9072858286629304, "grad_norm": 2.046875, "learning_rate": 1.2945468430414803e-06, "loss": 1.8121, "step": 2833 }, { "epoch": 0.9076060848678943, "grad_norm": 0.51953125, "learning_rate": 1.2856809989240326e-06, "loss": 2.3212, "step": 2834 }, { "epoch": 0.9079263410728583, "grad_norm": 0.484375, "learning_rate": 1.2768448172470444e-06, "loss": 2.4806, "step": 2835 }, { "epoch": 0.9082465972778223, "grad_norm": 0.5625, "learning_rate": 1.2680383090630077e-06, "loss": 2.4582, "step": 2836 }, { "epoch": 0.9085668534827862, "grad_norm": 0.55859375, "learning_rate": 1.2592614853873224e-06, "loss": 2.4116, "step": 2837 }, { "epoch": 0.9088871096877502, "grad_norm": 0.49609375, "learning_rate": 1.2505143571982515e-06, "loss": 2.1586, "step": 2838 }, { "epoch": 0.9092073658927142, "grad_norm": 0.640625, "learning_rate": 1.2417969354369147e-06, "loss": 2.3296, "step": 2839 }, { "epoch": 0.9095276220976781, "grad_norm": 0.69140625, "learning_rate": 1.2331092310072623e-06, "loss": 1.7858, "step": 2840 }, { "epoch": 0.9098478783026421, "grad_norm": 0.470703125, "learning_rate": 1.2244512547761016e-06, "loss": 2.1908, "step": 2841 }, { "epoch": 0.910168134507606, "grad_norm": 0.46484375, "learning_rate": 1.2158230175730201e-06, "loss": 2.1441, "step": 2842 }, { "epoch": 0.91048839071257, "grad_norm": 0.64453125, "learning_rate": 1.2072245301904345e-06, "loss": 2.2306, "step": 2843 }, { "epoch": 0.910808646917534, "grad_norm": 0.421875, "learning_rate": 1.1986558033835333e-06, "loss": 2.5195, "step": 2844 }, { "epoch": 0.911128903122498, "grad_norm": 0.427734375, "learning_rate": 1.190116847870293e-06, "loss": 2.6438, "step": 2845 }, { "epoch": 0.911449159327462, "grad_norm": 1.0, "learning_rate": 1.181607674331439e-06, "loss": 2.0309, "step": 2846 }, { "epoch": 0.911769415532426, "grad_norm": 0.6015625, "learning_rate": 1.1731282934104553e-06, "loss": 2.1805, "step": 2847 }, { "epoch": 0.9120896717373899, "grad_norm": 0.48046875, "learning_rate": 1.1646787157135463e-06, "loss": 2.2949, "step": 2848 }, { "epoch": 0.9124099279423539, "grad_norm": 0.443359375, "learning_rate": 1.1562589518096607e-06, "loss": 2.3779, "step": 2849 }, { "epoch": 0.9127301841473179, "grad_norm": 0.41015625, "learning_rate": 1.1478690122304293e-06, "loss": 1.8918, "step": 2850 }, { "epoch": 0.9130504403522818, "grad_norm": 0.6796875, "learning_rate": 1.1395089074701936e-06, "loss": 2.4792, "step": 2851 }, { "epoch": 0.9133706965572458, "grad_norm": 0.490234375, "learning_rate": 1.1311786479859781e-06, "loss": 2.6007, "step": 2852 }, { "epoch": 0.9136909527622098, "grad_norm": 0.361328125, "learning_rate": 1.12287824419747e-06, "loss": 2.0799, "step": 2853 }, { "epoch": 0.9140112089671737, "grad_norm": 0.498046875, "learning_rate": 1.114607706487006e-06, "loss": 2.4274, "step": 2854 }, { "epoch": 0.9143314651721377, "grad_norm": 0.62109375, "learning_rate": 1.1063670451995856e-06, "loss": 2.5174, "step": 2855 }, { "epoch": 0.9146517213771017, "grad_norm": 0.52734375, "learning_rate": 1.0981562706428166e-06, "loss": 2.1216, "step": 2856 }, { "epoch": 0.9149719775820656, "grad_norm": 0.6953125, "learning_rate": 1.0899753930869394e-06, "loss": 2.0923, "step": 2857 }, { "epoch": 0.9152922337870296, "grad_norm": 0.427734375, "learning_rate": 1.0818244227647905e-06, "loss": 2.3559, "step": 2858 }, { "epoch": 0.9156124899919936, "grad_norm": 2.15625, "learning_rate": 1.0737033698717979e-06, "loss": 2.3995, "step": 2859 }, { "epoch": 0.9159327461969575, "grad_norm": 0.48046875, "learning_rate": 1.0656122445659721e-06, "loss": 2.1872, "step": 2860 }, { "epoch": 0.9162530024019215, "grad_norm": 0.47265625, "learning_rate": 1.057551056967887e-06, "loss": 1.9036, "step": 2861 }, { "epoch": 0.9165732586068855, "grad_norm": 0.458984375, "learning_rate": 1.0495198171606657e-06, "loss": 2.4604, "step": 2862 }, { "epoch": 0.9168935148118494, "grad_norm": 0.60546875, "learning_rate": 1.0415185351899837e-06, "loss": 2.326, "step": 2863 }, { "epoch": 0.9172137710168135, "grad_norm": 0.46484375, "learning_rate": 1.0335472210640323e-06, "loss": 2.0918, "step": 2864 }, { "epoch": 0.9175340272217775, "grad_norm": 0.59765625, "learning_rate": 1.0256058847535221e-06, "loss": 2.2654, "step": 2865 }, { "epoch": 0.9178542834267414, "grad_norm": 0.4765625, "learning_rate": 1.0176945361916657e-06, "loss": 2.1514, "step": 2866 }, { "epoch": 0.9181745396317054, "grad_norm": 0.421875, "learning_rate": 1.0098131852741777e-06, "loss": 2.1784, "step": 2867 }, { "epoch": 0.9184947958366694, "grad_norm": 0.6171875, "learning_rate": 1.001961841859228e-06, "loss": 2.1977, "step": 2868 }, { "epoch": 0.9188150520416333, "grad_norm": 0.515625, "learning_rate": 9.941405157674778e-07, "loss": 2.4212, "step": 2869 }, { "epoch": 0.9191353082465973, "grad_norm": 0.51171875, "learning_rate": 9.863492167820232e-07, "loss": 1.8568, "step": 2870 }, { "epoch": 0.9194555644515613, "grad_norm": 0.43359375, "learning_rate": 9.785879546484077e-07, "loss": 2.2068, "step": 2871 }, { "epoch": 0.9197758206565252, "grad_norm": 0.48046875, "learning_rate": 9.708567390746077e-07, "loss": 2.2889, "step": 2872 }, { "epoch": 0.9200960768614892, "grad_norm": 0.55078125, "learning_rate": 9.631555797310122e-07, "loss": 2.3926, "step": 2873 }, { "epoch": 0.9204163330664532, "grad_norm": 0.490234375, "learning_rate": 9.554844862504158e-07, "loss": 2.2207, "step": 2874 }, { "epoch": 0.9207365892714171, "grad_norm": 0.478515625, "learning_rate": 9.47843468228013e-07, "loss": 2.3968, "step": 2875 }, { "epoch": 0.9210568454763811, "grad_norm": 0.4921875, "learning_rate": 9.402325352213582e-07, "loss": 2.2347, "step": 2876 }, { "epoch": 0.9213771016813451, "grad_norm": 0.466796875, "learning_rate": 9.326516967504056e-07, "loss": 2.4828, "step": 2877 }, { "epoch": 0.921697357886309, "grad_norm": 0.53515625, "learning_rate": 9.251009622974426e-07, "loss": 2.2806, "step": 2878 }, { "epoch": 0.922017614091273, "grad_norm": 0.3828125, "learning_rate": 9.175803413071088e-07, "loss": 2.2707, "step": 2879 }, { "epoch": 0.922337870296237, "grad_norm": 1.3515625, "learning_rate": 9.100898431863791e-07, "loss": 1.9029, "step": 2880 }, { "epoch": 0.9226581265012009, "grad_norm": 0.3828125, "learning_rate": 9.026294773045535e-07, "loss": 1.0364, "step": 2881 }, { "epoch": 0.922978382706165, "grad_norm": 0.56640625, "learning_rate": 8.95199252993223e-07, "loss": 2.0725, "step": 2882 }, { "epoch": 0.923298638911129, "grad_norm": 0.5859375, "learning_rate": 8.877991795463086e-07, "loss": 2.2058, "step": 2883 }, { "epoch": 0.9236188951160929, "grad_norm": 0.5078125, "learning_rate": 8.804292662199842e-07, "loss": 2.1161, "step": 2884 }, { "epoch": 0.9239391513210569, "grad_norm": 0.8828125, "learning_rate": 8.730895222327284e-07, "loss": 2.715, "step": 2885 }, { "epoch": 0.9242594075260209, "grad_norm": 0.29296875, "learning_rate": 8.657799567652614e-07, "loss": 1.7732, "step": 2886 }, { "epoch": 0.9245796637309848, "grad_norm": 0.462890625, "learning_rate": 8.585005789605638e-07, "loss": 2.3266, "step": 2887 }, { "epoch": 0.9248999199359488, "grad_norm": 0.4375, "learning_rate": 8.512513979238579e-07, "loss": 2.277, "step": 2888 }, { "epoch": 0.9252201761409127, "grad_norm": 0.50390625, "learning_rate": 8.440324227225959e-07, "loss": 2.2216, "step": 2889 }, { "epoch": 0.9255404323458767, "grad_norm": 0.34375, "learning_rate": 8.368436623864384e-07, "loss": 2.1062, "step": 2890 }, { "epoch": 0.9258606885508407, "grad_norm": 0.703125, "learning_rate": 8.296851259072702e-07, "loss": 2.6251, "step": 2891 }, { "epoch": 0.9261809447558046, "grad_norm": 0.5703125, "learning_rate": 8.225568222391539e-07, "loss": 3.0058, "step": 2892 }, { "epoch": 0.9265012009607686, "grad_norm": 0.46875, "learning_rate": 8.154587602983432e-07, "loss": 2.0105, "step": 2893 }, { "epoch": 0.9268214571657326, "grad_norm": 0.478515625, "learning_rate": 8.08390948963264e-07, "loss": 2.0824, "step": 2894 }, { "epoch": 0.9271417133706965, "grad_norm": 0.4921875, "learning_rate": 8.013533970745113e-07, "loss": 2.2465, "step": 2895 }, { "epoch": 0.9274619695756605, "grad_norm": 0.7265625, "learning_rate": 7.943461134348129e-07, "loss": 2.2219, "step": 2896 }, { "epoch": 0.9277822257806245, "grad_norm": 0.58203125, "learning_rate": 7.873691068090605e-07, "loss": 2.332, "step": 2897 }, { "epoch": 0.9281024819855884, "grad_norm": 0.45703125, "learning_rate": 7.804223859242482e-07, "loss": 2.7178, "step": 2898 }, { "epoch": 0.9284227381905524, "grad_norm": 0.515625, "learning_rate": 7.735059594695171e-07, "loss": 1.4854, "step": 2899 }, { "epoch": 0.9287429943955164, "grad_norm": 0.51953125, "learning_rate": 7.666198360960858e-07, "loss": 2.2354, "step": 2900 }, { "epoch": 0.9290632506004803, "grad_norm": 0.462890625, "learning_rate": 7.597640244172921e-07, "loss": 1.8512, "step": 2901 }, { "epoch": 0.9293835068054443, "grad_norm": 0.5078125, "learning_rate": 7.529385330085431e-07, "loss": 1.8892, "step": 2902 }, { "epoch": 0.9297037630104084, "grad_norm": 0.38671875, "learning_rate": 7.461433704073373e-07, "loss": 2.5126, "step": 2903 }, { "epoch": 0.9300240192153723, "grad_norm": 0.9296875, "learning_rate": 7.393785451132146e-07, "loss": 2.1277, "step": 2904 }, { "epoch": 0.9303442754203363, "grad_norm": 0.453125, "learning_rate": 7.326440655877925e-07, "loss": 2.4802, "step": 2905 }, { "epoch": 0.9306645316253003, "grad_norm": 0.65234375, "learning_rate": 7.259399402547134e-07, "loss": 2.1335, "step": 2906 }, { "epoch": 0.9309847878302642, "grad_norm": 1.46875, "learning_rate": 7.192661774996584e-07, "loss": 2.5345, "step": 2907 }, { "epoch": 0.9313050440352282, "grad_norm": 0.640625, "learning_rate": 7.126227856703332e-07, "loss": 2.3705, "step": 2908 }, { "epoch": 0.9316253002401922, "grad_norm": 0.609375, "learning_rate": 7.060097730764548e-07, "loss": 2.6065, "step": 2909 }, { "epoch": 0.9319455564451561, "grad_norm": 0.453125, "learning_rate": 6.994271479897314e-07, "loss": 2.1943, "step": 2910 }, { "epoch": 0.9322658126501201, "grad_norm": 0.47265625, "learning_rate": 6.928749186438738e-07, "loss": 1.9554, "step": 2911 }, { "epoch": 0.9325860688550841, "grad_norm": 0.494140625, "learning_rate": 6.863530932345625e-07, "loss": 2.3002, "step": 2912 }, { "epoch": 0.932906325060048, "grad_norm": 0.59765625, "learning_rate": 6.798616799194634e-07, "loss": 2.4129, "step": 2913 }, { "epoch": 0.933226581265012, "grad_norm": 0.57421875, "learning_rate": 6.734006868181847e-07, "loss": 2.1435, "step": 2914 }, { "epoch": 0.933546837469976, "grad_norm": 0.4921875, "learning_rate": 6.66970122012292e-07, "loss": 1.9209, "step": 2915 }, { "epoch": 0.9338670936749399, "grad_norm": 0.515625, "learning_rate": 6.605699935452902e-07, "loss": 2.4338, "step": 2916 }, { "epoch": 0.9341873498799039, "grad_norm": 0.44140625, "learning_rate": 6.542003094226201e-07, "loss": 2.2675, "step": 2917 }, { "epoch": 0.9345076060848679, "grad_norm": 0.431640625, "learning_rate": 6.478610776116278e-07, "loss": 1.4878, "step": 2918 }, { "epoch": 0.9348278622898318, "grad_norm": 0.6796875, "learning_rate": 6.415523060415846e-07, "loss": 2.2072, "step": 2919 }, { "epoch": 0.9351481184947958, "grad_norm": 0.458984375, "learning_rate": 6.352740026036475e-07, "loss": 2.3663, "step": 2920 }, { "epoch": 0.9354683746997599, "grad_norm": 0.5234375, "learning_rate": 6.290261751508708e-07, "loss": 2.3013, "step": 2921 }, { "epoch": 0.9357886309047238, "grad_norm": 0.5703125, "learning_rate": 6.228088314981867e-07, "loss": 2.0383, "step": 2922 }, { "epoch": 0.9361088871096878, "grad_norm": 0.53515625, "learning_rate": 6.16621979422402e-07, "loss": 1.8135, "step": 2923 }, { "epoch": 0.9364291433146518, "grad_norm": 0.94140625, "learning_rate": 6.104656266621766e-07, "loss": 2.2271, "step": 2924 }, { "epoch": 0.9367493995196157, "grad_norm": 0.5546875, "learning_rate": 6.043397809180257e-07, "loss": 2.2725, "step": 2925 }, { "epoch": 0.9370696557245797, "grad_norm": 0.58984375, "learning_rate": 5.982444498523005e-07, "loss": 2.1763, "step": 2926 }, { "epoch": 0.9373899119295437, "grad_norm": 0.5, "learning_rate": 5.921796410891994e-07, "loss": 2.3399, "step": 2927 }, { "epoch": 0.9377101681345076, "grad_norm": 0.474609375, "learning_rate": 5.861453622147157e-07, "loss": 1.8565, "step": 2928 }, { "epoch": 0.9380304243394716, "grad_norm": 0.57421875, "learning_rate": 5.801416207766808e-07, "loss": 2.3172, "step": 2929 }, { "epoch": 0.9383506805444356, "grad_norm": 1.4453125, "learning_rate": 5.741684242847156e-07, "loss": 1.864, "step": 2930 }, { "epoch": 0.9386709367493995, "grad_norm": 0.53125, "learning_rate": 5.682257802102408e-07, "loss": 2.4591, "step": 2931 }, { "epoch": 0.9389911929543635, "grad_norm": 0.498046875, "learning_rate": 5.623136959864494e-07, "loss": 2.759, "step": 2932 }, { "epoch": 0.9393114491593275, "grad_norm": 1.703125, "learning_rate": 5.564321790083316e-07, "loss": 1.5265, "step": 2933 }, { "epoch": 0.9396317053642914, "grad_norm": 0.54296875, "learning_rate": 5.505812366326219e-07, "loss": 2.2601, "step": 2934 }, { "epoch": 0.9399519615692554, "grad_norm": 0.50390625, "learning_rate": 5.447608761778189e-07, "loss": 1.9696, "step": 2935 }, { "epoch": 0.9402722177742193, "grad_norm": 0.423828125, "learning_rate": 5.389711049241741e-07, "loss": 2.2893, "step": 2936 }, { "epoch": 0.9405924739791833, "grad_norm": 0.515625, "learning_rate": 5.332119301136695e-07, "loss": 2.2355, "step": 2937 }, { "epoch": 0.9409127301841473, "grad_norm": 0.5859375, "learning_rate": 5.274833589500205e-07, "loss": 2.0032, "step": 2938 }, { "epoch": 0.9412329863891112, "grad_norm": 0.451171875, "learning_rate": 5.217853985986648e-07, "loss": 2.6002, "step": 2939 }, { "epoch": 0.9415532425940752, "grad_norm": 0.443359375, "learning_rate": 5.161180561867401e-07, "loss": 2.2874, "step": 2940 }, { "epoch": 0.9418734987990393, "grad_norm": 0.49609375, "learning_rate": 5.104813388031038e-07, "loss": 2.5509, "step": 2941 }, { "epoch": 0.9421937550040032, "grad_norm": 0.65625, "learning_rate": 5.048752534982909e-07, "loss": 2.1587, "step": 2942 }, { "epoch": 0.9425140112089672, "grad_norm": 0.451171875, "learning_rate": 4.992998072845311e-07, "loss": 2.4545, "step": 2943 }, { "epoch": 0.9428342674139312, "grad_norm": 0.443359375, "learning_rate": 4.937550071357239e-07, "loss": 1.585, "step": 2944 }, { "epoch": 0.9431545236188951, "grad_norm": 0.5390625, "learning_rate": 4.882408599874433e-07, "loss": 2.0142, "step": 2945 }, { "epoch": 0.9434747798238591, "grad_norm": 0.5703125, "learning_rate": 4.827573727369112e-07, "loss": 2.179, "step": 2946 }, { "epoch": 0.9437950360288231, "grad_norm": 0.7578125, "learning_rate": 4.773045522430103e-07, "loss": 2.4145, "step": 2947 }, { "epoch": 0.944115292233787, "grad_norm": 0.486328125, "learning_rate": 4.7188240532625714e-07, "loss": 2.4136, "step": 2948 }, { "epoch": 0.944435548438751, "grad_norm": 0.51171875, "learning_rate": 4.664909387688043e-07, "loss": 1.7031, "step": 2949 }, { "epoch": 0.944755804643715, "grad_norm": 0.4921875, "learning_rate": 4.6113015931442683e-07, "loss": 2.3827, "step": 2950 }, { "epoch": 0.9450760608486789, "grad_norm": 0.462890625, "learning_rate": 4.558000736685164e-07, "loss": 2.4576, "step": 2951 }, { "epoch": 0.9453963170536429, "grad_norm": 0.494140625, "learning_rate": 4.5050068849807334e-07, "loss": 2.6722, "step": 2952 }, { "epoch": 0.9457165732586069, "grad_norm": 0.48046875, "learning_rate": 4.4523201043169813e-07, "loss": 2.5232, "step": 2953 }, { "epoch": 0.9460368294635708, "grad_norm": 0.474609375, "learning_rate": 4.3999404605957196e-07, "loss": 2.327, "step": 2954 }, { "epoch": 0.9463570856685348, "grad_norm": 0.5859375, "learning_rate": 4.347868019334789e-07, "loss": 2.152, "step": 2955 }, { "epoch": 0.9466773418734988, "grad_norm": 0.6015625, "learning_rate": 4.296102845667532e-07, "loss": 2.3246, "step": 2956 }, { "epoch": 0.9469975980784627, "grad_norm": 0.8984375, "learning_rate": 4.244645004343156e-07, "loss": 2.4348, "step": 2957 }, { "epoch": 0.9473178542834267, "grad_norm": 0.546875, "learning_rate": 4.193494559726313e-07, "loss": 2.3624, "step": 2958 }, { "epoch": 0.9476381104883908, "grad_norm": 0.435546875, "learning_rate": 4.1426515757972684e-07, "loss": 2.3969, "step": 2959 }, { "epoch": 0.9479583666933546, "grad_norm": 1.015625, "learning_rate": 4.0921161161515674e-07, "loss": 2.1616, "step": 2960 }, { "epoch": 0.9482786228983187, "grad_norm": 0.34765625, "learning_rate": 4.041888244000286e-07, "loss": 1.9173, "step": 2961 }, { "epoch": 0.9485988791032827, "grad_norm": 0.52734375, "learning_rate": 3.991968022169529e-07, "loss": 2.3724, "step": 2962 }, { "epoch": 0.9489191353082466, "grad_norm": 0.380859375, "learning_rate": 3.9423555131007925e-07, "loss": 2.1529, "step": 2963 }, { "epoch": 0.9492393915132106, "grad_norm": 0.51953125, "learning_rate": 3.8930507788506046e-07, "loss": 2.1868, "step": 2964 }, { "epoch": 0.9495596477181746, "grad_norm": 0.4921875, "learning_rate": 3.8440538810904367e-07, "loss": 1.8231, "step": 2965 }, { "epoch": 0.9498799039231385, "grad_norm": 0.6171875, "learning_rate": 3.7953648811068777e-07, "loss": 2.235, "step": 2966 }, { "epoch": 0.9502001601281025, "grad_norm": 0.96484375, "learning_rate": 3.746983839801238e-07, "loss": 2.3122, "step": 2967 }, { "epoch": 0.9505204163330665, "grad_norm": 0.50390625, "learning_rate": 3.698910817689694e-07, "loss": 2.4373, "step": 2968 }, { "epoch": 0.9508406725380304, "grad_norm": 0.416015625, "learning_rate": 3.6511458749031723e-07, "loss": 1.7601, "step": 2969 }, { "epoch": 0.9511609287429944, "grad_norm": 0.7734375, "learning_rate": 3.6036890711871886e-07, "loss": 2.7486, "step": 2970 }, { "epoch": 0.9514811849479584, "grad_norm": 0.4296875, "learning_rate": 3.556540465901842e-07, "loss": 2.1883, "step": 2971 }, { "epoch": 0.9518014411529223, "grad_norm": 0.57421875, "learning_rate": 3.509700118021736e-07, "loss": 2.4816, "step": 2972 }, { "epoch": 0.9521216973578863, "grad_norm": 0.6484375, "learning_rate": 3.463168086135976e-07, "loss": 2.3101, "step": 2973 }, { "epoch": 0.9524419535628503, "grad_norm": 0.55859375, "learning_rate": 3.416944428447838e-07, "loss": 2.7864, "step": 2974 }, { "epoch": 0.9527622097678142, "grad_norm": 0.58203125, "learning_rate": 3.3710292027750724e-07, "loss": 2.1871, "step": 2975 }, { "epoch": 0.9530824659727782, "grad_norm": 0.55078125, "learning_rate": 3.3254224665494883e-07, "loss": 2.508, "step": 2976 }, { "epoch": 0.9534027221777422, "grad_norm": 0.65234375, "learning_rate": 3.2801242768171483e-07, "loss": 1.6672, "step": 2977 }, { "epoch": 0.9537229783827061, "grad_norm": 0.451171875, "learning_rate": 3.235134690238062e-07, "loss": 2.5607, "step": 2978 }, { "epoch": 0.9540432345876702, "grad_norm": 0.5703125, "learning_rate": 3.1904537630862984e-07, "loss": 2.4434, "step": 2979 }, { "epoch": 0.9543634907926342, "grad_norm": 0.56640625, "learning_rate": 3.146081551249874e-07, "loss": 2.1326, "step": 2980 }, { "epoch": 0.9546837469975981, "grad_norm": 0.6796875, "learning_rate": 3.102018110230587e-07, "loss": 2.3844, "step": 2981 }, { "epoch": 0.9550040032025621, "grad_norm": 0.3828125, "learning_rate": 3.058263495144015e-07, "loss": 2.5421, "step": 2982 }, { "epoch": 0.955324259407526, "grad_norm": 0.46484375, "learning_rate": 3.014817760719546e-07, "loss": 1.6864, "step": 2983 }, { "epoch": 0.95564451561249, "grad_norm": 0.55078125, "learning_rate": 2.971680961300127e-07, "loss": 2.5311, "step": 2984 }, { "epoch": 0.955964771817454, "grad_norm": 0.6875, "learning_rate": 2.9288531508422644e-07, "loss": 2.2674, "step": 2985 }, { "epoch": 0.9562850280224179, "grad_norm": 0.8828125, "learning_rate": 2.886334382916078e-07, "loss": 2.4979, "step": 2986 }, { "epoch": 0.9566052842273819, "grad_norm": 0.51171875, "learning_rate": 2.8441247107050264e-07, "loss": 2.3117, "step": 2987 }, { "epoch": 0.9569255404323459, "grad_norm": 0.546875, "learning_rate": 2.802224187005986e-07, "loss": 2.069, "step": 2988 }, { "epoch": 0.9572457966373098, "grad_norm": 1.1484375, "learning_rate": 2.7606328642291736e-07, "loss": 2.3921, "step": 2989 }, { "epoch": 0.9575660528422738, "grad_norm": 0.52734375, "learning_rate": 2.7193507943979457e-07, "loss": 2.7724, "step": 2990 }, { "epoch": 0.9578863090472378, "grad_norm": 0.51953125, "learning_rate": 2.678378029148998e-07, "loss": 2.1656, "step": 2991 }, { "epoch": 0.9582065652522017, "grad_norm": 0.51171875, "learning_rate": 2.6377146197319733e-07, "loss": 2.1326, "step": 2992 }, { "epoch": 0.9585268214571657, "grad_norm": 0.66015625, "learning_rate": 2.5973606170096585e-07, "loss": 2.2594, "step": 2993 }, { "epoch": 0.9588470776621297, "grad_norm": 0.671875, "learning_rate": 2.5573160714578436e-07, "loss": 2.2226, "step": 2994 }, { "epoch": 0.9591673338670936, "grad_norm": 0.44921875, "learning_rate": 2.517581033165184e-07, "loss": 2.1275, "step": 2995 }, { "epoch": 0.9594875900720576, "grad_norm": 0.5234375, "learning_rate": 2.4781555518332e-07, "loss": 2.3456, "step": 2996 }, { "epoch": 0.9598078462770216, "grad_norm": 0.6484375, "learning_rate": 2.439039676776306e-07, "loss": 2.5249, "step": 2997 }, { "epoch": 0.9601281024819855, "grad_norm": 0.373046875, "learning_rate": 2.4002334569215023e-07, "loss": 1.8815, "step": 2998 }, { "epoch": 0.9604483586869496, "grad_norm": 0.5546875, "learning_rate": 2.3617369408085732e-07, "loss": 2.5048, "step": 2999 }, { "epoch": 0.9607686148919136, "grad_norm": 0.53125, "learning_rate": 2.3235501765898615e-07, "loss": 2.205, "step": 3000 }, { "epoch": 0.9610888710968775, "grad_norm": 0.65625, "learning_rate": 2.285673212030326e-07, "loss": 2.0211, "step": 3001 }, { "epoch": 0.9614091273018415, "grad_norm": 0.50390625, "learning_rate": 2.2481060945073738e-07, "loss": 2.1957, "step": 3002 }, { "epoch": 0.9617293835068055, "grad_norm": 0.498046875, "learning_rate": 2.21084887101089e-07, "loss": 2.2572, "step": 3003 }, { "epoch": 0.9620496397117694, "grad_norm": 0.51953125, "learning_rate": 2.1739015881430125e-07, "loss": 2.4002, "step": 3004 }, { "epoch": 0.9623698959167334, "grad_norm": 0.5703125, "learning_rate": 2.137264292118385e-07, "loss": 1.8485, "step": 3005 }, { "epoch": 0.9626901521216974, "grad_norm": 2.34375, "learning_rate": 2.1009370287637664e-07, "loss": 2.4418, "step": 3006 }, { "epoch": 0.9630104083266613, "grad_norm": 0.5546875, "learning_rate": 2.0649198435181983e-07, "loss": 2.4171, "step": 3007 }, { "epoch": 0.9633306645316253, "grad_norm": 0.4609375, "learning_rate": 2.02921278143281e-07, "loss": 2.5708, "step": 3008 }, { "epoch": 0.9636509207365893, "grad_norm": 0.44140625, "learning_rate": 1.99381588717093e-07, "loss": 2.2956, "step": 3009 }, { "epoch": 0.9639711769415532, "grad_norm": 0.734375, "learning_rate": 1.9587292050077255e-07, "loss": 2.6983, "step": 3010 }, { "epoch": 0.9642914331465172, "grad_norm": 0.55859375, "learning_rate": 1.9239527788305622e-07, "loss": 2.1624, "step": 3011 }, { "epoch": 0.9646116893514812, "grad_norm": 0.7265625, "learning_rate": 1.889486652138589e-07, "loss": 1.8433, "step": 3012 }, { "epoch": 0.9649319455564451, "grad_norm": 0.58203125, "learning_rate": 1.855330868042876e-07, "loss": 2.5753, "step": 3013 }, { "epoch": 0.9652522017614091, "grad_norm": 0.43359375, "learning_rate": 1.8214854692662765e-07, "loss": 1.9499, "step": 3014 }, { "epoch": 0.9655724579663731, "grad_norm": 0.54296875, "learning_rate": 1.7879504981434536e-07, "loss": 1.9731, "step": 3015 }, { "epoch": 0.965892714171337, "grad_norm": 0.466796875, "learning_rate": 1.7547259966207708e-07, "loss": 2.0553, "step": 3016 }, { "epoch": 0.966212970376301, "grad_norm": 0.51953125, "learning_rate": 1.7218120062562071e-07, "loss": 2.3178, "step": 3017 }, { "epoch": 0.9665332265812651, "grad_norm": 0.48828125, "learning_rate": 1.6892085682193583e-07, "loss": 1.948, "step": 3018 }, { "epoch": 0.966853482786229, "grad_norm": 0.58984375, "learning_rate": 1.6569157232914079e-07, "loss": 2.2639, "step": 3019 }, { "epoch": 0.967173738991193, "grad_norm": 0.365234375, "learning_rate": 1.6249335118650456e-07, "loss": 2.1195, "step": 3020 }, { "epoch": 0.967493995196157, "grad_norm": 0.54296875, "learning_rate": 1.5932619739443543e-07, "loss": 2.0017, "step": 3021 }, { "epoch": 0.9678142514011209, "grad_norm": 0.443359375, "learning_rate": 1.5619011491448677e-07, "loss": 1.485, "step": 3022 }, { "epoch": 0.9681345076060849, "grad_norm": 0.4453125, "learning_rate": 1.5308510766934848e-07, "loss": 2.2819, "step": 3023 }, { "epoch": 0.9684547638110489, "grad_norm": 0.4375, "learning_rate": 1.5001117954283329e-07, "loss": 1.7135, "step": 3024 }, { "epoch": 0.9687750200160128, "grad_norm": 0.462890625, "learning_rate": 1.4696833437988778e-07, "loss": 2.3383, "step": 3025 }, { "epoch": 0.9690952762209768, "grad_norm": 0.921875, "learning_rate": 1.4395657598657296e-07, "loss": 1.7648, "step": 3026 }, { "epoch": 0.9694155324259408, "grad_norm": 0.5234375, "learning_rate": 1.4097590813007545e-07, "loss": 2.0177, "step": 3027 }, { "epoch": 0.9697357886309047, "grad_norm": 0.462890625, "learning_rate": 1.380263345386795e-07, "loss": 1.8389, "step": 3028 }, { "epoch": 0.9700560448358687, "grad_norm": 0.5234375, "learning_rate": 1.35107858901784e-07, "loss": 2.0451, "step": 3029 }, { "epoch": 0.9703763010408326, "grad_norm": 0.62890625, "learning_rate": 1.3222048486989102e-07, "loss": 2.4008, "step": 3030 }, { "epoch": 0.9706965572457966, "grad_norm": 0.484375, "learning_rate": 1.2936421605459494e-07, "loss": 2.2855, "step": 3031 }, { "epoch": 0.9710168134507606, "grad_norm": 0.54296875, "learning_rate": 1.2653905602858507e-07, "loss": 2.2895, "step": 3032 }, { "epoch": 0.9713370696557245, "grad_norm": 0.349609375, "learning_rate": 1.2374500832564306e-07, "loss": 2.4505, "step": 3033 }, { "epoch": 0.9716573258606885, "grad_norm": 0.54296875, "learning_rate": 1.209820764406261e-07, "loss": 2.6746, "step": 3034 }, { "epoch": 0.9719775820656525, "grad_norm": 0.640625, "learning_rate": 1.1825026382947801e-07, "loss": 2.5954, "step": 3035 }, { "epoch": 0.9722978382706164, "grad_norm": 0.71484375, "learning_rate": 1.1554957390921828e-07, "loss": 2.7764, "step": 3036 }, { "epoch": 0.9726180944755805, "grad_norm": 0.51953125, "learning_rate": 1.1288001005793358e-07, "loss": 2.1283, "step": 3037 }, { "epoch": 0.9729383506805445, "grad_norm": 0.412109375, "learning_rate": 1.1024157561477233e-07, "loss": 2.1614, "step": 3038 }, { "epoch": 0.9732586068855084, "grad_norm": 0.5546875, "learning_rate": 1.0763427387996128e-07, "loss": 2.4447, "step": 3039 }, { "epoch": 0.9735788630904724, "grad_norm": 0.52734375, "learning_rate": 1.0505810811476947e-07, "loss": 2.3156, "step": 3040 }, { "epoch": 0.9738991192954364, "grad_norm": 0.5234375, "learning_rate": 1.0251308154153317e-07, "loss": 2.5195, "step": 3041 }, { "epoch": 0.9742193755004003, "grad_norm": 0.640625, "learning_rate": 9.999919734362539e-08, "loss": 2.5644, "step": 3042 }, { "epoch": 0.9745396317053643, "grad_norm": 0.6328125, "learning_rate": 9.75164586654781e-08, "loss": 2.3122, "step": 3043 }, { "epoch": 0.9748598879103283, "grad_norm": 0.349609375, "learning_rate": 9.506486861255715e-08, "loss": 2.1659, "step": 3044 }, { "epoch": 0.9751801441152922, "grad_norm": 0.423828125, "learning_rate": 9.264443025137626e-08, "loss": 2.1065, "step": 3045 }, { "epoch": 0.9755004003202562, "grad_norm": 1.359375, "learning_rate": 9.025514660946921e-08, "loss": 2.4461, "step": 3046 }, { "epoch": 0.9758206565252202, "grad_norm": 0.35546875, "learning_rate": 8.78970206754176e-08, "loss": 2.4571, "step": 3047 }, { "epoch": 0.9761409127301841, "grad_norm": 0.4765625, "learning_rate": 8.55700553988148e-08, "loss": 1.889, "step": 3048 }, { "epoch": 0.9764611689351481, "grad_norm": 0.421875, "learning_rate": 8.327425369029085e-08, "loss": 2.25, "step": 3049 }, { "epoch": 0.9767814251401121, "grad_norm": 0.380859375, "learning_rate": 8.100961842148758e-08, "loss": 2.4113, "step": 3050 }, { "epoch": 0.977101681345076, "grad_norm": 0.546875, "learning_rate": 7.877615242506408e-08, "loss": 2.4718, "step": 3051 }, { "epoch": 0.97742193755004, "grad_norm": 0.54296875, "learning_rate": 7.657385849469678e-08, "loss": 2.5222, "step": 3052 }, { "epoch": 0.977742193755004, "grad_norm": 0.55859375, "learning_rate": 7.44027393850627e-08, "loss": 2.1825, "step": 3053 }, { "epoch": 0.9780624499599679, "grad_norm": 0.484375, "learning_rate": 7.226279781185341e-08, "loss": 1.9089, "step": 3054 }, { "epoch": 0.978382706164932, "grad_norm": 0.6953125, "learning_rate": 7.015403645176111e-08, "loss": 2.3184, "step": 3055 }, { "epoch": 0.978702962369896, "grad_norm": 0.486328125, "learning_rate": 6.80764579424703e-08, "loss": 1.9132, "step": 3056 }, { "epoch": 0.9790232185748599, "grad_norm": 0.921875, "learning_rate": 6.603006488266616e-08, "loss": 1.9538, "step": 3057 }, { "epoch": 0.9793434747798239, "grad_norm": 0.66796875, "learning_rate": 6.40148598320317e-08, "loss": 2.1982, "step": 3058 }, { "epoch": 0.9796637309847879, "grad_norm": 0.5390625, "learning_rate": 6.203084531123115e-08, "loss": 1.9992, "step": 3059 }, { "epoch": 0.9799839871897518, "grad_norm": 0.53515625, "learning_rate": 6.007802380191552e-08, "loss": 2.0144, "step": 3060 }, { "epoch": 0.9803042433947158, "grad_norm": 0.5703125, "learning_rate": 5.815639774672532e-08, "loss": 1.9409, "step": 3061 }, { "epoch": 0.9806244995996798, "grad_norm": 0.365234375, "learning_rate": 5.6265969549271216e-08, "loss": 2.4769, "step": 3062 }, { "epoch": 0.9809447558046437, "grad_norm": 0.37890625, "learning_rate": 5.44067415741506e-08, "loss": 2.4554, "step": 3063 }, { "epoch": 0.9812650120096077, "grad_norm": 0.54296875, "learning_rate": 5.257871614692822e-08, "loss": 2.1183, "step": 3064 }, { "epoch": 0.9815852682145717, "grad_norm": 0.75, "learning_rate": 5.07818955541417e-08, "loss": 2.0154, "step": 3065 }, { "epoch": 0.9819055244195356, "grad_norm": 0.482421875, "learning_rate": 4.901628204330155e-08, "loss": 2.1623, "step": 3066 }, { "epoch": 0.9822257806244996, "grad_norm": 0.625, "learning_rate": 4.728187782287452e-08, "loss": 1.8426, "step": 3067 }, { "epoch": 0.9825460368294636, "grad_norm": 0.466796875, "learning_rate": 4.5578685062297456e-08, "loss": 2.0737, "step": 3068 }, { "epoch": 0.9828662930344275, "grad_norm": 0.62109375, "learning_rate": 4.390670589196622e-08, "loss": 2.1993, "step": 3069 }, { "epoch": 0.9831865492393915, "grad_norm": 0.9140625, "learning_rate": 4.2265942403227345e-08, "loss": 2.5824, "step": 3070 }, { "epoch": 0.9835068054443555, "grad_norm": 0.70703125, "learning_rate": 4.065639664839471e-08, "loss": 2.3295, "step": 3071 }, { "epoch": 0.9838270616493194, "grad_norm": 0.58203125, "learning_rate": 3.907807064072178e-08, "loss": 1.9268, "step": 3072 }, { "epoch": 0.9841473178542834, "grad_norm": 0.490234375, "learning_rate": 3.7530966354418215e-08, "loss": 2.4965, "step": 3073 }, { "epoch": 0.9844675740592475, "grad_norm": 0.345703125, "learning_rate": 3.6015085724638856e-08, "loss": 2.1137, "step": 3074 }, { "epoch": 0.9847878302642114, "grad_norm": 0.326171875, "learning_rate": 3.453043064748362e-08, "loss": 1.6508, "step": 3075 }, { "epoch": 0.9851080864691754, "grad_norm": 0.478515625, "learning_rate": 3.307700297999483e-08, "loss": 2.1851, "step": 3076 }, { "epoch": 0.9854283426741393, "grad_norm": 0.314453125, "learning_rate": 3.1654804540157124e-08, "loss": 1.754, "step": 3077 }, { "epoch": 0.9857485988791033, "grad_norm": 0.37890625, "learning_rate": 3.02638371068864e-08, "loss": 2.1616, "step": 3078 }, { "epoch": 0.9860688550840673, "grad_norm": 0.53515625, "learning_rate": 2.890410242003816e-08, "loss": 2.733, "step": 3079 }, { "epoch": 0.9863891112890312, "grad_norm": 0.474609375, "learning_rate": 2.757560218040467e-08, "loss": 2.0874, "step": 3080 }, { "epoch": 0.9867093674939952, "grad_norm": 0.48828125, "learning_rate": 2.6278338049706696e-08, "loss": 2.343, "step": 3081 }, { "epoch": 0.9870296236989592, "grad_norm": 0.50390625, "learning_rate": 2.5012311650587927e-08, "loss": 2.3688, "step": 3082 }, { "epoch": 0.9873498799039231, "grad_norm": 0.47265625, "learning_rate": 2.377752456662885e-08, "loss": 1.8563, "step": 3083 }, { "epoch": 0.9876701361088871, "grad_norm": 0.53125, "learning_rate": 2.257397834233288e-08, "loss": 2.2285, "step": 3084 }, { "epoch": 0.9879903923138511, "grad_norm": 0.4921875, "learning_rate": 2.1401674483118027e-08, "loss": 2.2832, "step": 3085 }, { "epoch": 0.988310648518815, "grad_norm": 0.5625, "learning_rate": 2.026061445533356e-08, "loss": 2.1044, "step": 3086 }, { "epoch": 0.988630904723779, "grad_norm": 0.58984375, "learning_rate": 1.9150799686246112e-08, "loss": 2.234, "step": 3087 }, { "epoch": 0.988951160928743, "grad_norm": 0.41796875, "learning_rate": 1.8072231564036922e-08, "loss": 1.7073, "step": 3088 }, { "epoch": 0.9892714171337069, "grad_norm": 0.490234375, "learning_rate": 1.702491143780738e-08, "loss": 2.6741, "step": 3089 }, { "epoch": 0.9895916733386709, "grad_norm": 1.1484375, "learning_rate": 1.6008840617565134e-08, "loss": 2.2903, "step": 3090 }, { "epoch": 0.9899119295436349, "grad_norm": 0.52734375, "learning_rate": 1.5024020374243554e-08, "loss": 2.1201, "step": 3091 }, { "epoch": 0.9902321857485988, "grad_norm": 0.65234375, "learning_rate": 1.4070451939673934e-08, "loss": 2.3257, "step": 3092 }, { "epoch": 0.9905524419535628, "grad_norm": 0.39453125, "learning_rate": 1.3148136506604958e-08, "loss": 2.3063, "step": 3093 }, { "epoch": 0.9908726981585269, "grad_norm": 0.42578125, "learning_rate": 1.2257075228688797e-08, "loss": 2.2024, "step": 3094 }, { "epoch": 0.9911929543634908, "grad_norm": 0.42578125, "learning_rate": 1.1397269220486672e-08, "loss": 2.2727, "step": 3095 }, { "epoch": 0.9915132105684548, "grad_norm": 0.80078125, "learning_rate": 1.0568719557468853e-08, "loss": 1.8462, "step": 3096 }, { "epoch": 0.9918334667734188, "grad_norm": 0.44140625, "learning_rate": 9.771427276000778e-09, "loss": 1.2467, "step": 3097 }, { "epoch": 0.9921537229783827, "grad_norm": 0.27734375, "learning_rate": 9.00539337335693e-09, "loss": 1.7548, "step": 3098 }, { "epoch": 0.9924739791833467, "grad_norm": 0.55859375, "learning_rate": 8.270618807706965e-09, "loss": 1.8665, "step": 3099 }, { "epoch": 0.9927942353883107, "grad_norm": 0.400390625, "learning_rate": 7.56710449813236e-09, "loss": 1.8483, "step": 3100 }, { "epoch": 0.9931144915932746, "grad_norm": 0.75, "learning_rate": 6.894851324595886e-09, "loss": 2.3824, "step": 3101 }, { "epoch": 0.9934347477982386, "grad_norm": 0.48828125, "learning_rate": 6.2538601279776846e-09, "loss": 2.3081, "step": 3102 }, { "epoch": 0.9937550040032026, "grad_norm": 0.53515625, "learning_rate": 5.644131710039191e-09, "loss": 2.6313, "step": 3103 }, { "epoch": 0.9940752602081665, "grad_norm": 0.42578125, "learning_rate": 5.065666833442562e-09, "loss": 2.2497, "step": 3104 }, { "epoch": 0.9943955164131305, "grad_norm": 0.478515625, "learning_rate": 4.518466221750672e-09, "loss": 2.2095, "step": 3105 }, { "epoch": 0.9947157726180945, "grad_norm": 2.53125, "learning_rate": 4.002530559410467e-09, "loss": 2.0491, "step": 3106 }, { "epoch": 0.9950360288230584, "grad_norm": 0.482421875, "learning_rate": 3.5178604917668334e-09, "loss": 1.9228, "step": 3107 }, { "epoch": 0.9953562850280224, "grad_norm": 0.427734375, "learning_rate": 3.0644566250598305e-09, "loss": 2.6178, "step": 3108 }, { "epoch": 0.9956765412329864, "grad_norm": 0.578125, "learning_rate": 2.6423195264163593e-09, "loss": 2.0714, "step": 3109 }, { "epoch": 0.9959967974379503, "grad_norm": 0.5703125, "learning_rate": 2.2514497238557144e-09, "loss": 2.1848, "step": 3110 }, { "epoch": 0.9963170536429143, "grad_norm": 0.484375, "learning_rate": 1.891847706286809e-09, "loss": 2.4682, "step": 3111 }, { "epoch": 0.9966373098478783, "grad_norm": 0.462890625, "learning_rate": 1.5635139235109507e-09, "loss": 2.3354, "step": 3112 }, { "epoch": 0.9969575660528422, "grad_norm": 0.5546875, "learning_rate": 1.2664487862107389e-09, "loss": 1.8637, "step": 3113 }, { "epoch": 0.9972778222578063, "grad_norm": 0.408203125, "learning_rate": 1.0006526659667171e-09, "loss": 2.0926, "step": 3114 }, { "epoch": 0.9975980784627703, "grad_norm": 0.53515625, "learning_rate": 7.661258952434969e-10, "loss": 2.0889, "step": 3115 }, { "epoch": 0.9979183346677342, "grad_norm": 0.5703125, "learning_rate": 5.628687673897571e-10, "loss": 1.7024, "step": 3116 }, { "epoch": 0.9982385908726982, "grad_norm": 0.5234375, "learning_rate": 3.9088153664379457e-10, "loss": 2.477, "step": 3117 }, { "epoch": 0.9985588470776622, "grad_norm": 0.470703125, "learning_rate": 2.5016441813630054e-10, "loss": 2.363, "step": 3118 }, { "epoch": 0.9988791032826261, "grad_norm": 1.9765625, "learning_rate": 1.4071758787648214e-10, "loss": 2.3924, "step": 3119 }, { "epoch": 0.9991993594875901, "grad_norm": 1.265625, "learning_rate": 6.254118276316501e-11, "loss": 2.2253, "step": 3120 }, { "epoch": 0.9995196156925541, "grad_norm": 0.64453125, "learning_rate": 1.5635300579242008e-11, "loss": 2.4826, "step": 3121 }, { "epoch": 0.999839871897518, "grad_norm": 0.61328125, "learning_rate": 0.0, "loss": 2.4466, "step": 3122 }, { "epoch": 0.999839871897518, "step": 3122, "total_flos": 1.7832492586531553e+18, "train_loss": 2.216206028322468, "train_runtime": 33247.722, "train_samples_per_second": 0.751, "train_steps_per_second": 0.094 } ], "logging_steps": 1, "max_steps": 3122, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.7832492586531553e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }