mistral_lora / trainer_state.json
Ba2han's picture
Upload 9 files
2bec092 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.7806973218797373,
"eval_steps": 500,
"global_step": 1545,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0005053057099545225,
"grad_norm": 1.6384633779525757,
"learning_rate": 0.0,
"loss": 0.4704,
"step": 1
},
{
"epoch": 0.001010611419909045,
"grad_norm": 2.194593906402588,
"learning_rate": 6.41025641025641e-07,
"loss": 0.4286,
"step": 2
},
{
"epoch": 0.0015159171298635675,
"grad_norm": 3.6220381259918213,
"learning_rate": 1.282051282051282e-06,
"loss": 0.8297,
"step": 3
},
{
"epoch": 0.00202122283981809,
"grad_norm": 1.929093837738037,
"learning_rate": 1.9230769230769234e-06,
"loss": 0.6096,
"step": 4
},
{
"epoch": 0.0025265285497726125,
"grad_norm": 3.1244475841522217,
"learning_rate": 2.564102564102564e-06,
"loss": 0.8738,
"step": 5
},
{
"epoch": 0.003031834259727135,
"grad_norm": 3.0733749866485596,
"learning_rate": 3.205128205128205e-06,
"loss": 0.668,
"step": 6
},
{
"epoch": 0.0035371399696816574,
"grad_norm": 1.7501100301742554,
"learning_rate": 3.846153846153847e-06,
"loss": 0.5054,
"step": 7
},
{
"epoch": 0.00404244567963618,
"grad_norm": 2.8912413120269775,
"learning_rate": 4.487179487179488e-06,
"loss": 0.8734,
"step": 8
},
{
"epoch": 0.004547751389590703,
"grad_norm": 1.4918310642242432,
"learning_rate": 5.128205128205128e-06,
"loss": 0.8562,
"step": 9
},
{
"epoch": 0.005053057099545225,
"grad_norm": 2.0631442070007324,
"learning_rate": 5.76923076923077e-06,
"loss": 0.7117,
"step": 10
},
{
"epoch": 0.0055583628094997475,
"grad_norm": 1.4029316902160645,
"learning_rate": 6.41025641025641e-06,
"loss": 0.5705,
"step": 11
},
{
"epoch": 0.00606366851945427,
"grad_norm": 1.4444737434387207,
"learning_rate": 7.051282051282052e-06,
"loss": 0.661,
"step": 12
},
{
"epoch": 0.006568974229408792,
"grad_norm": 1.2494786977767944,
"learning_rate": 7.692307692307694e-06,
"loss": 0.7558,
"step": 13
},
{
"epoch": 0.007074279939363315,
"grad_norm": 1.3939992189407349,
"learning_rate": 8.333333333333334e-06,
"loss": 0.5186,
"step": 14
},
{
"epoch": 0.007579585649317837,
"grad_norm": 1.0526946783065796,
"learning_rate": 8.974358974358976e-06,
"loss": 0.543,
"step": 15
},
{
"epoch": 0.00808489135927236,
"grad_norm": 0.7698296904563904,
"learning_rate": 9.615384615384616e-06,
"loss": 0.366,
"step": 16
},
{
"epoch": 0.008590197069226882,
"grad_norm": 1.036993145942688,
"learning_rate": 1.0256410256410256e-05,
"loss": 0.4858,
"step": 17
},
{
"epoch": 0.009095502779181405,
"grad_norm": 1.1180205345153809,
"learning_rate": 1.0897435897435898e-05,
"loss": 0.5841,
"step": 18
},
{
"epoch": 0.009600808489135927,
"grad_norm": 0.8241035342216492,
"learning_rate": 1.153846153846154e-05,
"loss": 0.5563,
"step": 19
},
{
"epoch": 0.01010611419909045,
"grad_norm": 0.6611151695251465,
"learning_rate": 1.217948717948718e-05,
"loss": 0.3478,
"step": 20
},
{
"epoch": 0.010611419909044972,
"grad_norm": 0.4365403950214386,
"learning_rate": 1.282051282051282e-05,
"loss": 0.2442,
"step": 21
},
{
"epoch": 0.011116725618999495,
"grad_norm": 0.7351874709129333,
"learning_rate": 1.3461538461538462e-05,
"loss": 0.3456,
"step": 22
},
{
"epoch": 0.011622031328954016,
"grad_norm": 0.55842524766922,
"learning_rate": 1.4102564102564104e-05,
"loss": 0.3809,
"step": 23
},
{
"epoch": 0.01212733703890854,
"grad_norm": 0.9409593343734741,
"learning_rate": 1.4743589743589745e-05,
"loss": 0.4769,
"step": 24
},
{
"epoch": 0.012632642748863061,
"grad_norm": 0.803429901599884,
"learning_rate": 1.5384615384615387e-05,
"loss": 0.5099,
"step": 25
},
{
"epoch": 0.013137948458817585,
"grad_norm": 0.9147859215736389,
"learning_rate": 1.602564102564103e-05,
"loss": 0.5725,
"step": 26
},
{
"epoch": 0.013643254168772108,
"grad_norm": 0.6797801852226257,
"learning_rate": 1.6666666666666667e-05,
"loss": 0.3108,
"step": 27
},
{
"epoch": 0.01414855987872663,
"grad_norm": 0.6329852342605591,
"learning_rate": 1.730769230769231e-05,
"loss": 0.4665,
"step": 28
},
{
"epoch": 0.014653865588681153,
"grad_norm": 0.49528101086616516,
"learning_rate": 1.794871794871795e-05,
"loss": 0.3252,
"step": 29
},
{
"epoch": 0.015159171298635674,
"grad_norm": 0.6480412483215332,
"learning_rate": 1.858974358974359e-05,
"loss": 0.4001,
"step": 30
},
{
"epoch": 0.015664477008590198,
"grad_norm": 1.0041990280151367,
"learning_rate": 1.923076923076923e-05,
"loss": 0.635,
"step": 31
},
{
"epoch": 0.01616978271854472,
"grad_norm": 0.7632623314857483,
"learning_rate": 1.987179487179487e-05,
"loss": 0.3599,
"step": 32
},
{
"epoch": 0.01667508842849924,
"grad_norm": 0.8572860360145569,
"learning_rate": 2.0512820512820512e-05,
"loss": 0.4534,
"step": 33
},
{
"epoch": 0.017180394138453764,
"grad_norm": 0.9279261231422424,
"learning_rate": 2.1153846153846154e-05,
"loss": 0.4895,
"step": 34
},
{
"epoch": 0.017685699848408287,
"grad_norm": 0.5363617539405823,
"learning_rate": 2.1794871794871795e-05,
"loss": 0.284,
"step": 35
},
{
"epoch": 0.01819100555836281,
"grad_norm": 0.6156304478645325,
"learning_rate": 2.2435897435897437e-05,
"loss": 0.3305,
"step": 36
},
{
"epoch": 0.01869631126831733,
"grad_norm": 0.8967059850692749,
"learning_rate": 2.307692307692308e-05,
"loss": 0.5963,
"step": 37
},
{
"epoch": 0.019201616978271854,
"grad_norm": 0.8374812006950378,
"learning_rate": 2.3717948717948718e-05,
"loss": 0.3155,
"step": 38
},
{
"epoch": 0.019706922688226377,
"grad_norm": 0.7348405718803406,
"learning_rate": 2.435897435897436e-05,
"loss": 0.3585,
"step": 39
},
{
"epoch": 0.0202122283981809,
"grad_norm": 0.7170049548149109,
"learning_rate": 2.5e-05,
"loss": 0.3285,
"step": 40
},
{
"epoch": 0.020717534108135423,
"grad_norm": 0.7275897860527039,
"learning_rate": 2.564102564102564e-05,
"loss": 0.3612,
"step": 41
},
{
"epoch": 0.021222839818089943,
"grad_norm": 0.7469742298126221,
"learning_rate": 2.6282051282051285e-05,
"loss": 0.3367,
"step": 42
},
{
"epoch": 0.021728145528044467,
"grad_norm": 0.6935989856719971,
"learning_rate": 2.6923076923076923e-05,
"loss": 0.2455,
"step": 43
},
{
"epoch": 0.02223345123799899,
"grad_norm": 0.7077957391738892,
"learning_rate": 2.756410256410257e-05,
"loss": 0.4161,
"step": 44
},
{
"epoch": 0.022738756947953513,
"grad_norm": 0.38968655467033386,
"learning_rate": 2.8205128205128207e-05,
"loss": 0.1841,
"step": 45
},
{
"epoch": 0.023244062657908033,
"grad_norm": 0.9259281158447266,
"learning_rate": 2.8846153846153845e-05,
"loss": 0.4498,
"step": 46
},
{
"epoch": 0.023749368367862556,
"grad_norm": 1.0964964628219604,
"learning_rate": 2.948717948717949e-05,
"loss": 0.5429,
"step": 47
},
{
"epoch": 0.02425467407781708,
"grad_norm": 0.4879132807254791,
"learning_rate": 3.012820512820513e-05,
"loss": 0.2142,
"step": 48
},
{
"epoch": 0.024759979787771603,
"grad_norm": 0.6453592777252197,
"learning_rate": 3.0769230769230774e-05,
"loss": 0.2836,
"step": 49
},
{
"epoch": 0.025265285497726123,
"grad_norm": 0.5058504343032837,
"learning_rate": 3.141025641025641e-05,
"loss": 0.2157,
"step": 50
},
{
"epoch": 0.025770591207680646,
"grad_norm": 0.8575659990310669,
"learning_rate": 3.205128205128206e-05,
"loss": 0.6295,
"step": 51
},
{
"epoch": 0.02627589691763517,
"grad_norm": 0.48030632734298706,
"learning_rate": 3.269230769230769e-05,
"loss": 0.2874,
"step": 52
},
{
"epoch": 0.026781202627589692,
"grad_norm": 0.8366153240203857,
"learning_rate": 3.3333333333333335e-05,
"loss": 0.4651,
"step": 53
},
{
"epoch": 0.027286508337544216,
"grad_norm": 0.5325554609298706,
"learning_rate": 3.397435897435898e-05,
"loss": 0.2285,
"step": 54
},
{
"epoch": 0.027791814047498736,
"grad_norm": 0.764580488204956,
"learning_rate": 3.461538461538462e-05,
"loss": 0.428,
"step": 55
},
{
"epoch": 0.02829711975745326,
"grad_norm": 0.8604835867881775,
"learning_rate": 3.525641025641026e-05,
"loss": 0.5092,
"step": 56
},
{
"epoch": 0.028802425467407782,
"grad_norm": 2.482994794845581,
"learning_rate": 3.58974358974359e-05,
"loss": 0.4473,
"step": 57
},
{
"epoch": 0.029307731177362305,
"grad_norm": 0.8809303045272827,
"learning_rate": 3.653846153846154e-05,
"loss": 0.5762,
"step": 58
},
{
"epoch": 0.029813036887316825,
"grad_norm": 0.43776047229766846,
"learning_rate": 3.717948717948718e-05,
"loss": 0.1808,
"step": 59
},
{
"epoch": 0.03031834259727135,
"grad_norm": 0.6836262941360474,
"learning_rate": 3.782051282051282e-05,
"loss": 0.2987,
"step": 60
},
{
"epoch": 0.030823648307225872,
"grad_norm": 0.5660818815231323,
"learning_rate": 3.846153846153846e-05,
"loss": 0.2198,
"step": 61
},
{
"epoch": 0.031328954017180395,
"grad_norm": 0.4436876177787781,
"learning_rate": 3.9102564102564105e-05,
"loss": 0.1963,
"step": 62
},
{
"epoch": 0.03183425972713492,
"grad_norm": 0.7230072617530823,
"learning_rate": 3.974358974358974e-05,
"loss": 0.4539,
"step": 63
},
{
"epoch": 0.03233956543708944,
"grad_norm": 0.9129766225814819,
"learning_rate": 4.038461538461539e-05,
"loss": 0.3737,
"step": 64
},
{
"epoch": 0.032844871147043965,
"grad_norm": 1.1869587898254395,
"learning_rate": 4.1025641025641023e-05,
"loss": 0.696,
"step": 65
},
{
"epoch": 0.03335017685699848,
"grad_norm": 0.7398460507392883,
"learning_rate": 4.166666666666667e-05,
"loss": 0.5932,
"step": 66
},
{
"epoch": 0.033855482566953005,
"grad_norm": 0.7067481875419617,
"learning_rate": 4.230769230769231e-05,
"loss": 0.3213,
"step": 67
},
{
"epoch": 0.03436078827690753,
"grad_norm": 0.9902649521827698,
"learning_rate": 4.294871794871795e-05,
"loss": 0.416,
"step": 68
},
{
"epoch": 0.03486609398686205,
"grad_norm": 0.42111849784851074,
"learning_rate": 4.358974358974359e-05,
"loss": 0.2465,
"step": 69
},
{
"epoch": 0.035371399696816574,
"grad_norm": 0.9634829759597778,
"learning_rate": 4.423076923076923e-05,
"loss": 0.5916,
"step": 70
},
{
"epoch": 0.0358767054067711,
"grad_norm": 0.8240818977355957,
"learning_rate": 4.4871794871794874e-05,
"loss": 0.4181,
"step": 71
},
{
"epoch": 0.03638201111672562,
"grad_norm": 0.7216759324073792,
"learning_rate": 4.5512820512820516e-05,
"loss": 0.4966,
"step": 72
},
{
"epoch": 0.036887316826680144,
"grad_norm": 0.7505428194999695,
"learning_rate": 4.615384615384616e-05,
"loss": 0.3799,
"step": 73
},
{
"epoch": 0.03739262253663466,
"grad_norm": 1.2560772895812988,
"learning_rate": 4.67948717948718e-05,
"loss": 0.4691,
"step": 74
},
{
"epoch": 0.037897928246589184,
"grad_norm": 0.7000345587730408,
"learning_rate": 4.7435897435897435e-05,
"loss": 0.2898,
"step": 75
},
{
"epoch": 0.03840323395654371,
"grad_norm": 0.6096041202545166,
"learning_rate": 4.8076923076923084e-05,
"loss": 0.2929,
"step": 76
},
{
"epoch": 0.03890853966649823,
"grad_norm": 0.6045755743980408,
"learning_rate": 4.871794871794872e-05,
"loss": 0.282,
"step": 77
},
{
"epoch": 0.039413845376452754,
"grad_norm": 0.9146702885627747,
"learning_rate": 4.935897435897436e-05,
"loss": 0.4637,
"step": 78
},
{
"epoch": 0.03991915108640728,
"grad_norm": 0.9550428986549377,
"learning_rate": 5e-05,
"loss": 0.3772,
"step": 79
},
{
"epoch": 0.0404244567963618,
"grad_norm": 0.7160158157348633,
"learning_rate": 4.999998018159948e-05,
"loss": 0.3099,
"step": 80
},
{
"epoch": 0.040929762506316324,
"grad_norm": 0.6506277322769165,
"learning_rate": 4.999992072642933e-05,
"loss": 0.3599,
"step": 81
},
{
"epoch": 0.04143506821627085,
"grad_norm": 0.8049689531326294,
"learning_rate": 4.999982163458383e-05,
"loss": 0.3184,
"step": 82
},
{
"epoch": 0.04194037392622536,
"grad_norm": 1.0008819103240967,
"learning_rate": 4.999968290622007e-05,
"loss": 0.4903,
"step": 83
},
{
"epoch": 0.04244567963617989,
"grad_norm": 0.6113778352737427,
"learning_rate": 4.999950454155801e-05,
"loss": 0.2561,
"step": 84
},
{
"epoch": 0.04295098534613441,
"grad_norm": 1.4032593965530396,
"learning_rate": 4.999928654088044e-05,
"loss": 0.3617,
"step": 85
},
{
"epoch": 0.04345629105608893,
"grad_norm": 0.5701761841773987,
"learning_rate": 4.9999028904532996e-05,
"loss": 0.3419,
"step": 86
},
{
"epoch": 0.043961596766043456,
"grad_norm": 0.8148670792579651,
"learning_rate": 4.9998731632924144e-05,
"loss": 0.2048,
"step": 87
},
{
"epoch": 0.04446690247599798,
"grad_norm": 1.7208861112594604,
"learning_rate": 4.999839472652522e-05,
"loss": 0.4322,
"step": 88
},
{
"epoch": 0.0449722081859525,
"grad_norm": 2.0720438957214355,
"learning_rate": 4.9998018185870356e-05,
"loss": 0.2536,
"step": 89
},
{
"epoch": 0.045477513895907026,
"grad_norm": 1.085567831993103,
"learning_rate": 4.999760201155657e-05,
"loss": 0.5578,
"step": 90
},
{
"epoch": 0.04598281960586155,
"grad_norm": 0.5517830848693848,
"learning_rate": 4.999714620424367e-05,
"loss": 0.2543,
"step": 91
},
{
"epoch": 0.046488125315816066,
"grad_norm": 0.6557161808013916,
"learning_rate": 4.999665076465434e-05,
"loss": 0.4033,
"step": 92
},
{
"epoch": 0.04699343102577059,
"grad_norm": 0.8143996000289917,
"learning_rate": 4.9996115693574086e-05,
"loss": 0.3873,
"step": 93
},
{
"epoch": 0.04749873673572511,
"grad_norm": 0.9482316970825195,
"learning_rate": 4.999554099185124e-05,
"loss": 0.3902,
"step": 94
},
{
"epoch": 0.048004042445679636,
"grad_norm": 0.6633082032203674,
"learning_rate": 4.999492666039699e-05,
"loss": 0.3041,
"step": 95
},
{
"epoch": 0.04850934815563416,
"grad_norm": 1.107414960861206,
"learning_rate": 4.999427270018533e-05,
"loss": 0.4195,
"step": 96
},
{
"epoch": 0.04901465386558868,
"grad_norm": 0.4633142948150635,
"learning_rate": 4.99935791122531e-05,
"loss": 0.22,
"step": 97
},
{
"epoch": 0.049519959575543206,
"grad_norm": 0.7195435166358948,
"learning_rate": 4.999284589769996e-05,
"loss": 0.2776,
"step": 98
},
{
"epoch": 0.05002526528549773,
"grad_norm": 0.38064494729042053,
"learning_rate": 4.9992073057688403e-05,
"loss": 0.2126,
"step": 99
},
{
"epoch": 0.050530570995452245,
"grad_norm": 0.4273965060710907,
"learning_rate": 4.9991260593443745e-05,
"loss": 0.2464,
"step": 100
},
{
"epoch": 0.05103587670540677,
"grad_norm": 0.7326411008834839,
"learning_rate": 4.9990408506254134e-05,
"loss": 0.3473,
"step": 101
},
{
"epoch": 0.05154118241536129,
"grad_norm": 0.6799935102462769,
"learning_rate": 4.998951679747051e-05,
"loss": 0.3258,
"step": 102
},
{
"epoch": 0.052046488125315815,
"grad_norm": 0.6898297071456909,
"learning_rate": 4.998858546850668e-05,
"loss": 0.6305,
"step": 103
},
{
"epoch": 0.05255179383527034,
"grad_norm": 0.5631976127624512,
"learning_rate": 4.998761452083922e-05,
"loss": 0.3241,
"step": 104
},
{
"epoch": 0.05305709954522486,
"grad_norm": 0.4945349395275116,
"learning_rate": 4.998660395600754e-05,
"loss": 0.3002,
"step": 105
},
{
"epoch": 0.053562405255179385,
"grad_norm": 0.6329333186149597,
"learning_rate": 4.998555377561387e-05,
"loss": 0.3074,
"step": 106
},
{
"epoch": 0.05406771096513391,
"grad_norm": 0.532968282699585,
"learning_rate": 4.9984463981323246e-05,
"loss": 0.334,
"step": 107
},
{
"epoch": 0.05457301667508843,
"grad_norm": 0.7290483713150024,
"learning_rate": 4.998333457486349e-05,
"loss": 0.2753,
"step": 108
},
{
"epoch": 0.05507832238504295,
"grad_norm": 0.7623900771141052,
"learning_rate": 4.998216555802526e-05,
"loss": 0.4656,
"step": 109
},
{
"epoch": 0.05558362809499747,
"grad_norm": 0.6192227005958557,
"learning_rate": 4.9980956932661996e-05,
"loss": 0.258,
"step": 110
},
{
"epoch": 0.056088933804951994,
"grad_norm": 1.0014296770095825,
"learning_rate": 4.9979708700689934e-05,
"loss": 0.5037,
"step": 111
},
{
"epoch": 0.05659423951490652,
"grad_norm": 0.5563094615936279,
"learning_rate": 4.9978420864088115e-05,
"loss": 0.3379,
"step": 112
},
{
"epoch": 0.05709954522486104,
"grad_norm": 0.9516816139221191,
"learning_rate": 4.997709342489837e-05,
"loss": 0.4447,
"step": 113
},
{
"epoch": 0.057604850934815564,
"grad_norm": 0.7699498534202576,
"learning_rate": 4.997572638522531e-05,
"loss": 0.5482,
"step": 114
},
{
"epoch": 0.05811015664477009,
"grad_norm": 0.623145580291748,
"learning_rate": 4.9974319747236344e-05,
"loss": 0.3454,
"step": 115
},
{
"epoch": 0.05861546235472461,
"grad_norm": 0.62673419713974,
"learning_rate": 4.997287351316165e-05,
"loss": 0.4063,
"step": 116
},
{
"epoch": 0.059120768064679134,
"grad_norm": 0.6477133631706238,
"learning_rate": 4.997138768529421e-05,
"loss": 0.2508,
"step": 117
},
{
"epoch": 0.05962607377463365,
"grad_norm": 0.650576114654541,
"learning_rate": 4.9969862265989745e-05,
"loss": 0.3157,
"step": 118
},
{
"epoch": 0.060131379484588174,
"grad_norm": 0.6956886053085327,
"learning_rate": 4.9968297257666764e-05,
"loss": 0.2984,
"step": 119
},
{
"epoch": 0.0606366851945427,
"grad_norm": 1.090057611465454,
"learning_rate": 4.996669266280655e-05,
"loss": 0.5217,
"step": 120
},
{
"epoch": 0.06114199090449722,
"grad_norm": 0.6649771928787231,
"learning_rate": 4.996504848395314e-05,
"loss": 0.4094,
"step": 121
},
{
"epoch": 0.061647296614451744,
"grad_norm": 1.0914398431777954,
"learning_rate": 4.9963364723713344e-05,
"loss": 0.7907,
"step": 122
},
{
"epoch": 0.06215260232440627,
"grad_norm": 1.1362191438674927,
"learning_rate": 4.9961641384756705e-05,
"loss": 0.5202,
"step": 123
},
{
"epoch": 0.06265790803436079,
"grad_norm": 0.6446903347969055,
"learning_rate": 4.9959878469815534e-05,
"loss": 0.5434,
"step": 124
},
{
"epoch": 0.06316321374431531,
"grad_norm": 0.5553262233734131,
"learning_rate": 4.9958075981684884e-05,
"loss": 0.2622,
"step": 125
},
{
"epoch": 0.06366851945426984,
"grad_norm": 0.7034339904785156,
"learning_rate": 4.995623392322254e-05,
"loss": 0.4785,
"step": 126
},
{
"epoch": 0.06417382516422436,
"grad_norm": 0.8382132649421692,
"learning_rate": 4.995435229734905e-05,
"loss": 0.3813,
"step": 127
},
{
"epoch": 0.06467913087417888,
"grad_norm": 0.8590944409370422,
"learning_rate": 4.995243110704766e-05,
"loss": 0.6019,
"step": 128
},
{
"epoch": 0.0651844365841334,
"grad_norm": 0.8058075904846191,
"learning_rate": 4.9950470355364386e-05,
"loss": 0.4251,
"step": 129
},
{
"epoch": 0.06568974229408793,
"grad_norm": 0.5873439908027649,
"learning_rate": 4.994847004540793e-05,
"loss": 0.4432,
"step": 130
},
{
"epoch": 0.06619504800404244,
"grad_norm": 0.6679672598838806,
"learning_rate": 4.994643018034973e-05,
"loss": 0.4053,
"step": 131
},
{
"epoch": 0.06670035371399696,
"grad_norm": 0.8931362628936768,
"learning_rate": 4.994435076342394e-05,
"loss": 0.5092,
"step": 132
},
{
"epoch": 0.06720565942395149,
"grad_norm": 0.4417133033275604,
"learning_rate": 4.994223179792741e-05,
"loss": 0.2055,
"step": 133
},
{
"epoch": 0.06771096513390601,
"grad_norm": 0.3722580671310425,
"learning_rate": 4.9940073287219703e-05,
"loss": 0.2183,
"step": 134
},
{
"epoch": 0.06821627084386053,
"grad_norm": 1.3628448247909546,
"learning_rate": 4.9937875234723075e-05,
"loss": 0.4615,
"step": 135
},
{
"epoch": 0.06872157655381506,
"grad_norm": 0.6365114450454712,
"learning_rate": 4.993563764392248e-05,
"loss": 0.3071,
"step": 136
},
{
"epoch": 0.06922688226376958,
"grad_norm": 0.5635626912117004,
"learning_rate": 4.9933360518365566e-05,
"loss": 0.3514,
"step": 137
},
{
"epoch": 0.0697321879737241,
"grad_norm": 0.7801100611686707,
"learning_rate": 4.993104386166263e-05,
"loss": 0.3682,
"step": 138
},
{
"epoch": 0.07023749368367863,
"grad_norm": 0.7181437611579895,
"learning_rate": 4.9928687677486685e-05,
"loss": 0.3864,
"step": 139
},
{
"epoch": 0.07074279939363315,
"grad_norm": 0.7251635193824768,
"learning_rate": 4.992629196957339e-05,
"loss": 0.4515,
"step": 140
},
{
"epoch": 0.07124810510358767,
"grad_norm": 0.45708200335502625,
"learning_rate": 4.992385674172107e-05,
"loss": 0.2097,
"step": 141
},
{
"epoch": 0.0717534108135422,
"grad_norm": 0.5553367733955383,
"learning_rate": 4.9921381997790705e-05,
"loss": 0.246,
"step": 142
},
{
"epoch": 0.07225871652349672,
"grad_norm": 0.6126506924629211,
"learning_rate": 4.991886774170594e-05,
"loss": 0.2903,
"step": 143
},
{
"epoch": 0.07276402223345124,
"grad_norm": 0.8143837451934814,
"learning_rate": 4.991631397745307e-05,
"loss": 0.3333,
"step": 144
},
{
"epoch": 0.07326932794340577,
"grad_norm": 1.035240888595581,
"learning_rate": 4.991372070908099e-05,
"loss": 0.6579,
"step": 145
},
{
"epoch": 0.07377463365336029,
"grad_norm": 0.3664172887802124,
"learning_rate": 4.9911087940701274e-05,
"loss": 0.1993,
"step": 146
},
{
"epoch": 0.07427993936331481,
"grad_norm": 0.42232248187065125,
"learning_rate": 4.9908415676488094e-05,
"loss": 0.211,
"step": 147
},
{
"epoch": 0.07478524507326932,
"grad_norm": 0.7374719381332397,
"learning_rate": 4.990570392067826e-05,
"loss": 0.2841,
"step": 148
},
{
"epoch": 0.07529055078322384,
"grad_norm": 0.7315521240234375,
"learning_rate": 4.990295267757117e-05,
"loss": 0.4779,
"step": 149
},
{
"epoch": 0.07579585649317837,
"grad_norm": 0.8291402459144592,
"learning_rate": 4.9900161951528854e-05,
"loss": 0.5315,
"step": 150
},
{
"epoch": 0.07630116220313289,
"grad_norm": 0.9879725575447083,
"learning_rate": 4.9897331746975926e-05,
"loss": 0.5184,
"step": 151
},
{
"epoch": 0.07680646791308741,
"grad_norm": 0.6654685139656067,
"learning_rate": 4.98944620683996e-05,
"loss": 0.4268,
"step": 152
},
{
"epoch": 0.07731177362304194,
"grad_norm": 0.6749838590621948,
"learning_rate": 4.989155292034966e-05,
"loss": 0.3266,
"step": 153
},
{
"epoch": 0.07781707933299646,
"grad_norm": 0.5096444487571716,
"learning_rate": 4.98886043074385e-05,
"loss": 0.4308,
"step": 154
},
{
"epoch": 0.07832238504295098,
"grad_norm": 0.5235171914100647,
"learning_rate": 4.988561623434105e-05,
"loss": 0.3358,
"step": 155
},
{
"epoch": 0.07882769075290551,
"grad_norm": 0.39268213510513306,
"learning_rate": 4.988258870579481e-05,
"loss": 0.2653,
"step": 156
},
{
"epoch": 0.07933299646286003,
"grad_norm": 0.6037962436676025,
"learning_rate": 4.987952172659985e-05,
"loss": 0.1633,
"step": 157
},
{
"epoch": 0.07983830217281455,
"grad_norm": 0.4751133918762207,
"learning_rate": 4.9876415301618787e-05,
"loss": 0.2796,
"step": 158
},
{
"epoch": 0.08034360788276908,
"grad_norm": 0.7282652854919434,
"learning_rate": 4.987326943577675e-05,
"loss": 0.5218,
"step": 159
},
{
"epoch": 0.0808489135927236,
"grad_norm": 0.42741066217422485,
"learning_rate": 4.987008413406144e-05,
"loss": 0.1733,
"step": 160
},
{
"epoch": 0.08135421930267812,
"grad_norm": 0.5225434899330139,
"learning_rate": 4.9866859401523056e-05,
"loss": 0.1923,
"step": 161
},
{
"epoch": 0.08185952501263265,
"grad_norm": 0.5940921902656555,
"learning_rate": 4.986359524327433e-05,
"loss": 0.2548,
"step": 162
},
{
"epoch": 0.08236483072258717,
"grad_norm": 1.0237605571746826,
"learning_rate": 4.986029166449048e-05,
"loss": 0.3913,
"step": 163
},
{
"epoch": 0.0828701364325417,
"grad_norm": 0.44926929473876953,
"learning_rate": 4.985694867040924e-05,
"loss": 0.2304,
"step": 164
},
{
"epoch": 0.0833754421424962,
"grad_norm": 0.968037486076355,
"learning_rate": 4.9853566266330855e-05,
"loss": 0.4391,
"step": 165
},
{
"epoch": 0.08388074785245073,
"grad_norm": 0.6649966835975647,
"learning_rate": 4.9850144457618e-05,
"loss": 0.2662,
"step": 166
},
{
"epoch": 0.08438605356240525,
"grad_norm": 0.5354177951812744,
"learning_rate": 4.984668324969588e-05,
"loss": 0.223,
"step": 167
},
{
"epoch": 0.08489135927235977,
"grad_norm": 0.9306359887123108,
"learning_rate": 4.9843182648052136e-05,
"loss": 0.5532,
"step": 168
},
{
"epoch": 0.0853966649823143,
"grad_norm": 0.2804405987262726,
"learning_rate": 4.983964265823687e-05,
"loss": 0.1795,
"step": 169
},
{
"epoch": 0.08590197069226882,
"grad_norm": 0.6046093702316284,
"learning_rate": 4.9836063285862636e-05,
"loss": 0.3467,
"step": 170
},
{
"epoch": 0.08640727640222334,
"grad_norm": 0.3996451199054718,
"learning_rate": 4.983244453660445e-05,
"loss": 0.2116,
"step": 171
},
{
"epoch": 0.08691258211217787,
"grad_norm": 0.5045155882835388,
"learning_rate": 4.98287864161997e-05,
"loss": 0.2028,
"step": 172
},
{
"epoch": 0.08741788782213239,
"grad_norm": 0.7425013780593872,
"learning_rate": 4.982508893044827e-05,
"loss": 0.4183,
"step": 173
},
{
"epoch": 0.08792319353208691,
"grad_norm": 0.4544399380683899,
"learning_rate": 4.98213520852124e-05,
"loss": 0.2321,
"step": 174
},
{
"epoch": 0.08842849924204144,
"grad_norm": 0.7836821675300598,
"learning_rate": 4.981757588641676e-05,
"loss": 0.4456,
"step": 175
},
{
"epoch": 0.08893380495199596,
"grad_norm": 0.24565193057060242,
"learning_rate": 4.981376034004841e-05,
"loss": 0.1439,
"step": 176
},
{
"epoch": 0.08943911066195048,
"grad_norm": 0.3137032687664032,
"learning_rate": 4.980990545215678e-05,
"loss": 0.1513,
"step": 177
},
{
"epoch": 0.089944416371905,
"grad_norm": 0.6917077302932739,
"learning_rate": 4.98060112288537e-05,
"loss": 0.3292,
"step": 178
},
{
"epoch": 0.09044972208185953,
"grad_norm": 0.6623265147209167,
"learning_rate": 4.9802077676313354e-05,
"loss": 0.381,
"step": 179
},
{
"epoch": 0.09095502779181405,
"grad_norm": 0.6027669906616211,
"learning_rate": 4.979810480077227e-05,
"loss": 0.3129,
"step": 180
},
{
"epoch": 0.09146033350176858,
"grad_norm": 0.4134594798088074,
"learning_rate": 4.979409260852933e-05,
"loss": 0.1588,
"step": 181
},
{
"epoch": 0.0919656392117231,
"grad_norm": 0.33261945843696594,
"learning_rate": 4.979004110594576e-05,
"loss": 0.1806,
"step": 182
},
{
"epoch": 0.09247094492167761,
"grad_norm": 1.0738641023635864,
"learning_rate": 4.97859502994451e-05,
"loss": 0.4609,
"step": 183
},
{
"epoch": 0.09297625063163213,
"grad_norm": 1.1859742403030396,
"learning_rate": 4.978182019551321e-05,
"loss": 0.5995,
"step": 184
},
{
"epoch": 0.09348155634158666,
"grad_norm": 0.5822703838348389,
"learning_rate": 4.977765080069825e-05,
"loss": 0.2734,
"step": 185
},
{
"epoch": 0.09398686205154118,
"grad_norm": 0.9306544065475464,
"learning_rate": 4.977344212161069e-05,
"loss": 0.5029,
"step": 186
},
{
"epoch": 0.0944921677614957,
"grad_norm": 0.3687601685523987,
"learning_rate": 4.976919416492327e-05,
"loss": 0.2571,
"step": 187
},
{
"epoch": 0.09499747347145023,
"grad_norm": 0.7644825577735901,
"learning_rate": 4.9764906937370995e-05,
"loss": 0.4293,
"step": 188
},
{
"epoch": 0.09550277918140475,
"grad_norm": 0.6048409342765808,
"learning_rate": 4.976058044575116e-05,
"loss": 0.252,
"step": 189
},
{
"epoch": 0.09600808489135927,
"grad_norm": 0.89262455701828,
"learning_rate": 4.975621469692329e-05,
"loss": 0.4889,
"step": 190
},
{
"epoch": 0.0965133906013138,
"grad_norm": 0.7101221680641174,
"learning_rate": 4.9751809697809145e-05,
"loss": 0.2739,
"step": 191
},
{
"epoch": 0.09701869631126832,
"grad_norm": 0.29569289088249207,
"learning_rate": 4.974736545539275e-05,
"loss": 0.1573,
"step": 192
},
{
"epoch": 0.09752400202122284,
"grad_norm": 0.5595278739929199,
"learning_rate": 4.9742881976720314e-05,
"loss": 0.335,
"step": 193
},
{
"epoch": 0.09802930773117736,
"grad_norm": 0.6488416790962219,
"learning_rate": 4.973835926890027e-05,
"loss": 0.4142,
"step": 194
},
{
"epoch": 0.09853461344113189,
"grad_norm": 0.6213271021842957,
"learning_rate": 4.973379733910325e-05,
"loss": 0.3829,
"step": 195
},
{
"epoch": 0.09903991915108641,
"grad_norm": 0.5019063353538513,
"learning_rate": 4.972919619456205e-05,
"loss": 0.275,
"step": 196
},
{
"epoch": 0.09954522486104093,
"grad_norm": 0.9575361609458923,
"learning_rate": 4.972455584257168e-05,
"loss": 0.507,
"step": 197
},
{
"epoch": 0.10005053057099546,
"grad_norm": 0.36353740096092224,
"learning_rate": 4.9719876290489264e-05,
"loss": 0.2864,
"step": 198
},
{
"epoch": 0.10055583628094998,
"grad_norm": 0.46957746148109436,
"learning_rate": 4.971515754573412e-05,
"loss": 0.2227,
"step": 199
},
{
"epoch": 0.10106114199090449,
"grad_norm": 0.7227563858032227,
"learning_rate": 4.9710399615787676e-05,
"loss": 0.5132,
"step": 200
},
{
"epoch": 0.10156644770085901,
"grad_norm": 0.4742310643196106,
"learning_rate": 4.9705602508193496e-05,
"loss": 0.1944,
"step": 201
},
{
"epoch": 0.10207175341081354,
"grad_norm": 0.7540463209152222,
"learning_rate": 4.970076623055727e-05,
"loss": 0.2761,
"step": 202
},
{
"epoch": 0.10257705912076806,
"grad_norm": 0.7352909445762634,
"learning_rate": 4.9695890790546765e-05,
"loss": 0.486,
"step": 203
},
{
"epoch": 0.10308236483072258,
"grad_norm": 0.6061909794807434,
"learning_rate": 4.969097619589187e-05,
"loss": 0.5126,
"step": 204
},
{
"epoch": 0.10358767054067711,
"grad_norm": 0.8726075291633606,
"learning_rate": 4.968602245438453e-05,
"loss": 0.5034,
"step": 205
},
{
"epoch": 0.10409297625063163,
"grad_norm": 0.9859480857849121,
"learning_rate": 4.968102957387877e-05,
"loss": 0.4496,
"step": 206
},
{
"epoch": 0.10459828196058615,
"grad_norm": 0.5186076164245605,
"learning_rate": 4.967599756229065e-05,
"loss": 0.3507,
"step": 207
},
{
"epoch": 0.10510358767054068,
"grad_norm": 0.4728350043296814,
"learning_rate": 4.967092642759829e-05,
"loss": 0.1672,
"step": 208
},
{
"epoch": 0.1056088933804952,
"grad_norm": 0.6087246537208557,
"learning_rate": 4.966581617784184e-05,
"loss": 0.3796,
"step": 209
},
{
"epoch": 0.10611419909044972,
"grad_norm": 0.8859485983848572,
"learning_rate": 4.966066682112345e-05,
"loss": 0.5639,
"step": 210
},
{
"epoch": 0.10661950480040425,
"grad_norm": 0.8429753184318542,
"learning_rate": 4.9655478365607276e-05,
"loss": 0.2432,
"step": 211
},
{
"epoch": 0.10712481051035877,
"grad_norm": 0.609099805355072,
"learning_rate": 4.9650250819519485e-05,
"loss": 0.3056,
"step": 212
},
{
"epoch": 0.1076301162203133,
"grad_norm": 0.40738436579704285,
"learning_rate": 4.964498419114819e-05,
"loss": 0.1862,
"step": 213
},
{
"epoch": 0.10813542193026782,
"grad_norm": 0.7260504364967346,
"learning_rate": 4.963967848884349e-05,
"loss": 0.5646,
"step": 214
},
{
"epoch": 0.10864072764022234,
"grad_norm": 1.013758659362793,
"learning_rate": 4.963433372101743e-05,
"loss": 0.4497,
"step": 215
},
{
"epoch": 0.10914603335017686,
"grad_norm": 0.41159892082214355,
"learning_rate": 4.962894989614398e-05,
"loss": 0.3157,
"step": 216
},
{
"epoch": 0.10965133906013137,
"grad_norm": 0.42557021975517273,
"learning_rate": 4.9623527022759054e-05,
"loss": 0.1879,
"step": 217
},
{
"epoch": 0.1101566447700859,
"grad_norm": 0.7539249658584595,
"learning_rate": 4.9618065109460465e-05,
"loss": 0.4206,
"step": 218
},
{
"epoch": 0.11066195048004042,
"grad_norm": 0.6201863884925842,
"learning_rate": 4.961256416490792e-05,
"loss": 0.2961,
"step": 219
},
{
"epoch": 0.11116725618999494,
"grad_norm": 0.7610240578651428,
"learning_rate": 4.9607024197823014e-05,
"loss": 0.4731,
"step": 220
},
{
"epoch": 0.11167256189994947,
"grad_norm": 0.4443597197532654,
"learning_rate": 4.9601445216989206e-05,
"loss": 0.2716,
"step": 221
},
{
"epoch": 0.11217786760990399,
"grad_norm": 0.47122883796691895,
"learning_rate": 4.9595827231251824e-05,
"loss": 0.1896,
"step": 222
},
{
"epoch": 0.11268317331985851,
"grad_norm": 0.6130548715591431,
"learning_rate": 4.959017024951802e-05,
"loss": 0.3805,
"step": 223
},
{
"epoch": 0.11318847902981304,
"grad_norm": 0.6293324828147888,
"learning_rate": 4.958447428075678e-05,
"loss": 0.5559,
"step": 224
},
{
"epoch": 0.11369378473976756,
"grad_norm": 0.5696532726287842,
"learning_rate": 4.957873933399891e-05,
"loss": 0.2836,
"step": 225
},
{
"epoch": 0.11419909044972208,
"grad_norm": 0.5803974270820618,
"learning_rate": 4.9572965418336995e-05,
"loss": 0.3246,
"step": 226
},
{
"epoch": 0.1147043961596766,
"grad_norm": 0.6962433457374573,
"learning_rate": 4.956715254292543e-05,
"loss": 0.4295,
"step": 227
},
{
"epoch": 0.11520970186963113,
"grad_norm": 0.8711988925933838,
"learning_rate": 4.9561300716980365e-05,
"loss": 0.6052,
"step": 228
},
{
"epoch": 0.11571500757958565,
"grad_norm": 0.39029747247695923,
"learning_rate": 4.95554099497797e-05,
"loss": 0.2225,
"step": 229
},
{
"epoch": 0.11622031328954018,
"grad_norm": 0.7048918008804321,
"learning_rate": 4.9549480250663086e-05,
"loss": 0.3617,
"step": 230
},
{
"epoch": 0.1167256189994947,
"grad_norm": 0.8495561480522156,
"learning_rate": 4.954351162903188e-05,
"loss": 0.47,
"step": 231
},
{
"epoch": 0.11723092470944922,
"grad_norm": 0.8110531568527222,
"learning_rate": 4.953750409434919e-05,
"loss": 0.5883,
"step": 232
},
{
"epoch": 0.11773623041940375,
"grad_norm": 0.5132966637611389,
"learning_rate": 4.953145765613978e-05,
"loss": 0.2866,
"step": 233
},
{
"epoch": 0.11824153612935827,
"grad_norm": 0.6017918586730957,
"learning_rate": 4.952537232399012e-05,
"loss": 0.3651,
"step": 234
},
{
"epoch": 0.11874684183931278,
"grad_norm": 0.5894491672515869,
"learning_rate": 4.951924810754831e-05,
"loss": 0.2754,
"step": 235
},
{
"epoch": 0.1192521475492673,
"grad_norm": 0.46862131357192993,
"learning_rate": 4.9513085016524153e-05,
"loss": 0.3216,
"step": 236
},
{
"epoch": 0.11975745325922182,
"grad_norm": 0.6514838933944702,
"learning_rate": 4.9506883060689035e-05,
"loss": 0.2915,
"step": 237
},
{
"epoch": 0.12026275896917635,
"grad_norm": 1.0483429431915283,
"learning_rate": 4.950064224987599e-05,
"loss": 0.6246,
"step": 238
},
{
"epoch": 0.12076806467913087,
"grad_norm": 0.3560936748981476,
"learning_rate": 4.949436259397966e-05,
"loss": 0.2231,
"step": 239
},
{
"epoch": 0.1212733703890854,
"grad_norm": 0.7140908241271973,
"learning_rate": 4.948804410295625e-05,
"loss": 0.8288,
"step": 240
},
{
"epoch": 0.12177867609903992,
"grad_norm": 0.8638331294059753,
"learning_rate": 4.9481686786823554e-05,
"loss": 0.4715,
"step": 241
},
{
"epoch": 0.12228398180899444,
"grad_norm": 0.8147590160369873,
"learning_rate": 4.9475290655660926e-05,
"loss": 0.3322,
"step": 242
},
{
"epoch": 0.12278928751894896,
"grad_norm": 1.1923398971557617,
"learning_rate": 4.946885571960925e-05,
"loss": 0.4714,
"step": 243
},
{
"epoch": 0.12329459322890349,
"grad_norm": 0.4850132167339325,
"learning_rate": 4.946238198887093e-05,
"loss": 0.2975,
"step": 244
},
{
"epoch": 0.12379989893885801,
"grad_norm": 0.4749908447265625,
"learning_rate": 4.9455869473709895e-05,
"loss": 0.3996,
"step": 245
},
{
"epoch": 0.12430520464881253,
"grad_norm": 0.4123810827732086,
"learning_rate": 4.9449318184451545e-05,
"loss": 0.1817,
"step": 246
},
{
"epoch": 0.12481051035876706,
"grad_norm": 0.603053629398346,
"learning_rate": 4.944272813148277e-05,
"loss": 0.4783,
"step": 247
},
{
"epoch": 0.12531581606872158,
"grad_norm": 0.2952713668346405,
"learning_rate": 4.943609932525193e-05,
"loss": 0.1619,
"step": 248
},
{
"epoch": 0.1258211217786761,
"grad_norm": 0.7722076773643494,
"learning_rate": 4.942943177626879e-05,
"loss": 0.4847,
"step": 249
},
{
"epoch": 0.12632642748863063,
"grad_norm": 0.4615485966205597,
"learning_rate": 4.942272549510457e-05,
"loss": 0.2843,
"step": 250
},
{
"epoch": 0.12683173319858515,
"grad_norm": 0.501020073890686,
"learning_rate": 4.94159804923919e-05,
"loss": 0.2411,
"step": 251
},
{
"epoch": 0.12733703890853967,
"grad_norm": 0.9735652208328247,
"learning_rate": 4.9409196778824776e-05,
"loss": 0.546,
"step": 252
},
{
"epoch": 0.1278423446184942,
"grad_norm": 0.9183477163314819,
"learning_rate": 4.94023743651586e-05,
"loss": 0.5422,
"step": 253
},
{
"epoch": 0.12834765032844872,
"grad_norm": 0.25691601634025574,
"learning_rate": 4.939551326221012e-05,
"loss": 0.1477,
"step": 254
},
{
"epoch": 0.12885295603840324,
"grad_norm": 1.1621726751327515,
"learning_rate": 4.938861348085742e-05,
"loss": 0.518,
"step": 255
},
{
"epoch": 0.12935826174835777,
"grad_norm": 1.0264886617660522,
"learning_rate": 4.9381675032039896e-05,
"loss": 0.513,
"step": 256
},
{
"epoch": 0.1298635674583123,
"grad_norm": 0.46496647596359253,
"learning_rate": 4.9374697926758275e-05,
"loss": 0.2556,
"step": 257
},
{
"epoch": 0.1303688731682668,
"grad_norm": 0.2897842526435852,
"learning_rate": 4.936768217607457e-05,
"loss": 0.1476,
"step": 258
},
{
"epoch": 0.13087417887822134,
"grad_norm": 1.0029571056365967,
"learning_rate": 4.9360627791112046e-05,
"loss": 0.6114,
"step": 259
},
{
"epoch": 0.13137948458817586,
"grad_norm": 0.42095255851745605,
"learning_rate": 4.935353478305524e-05,
"loss": 0.3028,
"step": 260
},
{
"epoch": 0.13188479029813036,
"grad_norm": 1.4107495546340942,
"learning_rate": 4.9346403163149924e-05,
"loss": 0.677,
"step": 261
},
{
"epoch": 0.13239009600808488,
"grad_norm": 0.33536162972450256,
"learning_rate": 4.9339232942703067e-05,
"loss": 0.1942,
"step": 262
},
{
"epoch": 0.1328954017180394,
"grad_norm": 0.4370989203453064,
"learning_rate": 4.933202413308286e-05,
"loss": 0.3653,
"step": 263
},
{
"epoch": 0.13340070742799393,
"grad_norm": 1.1874058246612549,
"learning_rate": 4.932477674571867e-05,
"loss": 0.5772,
"step": 264
},
{
"epoch": 0.13390601313794845,
"grad_norm": 0.6043871641159058,
"learning_rate": 4.931749079210103e-05,
"loss": 0.2519,
"step": 265
},
{
"epoch": 0.13441131884790297,
"grad_norm": 0.5337929129600525,
"learning_rate": 4.931016628378161e-05,
"loss": 0.2081,
"step": 266
},
{
"epoch": 0.1349166245578575,
"grad_norm": 0.6960159540176392,
"learning_rate": 4.930280323237321e-05,
"loss": 0.3724,
"step": 267
},
{
"epoch": 0.13542193026781202,
"grad_norm": 0.3179565966129303,
"learning_rate": 4.9295401649549755e-05,
"loss": 0.2135,
"step": 268
},
{
"epoch": 0.13592723597776654,
"grad_norm": 0.4810411036014557,
"learning_rate": 4.9287961547046234e-05,
"loss": 0.2307,
"step": 269
},
{
"epoch": 0.13643254168772107,
"grad_norm": 0.4444022476673126,
"learning_rate": 4.928048293665873e-05,
"loss": 0.3267,
"step": 270
},
{
"epoch": 0.1369378473976756,
"grad_norm": 0.5844589471817017,
"learning_rate": 4.9272965830244366e-05,
"loss": 0.2969,
"step": 271
},
{
"epoch": 0.1374431531076301,
"grad_norm": 0.37751492857933044,
"learning_rate": 4.9265410239721313e-05,
"loss": 0.2331,
"step": 272
},
{
"epoch": 0.13794845881758463,
"grad_norm": 0.9071863293647766,
"learning_rate": 4.925781617706874e-05,
"loss": 0.486,
"step": 273
},
{
"epoch": 0.13845376452753916,
"grad_norm": 0.4862212538719177,
"learning_rate": 4.925018365432682e-05,
"loss": 0.234,
"step": 274
},
{
"epoch": 0.13895907023749368,
"grad_norm": 0.2530370056629181,
"learning_rate": 4.92425126835967e-05,
"loss": 0.1656,
"step": 275
},
{
"epoch": 0.1394643759474482,
"grad_norm": 0.8576244711875916,
"learning_rate": 4.923480327704051e-05,
"loss": 0.6208,
"step": 276
},
{
"epoch": 0.13996968165740273,
"grad_norm": 0.6219656467437744,
"learning_rate": 4.922705544688127e-05,
"loss": 0.3705,
"step": 277
},
{
"epoch": 0.14047498736735725,
"grad_norm": 0.3833385705947876,
"learning_rate": 4.9219269205402976e-05,
"loss": 0.2777,
"step": 278
},
{
"epoch": 0.14098029307731177,
"grad_norm": 0.5600672960281372,
"learning_rate": 4.921144456495048e-05,
"loss": 0.5572,
"step": 279
},
{
"epoch": 0.1414855987872663,
"grad_norm": 0.5854461193084717,
"learning_rate": 4.9203581537929536e-05,
"loss": 0.3957,
"step": 280
},
{
"epoch": 0.14199090449722082,
"grad_norm": 0.6858914494514465,
"learning_rate": 4.919568013680676e-05,
"loss": 0.3994,
"step": 281
},
{
"epoch": 0.14249621020717534,
"grad_norm": 0.8274922370910645,
"learning_rate": 4.918774037410958e-05,
"loss": 0.3523,
"step": 282
},
{
"epoch": 0.14300151591712987,
"grad_norm": 0.423542320728302,
"learning_rate": 4.9179762262426286e-05,
"loss": 0.2194,
"step": 283
},
{
"epoch": 0.1435068216270844,
"grad_norm": 0.7165141105651855,
"learning_rate": 4.917174581440595e-05,
"loss": 0.4029,
"step": 284
},
{
"epoch": 0.14401212733703891,
"grad_norm": 0.900047242641449,
"learning_rate": 4.916369104275842e-05,
"loss": 0.5053,
"step": 285
},
{
"epoch": 0.14451743304699344,
"grad_norm": 0.466221421957016,
"learning_rate": 4.9155597960254316e-05,
"loss": 0.3247,
"step": 286
},
{
"epoch": 0.14502273875694796,
"grad_norm": 0.6081900596618652,
"learning_rate": 4.914746657972499e-05,
"loss": 0.3749,
"step": 287
},
{
"epoch": 0.14552804446690248,
"grad_norm": 0.8832564949989319,
"learning_rate": 4.913929691406253e-05,
"loss": 0.3952,
"step": 288
},
{
"epoch": 0.146033350176857,
"grad_norm": 0.6159337759017944,
"learning_rate": 4.91310889762197e-05,
"loss": 0.4868,
"step": 289
},
{
"epoch": 0.14653865588681153,
"grad_norm": 0.6993011832237244,
"learning_rate": 4.9122842779209954e-05,
"loss": 0.4032,
"step": 290
},
{
"epoch": 0.14704396159676605,
"grad_norm": 0.7520043253898621,
"learning_rate": 4.911455833610742e-05,
"loss": 0.2641,
"step": 291
},
{
"epoch": 0.14754926730672058,
"grad_norm": 0.4298335313796997,
"learning_rate": 4.910623566004684e-05,
"loss": 0.1774,
"step": 292
},
{
"epoch": 0.1480545730166751,
"grad_norm": 0.5507643818855286,
"learning_rate": 4.909787476422358e-05,
"loss": 0.305,
"step": 293
},
{
"epoch": 0.14855987872662962,
"grad_norm": 0.6847674250602722,
"learning_rate": 4.9089475661893625e-05,
"loss": 0.3482,
"step": 294
},
{
"epoch": 0.14906518443658415,
"grad_norm": 0.48942941427230835,
"learning_rate": 4.908103836637351e-05,
"loss": 0.2464,
"step": 295
},
{
"epoch": 0.14957049014653864,
"grad_norm": 0.39289000630378723,
"learning_rate": 4.907256289104032e-05,
"loss": 0.195,
"step": 296
},
{
"epoch": 0.15007579585649317,
"grad_norm": 0.599053144454956,
"learning_rate": 4.906404924933169e-05,
"loss": 0.2337,
"step": 297
},
{
"epoch": 0.1505811015664477,
"grad_norm": 0.6427844762802124,
"learning_rate": 4.9055497454745755e-05,
"loss": 0.3626,
"step": 298
},
{
"epoch": 0.1510864072764022,
"grad_norm": 0.875948429107666,
"learning_rate": 4.904690752084117e-05,
"loss": 0.4159,
"step": 299
},
{
"epoch": 0.15159171298635674,
"grad_norm": 0.7781713604927063,
"learning_rate": 4.9038279461237016e-05,
"loss": 0.3178,
"step": 300
},
{
"epoch": 0.15209701869631126,
"grad_norm": 0.5986030697822571,
"learning_rate": 4.902961328961283e-05,
"loss": 0.5787,
"step": 301
},
{
"epoch": 0.15260232440626578,
"grad_norm": 0.6488214135169983,
"learning_rate": 4.902090901970861e-05,
"loss": 0.3557,
"step": 302
},
{
"epoch": 0.1531076301162203,
"grad_norm": 0.6705836057662964,
"learning_rate": 4.9012166665324714e-05,
"loss": 0.306,
"step": 303
},
{
"epoch": 0.15361293582617483,
"grad_norm": 0.5052811503410339,
"learning_rate": 4.900338624032191e-05,
"loss": 0.2413,
"step": 304
},
{
"epoch": 0.15411824153612935,
"grad_norm": 0.5678831934928894,
"learning_rate": 4.899456775862131e-05,
"loss": 0.3372,
"step": 305
},
{
"epoch": 0.15462354724608388,
"grad_norm": 0.4097347855567932,
"learning_rate": 4.898571123420438e-05,
"loss": 0.2792,
"step": 306
},
{
"epoch": 0.1551288529560384,
"grad_norm": 0.3428628444671631,
"learning_rate": 4.897681668111288e-05,
"loss": 0.2215,
"step": 307
},
{
"epoch": 0.15563415866599292,
"grad_norm": 0.5401535630226135,
"learning_rate": 4.896788411344888e-05,
"loss": 0.2768,
"step": 308
},
{
"epoch": 0.15613946437594745,
"grad_norm": 0.65704745054245,
"learning_rate": 4.895891354537472e-05,
"loss": 0.4244,
"step": 309
},
{
"epoch": 0.15664477008590197,
"grad_norm": 0.8244349360466003,
"learning_rate": 4.894990499111297e-05,
"loss": 0.4721,
"step": 310
},
{
"epoch": 0.1571500757958565,
"grad_norm": 1.0010887384414673,
"learning_rate": 4.894085846494647e-05,
"loss": 0.6647,
"step": 311
},
{
"epoch": 0.15765538150581102,
"grad_norm": 0.5577598214149475,
"learning_rate": 4.893177398121821e-05,
"loss": 0.458,
"step": 312
},
{
"epoch": 0.15816068721576554,
"grad_norm": 0.9697132110595703,
"learning_rate": 4.8922651554331386e-05,
"loss": 0.583,
"step": 313
},
{
"epoch": 0.15866599292572006,
"grad_norm": 0.7918589115142822,
"learning_rate": 4.891349119874936e-05,
"loss": 0.3947,
"step": 314
},
{
"epoch": 0.15917129863567459,
"grad_norm": 0.5606548190116882,
"learning_rate": 4.890429292899562e-05,
"loss": 0.2518,
"step": 315
},
{
"epoch": 0.1596766043456291,
"grad_norm": 0.46425801515579224,
"learning_rate": 4.889505675965376e-05,
"loss": 0.231,
"step": 316
},
{
"epoch": 0.16018191005558363,
"grad_norm": 0.5073766708374023,
"learning_rate": 4.888578270536747e-05,
"loss": 0.404,
"step": 317
},
{
"epoch": 0.16068721576553815,
"grad_norm": 0.8138315677642822,
"learning_rate": 4.88764707808405e-05,
"loss": 0.3759,
"step": 318
},
{
"epoch": 0.16119252147549268,
"grad_norm": 0.6912230849266052,
"learning_rate": 4.886712100083665e-05,
"loss": 0.4989,
"step": 319
},
{
"epoch": 0.1616978271854472,
"grad_norm": 0.43325480818748474,
"learning_rate": 4.8857733380179735e-05,
"loss": 0.3023,
"step": 320
},
{
"epoch": 0.16220313289540172,
"grad_norm": 0.4167453944683075,
"learning_rate": 4.8848307933753566e-05,
"loss": 0.2701,
"step": 321
},
{
"epoch": 0.16270843860535625,
"grad_norm": 0.3875753581523895,
"learning_rate": 4.883884467650193e-05,
"loss": 0.2084,
"step": 322
},
{
"epoch": 0.16321374431531077,
"grad_norm": 0.4990168511867523,
"learning_rate": 4.882934362342854e-05,
"loss": 0.6131,
"step": 323
},
{
"epoch": 0.1637190500252653,
"grad_norm": 0.5788955092430115,
"learning_rate": 4.881980478959707e-05,
"loss": 0.2931,
"step": 324
},
{
"epoch": 0.16422435573521982,
"grad_norm": 0.7873169183731079,
"learning_rate": 4.881022819013106e-05,
"loss": 0.3825,
"step": 325
},
{
"epoch": 0.16472966144517434,
"grad_norm": 0.30970779061317444,
"learning_rate": 4.8800613840213936e-05,
"loss": 0.1879,
"step": 326
},
{
"epoch": 0.16523496715512886,
"grad_norm": 0.43223828077316284,
"learning_rate": 4.8790961755089006e-05,
"loss": 0.194,
"step": 327
},
{
"epoch": 0.1657402728650834,
"grad_norm": 0.4357849061489105,
"learning_rate": 4.878127195005936e-05,
"loss": 0.2503,
"step": 328
},
{
"epoch": 0.1662455785750379,
"grad_norm": 0.6348159313201904,
"learning_rate": 4.8771544440487916e-05,
"loss": 0.2907,
"step": 329
},
{
"epoch": 0.1667508842849924,
"grad_norm": 0.4712270200252533,
"learning_rate": 4.8761779241797366e-05,
"loss": 0.2412,
"step": 330
},
{
"epoch": 0.16725618999494693,
"grad_norm": 0.7354241609573364,
"learning_rate": 4.875197636947018e-05,
"loss": 0.4536,
"step": 331
},
{
"epoch": 0.16776149570490145,
"grad_norm": 0.3180365264415741,
"learning_rate": 4.874213583904851e-05,
"loss": 0.1503,
"step": 332
},
{
"epoch": 0.16826680141485598,
"grad_norm": 0.9528758525848389,
"learning_rate": 4.873225766613425e-05,
"loss": 0.4599,
"step": 333
},
{
"epoch": 0.1687721071248105,
"grad_norm": 0.5695481300354004,
"learning_rate": 4.872234186638898e-05,
"loss": 0.3238,
"step": 334
},
{
"epoch": 0.16927741283476502,
"grad_norm": 0.9408870935440063,
"learning_rate": 4.871238845553391e-05,
"loss": 0.4594,
"step": 335
},
{
"epoch": 0.16978271854471955,
"grad_norm": 0.5320512652397156,
"learning_rate": 4.87023974493499e-05,
"loss": 0.3218,
"step": 336
},
{
"epoch": 0.17028802425467407,
"grad_norm": 0.6941995024681091,
"learning_rate": 4.8692368863677416e-05,
"loss": 0.4405,
"step": 337
},
{
"epoch": 0.1707933299646286,
"grad_norm": 0.5261353850364685,
"learning_rate": 4.8682302714416494e-05,
"loss": 0.2217,
"step": 338
},
{
"epoch": 0.17129863567458312,
"grad_norm": 0.6723947525024414,
"learning_rate": 4.8672199017526726e-05,
"loss": 0.3233,
"step": 339
},
{
"epoch": 0.17180394138453764,
"grad_norm": 0.499994695186615,
"learning_rate": 4.866205778902725e-05,
"loss": 0.2603,
"step": 340
},
{
"epoch": 0.17230924709449216,
"grad_norm": 0.3436490595340729,
"learning_rate": 4.86518790449967e-05,
"loss": 0.2763,
"step": 341
},
{
"epoch": 0.17281455280444669,
"grad_norm": 0.6315925717353821,
"learning_rate": 4.8641662801573185e-05,
"loss": 0.2521,
"step": 342
},
{
"epoch": 0.1733198585144012,
"grad_norm": 0.5178197026252747,
"learning_rate": 4.863140907495428e-05,
"loss": 0.222,
"step": 343
},
{
"epoch": 0.17382516422435573,
"grad_norm": 0.5388314127922058,
"learning_rate": 4.862111788139697e-05,
"loss": 0.3894,
"step": 344
},
{
"epoch": 0.17433046993431026,
"grad_norm": 0.7556846737861633,
"learning_rate": 4.861078923721767e-05,
"loss": 0.3644,
"step": 345
},
{
"epoch": 0.17483577564426478,
"grad_norm": 0.827375590801239,
"learning_rate": 4.860042315879214e-05,
"loss": 0.5613,
"step": 346
},
{
"epoch": 0.1753410813542193,
"grad_norm": 0.550095796585083,
"learning_rate": 4.8590019662555534e-05,
"loss": 0.3149,
"step": 347
},
{
"epoch": 0.17584638706417383,
"grad_norm": 0.41516733169555664,
"learning_rate": 4.857957876500227e-05,
"loss": 0.254,
"step": 348
},
{
"epoch": 0.17635169277412835,
"grad_norm": 0.6708664298057556,
"learning_rate": 4.8569100482686125e-05,
"loss": 0.7031,
"step": 349
},
{
"epoch": 0.17685699848408287,
"grad_norm": 0.8377763032913208,
"learning_rate": 4.855858483222011e-05,
"loss": 0.4918,
"step": 350
},
{
"epoch": 0.1773623041940374,
"grad_norm": 0.5428167581558228,
"learning_rate": 4.8548031830276496e-05,
"loss": 0.3071,
"step": 351
},
{
"epoch": 0.17786760990399192,
"grad_norm": 0.6892185807228088,
"learning_rate": 4.8537441493586786e-05,
"loss": 0.3797,
"step": 352
},
{
"epoch": 0.17837291561394644,
"grad_norm": 0.5104579925537109,
"learning_rate": 4.852681383894164e-05,
"loss": 0.3025,
"step": 353
},
{
"epoch": 0.17887822132390097,
"grad_norm": 0.42806196212768555,
"learning_rate": 4.8516148883190926e-05,
"loss": 0.2176,
"step": 354
},
{
"epoch": 0.1793835270338555,
"grad_norm": 0.6678165197372437,
"learning_rate": 4.850544664324363e-05,
"loss": 0.3057,
"step": 355
},
{
"epoch": 0.17988883274381,
"grad_norm": 0.33582162857055664,
"learning_rate": 4.849470713606785e-05,
"loss": 0.1869,
"step": 356
},
{
"epoch": 0.18039413845376454,
"grad_norm": 0.8661889433860779,
"learning_rate": 4.8483930378690765e-05,
"loss": 0.4672,
"step": 357
},
{
"epoch": 0.18089944416371906,
"grad_norm": 0.6204270124435425,
"learning_rate": 4.847311638819865e-05,
"loss": 0.2449,
"step": 358
},
{
"epoch": 0.18140474987367358,
"grad_norm": 0.8477338552474976,
"learning_rate": 4.8462265181736756e-05,
"loss": 0.3236,
"step": 359
},
{
"epoch": 0.1819100555836281,
"grad_norm": 0.5757856965065002,
"learning_rate": 4.845137677650937e-05,
"loss": 0.2292,
"step": 360
},
{
"epoch": 0.18241536129358263,
"grad_norm": 0.3746514320373535,
"learning_rate": 4.8440451189779776e-05,
"loss": 0.2001,
"step": 361
},
{
"epoch": 0.18292066700353715,
"grad_norm": 0.5158061981201172,
"learning_rate": 4.8429488438870174e-05,
"loss": 0.4076,
"step": 362
},
{
"epoch": 0.18342597271349168,
"grad_norm": 0.6773928999900818,
"learning_rate": 4.841848854116169e-05,
"loss": 0.3376,
"step": 363
},
{
"epoch": 0.1839312784234462,
"grad_norm": 0.6250408291816711,
"learning_rate": 4.840745151409437e-05,
"loss": 0.3262,
"step": 364
},
{
"epoch": 0.1844365841334007,
"grad_norm": 0.43249404430389404,
"learning_rate": 4.83963773751671e-05,
"loss": 0.2738,
"step": 365
},
{
"epoch": 0.18494188984335522,
"grad_norm": 0.5019465684890747,
"learning_rate": 4.838526614193762e-05,
"loss": 0.4144,
"step": 366
},
{
"epoch": 0.18544719555330974,
"grad_norm": 0.45221251249313354,
"learning_rate": 4.837411783202249e-05,
"loss": 0.2398,
"step": 367
},
{
"epoch": 0.18595250126326426,
"grad_norm": 0.36865469813346863,
"learning_rate": 4.836293246309702e-05,
"loss": 0.222,
"step": 368
},
{
"epoch": 0.1864578069732188,
"grad_norm": 0.43405380845069885,
"learning_rate": 4.8351710052895324e-05,
"loss": 0.2316,
"step": 369
},
{
"epoch": 0.1869631126831733,
"grad_norm": 0.7884693145751953,
"learning_rate": 4.834045061921022e-05,
"loss": 0.3766,
"step": 370
},
{
"epoch": 0.18746841839312783,
"grad_norm": 0.8879896402359009,
"learning_rate": 4.83291541798932e-05,
"loss": 0.3775,
"step": 371
},
{
"epoch": 0.18797372410308236,
"grad_norm": 0.9513370394706726,
"learning_rate": 4.831782075285448e-05,
"loss": 0.3324,
"step": 372
},
{
"epoch": 0.18847902981303688,
"grad_norm": 0.4089733064174652,
"learning_rate": 4.830645035606287e-05,
"loss": 0.1853,
"step": 373
},
{
"epoch": 0.1889843355229914,
"grad_norm": 0.5456333756446838,
"learning_rate": 4.829504300754584e-05,
"loss": 0.2975,
"step": 374
},
{
"epoch": 0.18948964123294593,
"grad_norm": 0.5223088264465332,
"learning_rate": 4.82835987253894e-05,
"loss": 0.308,
"step": 375
},
{
"epoch": 0.18999494694290045,
"grad_norm": 0.42592930793762207,
"learning_rate": 4.8272117527738146e-05,
"loss": 0.2699,
"step": 376
},
{
"epoch": 0.19050025265285497,
"grad_norm": 0.30264854431152344,
"learning_rate": 4.8260599432795205e-05,
"loss": 0.1962,
"step": 377
},
{
"epoch": 0.1910055583628095,
"grad_norm": 0.5884383916854858,
"learning_rate": 4.824904445882218e-05,
"loss": 0.2513,
"step": 378
},
{
"epoch": 0.19151086407276402,
"grad_norm": 0.5112461447715759,
"learning_rate": 4.823745262413917e-05,
"loss": 0.2216,
"step": 379
},
{
"epoch": 0.19201616978271854,
"grad_norm": 0.5951370000839233,
"learning_rate": 4.822582394712471e-05,
"loss": 0.2971,
"step": 380
},
{
"epoch": 0.19252147549267307,
"grad_norm": 0.4729789197444916,
"learning_rate": 4.821415844621572e-05,
"loss": 0.3288,
"step": 381
},
{
"epoch": 0.1930267812026276,
"grad_norm": 0.8325059413909912,
"learning_rate": 4.820245613990755e-05,
"loss": 0.5792,
"step": 382
},
{
"epoch": 0.1935320869125821,
"grad_norm": 0.6559721827507019,
"learning_rate": 4.819071704675386e-05,
"loss": 0.3209,
"step": 383
},
{
"epoch": 0.19403739262253664,
"grad_norm": 0.5043466091156006,
"learning_rate": 4.817894118536667e-05,
"loss": 0.2815,
"step": 384
},
{
"epoch": 0.19454269833249116,
"grad_norm": 0.6874819397926331,
"learning_rate": 4.8167128574416266e-05,
"loss": 0.4657,
"step": 385
},
{
"epoch": 0.19504800404244568,
"grad_norm": 0.7303402423858643,
"learning_rate": 4.8155279232631214e-05,
"loss": 0.4223,
"step": 386
},
{
"epoch": 0.1955533097524002,
"grad_norm": 0.5754918456077576,
"learning_rate": 4.8143393178798326e-05,
"loss": 0.4584,
"step": 387
},
{
"epoch": 0.19605861546235473,
"grad_norm": 0.613545298576355,
"learning_rate": 4.81314704317626e-05,
"loss": 0.2793,
"step": 388
},
{
"epoch": 0.19656392117230925,
"grad_norm": 0.49800634384155273,
"learning_rate": 4.811951101042722e-05,
"loss": 0.3728,
"step": 389
},
{
"epoch": 0.19706922688226378,
"grad_norm": 0.3457117974758148,
"learning_rate": 4.8107514933753507e-05,
"loss": 0.3332,
"step": 390
},
{
"epoch": 0.1975745325922183,
"grad_norm": 0.4752323031425476,
"learning_rate": 4.8095482220760914e-05,
"loss": 0.249,
"step": 391
},
{
"epoch": 0.19807983830217282,
"grad_norm": 0.6126275062561035,
"learning_rate": 4.808341289052697e-05,
"loss": 0.334,
"step": 392
},
{
"epoch": 0.19858514401212735,
"grad_norm": 0.7043460011482239,
"learning_rate": 4.807130696218726e-05,
"loss": 0.3765,
"step": 393
},
{
"epoch": 0.19909044972208187,
"grad_norm": 0.410325825214386,
"learning_rate": 4.805916445493538e-05,
"loss": 0.214,
"step": 394
},
{
"epoch": 0.1995957554320364,
"grad_norm": 0.794836699962616,
"learning_rate": 4.8046985388022955e-05,
"loss": 0.3224,
"step": 395
},
{
"epoch": 0.20010106114199092,
"grad_norm": 0.3947627544403076,
"learning_rate": 4.8034769780759556e-05,
"loss": 0.2713,
"step": 396
},
{
"epoch": 0.20060636685194544,
"grad_norm": 0.6008957028388977,
"learning_rate": 4.8022517652512674e-05,
"loss": 0.3625,
"step": 397
},
{
"epoch": 0.20111167256189996,
"grad_norm": 0.796962559223175,
"learning_rate": 4.801022902270773e-05,
"loss": 0.4219,
"step": 398
},
{
"epoch": 0.20161697827185449,
"grad_norm": 0.5383734107017517,
"learning_rate": 4.799790391082799e-05,
"loss": 0.2021,
"step": 399
},
{
"epoch": 0.20212228398180898,
"grad_norm": 0.28900346159935,
"learning_rate": 4.7985542336414576e-05,
"loss": 0.1807,
"step": 400
},
{
"epoch": 0.2026275896917635,
"grad_norm": 0.883991539478302,
"learning_rate": 4.797314431906642e-05,
"loss": 0.4265,
"step": 401
},
{
"epoch": 0.20313289540171803,
"grad_norm": 0.44878625869750977,
"learning_rate": 4.796070987844024e-05,
"loss": 0.2376,
"step": 402
},
{
"epoch": 0.20363820111167255,
"grad_norm": 0.3813835680484772,
"learning_rate": 4.7948239034250485e-05,
"loss": 0.1876,
"step": 403
},
{
"epoch": 0.20414350682162707,
"grad_norm": 0.396606981754303,
"learning_rate": 4.793573180626934e-05,
"loss": 0.2111,
"step": 404
},
{
"epoch": 0.2046488125315816,
"grad_norm": 0.6839037537574768,
"learning_rate": 4.792318821432664e-05,
"loss": 0.4695,
"step": 405
},
{
"epoch": 0.20515411824153612,
"grad_norm": 0.39374062418937683,
"learning_rate": 4.791060827830993e-05,
"loss": 0.2562,
"step": 406
},
{
"epoch": 0.20565942395149064,
"grad_norm": 0.5716267824172974,
"learning_rate": 4.789799201816433e-05,
"loss": 0.3251,
"step": 407
},
{
"epoch": 0.20616472966144517,
"grad_norm": 0.4453814625740051,
"learning_rate": 4.788533945389257e-05,
"loss": 0.2136,
"step": 408
},
{
"epoch": 0.2066700353713997,
"grad_norm": 0.6336759924888611,
"learning_rate": 4.787265060555495e-05,
"loss": 0.3443,
"step": 409
},
{
"epoch": 0.20717534108135421,
"grad_norm": 0.7093701362609863,
"learning_rate": 4.785992549326925e-05,
"loss": 0.4021,
"step": 410
},
{
"epoch": 0.20768064679130874,
"grad_norm": 0.38103368878364563,
"learning_rate": 4.784716413721082e-05,
"loss": 0.2219,
"step": 411
},
{
"epoch": 0.20818595250126326,
"grad_norm": 0.49759477376937866,
"learning_rate": 4.783436655761241e-05,
"loss": 0.2649,
"step": 412
},
{
"epoch": 0.20869125821121778,
"grad_norm": 0.7099438309669495,
"learning_rate": 4.7821532774764235e-05,
"loss": 0.2655,
"step": 413
},
{
"epoch": 0.2091965639211723,
"grad_norm": 0.6431573629379272,
"learning_rate": 4.780866280901389e-05,
"loss": 0.2983,
"step": 414
},
{
"epoch": 0.20970186963112683,
"grad_norm": 0.6158983111381531,
"learning_rate": 4.779575668076636e-05,
"loss": 0.2617,
"step": 415
},
{
"epoch": 0.21020717534108135,
"grad_norm": 0.29485249519348145,
"learning_rate": 4.778281441048392e-05,
"loss": 0.1779,
"step": 416
},
{
"epoch": 0.21071248105103588,
"grad_norm": 0.45584672689437866,
"learning_rate": 4.7769836018686214e-05,
"loss": 0.2727,
"step": 417
},
{
"epoch": 0.2112177867609904,
"grad_norm": 0.5246008634567261,
"learning_rate": 4.77568215259501e-05,
"loss": 0.2185,
"step": 418
},
{
"epoch": 0.21172309247094492,
"grad_norm": 0.9197020530700684,
"learning_rate": 4.7743770952909694e-05,
"loss": 0.3985,
"step": 419
},
{
"epoch": 0.21222839818089945,
"grad_norm": 0.6227766871452332,
"learning_rate": 4.773068432025632e-05,
"loss": 0.3579,
"step": 420
},
{
"epoch": 0.21273370389085397,
"grad_norm": 0.42557641863822937,
"learning_rate": 4.771756164873847e-05,
"loss": 0.355,
"step": 421
},
{
"epoch": 0.2132390096008085,
"grad_norm": 0.2553727626800537,
"learning_rate": 4.770440295916176e-05,
"loss": 0.1222,
"step": 422
},
{
"epoch": 0.21374431531076302,
"grad_norm": 0.6675715446472168,
"learning_rate": 4.769120827238893e-05,
"loss": 0.31,
"step": 423
},
{
"epoch": 0.21424962102071754,
"grad_norm": 0.7006489634513855,
"learning_rate": 4.76779776093398e-05,
"loss": 0.5355,
"step": 424
},
{
"epoch": 0.21475492673067206,
"grad_norm": 0.49595072865486145,
"learning_rate": 4.766471099099119e-05,
"loss": 0.2717,
"step": 425
},
{
"epoch": 0.2152602324406266,
"grad_norm": 0.7701184153556824,
"learning_rate": 4.765140843837697e-05,
"loss": 0.5776,
"step": 426
},
{
"epoch": 0.2157655381505811,
"grad_norm": 0.5482222437858582,
"learning_rate": 4.763806997258798e-05,
"loss": 0.3752,
"step": 427
},
{
"epoch": 0.21627084386053563,
"grad_norm": 0.7224484086036682,
"learning_rate": 4.762469561477195e-05,
"loss": 0.4275,
"step": 428
},
{
"epoch": 0.21677614957049016,
"grad_norm": 0.5047935843467712,
"learning_rate": 4.7611285386133584e-05,
"loss": 0.3352,
"step": 429
},
{
"epoch": 0.21728145528044468,
"grad_norm": 0.7135103344917297,
"learning_rate": 4.7597839307934397e-05,
"loss": 0.4041,
"step": 430
},
{
"epoch": 0.2177867609903992,
"grad_norm": 0.622363805770874,
"learning_rate": 4.7584357401492784e-05,
"loss": 0.4279,
"step": 431
},
{
"epoch": 0.21829206670035373,
"grad_norm": 0.616588830947876,
"learning_rate": 4.7570839688183934e-05,
"loss": 0.5381,
"step": 432
},
{
"epoch": 0.21879737241030825,
"grad_norm": 0.4345376789569855,
"learning_rate": 4.7557286189439794e-05,
"loss": 0.2053,
"step": 433
},
{
"epoch": 0.21930267812026275,
"grad_norm": 0.5042337775230408,
"learning_rate": 4.754369692674906e-05,
"loss": 0.3279,
"step": 434
},
{
"epoch": 0.21980798383021727,
"grad_norm": 0.6724066734313965,
"learning_rate": 4.7530071921657134e-05,
"loss": 0.4296,
"step": 435
},
{
"epoch": 0.2203132895401718,
"grad_norm": 0.607983410358429,
"learning_rate": 4.751641119576607e-05,
"loss": 0.3913,
"step": 436
},
{
"epoch": 0.22081859525012631,
"grad_norm": 0.7748120427131653,
"learning_rate": 4.750271477073458e-05,
"loss": 0.2699,
"step": 437
},
{
"epoch": 0.22132390096008084,
"grad_norm": 0.9839958548545837,
"learning_rate": 4.748898266827795e-05,
"loss": 0.365,
"step": 438
},
{
"epoch": 0.22182920667003536,
"grad_norm": 0.8189884424209595,
"learning_rate": 4.747521491016805e-05,
"loss": 0.4648,
"step": 439
},
{
"epoch": 0.22233451237998988,
"grad_norm": 0.43248504400253296,
"learning_rate": 4.7461411518233286e-05,
"loss": 0.2829,
"step": 440
},
{
"epoch": 0.2228398180899444,
"grad_norm": 0.6537896990776062,
"learning_rate": 4.744757251435853e-05,
"loss": 0.4676,
"step": 441
},
{
"epoch": 0.22334512379989893,
"grad_norm": 0.733722448348999,
"learning_rate": 4.7433697920485156e-05,
"loss": 0.3722,
"step": 442
},
{
"epoch": 0.22385042950985345,
"grad_norm": 0.38846614956855774,
"learning_rate": 4.741978775861092e-05,
"loss": 0.1931,
"step": 443
},
{
"epoch": 0.22435573521980798,
"grad_norm": 0.5748143196105957,
"learning_rate": 4.7405842050790014e-05,
"loss": 0.3245,
"step": 444
},
{
"epoch": 0.2248610409297625,
"grad_norm": 0.674279510974884,
"learning_rate": 4.739186081913297e-05,
"loss": 0.2701,
"step": 445
},
{
"epoch": 0.22536634663971702,
"grad_norm": 0.4921652376651764,
"learning_rate": 4.737784408580663e-05,
"loss": 0.3256,
"step": 446
},
{
"epoch": 0.22587165234967155,
"grad_norm": 0.4046092629432678,
"learning_rate": 4.736379187303414e-05,
"loss": 0.2572,
"step": 447
},
{
"epoch": 0.22637695805962607,
"grad_norm": 0.935234785079956,
"learning_rate": 4.734970420309488e-05,
"loss": 0.5253,
"step": 448
},
{
"epoch": 0.2268822637695806,
"grad_norm": 0.44825106859207153,
"learning_rate": 4.733558109832447e-05,
"loss": 0.254,
"step": 449
},
{
"epoch": 0.22738756947953512,
"grad_norm": 0.45430219173431396,
"learning_rate": 4.732142258111468e-05,
"loss": 0.2459,
"step": 450
},
{
"epoch": 0.22789287518948964,
"grad_norm": 0.44165119528770447,
"learning_rate": 4.730722867391346e-05,
"loss": 0.23,
"step": 451
},
{
"epoch": 0.22839818089944416,
"grad_norm": 0.4309244155883789,
"learning_rate": 4.729299939922485e-05,
"loss": 0.2213,
"step": 452
},
{
"epoch": 0.2289034866093987,
"grad_norm": 0.6634194254875183,
"learning_rate": 4.727873477960896e-05,
"loss": 0.3263,
"step": 453
},
{
"epoch": 0.2294087923193532,
"grad_norm": 0.2821512222290039,
"learning_rate": 4.726443483768195e-05,
"loss": 0.1682,
"step": 454
},
{
"epoch": 0.22991409802930773,
"grad_norm": 0.7871589064598083,
"learning_rate": 4.7250099596115985e-05,
"loss": 0.4115,
"step": 455
},
{
"epoch": 0.23041940373926226,
"grad_norm": 0.3980810344219208,
"learning_rate": 4.723572907763917e-05,
"loss": 0.2222,
"step": 456
},
{
"epoch": 0.23092470944921678,
"grad_norm": 0.34110260009765625,
"learning_rate": 4.722132330503558e-05,
"loss": 0.2064,
"step": 457
},
{
"epoch": 0.2314300151591713,
"grad_norm": 0.5213835835456848,
"learning_rate": 4.7206882301145164e-05,
"loss": 0.3346,
"step": 458
},
{
"epoch": 0.23193532086912583,
"grad_norm": 0.6506749987602234,
"learning_rate": 4.719240608886372e-05,
"loss": 0.4456,
"step": 459
},
{
"epoch": 0.23244062657908035,
"grad_norm": 0.7075578570365906,
"learning_rate": 4.7177894691142874e-05,
"loss": 0.3906,
"step": 460
},
{
"epoch": 0.23294593228903487,
"grad_norm": 0.6361711025238037,
"learning_rate": 4.7163348130990055e-05,
"loss": 0.497,
"step": 461
},
{
"epoch": 0.2334512379989894,
"grad_norm": 0.4896164536476135,
"learning_rate": 4.714876643146842e-05,
"loss": 0.2448,
"step": 462
},
{
"epoch": 0.23395654370894392,
"grad_norm": 0.36150890588760376,
"learning_rate": 4.7134149615696846e-05,
"loss": 0.2344,
"step": 463
},
{
"epoch": 0.23446184941889844,
"grad_norm": 0.5678621530532837,
"learning_rate": 4.711949770684989e-05,
"loss": 0.594,
"step": 464
},
{
"epoch": 0.23496715512885297,
"grad_norm": 0.42508170008659363,
"learning_rate": 4.7104810728157745e-05,
"loss": 0.2382,
"step": 465
},
{
"epoch": 0.2354724608388075,
"grad_norm": 0.6501306295394897,
"learning_rate": 4.709008870290619e-05,
"loss": 0.3541,
"step": 466
},
{
"epoch": 0.235977766548762,
"grad_norm": 0.6006211638450623,
"learning_rate": 4.7075331654436606e-05,
"loss": 0.4746,
"step": 467
},
{
"epoch": 0.23648307225871654,
"grad_norm": 0.349697083234787,
"learning_rate": 4.706053960614586e-05,
"loss": 0.1804,
"step": 468
},
{
"epoch": 0.23698837796867103,
"grad_norm": 0.4210861325263977,
"learning_rate": 4.7045712581486345e-05,
"loss": 0.3136,
"step": 469
},
{
"epoch": 0.23749368367862556,
"grad_norm": 0.42396318912506104,
"learning_rate": 4.703085060396589e-05,
"loss": 0.217,
"step": 470
},
{
"epoch": 0.23799898938858008,
"grad_norm": 0.4881645739078522,
"learning_rate": 4.701595369714775e-05,
"loss": 0.3285,
"step": 471
},
{
"epoch": 0.2385042950985346,
"grad_norm": 0.33498990535736084,
"learning_rate": 4.7001021884650545e-05,
"loss": 0.2067,
"step": 472
},
{
"epoch": 0.23900960080848913,
"grad_norm": 0.5688101649284363,
"learning_rate": 4.6986055190148255e-05,
"loss": 0.2357,
"step": 473
},
{
"epoch": 0.23951490651844365,
"grad_norm": 0.4370393753051758,
"learning_rate": 4.697105363737014e-05,
"loss": 0.213,
"step": 474
},
{
"epoch": 0.24002021222839817,
"grad_norm": 0.31808745861053467,
"learning_rate": 4.695601725010077e-05,
"loss": 0.2219,
"step": 475
},
{
"epoch": 0.2405255179383527,
"grad_norm": 0.533358633518219,
"learning_rate": 4.694094605217989e-05,
"loss": 0.2173,
"step": 476
},
{
"epoch": 0.24103082364830722,
"grad_norm": 0.4781337380409241,
"learning_rate": 4.6925840067502475e-05,
"loss": 0.3474,
"step": 477
},
{
"epoch": 0.24153612935826174,
"grad_norm": 0.43365997076034546,
"learning_rate": 4.691069932001864e-05,
"loss": 0.1846,
"step": 478
},
{
"epoch": 0.24204143506821627,
"grad_norm": 0.6059222221374512,
"learning_rate": 4.689552383373361e-05,
"loss": 0.3881,
"step": 479
},
{
"epoch": 0.2425467407781708,
"grad_norm": 0.5931808352470398,
"learning_rate": 4.6880313632707714e-05,
"loss": 0.2737,
"step": 480
},
{
"epoch": 0.2430520464881253,
"grad_norm": 0.4886901378631592,
"learning_rate": 4.6865068741056276e-05,
"loss": 0.2409,
"step": 481
},
{
"epoch": 0.24355735219807984,
"grad_norm": 0.8345271348953247,
"learning_rate": 4.684978918294967e-05,
"loss": 0.3453,
"step": 482
},
{
"epoch": 0.24406265790803436,
"grad_norm": 0.5799844861030579,
"learning_rate": 4.683447498261318e-05,
"loss": 0.403,
"step": 483
},
{
"epoch": 0.24456796361798888,
"grad_norm": 0.5229687690734863,
"learning_rate": 4.6819126164327067e-05,
"loss": 0.3206,
"step": 484
},
{
"epoch": 0.2450732693279434,
"grad_norm": 0.8277320861816406,
"learning_rate": 4.680374275242644e-05,
"loss": 0.5465,
"step": 485
},
{
"epoch": 0.24557857503789793,
"grad_norm": 0.6758053302764893,
"learning_rate": 4.678832477130127e-05,
"loss": 0.3294,
"step": 486
},
{
"epoch": 0.24608388074785245,
"grad_norm": 0.5581978559494019,
"learning_rate": 4.677287224539634e-05,
"loss": 0.2746,
"step": 487
},
{
"epoch": 0.24658918645780697,
"grad_norm": 0.21748855710029602,
"learning_rate": 4.675738519921119e-05,
"loss": 0.1429,
"step": 488
},
{
"epoch": 0.2470944921677615,
"grad_norm": 1.0185459852218628,
"learning_rate": 4.6741863657300114e-05,
"loss": 0.4322,
"step": 489
},
{
"epoch": 0.24759979787771602,
"grad_norm": 0.37968578934669495,
"learning_rate": 4.672630764427206e-05,
"loss": 0.2318,
"step": 490
},
{
"epoch": 0.24810510358767054,
"grad_norm": 0.32278138399124146,
"learning_rate": 4.671071718479067e-05,
"loss": 0.2341,
"step": 491
},
{
"epoch": 0.24861040929762507,
"grad_norm": 0.49912557005882263,
"learning_rate": 4.669509230357418e-05,
"loss": 0.3063,
"step": 492
},
{
"epoch": 0.2491157150075796,
"grad_norm": 0.6750683188438416,
"learning_rate": 4.667943302539539e-05,
"loss": 0.3489,
"step": 493
},
{
"epoch": 0.24962102071753411,
"grad_norm": 0.6013567447662354,
"learning_rate": 4.666373937508166e-05,
"loss": 0.3084,
"step": 494
},
{
"epoch": 0.2501263264274886,
"grad_norm": 0.4183938503265381,
"learning_rate": 4.6648011377514834e-05,
"loss": 0.2177,
"step": 495
},
{
"epoch": 0.25063163213744316,
"grad_norm": 0.5014682412147522,
"learning_rate": 4.6632249057631205e-05,
"loss": 0.189,
"step": 496
},
{
"epoch": 0.25113693784739766,
"grad_norm": 0.7215374708175659,
"learning_rate": 4.661645244042149e-05,
"loss": 0.4073,
"step": 497
},
{
"epoch": 0.2516422435573522,
"grad_norm": 0.5658249855041504,
"learning_rate": 4.660062155093079e-05,
"loss": 0.3558,
"step": 498
},
{
"epoch": 0.2521475492673067,
"grad_norm": 0.5246372818946838,
"learning_rate": 4.6584756414258534e-05,
"loss": 0.2615,
"step": 499
},
{
"epoch": 0.25265285497726125,
"grad_norm": 0.4146447777748108,
"learning_rate": 4.656885705555846e-05,
"loss": 0.1839,
"step": 500
},
{
"epoch": 0.25315816068721575,
"grad_norm": 0.7092597484588623,
"learning_rate": 4.6552923500038544e-05,
"loss": 0.4383,
"step": 501
},
{
"epoch": 0.2536634663971703,
"grad_norm": 0.44563567638397217,
"learning_rate": 4.6536955772961e-05,
"loss": 0.2642,
"step": 502
},
{
"epoch": 0.2541687721071248,
"grad_norm": 0.4488155245780945,
"learning_rate": 4.652095389964222e-05,
"loss": 0.3054,
"step": 503
},
{
"epoch": 0.25467407781707935,
"grad_norm": 0.828281044960022,
"learning_rate": 4.6504917905452704e-05,
"loss": 0.5113,
"step": 504
},
{
"epoch": 0.25517938352703384,
"grad_norm": 1.007767677307129,
"learning_rate": 4.6488847815817096e-05,
"loss": 0.6539,
"step": 505
},
{
"epoch": 0.2556846892369884,
"grad_norm": 0.3789096176624298,
"learning_rate": 4.647274365621407e-05,
"loss": 0.277,
"step": 506
},
{
"epoch": 0.2561899949469429,
"grad_norm": 0.37934887409210205,
"learning_rate": 4.6456605452176306e-05,
"loss": 0.2674,
"step": 507
},
{
"epoch": 0.25669530065689744,
"grad_norm": 0.39194759726524353,
"learning_rate": 4.644043322929049e-05,
"loss": 0.2938,
"step": 508
},
{
"epoch": 0.25720060636685194,
"grad_norm": 0.5336602926254272,
"learning_rate": 4.642422701319723e-05,
"loss": 0.4041,
"step": 509
},
{
"epoch": 0.2577059120768065,
"grad_norm": 0.3528728783130646,
"learning_rate": 4.640798682959101e-05,
"loss": 0.2254,
"step": 510
},
{
"epoch": 0.258211217786761,
"grad_norm": 0.42827707529067993,
"learning_rate": 4.6391712704220215e-05,
"loss": 0.1916,
"step": 511
},
{
"epoch": 0.25871652349671553,
"grad_norm": 0.4275967478752136,
"learning_rate": 4.637540466288699e-05,
"loss": 0.2075,
"step": 512
},
{
"epoch": 0.25922182920667003,
"grad_norm": 0.8210417032241821,
"learning_rate": 4.6359062731447296e-05,
"loss": 0.5352,
"step": 513
},
{
"epoch": 0.2597271349166246,
"grad_norm": 0.4613182544708252,
"learning_rate": 4.63426869358108e-05,
"loss": 0.2759,
"step": 514
},
{
"epoch": 0.2602324406265791,
"grad_norm": 0.3744739890098572,
"learning_rate": 4.632627730194087e-05,
"loss": 0.2091,
"step": 515
},
{
"epoch": 0.2607377463365336,
"grad_norm": 0.43345507979393005,
"learning_rate": 4.630983385585452e-05,
"loss": 0.3114,
"step": 516
},
{
"epoch": 0.2612430520464881,
"grad_norm": 1.5208643674850464,
"learning_rate": 4.6293356623622376e-05,
"loss": 0.5386,
"step": 517
},
{
"epoch": 0.2617483577564427,
"grad_norm": 0.3538508117198944,
"learning_rate": 4.627684563136863e-05,
"loss": 0.2279,
"step": 518
},
{
"epoch": 0.26225366346639717,
"grad_norm": 0.5201990008354187,
"learning_rate": 4.6260300905271e-05,
"loss": 0.2865,
"step": 519
},
{
"epoch": 0.2627589691763517,
"grad_norm": 0.8713021874427795,
"learning_rate": 4.624372247156068e-05,
"loss": 0.5142,
"step": 520
},
{
"epoch": 0.2632642748863062,
"grad_norm": 0.8699602484703064,
"learning_rate": 4.622711035652232e-05,
"loss": 0.6462,
"step": 521
},
{
"epoch": 0.2637695805962607,
"grad_norm": 0.3401387333869934,
"learning_rate": 4.6210464586493963e-05,
"loss": 0.1582,
"step": 522
},
{
"epoch": 0.26427488630621526,
"grad_norm": 0.5219199061393738,
"learning_rate": 4.6193785187867005e-05,
"loss": 0.3941,
"step": 523
},
{
"epoch": 0.26478019201616976,
"grad_norm": 0.4327852129936218,
"learning_rate": 4.617707218708617e-05,
"loss": 0.2176,
"step": 524
},
{
"epoch": 0.2652854977261243,
"grad_norm": 0.4247454106807709,
"learning_rate": 4.6160325610649465e-05,
"loss": 0.1823,
"step": 525
},
{
"epoch": 0.2657908034360788,
"grad_norm": 0.8779489398002625,
"learning_rate": 4.6143545485108094e-05,
"loss": 0.5002,
"step": 526
},
{
"epoch": 0.26629610914603336,
"grad_norm": 0.7940444946289062,
"learning_rate": 4.6126731837066505e-05,
"loss": 0.334,
"step": 527
},
{
"epoch": 0.26680141485598785,
"grad_norm": 0.7262102365493774,
"learning_rate": 4.610988469318225e-05,
"loss": 0.3101,
"step": 528
},
{
"epoch": 0.2673067205659424,
"grad_norm": 0.36386728286743164,
"learning_rate": 4.6093004080166e-05,
"loss": 0.2009,
"step": 529
},
{
"epoch": 0.2678120262758969,
"grad_norm": 0.31724250316619873,
"learning_rate": 4.607609002478152e-05,
"loss": 0.1564,
"step": 530
},
{
"epoch": 0.26831733198585145,
"grad_norm": 0.9002480506896973,
"learning_rate": 4.605914255384554e-05,
"loss": 0.3996,
"step": 531
},
{
"epoch": 0.26882263769580594,
"grad_norm": 0.28965526819229126,
"learning_rate": 4.604216169422783e-05,
"loss": 0.1917,
"step": 532
},
{
"epoch": 0.2693279434057605,
"grad_norm": 0.3934868574142456,
"learning_rate": 4.6025147472851036e-05,
"loss": 0.2175,
"step": 533
},
{
"epoch": 0.269833249115715,
"grad_norm": 1.457297682762146,
"learning_rate": 4.600809991669076e-05,
"loss": 0.5182,
"step": 534
},
{
"epoch": 0.27033855482566954,
"grad_norm": 0.6048038601875305,
"learning_rate": 4.599101905277541e-05,
"loss": 0.3971,
"step": 535
},
{
"epoch": 0.27084386053562404,
"grad_norm": 0.4743158519268036,
"learning_rate": 4.597390490818623e-05,
"loss": 0.5295,
"step": 536
},
{
"epoch": 0.2713491662455786,
"grad_norm": 0.7490699291229248,
"learning_rate": 4.59567575100572e-05,
"loss": 0.3392,
"step": 537
},
{
"epoch": 0.2718544719555331,
"grad_norm": 0.66379314661026,
"learning_rate": 4.5939576885575065e-05,
"loss": 0.2745,
"step": 538
},
{
"epoch": 0.27235977766548763,
"grad_norm": 0.3140435218811035,
"learning_rate": 4.59223630619792e-05,
"loss": 0.2245,
"step": 539
},
{
"epoch": 0.27286508337544213,
"grad_norm": 0.6745867133140564,
"learning_rate": 4.590511606656165e-05,
"loss": 0.3131,
"step": 540
},
{
"epoch": 0.2733703890853967,
"grad_norm": 0.7853817939758301,
"learning_rate": 4.588783592666704e-05,
"loss": 0.5628,
"step": 541
},
{
"epoch": 0.2738756947953512,
"grad_norm": 0.6165689826011658,
"learning_rate": 4.5870522669692574e-05,
"loss": 0.2666,
"step": 542
},
{
"epoch": 0.2743810005053057,
"grad_norm": 0.5359889268875122,
"learning_rate": 4.58531763230879e-05,
"loss": 0.3545,
"step": 543
},
{
"epoch": 0.2748863062152602,
"grad_norm": 0.6092820763587952,
"learning_rate": 4.5835796914355195e-05,
"loss": 0.3448,
"step": 544
},
{
"epoch": 0.2753916119252148,
"grad_norm": 0.6095251441001892,
"learning_rate": 4.5818384471049006e-05,
"loss": 0.2423,
"step": 545
},
{
"epoch": 0.27589691763516927,
"grad_norm": 0.5478048920631409,
"learning_rate": 4.5800939020776286e-05,
"loss": 0.3708,
"step": 546
},
{
"epoch": 0.2764022233451238,
"grad_norm": 0.3708425462245941,
"learning_rate": 4.5783460591196305e-05,
"loss": 0.228,
"step": 547
},
{
"epoch": 0.2769075290550783,
"grad_norm": 0.7903876304626465,
"learning_rate": 4.576594921002062e-05,
"loss": 0.3286,
"step": 548
},
{
"epoch": 0.27741283476503287,
"grad_norm": 0.40874460339546204,
"learning_rate": 4.5748404905013045e-05,
"loss": 0.3423,
"step": 549
},
{
"epoch": 0.27791814047498736,
"grad_norm": 0.2980653941631317,
"learning_rate": 4.5730827703989574e-05,
"loss": 0.2125,
"step": 550
},
{
"epoch": 0.2784234461849419,
"grad_norm": 0.6296504139900208,
"learning_rate": 4.5713217634818384e-05,
"loss": 0.4838,
"step": 551
},
{
"epoch": 0.2789287518948964,
"grad_norm": 0.3194442391395569,
"learning_rate": 4.569557472541973e-05,
"loss": 0.2049,
"step": 552
},
{
"epoch": 0.27943405760485096,
"grad_norm": 0.4441561698913574,
"learning_rate": 4.5677899003765965e-05,
"loss": 0.2307,
"step": 553
},
{
"epoch": 0.27993936331480546,
"grad_norm": 0.6459462642669678,
"learning_rate": 4.566019049788145e-05,
"loss": 0.4587,
"step": 554
},
{
"epoch": 0.28044466902476,
"grad_norm": 0.5939819812774658,
"learning_rate": 4.564244923584251e-05,
"loss": 0.3031,
"step": 555
},
{
"epoch": 0.2809499747347145,
"grad_norm": 0.5529703497886658,
"learning_rate": 4.562467524577744e-05,
"loss": 0.2677,
"step": 556
},
{
"epoch": 0.281455280444669,
"grad_norm": 0.8261821269989014,
"learning_rate": 4.56068685558664e-05,
"loss": 0.3718,
"step": 557
},
{
"epoch": 0.28196058615462355,
"grad_norm": 0.42143872380256653,
"learning_rate": 4.558902919434139e-05,
"loss": 0.3227,
"step": 558
},
{
"epoch": 0.28246589186457804,
"grad_norm": 0.5339514017105103,
"learning_rate": 4.557115718948622e-05,
"loss": 0.2646,
"step": 559
},
{
"epoch": 0.2829711975745326,
"grad_norm": 0.4582768380641937,
"learning_rate": 4.5553252569636466e-05,
"loss": 0.252,
"step": 560
},
{
"epoch": 0.2834765032844871,
"grad_norm": 0.4415362477302551,
"learning_rate": 4.5535315363179395e-05,
"loss": 0.2225,
"step": 561
},
{
"epoch": 0.28398180899444164,
"grad_norm": 0.6997858881950378,
"learning_rate": 4.551734559855394e-05,
"loss": 0.4024,
"step": 562
},
{
"epoch": 0.28448711470439614,
"grad_norm": 0.34175369143486023,
"learning_rate": 4.549934330425067e-05,
"loss": 0.2033,
"step": 563
},
{
"epoch": 0.2849924204143507,
"grad_norm": 0.589993953704834,
"learning_rate": 4.548130850881171e-05,
"loss": 0.3179,
"step": 564
},
{
"epoch": 0.2854977261243052,
"grad_norm": 0.4499538242816925,
"learning_rate": 4.546324124083073e-05,
"loss": 0.3204,
"step": 565
},
{
"epoch": 0.28600303183425974,
"grad_norm": 0.33767470717430115,
"learning_rate": 4.544514152895289e-05,
"loss": 0.1882,
"step": 566
},
{
"epoch": 0.28650833754421423,
"grad_norm": 0.2902340292930603,
"learning_rate": 4.542700940187476e-05,
"loss": 0.1537,
"step": 567
},
{
"epoch": 0.2870136432541688,
"grad_norm": 0.9272497892379761,
"learning_rate": 4.5408844888344315e-05,
"loss": 0.6278,
"step": 568
},
{
"epoch": 0.2875189489641233,
"grad_norm": 0.4656500816345215,
"learning_rate": 4.5390648017160904e-05,
"loss": 0.2201,
"step": 569
},
{
"epoch": 0.28802425467407783,
"grad_norm": 0.6928651928901672,
"learning_rate": 4.5372418817175146e-05,
"loss": 0.3212,
"step": 570
},
{
"epoch": 0.2885295603840323,
"grad_norm": 0.40516453981399536,
"learning_rate": 4.535415731728892e-05,
"loss": 0.2295,
"step": 571
},
{
"epoch": 0.2890348660939869,
"grad_norm": 0.8511061072349548,
"learning_rate": 4.533586354645534e-05,
"loss": 0.6501,
"step": 572
},
{
"epoch": 0.28954017180394137,
"grad_norm": 0.3017459809780121,
"learning_rate": 4.531753753367867e-05,
"loss": 0.1855,
"step": 573
},
{
"epoch": 0.2900454775138959,
"grad_norm": 0.6671580076217651,
"learning_rate": 4.5299179308014266e-05,
"loss": 0.5083,
"step": 574
},
{
"epoch": 0.2905507832238504,
"grad_norm": 0.323030561208725,
"learning_rate": 4.52807888985686e-05,
"loss": 0.228,
"step": 575
},
{
"epoch": 0.29105608893380497,
"grad_norm": 0.5347647666931152,
"learning_rate": 4.526236633449916e-05,
"loss": 0.2377,
"step": 576
},
{
"epoch": 0.29156139464375946,
"grad_norm": 0.4077557325363159,
"learning_rate": 4.524391164501439e-05,
"loss": 0.2062,
"step": 577
},
{
"epoch": 0.292066700353714,
"grad_norm": 0.2940428853034973,
"learning_rate": 4.522542485937369e-05,
"loss": 0.2001,
"step": 578
},
{
"epoch": 0.2925720060636685,
"grad_norm": 0.48638758063316345,
"learning_rate": 4.520690600688734e-05,
"loss": 0.4731,
"step": 579
},
{
"epoch": 0.29307731177362306,
"grad_norm": 0.6385740637779236,
"learning_rate": 4.5188355116916465e-05,
"loss": 0.3074,
"step": 580
},
{
"epoch": 0.29358261748357756,
"grad_norm": 0.6961050033569336,
"learning_rate": 4.516977221887299e-05,
"loss": 0.4279,
"step": 581
},
{
"epoch": 0.2940879231935321,
"grad_norm": 0.3082885444164276,
"learning_rate": 4.515115734221956e-05,
"loss": 0.1595,
"step": 582
},
{
"epoch": 0.2945932289034866,
"grad_norm": 0.27871477603912354,
"learning_rate": 4.5132510516469553e-05,
"loss": 0.3318,
"step": 583
},
{
"epoch": 0.29509853461344115,
"grad_norm": 0.464947909116745,
"learning_rate": 4.5113831771187e-05,
"loss": 0.2083,
"step": 584
},
{
"epoch": 0.29560384032339565,
"grad_norm": 0.6374766230583191,
"learning_rate": 4.509512113598652e-05,
"loss": 0.3364,
"step": 585
},
{
"epoch": 0.2961091460333502,
"grad_norm": 0.6303216814994812,
"learning_rate": 4.507637864053329e-05,
"loss": 0.4063,
"step": 586
},
{
"epoch": 0.2966144517433047,
"grad_norm": 0.7377685904502869,
"learning_rate": 4.5057604314543033e-05,
"loss": 0.3406,
"step": 587
},
{
"epoch": 0.29711975745325925,
"grad_norm": 0.6760829091072083,
"learning_rate": 4.503879818778191e-05,
"loss": 0.3499,
"step": 588
},
{
"epoch": 0.29762506316321374,
"grad_norm": 0.8320351839065552,
"learning_rate": 4.501996029006651e-05,
"loss": 0.3893,
"step": 589
},
{
"epoch": 0.2981303688731683,
"grad_norm": 0.657802402973175,
"learning_rate": 4.500109065126379e-05,
"loss": 0.4104,
"step": 590
},
{
"epoch": 0.2986356745831228,
"grad_norm": 0.4237852394580841,
"learning_rate": 4.4982189301291037e-05,
"loss": 0.2241,
"step": 591
},
{
"epoch": 0.2991409802930773,
"grad_norm": 0.4519661068916321,
"learning_rate": 4.496325627011581e-05,
"loss": 0.2499,
"step": 592
},
{
"epoch": 0.29964628600303184,
"grad_norm": 0.9270481467247009,
"learning_rate": 4.49442915877559e-05,
"loss": 0.3713,
"step": 593
},
{
"epoch": 0.30015159171298633,
"grad_norm": 1.016149640083313,
"learning_rate": 4.492529528427929e-05,
"loss": 0.8942,
"step": 594
},
{
"epoch": 0.3006568974229409,
"grad_norm": 0.617546558380127,
"learning_rate": 4.490626738980408e-05,
"loss": 0.361,
"step": 595
},
{
"epoch": 0.3011622031328954,
"grad_norm": 0.873954176902771,
"learning_rate": 4.488720793449847e-05,
"loss": 0.3791,
"step": 596
},
{
"epoch": 0.30166750884284993,
"grad_norm": 0.4339880347251892,
"learning_rate": 4.486811694858069e-05,
"loss": 0.197,
"step": 597
},
{
"epoch": 0.3021728145528044,
"grad_norm": 0.8034185171127319,
"learning_rate": 4.484899446231896e-05,
"loss": 0.3023,
"step": 598
},
{
"epoch": 0.302678120262759,
"grad_norm": 0.6054482460021973,
"learning_rate": 4.4829840506031454e-05,
"loss": 0.2276,
"step": 599
},
{
"epoch": 0.30318342597271347,
"grad_norm": 0.37860724329948425,
"learning_rate": 4.481065511008623e-05,
"loss": 0.2614,
"step": 600
},
{
"epoch": 0.303688731682668,
"grad_norm": 0.6753828525543213,
"learning_rate": 4.47914383049012e-05,
"loss": 0.4661,
"step": 601
},
{
"epoch": 0.3041940373926225,
"grad_norm": 0.8002330660820007,
"learning_rate": 4.477219012094407e-05,
"loss": 0.7009,
"step": 602
},
{
"epoch": 0.30469934310257707,
"grad_norm": 0.45732784271240234,
"learning_rate": 4.475291058873231e-05,
"loss": 0.2619,
"step": 603
},
{
"epoch": 0.30520464881253156,
"grad_norm": 0.6500813364982605,
"learning_rate": 4.473359973883305e-05,
"loss": 0.6237,
"step": 604
},
{
"epoch": 0.3057099545224861,
"grad_norm": 0.4476456642150879,
"learning_rate": 4.471425760186313e-05,
"loss": 0.2877,
"step": 605
},
{
"epoch": 0.3062152602324406,
"grad_norm": 0.2967967689037323,
"learning_rate": 4.4694884208488955e-05,
"loss": 0.2686,
"step": 606
},
{
"epoch": 0.30672056594239516,
"grad_norm": 0.320054829120636,
"learning_rate": 4.4675479589426496e-05,
"loss": 0.1553,
"step": 607
},
{
"epoch": 0.30722587165234966,
"grad_norm": 0.7721436619758606,
"learning_rate": 4.465604377544124e-05,
"loss": 0.3458,
"step": 608
},
{
"epoch": 0.3077311773623042,
"grad_norm": 0.25658470392227173,
"learning_rate": 4.463657679734813e-05,
"loss": 0.1976,
"step": 609
},
{
"epoch": 0.3082364830722587,
"grad_norm": 0.5959826707839966,
"learning_rate": 4.4617078686011506e-05,
"loss": 0.3751,
"step": 610
},
{
"epoch": 0.30874178878221326,
"grad_norm": 0.7740558981895447,
"learning_rate": 4.459754947234508e-05,
"loss": 0.5662,
"step": 611
},
{
"epoch": 0.30924709449216775,
"grad_norm": 0.4466153085231781,
"learning_rate": 4.4577989187311886e-05,
"loss": 0.2398,
"step": 612
},
{
"epoch": 0.3097524002021223,
"grad_norm": 0.40263572335243225,
"learning_rate": 4.4558397861924195e-05,
"loss": 0.2314,
"step": 613
},
{
"epoch": 0.3102577059120768,
"grad_norm": 0.5410189032554626,
"learning_rate": 4.453877552724352e-05,
"loss": 0.3845,
"step": 614
},
{
"epoch": 0.31076301162203135,
"grad_norm": 0.32013383507728577,
"learning_rate": 4.451912221438051e-05,
"loss": 0.1808,
"step": 615
},
{
"epoch": 0.31126831733198584,
"grad_norm": 0.8941938877105713,
"learning_rate": 4.449943795449494e-05,
"loss": 0.4481,
"step": 616
},
{
"epoch": 0.3117736230419404,
"grad_norm": 0.5344945788383484,
"learning_rate": 4.4479722778795675e-05,
"loss": 0.3771,
"step": 617
},
{
"epoch": 0.3122789287518949,
"grad_norm": 0.5631364583969116,
"learning_rate": 4.4459976718540555e-05,
"loss": 0.3271,
"step": 618
},
{
"epoch": 0.31278423446184944,
"grad_norm": 0.33273768424987793,
"learning_rate": 4.444019980503641e-05,
"loss": 0.1694,
"step": 619
},
{
"epoch": 0.31328954017180394,
"grad_norm": 0.7633798122406006,
"learning_rate": 4.4420392069638994e-05,
"loss": 0.3704,
"step": 620
},
{
"epoch": 0.3137948458817585,
"grad_norm": 0.45242956280708313,
"learning_rate": 4.44005535437529e-05,
"loss": 0.1935,
"step": 621
},
{
"epoch": 0.314300151591713,
"grad_norm": 0.8318801522254944,
"learning_rate": 4.438068425883158e-05,
"loss": 0.3171,
"step": 622
},
{
"epoch": 0.31480545730166754,
"grad_norm": 0.5340676307678223,
"learning_rate": 4.4360784246377205e-05,
"loss": 0.2594,
"step": 623
},
{
"epoch": 0.31531076301162203,
"grad_norm": 0.9593076705932617,
"learning_rate": 4.434085353794071e-05,
"loss": 0.813,
"step": 624
},
{
"epoch": 0.3158160687215765,
"grad_norm": 0.7513559460639954,
"learning_rate": 4.432089216512166e-05,
"loss": 0.3647,
"step": 625
},
{
"epoch": 0.3163213744315311,
"grad_norm": 0.5096583366394043,
"learning_rate": 4.430090015956826e-05,
"loss": 0.2323,
"step": 626
},
{
"epoch": 0.3168266801414856,
"grad_norm": 0.587281346321106,
"learning_rate": 4.4280877552977276e-05,
"loss": 0.2932,
"step": 627
},
{
"epoch": 0.3173319858514401,
"grad_norm": 0.575646698474884,
"learning_rate": 4.426082437709399e-05,
"loss": 0.532,
"step": 628
},
{
"epoch": 0.3178372915613946,
"grad_norm": 0.38853687047958374,
"learning_rate": 4.4240740663712155e-05,
"loss": 0.1897,
"step": 629
},
{
"epoch": 0.31834259727134917,
"grad_norm": 0.7989168167114258,
"learning_rate": 4.4220626444673926e-05,
"loss": 0.4079,
"step": 630
},
{
"epoch": 0.31884790298130367,
"grad_norm": 0.783780574798584,
"learning_rate": 4.420048175186985e-05,
"loss": 0.3934,
"step": 631
},
{
"epoch": 0.3193532086912582,
"grad_norm": 0.7664374113082886,
"learning_rate": 4.418030661723877e-05,
"loss": 0.2589,
"step": 632
},
{
"epoch": 0.3198585144012127,
"grad_norm": 0.9156386852264404,
"learning_rate": 4.416010107276779e-05,
"loss": 0.3592,
"step": 633
},
{
"epoch": 0.32036382011116726,
"grad_norm": 0.4596467912197113,
"learning_rate": 4.413986515049224e-05,
"loss": 0.3141,
"step": 634
},
{
"epoch": 0.32086912582112176,
"grad_norm": 0.5608931183815002,
"learning_rate": 4.4119598882495606e-05,
"loss": 0.2723,
"step": 635
},
{
"epoch": 0.3213744315310763,
"grad_norm": 0.6327331066131592,
"learning_rate": 4.40993023009095e-05,
"loss": 0.273,
"step": 636
},
{
"epoch": 0.3218797372410308,
"grad_norm": 0.46783167123794556,
"learning_rate": 4.4078975437913575e-05,
"loss": 0.3164,
"step": 637
},
{
"epoch": 0.32238504295098536,
"grad_norm": 0.9565727710723877,
"learning_rate": 4.4058618325735506e-05,
"loss": 0.5771,
"step": 638
},
{
"epoch": 0.32289034866093985,
"grad_norm": 0.8507360816001892,
"learning_rate": 4.403823099665093e-05,
"loss": 0.5183,
"step": 639
},
{
"epoch": 0.3233956543708944,
"grad_norm": 0.8383386135101318,
"learning_rate": 4.401781348298338e-05,
"loss": 0.4447,
"step": 640
},
{
"epoch": 0.3239009600808489,
"grad_norm": 0.4394550621509552,
"learning_rate": 4.3997365817104255e-05,
"loss": 0.2804,
"step": 641
},
{
"epoch": 0.32440626579080345,
"grad_norm": 0.3480008840560913,
"learning_rate": 4.397688803143276e-05,
"loss": 0.2214,
"step": 642
},
{
"epoch": 0.32491157150075795,
"grad_norm": 0.7126933336257935,
"learning_rate": 4.395638015843585e-05,
"loss": 0.5511,
"step": 643
},
{
"epoch": 0.3254168772107125,
"grad_norm": 0.681785523891449,
"learning_rate": 4.393584223062819e-05,
"loss": 0.5339,
"step": 644
},
{
"epoch": 0.325922182920667,
"grad_norm": 0.4552098214626312,
"learning_rate": 4.391527428057207e-05,
"loss": 0.3332,
"step": 645
},
{
"epoch": 0.32642748863062154,
"grad_norm": 0.3405364751815796,
"learning_rate": 4.3894676340877425e-05,
"loss": 0.1796,
"step": 646
},
{
"epoch": 0.32693279434057604,
"grad_norm": 0.3657936155796051,
"learning_rate": 4.387404844420171e-05,
"loss": 0.2149,
"step": 647
},
{
"epoch": 0.3274381000505306,
"grad_norm": 0.7833860516548157,
"learning_rate": 4.385339062324986e-05,
"loss": 0.5498,
"step": 648
},
{
"epoch": 0.3279434057604851,
"grad_norm": 0.7392014265060425,
"learning_rate": 4.3832702910774294e-05,
"loss": 0.424,
"step": 649
},
{
"epoch": 0.32844871147043964,
"grad_norm": 0.37213972210884094,
"learning_rate": 4.381198533957478e-05,
"loss": 0.2245,
"step": 650
},
{
"epoch": 0.32895401718039413,
"grad_norm": 0.7969143986701965,
"learning_rate": 4.379123794249846e-05,
"loss": 0.5756,
"step": 651
},
{
"epoch": 0.3294593228903487,
"grad_norm": 0.46545347571372986,
"learning_rate": 4.377046075243976e-05,
"loss": 0.2736,
"step": 652
},
{
"epoch": 0.3299646286003032,
"grad_norm": 0.7078908681869507,
"learning_rate": 4.3749653802340315e-05,
"loss": 0.454,
"step": 653
},
{
"epoch": 0.33046993431025773,
"grad_norm": 0.43309125304222107,
"learning_rate": 4.372881712518898e-05,
"loss": 0.2874,
"step": 654
},
{
"epoch": 0.3309752400202122,
"grad_norm": 0.8200904726982117,
"learning_rate": 4.3707950754021715e-05,
"loss": 0.4872,
"step": 655
},
{
"epoch": 0.3314805457301668,
"grad_norm": 0.46345481276512146,
"learning_rate": 4.368705472192157e-05,
"loss": 0.3902,
"step": 656
},
{
"epoch": 0.33198585144012127,
"grad_norm": 0.6028680205345154,
"learning_rate": 4.366612906201862e-05,
"loss": 0.3987,
"step": 657
},
{
"epoch": 0.3324911571500758,
"grad_norm": 1.018164873123169,
"learning_rate": 4.364517380748991e-05,
"loss": 0.5398,
"step": 658
},
{
"epoch": 0.3329964628600303,
"grad_norm": 0.5439718961715698,
"learning_rate": 4.362418899155941e-05,
"loss": 0.2593,
"step": 659
},
{
"epoch": 0.3335017685699848,
"grad_norm": 0.5061584115028381,
"learning_rate": 4.360317464749798e-05,
"loss": 0.1962,
"step": 660
},
{
"epoch": 0.33400707427993936,
"grad_norm": 0.5780636072158813,
"learning_rate": 4.358213080862324e-05,
"loss": 0.2549,
"step": 661
},
{
"epoch": 0.33451237998989386,
"grad_norm": 0.6662091016769409,
"learning_rate": 4.356105750829963e-05,
"loss": 0.4128,
"step": 662
},
{
"epoch": 0.3350176856998484,
"grad_norm": 0.473575234413147,
"learning_rate": 4.3539954779938276e-05,
"loss": 0.2602,
"step": 663
},
{
"epoch": 0.3355229914098029,
"grad_norm": 0.6012740135192871,
"learning_rate": 4.351882265699696e-05,
"loss": 0.5341,
"step": 664
},
{
"epoch": 0.33602829711975746,
"grad_norm": 0.48335346579551697,
"learning_rate": 4.3497661172980074e-05,
"loss": 0.2869,
"step": 665
},
{
"epoch": 0.33653360282971195,
"grad_norm": 0.3564821481704712,
"learning_rate": 4.347647036143856e-05,
"loss": 0.2183,
"step": 666
},
{
"epoch": 0.3370389085396665,
"grad_norm": 0.4883384704589844,
"learning_rate": 4.345525025596986e-05,
"loss": 0.5619,
"step": 667
},
{
"epoch": 0.337544214249621,
"grad_norm": 0.31306517124176025,
"learning_rate": 4.343400089021785e-05,
"loss": 0.1676,
"step": 668
},
{
"epoch": 0.33804951995957555,
"grad_norm": 0.7773547172546387,
"learning_rate": 4.341272229787281e-05,
"loss": 0.3485,
"step": 669
},
{
"epoch": 0.33855482566953005,
"grad_norm": 0.349680632352829,
"learning_rate": 4.339141451267136e-05,
"loss": 0.2042,
"step": 670
},
{
"epoch": 0.3390601313794846,
"grad_norm": 0.8842143416404724,
"learning_rate": 4.337007756839638e-05,
"loss": 0.4101,
"step": 671
},
{
"epoch": 0.3395654370894391,
"grad_norm": 0.4938321113586426,
"learning_rate": 4.334871149887701e-05,
"loss": 0.2784,
"step": 672
},
{
"epoch": 0.34007074279939364,
"grad_norm": 0.6494061350822449,
"learning_rate": 4.332731633798857e-05,
"loss": 0.4212,
"step": 673
},
{
"epoch": 0.34057604850934814,
"grad_norm": 0.33829325437545776,
"learning_rate": 4.3305892119652457e-05,
"loss": 0.2634,
"step": 674
},
{
"epoch": 0.3410813542193027,
"grad_norm": 0.5153347849845886,
"learning_rate": 4.32844388778362e-05,
"loss": 0.235,
"step": 675
},
{
"epoch": 0.3415866599292572,
"grad_norm": 0.4803626835346222,
"learning_rate": 4.326295664655329e-05,
"loss": 0.3195,
"step": 676
},
{
"epoch": 0.34209196563921174,
"grad_norm": 0.5081639289855957,
"learning_rate": 4.3241445459863225e-05,
"loss": 0.2381,
"step": 677
},
{
"epoch": 0.34259727134916623,
"grad_norm": 0.5200014710426331,
"learning_rate": 4.321990535187138e-05,
"loss": 0.3903,
"step": 678
},
{
"epoch": 0.3431025770591208,
"grad_norm": 0.5185570120811462,
"learning_rate": 4.319833635672899e-05,
"loss": 0.1852,
"step": 679
},
{
"epoch": 0.3436078827690753,
"grad_norm": 0.38153204321861267,
"learning_rate": 4.317673850863311e-05,
"loss": 0.1491,
"step": 680
},
{
"epoch": 0.34411318847902983,
"grad_norm": 0.4061324894428253,
"learning_rate": 4.31551118418265e-05,
"loss": 0.3238,
"step": 681
},
{
"epoch": 0.3446184941889843,
"grad_norm": 0.45140764117240906,
"learning_rate": 4.313345639059766e-05,
"loss": 0.2929,
"step": 682
},
{
"epoch": 0.3451237998989389,
"grad_norm": 0.5511850714683533,
"learning_rate": 4.3111772189280686e-05,
"loss": 0.2723,
"step": 683
},
{
"epoch": 0.34562910560889337,
"grad_norm": 0.5179259777069092,
"learning_rate": 4.309005927225528e-05,
"loss": 0.2211,
"step": 684
},
{
"epoch": 0.3461344113188479,
"grad_norm": 0.49272972345352173,
"learning_rate": 4.306831767394666e-05,
"loss": 0.2261,
"step": 685
},
{
"epoch": 0.3466397170288024,
"grad_norm": 0.6372472047805786,
"learning_rate": 4.3046547428825524e-05,
"loss": 0.3174,
"step": 686
},
{
"epoch": 0.34714502273875697,
"grad_norm": 1.0002518892288208,
"learning_rate": 4.302474857140799e-05,
"loss": 0.4862,
"step": 687
},
{
"epoch": 0.34765032844871147,
"grad_norm": 0.46908167004585266,
"learning_rate": 4.3002921136255535e-05,
"loss": 0.2217,
"step": 688
},
{
"epoch": 0.348155634158666,
"grad_norm": 0.31772932410240173,
"learning_rate": 4.298106515797496e-05,
"loss": 0.2601,
"step": 689
},
{
"epoch": 0.3486609398686205,
"grad_norm": 0.40609073638916016,
"learning_rate": 4.295918067121828e-05,
"loss": 0.2816,
"step": 690
},
{
"epoch": 0.34916624557857506,
"grad_norm": 0.5628025531768799,
"learning_rate": 4.293726771068275e-05,
"loss": 0.2602,
"step": 691
},
{
"epoch": 0.34967155128852956,
"grad_norm": 0.42692917585372925,
"learning_rate": 4.2915326311110766e-05,
"loss": 0.2569,
"step": 692
},
{
"epoch": 0.3501768569984841,
"grad_norm": 0.41802138090133667,
"learning_rate": 4.289335650728979e-05,
"loss": 0.3457,
"step": 693
},
{
"epoch": 0.3506821627084386,
"grad_norm": 0.5327426195144653,
"learning_rate": 4.287135833405235e-05,
"loss": 0.5488,
"step": 694
},
{
"epoch": 0.3511874684183931,
"grad_norm": 0.36314961314201355,
"learning_rate": 4.2849331826275905e-05,
"loss": 0.3411,
"step": 695
},
{
"epoch": 0.35169277412834765,
"grad_norm": 0.3764992356300354,
"learning_rate": 4.282727701888289e-05,
"loss": 0.2655,
"step": 696
},
{
"epoch": 0.35219807983830215,
"grad_norm": 0.4138118028640747,
"learning_rate": 4.280519394684059e-05,
"loss": 0.2747,
"step": 697
},
{
"epoch": 0.3527033855482567,
"grad_norm": 0.31555989384651184,
"learning_rate": 4.2783082645161073e-05,
"loss": 0.2322,
"step": 698
},
{
"epoch": 0.3532086912582112,
"grad_norm": 0.5712180733680725,
"learning_rate": 4.276094314890122e-05,
"loss": 0.3735,
"step": 699
},
{
"epoch": 0.35371399696816574,
"grad_norm": 0.764113187789917,
"learning_rate": 4.273877549316256e-05,
"loss": 0.4034,
"step": 700
},
{
"epoch": 0.35421930267812024,
"grad_norm": 0.8156772255897522,
"learning_rate": 4.27165797130913e-05,
"loss": 0.3585,
"step": 701
},
{
"epoch": 0.3547246083880748,
"grad_norm": 0.5966439843177795,
"learning_rate": 4.269435584387823e-05,
"loss": 0.413,
"step": 702
},
{
"epoch": 0.3552299140980293,
"grad_norm": 0.621375560760498,
"learning_rate": 4.267210392075867e-05,
"loss": 0.4242,
"step": 703
},
{
"epoch": 0.35573521980798384,
"grad_norm": 0.6661776304244995,
"learning_rate": 4.2649823979012426e-05,
"loss": 0.3709,
"step": 704
},
{
"epoch": 0.35624052551793833,
"grad_norm": 0.8590171933174133,
"learning_rate": 4.262751605396372e-05,
"loss": 0.4686,
"step": 705
},
{
"epoch": 0.3567458312278929,
"grad_norm": 0.5522347092628479,
"learning_rate": 4.260518018098114e-05,
"loss": 0.2676,
"step": 706
},
{
"epoch": 0.3572511369378474,
"grad_norm": 0.5579246282577515,
"learning_rate": 4.25828163954776e-05,
"loss": 0.3184,
"step": 707
},
{
"epoch": 0.35775644264780193,
"grad_norm": 0.6939714550971985,
"learning_rate": 4.2560424732910246e-05,
"loss": 0.3246,
"step": 708
},
{
"epoch": 0.3582617483577564,
"grad_norm": 0.5740989446640015,
"learning_rate": 4.2538005228780426e-05,
"loss": 0.261,
"step": 709
},
{
"epoch": 0.358767054067711,
"grad_norm": 0.5999382734298706,
"learning_rate": 4.251555791863365e-05,
"loss": 0.3403,
"step": 710
},
{
"epoch": 0.3592723597776655,
"grad_norm": 0.7810065746307373,
"learning_rate": 4.2493082838059506e-05,
"loss": 0.2779,
"step": 711
},
{
"epoch": 0.35977766548762,
"grad_norm": 0.6728970408439636,
"learning_rate": 4.2470580022691584e-05,
"loss": 0.3938,
"step": 712
},
{
"epoch": 0.3602829711975745,
"grad_norm": 0.42731353640556335,
"learning_rate": 4.2448049508207495e-05,
"loss": 0.3949,
"step": 713
},
{
"epoch": 0.36078827690752907,
"grad_norm": 0.5872007608413696,
"learning_rate": 4.242549133032872e-05,
"loss": 0.2941,
"step": 714
},
{
"epoch": 0.36129358261748357,
"grad_norm": 0.7026878595352173,
"learning_rate": 4.2402905524820625e-05,
"loss": 0.3886,
"step": 715
},
{
"epoch": 0.3617988883274381,
"grad_norm": 0.8333798050880432,
"learning_rate": 4.238029212749237e-05,
"loss": 0.4645,
"step": 716
},
{
"epoch": 0.3623041940373926,
"grad_norm": 0.4653269946575165,
"learning_rate": 4.235765117419688e-05,
"loss": 0.2918,
"step": 717
},
{
"epoch": 0.36280949974734716,
"grad_norm": 0.46014687418937683,
"learning_rate": 4.233498270083074e-05,
"loss": 0.1886,
"step": 718
},
{
"epoch": 0.36331480545730166,
"grad_norm": 0.5073797106742859,
"learning_rate": 4.231228674333417e-05,
"loss": 0.3901,
"step": 719
},
{
"epoch": 0.3638201111672562,
"grad_norm": 0.4070146679878235,
"learning_rate": 4.2289563337691004e-05,
"loss": 0.3156,
"step": 720
},
{
"epoch": 0.3643254168772107,
"grad_norm": 0.5377737283706665,
"learning_rate": 4.226681251992854e-05,
"loss": 0.3167,
"step": 721
},
{
"epoch": 0.36483072258716526,
"grad_norm": 0.590934693813324,
"learning_rate": 4.224403432611758e-05,
"loss": 0.2468,
"step": 722
},
{
"epoch": 0.36533602829711975,
"grad_norm": 0.47819361090660095,
"learning_rate": 4.22212287923723e-05,
"loss": 0.2129,
"step": 723
},
{
"epoch": 0.3658413340070743,
"grad_norm": 0.39465945959091187,
"learning_rate": 4.219839595485026e-05,
"loss": 0.1701,
"step": 724
},
{
"epoch": 0.3663466397170288,
"grad_norm": 0.5328222513198853,
"learning_rate": 4.2175535849752254e-05,
"loss": 0.2442,
"step": 725
},
{
"epoch": 0.36685194542698335,
"grad_norm": 0.4375799596309662,
"learning_rate": 4.215264851332236e-05,
"loss": 0.2072,
"step": 726
},
{
"epoch": 0.36735725113693785,
"grad_norm": 0.5246615409851074,
"learning_rate": 4.212973398184781e-05,
"loss": 0.2764,
"step": 727
},
{
"epoch": 0.3678625568468924,
"grad_norm": 0.5006076097488403,
"learning_rate": 4.210679229165894e-05,
"loss": 0.2225,
"step": 728
},
{
"epoch": 0.3683678625568469,
"grad_norm": 0.2846226990222931,
"learning_rate": 4.2083823479129175e-05,
"loss": 0.147,
"step": 729
},
{
"epoch": 0.3688731682668014,
"grad_norm": 0.4136352837085724,
"learning_rate": 4.206082758067491e-05,
"loss": 0.1903,
"step": 730
},
{
"epoch": 0.36937847397675594,
"grad_norm": 0.8793293833732605,
"learning_rate": 4.203780463275551e-05,
"loss": 0.4302,
"step": 731
},
{
"epoch": 0.36988377968671043,
"grad_norm": 0.45669475197792053,
"learning_rate": 4.2014754671873205e-05,
"loss": 0.2598,
"step": 732
},
{
"epoch": 0.370389085396665,
"grad_norm": 0.6066158413887024,
"learning_rate": 4.199167773457308e-05,
"loss": 0.2745,
"step": 733
},
{
"epoch": 0.3708943911066195,
"grad_norm": 0.6452160477638245,
"learning_rate": 4.196857385744295e-05,
"loss": 0.2735,
"step": 734
},
{
"epoch": 0.37139969681657403,
"grad_norm": 0.9064080715179443,
"learning_rate": 4.194544307711338e-05,
"loss": 0.5442,
"step": 735
},
{
"epoch": 0.3719050025265285,
"grad_norm": 0.5397379398345947,
"learning_rate": 4.192228543025759e-05,
"loss": 0.3049,
"step": 736
},
{
"epoch": 0.3724103082364831,
"grad_norm": 0.7189819812774658,
"learning_rate": 4.189910095359135e-05,
"loss": 0.4373,
"step": 737
},
{
"epoch": 0.3729156139464376,
"grad_norm": 0.4282380938529968,
"learning_rate": 4.187588968387303e-05,
"loss": 0.2361,
"step": 738
},
{
"epoch": 0.3734209196563921,
"grad_norm": 0.6697452068328857,
"learning_rate": 4.185265165790343e-05,
"loss": 0.7482,
"step": 739
},
{
"epoch": 0.3739262253663466,
"grad_norm": 0.8984920978546143,
"learning_rate": 4.182938691252579e-05,
"loss": 0.3867,
"step": 740
},
{
"epoch": 0.37443153107630117,
"grad_norm": 0.8725920915603638,
"learning_rate": 4.180609548462573e-05,
"loss": 0.3112,
"step": 741
},
{
"epoch": 0.37493683678625567,
"grad_norm": 0.548840343952179,
"learning_rate": 4.178277741113114e-05,
"loss": 0.2397,
"step": 742
},
{
"epoch": 0.3754421424962102,
"grad_norm": 0.38334935903549194,
"learning_rate": 4.175943272901218e-05,
"loss": 0.2286,
"step": 743
},
{
"epoch": 0.3759474482061647,
"grad_norm": 0.3433682918548584,
"learning_rate": 4.173606147528119e-05,
"loss": 0.2938,
"step": 744
},
{
"epoch": 0.37645275391611926,
"grad_norm": 0.3817785680294037,
"learning_rate": 4.171266368699265e-05,
"loss": 0.1541,
"step": 745
},
{
"epoch": 0.37695805962607376,
"grad_norm": 0.3005577325820923,
"learning_rate": 4.1689239401243086e-05,
"loss": 0.1874,
"step": 746
},
{
"epoch": 0.3774633653360283,
"grad_norm": 0.41856083273887634,
"learning_rate": 4.166578865517105e-05,
"loss": 0.3474,
"step": 747
},
{
"epoch": 0.3779686710459828,
"grad_norm": 0.6588180661201477,
"learning_rate": 4.164231148595705e-05,
"loss": 0.3855,
"step": 748
},
{
"epoch": 0.37847397675593736,
"grad_norm": 0.5970372557640076,
"learning_rate": 4.1618807930823474e-05,
"loss": 0.3084,
"step": 749
},
{
"epoch": 0.37897928246589185,
"grad_norm": 0.29243141412734985,
"learning_rate": 4.159527802703456e-05,
"loss": 0.1611,
"step": 750
},
{
"epoch": 0.3794845881758464,
"grad_norm": 0.5415188670158386,
"learning_rate": 4.1571721811896304e-05,
"loss": 0.2595,
"step": 751
},
{
"epoch": 0.3799898938858009,
"grad_norm": 0.33659324049949646,
"learning_rate": 4.154813932275643e-05,
"loss": 0.1616,
"step": 752
},
{
"epoch": 0.38049519959575545,
"grad_norm": 0.542412281036377,
"learning_rate": 4.152453059700431e-05,
"loss": 0.5679,
"step": 753
},
{
"epoch": 0.38100050530570995,
"grad_norm": 0.43469804525375366,
"learning_rate": 4.150089567207094e-05,
"loss": 0.1785,
"step": 754
},
{
"epoch": 0.3815058110156645,
"grad_norm": 0.27857232093811035,
"learning_rate": 4.14772345854288e-05,
"loss": 0.1647,
"step": 755
},
{
"epoch": 0.382011116725619,
"grad_norm": 0.45017367601394653,
"learning_rate": 4.145354737459191e-05,
"loss": 0.2412,
"step": 756
},
{
"epoch": 0.38251642243557354,
"grad_norm": 0.6093769669532776,
"learning_rate": 4.142983407711567e-05,
"loss": 0.3251,
"step": 757
},
{
"epoch": 0.38302172814552804,
"grad_norm": 0.9887945652008057,
"learning_rate": 4.1406094730596844e-05,
"loss": 0.4258,
"step": 758
},
{
"epoch": 0.3835270338554826,
"grad_norm": 0.3608369827270508,
"learning_rate": 4.138232937267351e-05,
"loss": 0.1477,
"step": 759
},
{
"epoch": 0.3840323395654371,
"grad_norm": 0.6515466570854187,
"learning_rate": 4.135853804102497e-05,
"loss": 0.346,
"step": 760
},
{
"epoch": 0.38453764527539164,
"grad_norm": 0.6651821732521057,
"learning_rate": 4.133472077337173e-05,
"loss": 0.4113,
"step": 761
},
{
"epoch": 0.38504295098534613,
"grad_norm": 0.43307211995124817,
"learning_rate": 4.131087760747538e-05,
"loss": 0.1571,
"step": 762
},
{
"epoch": 0.3855482566953007,
"grad_norm": 0.928577721118927,
"learning_rate": 4.128700858113862e-05,
"loss": 0.4672,
"step": 763
},
{
"epoch": 0.3860535624052552,
"grad_norm": 0.5100760459899902,
"learning_rate": 4.1263113732205106e-05,
"loss": 0.3695,
"step": 764
},
{
"epoch": 0.3865588681152097,
"grad_norm": 0.5143333673477173,
"learning_rate": 4.123919309855945e-05,
"loss": 0.2731,
"step": 765
},
{
"epoch": 0.3870641738251642,
"grad_norm": 0.705123245716095,
"learning_rate": 4.121524671812716e-05,
"loss": 0.2643,
"step": 766
},
{
"epoch": 0.3875694795351187,
"grad_norm": 0.44116538763046265,
"learning_rate": 4.1191274628874545e-05,
"loss": 0.2125,
"step": 767
},
{
"epoch": 0.3880747852450733,
"grad_norm": 1.2351047992706299,
"learning_rate": 4.116727686880869e-05,
"loss": 0.4035,
"step": 768
},
{
"epoch": 0.38858009095502777,
"grad_norm": 0.646878182888031,
"learning_rate": 4.114325347597736e-05,
"loss": 0.2416,
"step": 769
},
{
"epoch": 0.3890853966649823,
"grad_norm": 0.19294582307338715,
"learning_rate": 4.111920448846898e-05,
"loss": 0.0957,
"step": 770
},
{
"epoch": 0.3895907023749368,
"grad_norm": 0.45887112617492676,
"learning_rate": 4.1095129944412556e-05,
"loss": 0.4743,
"step": 771
},
{
"epoch": 0.39009600808489137,
"grad_norm": 0.3489443063735962,
"learning_rate": 4.107102988197758e-05,
"loss": 0.1996,
"step": 772
},
{
"epoch": 0.39060131379484586,
"grad_norm": 0.6400259137153625,
"learning_rate": 4.1046904339374056e-05,
"loss": 0.3377,
"step": 773
},
{
"epoch": 0.3911066195048004,
"grad_norm": 0.8612831830978394,
"learning_rate": 4.1022753354852345e-05,
"loss": 0.3707,
"step": 774
},
{
"epoch": 0.3916119252147549,
"grad_norm": 0.6197051405906677,
"learning_rate": 4.099857696670315e-05,
"loss": 0.3053,
"step": 775
},
{
"epoch": 0.39211723092470946,
"grad_norm": 0.5743224620819092,
"learning_rate": 4.0974375213257475e-05,
"loss": 0.3574,
"step": 776
},
{
"epoch": 0.39262253663466395,
"grad_norm": 0.5291123986244202,
"learning_rate": 4.0950148132886504e-05,
"loss": 0.2928,
"step": 777
},
{
"epoch": 0.3931278423446185,
"grad_norm": 0.9359875917434692,
"learning_rate": 4.0925895764001615e-05,
"loss": 0.5865,
"step": 778
},
{
"epoch": 0.393633148054573,
"grad_norm": 0.40809711813926697,
"learning_rate": 4.090161814505424e-05,
"loss": 0.2414,
"step": 779
},
{
"epoch": 0.39413845376452755,
"grad_norm": 0.3236285448074341,
"learning_rate": 4.087731531453588e-05,
"loss": 0.1779,
"step": 780
},
{
"epoch": 0.39464375947448205,
"grad_norm": 0.8675460815429688,
"learning_rate": 4.0852987310978e-05,
"loss": 0.4552,
"step": 781
},
{
"epoch": 0.3951490651844366,
"grad_norm": 0.3517732620239258,
"learning_rate": 4.082863417295195e-05,
"loss": 0.1672,
"step": 782
},
{
"epoch": 0.3956543708943911,
"grad_norm": 0.5657680630683899,
"learning_rate": 4.080425593906896e-05,
"loss": 0.3311,
"step": 783
},
{
"epoch": 0.39615967660434565,
"grad_norm": 0.7513840198516846,
"learning_rate": 4.077985264798003e-05,
"loss": 0.3397,
"step": 784
},
{
"epoch": 0.39666498231430014,
"grad_norm": 0.41284772753715515,
"learning_rate": 4.075542433837591e-05,
"loss": 0.1583,
"step": 785
},
{
"epoch": 0.3971702880242547,
"grad_norm": 0.41029801964759827,
"learning_rate": 4.0730971048987e-05,
"loss": 0.3041,
"step": 786
},
{
"epoch": 0.3976755937342092,
"grad_norm": 0.7941258549690247,
"learning_rate": 4.070649281858329e-05,
"loss": 0.3464,
"step": 787
},
{
"epoch": 0.39818089944416374,
"grad_norm": 1.075421929359436,
"learning_rate": 4.068198968597435e-05,
"loss": 0.5832,
"step": 788
},
{
"epoch": 0.39868620515411823,
"grad_norm": 0.46112877130508423,
"learning_rate": 4.0657461690009195e-05,
"loss": 0.322,
"step": 789
},
{
"epoch": 0.3991915108640728,
"grad_norm": 0.528905987739563,
"learning_rate": 4.063290886957629e-05,
"loss": 0.5137,
"step": 790
},
{
"epoch": 0.3996968165740273,
"grad_norm": 0.43017876148223877,
"learning_rate": 4.0608331263603437e-05,
"loss": 0.2687,
"step": 791
},
{
"epoch": 0.40020212228398183,
"grad_norm": 0.49231600761413574,
"learning_rate": 4.058372891105775e-05,
"loss": 0.2438,
"step": 792
},
{
"epoch": 0.4007074279939363,
"grad_norm": 0.7696442008018494,
"learning_rate": 4.055910185094557e-05,
"loss": 0.3306,
"step": 793
},
{
"epoch": 0.4012127337038909,
"grad_norm": 0.6930801272392273,
"learning_rate": 4.0534450122312404e-05,
"loss": 0.6515,
"step": 794
},
{
"epoch": 0.4017180394138454,
"grad_norm": 0.28239572048187256,
"learning_rate": 4.0509773764242886e-05,
"loss": 0.2526,
"step": 795
},
{
"epoch": 0.4022233451237999,
"grad_norm": 0.5681050419807434,
"learning_rate": 4.048507281586069e-05,
"loss": 0.2447,
"step": 796
},
{
"epoch": 0.4027286508337544,
"grad_norm": 0.3820132911205292,
"learning_rate": 4.046034731632847e-05,
"loss": 0.2624,
"step": 797
},
{
"epoch": 0.40323395654370897,
"grad_norm": 0.9789735674858093,
"learning_rate": 4.043559730484784e-05,
"loss": 0.5155,
"step": 798
},
{
"epoch": 0.40373926225366347,
"grad_norm": 0.4344816505908966,
"learning_rate": 4.0410822820659224e-05,
"loss": 0.2145,
"step": 799
},
{
"epoch": 0.40424456796361796,
"grad_norm": 0.33764201402664185,
"learning_rate": 4.0386023903041895e-05,
"loss": 0.2634,
"step": 800
},
{
"epoch": 0.4047498736735725,
"grad_norm": 0.2780141830444336,
"learning_rate": 4.036120059131382e-05,
"loss": 0.1559,
"step": 801
},
{
"epoch": 0.405255179383527,
"grad_norm": 0.5865265130996704,
"learning_rate": 4.0336352924831674e-05,
"loss": 0.3824,
"step": 802
},
{
"epoch": 0.40576048509348156,
"grad_norm": 0.5158589482307434,
"learning_rate": 4.0311480942990756e-05,
"loss": 0.3323,
"step": 803
},
{
"epoch": 0.40626579080343606,
"grad_norm": 0.5034775733947754,
"learning_rate": 4.0286584685224885e-05,
"loss": 0.2576,
"step": 804
},
{
"epoch": 0.4067710965133906,
"grad_norm": 0.5502609610557556,
"learning_rate": 4.026166419100638e-05,
"loss": 0.3489,
"step": 805
},
{
"epoch": 0.4072764022233451,
"grad_norm": 0.4883863031864166,
"learning_rate": 4.023671949984598e-05,
"loss": 0.2665,
"step": 806
},
{
"epoch": 0.40778170793329965,
"grad_norm": 0.40211907029151917,
"learning_rate": 4.02117506512928e-05,
"loss": 0.2291,
"step": 807
},
{
"epoch": 0.40828701364325415,
"grad_norm": 0.35028353333473206,
"learning_rate": 4.0186757684934275e-05,
"loss": 0.2617,
"step": 808
},
{
"epoch": 0.4087923193532087,
"grad_norm": 0.6691428422927856,
"learning_rate": 4.0161740640396015e-05,
"loss": 0.4747,
"step": 809
},
{
"epoch": 0.4092976250631632,
"grad_norm": 0.3790917694568634,
"learning_rate": 4.0136699557341874e-05,
"loss": 0.1726,
"step": 810
},
{
"epoch": 0.40980293077311775,
"grad_norm": 0.2536380887031555,
"learning_rate": 4.011163447547378e-05,
"loss": 0.1844,
"step": 811
},
{
"epoch": 0.41030823648307224,
"grad_norm": 0.7773703336715698,
"learning_rate": 4.0086545434531717e-05,
"loss": 0.2856,
"step": 812
},
{
"epoch": 0.4108135421930268,
"grad_norm": 0.36982572078704834,
"learning_rate": 4.0061432474293656e-05,
"loss": 0.1549,
"step": 813
},
{
"epoch": 0.4113188479029813,
"grad_norm": 0.3912838399410248,
"learning_rate": 4.00362956345755e-05,
"loss": 0.2821,
"step": 814
},
{
"epoch": 0.41182415361293584,
"grad_norm": 0.4090442359447479,
"learning_rate": 4.001113495523101e-05,
"loss": 0.2502,
"step": 815
},
{
"epoch": 0.41232945932289033,
"grad_norm": 0.860283374786377,
"learning_rate": 3.998595047615172e-05,
"loss": 0.3686,
"step": 816
},
{
"epoch": 0.4128347650328449,
"grad_norm": 0.3724400997161865,
"learning_rate": 3.9960742237266946e-05,
"loss": 0.1763,
"step": 817
},
{
"epoch": 0.4133400707427994,
"grad_norm": 0.2720755636692047,
"learning_rate": 3.993551027854362e-05,
"loss": 0.1704,
"step": 818
},
{
"epoch": 0.41384537645275393,
"grad_norm": 0.37592801451683044,
"learning_rate": 3.991025463998632e-05,
"loss": 0.2151,
"step": 819
},
{
"epoch": 0.41435068216270843,
"grad_norm": 0.5671641826629639,
"learning_rate": 3.9884975361637146e-05,
"loss": 0.5564,
"step": 820
},
{
"epoch": 0.414855987872663,
"grad_norm": 0.6280914545059204,
"learning_rate": 3.9859672483575694e-05,
"loss": 0.3185,
"step": 821
},
{
"epoch": 0.4153612935826175,
"grad_norm": 0.3018336892127991,
"learning_rate": 3.983434604591898e-05,
"loss": 0.235,
"step": 822
},
{
"epoch": 0.415866599292572,
"grad_norm": 0.7508260607719421,
"learning_rate": 3.980899608882134e-05,
"loss": 0.3689,
"step": 823
},
{
"epoch": 0.4163719050025265,
"grad_norm": 0.7853586077690125,
"learning_rate": 3.9783622652474436e-05,
"loss": 0.3986,
"step": 824
},
{
"epoch": 0.41687721071248107,
"grad_norm": 0.6880490779876709,
"learning_rate": 3.975822577710714e-05,
"loss": 0.3101,
"step": 825
},
{
"epoch": 0.41738251642243557,
"grad_norm": 0.686315655708313,
"learning_rate": 3.973280550298549e-05,
"loss": 0.3909,
"step": 826
},
{
"epoch": 0.4178878221323901,
"grad_norm": 0.22469963133335114,
"learning_rate": 3.970736187041262e-05,
"loss": 0.1319,
"step": 827
},
{
"epoch": 0.4183931278423446,
"grad_norm": 0.5535637736320496,
"learning_rate": 3.968189491972869e-05,
"loss": 0.3014,
"step": 828
},
{
"epoch": 0.41889843355229917,
"grad_norm": 0.5633406639099121,
"learning_rate": 3.9656404691310846e-05,
"loss": 0.3009,
"step": 829
},
{
"epoch": 0.41940373926225366,
"grad_norm": 0.6593132615089417,
"learning_rate": 3.963089122557313e-05,
"loss": 0.5059,
"step": 830
},
{
"epoch": 0.4199090449722082,
"grad_norm": 0.3774155378341675,
"learning_rate": 3.960535456296643e-05,
"loss": 0.3018,
"step": 831
},
{
"epoch": 0.4204143506821627,
"grad_norm": 0.39002594351768494,
"learning_rate": 3.957979474397841e-05,
"loss": 0.3974,
"step": 832
},
{
"epoch": 0.42091965639211726,
"grad_norm": 0.5461925268173218,
"learning_rate": 3.9554211809133446e-05,
"loss": 0.3468,
"step": 833
},
{
"epoch": 0.42142496210207175,
"grad_norm": 0.34177666902542114,
"learning_rate": 3.9528605798992566e-05,
"loss": 0.1479,
"step": 834
},
{
"epoch": 0.42193026781202625,
"grad_norm": 0.5321192741394043,
"learning_rate": 3.950297675415339e-05,
"loss": 0.3468,
"step": 835
},
{
"epoch": 0.4224355735219808,
"grad_norm": 0.4154065251350403,
"learning_rate": 3.9477324715250045e-05,
"loss": 0.3419,
"step": 836
},
{
"epoch": 0.4229408792319353,
"grad_norm": 0.4703141450881958,
"learning_rate": 3.945164972295313e-05,
"loss": 0.3884,
"step": 837
},
{
"epoch": 0.42344618494188985,
"grad_norm": 0.43752673268318176,
"learning_rate": 3.942595181796961e-05,
"loss": 0.2079,
"step": 838
},
{
"epoch": 0.42395149065184434,
"grad_norm": 0.3551587760448456,
"learning_rate": 3.940023104104281e-05,
"loss": 0.2126,
"step": 839
},
{
"epoch": 0.4244567963617989,
"grad_norm": 0.8521201610565186,
"learning_rate": 3.9374487432952304e-05,
"loss": 0.4272,
"step": 840
},
{
"epoch": 0.4249621020717534,
"grad_norm": 0.477527916431427,
"learning_rate": 3.934872103451386e-05,
"loss": 0.2675,
"step": 841
},
{
"epoch": 0.42546740778170794,
"grad_norm": 0.46821895241737366,
"learning_rate": 3.932293188657937e-05,
"loss": 0.2343,
"step": 842
},
{
"epoch": 0.42597271349166244,
"grad_norm": 0.7220411896705627,
"learning_rate": 3.929712003003681e-05,
"loss": 0.4072,
"step": 843
},
{
"epoch": 0.426478019201617,
"grad_norm": 0.7784841656684875,
"learning_rate": 3.927128550581019e-05,
"loss": 0.5188,
"step": 844
},
{
"epoch": 0.4269833249115715,
"grad_norm": 0.8209807872772217,
"learning_rate": 3.924542835485938e-05,
"loss": 0.6626,
"step": 845
},
{
"epoch": 0.42748863062152603,
"grad_norm": 0.4226566553115845,
"learning_rate": 3.9219548618180194e-05,
"loss": 0.2538,
"step": 846
},
{
"epoch": 0.42799393633148053,
"grad_norm": 0.45771926641464233,
"learning_rate": 3.919364633680423e-05,
"loss": 0.3554,
"step": 847
},
{
"epoch": 0.4284992420414351,
"grad_norm": 0.4026903808116913,
"learning_rate": 3.916772155179882e-05,
"loss": 0.2283,
"step": 848
},
{
"epoch": 0.4290045477513896,
"grad_norm": 0.474115252494812,
"learning_rate": 3.9141774304266995e-05,
"loss": 0.3667,
"step": 849
},
{
"epoch": 0.4295098534613441,
"grad_norm": 0.5068637132644653,
"learning_rate": 3.91158046353474e-05,
"loss": 0.3598,
"step": 850
},
{
"epoch": 0.4300151591712986,
"grad_norm": 0.63497394323349,
"learning_rate": 3.9089812586214195e-05,
"loss": 0.3532,
"step": 851
},
{
"epoch": 0.4305204648812532,
"grad_norm": 1.019968032836914,
"learning_rate": 3.906379819807707e-05,
"loss": 0.5262,
"step": 852
},
{
"epoch": 0.43102577059120767,
"grad_norm": 0.7702900171279907,
"learning_rate": 3.903776151218109e-05,
"loss": 0.4335,
"step": 853
},
{
"epoch": 0.4315310763011622,
"grad_norm": 0.4589867889881134,
"learning_rate": 3.901170256980671e-05,
"loss": 0.1948,
"step": 854
},
{
"epoch": 0.4320363820111167,
"grad_norm": 0.3241965174674988,
"learning_rate": 3.8985621412269644e-05,
"loss": 0.3025,
"step": 855
},
{
"epoch": 0.43254168772107127,
"grad_norm": 0.5903672575950623,
"learning_rate": 3.895951808092084e-05,
"loss": 0.2925,
"step": 856
},
{
"epoch": 0.43304699343102576,
"grad_norm": 0.42580774426460266,
"learning_rate": 3.893339261714642e-05,
"loss": 0.4694,
"step": 857
},
{
"epoch": 0.4335522991409803,
"grad_norm": 0.4376884698867798,
"learning_rate": 3.890724506236753e-05,
"loss": 0.2743,
"step": 858
},
{
"epoch": 0.4340576048509348,
"grad_norm": 0.5972741842269897,
"learning_rate": 3.8881075458040426e-05,
"loss": 0.3676,
"step": 859
},
{
"epoch": 0.43456291056088936,
"grad_norm": 0.6178309321403503,
"learning_rate": 3.8854883845656275e-05,
"loss": 0.3012,
"step": 860
},
{
"epoch": 0.43506821627084385,
"grad_norm": 0.8123751878738403,
"learning_rate": 3.882867026674114e-05,
"loss": 0.4476,
"step": 861
},
{
"epoch": 0.4355735219807984,
"grad_norm": 0.5609534978866577,
"learning_rate": 3.880243476285593e-05,
"loss": 0.3029,
"step": 862
},
{
"epoch": 0.4360788276907529,
"grad_norm": 0.5837926268577576,
"learning_rate": 3.877617737559629e-05,
"loss": 0.4239,
"step": 863
},
{
"epoch": 0.43658413340070745,
"grad_norm": 0.40057340264320374,
"learning_rate": 3.874989814659258e-05,
"loss": 0.1811,
"step": 864
},
{
"epoch": 0.43708943911066195,
"grad_norm": 0.45030128955841064,
"learning_rate": 3.872359711750978e-05,
"loss": 0.2334,
"step": 865
},
{
"epoch": 0.4375947448206165,
"grad_norm": 0.3968111574649811,
"learning_rate": 3.8697274330047427e-05,
"loss": 0.2225,
"step": 866
},
{
"epoch": 0.438100050530571,
"grad_norm": 0.6887713074684143,
"learning_rate": 3.8670929825939586e-05,
"loss": 0.5391,
"step": 867
},
{
"epoch": 0.4386053562405255,
"grad_norm": 1.0266462564468384,
"learning_rate": 3.864456364695472e-05,
"loss": 0.4051,
"step": 868
},
{
"epoch": 0.43911066195048004,
"grad_norm": 0.40764105319976807,
"learning_rate": 3.8618175834895664e-05,
"loss": 0.2342,
"step": 869
},
{
"epoch": 0.43961596766043454,
"grad_norm": 0.3224446475505829,
"learning_rate": 3.8591766431599566e-05,
"loss": 0.1845,
"step": 870
},
{
"epoch": 0.4401212733703891,
"grad_norm": 0.8195884227752686,
"learning_rate": 3.856533547893779e-05,
"loss": 0.4636,
"step": 871
},
{
"epoch": 0.4406265790803436,
"grad_norm": 0.4880749583244324,
"learning_rate": 3.853888301881587e-05,
"loss": 0.3212,
"step": 872
},
{
"epoch": 0.44113188479029813,
"grad_norm": 0.6332755088806152,
"learning_rate": 3.8512409093173453e-05,
"loss": 0.3923,
"step": 873
},
{
"epoch": 0.44163719050025263,
"grad_norm": 0.41636690497398376,
"learning_rate": 3.8485913743984206e-05,
"loss": 0.2354,
"step": 874
},
{
"epoch": 0.4421424962102072,
"grad_norm": 0.5678810477256775,
"learning_rate": 3.845939701325575e-05,
"loss": 0.2633,
"step": 875
},
{
"epoch": 0.4426478019201617,
"grad_norm": 0.3602360188961029,
"learning_rate": 3.8432858943029625e-05,
"loss": 0.2014,
"step": 876
},
{
"epoch": 0.4431531076301162,
"grad_norm": 0.48009565472602844,
"learning_rate": 3.840629957538121e-05,
"loss": 0.3181,
"step": 877
},
{
"epoch": 0.4436584133400707,
"grad_norm": 0.3176063001155853,
"learning_rate": 3.837971895241964e-05,
"loss": 0.1613,
"step": 878
},
{
"epoch": 0.4441637190500253,
"grad_norm": 0.5749449133872986,
"learning_rate": 3.835311711628774e-05,
"loss": 0.2763,
"step": 879
},
{
"epoch": 0.44466902475997977,
"grad_norm": 0.40219807624816895,
"learning_rate": 3.832649410916198e-05,
"loss": 0.1707,
"step": 880
},
{
"epoch": 0.4451743304699343,
"grad_norm": 0.6188439130783081,
"learning_rate": 3.8299849973252386e-05,
"loss": 0.2757,
"step": 881
},
{
"epoch": 0.4456796361798888,
"grad_norm": 0.6787897348403931,
"learning_rate": 3.827318475080252e-05,
"loss": 0.2632,
"step": 882
},
{
"epoch": 0.44618494188984337,
"grad_norm": 0.46739840507507324,
"learning_rate": 3.8246498484089306e-05,
"loss": 0.2511,
"step": 883
},
{
"epoch": 0.44669024759979786,
"grad_norm": 0.5028379559516907,
"learning_rate": 3.82197912154231e-05,
"loss": 0.3607,
"step": 884
},
{
"epoch": 0.4471955533097524,
"grad_norm": 0.45272600650787354,
"learning_rate": 3.819306298714752e-05,
"loss": 0.2706,
"step": 885
},
{
"epoch": 0.4477008590197069,
"grad_norm": 0.37481459975242615,
"learning_rate": 3.816631384163944e-05,
"loss": 0.1997,
"step": 886
},
{
"epoch": 0.44820616472966146,
"grad_norm": 0.7351203560829163,
"learning_rate": 3.813954382130886e-05,
"loss": 0.2668,
"step": 887
},
{
"epoch": 0.44871147043961596,
"grad_norm": 0.7236170768737793,
"learning_rate": 3.8112752968598907e-05,
"loss": 0.3759,
"step": 888
},
{
"epoch": 0.4492167761495705,
"grad_norm": 0.36974865198135376,
"learning_rate": 3.808594132598574e-05,
"loss": 0.2733,
"step": 889
},
{
"epoch": 0.449722081859525,
"grad_norm": 0.76992267370224,
"learning_rate": 3.805910893597845e-05,
"loss": 0.4098,
"step": 890
},
{
"epoch": 0.45022738756947955,
"grad_norm": 0.31903111934661865,
"learning_rate": 3.803225584111907e-05,
"loss": 0.1713,
"step": 891
},
{
"epoch": 0.45073269327943405,
"grad_norm": 0.7338394522666931,
"learning_rate": 3.8005382083982386e-05,
"loss": 0.7272,
"step": 892
},
{
"epoch": 0.4512379989893886,
"grad_norm": 0.9718958139419556,
"learning_rate": 3.797848770717603e-05,
"loss": 0.4103,
"step": 893
},
{
"epoch": 0.4517433046993431,
"grad_norm": 0.24202537536621094,
"learning_rate": 3.795157275334027e-05,
"loss": 0.2362,
"step": 894
},
{
"epoch": 0.45224861040929765,
"grad_norm": 1.0490413904190063,
"learning_rate": 3.7924637265148e-05,
"loss": 0.4398,
"step": 895
},
{
"epoch": 0.45275391611925214,
"grad_norm": 0.46118178963661194,
"learning_rate": 3.789768128530471e-05,
"loss": 0.3123,
"step": 896
},
{
"epoch": 0.4532592218292067,
"grad_norm": 0.3323919475078583,
"learning_rate": 3.787070485654833e-05,
"loss": 0.1714,
"step": 897
},
{
"epoch": 0.4537645275391612,
"grad_norm": 0.8963467478752136,
"learning_rate": 3.784370802164925e-05,
"loss": 0.5634,
"step": 898
},
{
"epoch": 0.45426983324911574,
"grad_norm": 0.9990172386169434,
"learning_rate": 3.781669082341018e-05,
"loss": 0.7088,
"step": 899
},
{
"epoch": 0.45477513895907024,
"grad_norm": 0.5156543850898743,
"learning_rate": 3.7789653304666136e-05,
"loss": 0.2425,
"step": 900
},
{
"epoch": 0.4552804446690248,
"grad_norm": 0.3596484363079071,
"learning_rate": 3.776259550828437e-05,
"loss": 0.1751,
"step": 901
},
{
"epoch": 0.4557857503789793,
"grad_norm": 0.7215254902839661,
"learning_rate": 3.773551747716423e-05,
"loss": 0.373,
"step": 902
},
{
"epoch": 0.4562910560889338,
"grad_norm": 0.36745965480804443,
"learning_rate": 3.77084192542372e-05,
"loss": 0.1979,
"step": 903
},
{
"epoch": 0.45679636179888833,
"grad_norm": 0.3110678791999817,
"learning_rate": 3.768130088246674e-05,
"loss": 0.1676,
"step": 904
},
{
"epoch": 0.4573016675088428,
"grad_norm": 0.6911036968231201,
"learning_rate": 3.765416240484828e-05,
"loss": 0.331,
"step": 905
},
{
"epoch": 0.4578069732187974,
"grad_norm": 0.48644575476646423,
"learning_rate": 3.7627003864409116e-05,
"loss": 0.4377,
"step": 906
},
{
"epoch": 0.45831227892875187,
"grad_norm": 0.5354007482528687,
"learning_rate": 3.759982530420835e-05,
"loss": 0.2489,
"step": 907
},
{
"epoch": 0.4588175846387064,
"grad_norm": 0.4465010464191437,
"learning_rate": 3.757262676733684e-05,
"loss": 0.2772,
"step": 908
},
{
"epoch": 0.4593228903486609,
"grad_norm": 0.47116124629974365,
"learning_rate": 3.754540829691708e-05,
"loss": 0.202,
"step": 909
},
{
"epoch": 0.45982819605861547,
"grad_norm": 0.635790228843689,
"learning_rate": 3.751816993610323e-05,
"loss": 0.2873,
"step": 910
},
{
"epoch": 0.46033350176856996,
"grad_norm": 0.3087238371372223,
"learning_rate": 3.749091172808093e-05,
"loss": 0.19,
"step": 911
},
{
"epoch": 0.4608388074785245,
"grad_norm": 0.3472057282924652,
"learning_rate": 3.74636337160673e-05,
"loss": 0.2691,
"step": 912
},
{
"epoch": 0.461344113188479,
"grad_norm": 0.5002533197402954,
"learning_rate": 3.743633594331089e-05,
"loss": 0.2269,
"step": 913
},
{
"epoch": 0.46184941889843356,
"grad_norm": 0.5435811877250671,
"learning_rate": 3.740901845309152e-05,
"loss": 0.2686,
"step": 914
},
{
"epoch": 0.46235472460838806,
"grad_norm": 0.5896182060241699,
"learning_rate": 3.738168128872033e-05,
"loss": 0.3024,
"step": 915
},
{
"epoch": 0.4628600303183426,
"grad_norm": 0.6780158281326294,
"learning_rate": 3.735432449353963e-05,
"loss": 0.3342,
"step": 916
},
{
"epoch": 0.4633653360282971,
"grad_norm": 0.30589649081230164,
"learning_rate": 3.7326948110922855e-05,
"loss": 0.1874,
"step": 917
},
{
"epoch": 0.46387064173825165,
"grad_norm": 0.729263424873352,
"learning_rate": 3.729955218427449e-05,
"loss": 0.2734,
"step": 918
},
{
"epoch": 0.46437594744820615,
"grad_norm": 0.4440087080001831,
"learning_rate": 3.7272136757029995e-05,
"loss": 0.2259,
"step": 919
},
{
"epoch": 0.4648812531581607,
"grad_norm": 0.9144408702850342,
"learning_rate": 3.724470187265579e-05,
"loss": 0.544,
"step": 920
},
{
"epoch": 0.4653865588681152,
"grad_norm": 0.6351197957992554,
"learning_rate": 3.721724757464911e-05,
"loss": 0.3172,
"step": 921
},
{
"epoch": 0.46589186457806975,
"grad_norm": 0.7643218040466309,
"learning_rate": 3.718977390653796e-05,
"loss": 0.5287,
"step": 922
},
{
"epoch": 0.46639717028802424,
"grad_norm": 0.5360459685325623,
"learning_rate": 3.7162280911881085e-05,
"loss": 0.1788,
"step": 923
},
{
"epoch": 0.4669024759979788,
"grad_norm": 0.4476550221443176,
"learning_rate": 3.713476863426787e-05,
"loss": 0.2565,
"step": 924
},
{
"epoch": 0.4674077817079333,
"grad_norm": 0.3940451443195343,
"learning_rate": 3.710723711731825e-05,
"loss": 0.2263,
"step": 925
},
{
"epoch": 0.46791308741788784,
"grad_norm": 0.5123595595359802,
"learning_rate": 3.7079686404682674e-05,
"loss": 0.2216,
"step": 926
},
{
"epoch": 0.46841839312784234,
"grad_norm": 0.36884805560112,
"learning_rate": 3.7052116540042025e-05,
"loss": 0.2212,
"step": 927
},
{
"epoch": 0.4689236988377969,
"grad_norm": 0.6213935017585754,
"learning_rate": 3.7024527567107564e-05,
"loss": 0.3156,
"step": 928
},
{
"epoch": 0.4694290045477514,
"grad_norm": 0.40270182490348816,
"learning_rate": 3.6996919529620825e-05,
"loss": 0.1813,
"step": 929
},
{
"epoch": 0.46993431025770593,
"grad_norm": 0.9014922976493835,
"learning_rate": 3.696929247135359e-05,
"loss": 0.46,
"step": 930
},
{
"epoch": 0.47043961596766043,
"grad_norm": 0.5056794881820679,
"learning_rate": 3.694164643610777e-05,
"loss": 0.2321,
"step": 931
},
{
"epoch": 0.470944921677615,
"grad_norm": 0.47349992394447327,
"learning_rate": 3.6913981467715406e-05,
"loss": 0.1869,
"step": 932
},
{
"epoch": 0.4714502273875695,
"grad_norm": 0.7985735535621643,
"learning_rate": 3.6886297610038514e-05,
"loss": 0.37,
"step": 933
},
{
"epoch": 0.471955533097524,
"grad_norm": 0.9785382151603699,
"learning_rate": 3.6858594906969077e-05,
"loss": 0.4063,
"step": 934
},
{
"epoch": 0.4724608388074785,
"grad_norm": 0.32149001955986023,
"learning_rate": 3.6830873402428964e-05,
"loss": 0.2423,
"step": 935
},
{
"epoch": 0.4729661445174331,
"grad_norm": 0.8392232060432434,
"learning_rate": 3.6803133140369836e-05,
"loss": 0.7566,
"step": 936
},
{
"epoch": 0.47347145022738757,
"grad_norm": 0.383240669965744,
"learning_rate": 3.6775374164773105e-05,
"loss": 0.1921,
"step": 937
},
{
"epoch": 0.47397675593734206,
"grad_norm": 0.6196987628936768,
"learning_rate": 3.6747596519649864e-05,
"loss": 0.4214,
"step": 938
},
{
"epoch": 0.4744820616472966,
"grad_norm": 0.5443540811538696,
"learning_rate": 3.671980024904078e-05,
"loss": 0.2819,
"step": 939
},
{
"epoch": 0.4749873673572511,
"grad_norm": 0.5839280486106873,
"learning_rate": 3.669198539701606e-05,
"loss": 0.345,
"step": 940
},
{
"epoch": 0.47549267306720566,
"grad_norm": 0.43209922313690186,
"learning_rate": 3.666415200767538e-05,
"loss": 0.2777,
"step": 941
},
{
"epoch": 0.47599797877716016,
"grad_norm": 0.4791796803474426,
"learning_rate": 3.6636300125147804e-05,
"loss": 0.2516,
"step": 942
},
{
"epoch": 0.4765032844871147,
"grad_norm": 0.5791419148445129,
"learning_rate": 3.660842979359171e-05,
"loss": 0.4067,
"step": 943
},
{
"epoch": 0.4770085901970692,
"grad_norm": 0.4313831329345703,
"learning_rate": 3.6580541057194726e-05,
"loss": 0.1971,
"step": 944
},
{
"epoch": 0.47751389590702376,
"grad_norm": 0.3102171719074249,
"learning_rate": 3.655263396017367e-05,
"loss": 0.2316,
"step": 945
},
{
"epoch": 0.47801920161697825,
"grad_norm": 0.31960973143577576,
"learning_rate": 3.652470854677446e-05,
"loss": 0.2716,
"step": 946
},
{
"epoch": 0.4785245073269328,
"grad_norm": 0.24606995284557343,
"learning_rate": 3.649676486127206e-05,
"loss": 0.1721,
"step": 947
},
{
"epoch": 0.4790298130368873,
"grad_norm": 0.5517093539237976,
"learning_rate": 3.64688029479704e-05,
"loss": 0.3408,
"step": 948
},
{
"epoch": 0.47953511874684185,
"grad_norm": 0.6657114028930664,
"learning_rate": 3.644082285120231e-05,
"loss": 0.5508,
"step": 949
},
{
"epoch": 0.48004042445679634,
"grad_norm": 0.39712589979171753,
"learning_rate": 3.641282461532946e-05,
"loss": 0.2008,
"step": 950
},
{
"epoch": 0.4805457301667509,
"grad_norm": 0.5426539182662964,
"learning_rate": 3.6384808284742254e-05,
"loss": 0.3242,
"step": 951
},
{
"epoch": 0.4810510358767054,
"grad_norm": 0.4150180518627167,
"learning_rate": 3.635677390385982e-05,
"loss": 0.2666,
"step": 952
},
{
"epoch": 0.48155634158665994,
"grad_norm": 0.3427567780017853,
"learning_rate": 3.632872151712988e-05,
"loss": 0.2593,
"step": 953
},
{
"epoch": 0.48206164729661444,
"grad_norm": 0.5839005708694458,
"learning_rate": 3.63006511690287e-05,
"loss": 0.4472,
"step": 954
},
{
"epoch": 0.482566953006569,
"grad_norm": 0.6474454998970032,
"learning_rate": 3.627256290406103e-05,
"loss": 0.3743,
"step": 955
},
{
"epoch": 0.4830722587165235,
"grad_norm": 0.361176073551178,
"learning_rate": 3.624445676676004e-05,
"loss": 0.2014,
"step": 956
},
{
"epoch": 0.48357756442647803,
"grad_norm": 0.6297400593757629,
"learning_rate": 3.621633280168723e-05,
"loss": 0.5208,
"step": 957
},
{
"epoch": 0.48408287013643253,
"grad_norm": 0.38691607117652893,
"learning_rate": 3.618819105343234e-05,
"loss": 0.3079,
"step": 958
},
{
"epoch": 0.4845881758463871,
"grad_norm": 0.48631709814071655,
"learning_rate": 3.6160031566613336e-05,
"loss": 0.3092,
"step": 959
},
{
"epoch": 0.4850934815563416,
"grad_norm": 0.30502668023109436,
"learning_rate": 3.6131854385876314e-05,
"loss": 0.2976,
"step": 960
},
{
"epoch": 0.48559878726629613,
"grad_norm": 0.6193214654922485,
"learning_rate": 3.6103659555895376e-05,
"loss": 0.3162,
"step": 961
},
{
"epoch": 0.4861040929762506,
"grad_norm": 0.5869472026824951,
"learning_rate": 3.607544712137265e-05,
"loss": 0.5059,
"step": 962
},
{
"epoch": 0.4866093986862052,
"grad_norm": 0.6832082867622375,
"learning_rate": 3.604721712703817e-05,
"loss": 0.2766,
"step": 963
},
{
"epoch": 0.48711470439615967,
"grad_norm": 0.8644431829452515,
"learning_rate": 3.601896961764978e-05,
"loss": 0.632,
"step": 964
},
{
"epoch": 0.4876200101061142,
"grad_norm": 0.7030072212219238,
"learning_rate": 3.599070463799315e-05,
"loss": 0.4698,
"step": 965
},
{
"epoch": 0.4881253158160687,
"grad_norm": 0.49395883083343506,
"learning_rate": 3.5962422232881585e-05,
"loss": 0.1997,
"step": 966
},
{
"epoch": 0.48863062152602327,
"grad_norm": 0.3058975338935852,
"learning_rate": 3.5934122447156064e-05,
"loss": 0.1957,
"step": 967
},
{
"epoch": 0.48913592723597776,
"grad_norm": 0.6280907392501831,
"learning_rate": 3.59058053256851e-05,
"loss": 0.2984,
"step": 968
},
{
"epoch": 0.4896412329459323,
"grad_norm": 0.3592299818992615,
"learning_rate": 3.5877470913364695e-05,
"loss": 0.1573,
"step": 969
},
{
"epoch": 0.4901465386558868,
"grad_norm": 0.3483614921569824,
"learning_rate": 3.5849119255118274e-05,
"loss": 0.1568,
"step": 970
},
{
"epoch": 0.49065184436584136,
"grad_norm": 0.2722894847393036,
"learning_rate": 3.58207503958966e-05,
"loss": 0.1799,
"step": 971
},
{
"epoch": 0.49115715007579586,
"grad_norm": 0.4917094111442566,
"learning_rate": 3.57923643806777e-05,
"loss": 0.2378,
"step": 972
},
{
"epoch": 0.49166245578575035,
"grad_norm": 0.36408787965774536,
"learning_rate": 3.57639612544668e-05,
"loss": 0.1882,
"step": 973
},
{
"epoch": 0.4921677614957049,
"grad_norm": 0.6604106426239014,
"learning_rate": 3.573554106229629e-05,
"loss": 0.5856,
"step": 974
},
{
"epoch": 0.4926730672056594,
"grad_norm": 0.6453742384910583,
"learning_rate": 3.5707103849225554e-05,
"loss": 0.5509,
"step": 975
},
{
"epoch": 0.49317837291561395,
"grad_norm": 0.431162565946579,
"learning_rate": 3.5678649660341026e-05,
"loss": 0.2175,
"step": 976
},
{
"epoch": 0.49368367862556845,
"grad_norm": 0.510044515132904,
"learning_rate": 3.565017854075602e-05,
"loss": 0.2819,
"step": 977
},
{
"epoch": 0.494188984335523,
"grad_norm": 0.6410518288612366,
"learning_rate": 3.5621690535610684e-05,
"loss": 0.3669,
"step": 978
},
{
"epoch": 0.4946942900454775,
"grad_norm": 0.2983573377132416,
"learning_rate": 3.559318569007198e-05,
"loss": 0.2469,
"step": 979
},
{
"epoch": 0.49519959575543204,
"grad_norm": 0.4974038898944855,
"learning_rate": 3.5564664049333527e-05,
"loss": 0.2096,
"step": 980
},
{
"epoch": 0.49570490146538654,
"grad_norm": 0.685854434967041,
"learning_rate": 3.5536125658615586e-05,
"loss": 0.2544,
"step": 981
},
{
"epoch": 0.4962102071753411,
"grad_norm": 0.5799931287765503,
"learning_rate": 3.5507570563164996e-05,
"loss": 0.3086,
"step": 982
},
{
"epoch": 0.4967155128852956,
"grad_norm": 0.2882770597934723,
"learning_rate": 3.5478998808255045e-05,
"loss": 0.2204,
"step": 983
},
{
"epoch": 0.49722081859525014,
"grad_norm": 0.40922459959983826,
"learning_rate": 3.545041043918546e-05,
"loss": 0.1892,
"step": 984
},
{
"epoch": 0.49772612430520463,
"grad_norm": 0.3756551146507263,
"learning_rate": 3.542180550128229e-05,
"loss": 0.2262,
"step": 985
},
{
"epoch": 0.4982314300151592,
"grad_norm": 0.8342236280441284,
"learning_rate": 3.5393184039897876e-05,
"loss": 0.4221,
"step": 986
},
{
"epoch": 0.4987367357251137,
"grad_norm": 0.7572706341743469,
"learning_rate": 3.5364546100410744e-05,
"loss": 0.3071,
"step": 987
},
{
"epoch": 0.49924204143506823,
"grad_norm": 0.44951027631759644,
"learning_rate": 3.533589172822554e-05,
"loss": 0.3021,
"step": 988
},
{
"epoch": 0.4997473471450227,
"grad_norm": 0.5929198265075684,
"learning_rate": 3.530722096877298e-05,
"loss": 0.3193,
"step": 989
},
{
"epoch": 0.5002526528549772,
"grad_norm": 0.6089531779289246,
"learning_rate": 3.527853386750974e-05,
"loss": 0.3985,
"step": 990
},
{
"epoch": 0.5007579585649318,
"grad_norm": 0.6241928935050964,
"learning_rate": 3.524983046991842e-05,
"loss": 0.4037,
"step": 991
},
{
"epoch": 0.5012632642748863,
"grad_norm": 0.5915480852127075,
"learning_rate": 3.522111082150746e-05,
"loss": 0.3821,
"step": 992
},
{
"epoch": 0.5017685699848409,
"grad_norm": 0.956665575504303,
"learning_rate": 3.519237496781106e-05,
"loss": 0.5615,
"step": 993
},
{
"epoch": 0.5022738756947953,
"grad_norm": 0.7047909498214722,
"learning_rate": 3.5163622954389105e-05,
"loss": 0.3062,
"step": 994
},
{
"epoch": 0.5027791814047499,
"grad_norm": 0.31413036584854126,
"learning_rate": 3.5134854826827114e-05,
"loss": 0.2965,
"step": 995
},
{
"epoch": 0.5032844871147044,
"grad_norm": 0.6101884245872498,
"learning_rate": 3.510607063073614e-05,
"loss": 0.4699,
"step": 996
},
{
"epoch": 0.503789792824659,
"grad_norm": 0.6464351415634155,
"learning_rate": 3.5077270411752746e-05,
"loss": 0.3672,
"step": 997
},
{
"epoch": 0.5042950985346134,
"grad_norm": 0.5930270552635193,
"learning_rate": 3.504845421553884e-05,
"loss": 0.1938,
"step": 998
},
{
"epoch": 0.504800404244568,
"grad_norm": 0.614264190196991,
"learning_rate": 3.501962208778172e-05,
"loss": 0.3049,
"step": 999
},
{
"epoch": 0.5053057099545225,
"grad_norm": 0.5080682635307312,
"learning_rate": 3.4990774074193906e-05,
"loss": 0.3748,
"step": 1000
},
{
"epoch": 0.5058110156644771,
"grad_norm": 0.30347660183906555,
"learning_rate": 3.496191022051312e-05,
"loss": 0.2212,
"step": 1001
},
{
"epoch": 0.5063163213744315,
"grad_norm": 0.2643427848815918,
"learning_rate": 3.49330305725022e-05,
"loss": 0.1181,
"step": 1002
},
{
"epoch": 0.506821627084386,
"grad_norm": 0.6040931940078735,
"learning_rate": 3.490413517594901e-05,
"loss": 0.4396,
"step": 1003
},
{
"epoch": 0.5073269327943406,
"grad_norm": 0.4380238652229309,
"learning_rate": 3.487522407666641e-05,
"loss": 0.2364,
"step": 1004
},
{
"epoch": 0.5078322385042952,
"grad_norm": 1.0693117380142212,
"learning_rate": 3.4846297320492124e-05,
"loss": 0.6679,
"step": 1005
},
{
"epoch": 0.5083375442142496,
"grad_norm": 0.4763444662094116,
"learning_rate": 3.4817354953288715e-05,
"loss": 0.3052,
"step": 1006
},
{
"epoch": 0.5088428499242041,
"grad_norm": 0.41242119669914246,
"learning_rate": 3.478839702094351e-05,
"loss": 0.2532,
"step": 1007
},
{
"epoch": 0.5093481556341587,
"grad_norm": 0.41044360399246216,
"learning_rate": 3.47594235693685e-05,
"loss": 0.4045,
"step": 1008
},
{
"epoch": 0.5098534613441131,
"grad_norm": 0.5892174243927002,
"learning_rate": 3.473043464450027e-05,
"loss": 0.3431,
"step": 1009
},
{
"epoch": 0.5103587670540677,
"grad_norm": 0.5369464755058289,
"learning_rate": 3.470143029229996e-05,
"loss": 0.2737,
"step": 1010
},
{
"epoch": 0.5108640727640222,
"grad_norm": 0.39442282915115356,
"learning_rate": 3.4672410558753155e-05,
"loss": 0.2475,
"step": 1011
},
{
"epoch": 0.5113693784739768,
"grad_norm": 0.4975070059299469,
"learning_rate": 3.464337548986983e-05,
"loss": 0.2419,
"step": 1012
},
{
"epoch": 0.5118746841839312,
"grad_norm": 0.6060254573822021,
"learning_rate": 3.4614325131684275e-05,
"loss": 0.387,
"step": 1013
},
{
"epoch": 0.5123799898938858,
"grad_norm": 0.3599434196949005,
"learning_rate": 3.458525953025503e-05,
"loss": 0.1805,
"step": 1014
},
{
"epoch": 0.5128852956038403,
"grad_norm": 0.457060843706131,
"learning_rate": 3.4556178731664776e-05,
"loss": 0.2104,
"step": 1015
},
{
"epoch": 0.5133906013137949,
"grad_norm": 0.6781561970710754,
"learning_rate": 3.4527082782020323e-05,
"loss": 0.4543,
"step": 1016
},
{
"epoch": 0.5138959070237493,
"grad_norm": 0.4415772259235382,
"learning_rate": 3.449797172745247e-05,
"loss": 0.2245,
"step": 1017
},
{
"epoch": 0.5144012127337039,
"grad_norm": 0.3077252507209778,
"learning_rate": 3.4468845614116e-05,
"loss": 0.2002,
"step": 1018
},
{
"epoch": 0.5149065184436584,
"grad_norm": 0.7845990657806396,
"learning_rate": 3.443970448818954e-05,
"loss": 0.5271,
"step": 1019
},
{
"epoch": 0.515411824153613,
"grad_norm": 0.44302085041999817,
"learning_rate": 3.441054839587553e-05,
"loss": 0.1878,
"step": 1020
},
{
"epoch": 0.5159171298635674,
"grad_norm": 1.1317225694656372,
"learning_rate": 3.438137738340014e-05,
"loss": 0.4688,
"step": 1021
},
{
"epoch": 0.516422435573522,
"grad_norm": 0.7646364569664001,
"learning_rate": 3.435219149701319e-05,
"loss": 0.4279,
"step": 1022
},
{
"epoch": 0.5169277412834765,
"grad_norm": 0.3174137771129608,
"learning_rate": 3.43229907829881e-05,
"loss": 0.207,
"step": 1023
},
{
"epoch": 0.5174330469934311,
"grad_norm": 0.25344839692115784,
"learning_rate": 3.429377528762177e-05,
"loss": 0.174,
"step": 1024
},
{
"epoch": 0.5179383527033855,
"grad_norm": 0.4643212854862213,
"learning_rate": 3.4264545057234565e-05,
"loss": 0.3102,
"step": 1025
},
{
"epoch": 0.5184436584133401,
"grad_norm": 0.3090815842151642,
"learning_rate": 3.423530013817019e-05,
"loss": 0.1688,
"step": 1026
},
{
"epoch": 0.5189489641232946,
"grad_norm": 0.7463130950927734,
"learning_rate": 3.4206040576795644e-05,
"loss": 0.6639,
"step": 1027
},
{
"epoch": 0.5194542698332492,
"grad_norm": 0.4260081648826599,
"learning_rate": 3.417676641950115e-05,
"loss": 0.1466,
"step": 1028
},
{
"epoch": 0.5199595755432036,
"grad_norm": 0.8411809802055359,
"learning_rate": 3.414747771270006e-05,
"loss": 0.5472,
"step": 1029
},
{
"epoch": 0.5204648812531582,
"grad_norm": 0.590641438961029,
"learning_rate": 3.411817450282881e-05,
"loss": 0.402,
"step": 1030
},
{
"epoch": 0.5209701869631127,
"grad_norm": 0.9213524460792542,
"learning_rate": 3.408885683634681e-05,
"loss": 0.5014,
"step": 1031
},
{
"epoch": 0.5214754926730673,
"grad_norm": 0.3281521201133728,
"learning_rate": 3.4059524759736404e-05,
"loss": 0.2153,
"step": 1032
},
{
"epoch": 0.5219807983830217,
"grad_norm": 0.4002426266670227,
"learning_rate": 3.4030178319502784e-05,
"loss": 0.2112,
"step": 1033
},
{
"epoch": 0.5224861040929762,
"grad_norm": 0.27855736017227173,
"learning_rate": 3.40008175621739e-05,
"loss": 0.1904,
"step": 1034
},
{
"epoch": 0.5229914098029308,
"grad_norm": 1.192247986793518,
"learning_rate": 3.3971442534300414e-05,
"loss": 0.5124,
"step": 1035
},
{
"epoch": 0.5234967155128853,
"grad_norm": 0.3181244730949402,
"learning_rate": 3.3942053282455625e-05,
"loss": 0.2413,
"step": 1036
},
{
"epoch": 0.5240020212228398,
"grad_norm": 0.5175408720970154,
"learning_rate": 3.3912649853235355e-05,
"loss": 0.2505,
"step": 1037
},
{
"epoch": 0.5245073269327943,
"grad_norm": 0.7706686854362488,
"learning_rate": 3.3883232293257926e-05,
"loss": 0.4528,
"step": 1038
},
{
"epoch": 0.5250126326427489,
"grad_norm": 0.6226242184638977,
"learning_rate": 3.385380064916405e-05,
"loss": 0.2895,
"step": 1039
},
{
"epoch": 0.5255179383527034,
"grad_norm": 0.6919043064117432,
"learning_rate": 3.382435496761679e-05,
"loss": 0.5499,
"step": 1040
},
{
"epoch": 0.5260232440626579,
"grad_norm": 0.7153246402740479,
"learning_rate": 3.379489529530144e-05,
"loss": 0.4498,
"step": 1041
},
{
"epoch": 0.5265285497726124,
"grad_norm": 0.3823365271091461,
"learning_rate": 3.376542167892548e-05,
"loss": 0.3495,
"step": 1042
},
{
"epoch": 0.527033855482567,
"grad_norm": 0.7256558537483215,
"learning_rate": 3.373593416521852e-05,
"loss": 0.4197,
"step": 1043
},
{
"epoch": 0.5275391611925214,
"grad_norm": 0.6454598903656006,
"learning_rate": 3.3706432800932184e-05,
"loss": 0.4144,
"step": 1044
},
{
"epoch": 0.528044466902476,
"grad_norm": 0.3100246489048004,
"learning_rate": 3.367691763284005e-05,
"loss": 0.1587,
"step": 1045
},
{
"epoch": 0.5285497726124305,
"grad_norm": 0.44669437408447266,
"learning_rate": 3.3647388707737606e-05,
"loss": 0.2715,
"step": 1046
},
{
"epoch": 0.5290550783223851,
"grad_norm": 0.31795749068260193,
"learning_rate": 3.3617846072442135e-05,
"loss": 0.2045,
"step": 1047
},
{
"epoch": 0.5295603840323395,
"grad_norm": 0.34143051505088806,
"learning_rate": 3.3588289773792646e-05,
"loss": 0.2189,
"step": 1048
},
{
"epoch": 0.5300656897422941,
"grad_norm": 0.5310137271881104,
"learning_rate": 3.3558719858649834e-05,
"loss": 0.2596,
"step": 1049
},
{
"epoch": 0.5305709954522486,
"grad_norm": 0.7323408722877502,
"learning_rate": 3.352913637389598e-05,
"loss": 0.3134,
"step": 1050
},
{
"epoch": 0.5310763011622032,
"grad_norm": 0.3827390968799591,
"learning_rate": 3.349953936643486e-05,
"loss": 0.2751,
"step": 1051
},
{
"epoch": 0.5315816068721576,
"grad_norm": 0.701563835144043,
"learning_rate": 3.3469928883191706e-05,
"loss": 0.3593,
"step": 1052
},
{
"epoch": 0.5320869125821122,
"grad_norm": 0.68122398853302,
"learning_rate": 3.344030497111312e-05,
"loss": 0.6736,
"step": 1053
},
{
"epoch": 0.5325922182920667,
"grad_norm": 0.47403842210769653,
"learning_rate": 3.3410667677166966e-05,
"loss": 0.3286,
"step": 1054
},
{
"epoch": 0.5330975240020213,
"grad_norm": 0.24394340813159943,
"learning_rate": 3.338101704834236e-05,
"loss": 0.161,
"step": 1055
},
{
"epoch": 0.5336028297119757,
"grad_norm": 0.6971885561943054,
"learning_rate": 3.3351353131649546e-05,
"loss": 0.2709,
"step": 1056
},
{
"epoch": 0.5341081354219303,
"grad_norm": 0.7203572988510132,
"learning_rate": 3.332167597411983e-05,
"loss": 0.3845,
"step": 1057
},
{
"epoch": 0.5346134411318848,
"grad_norm": 0.8475279211997986,
"learning_rate": 3.3291985622805516e-05,
"loss": 0.436,
"step": 1058
},
{
"epoch": 0.5351187468418394,
"grad_norm": 0.46801191568374634,
"learning_rate": 3.326228212477982e-05,
"loss": 0.2384,
"step": 1059
},
{
"epoch": 0.5356240525517938,
"grad_norm": 0.4841732680797577,
"learning_rate": 3.3232565527136815e-05,
"loss": 0.3124,
"step": 1060
},
{
"epoch": 0.5361293582617483,
"grad_norm": 0.2917931377887726,
"learning_rate": 3.3202835876991346e-05,
"loss": 0.178,
"step": 1061
},
{
"epoch": 0.5366346639717029,
"grad_norm": 0.5136047005653381,
"learning_rate": 3.317309322147891e-05,
"loss": 0.2269,
"step": 1062
},
{
"epoch": 0.5371399696816574,
"grad_norm": 0.504074215888977,
"learning_rate": 3.3143337607755694e-05,
"loss": 0.2895,
"step": 1063
},
{
"epoch": 0.5376452753916119,
"grad_norm": 0.44980528950691223,
"learning_rate": 3.311356908299836e-05,
"loss": 0.346,
"step": 1064
},
{
"epoch": 0.5381505811015664,
"grad_norm": 0.45925334095954895,
"learning_rate": 3.30837876944041e-05,
"loss": 0.2025,
"step": 1065
},
{
"epoch": 0.538655886811521,
"grad_norm": 0.48478057980537415,
"learning_rate": 3.305399348919045e-05,
"loss": 0.3748,
"step": 1066
},
{
"epoch": 0.5391611925214755,
"grad_norm": 0.376342236995697,
"learning_rate": 3.3024186514595294e-05,
"loss": 0.2637,
"step": 1067
},
{
"epoch": 0.53966649823143,
"grad_norm": 0.8223666548728943,
"learning_rate": 3.299436681787677e-05,
"loss": 0.3113,
"step": 1068
},
{
"epoch": 0.5401718039413845,
"grad_norm": 0.3819112181663513,
"learning_rate": 3.296453444631316e-05,
"loss": 0.2424,
"step": 1069
},
{
"epoch": 0.5406771096513391,
"grad_norm": 0.5420382022857666,
"learning_rate": 3.293468944720287e-05,
"loss": 0.2229,
"step": 1070
},
{
"epoch": 0.5411824153612936,
"grad_norm": 0.497595876455307,
"learning_rate": 3.290483186786428e-05,
"loss": 0.2529,
"step": 1071
},
{
"epoch": 0.5416877210712481,
"grad_norm": 0.7739394307136536,
"learning_rate": 3.287496175563578e-05,
"loss": 0.3402,
"step": 1072
},
{
"epoch": 0.5421930267812026,
"grad_norm": 0.5213684439659119,
"learning_rate": 3.284507915787559e-05,
"loss": 0.2469,
"step": 1073
},
{
"epoch": 0.5426983324911572,
"grad_norm": 0.45709890127182007,
"learning_rate": 3.281518412196172e-05,
"loss": 0.446,
"step": 1074
},
{
"epoch": 0.5432036382011117,
"grad_norm": 0.47826409339904785,
"learning_rate": 3.2785276695291935e-05,
"loss": 0.3853,
"step": 1075
},
{
"epoch": 0.5437089439110662,
"grad_norm": 0.5849752426147461,
"learning_rate": 3.27553569252836e-05,
"loss": 0.3097,
"step": 1076
},
{
"epoch": 0.5442142496210207,
"grad_norm": 0.3121519982814789,
"learning_rate": 3.272542485937369e-05,
"loss": 0.1677,
"step": 1077
},
{
"epoch": 0.5447195553309753,
"grad_norm": 0.5105435252189636,
"learning_rate": 3.269548054501865e-05,
"loss": 0.2058,
"step": 1078
},
{
"epoch": 0.5452248610409297,
"grad_norm": 0.5398941040039062,
"learning_rate": 3.2665524029694366e-05,
"loss": 0.3785,
"step": 1079
},
{
"epoch": 0.5457301667508843,
"grad_norm": 0.5803938508033752,
"learning_rate": 3.263555536089604e-05,
"loss": 0.4135,
"step": 1080
},
{
"epoch": 0.5462354724608388,
"grad_norm": 0.4731886088848114,
"learning_rate": 3.260557458613817e-05,
"loss": 0.3249,
"step": 1081
},
{
"epoch": 0.5467407781707934,
"grad_norm": 0.29429805278778076,
"learning_rate": 3.257558175295443e-05,
"loss": 0.1736,
"step": 1082
},
{
"epoch": 0.5472460838807478,
"grad_norm": 0.4236926734447479,
"learning_rate": 3.254557690889762e-05,
"loss": 0.2494,
"step": 1083
},
{
"epoch": 0.5477513895907024,
"grad_norm": 0.5632703900337219,
"learning_rate": 3.251556010153958e-05,
"loss": 0.3988,
"step": 1084
},
{
"epoch": 0.5482566953006569,
"grad_norm": 0.590407133102417,
"learning_rate": 3.248553137847112e-05,
"loss": 0.3471,
"step": 1085
},
{
"epoch": 0.5487620010106115,
"grad_norm": 0.5127763748168945,
"learning_rate": 3.245549078730195e-05,
"loss": 0.3066,
"step": 1086
},
{
"epoch": 0.5492673067205659,
"grad_norm": 0.5116364359855652,
"learning_rate": 3.242543837566057e-05,
"loss": 0.5509,
"step": 1087
},
{
"epoch": 0.5497726124305204,
"grad_norm": 0.40164387226104736,
"learning_rate": 3.239537419119425e-05,
"loss": 0.1977,
"step": 1088
},
{
"epoch": 0.550277918140475,
"grad_norm": 0.646152675151825,
"learning_rate": 3.236529828156891e-05,
"loss": 0.4706,
"step": 1089
},
{
"epoch": 0.5507832238504295,
"grad_norm": 0.5561092495918274,
"learning_rate": 3.233521069446907e-05,
"loss": 0.1772,
"step": 1090
},
{
"epoch": 0.551288529560384,
"grad_norm": 0.5761879682540894,
"learning_rate": 3.230511147759775e-05,
"loss": 0.2085,
"step": 1091
},
{
"epoch": 0.5517938352703385,
"grad_norm": 0.4727552533149719,
"learning_rate": 3.227500067867642e-05,
"loss": 0.2691,
"step": 1092
},
{
"epoch": 0.5522991409802931,
"grad_norm": 0.9213289618492126,
"learning_rate": 3.2244878345444905e-05,
"loss": 0.4643,
"step": 1093
},
{
"epoch": 0.5528044466902476,
"grad_norm": 0.8000922203063965,
"learning_rate": 3.221474452566133e-05,
"loss": 0.4779,
"step": 1094
},
{
"epoch": 0.5533097524002021,
"grad_norm": 0.3150307238101959,
"learning_rate": 3.218459926710202e-05,
"loss": 0.1698,
"step": 1095
},
{
"epoch": 0.5538150581101566,
"grad_norm": 0.5562626123428345,
"learning_rate": 3.215444261756144e-05,
"loss": 0.2622,
"step": 1096
},
{
"epoch": 0.5543203638201112,
"grad_norm": 0.576187014579773,
"learning_rate": 3.212427462485212e-05,
"loss": 0.2248,
"step": 1097
},
{
"epoch": 0.5548256695300657,
"grad_norm": 0.4368909001350403,
"learning_rate": 3.209409533680457e-05,
"loss": 0.2079,
"step": 1098
},
{
"epoch": 0.5553309752400202,
"grad_norm": 0.5242764353752136,
"learning_rate": 3.206390480126718e-05,
"loss": 0.3002,
"step": 1099
},
{
"epoch": 0.5558362809499747,
"grad_norm": 0.5973017811775208,
"learning_rate": 3.203370306610624e-05,
"loss": 0.2916,
"step": 1100
},
{
"epoch": 0.5563415866599293,
"grad_norm": 0.5519207119941711,
"learning_rate": 3.200349017920573e-05,
"loss": 0.2784,
"step": 1101
},
{
"epoch": 0.5568468923698838,
"grad_norm": 0.5289061069488525,
"learning_rate": 3.1973266188467346e-05,
"loss": 0.2497,
"step": 1102
},
{
"epoch": 0.5573521980798383,
"grad_norm": 0.5926082134246826,
"learning_rate": 3.1943031141810385e-05,
"loss": 0.3212,
"step": 1103
},
{
"epoch": 0.5578575037897928,
"grad_norm": 0.5896833539009094,
"learning_rate": 3.191278508717166e-05,
"loss": 0.3915,
"step": 1104
},
{
"epoch": 0.5583628094997474,
"grad_norm": 0.6870902180671692,
"learning_rate": 3.188252807250545e-05,
"loss": 0.3646,
"step": 1105
},
{
"epoch": 0.5588681152097019,
"grad_norm": 0.3682582676410675,
"learning_rate": 3.18522601457834e-05,
"loss": 0.2704,
"step": 1106
},
{
"epoch": 0.5593734209196564,
"grad_norm": 0.47527265548706055,
"learning_rate": 3.182198135499447e-05,
"loss": 0.2499,
"step": 1107
},
{
"epoch": 0.5598787266296109,
"grad_norm": 0.4023267924785614,
"learning_rate": 3.179169174814483e-05,
"loss": 0.1694,
"step": 1108
},
{
"epoch": 0.5603840323395655,
"grad_norm": 0.6233897805213928,
"learning_rate": 3.176139137325781e-05,
"loss": 0.4744,
"step": 1109
},
{
"epoch": 0.56088933804952,
"grad_norm": 0.2687748372554779,
"learning_rate": 3.173108027837379e-05,
"loss": 0.1979,
"step": 1110
},
{
"epoch": 0.5613946437594745,
"grad_norm": 0.4329420328140259,
"learning_rate": 3.1700758511550185e-05,
"loss": 0.2352,
"step": 1111
},
{
"epoch": 0.561899949469429,
"grad_norm": 0.36293160915374756,
"learning_rate": 3.1670426120861314e-05,
"loss": 0.2382,
"step": 1112
},
{
"epoch": 0.5624052551793836,
"grad_norm": 0.6872026324272156,
"learning_rate": 3.1640083154398306e-05,
"loss": 0.3837,
"step": 1113
},
{
"epoch": 0.562910560889338,
"grad_norm": 0.346775621175766,
"learning_rate": 3.160972966026911e-05,
"loss": 0.2396,
"step": 1114
},
{
"epoch": 0.5634158665992925,
"grad_norm": 0.6277110576629639,
"learning_rate": 3.1579365686598336e-05,
"loss": 0.3142,
"step": 1115
},
{
"epoch": 0.5639211723092471,
"grad_norm": 0.7550572156906128,
"learning_rate": 3.15489912815272e-05,
"loss": 0.2231,
"step": 1116
},
{
"epoch": 0.5644264780192016,
"grad_norm": 0.2056116759777069,
"learning_rate": 3.15186064932135e-05,
"loss": 0.1063,
"step": 1117
},
{
"epoch": 0.5649317837291561,
"grad_norm": 0.7212878465652466,
"learning_rate": 3.148821136983144e-05,
"loss": 0.3422,
"step": 1118
},
{
"epoch": 0.5654370894391106,
"grad_norm": 0.2987580895423889,
"learning_rate": 3.1457805959571665e-05,
"loss": 0.2378,
"step": 1119
},
{
"epoch": 0.5659423951490652,
"grad_norm": 0.45587587356567383,
"learning_rate": 3.142739031064108e-05,
"loss": 0.2602,
"step": 1120
},
{
"epoch": 0.5664477008590197,
"grad_norm": 0.943342387676239,
"learning_rate": 3.139696447126286e-05,
"loss": 0.3641,
"step": 1121
},
{
"epoch": 0.5669530065689742,
"grad_norm": 0.8414391279220581,
"learning_rate": 3.1366528489676315e-05,
"loss": 0.5864,
"step": 1122
},
{
"epoch": 0.5674583122789287,
"grad_norm": 0.5744806528091431,
"learning_rate": 3.133608241413685e-05,
"loss": 0.3196,
"step": 1123
},
{
"epoch": 0.5679636179888833,
"grad_norm": 0.7059041857719421,
"learning_rate": 3.130562629291586e-05,
"loss": 0.4094,
"step": 1124
},
{
"epoch": 0.5684689236988378,
"grad_norm": 0.4427710175514221,
"learning_rate": 3.127516017430067e-05,
"loss": 0.2116,
"step": 1125
},
{
"epoch": 0.5689742294087923,
"grad_norm": 0.8828384280204773,
"learning_rate": 3.124468410659448e-05,
"loss": 0.3982,
"step": 1126
},
{
"epoch": 0.5694795351187468,
"grad_norm": 0.9311108589172363,
"learning_rate": 3.121419813811622e-05,
"loss": 0.4192,
"step": 1127
},
{
"epoch": 0.5699848408287014,
"grad_norm": 0.8841239809989929,
"learning_rate": 3.118370231720055e-05,
"loss": 0.4229,
"step": 1128
},
{
"epoch": 0.5704901465386559,
"grad_norm": 0.549286425113678,
"learning_rate": 3.1153196692197745e-05,
"loss": 0.3589,
"step": 1129
},
{
"epoch": 0.5709954522486104,
"grad_norm": 0.4687441885471344,
"learning_rate": 3.112268131147361e-05,
"loss": 0.2813,
"step": 1130
},
{
"epoch": 0.5715007579585649,
"grad_norm": 0.5466412305831909,
"learning_rate": 3.109215622340944e-05,
"loss": 0.2345,
"step": 1131
},
{
"epoch": 0.5720060636685195,
"grad_norm": 0.39715632796287537,
"learning_rate": 3.10616214764019e-05,
"loss": 0.2707,
"step": 1132
},
{
"epoch": 0.572511369378474,
"grad_norm": 0.3774857521057129,
"learning_rate": 3.103107711886299e-05,
"loss": 0.3072,
"step": 1133
},
{
"epoch": 0.5730166750884285,
"grad_norm": 0.5280380845069885,
"learning_rate": 3.1000523199219925e-05,
"loss": 0.2819,
"step": 1134
},
{
"epoch": 0.573521980798383,
"grad_norm": 0.4964953064918518,
"learning_rate": 3.096995976591508e-05,
"loss": 0.2877,
"step": 1135
},
{
"epoch": 0.5740272865083376,
"grad_norm": 0.4315342903137207,
"learning_rate": 3.093938686740594e-05,
"loss": 0.2174,
"step": 1136
},
{
"epoch": 0.5745325922182921,
"grad_norm": 0.5171803832054138,
"learning_rate": 3.0908804552164985e-05,
"loss": 0.4131,
"step": 1137
},
{
"epoch": 0.5750378979282466,
"grad_norm": 0.43991976976394653,
"learning_rate": 3.08782128686796e-05,
"loss": 0.2688,
"step": 1138
},
{
"epoch": 0.5755432036382011,
"grad_norm": 0.45389074087142944,
"learning_rate": 3.084761186545206e-05,
"loss": 0.3958,
"step": 1139
},
{
"epoch": 0.5760485093481557,
"grad_norm": 0.4297245740890503,
"learning_rate": 3.081700159099939e-05,
"loss": 0.2264,
"step": 1140
},
{
"epoch": 0.5765538150581102,
"grad_norm": 0.6526983380317688,
"learning_rate": 3.0786382093853336e-05,
"loss": 0.5204,
"step": 1141
},
{
"epoch": 0.5770591207680646,
"grad_norm": 0.46042600274086,
"learning_rate": 3.0755753422560254e-05,
"loss": 0.2465,
"step": 1142
},
{
"epoch": 0.5775644264780192,
"grad_norm": 0.8101718425750732,
"learning_rate": 3.0725115625681035e-05,
"loss": 0.6224,
"step": 1143
},
{
"epoch": 0.5780697321879738,
"grad_norm": 0.6587077379226685,
"learning_rate": 3.069446875179106e-05,
"loss": 0.5301,
"step": 1144
},
{
"epoch": 0.5785750378979283,
"grad_norm": 0.49807339906692505,
"learning_rate": 3.0663812849480075e-05,
"loss": 0.2905,
"step": 1145
},
{
"epoch": 0.5790803436078827,
"grad_norm": 0.514011800289154,
"learning_rate": 3.063314796735218e-05,
"loss": 0.2377,
"step": 1146
},
{
"epoch": 0.5795856493178373,
"grad_norm": 0.5293000340461731,
"learning_rate": 3.060247415402567e-05,
"loss": 0.3341,
"step": 1147
},
{
"epoch": 0.5800909550277918,
"grad_norm": 0.705506443977356,
"learning_rate": 3.057179145813302e-05,
"loss": 0.3912,
"step": 1148
},
{
"epoch": 0.5805962607377463,
"grad_norm": 0.6125894784927368,
"learning_rate": 3.0541099928320804e-05,
"loss": 0.2844,
"step": 1149
},
{
"epoch": 0.5811015664477008,
"grad_norm": 0.6330406069755554,
"learning_rate": 3.0510399613249556e-05,
"loss": 0.3338,
"step": 1150
},
{
"epoch": 0.5816068721576554,
"grad_norm": 0.6614536046981812,
"learning_rate": 3.0479690561593793e-05,
"loss": 0.2598,
"step": 1151
},
{
"epoch": 0.5821121778676099,
"grad_norm": 0.37762120366096497,
"learning_rate": 3.0448972822041843e-05,
"loss": 0.2355,
"step": 1152
},
{
"epoch": 0.5826174835775644,
"grad_norm": 0.6358450055122375,
"learning_rate": 3.0418246443295823e-05,
"loss": 0.2811,
"step": 1153
},
{
"epoch": 0.5831227892875189,
"grad_norm": 0.3570258617401123,
"learning_rate": 3.0387511474071557e-05,
"loss": 0.172,
"step": 1154
},
{
"epoch": 0.5836280949974735,
"grad_norm": 0.3080599308013916,
"learning_rate": 3.0356767963098464e-05,
"loss": 0.166,
"step": 1155
},
{
"epoch": 0.584133400707428,
"grad_norm": 0.3466140031814575,
"learning_rate": 3.0326015959119535e-05,
"loss": 0.1429,
"step": 1156
},
{
"epoch": 0.5846387064173825,
"grad_norm": 0.4894466996192932,
"learning_rate": 3.0295255510891213e-05,
"loss": 0.4822,
"step": 1157
},
{
"epoch": 0.585144012127337,
"grad_norm": 0.4862903952598572,
"learning_rate": 3.0264486667183322e-05,
"loss": 0.2294,
"step": 1158
},
{
"epoch": 0.5856493178372916,
"grad_norm": 0.4425124228000641,
"learning_rate": 3.023370947677901e-05,
"loss": 0.235,
"step": 1159
},
{
"epoch": 0.5861546235472461,
"grad_norm": 0.4671436846256256,
"learning_rate": 3.020292398847464e-05,
"loss": 0.2426,
"step": 1160
},
{
"epoch": 0.5866599292572006,
"grad_norm": 1.0623973608016968,
"learning_rate": 3.0172130251079757e-05,
"loss": 0.5149,
"step": 1161
},
{
"epoch": 0.5871652349671551,
"grad_norm": 0.6807923913002014,
"learning_rate": 3.014132831341696e-05,
"loss": 0.3281,
"step": 1162
},
{
"epoch": 0.5876705406771097,
"grad_norm": 0.3466711640357971,
"learning_rate": 3.0110518224321865e-05,
"loss": 0.1801,
"step": 1163
},
{
"epoch": 0.5881758463870642,
"grad_norm": 0.6598278284072876,
"learning_rate": 3.007970003264301e-05,
"loss": 0.3268,
"step": 1164
},
{
"epoch": 0.5886811520970187,
"grad_norm": 0.5974249243736267,
"learning_rate": 3.0048873787241765e-05,
"loss": 0.2778,
"step": 1165
},
{
"epoch": 0.5891864578069732,
"grad_norm": 0.35448309779167175,
"learning_rate": 3.0018039536992298e-05,
"loss": 0.1175,
"step": 1166
},
{
"epoch": 0.5896917635169278,
"grad_norm": 0.6994403004646301,
"learning_rate": 2.9987197330781435e-05,
"loss": 0.2628,
"step": 1167
},
{
"epoch": 0.5901970692268823,
"grad_norm": 0.5090075731277466,
"learning_rate": 2.9956347217508633e-05,
"loss": 0.2562,
"step": 1168
},
{
"epoch": 0.5907023749368367,
"grad_norm": 0.29259657859802246,
"learning_rate": 2.9925489246085897e-05,
"loss": 0.1874,
"step": 1169
},
{
"epoch": 0.5912076806467913,
"grad_norm": 0.30072706937789917,
"learning_rate": 2.989462346543766e-05,
"loss": 0.1683,
"step": 1170
},
{
"epoch": 0.5917129863567459,
"grad_norm": 0.258619099855423,
"learning_rate": 2.9863749924500765e-05,
"loss": 0.1159,
"step": 1171
},
{
"epoch": 0.5922182920667004,
"grad_norm": 0.6549366116523743,
"learning_rate": 2.9832868672224347e-05,
"loss": 0.2978,
"step": 1172
},
{
"epoch": 0.5927235977766548,
"grad_norm": 0.9457119703292847,
"learning_rate": 2.9801979757569774e-05,
"loss": 0.3657,
"step": 1173
},
{
"epoch": 0.5932289034866094,
"grad_norm": 0.40999117493629456,
"learning_rate": 2.9771083229510544e-05,
"loss": 0.2076,
"step": 1174
},
{
"epoch": 0.593734209196564,
"grad_norm": 0.629391074180603,
"learning_rate": 2.974017913703224e-05,
"loss": 0.3435,
"step": 1175
},
{
"epoch": 0.5942395149065185,
"grad_norm": 0.42947840690612793,
"learning_rate": 2.970926752913245e-05,
"loss": 0.2741,
"step": 1176
},
{
"epoch": 0.5947448206164729,
"grad_norm": 0.7551778554916382,
"learning_rate": 2.9678348454820648e-05,
"loss": 0.2764,
"step": 1177
},
{
"epoch": 0.5952501263264275,
"grad_norm": 0.4660588800907135,
"learning_rate": 2.9647421963118177e-05,
"loss": 0.1889,
"step": 1178
},
{
"epoch": 0.595755432036382,
"grad_norm": 0.5166782140731812,
"learning_rate": 2.9616488103058116e-05,
"loss": 0.3303,
"step": 1179
},
{
"epoch": 0.5962607377463366,
"grad_norm": 0.6265774965286255,
"learning_rate": 2.9585546923685227e-05,
"loss": 0.2055,
"step": 1180
},
{
"epoch": 0.596766043456291,
"grad_norm": 0.4520041346549988,
"learning_rate": 2.9554598474055905e-05,
"loss": 0.2479,
"step": 1181
},
{
"epoch": 0.5972713491662456,
"grad_norm": 0.9512003064155579,
"learning_rate": 2.952364280323803e-05,
"loss": 0.4837,
"step": 1182
},
{
"epoch": 0.5977766548762001,
"grad_norm": 0.33608609437942505,
"learning_rate": 2.9492679960310975e-05,
"loss": 0.2544,
"step": 1183
},
{
"epoch": 0.5982819605861546,
"grad_norm": 0.49464377760887146,
"learning_rate": 2.9461709994365446e-05,
"loss": 0.1771,
"step": 1184
},
{
"epoch": 0.5987872662961091,
"grad_norm": 0.7408649921417236,
"learning_rate": 2.9430732954503454e-05,
"loss": 0.3388,
"step": 1185
},
{
"epoch": 0.5992925720060637,
"grad_norm": 0.6129613518714905,
"learning_rate": 2.9399748889838246e-05,
"loss": 0.3103,
"step": 1186
},
{
"epoch": 0.5997978777160182,
"grad_norm": 0.5622842311859131,
"learning_rate": 2.9368757849494183e-05,
"loss": 0.3837,
"step": 1187
},
{
"epoch": 0.6003031834259727,
"grad_norm": 0.7855788469314575,
"learning_rate": 2.9337759882606684e-05,
"loss": 0.3053,
"step": 1188
},
{
"epoch": 0.6008084891359272,
"grad_norm": 0.424996942281723,
"learning_rate": 2.930675503832217e-05,
"loss": 0.2624,
"step": 1189
},
{
"epoch": 0.6013137948458818,
"grad_norm": 0.748625636100769,
"learning_rate": 2.9275743365797954e-05,
"loss": 0.3661,
"step": 1190
},
{
"epoch": 0.6018191005558363,
"grad_norm": 0.47824960947036743,
"learning_rate": 2.9244724914202176e-05,
"loss": 0.2256,
"step": 1191
},
{
"epoch": 0.6023244062657908,
"grad_norm": 0.5377376079559326,
"learning_rate": 2.9213699732713717e-05,
"loss": 0.2619,
"step": 1192
},
{
"epoch": 0.6028297119757453,
"grad_norm": 0.5620574951171875,
"learning_rate": 2.9182667870522147e-05,
"loss": 0.3563,
"step": 1193
},
{
"epoch": 0.6033350176856999,
"grad_norm": 0.2533213198184967,
"learning_rate": 2.91516293768276e-05,
"loss": 0.1327,
"step": 1194
},
{
"epoch": 0.6038403233956544,
"grad_norm": 0.4888068437576294,
"learning_rate": 2.9120584300840746e-05,
"loss": 0.2132,
"step": 1195
},
{
"epoch": 0.6043456291056089,
"grad_norm": 0.357085645198822,
"learning_rate": 2.90895326917827e-05,
"loss": 0.1772,
"step": 1196
},
{
"epoch": 0.6048509348155634,
"grad_norm": 0.30169978737831116,
"learning_rate": 2.9058474598884893e-05,
"loss": 0.202,
"step": 1197
},
{
"epoch": 0.605356240525518,
"grad_norm": 0.3223772644996643,
"learning_rate": 2.902741007138909e-05,
"loss": 0.1932,
"step": 1198
},
{
"epoch": 0.6058615462354725,
"grad_norm": 0.42703449726104736,
"learning_rate": 2.899633915854721e-05,
"loss": 0.1927,
"step": 1199
},
{
"epoch": 0.6063668519454269,
"grad_norm": 0.39132627844810486,
"learning_rate": 2.896526190962132e-05,
"loss": 0.2409,
"step": 1200
},
{
"epoch": 0.6068721576553815,
"grad_norm": 0.6942405104637146,
"learning_rate": 2.8934178373883543e-05,
"loss": 0.4429,
"step": 1201
},
{
"epoch": 0.607377463365336,
"grad_norm": 0.6032945513725281,
"learning_rate": 2.890308860061594e-05,
"loss": 0.4111,
"step": 1202
},
{
"epoch": 0.6078827690752906,
"grad_norm": 0.679614782333374,
"learning_rate": 2.8871992639110484e-05,
"loss": 0.3658,
"step": 1203
},
{
"epoch": 0.608388074785245,
"grad_norm": 0.714556097984314,
"learning_rate": 2.8840890538668952e-05,
"loss": 0.3826,
"step": 1204
},
{
"epoch": 0.6088933804951996,
"grad_norm": 0.5705816745758057,
"learning_rate": 2.8809782348602864e-05,
"loss": 0.3086,
"step": 1205
},
{
"epoch": 0.6093986862051541,
"grad_norm": 0.7331538200378418,
"learning_rate": 2.877866811823337e-05,
"loss": 0.4354,
"step": 1206
},
{
"epoch": 0.6099039919151087,
"grad_norm": 0.5332103967666626,
"learning_rate": 2.8747547896891218e-05,
"loss": 0.254,
"step": 1207
},
{
"epoch": 0.6104092976250631,
"grad_norm": 0.3042926490306854,
"learning_rate": 2.871642173391666e-05,
"loss": 0.2578,
"step": 1208
},
{
"epoch": 0.6109146033350177,
"grad_norm": 0.5534253716468811,
"learning_rate": 2.868528967865934e-05,
"loss": 0.3656,
"step": 1209
},
{
"epoch": 0.6114199090449722,
"grad_norm": 0.6421956419944763,
"learning_rate": 2.8654151780478265e-05,
"loss": 0.2551,
"step": 1210
},
{
"epoch": 0.6119252147549268,
"grad_norm": 0.5455820560455322,
"learning_rate": 2.862300808874171e-05,
"loss": 0.3176,
"step": 1211
},
{
"epoch": 0.6124305204648812,
"grad_norm": 0.4438186585903168,
"learning_rate": 2.859185865282712e-05,
"loss": 0.2395,
"step": 1212
},
{
"epoch": 0.6129358261748358,
"grad_norm": 0.3606109917163849,
"learning_rate": 2.8560703522121064e-05,
"loss": 0.2739,
"step": 1213
},
{
"epoch": 0.6134411318847903,
"grad_norm": 0.4127027094364166,
"learning_rate": 2.8529542746019115e-05,
"loss": 0.2952,
"step": 1214
},
{
"epoch": 0.6139464375947449,
"grad_norm": 0.5099959969520569,
"learning_rate": 2.8498376373925834e-05,
"loss": 0.4231,
"step": 1215
},
{
"epoch": 0.6144517433046993,
"grad_norm": 0.68137127161026,
"learning_rate": 2.8467204455254614e-05,
"loss": 0.2562,
"step": 1216
},
{
"epoch": 0.6149570490146539,
"grad_norm": 0.8698126673698425,
"learning_rate": 2.8436027039427667e-05,
"loss": 0.4673,
"step": 1217
},
{
"epoch": 0.6154623547246084,
"grad_norm": 0.3290408253669739,
"learning_rate": 2.8404844175875912e-05,
"loss": 0.1761,
"step": 1218
},
{
"epoch": 0.6159676604345629,
"grad_norm": 0.6963320970535278,
"learning_rate": 2.8373655914038904e-05,
"loss": 0.3319,
"step": 1219
},
{
"epoch": 0.6164729661445174,
"grad_norm": 0.32454270124435425,
"learning_rate": 2.8342462303364777e-05,
"loss": 0.1979,
"step": 1220
},
{
"epoch": 0.616978271854472,
"grad_norm": 0.40933623909950256,
"learning_rate": 2.831126339331011e-05,
"loss": 0.1951,
"step": 1221
},
{
"epoch": 0.6174835775644265,
"grad_norm": 0.665077269077301,
"learning_rate": 2.8280059233339913e-05,
"loss": 0.5259,
"step": 1222
},
{
"epoch": 0.617988883274381,
"grad_norm": 0.48233264684677124,
"learning_rate": 2.8248849872927507e-05,
"loss": 0.2227,
"step": 1223
},
{
"epoch": 0.6184941889843355,
"grad_norm": 0.3461703062057495,
"learning_rate": 2.8217635361554456e-05,
"loss": 0.1607,
"step": 1224
},
{
"epoch": 0.61899949469429,
"grad_norm": 0.4308650493621826,
"learning_rate": 2.8186415748710504e-05,
"loss": 0.2759,
"step": 1225
},
{
"epoch": 0.6195048004042446,
"grad_norm": 0.8697786331176758,
"learning_rate": 2.8155191083893467e-05,
"loss": 0.4196,
"step": 1226
},
{
"epoch": 0.620010106114199,
"grad_norm": 0.5697680115699768,
"learning_rate": 2.8123961416609174e-05,
"loss": 0.3709,
"step": 1227
},
{
"epoch": 0.6205154118241536,
"grad_norm": 0.6422898769378662,
"learning_rate": 2.8092726796371406e-05,
"loss": 0.2012,
"step": 1228
},
{
"epoch": 0.6210207175341081,
"grad_norm": 0.271230012178421,
"learning_rate": 2.806148727270176e-05,
"loss": 0.136,
"step": 1229
},
{
"epoch": 0.6215260232440627,
"grad_norm": 0.3259721100330353,
"learning_rate": 2.8030242895129643e-05,
"loss": 0.239,
"step": 1230
},
{
"epoch": 0.6220313289540171,
"grad_norm": 0.49231839179992676,
"learning_rate": 2.7998993713192128e-05,
"loss": 0.2351,
"step": 1231
},
{
"epoch": 0.6225366346639717,
"grad_norm": 0.6262175440788269,
"learning_rate": 2.7967739776433933e-05,
"loss": 0.2764,
"step": 1232
},
{
"epoch": 0.6230419403739262,
"grad_norm": 0.4242885410785675,
"learning_rate": 2.79364811344073e-05,
"loss": 0.1978,
"step": 1233
},
{
"epoch": 0.6235472460838808,
"grad_norm": 0.5172401070594788,
"learning_rate": 2.7905217836671916e-05,
"loss": 0.3884,
"step": 1234
},
{
"epoch": 0.6240525517938352,
"grad_norm": 0.5422953963279724,
"learning_rate": 2.787394993279488e-05,
"loss": 0.318,
"step": 1235
},
{
"epoch": 0.6245578575037898,
"grad_norm": 0.8409627676010132,
"learning_rate": 2.7842677472350577e-05,
"loss": 0.3753,
"step": 1236
},
{
"epoch": 0.6250631632137443,
"grad_norm": 0.9036238789558411,
"learning_rate": 2.7811400504920622e-05,
"loss": 0.5445,
"step": 1237
},
{
"epoch": 0.6255684689236989,
"grad_norm": 0.5030844807624817,
"learning_rate": 2.778011908009376e-05,
"loss": 0.2084,
"step": 1238
},
{
"epoch": 0.6260737746336533,
"grad_norm": 0.544114351272583,
"learning_rate": 2.7748833247465827e-05,
"loss": 0.3824,
"step": 1239
},
{
"epoch": 0.6265790803436079,
"grad_norm": 0.5116641521453857,
"learning_rate": 2.7717543056639637e-05,
"loss": 0.4358,
"step": 1240
},
{
"epoch": 0.6270843860535624,
"grad_norm": 0.49113696813583374,
"learning_rate": 2.7686248557224904e-05,
"loss": 0.2644,
"step": 1241
},
{
"epoch": 0.627589691763517,
"grad_norm": 0.3701680600643158,
"learning_rate": 2.765494979883819e-05,
"loss": 0.1596,
"step": 1242
},
{
"epoch": 0.6280949974734714,
"grad_norm": 0.5240901708602905,
"learning_rate": 2.762364683110279e-05,
"loss": 0.3679,
"step": 1243
},
{
"epoch": 0.628600303183426,
"grad_norm": 0.49889007210731506,
"learning_rate": 2.7592339703648694e-05,
"loss": 0.6432,
"step": 1244
},
{
"epoch": 0.6291056088933805,
"grad_norm": 0.3745577037334442,
"learning_rate": 2.7561028466112482e-05,
"loss": 0.154,
"step": 1245
},
{
"epoch": 0.6296109146033351,
"grad_norm": 0.5030989050865173,
"learning_rate": 2.7529713168137233e-05,
"loss": 0.2932,
"step": 1246
},
{
"epoch": 0.6301162203132895,
"grad_norm": 1.0743045806884766,
"learning_rate": 2.7498393859372484e-05,
"loss": 0.2168,
"step": 1247
},
{
"epoch": 0.6306215260232441,
"grad_norm": 0.46885451674461365,
"learning_rate": 2.746707058947412e-05,
"loss": 0.2967,
"step": 1248
},
{
"epoch": 0.6311268317331986,
"grad_norm": 0.3801758885383606,
"learning_rate": 2.743574340810431e-05,
"loss": 0.2834,
"step": 1249
},
{
"epoch": 0.631632137443153,
"grad_norm": 0.700106143951416,
"learning_rate": 2.7404412364931427e-05,
"loss": 0.5024,
"step": 1250
},
{
"epoch": 0.6321374431531076,
"grad_norm": 0.8579809069633484,
"learning_rate": 2.7373077509629957e-05,
"loss": 0.313,
"step": 1251
},
{
"epoch": 0.6326427488630622,
"grad_norm": 0.3703233599662781,
"learning_rate": 2.7341738891880452e-05,
"loss": 0.2095,
"step": 1252
},
{
"epoch": 0.6331480545730167,
"grad_norm": 0.2917376160621643,
"learning_rate": 2.73103965613694e-05,
"loss": 0.1726,
"step": 1253
},
{
"epoch": 0.6336533602829711,
"grad_norm": 0.5064457058906555,
"learning_rate": 2.7279050567789194e-05,
"loss": 0.2515,
"step": 1254
},
{
"epoch": 0.6341586659929257,
"grad_norm": 0.47626200318336487,
"learning_rate": 2.7247700960838034e-05,
"loss": 0.4029,
"step": 1255
},
{
"epoch": 0.6346639717028802,
"grad_norm": 0.3616740107536316,
"learning_rate": 2.7216347790219843e-05,
"loss": 0.2887,
"step": 1256
},
{
"epoch": 0.6351692774128348,
"grad_norm": 0.2952771484851837,
"learning_rate": 2.7184991105644192e-05,
"loss": 0.2067,
"step": 1257
},
{
"epoch": 0.6356745831227892,
"grad_norm": 0.2758321166038513,
"learning_rate": 2.715363095682623e-05,
"loss": 0.1615,
"step": 1258
},
{
"epoch": 0.6361798888327438,
"grad_norm": 0.5239960551261902,
"learning_rate": 2.7122267393486607e-05,
"loss": 0.3267,
"step": 1259
},
{
"epoch": 0.6366851945426983,
"grad_norm": 0.4696899652481079,
"learning_rate": 2.7090900465351355e-05,
"loss": 0.3621,
"step": 1260
},
{
"epoch": 0.6371905002526529,
"grad_norm": 0.40326637029647827,
"learning_rate": 2.7059530222151873e-05,
"loss": 0.1739,
"step": 1261
},
{
"epoch": 0.6376958059626073,
"grad_norm": 0.5829006433486938,
"learning_rate": 2.7028156713624814e-05,
"loss": 0.3423,
"step": 1262
},
{
"epoch": 0.6382011116725619,
"grad_norm": 0.3605864644050598,
"learning_rate": 2.6996779989511985e-05,
"loss": 0.1829,
"step": 1263
},
{
"epoch": 0.6387064173825164,
"grad_norm": 0.38949474692344666,
"learning_rate": 2.6965400099560308e-05,
"loss": 0.1814,
"step": 1264
},
{
"epoch": 0.639211723092471,
"grad_norm": 0.835013747215271,
"learning_rate": 2.6934017093521717e-05,
"loss": 0.5331,
"step": 1265
},
{
"epoch": 0.6397170288024254,
"grad_norm": 0.3474362790584564,
"learning_rate": 2.6902631021153096e-05,
"loss": 0.1773,
"step": 1266
},
{
"epoch": 0.64022233451238,
"grad_norm": 0.7530576586723328,
"learning_rate": 2.6871241932216195e-05,
"loss": 0.6231,
"step": 1267
},
{
"epoch": 0.6407276402223345,
"grad_norm": 0.46543627977371216,
"learning_rate": 2.6839849876477518e-05,
"loss": 0.2082,
"step": 1268
},
{
"epoch": 0.6412329459322891,
"grad_norm": 0.6920028328895569,
"learning_rate": 2.6808454903708315e-05,
"loss": 0.3675,
"step": 1269
},
{
"epoch": 0.6417382516422435,
"grad_norm": 0.639060378074646,
"learning_rate": 2.6777057063684406e-05,
"loss": 0.4416,
"step": 1270
},
{
"epoch": 0.6422435573521981,
"grad_norm": 0.8826596736907959,
"learning_rate": 2.674565640618621e-05,
"loss": 0.3633,
"step": 1271
},
{
"epoch": 0.6427488630621526,
"grad_norm": 0.6439775824546814,
"learning_rate": 2.6714252980998595e-05,
"loss": 0.3737,
"step": 1272
},
{
"epoch": 0.6432541687721072,
"grad_norm": 0.3467325270175934,
"learning_rate": 2.6682846837910796e-05,
"loss": 0.1777,
"step": 1273
},
{
"epoch": 0.6437594744820616,
"grad_norm": 1.040595293045044,
"learning_rate": 2.6651438026716396e-05,
"loss": 0.4716,
"step": 1274
},
{
"epoch": 0.6442647801920162,
"grad_norm": 0.6756473779678345,
"learning_rate": 2.6620026597213167e-05,
"loss": 0.3154,
"step": 1275
},
{
"epoch": 0.6447700859019707,
"grad_norm": 0.6702762246131897,
"learning_rate": 2.658861259920306e-05,
"loss": 0.366,
"step": 1276
},
{
"epoch": 0.6452753916119253,
"grad_norm": 0.34285977482795715,
"learning_rate": 2.6557196082492105e-05,
"loss": 0.1714,
"step": 1277
},
{
"epoch": 0.6457806973218797,
"grad_norm": 0.5991812348365784,
"learning_rate": 2.6525777096890286e-05,
"loss": 0.3714,
"step": 1278
},
{
"epoch": 0.6462860030318343,
"grad_norm": 0.39086848497390747,
"learning_rate": 2.6494355692211535e-05,
"loss": 0.1841,
"step": 1279
},
{
"epoch": 0.6467913087417888,
"grad_norm": 0.3882160186767578,
"learning_rate": 2.6462931918273624e-05,
"loss": 0.2613,
"step": 1280
},
{
"epoch": 0.6472966144517434,
"grad_norm": 0.7259517908096313,
"learning_rate": 2.643150582489805e-05,
"loss": 0.3005,
"step": 1281
},
{
"epoch": 0.6478019201616978,
"grad_norm": 0.6918402314186096,
"learning_rate": 2.6400077461910013e-05,
"loss": 0.4195,
"step": 1282
},
{
"epoch": 0.6483072258716523,
"grad_norm": 0.6692409515380859,
"learning_rate": 2.6368646879138304e-05,
"loss": 0.4622,
"step": 1283
},
{
"epoch": 0.6488125315816069,
"grad_norm": 0.6194551587104797,
"learning_rate": 2.633721412641524e-05,
"loss": 0.4571,
"step": 1284
},
{
"epoch": 0.6493178372915613,
"grad_norm": 0.4386284351348877,
"learning_rate": 2.6305779253576556e-05,
"loss": 0.1687,
"step": 1285
},
{
"epoch": 0.6498231430015159,
"grad_norm": 0.43226057291030884,
"learning_rate": 2.6274342310461382e-05,
"loss": 0.3734,
"step": 1286
},
{
"epoch": 0.6503284487114704,
"grad_norm": 0.3043833076953888,
"learning_rate": 2.624290334691209e-05,
"loss": 0.1776,
"step": 1287
},
{
"epoch": 0.650833754421425,
"grad_norm": 0.4461258053779602,
"learning_rate": 2.6211462412774298e-05,
"loss": 0.2532,
"step": 1288
},
{
"epoch": 0.6513390601313794,
"grad_norm": 0.5978369116783142,
"learning_rate": 2.6180019557896725e-05,
"loss": 0.4937,
"step": 1289
},
{
"epoch": 0.651844365841334,
"grad_norm": 0.4691097140312195,
"learning_rate": 2.614857483213113e-05,
"loss": 0.3146,
"step": 1290
},
{
"epoch": 0.6523496715512885,
"grad_norm": 0.48737433552742004,
"learning_rate": 2.6117128285332248e-05,
"loss": 0.2174,
"step": 1291
},
{
"epoch": 0.6528549772612431,
"grad_norm": 0.692764937877655,
"learning_rate": 2.6085679967357707e-05,
"loss": 0.4083,
"step": 1292
},
{
"epoch": 0.6533602829711975,
"grad_norm": 0.6832363605499268,
"learning_rate": 2.6054229928067923e-05,
"loss": 0.7104,
"step": 1293
},
{
"epoch": 0.6538655886811521,
"grad_norm": 0.2708096504211426,
"learning_rate": 2.6022778217326077e-05,
"loss": 0.1944,
"step": 1294
},
{
"epoch": 0.6543708943911066,
"grad_norm": 0.2859404385089874,
"learning_rate": 2.5991324884997952e-05,
"loss": 0.1835,
"step": 1295
},
{
"epoch": 0.6548762001010612,
"grad_norm": 0.6921172142028809,
"learning_rate": 2.595986998095194e-05,
"loss": 0.3649,
"step": 1296
},
{
"epoch": 0.6553815058110156,
"grad_norm": 0.270704984664917,
"learning_rate": 2.5928413555058904e-05,
"loss": 0.1674,
"step": 1297
},
{
"epoch": 0.6558868115209702,
"grad_norm": 0.39965078234672546,
"learning_rate": 2.5896955657192142e-05,
"loss": 0.29,
"step": 1298
},
{
"epoch": 0.6563921172309247,
"grad_norm": 0.6206263303756714,
"learning_rate": 2.5865496337227256e-05,
"loss": 0.2713,
"step": 1299
},
{
"epoch": 0.6568974229408793,
"grad_norm": 0.4320612847805023,
"learning_rate": 2.583403564504213e-05,
"loss": 0.2148,
"step": 1300
},
{
"epoch": 0.6574027286508337,
"grad_norm": 0.4375782012939453,
"learning_rate": 2.5802573630516806e-05,
"loss": 0.481,
"step": 1301
},
{
"epoch": 0.6579080343607883,
"grad_norm": 0.40101978182792664,
"learning_rate": 2.5771110343533428e-05,
"loss": 0.2242,
"step": 1302
},
{
"epoch": 0.6584133400707428,
"grad_norm": 0.7650478482246399,
"learning_rate": 2.5739645833976154e-05,
"loss": 0.337,
"step": 1303
},
{
"epoch": 0.6589186457806974,
"grad_norm": 0.6947590708732605,
"learning_rate": 2.5708180151731103e-05,
"loss": 0.4345,
"step": 1304
},
{
"epoch": 0.6594239514906518,
"grad_norm": 0.42338618636131287,
"learning_rate": 2.567671334668621e-05,
"loss": 0.2952,
"step": 1305
},
{
"epoch": 0.6599292572006064,
"grad_norm": 0.748132586479187,
"learning_rate": 2.5645245468731234e-05,
"loss": 0.8043,
"step": 1306
},
{
"epoch": 0.6604345629105609,
"grad_norm": 0.40081891417503357,
"learning_rate": 2.5613776567757595e-05,
"loss": 0.4187,
"step": 1307
},
{
"epoch": 0.6609398686205155,
"grad_norm": 0.5815805792808533,
"learning_rate": 2.5582306693658374e-05,
"loss": 0.251,
"step": 1308
},
{
"epoch": 0.6614451743304699,
"grad_norm": 0.4176938533782959,
"learning_rate": 2.555083589632818e-05,
"loss": 0.2369,
"step": 1309
},
{
"epoch": 0.6619504800404244,
"grad_norm": 0.2541157007217407,
"learning_rate": 2.5519364225663055e-05,
"loss": 0.1507,
"step": 1310
},
{
"epoch": 0.662455785750379,
"grad_norm": 0.5441336631774902,
"learning_rate": 2.5487891731560482e-05,
"loss": 0.2521,
"step": 1311
},
{
"epoch": 0.6629610914603336,
"grad_norm": 0.4862014055252075,
"learning_rate": 2.54564184639192e-05,
"loss": 0.2766,
"step": 1312
},
{
"epoch": 0.663466397170288,
"grad_norm": 0.7492696046829224,
"learning_rate": 2.5424944472639216e-05,
"loss": 0.4431,
"step": 1313
},
{
"epoch": 0.6639717028802425,
"grad_norm": 0.45778077840805054,
"learning_rate": 2.5393469807621645e-05,
"loss": 0.2535,
"step": 1314
},
{
"epoch": 0.6644770085901971,
"grad_norm": 0.8111942410469055,
"learning_rate": 2.5361994518768695e-05,
"loss": 0.3406,
"step": 1315
},
{
"epoch": 0.6649823143001516,
"grad_norm": 0.48901939392089844,
"learning_rate": 2.5330518655983558e-05,
"loss": 0.2384,
"step": 1316
},
{
"epoch": 0.6654876200101061,
"grad_norm": 0.67337566614151,
"learning_rate": 2.529904226917033e-05,
"loss": 0.4046,
"step": 1317
},
{
"epoch": 0.6659929257200606,
"grad_norm": 0.5460554957389832,
"learning_rate": 2.5267565408233952e-05,
"loss": 0.286,
"step": 1318
},
{
"epoch": 0.6664982314300152,
"grad_norm": 0.34170493483543396,
"learning_rate": 2.5236088123080092e-05,
"loss": 0.2063,
"step": 1319
},
{
"epoch": 0.6670035371399696,
"grad_norm": 0.47587305307388306,
"learning_rate": 2.5204610463615115e-05,
"loss": 0.2657,
"step": 1320
},
{
"epoch": 0.6675088428499242,
"grad_norm": 0.4068621098995209,
"learning_rate": 2.5173132479745975e-05,
"loss": 0.3323,
"step": 1321
},
{
"epoch": 0.6680141485598787,
"grad_norm": 0.4757383167743683,
"learning_rate": 2.514165422138012e-05,
"loss": 0.3589,
"step": 1322
},
{
"epoch": 0.6685194542698333,
"grad_norm": 0.3725772202014923,
"learning_rate": 2.5110175738425462e-05,
"loss": 0.1664,
"step": 1323
},
{
"epoch": 0.6690247599797877,
"grad_norm": 0.6302722692489624,
"learning_rate": 2.507869708079025e-05,
"loss": 0.4999,
"step": 1324
},
{
"epoch": 0.6695300656897423,
"grad_norm": 0.44832542538642883,
"learning_rate": 2.5047218298383015e-05,
"loss": 0.3016,
"step": 1325
},
{
"epoch": 0.6700353713996968,
"grad_norm": 0.4571584463119507,
"learning_rate": 2.5015739441112484e-05,
"loss": 0.1797,
"step": 1326
},
{
"epoch": 0.6705406771096514,
"grad_norm": 0.6637290120124817,
"learning_rate": 2.4984260558887522e-05,
"loss": 0.5544,
"step": 1327
},
{
"epoch": 0.6710459828196058,
"grad_norm": 0.337741494178772,
"learning_rate": 2.495278170161699e-05,
"loss": 0.1652,
"step": 1328
},
{
"epoch": 0.6715512885295604,
"grad_norm": 0.6732521653175354,
"learning_rate": 2.4921302919209763e-05,
"loss": 0.5967,
"step": 1329
},
{
"epoch": 0.6720565942395149,
"grad_norm": 0.8650082945823669,
"learning_rate": 2.4889824261574547e-05,
"loss": 0.4235,
"step": 1330
},
{
"epoch": 0.6725618999494695,
"grad_norm": 0.7857891321182251,
"learning_rate": 2.4858345778619887e-05,
"loss": 0.5156,
"step": 1331
},
{
"epoch": 0.6730672056594239,
"grad_norm": 0.42698702216148376,
"learning_rate": 2.4826867520254034e-05,
"loss": 0.2091,
"step": 1332
},
{
"epoch": 0.6735725113693785,
"grad_norm": 0.32361331582069397,
"learning_rate": 2.4795389536384887e-05,
"loss": 0.1365,
"step": 1333
},
{
"epoch": 0.674077817079333,
"grad_norm": 0.7567768096923828,
"learning_rate": 2.4763911876919917e-05,
"loss": 0.4576,
"step": 1334
},
{
"epoch": 0.6745831227892876,
"grad_norm": 0.2475138008594513,
"learning_rate": 2.4732434591766057e-05,
"loss": 0.1624,
"step": 1335
},
{
"epoch": 0.675088428499242,
"grad_norm": 0.8214117288589478,
"learning_rate": 2.4700957730829675e-05,
"loss": 0.3756,
"step": 1336
},
{
"epoch": 0.6755937342091966,
"grad_norm": 0.6251058578491211,
"learning_rate": 2.4669481344016444e-05,
"loss": 0.4255,
"step": 1337
},
{
"epoch": 0.6760990399191511,
"grad_norm": 0.45437926054000854,
"learning_rate": 2.4638005481231304e-05,
"loss": 0.5109,
"step": 1338
},
{
"epoch": 0.6766043456291057,
"grad_norm": 0.7901365756988525,
"learning_rate": 2.4606530192378358e-05,
"loss": 0.4083,
"step": 1339
},
{
"epoch": 0.6771096513390601,
"grad_norm": 0.47076261043548584,
"learning_rate": 2.4575055527360793e-05,
"loss": 0.33,
"step": 1340
},
{
"epoch": 0.6776149570490146,
"grad_norm": 0.7028003334999084,
"learning_rate": 2.45435815360808e-05,
"loss": 0.4038,
"step": 1341
},
{
"epoch": 0.6781202627589692,
"grad_norm": 0.3465520441532135,
"learning_rate": 2.4512108268439524e-05,
"loss": 0.2881,
"step": 1342
},
{
"epoch": 0.6786255684689237,
"grad_norm": 0.6936810612678528,
"learning_rate": 2.4480635774336947e-05,
"loss": 0.4991,
"step": 1343
},
{
"epoch": 0.6791308741788782,
"grad_norm": 0.264642596244812,
"learning_rate": 2.4449164103671833e-05,
"loss": 0.1572,
"step": 1344
},
{
"epoch": 0.6796361798888327,
"grad_norm": 0.6691703200340271,
"learning_rate": 2.441769330634163e-05,
"loss": 0.4967,
"step": 1345
},
{
"epoch": 0.6801414855987873,
"grad_norm": 0.8646639585494995,
"learning_rate": 2.4386223432242407e-05,
"loss": 0.3253,
"step": 1346
},
{
"epoch": 0.6806467913087418,
"grad_norm": 0.618078887462616,
"learning_rate": 2.435475453126877e-05,
"loss": 0.1929,
"step": 1347
},
{
"epoch": 0.6811520970186963,
"grad_norm": 0.8526009917259216,
"learning_rate": 2.432328665331379e-05,
"loss": 0.4779,
"step": 1348
},
{
"epoch": 0.6816574027286508,
"grad_norm": 0.8699708580970764,
"learning_rate": 2.4291819848268906e-05,
"loss": 0.3817,
"step": 1349
},
{
"epoch": 0.6821627084386054,
"grad_norm": 0.4946329891681671,
"learning_rate": 2.4260354166023848e-05,
"loss": 0.2458,
"step": 1350
},
{
"epoch": 0.6826680141485599,
"grad_norm": 0.5863744020462036,
"learning_rate": 2.4228889656466578e-05,
"loss": 0.5891,
"step": 1351
},
{
"epoch": 0.6831733198585144,
"grad_norm": 0.5861146450042725,
"learning_rate": 2.41974263694832e-05,
"loss": 0.3156,
"step": 1352
},
{
"epoch": 0.6836786255684689,
"grad_norm": 0.590474009513855,
"learning_rate": 2.416596435495787e-05,
"loss": 0.3012,
"step": 1353
},
{
"epoch": 0.6841839312784235,
"grad_norm": 0.4421933591365814,
"learning_rate": 2.4134503662772753e-05,
"loss": 0.3528,
"step": 1354
},
{
"epoch": 0.6846892369883779,
"grad_norm": 0.3189122974872589,
"learning_rate": 2.4103044342807867e-05,
"loss": 0.1986,
"step": 1355
},
{
"epoch": 0.6851945426983325,
"grad_norm": 0.6435018181800842,
"learning_rate": 2.4071586444941098e-05,
"loss": 0.3458,
"step": 1356
},
{
"epoch": 0.685699848408287,
"grad_norm": 0.8088643550872803,
"learning_rate": 2.404013001904807e-05,
"loss": 0.5646,
"step": 1357
},
{
"epoch": 0.6862051541182416,
"grad_norm": 0.4556143879890442,
"learning_rate": 2.4008675115002054e-05,
"loss": 0.3147,
"step": 1358
},
{
"epoch": 0.686710459828196,
"grad_norm": 0.4295537769794464,
"learning_rate": 2.3977221782673933e-05,
"loss": 0.2078,
"step": 1359
},
{
"epoch": 0.6872157655381506,
"grad_norm": 0.5436996817588806,
"learning_rate": 2.394577007193208e-05,
"loss": 0.3047,
"step": 1360
},
{
"epoch": 0.6877210712481051,
"grad_norm": 0.6986174583435059,
"learning_rate": 2.39143200326423e-05,
"loss": 0.4117,
"step": 1361
},
{
"epoch": 0.6882263769580597,
"grad_norm": 0.5140484571456909,
"learning_rate": 2.3882871714667754e-05,
"loss": 0.2067,
"step": 1362
},
{
"epoch": 0.6887316826680141,
"grad_norm": 0.9210074543952942,
"learning_rate": 2.3851425167868874e-05,
"loss": 0.4738,
"step": 1363
},
{
"epoch": 0.6892369883779687,
"grad_norm": 0.5711868405342102,
"learning_rate": 2.3819980442103288e-05,
"loss": 0.2715,
"step": 1364
},
{
"epoch": 0.6897422940879232,
"grad_norm": 0.691596508026123,
"learning_rate": 2.3788537587225705e-05,
"loss": 0.4291,
"step": 1365
},
{
"epoch": 0.6902475997978778,
"grad_norm": 0.6653650999069214,
"learning_rate": 2.3757096653087914e-05,
"loss": 0.2889,
"step": 1366
},
{
"epoch": 0.6907529055078322,
"grad_norm": 0.5367448329925537,
"learning_rate": 2.3725657689538627e-05,
"loss": 0.3765,
"step": 1367
},
{
"epoch": 0.6912582112177867,
"grad_norm": 0.4669603705406189,
"learning_rate": 2.3694220746423447e-05,
"loss": 0.2137,
"step": 1368
},
{
"epoch": 0.6917635169277413,
"grad_norm": 0.5713473558425903,
"learning_rate": 2.3662785873584773e-05,
"loss": 0.3873,
"step": 1369
},
{
"epoch": 0.6922688226376958,
"grad_norm": 0.444577693939209,
"learning_rate": 2.36313531208617e-05,
"loss": 0.257,
"step": 1370
},
{
"epoch": 0.6927741283476503,
"grad_norm": 0.44812366366386414,
"learning_rate": 2.3599922538089993e-05,
"loss": 0.2485,
"step": 1371
},
{
"epoch": 0.6932794340576048,
"grad_norm": 0.39372605085372925,
"learning_rate": 2.3568494175101958e-05,
"loss": 0.2552,
"step": 1372
},
{
"epoch": 0.6937847397675594,
"grad_norm": 0.8668807744979858,
"learning_rate": 2.353706808172638e-05,
"loss": 0.4462,
"step": 1373
},
{
"epoch": 0.6942900454775139,
"grad_norm": 0.37225598096847534,
"learning_rate": 2.3505644307788467e-05,
"loss": 0.2313,
"step": 1374
},
{
"epoch": 0.6947953511874684,
"grad_norm": 0.2703179717063904,
"learning_rate": 2.3474222903109723e-05,
"loss": 0.1954,
"step": 1375
},
{
"epoch": 0.6953006568974229,
"grad_norm": 0.46796905994415283,
"learning_rate": 2.34428039175079e-05,
"loss": 0.2335,
"step": 1376
},
{
"epoch": 0.6958059626073775,
"grad_norm": 0.4360145926475525,
"learning_rate": 2.3411387400796934e-05,
"loss": 0.2707,
"step": 1377
},
{
"epoch": 0.696311268317332,
"grad_norm": 0.4478427767753601,
"learning_rate": 2.3379973402786832e-05,
"loss": 0.2095,
"step": 1378
},
{
"epoch": 0.6968165740272865,
"grad_norm": 0.544204831123352,
"learning_rate": 2.3348561973283613e-05,
"loss": 0.2305,
"step": 1379
},
{
"epoch": 0.697321879737241,
"grad_norm": 0.7342060208320618,
"learning_rate": 2.3317153162089206e-05,
"loss": 0.3866,
"step": 1380
},
{
"epoch": 0.6978271854471956,
"grad_norm": 0.33234989643096924,
"learning_rate": 2.328574701900141e-05,
"loss": 0.1789,
"step": 1381
},
{
"epoch": 0.6983324911571501,
"grad_norm": 0.541576623916626,
"learning_rate": 2.325434359381379e-05,
"loss": 0.2606,
"step": 1382
},
{
"epoch": 0.6988377968671046,
"grad_norm": 0.5520817041397095,
"learning_rate": 2.3222942936315606e-05,
"loss": 0.2925,
"step": 1383
},
{
"epoch": 0.6993431025770591,
"grad_norm": 0.5592846870422363,
"learning_rate": 2.31915450962917e-05,
"loss": 0.2995,
"step": 1384
},
{
"epoch": 0.6998484082870137,
"grad_norm": 0.18375256657600403,
"learning_rate": 2.3160150123522485e-05,
"loss": 0.0958,
"step": 1385
},
{
"epoch": 0.7003537139969682,
"grad_norm": 0.5363619923591614,
"learning_rate": 2.3128758067783808e-05,
"loss": 0.2511,
"step": 1386
},
{
"epoch": 0.7008590197069227,
"grad_norm": 0.4795052111148834,
"learning_rate": 2.30973689788469e-05,
"loss": 0.2263,
"step": 1387
},
{
"epoch": 0.7013643254168772,
"grad_norm": 0.6192831993103027,
"learning_rate": 2.306598290647829e-05,
"loss": 0.3247,
"step": 1388
},
{
"epoch": 0.7018696311268318,
"grad_norm": 0.4603809416294098,
"learning_rate": 2.30345999004397e-05,
"loss": 0.2794,
"step": 1389
},
{
"epoch": 0.7023749368367862,
"grad_norm": 0.5347074270248413,
"learning_rate": 2.300322001048802e-05,
"loss": 0.3136,
"step": 1390
},
{
"epoch": 0.7028802425467408,
"grad_norm": 0.6828907132148743,
"learning_rate": 2.2971843286375192e-05,
"loss": 0.2847,
"step": 1391
},
{
"epoch": 0.7033855482566953,
"grad_norm": 0.6386706233024597,
"learning_rate": 2.2940469777848123e-05,
"loss": 0.4388,
"step": 1392
},
{
"epoch": 0.7038908539666499,
"grad_norm": 0.4837972819805145,
"learning_rate": 2.2909099534648658e-05,
"loss": 0.2637,
"step": 1393
},
{
"epoch": 0.7043961596766043,
"grad_norm": 0.4311538338661194,
"learning_rate": 2.2877732606513406e-05,
"loss": 0.2201,
"step": 1394
},
{
"epoch": 0.7049014653865588,
"grad_norm": 0.7682383060455322,
"learning_rate": 2.2846369043173775e-05,
"loss": 0.3609,
"step": 1395
},
{
"epoch": 0.7054067710965134,
"grad_norm": 0.4595930576324463,
"learning_rate": 2.281500889435581e-05,
"loss": 0.3006,
"step": 1396
},
{
"epoch": 0.705912076806468,
"grad_norm": 0.45893973112106323,
"learning_rate": 2.278365220978016e-05,
"loss": 0.2403,
"step": 1397
},
{
"epoch": 0.7064173825164224,
"grad_norm": 0.24530524015426636,
"learning_rate": 2.2752299039161975e-05,
"loss": 0.1777,
"step": 1398
},
{
"epoch": 0.7069226882263769,
"grad_norm": 0.6683110594749451,
"learning_rate": 2.272094943221081e-05,
"loss": 0.3115,
"step": 1399
},
{
"epoch": 0.7074279939363315,
"grad_norm": 0.5461668372154236,
"learning_rate": 2.2689603438630606e-05,
"loss": 0.3772,
"step": 1400
},
{
"epoch": 0.707933299646286,
"grad_norm": 0.2557353079319,
"learning_rate": 2.2658261108119554e-05,
"loss": 0.1629,
"step": 1401
},
{
"epoch": 0.7084386053562405,
"grad_norm": 0.43094396591186523,
"learning_rate": 2.262692249037004e-05,
"loss": 0.1927,
"step": 1402
},
{
"epoch": 0.708943911066195,
"grad_norm": 0.43190810084342957,
"learning_rate": 2.259558763506858e-05,
"loss": 0.3081,
"step": 1403
},
{
"epoch": 0.7094492167761496,
"grad_norm": 0.32105064392089844,
"learning_rate": 2.2564256591895693e-05,
"loss": 0.1161,
"step": 1404
},
{
"epoch": 0.7099545224861041,
"grad_norm": 0.45100855827331543,
"learning_rate": 2.2532929410525887e-05,
"loss": 0.2658,
"step": 1405
},
{
"epoch": 0.7104598281960586,
"grad_norm": 0.41769468784332275,
"learning_rate": 2.2501606140627518e-05,
"loss": 0.3101,
"step": 1406
},
{
"epoch": 0.7109651339060131,
"grad_norm": 0.5544657111167908,
"learning_rate": 2.2470286831862766e-05,
"loss": 0.38,
"step": 1407
},
{
"epoch": 0.7114704396159677,
"grad_norm": 0.45244771242141724,
"learning_rate": 2.2438971533887527e-05,
"loss": 0.3448,
"step": 1408
},
{
"epoch": 0.7119757453259222,
"grad_norm": 0.3420441150665283,
"learning_rate": 2.240766029635131e-05,
"loss": 0.2201,
"step": 1409
},
{
"epoch": 0.7124810510358767,
"grad_norm": 0.6040711998939514,
"learning_rate": 2.2376353168897214e-05,
"loss": 0.4204,
"step": 1410
},
{
"epoch": 0.7129863567458312,
"grad_norm": 0.6188685894012451,
"learning_rate": 2.2345050201161814e-05,
"loss": 0.2601,
"step": 1411
},
{
"epoch": 0.7134916624557858,
"grad_norm": 0.4009005129337311,
"learning_rate": 2.2313751442775098e-05,
"loss": 0.2132,
"step": 1412
},
{
"epoch": 0.7139969681657403,
"grad_norm": 1.1570433378219604,
"learning_rate": 2.2282456943360375e-05,
"loss": 0.7633,
"step": 1413
},
{
"epoch": 0.7145022738756948,
"grad_norm": 0.5114752054214478,
"learning_rate": 2.225116675253418e-05,
"loss": 0.2895,
"step": 1414
},
{
"epoch": 0.7150075795856493,
"grad_norm": 0.3823153078556061,
"learning_rate": 2.2219880919906244e-05,
"loss": 0.2226,
"step": 1415
},
{
"epoch": 0.7155128852956039,
"grad_norm": 0.41869547963142395,
"learning_rate": 2.2188599495079384e-05,
"loss": 0.2079,
"step": 1416
},
{
"epoch": 0.7160181910055584,
"grad_norm": 0.43806976079940796,
"learning_rate": 2.2157322527649422e-05,
"loss": 0.1898,
"step": 1417
},
{
"epoch": 0.7165234967155129,
"grad_norm": 0.8781391382217407,
"learning_rate": 2.2126050067205127e-05,
"loss": 0.3187,
"step": 1418
},
{
"epoch": 0.7170288024254674,
"grad_norm": 0.5207377672195435,
"learning_rate": 2.209478216332809e-05,
"loss": 0.2541,
"step": 1419
},
{
"epoch": 0.717534108135422,
"grad_norm": 0.7016043663024902,
"learning_rate": 2.206351886559271e-05,
"loss": 0.2743,
"step": 1420
},
{
"epoch": 0.7180394138453765,
"grad_norm": 0.24266977608203888,
"learning_rate": 2.2032260223566072e-05,
"loss": 0.1328,
"step": 1421
},
{
"epoch": 0.718544719555331,
"grad_norm": 0.9864704608917236,
"learning_rate": 2.2001006286807875e-05,
"loss": 0.3871,
"step": 1422
},
{
"epoch": 0.7190500252652855,
"grad_norm": 0.3588007092475891,
"learning_rate": 2.1969757104870367e-05,
"loss": 0.143,
"step": 1423
},
{
"epoch": 0.71955533097524,
"grad_norm": 0.716530978679657,
"learning_rate": 2.1938512727298246e-05,
"loss": 0.3186,
"step": 1424
},
{
"epoch": 0.7200606366851945,
"grad_norm": 0.5047246813774109,
"learning_rate": 2.19072732036286e-05,
"loss": 0.1932,
"step": 1425
},
{
"epoch": 0.720565942395149,
"grad_norm": 0.6216627955436707,
"learning_rate": 2.1876038583390825e-05,
"loss": 0.4384,
"step": 1426
},
{
"epoch": 0.7210712481051036,
"grad_norm": 0.5591406226158142,
"learning_rate": 2.1844808916106536e-05,
"loss": 0.2485,
"step": 1427
},
{
"epoch": 0.7215765538150581,
"grad_norm": 0.6369313597679138,
"learning_rate": 2.1813584251289505e-05,
"loss": 0.4893,
"step": 1428
},
{
"epoch": 0.7220818595250126,
"grad_norm": 0.8026990294456482,
"learning_rate": 2.1782364638445546e-05,
"loss": 0.6091,
"step": 1429
},
{
"epoch": 0.7225871652349671,
"grad_norm": 0.45339179039001465,
"learning_rate": 2.17511501270725e-05,
"loss": 0.3557,
"step": 1430
},
{
"epoch": 0.7230924709449217,
"grad_norm": 0.5372389554977417,
"learning_rate": 2.171994076666009e-05,
"loss": 0.5554,
"step": 1431
},
{
"epoch": 0.7235977766548762,
"grad_norm": 0.45178845524787903,
"learning_rate": 2.168873660668989e-05,
"loss": 0.2379,
"step": 1432
},
{
"epoch": 0.7241030823648307,
"grad_norm": 0.4662405550479889,
"learning_rate": 2.1657537696635226e-05,
"loss": 0.2246,
"step": 1433
},
{
"epoch": 0.7246083880747852,
"grad_norm": 0.7474414706230164,
"learning_rate": 2.16263440859611e-05,
"loss": 0.4195,
"step": 1434
},
{
"epoch": 0.7251136937847398,
"grad_norm": 0.6440469622612,
"learning_rate": 2.1595155824124097e-05,
"loss": 0.3275,
"step": 1435
},
{
"epoch": 0.7256189994946943,
"grad_norm": 0.40439724922180176,
"learning_rate": 2.1563972960572343e-05,
"loss": 0.1541,
"step": 1436
},
{
"epoch": 0.7261243052046488,
"grad_norm": 0.7414180636405945,
"learning_rate": 2.1532795544745392e-05,
"loss": 0.3458,
"step": 1437
},
{
"epoch": 0.7266296109146033,
"grad_norm": 0.5826343894004822,
"learning_rate": 2.1501623626074175e-05,
"loss": 0.2742,
"step": 1438
},
{
"epoch": 0.7271349166245579,
"grad_norm": 0.38583678007125854,
"learning_rate": 2.1470457253980887e-05,
"loss": 0.1891,
"step": 1439
},
{
"epoch": 0.7276402223345124,
"grad_norm": 0.4571583867073059,
"learning_rate": 2.143929647787894e-05,
"loss": 0.2069,
"step": 1440
},
{
"epoch": 0.7281455280444669,
"grad_norm": 0.37782159447669983,
"learning_rate": 2.1408141347172877e-05,
"loss": 0.2598,
"step": 1441
},
{
"epoch": 0.7286508337544214,
"grad_norm": 1.0491501092910767,
"learning_rate": 2.13769919112583e-05,
"loss": 0.4662,
"step": 1442
},
{
"epoch": 0.729156139464376,
"grad_norm": 0.4999569058418274,
"learning_rate": 2.134584821952174e-05,
"loss": 0.2569,
"step": 1443
},
{
"epoch": 0.7296614451743305,
"grad_norm": 0.422539621591568,
"learning_rate": 2.131471032134067e-05,
"loss": 0.2263,
"step": 1444
},
{
"epoch": 0.730166750884285,
"grad_norm": 0.7593914866447449,
"learning_rate": 2.128357826608335e-05,
"loss": 0.3716,
"step": 1445
},
{
"epoch": 0.7306720565942395,
"grad_norm": 0.6366856098175049,
"learning_rate": 2.125245210310878e-05,
"loss": 0.2571,
"step": 1446
},
{
"epoch": 0.7311773623041941,
"grad_norm": 0.46671852469444275,
"learning_rate": 2.122133188176664e-05,
"loss": 0.4696,
"step": 1447
},
{
"epoch": 0.7316826680141486,
"grad_norm": 0.37286749482154846,
"learning_rate": 2.1190217651397145e-05,
"loss": 0.2091,
"step": 1448
},
{
"epoch": 0.732187973724103,
"grad_norm": 0.8244792819023132,
"learning_rate": 2.115910946133105e-05,
"loss": 0.2842,
"step": 1449
},
{
"epoch": 0.7326932794340576,
"grad_norm": 0.9506634473800659,
"learning_rate": 2.112800736088952e-05,
"loss": 0.365,
"step": 1450
},
{
"epoch": 0.7331985851440121,
"grad_norm": 0.9866954684257507,
"learning_rate": 2.1096911399384066e-05,
"loss": 0.3483,
"step": 1451
},
{
"epoch": 0.7337038908539667,
"grad_norm": 0.4005601406097412,
"learning_rate": 2.1065821626116466e-05,
"loss": 0.2033,
"step": 1452
},
{
"epoch": 0.7342091965639211,
"grad_norm": 0.43610894680023193,
"learning_rate": 2.1034738090378688e-05,
"loss": 0.2345,
"step": 1453
},
{
"epoch": 0.7347145022738757,
"grad_norm": 0.2879337668418884,
"learning_rate": 2.10036608414528e-05,
"loss": 0.1507,
"step": 1454
},
{
"epoch": 0.7352198079838302,
"grad_norm": 0.8920745849609375,
"learning_rate": 2.0972589928610917e-05,
"loss": 0.3903,
"step": 1455
},
{
"epoch": 0.7357251136937848,
"grad_norm": 0.8228932619094849,
"learning_rate": 2.0941525401115106e-05,
"loss": 0.3689,
"step": 1456
},
{
"epoch": 0.7362304194037392,
"grad_norm": 0.3942200839519501,
"learning_rate": 2.0910467308217314e-05,
"loss": 0.1604,
"step": 1457
},
{
"epoch": 0.7367357251136938,
"grad_norm": 0.47732532024383545,
"learning_rate": 2.0879415699159256e-05,
"loss": 0.3561,
"step": 1458
},
{
"epoch": 0.7372410308236483,
"grad_norm": 0.6839754581451416,
"learning_rate": 2.0848370623172405e-05,
"loss": 0.3294,
"step": 1459
},
{
"epoch": 0.7377463365336028,
"grad_norm": 0.35987260937690735,
"learning_rate": 2.081733212947786e-05,
"loss": 0.2124,
"step": 1460
},
{
"epoch": 0.7382516422435573,
"grad_norm": 0.5114133954048157,
"learning_rate": 2.078630026728628e-05,
"loss": 0.3365,
"step": 1461
},
{
"epoch": 0.7387569479535119,
"grad_norm": 0.6798128485679626,
"learning_rate": 2.0755275085797833e-05,
"loss": 0.3135,
"step": 1462
},
{
"epoch": 0.7392622536634664,
"grad_norm": 0.7099087238311768,
"learning_rate": 2.0724256634202052e-05,
"loss": 0.2584,
"step": 1463
},
{
"epoch": 0.7397675593734209,
"grad_norm": 0.6906381845474243,
"learning_rate": 2.069324496167783e-05,
"loss": 0.2981,
"step": 1464
},
{
"epoch": 0.7402728650833754,
"grad_norm": 0.2697136402130127,
"learning_rate": 2.0662240117393318e-05,
"loss": 0.174,
"step": 1465
},
{
"epoch": 0.74077817079333,
"grad_norm": 0.5975183248519897,
"learning_rate": 2.0631242150505827e-05,
"loss": 0.3088,
"step": 1466
},
{
"epoch": 0.7412834765032845,
"grad_norm": 0.4744568467140198,
"learning_rate": 2.060025111016176e-05,
"loss": 0.1892,
"step": 1467
},
{
"epoch": 0.741788782213239,
"grad_norm": 0.7941866517066956,
"learning_rate": 2.056926704549655e-05,
"loss": 0.3482,
"step": 1468
},
{
"epoch": 0.7422940879231935,
"grad_norm": 0.827198326587677,
"learning_rate": 2.0538290005634564e-05,
"loss": 0.3751,
"step": 1469
},
{
"epoch": 0.7427993936331481,
"grad_norm": 0.38413000106811523,
"learning_rate": 2.050732003968903e-05,
"loss": 0.3045,
"step": 1470
},
{
"epoch": 0.7433046993431026,
"grad_norm": 0.4320742189884186,
"learning_rate": 2.0476357196761967e-05,
"loss": 0.2605,
"step": 1471
},
{
"epoch": 0.743810005053057,
"grad_norm": 0.6644576191902161,
"learning_rate": 2.0445401525944107e-05,
"loss": 0.2997,
"step": 1472
},
{
"epoch": 0.7443153107630116,
"grad_norm": 0.6894829869270325,
"learning_rate": 2.0414453076314776e-05,
"loss": 0.4918,
"step": 1473
},
{
"epoch": 0.7448206164729662,
"grad_norm": 0.35164880752563477,
"learning_rate": 2.0383511896941893e-05,
"loss": 0.3007,
"step": 1474
},
{
"epoch": 0.7453259221829207,
"grad_norm": 0.6786766648292542,
"learning_rate": 2.0352578036881826e-05,
"loss": 0.4865,
"step": 1475
},
{
"epoch": 0.7458312278928751,
"grad_norm": 0.4968150556087494,
"learning_rate": 2.0321651545179348e-05,
"loss": 0.2479,
"step": 1476
},
{
"epoch": 0.7463365336028297,
"grad_norm": 0.8302887678146362,
"learning_rate": 2.029073247086756e-05,
"loss": 0.2688,
"step": 1477
},
{
"epoch": 0.7468418393127843,
"grad_norm": 0.44390469789505005,
"learning_rate": 2.0259820862967765e-05,
"loss": 0.3654,
"step": 1478
},
{
"epoch": 0.7473471450227388,
"grad_norm": 0.41298776865005493,
"learning_rate": 2.0228916770489466e-05,
"loss": 0.2194,
"step": 1479
},
{
"epoch": 0.7478524507326932,
"grad_norm": 0.46970030665397644,
"learning_rate": 2.019802024243023e-05,
"loss": 0.2061,
"step": 1480
},
{
"epoch": 0.7483577564426478,
"grad_norm": 0.510265052318573,
"learning_rate": 2.0167131327775652e-05,
"loss": 0.237,
"step": 1481
},
{
"epoch": 0.7488630621526023,
"grad_norm": 0.2790643870830536,
"learning_rate": 2.013625007549924e-05,
"loss": 0.2037,
"step": 1482
},
{
"epoch": 0.7493683678625569,
"grad_norm": 0.5583669543266296,
"learning_rate": 2.0105376534562344e-05,
"loss": 0.1887,
"step": 1483
},
{
"epoch": 0.7498736735725113,
"grad_norm": 0.4619552493095398,
"learning_rate": 2.0074510753914112e-05,
"loss": 0.1662,
"step": 1484
},
{
"epoch": 0.7503789792824659,
"grad_norm": 0.4680952727794647,
"learning_rate": 2.004365278249137e-05,
"loss": 0.3742,
"step": 1485
},
{
"epoch": 0.7508842849924204,
"grad_norm": 0.524354875087738,
"learning_rate": 2.0012802669218568e-05,
"loss": 0.2715,
"step": 1486
},
{
"epoch": 0.751389590702375,
"grad_norm": 0.21866445243358612,
"learning_rate": 1.998196046300771e-05,
"loss": 0.1283,
"step": 1487
},
{
"epoch": 0.7518948964123294,
"grad_norm": 0.3985232412815094,
"learning_rate": 1.9951126212758238e-05,
"loss": 0.2006,
"step": 1488
},
{
"epoch": 0.752400202122284,
"grad_norm": 0.4596607983112335,
"learning_rate": 1.9920299967356997e-05,
"loss": 0.4051,
"step": 1489
},
{
"epoch": 0.7529055078322385,
"grad_norm": 0.3508613705635071,
"learning_rate": 1.9889481775678134e-05,
"loss": 0.2223,
"step": 1490
},
{
"epoch": 0.7534108135421931,
"grad_norm": 0.2925485372543335,
"learning_rate": 1.9858671686583037e-05,
"loss": 0.1573,
"step": 1491
},
{
"epoch": 0.7539161192521475,
"grad_norm": 0.6288982033729553,
"learning_rate": 1.9827869748920252e-05,
"loss": 0.4522,
"step": 1492
},
{
"epoch": 0.7544214249621021,
"grad_norm": 0.6666999459266663,
"learning_rate": 1.9797076011525365e-05,
"loss": 0.5193,
"step": 1493
},
{
"epoch": 0.7549267306720566,
"grad_norm": 0.5883360505104065,
"learning_rate": 1.9766290523221e-05,
"loss": 0.3678,
"step": 1494
},
{
"epoch": 0.7554320363820111,
"grad_norm": 1.0901826620101929,
"learning_rate": 1.973551333281668e-05,
"loss": 0.638,
"step": 1495
},
{
"epoch": 0.7559373420919656,
"grad_norm": 0.41034024953842163,
"learning_rate": 1.9704744489108792e-05,
"loss": 0.1508,
"step": 1496
},
{
"epoch": 0.7564426478019202,
"grad_norm": 0.7889742851257324,
"learning_rate": 1.9673984040880468e-05,
"loss": 0.44,
"step": 1497
},
{
"epoch": 0.7569479535118747,
"grad_norm": 0.4004971385002136,
"learning_rate": 1.964323203690154e-05,
"loss": 0.1833,
"step": 1498
},
{
"epoch": 0.7574532592218292,
"grad_norm": 1.0051130056381226,
"learning_rate": 1.9612488525928452e-05,
"loss": 0.2733,
"step": 1499
},
{
"epoch": 0.7579585649317837,
"grad_norm": 0.49123916029930115,
"learning_rate": 1.958175355670418e-05,
"loss": 0.2622,
"step": 1500
},
{
"epoch": 0.7584638706417383,
"grad_norm": 0.581283688545227,
"learning_rate": 1.955102717795817e-05,
"loss": 0.1906,
"step": 1501
},
{
"epoch": 0.7589691763516928,
"grad_norm": 0.7873732447624207,
"learning_rate": 1.9520309438406217e-05,
"loss": 0.2679,
"step": 1502
},
{
"epoch": 0.7594744820616472,
"grad_norm": 0.5490037798881531,
"learning_rate": 1.9489600386750447e-05,
"loss": 0.2416,
"step": 1503
},
{
"epoch": 0.7599797877716018,
"grad_norm": 0.5317832231521606,
"learning_rate": 1.9458900071679202e-05,
"loss": 0.2828,
"step": 1504
},
{
"epoch": 0.7604850934815564,
"grad_norm": 0.6812729835510254,
"learning_rate": 1.9428208541866973e-05,
"loss": 0.3149,
"step": 1505
},
{
"epoch": 0.7609903991915109,
"grad_norm": 0.6650682091712952,
"learning_rate": 1.9397525845974336e-05,
"loss": 0.4121,
"step": 1506
},
{
"epoch": 0.7614957049014653,
"grad_norm": 0.5870944261550903,
"learning_rate": 1.9366852032647826e-05,
"loss": 0.3445,
"step": 1507
},
{
"epoch": 0.7620010106114199,
"grad_norm": 0.5600792765617371,
"learning_rate": 1.9336187150519924e-05,
"loss": 0.2882,
"step": 1508
},
{
"epoch": 0.7625063163213744,
"grad_norm": 0.45185530185699463,
"learning_rate": 1.9305531248208946e-05,
"loss": 0.1826,
"step": 1509
},
{
"epoch": 0.763011622031329,
"grad_norm": 0.7576361894607544,
"learning_rate": 1.9274884374318967e-05,
"loss": 0.3394,
"step": 1510
},
{
"epoch": 0.7635169277412834,
"grad_norm": 0.3762153089046478,
"learning_rate": 1.9244246577439752e-05,
"loss": 0.1961,
"step": 1511
},
{
"epoch": 0.764022233451238,
"grad_norm": 0.6146969199180603,
"learning_rate": 1.9213617906146663e-05,
"loss": 0.3267,
"step": 1512
},
{
"epoch": 0.7645275391611925,
"grad_norm": 0.5639180541038513,
"learning_rate": 1.9182998409000606e-05,
"loss": 0.2919,
"step": 1513
},
{
"epoch": 0.7650328448711471,
"grad_norm": 0.5925596952438354,
"learning_rate": 1.9152388134547945e-05,
"loss": 0.2764,
"step": 1514
},
{
"epoch": 0.7655381505811015,
"grad_norm": 0.41065713763237,
"learning_rate": 1.91217871313204e-05,
"loss": 0.188,
"step": 1515
},
{
"epoch": 0.7660434562910561,
"grad_norm": 0.38348203897476196,
"learning_rate": 1.909119544783503e-05,
"loss": 0.2152,
"step": 1516
},
{
"epoch": 0.7665487620010106,
"grad_norm": 0.4948267340660095,
"learning_rate": 1.9060613132594067e-05,
"loss": 0.2266,
"step": 1517
},
{
"epoch": 0.7670540677109652,
"grad_norm": 0.5641569495201111,
"learning_rate": 1.9030040234084926e-05,
"loss": 0.2611,
"step": 1518
},
{
"epoch": 0.7675593734209196,
"grad_norm": 0.4439907670021057,
"learning_rate": 1.8999476800780087e-05,
"loss": 0.2271,
"step": 1519
},
{
"epoch": 0.7680646791308742,
"grad_norm": 0.7098090648651123,
"learning_rate": 1.8968922881137012e-05,
"loss": 0.4192,
"step": 1520
},
{
"epoch": 0.7685699848408287,
"grad_norm": 0.6672466397285461,
"learning_rate": 1.8938378523598103e-05,
"loss": 0.3129,
"step": 1521
},
{
"epoch": 0.7690752905507833,
"grad_norm": 0.672419011592865,
"learning_rate": 1.8907843776590567e-05,
"loss": 0.3302,
"step": 1522
},
{
"epoch": 0.7695805962607377,
"grad_norm": 0.5676461458206177,
"learning_rate": 1.8877318688526393e-05,
"loss": 0.2522,
"step": 1523
},
{
"epoch": 0.7700859019706923,
"grad_norm": 0.4994024932384491,
"learning_rate": 1.8846803307802265e-05,
"loss": 0.1974,
"step": 1524
},
{
"epoch": 0.7705912076806468,
"grad_norm": 0.4182925820350647,
"learning_rate": 1.8816297682799452e-05,
"loss": 0.1786,
"step": 1525
},
{
"epoch": 0.7710965133906014,
"grad_norm": 0.6781833171844482,
"learning_rate": 1.8785801861883788e-05,
"loss": 0.4861,
"step": 1526
},
{
"epoch": 0.7716018191005558,
"grad_norm": 0.6011713147163391,
"learning_rate": 1.875531589340553e-05,
"loss": 0.2655,
"step": 1527
},
{
"epoch": 0.7721071248105104,
"grad_norm": 0.6764652729034424,
"learning_rate": 1.8724839825699332e-05,
"loss": 0.5056,
"step": 1528
},
{
"epoch": 0.7726124305204649,
"grad_norm": 0.5193274021148682,
"learning_rate": 1.869437370708415e-05,
"loss": 0.2291,
"step": 1529
},
{
"epoch": 0.7731177362304194,
"grad_norm": 0.5534316301345825,
"learning_rate": 1.8663917585863154e-05,
"loss": 0.2526,
"step": 1530
},
{
"epoch": 0.7736230419403739,
"grad_norm": 0.4299106299877167,
"learning_rate": 1.8633471510323694e-05,
"loss": 0.1945,
"step": 1531
},
{
"epoch": 0.7741283476503285,
"grad_norm": 0.5962615013122559,
"learning_rate": 1.8603035528737146e-05,
"loss": 0.3837,
"step": 1532
},
{
"epoch": 0.774633653360283,
"grad_norm": 0.6587054133415222,
"learning_rate": 1.8572609689358923e-05,
"loss": 0.3066,
"step": 1533
},
{
"epoch": 0.7751389590702374,
"grad_norm": 0.919040858745575,
"learning_rate": 1.854219404042834e-05,
"loss": 0.369,
"step": 1534
},
{
"epoch": 0.775644264780192,
"grad_norm": 0.4253031611442566,
"learning_rate": 1.8511788630168557e-05,
"loss": 0.199,
"step": 1535
},
{
"epoch": 0.7761495704901465,
"grad_norm": 0.6377377510070801,
"learning_rate": 1.848139350678651e-05,
"loss": 0.3452,
"step": 1536
},
{
"epoch": 0.7766548762001011,
"grad_norm": 0.9110857248306274,
"learning_rate": 1.8451008718472802e-05,
"loss": 0.3778,
"step": 1537
},
{
"epoch": 0.7771601819100555,
"grad_norm": 0.42227810621261597,
"learning_rate": 1.8420634313401673e-05,
"loss": 0.1541,
"step": 1538
},
{
"epoch": 0.7776654876200101,
"grad_norm": 0.5211955904960632,
"learning_rate": 1.8390270339730892e-05,
"loss": 0.3227,
"step": 1539
},
{
"epoch": 0.7781707933299646,
"grad_norm": 0.366298645734787,
"learning_rate": 1.835991684560169e-05,
"loss": 0.1747,
"step": 1540
},
{
"epoch": 0.7786760990399192,
"grad_norm": 0.4732970893383026,
"learning_rate": 1.8329573879138695e-05,
"loss": 0.2065,
"step": 1541
},
{
"epoch": 0.7791814047498736,
"grad_norm": 0.43378695845603943,
"learning_rate": 1.8299241488449814e-05,
"loss": 0.1849,
"step": 1542
},
{
"epoch": 0.7796867104598282,
"grad_norm": 0.6272542476654053,
"learning_rate": 1.826891972162621e-05,
"loss": 0.2282,
"step": 1543
},
{
"epoch": 0.7801920161697827,
"grad_norm": 0.46285495162010193,
"learning_rate": 1.82386086267422e-05,
"loss": 0.2114,
"step": 1544
},
{
"epoch": 0.7806973218797373,
"grad_norm": 0.5049265623092651,
"learning_rate": 1.8208308251855176e-05,
"loss": 0.5401,
"step": 1545
}
],
"logging_steps": 1,
"max_steps": 2573,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 515,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 2.4075086284361073e+18,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}