Hebrew-Mistral-7B / trainer_state.json
yam-peleg's picture
Upload 13 files
133e7da verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.032734074612906575,
"eval_steps": 500,
"global_step": 630,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0,
"grad_norm": 2.967684234392304,
"learning_rate": 8.722043470761813e-09,
"loss": 1.4217,
"step": 1
},
{
"epoch": 0.0,
"grad_norm": 3.0271119924702425,
"learning_rate": 1.7444086941523626e-08,
"loss": 1.3854,
"step": 2
},
{
"epoch": 0.0,
"grad_norm": 3.0212962641045844,
"learning_rate": 2.6166130412285438e-08,
"loss": 1.421,
"step": 3
},
{
"epoch": 0.0,
"grad_norm": 2.969098210999342,
"learning_rate": 3.488817388304725e-08,
"loss": 1.3894,
"step": 4
},
{
"epoch": 0.0,
"grad_norm": 2.9600462061407877,
"learning_rate": 4.361021735380907e-08,
"loss": 1.3803,
"step": 5
},
{
"epoch": 0.0,
"grad_norm": 3.074991458239833,
"learning_rate": 5.2332260824570876e-08,
"loss": 1.3815,
"step": 6
},
{
"epoch": 0.0,
"grad_norm": 3.026445479098331,
"learning_rate": 6.105430429533269e-08,
"loss": 1.3684,
"step": 7
},
{
"epoch": 0.0,
"grad_norm": 2.9268992001978744,
"learning_rate": 6.97763477660945e-08,
"loss": 1.4307,
"step": 8
},
{
"epoch": 0.0,
"grad_norm": 3.020642302186043,
"learning_rate": 7.849839123685631e-08,
"loss": 1.3318,
"step": 9
},
{
"epoch": 0.0,
"grad_norm": 3.0765469986494853,
"learning_rate": 8.722043470761814e-08,
"loss": 1.3942,
"step": 10
},
{
"epoch": 0.0,
"grad_norm": 3.081729945762728,
"learning_rate": 9.594247817837994e-08,
"loss": 1.3555,
"step": 11
},
{
"epoch": 0.0,
"grad_norm": 3.0021774146670808,
"learning_rate": 1.0466452164914175e-07,
"loss": 1.4366,
"step": 12
},
{
"epoch": 0.0,
"grad_norm": 3.059550099782956,
"learning_rate": 1.1338656511990357e-07,
"loss": 1.3999,
"step": 13
},
{
"epoch": 0.0,
"grad_norm": 2.9198920843423073,
"learning_rate": 1.2210860859066538e-07,
"loss": 1.3949,
"step": 14
},
{
"epoch": 0.0,
"grad_norm": 3.013793892067568,
"learning_rate": 1.308306520614272e-07,
"loss": 1.4062,
"step": 15
},
{
"epoch": 0.0,
"grad_norm": 2.9189606946894022,
"learning_rate": 1.39552695532189e-07,
"loss": 1.401,
"step": 16
},
{
"epoch": 0.0,
"grad_norm": 2.9989048069702475,
"learning_rate": 1.4827473900295083e-07,
"loss": 1.39,
"step": 17
},
{
"epoch": 0.0,
"grad_norm": 2.8873329349365613,
"learning_rate": 1.5699678247371262e-07,
"loss": 1.4061,
"step": 18
},
{
"epoch": 0.0,
"grad_norm": 2.8867416403445345,
"learning_rate": 1.6571882594447446e-07,
"loss": 1.4032,
"step": 19
},
{
"epoch": 0.0,
"grad_norm": 2.9725069440958127,
"learning_rate": 1.7444086941523627e-07,
"loss": 1.3772,
"step": 20
},
{
"epoch": 0.0,
"grad_norm": 3.0049844241627675,
"learning_rate": 1.8316291288599806e-07,
"loss": 1.3907,
"step": 21
},
{
"epoch": 0.0,
"grad_norm": 2.878905929699835,
"learning_rate": 1.9188495635675987e-07,
"loss": 1.4029,
"step": 22
},
{
"epoch": 0.0,
"grad_norm": 2.8262523355216014,
"learning_rate": 2.006069998275217e-07,
"loss": 1.3488,
"step": 23
},
{
"epoch": 0.0,
"grad_norm": 2.913459302454502,
"learning_rate": 2.093290432982835e-07,
"loss": 1.3888,
"step": 24
},
{
"epoch": 0.0,
"grad_norm": 2.924777837323026,
"learning_rate": 2.1805108676904532e-07,
"loss": 1.3423,
"step": 25
},
{
"epoch": 0.0,
"grad_norm": 3.0238436977422207,
"learning_rate": 2.2677313023980713e-07,
"loss": 1.4083,
"step": 26
},
{
"epoch": 0.0,
"grad_norm": 2.9404632380821036,
"learning_rate": 2.3549517371056895e-07,
"loss": 1.3325,
"step": 27
},
{
"epoch": 0.0,
"grad_norm": 2.9624670572200222,
"learning_rate": 2.4421721718133076e-07,
"loss": 1.4221,
"step": 28
},
{
"epoch": 0.0,
"grad_norm": 2.9307501329723236,
"learning_rate": 2.5293926065209255e-07,
"loss": 1.3827,
"step": 29
},
{
"epoch": 0.0,
"grad_norm": 2.8568182093611365,
"learning_rate": 2.616613041228544e-07,
"loss": 1.394,
"step": 30
},
{
"epoch": 0.0,
"grad_norm": 3.0056868600386544,
"learning_rate": 2.703833475936162e-07,
"loss": 1.3585,
"step": 31
},
{
"epoch": 0.0,
"grad_norm": 2.928843591728527,
"learning_rate": 2.79105391064378e-07,
"loss": 1.3927,
"step": 32
},
{
"epoch": 0.0,
"grad_norm": 2.979059280785437,
"learning_rate": 2.878274345351398e-07,
"loss": 1.4622,
"step": 33
},
{
"epoch": 0.0,
"grad_norm": 2.948156874136048,
"learning_rate": 2.9654947800590165e-07,
"loss": 1.3583,
"step": 34
},
{
"epoch": 0.0,
"grad_norm": 2.883247676830032,
"learning_rate": 3.0527152147666344e-07,
"loss": 1.3533,
"step": 35
},
{
"epoch": 0.0,
"grad_norm": 3.058107134456641,
"learning_rate": 3.1399356494742523e-07,
"loss": 1.3904,
"step": 36
},
{
"epoch": 0.0,
"grad_norm": 3.0535192707251237,
"learning_rate": 3.2271560841818707e-07,
"loss": 1.3978,
"step": 37
},
{
"epoch": 0.0,
"grad_norm": 2.897725634603183,
"learning_rate": 3.314376518889489e-07,
"loss": 1.3422,
"step": 38
},
{
"epoch": 0.0,
"grad_norm": 2.869420148217502,
"learning_rate": 3.401596953597107e-07,
"loss": 1.4465,
"step": 39
},
{
"epoch": 0.0,
"grad_norm": 2.9815636403050676,
"learning_rate": 3.4888173883047254e-07,
"loss": 1.407,
"step": 40
},
{
"epoch": 0.0,
"grad_norm": 2.9486391625075123,
"learning_rate": 3.576037823012343e-07,
"loss": 1.391,
"step": 41
},
{
"epoch": 0.0,
"grad_norm": 2.9396534742524283,
"learning_rate": 3.663258257719961e-07,
"loss": 1.4191,
"step": 42
},
{
"epoch": 0.0,
"grad_norm": 2.954105365539368,
"learning_rate": 3.7504786924275796e-07,
"loss": 1.4039,
"step": 43
},
{
"epoch": 0.0,
"grad_norm": 2.9899210526299482,
"learning_rate": 3.8376991271351975e-07,
"loss": 1.3606,
"step": 44
},
{
"epoch": 0.0,
"grad_norm": 2.9715202250605515,
"learning_rate": 3.924919561842816e-07,
"loss": 1.3561,
"step": 45
},
{
"epoch": 0.0,
"grad_norm": 2.8726950910953133,
"learning_rate": 4.012139996550434e-07,
"loss": 1.4004,
"step": 46
},
{
"epoch": 0.0,
"grad_norm": 2.9356806631744568,
"learning_rate": 4.0993604312580517e-07,
"loss": 1.3626,
"step": 47
},
{
"epoch": 0.0,
"grad_norm": 2.828385414972167,
"learning_rate": 4.18658086596567e-07,
"loss": 1.416,
"step": 48
},
{
"epoch": 0.0,
"grad_norm": 2.860775941749999,
"learning_rate": 4.273801300673288e-07,
"loss": 1.4232,
"step": 49
},
{
"epoch": 0.0,
"grad_norm": 2.908555437519401,
"learning_rate": 4.3610217353809064e-07,
"loss": 1.371,
"step": 50
},
{
"epoch": 0.0,
"grad_norm": 3.0110809579629283,
"learning_rate": 4.448242170088525e-07,
"loss": 1.3886,
"step": 51
},
{
"epoch": 0.0,
"grad_norm": 2.8879956038976062,
"learning_rate": 4.5354626047961427e-07,
"loss": 1.3677,
"step": 52
},
{
"epoch": 0.0,
"grad_norm": 2.9475494022582405,
"learning_rate": 4.622683039503761e-07,
"loss": 1.3596,
"step": 53
},
{
"epoch": 0.0,
"grad_norm": 2.9208201499436677,
"learning_rate": 4.709903474211379e-07,
"loss": 1.3795,
"step": 54
},
{
"epoch": 0.0,
"grad_norm": 2.9250691956165293,
"learning_rate": 4.797123908918997e-07,
"loss": 1.3521,
"step": 55
},
{
"epoch": 0.0,
"grad_norm": 3.0059604294844795,
"learning_rate": 4.884344343626615e-07,
"loss": 1.3872,
"step": 56
},
{
"epoch": 0.0,
"grad_norm": 2.9408348770709347,
"learning_rate": 4.971564778334233e-07,
"loss": 1.41,
"step": 57
},
{
"epoch": 0.0,
"grad_norm": 2.9941841393676247,
"learning_rate": 5.058785213041851e-07,
"loss": 1.4293,
"step": 58
},
{
"epoch": 0.0,
"grad_norm": 2.9287843738148744,
"learning_rate": 5.146005647749469e-07,
"loss": 1.411,
"step": 59
},
{
"epoch": 0.0,
"grad_norm": 2.9468093301597533,
"learning_rate": 5.233226082457088e-07,
"loss": 1.3508,
"step": 60
},
{
"epoch": 0.0,
"grad_norm": 2.854893968299532,
"learning_rate": 5.320446517164706e-07,
"loss": 1.4042,
"step": 61
},
{
"epoch": 0.0,
"grad_norm": 2.9224674242946387,
"learning_rate": 5.407666951872324e-07,
"loss": 1.4182,
"step": 62
},
{
"epoch": 0.0,
"grad_norm": 3.025591867750147,
"learning_rate": 5.494887386579943e-07,
"loss": 1.4162,
"step": 63
},
{
"epoch": 0.0,
"grad_norm": 2.9351339151935045,
"learning_rate": 5.58210782128756e-07,
"loss": 1.2956,
"step": 64
},
{
"epoch": 0.0,
"grad_norm": 2.9740598113404677,
"learning_rate": 5.669328255995178e-07,
"loss": 1.4181,
"step": 65
},
{
"epoch": 0.0,
"grad_norm": 2.8698658052840464,
"learning_rate": 5.756548690702796e-07,
"loss": 1.3885,
"step": 66
},
{
"epoch": 0.0,
"grad_norm": 2.847745090454755,
"learning_rate": 5.843769125410415e-07,
"loss": 1.4268,
"step": 67
},
{
"epoch": 0.0,
"grad_norm": 3.53006793293518,
"learning_rate": 5.930989560118033e-07,
"loss": 1.4369,
"step": 68
},
{
"epoch": 0.0,
"grad_norm": 3.02012038605618,
"learning_rate": 6.01820999482565e-07,
"loss": 1.406,
"step": 69
},
{
"epoch": 0.0,
"grad_norm": 2.8502665195139074,
"learning_rate": 6.105430429533269e-07,
"loss": 1.4273,
"step": 70
},
{
"epoch": 0.0,
"grad_norm": 2.966110454182966,
"learning_rate": 6.192650864240887e-07,
"loss": 1.3822,
"step": 71
},
{
"epoch": 0.0,
"grad_norm": 2.9117409152319604,
"learning_rate": 6.279871298948505e-07,
"loss": 1.3758,
"step": 72
},
{
"epoch": 0.0,
"grad_norm": 2.9583103749551247,
"learning_rate": 6.367091733656124e-07,
"loss": 1.3821,
"step": 73
},
{
"epoch": 0.0,
"grad_norm": 2.9976941225678524,
"learning_rate": 6.454312168363741e-07,
"loss": 1.3818,
"step": 74
},
{
"epoch": 0.0,
"grad_norm": 2.9208562033277237,
"learning_rate": 6.541532603071359e-07,
"loss": 1.4249,
"step": 75
},
{
"epoch": 0.0,
"grad_norm": 2.9753607287344868,
"learning_rate": 6.628753037778978e-07,
"loss": 1.394,
"step": 76
},
{
"epoch": 0.0,
"grad_norm": 2.852059681146882,
"learning_rate": 6.715973472486596e-07,
"loss": 1.3735,
"step": 77
},
{
"epoch": 0.0,
"grad_norm": 2.9721541467116075,
"learning_rate": 6.803193907194214e-07,
"loss": 1.3284,
"step": 78
},
{
"epoch": 0.0,
"grad_norm": 2.959460844473033,
"learning_rate": 6.890414341901832e-07,
"loss": 1.3998,
"step": 79
},
{
"epoch": 0.0,
"grad_norm": 2.8741209861248747,
"learning_rate": 6.977634776609451e-07,
"loss": 1.3696,
"step": 80
},
{
"epoch": 0.0,
"grad_norm": 2.916879772765294,
"learning_rate": 7.064855211317069e-07,
"loss": 1.4736,
"step": 81
},
{
"epoch": 0.0,
"grad_norm": 3.069300514821944,
"learning_rate": 7.152075646024686e-07,
"loss": 1.3341,
"step": 82
},
{
"epoch": 0.0,
"grad_norm": 2.966078167553101,
"learning_rate": 7.239296080732305e-07,
"loss": 1.3705,
"step": 83
},
{
"epoch": 0.0,
"grad_norm": 2.891366947828864,
"learning_rate": 7.326516515439922e-07,
"loss": 1.4036,
"step": 84
},
{
"epoch": 0.0,
"grad_norm": 3.0015624379292687,
"learning_rate": 7.41373695014754e-07,
"loss": 1.3548,
"step": 85
},
{
"epoch": 0.0,
"grad_norm": 2.9762714989519297,
"learning_rate": 7.500957384855159e-07,
"loss": 1.4038,
"step": 86
},
{
"epoch": 0.0,
"grad_norm": 3.0417489896152325,
"learning_rate": 7.588177819562777e-07,
"loss": 1.3972,
"step": 87
},
{
"epoch": 0.0,
"grad_norm": 2.957367413809754,
"learning_rate": 7.675398254270395e-07,
"loss": 1.376,
"step": 88
},
{
"epoch": 0.0,
"grad_norm": 2.9811178726545466,
"learning_rate": 7.762618688978014e-07,
"loss": 1.366,
"step": 89
},
{
"epoch": 0.0,
"grad_norm": 2.9882512869731994,
"learning_rate": 7.849839123685632e-07,
"loss": 1.3835,
"step": 90
},
{
"epoch": 0.0,
"grad_norm": 2.878782548871192,
"learning_rate": 7.93705955839325e-07,
"loss": 1.3372,
"step": 91
},
{
"epoch": 0.0,
"grad_norm": 2.9971417051306344,
"learning_rate": 8.024279993100868e-07,
"loss": 1.4037,
"step": 92
},
{
"epoch": 0.0,
"grad_norm": 3.0231933357399066,
"learning_rate": 8.111500427808487e-07,
"loss": 1.3692,
"step": 93
},
{
"epoch": 0.0,
"grad_norm": 2.8777305685815024,
"learning_rate": 8.198720862516103e-07,
"loss": 1.3493,
"step": 94
},
{
"epoch": 0.0,
"grad_norm": 2.990027107011572,
"learning_rate": 8.285941297223721e-07,
"loss": 1.3586,
"step": 95
},
{
"epoch": 0.0,
"grad_norm": 2.818072292309908,
"learning_rate": 8.37316173193134e-07,
"loss": 1.3612,
"step": 96
},
{
"epoch": 0.01,
"grad_norm": 3.1271514088993144,
"learning_rate": 8.460382166638958e-07,
"loss": 1.336,
"step": 97
},
{
"epoch": 0.01,
"grad_norm": 3.0169265133833396,
"learning_rate": 8.547602601346576e-07,
"loss": 1.3665,
"step": 98
},
{
"epoch": 0.01,
"grad_norm": 2.9482660829507994,
"learning_rate": 8.634823036054195e-07,
"loss": 1.3551,
"step": 99
},
{
"epoch": 0.01,
"grad_norm": 3.087617940804514,
"learning_rate": 8.722043470761813e-07,
"loss": 1.3659,
"step": 100
},
{
"epoch": 0.01,
"grad_norm": 2.894747860556493,
"learning_rate": 8.72204341205319e-07,
"loss": 1.4032,
"step": 101
},
{
"epoch": 0.01,
"grad_norm": 2.9567796592966546,
"learning_rate": 8.722043235927325e-07,
"loss": 1.3568,
"step": 102
},
{
"epoch": 0.01,
"grad_norm": 3.130564519854559,
"learning_rate": 8.72204294238422e-07,
"loss": 1.3567,
"step": 103
},
{
"epoch": 0.01,
"grad_norm": 2.98891915499642,
"learning_rate": 8.722042531423884e-07,
"loss": 1.3865,
"step": 104
},
{
"epoch": 0.01,
"grad_norm": 3.036577692929289,
"learning_rate": 8.722042003046327e-07,
"loss": 1.3901,
"step": 105
},
{
"epoch": 0.01,
"grad_norm": 3.001761214468399,
"learning_rate": 8.722041357251567e-07,
"loss": 1.4117,
"step": 106
},
{
"epoch": 0.01,
"grad_norm": 3.006505553779288,
"learning_rate": 8.722040594039618e-07,
"loss": 1.4083,
"step": 107
},
{
"epoch": 0.01,
"grad_norm": 2.970081989296889,
"learning_rate": 8.722039713410501e-07,
"loss": 1.3163,
"step": 108
},
{
"epoch": 0.01,
"grad_norm": 2.9452680507051863,
"learning_rate": 8.72203871536424e-07,
"loss": 1.3367,
"step": 109
},
{
"epoch": 0.01,
"grad_norm": 3.0057027079906216,
"learning_rate": 8.722037599900863e-07,
"loss": 1.3893,
"step": 110
},
{
"epoch": 0.01,
"grad_norm": 2.9731127424958697,
"learning_rate": 8.722036367020397e-07,
"loss": 1.3604,
"step": 111
},
{
"epoch": 0.01,
"grad_norm": 2.884218184477182,
"learning_rate": 8.722035016722879e-07,
"loss": 1.4651,
"step": 112
},
{
"epoch": 0.01,
"grad_norm": 2.8763866239897644,
"learning_rate": 8.722033549008343e-07,
"loss": 1.3947,
"step": 113
},
{
"epoch": 0.01,
"grad_norm": 2.927176134008426,
"learning_rate": 8.722031963876829e-07,
"loss": 1.3667,
"step": 114
},
{
"epoch": 0.01,
"grad_norm": 2.9924621440798664,
"learning_rate": 8.72203026132838e-07,
"loss": 1.3632,
"step": 115
},
{
"epoch": 0.01,
"grad_norm": 3.0499504528400943,
"learning_rate": 8.72202844136304e-07,
"loss": 1.4176,
"step": 116
},
{
"epoch": 0.01,
"grad_norm": 2.9684475725772392,
"learning_rate": 8.722026503980863e-07,
"loss": 1.4146,
"step": 117
},
{
"epoch": 0.01,
"grad_norm": 2.9148264494394662,
"learning_rate": 8.722024449181895e-07,
"loss": 1.4205,
"step": 118
},
{
"epoch": 0.01,
"grad_norm": 2.947431519938494,
"learning_rate": 8.722022276966194e-07,
"loss": 1.3281,
"step": 119
},
{
"epoch": 0.01,
"grad_norm": 2.955626479582277,
"learning_rate": 8.72201998733382e-07,
"loss": 1.3465,
"step": 120
},
{
"epoch": 0.01,
"grad_norm": 3.028540598737732,
"learning_rate": 8.722017580284832e-07,
"loss": 1.3472,
"step": 121
},
{
"epoch": 0.01,
"grad_norm": 2.9197378030040753,
"learning_rate": 8.722015055819296e-07,
"loss": 1.381,
"step": 122
},
{
"epoch": 0.01,
"grad_norm": 3.0574320249257227,
"learning_rate": 8.722012413937282e-07,
"loss": 1.4225,
"step": 123
},
{
"epoch": 0.01,
"grad_norm": 2.913538682906452,
"learning_rate": 8.722009654638856e-07,
"loss": 1.3536,
"step": 124
},
{
"epoch": 0.01,
"grad_norm": 2.886066575609779,
"learning_rate": 8.722006777924096e-07,
"loss": 1.3736,
"step": 125
},
{
"epoch": 0.01,
"grad_norm": 2.957758024407401,
"learning_rate": 8.722003783793081e-07,
"loss": 1.3973,
"step": 126
},
{
"epoch": 0.01,
"grad_norm": 2.958995782934072,
"learning_rate": 8.722000672245888e-07,
"loss": 1.3954,
"step": 127
},
{
"epoch": 0.01,
"grad_norm": 2.987991679162308,
"learning_rate": 8.721997443282602e-07,
"loss": 1.3757,
"step": 128
},
{
"epoch": 0.01,
"grad_norm": 3.0909686056252434,
"learning_rate": 8.721994096903311e-07,
"loss": 1.3462,
"step": 129
},
{
"epoch": 0.01,
"grad_norm": 2.9477825764652494,
"learning_rate": 8.721990633108104e-07,
"loss": 1.4295,
"step": 130
},
{
"epoch": 0.01,
"grad_norm": 2.947921751933276,
"learning_rate": 8.721987051897074e-07,
"loss": 1.3854,
"step": 131
},
{
"epoch": 0.01,
"grad_norm": 2.8440480288328427,
"learning_rate": 8.721983353270319e-07,
"loss": 1.4106,
"step": 132
},
{
"epoch": 0.01,
"grad_norm": 2.9792049006251906,
"learning_rate": 8.721979537227935e-07,
"loss": 1.3913,
"step": 133
},
{
"epoch": 0.01,
"grad_norm": 2.9238731192373746,
"learning_rate": 8.721975603770031e-07,
"loss": 1.3695,
"step": 134
},
{
"epoch": 0.01,
"grad_norm": 3.0773774260583155,
"learning_rate": 8.721971552896706e-07,
"loss": 1.3629,
"step": 135
},
{
"epoch": 0.01,
"grad_norm": 2.9679219250269044,
"learning_rate": 8.721967384608074e-07,
"loss": 1.4205,
"step": 136
},
{
"epoch": 0.01,
"grad_norm": 3.030565370577699,
"learning_rate": 8.721963098904246e-07,
"loss": 1.4311,
"step": 137
},
{
"epoch": 0.01,
"grad_norm": 3.0375097512582503,
"learning_rate": 8.721958695785336e-07,
"loss": 1.4069,
"step": 138
},
{
"epoch": 0.01,
"grad_norm": 2.92214551378445,
"learning_rate": 8.721954175251462e-07,
"loss": 1.422,
"step": 139
},
{
"epoch": 0.01,
"grad_norm": 3.1283303266405578,
"learning_rate": 8.721949537302749e-07,
"loss": 1.432,
"step": 140
},
{
"epoch": 0.01,
"grad_norm": 3.1049340924381705,
"learning_rate": 8.72194478193932e-07,
"loss": 1.3815,
"step": 141
},
{
"epoch": 0.01,
"grad_norm": 2.869253522521714,
"learning_rate": 8.721939909161303e-07,
"loss": 1.391,
"step": 142
},
{
"epoch": 0.01,
"grad_norm": 2.8752461370622306,
"learning_rate": 8.721934918968828e-07,
"loss": 1.3769,
"step": 143
},
{
"epoch": 0.01,
"grad_norm": 2.909751532098998,
"learning_rate": 8.721929811362032e-07,
"loss": 1.3995,
"step": 144
},
{
"epoch": 0.01,
"grad_norm": 3.0462324191538572,
"learning_rate": 8.72192458634105e-07,
"loss": 1.3689,
"step": 145
},
{
"epoch": 0.01,
"grad_norm": 2.9750194133891363,
"learning_rate": 8.721919243906024e-07,
"loss": 1.3707,
"step": 146
},
{
"epoch": 0.01,
"grad_norm": 2.932835851287147,
"learning_rate": 8.721913784057099e-07,
"loss": 1.3676,
"step": 147
},
{
"epoch": 0.01,
"grad_norm": 2.9429516387295926,
"learning_rate": 8.721908206794419e-07,
"loss": 1.3731,
"step": 148
},
{
"epoch": 0.01,
"grad_norm": 2.994182650351975,
"learning_rate": 8.721902512118136e-07,
"loss": 1.3542,
"step": 149
},
{
"epoch": 0.01,
"grad_norm": 2.876833046636617,
"learning_rate": 8.721896700028404e-07,
"loss": 1.4124,
"step": 150
},
{
"epoch": 0.01,
"grad_norm": 2.945327518176284,
"learning_rate": 8.721890770525377e-07,
"loss": 1.4137,
"step": 151
},
{
"epoch": 0.01,
"grad_norm": 21.047240653653276,
"learning_rate": 8.721884723609218e-07,
"loss": 1.4264,
"step": 152
},
{
"epoch": 0.01,
"grad_norm": 3.029952956834553,
"learning_rate": 8.721878559280086e-07,
"loss": 1.4372,
"step": 153
},
{
"epoch": 0.01,
"grad_norm": 3.137230528895406,
"learning_rate": 8.721872277538151e-07,
"loss": 1.4019,
"step": 154
},
{
"epoch": 0.01,
"grad_norm": 2.9696146910825694,
"learning_rate": 8.72186587838358e-07,
"loss": 1.4515,
"step": 155
},
{
"epoch": 0.01,
"grad_norm": 2.980760336638325,
"learning_rate": 8.721859361816546e-07,
"loss": 1.4203,
"step": 156
},
{
"epoch": 0.01,
"grad_norm": 3.086006727040003,
"learning_rate": 8.721852727837222e-07,
"loss": 1.3712,
"step": 157
},
{
"epoch": 0.01,
"grad_norm": 3.003419403761712,
"learning_rate": 8.72184597644579e-07,
"loss": 1.4107,
"step": 158
},
{
"epoch": 0.01,
"grad_norm": 3.0463864266769773,
"learning_rate": 8.72183910764243e-07,
"loss": 1.4082,
"step": 159
},
{
"epoch": 0.01,
"grad_norm": 2.950362004991645,
"learning_rate": 8.721832121427326e-07,
"loss": 1.352,
"step": 160
},
{
"epoch": 0.01,
"grad_norm": 2.8779668562920815,
"learning_rate": 8.721825017800669e-07,
"loss": 1.4236,
"step": 161
},
{
"epoch": 0.01,
"grad_norm": 3.0172810234945455,
"learning_rate": 8.721817796762648e-07,
"loss": 1.3871,
"step": 162
},
{
"epoch": 0.01,
"grad_norm": 2.9726094865888224,
"learning_rate": 8.721810458313457e-07,
"loss": 1.349,
"step": 163
},
{
"epoch": 0.01,
"grad_norm": 3.0322153773349334,
"learning_rate": 8.721803002453297e-07,
"loss": 1.3935,
"step": 164
},
{
"epoch": 0.01,
"grad_norm": 3.0249194383352283,
"learning_rate": 8.721795429182364e-07,
"loss": 1.3849,
"step": 165
},
{
"epoch": 0.01,
"grad_norm": 2.9938901642866718,
"learning_rate": 8.721787738500866e-07,
"loss": 1.4267,
"step": 166
},
{
"epoch": 0.01,
"grad_norm": 2.923274256584191,
"learning_rate": 8.721779930409007e-07,
"loss": 1.4283,
"step": 167
},
{
"epoch": 0.01,
"grad_norm": 2.8466301519873785,
"learning_rate": 8.721772004906999e-07,
"loss": 1.3842,
"step": 168
},
{
"epoch": 0.01,
"grad_norm": 3.0052783960700165,
"learning_rate": 8.721763961995056e-07,
"loss": 1.4335,
"step": 169
},
{
"epoch": 0.01,
"grad_norm": 3.0935980229307614,
"learning_rate": 8.721755801673391e-07,
"loss": 1.3751,
"step": 170
},
{
"epoch": 0.01,
"grad_norm": 3.0201926532008505,
"learning_rate": 8.721747523942229e-07,
"loss": 1.383,
"step": 171
},
{
"epoch": 0.01,
"grad_norm": 2.964731945306275,
"learning_rate": 8.721739128801788e-07,
"loss": 1.3359,
"step": 172
},
{
"epoch": 0.01,
"grad_norm": 2.928301273992501,
"learning_rate": 8.721730616252297e-07,
"loss": 1.3461,
"step": 173
},
{
"epoch": 0.01,
"grad_norm": 2.9759904501938617,
"learning_rate": 8.721721986293985e-07,
"loss": 1.3644,
"step": 174
},
{
"epoch": 0.01,
"grad_norm": 3.040389211247673,
"learning_rate": 8.721713238927082e-07,
"loss": 1.4341,
"step": 175
},
{
"epoch": 0.01,
"grad_norm": 3.005645814777094,
"learning_rate": 8.721704374151826e-07,
"loss": 1.3967,
"step": 176
},
{
"epoch": 0.01,
"grad_norm": 3.1272949994089823,
"learning_rate": 8.721695391968456e-07,
"loss": 1.3796,
"step": 177
},
{
"epoch": 0.01,
"grad_norm": 3.0684527006439533,
"learning_rate": 8.721686292377211e-07,
"loss": 1.3905,
"step": 178
},
{
"epoch": 0.01,
"grad_norm": 2.9348716760076603,
"learning_rate": 8.721677075378338e-07,
"loss": 1.3905,
"step": 179
},
{
"epoch": 0.01,
"grad_norm": 2.9873597802095304,
"learning_rate": 8.721667740972085e-07,
"loss": 1.4103,
"step": 180
},
{
"epoch": 0.01,
"grad_norm": 2.979793648840201,
"learning_rate": 8.721658289158703e-07,
"loss": 1.3622,
"step": 181
},
{
"epoch": 0.01,
"grad_norm": 2.937468774579186,
"learning_rate": 8.721648719938447e-07,
"loss": 1.414,
"step": 182
},
{
"epoch": 0.01,
"grad_norm": 2.879104091071243,
"learning_rate": 8.721639033311573e-07,
"loss": 1.3108,
"step": 183
},
{
"epoch": 0.01,
"grad_norm": 3.0663878291218203,
"learning_rate": 8.721629229278344e-07,
"loss": 1.3543,
"step": 184
},
{
"epoch": 0.01,
"grad_norm": 2.9407287447315826,
"learning_rate": 8.721619307839025e-07,
"loss": 1.3753,
"step": 185
},
{
"epoch": 0.01,
"grad_norm": 2.967538550932994,
"learning_rate": 8.721609268993879e-07,
"loss": 1.3973,
"step": 186
},
{
"epoch": 0.01,
"grad_norm": 3.057519293009879,
"learning_rate": 8.721599112743179e-07,
"loss": 1.4036,
"step": 187
},
{
"epoch": 0.01,
"grad_norm": 2.936392616519391,
"learning_rate": 8.721588839087197e-07,
"loss": 1.4852,
"step": 188
},
{
"epoch": 0.01,
"grad_norm": 2.9670464594249197,
"learning_rate": 8.721578448026212e-07,
"loss": 1.3643,
"step": 189
},
{
"epoch": 0.01,
"grad_norm": 3.0273720809079663,
"learning_rate": 8.721567939560502e-07,
"loss": 1.4109,
"step": 190
},
{
"epoch": 0.01,
"grad_norm": 3.0651462806238854,
"learning_rate": 8.721557313690349e-07,
"loss": 1.3599,
"step": 191
},
{
"epoch": 0.01,
"grad_norm": 2.9487059919929326,
"learning_rate": 8.721546570416042e-07,
"loss": 1.3377,
"step": 192
},
{
"epoch": 0.01,
"grad_norm": 3.0998938976371146,
"learning_rate": 8.721535709737867e-07,
"loss": 1.3685,
"step": 193
},
{
"epoch": 0.01,
"grad_norm": 2.940826121224176,
"learning_rate": 8.721524731656118e-07,
"loss": 1.4174,
"step": 194
},
{
"epoch": 0.01,
"grad_norm": 3.014412763659776,
"learning_rate": 8.721513636171093e-07,
"loss": 1.3758,
"step": 195
},
{
"epoch": 0.01,
"grad_norm": 3.0608004542672678,
"learning_rate": 8.721502423283086e-07,
"loss": 1.3716,
"step": 196
},
{
"epoch": 0.01,
"grad_norm": 2.94244083669587,
"learning_rate": 8.721491092992403e-07,
"loss": 1.3937,
"step": 197
},
{
"epoch": 0.01,
"grad_norm": 2.9911802591493144,
"learning_rate": 8.721479645299345e-07,
"loss": 1.4164,
"step": 198
},
{
"epoch": 0.01,
"grad_norm": 3.0344694715702065,
"learning_rate": 8.721468080204223e-07,
"loss": 1.4167,
"step": 199
},
{
"epoch": 0.01,
"grad_norm": 2.865667879330454,
"learning_rate": 8.72145639770735e-07,
"loss": 1.4041,
"step": 200
},
{
"epoch": 0.01,
"grad_norm": 3.072739743995507,
"learning_rate": 8.721444597809037e-07,
"loss": 1.4133,
"step": 201
},
{
"epoch": 0.01,
"grad_norm": 3.0276333439793843,
"learning_rate": 8.721432680509603e-07,
"loss": 1.3605,
"step": 202
},
{
"epoch": 0.01,
"grad_norm": 3.0465572110487686,
"learning_rate": 8.721420645809369e-07,
"loss": 1.3134,
"step": 203
},
{
"epoch": 0.01,
"grad_norm": 2.9542599421921376,
"learning_rate": 8.721408493708659e-07,
"loss": 1.4148,
"step": 204
},
{
"epoch": 0.01,
"grad_norm": 2.90720227559915,
"learning_rate": 8.721396224207801e-07,
"loss": 1.3997,
"step": 205
},
{
"epoch": 0.01,
"grad_norm": 3.0246259249156617,
"learning_rate": 8.721383837307123e-07,
"loss": 1.4238,
"step": 206
},
{
"epoch": 0.01,
"grad_norm": 2.9975853807488453,
"learning_rate": 8.721371333006962e-07,
"loss": 1.3879,
"step": 207
},
{
"epoch": 0.01,
"grad_norm": 3.0021383888830258,
"learning_rate": 8.721358711307651e-07,
"loss": 1.3349,
"step": 208
},
{
"epoch": 0.01,
"grad_norm": 3.0201205392426296,
"learning_rate": 8.721345972209533e-07,
"loss": 1.3692,
"step": 209
},
{
"epoch": 0.01,
"grad_norm": 2.972139743981842,
"learning_rate": 8.721333115712948e-07,
"loss": 1.3856,
"step": 210
},
{
"epoch": 0.01,
"grad_norm": 2.9209251468195276,
"learning_rate": 8.721320141818245e-07,
"loss": 1.3726,
"step": 211
},
{
"epoch": 0.01,
"grad_norm": 2.9320196507721277,
"learning_rate": 8.721307050525772e-07,
"loss": 1.4143,
"step": 212
},
{
"epoch": 0.01,
"grad_norm": 3.0002549106343337,
"learning_rate": 8.72129384183588e-07,
"loss": 1.3897,
"step": 213
},
{
"epoch": 0.01,
"grad_norm": 2.9582570275362206,
"learning_rate": 8.721280515748928e-07,
"loss": 1.3756,
"step": 214
},
{
"epoch": 0.01,
"grad_norm": 3.0145408653891526,
"learning_rate": 8.721267072265271e-07,
"loss": 1.3929,
"step": 215
},
{
"epoch": 0.01,
"grad_norm": 3.0495215598431553,
"learning_rate": 8.721253511385274e-07,
"loss": 1.4061,
"step": 216
},
{
"epoch": 0.01,
"grad_norm": 2.9254365712957613,
"learning_rate": 8.721239833109302e-07,
"loss": 1.3903,
"step": 217
},
{
"epoch": 0.01,
"grad_norm": 2.997455071778174,
"learning_rate": 8.72122603743772e-07,
"loss": 1.4246,
"step": 218
},
{
"epoch": 0.01,
"grad_norm": 2.9195114563849627,
"learning_rate": 8.721212124370902e-07,
"loss": 1.3968,
"step": 219
},
{
"epoch": 0.01,
"grad_norm": 3.0160422542520706,
"learning_rate": 8.721198093909225e-07,
"loss": 1.4347,
"step": 220
},
{
"epoch": 0.01,
"grad_norm": 3.01404121750226,
"learning_rate": 8.721183946053062e-07,
"loss": 1.3945,
"step": 221
},
{
"epoch": 0.01,
"grad_norm": 3.0611502119276692,
"learning_rate": 8.721169680802796e-07,
"loss": 1.3975,
"step": 222
},
{
"epoch": 0.01,
"grad_norm": 3.0684020412598727,
"learning_rate": 8.721155298158811e-07,
"loss": 1.373,
"step": 223
},
{
"epoch": 0.01,
"grad_norm": 2.9440206694677027,
"learning_rate": 8.721140798121494e-07,
"loss": 1.3432,
"step": 224
},
{
"epoch": 0.01,
"grad_norm": 2.974581850771121,
"learning_rate": 8.721126180691237e-07,
"loss": 1.3095,
"step": 225
},
{
"epoch": 0.01,
"grad_norm": 2.9933489105960844,
"learning_rate": 8.721111445868431e-07,
"loss": 1.3885,
"step": 226
},
{
"epoch": 0.01,
"grad_norm": 2.9831661987665528,
"learning_rate": 8.721096593653475e-07,
"loss": 1.3126,
"step": 227
},
{
"epoch": 0.01,
"grad_norm": 3.100080295310524,
"learning_rate": 8.721081624046766e-07,
"loss": 1.3567,
"step": 228
},
{
"epoch": 0.01,
"grad_norm": 2.989885190608965,
"learning_rate": 8.72106653704871e-07,
"loss": 1.3899,
"step": 229
},
{
"epoch": 0.01,
"grad_norm": 3.0712176271885023,
"learning_rate": 8.721051332659713e-07,
"loss": 1.4208,
"step": 230
},
{
"epoch": 0.01,
"grad_norm": 3.062312373029536,
"learning_rate": 8.721036010880183e-07,
"loss": 1.4147,
"step": 231
},
{
"epoch": 0.01,
"grad_norm": 2.9701616634317083,
"learning_rate": 8.721020571710533e-07,
"loss": 1.434,
"step": 232
},
{
"epoch": 0.01,
"grad_norm": 2.9572643731393646,
"learning_rate": 8.721005015151179e-07,
"loss": 1.3795,
"step": 233
},
{
"epoch": 0.01,
"grad_norm": 2.997406682050713,
"learning_rate": 8.720989341202539e-07,
"loss": 1.4501,
"step": 234
},
{
"epoch": 0.01,
"grad_norm": 2.9106126480356522,
"learning_rate": 8.720973549865035e-07,
"loss": 1.3684,
"step": 235
},
{
"epoch": 0.01,
"grad_norm": 2.9444070042644817,
"learning_rate": 8.720957641139094e-07,
"loss": 1.4213,
"step": 236
},
{
"epoch": 0.01,
"grad_norm": 2.8728037311842822,
"learning_rate": 8.720941615025142e-07,
"loss": 1.3519,
"step": 237
},
{
"epoch": 0.01,
"grad_norm": 3.015889105815668,
"learning_rate": 8.720925471523613e-07,
"loss": 1.4162,
"step": 238
},
{
"epoch": 0.01,
"grad_norm": 2.9419377055914744,
"learning_rate": 8.72090921063494e-07,
"loss": 1.3357,
"step": 239
},
{
"epoch": 0.01,
"grad_norm": 2.89008663153287,
"learning_rate": 8.720892832359559e-07,
"loss": 1.3647,
"step": 240
},
{
"epoch": 0.01,
"grad_norm": 2.9219232048658736,
"learning_rate": 8.720876336697914e-07,
"loss": 1.4069,
"step": 241
},
{
"epoch": 0.01,
"grad_norm": 2.96537590149616,
"learning_rate": 8.72085972365045e-07,
"loss": 1.4118,
"step": 242
},
{
"epoch": 0.01,
"grad_norm": 2.883655732971505,
"learning_rate": 8.720842993217609e-07,
"loss": 1.4136,
"step": 243
},
{
"epoch": 0.01,
"grad_norm": 2.9788747864733764,
"learning_rate": 8.720826145399848e-07,
"loss": 1.3976,
"step": 244
},
{
"epoch": 0.01,
"grad_norm": 2.875570982035785,
"learning_rate": 8.720809180197616e-07,
"loss": 1.426,
"step": 245
},
{
"epoch": 0.01,
"grad_norm": 2.9926412719867304,
"learning_rate": 8.720792097611372e-07,
"loss": 1.3629,
"step": 246
},
{
"epoch": 0.01,
"grad_norm": 2.958723584893194,
"learning_rate": 8.720774897641574e-07,
"loss": 1.3918,
"step": 247
},
{
"epoch": 0.01,
"grad_norm": 2.968992238648431,
"learning_rate": 8.720757580288688e-07,
"loss": 1.4241,
"step": 248
},
{
"epoch": 0.01,
"grad_norm": 2.889688463405204,
"learning_rate": 8.720740145553177e-07,
"loss": 1.4101,
"step": 249
},
{
"epoch": 0.01,
"grad_norm": 2.9993006762652312,
"learning_rate": 8.720722593435512e-07,
"loss": 1.3857,
"step": 250
},
{
"epoch": 0.01,
"grad_norm": 2.980847240255761,
"learning_rate": 8.720704923936167e-07,
"loss": 1.4077,
"step": 251
},
{
"epoch": 0.01,
"grad_norm": 2.884007977441845,
"learning_rate": 8.720687137055615e-07,
"loss": 1.3822,
"step": 252
},
{
"epoch": 0.01,
"grad_norm": 2.9646728227580645,
"learning_rate": 8.720669232794336e-07,
"loss": 1.3737,
"step": 253
},
{
"epoch": 0.01,
"grad_norm": 2.893382783809774,
"learning_rate": 8.720651211152813e-07,
"loss": 1.3762,
"step": 254
},
{
"epoch": 0.01,
"grad_norm": 3.024418707419624,
"learning_rate": 8.72063307213153e-07,
"loss": 1.3546,
"step": 255
},
{
"epoch": 0.01,
"grad_norm": 3.0252657870696495,
"learning_rate": 8.720614815730977e-07,
"loss": 1.3661,
"step": 256
},
{
"epoch": 0.01,
"grad_norm": 3.0208601885030606,
"learning_rate": 8.720596441951642e-07,
"loss": 1.4182,
"step": 257
},
{
"epoch": 0.01,
"grad_norm": 3.0552975630549954,
"learning_rate": 8.720577950794024e-07,
"loss": 1.38,
"step": 258
},
{
"epoch": 0.01,
"grad_norm": 2.916749346833794,
"learning_rate": 8.720559342258619e-07,
"loss": 1.4049,
"step": 259
},
{
"epoch": 0.01,
"grad_norm": 3.035247531851327,
"learning_rate": 8.720540616345928e-07,
"loss": 1.4256,
"step": 260
},
{
"epoch": 0.01,
"grad_norm": 2.9902303644665205,
"learning_rate": 8.720521773056454e-07,
"loss": 1.3356,
"step": 261
},
{
"epoch": 0.01,
"grad_norm": 3.119044393884763,
"learning_rate": 8.720502812390706e-07,
"loss": 1.4103,
"step": 262
},
{
"epoch": 0.01,
"grad_norm": 2.9552269954583803,
"learning_rate": 8.720483734349194e-07,
"loss": 1.3855,
"step": 263
},
{
"epoch": 0.01,
"grad_norm": 3.017213443982555,
"learning_rate": 8.720464538932433e-07,
"loss": 1.3902,
"step": 264
},
{
"epoch": 0.01,
"grad_norm": 2.925567645830183,
"learning_rate": 8.720445226140937e-07,
"loss": 1.4519,
"step": 265
},
{
"epoch": 0.01,
"grad_norm": 2.983266195022755,
"learning_rate": 8.720425795975228e-07,
"loss": 1.3971,
"step": 266
},
{
"epoch": 0.01,
"grad_norm": 2.992218221530755,
"learning_rate": 8.720406248435828e-07,
"loss": 1.4231,
"step": 267
},
{
"epoch": 0.01,
"grad_norm": 2.9250759809857882,
"learning_rate": 8.720386583523264e-07,
"loss": 1.3877,
"step": 268
},
{
"epoch": 0.01,
"grad_norm": 3.022334104434051,
"learning_rate": 8.720366801238065e-07,
"loss": 1.4133,
"step": 269
},
{
"epoch": 0.01,
"grad_norm": 2.90248743689466,
"learning_rate": 8.720346901580765e-07,
"loss": 1.3889,
"step": 270
},
{
"epoch": 0.01,
"grad_norm": 3.006981215139682,
"learning_rate": 8.720326884551899e-07,
"loss": 1.3657,
"step": 271
},
{
"epoch": 0.01,
"grad_norm": 2.996511837199606,
"learning_rate": 8.720306750152005e-07,
"loss": 1.3918,
"step": 272
},
{
"epoch": 0.01,
"grad_norm": 2.9853146230235317,
"learning_rate": 8.720286498381625e-07,
"loss": 1.3983,
"step": 273
},
{
"epoch": 0.01,
"grad_norm": 2.936322205558776,
"learning_rate": 8.720266129241307e-07,
"loss": 1.3549,
"step": 274
},
{
"epoch": 0.01,
"grad_norm": 3.0336499138604984,
"learning_rate": 8.720245642731596e-07,
"loss": 1.3614,
"step": 275
},
{
"epoch": 0.01,
"grad_norm": 2.9761367095195514,
"learning_rate": 8.720225038853046e-07,
"loss": 1.4223,
"step": 276
},
{
"epoch": 0.01,
"grad_norm": 2.8383029378391256,
"learning_rate": 8.72020431760621e-07,
"loss": 1.3027,
"step": 277
},
{
"epoch": 0.01,
"grad_norm": 3.0012216332284964,
"learning_rate": 8.720183478991647e-07,
"loss": 1.307,
"step": 278
},
{
"epoch": 0.01,
"grad_norm": 3.202220665410359,
"learning_rate": 8.720162523009919e-07,
"loss": 1.3495,
"step": 279
},
{
"epoch": 0.01,
"grad_norm": 2.926537743004413,
"learning_rate": 8.720141449661587e-07,
"loss": 1.346,
"step": 280
},
{
"epoch": 0.01,
"grad_norm": 3.025411005245412,
"learning_rate": 8.720120258947223e-07,
"loss": 1.3581,
"step": 281
},
{
"epoch": 0.01,
"grad_norm": 3.010956090798263,
"learning_rate": 8.720098950867392e-07,
"loss": 1.3634,
"step": 282
},
{
"epoch": 0.01,
"grad_norm": 3.0383205174055727,
"learning_rate": 8.720077525422671e-07,
"loss": 1.3642,
"step": 283
},
{
"epoch": 0.01,
"grad_norm": 2.967895550740301,
"learning_rate": 8.720055982613638e-07,
"loss": 1.3841,
"step": 284
},
{
"epoch": 0.01,
"grad_norm": 2.9365373503076246,
"learning_rate": 8.720034322440872e-07,
"loss": 1.3527,
"step": 285
},
{
"epoch": 0.01,
"grad_norm": 3.1104370711279214,
"learning_rate": 8.720012544904955e-07,
"loss": 1.3483,
"step": 286
},
{
"epoch": 0.01,
"grad_norm": 2.952339289555111,
"learning_rate": 8.719990650006473e-07,
"loss": 1.3956,
"step": 287
},
{
"epoch": 0.01,
"grad_norm": 2.942959986729864,
"learning_rate": 8.719968637746018e-07,
"loss": 1.4256,
"step": 288
},
{
"epoch": 0.02,
"grad_norm": 3.1019407497257507,
"learning_rate": 8.71994650812418e-07,
"loss": 1.3786,
"step": 289
},
{
"epoch": 0.02,
"grad_norm": 3.0501476908245984,
"learning_rate": 8.719924261141557e-07,
"loss": 1.4158,
"step": 290
},
{
"epoch": 0.02,
"grad_norm": 2.9781394711393507,
"learning_rate": 8.719901896798748e-07,
"loss": 1.427,
"step": 291
},
{
"epoch": 0.02,
"grad_norm": 3.0411329565229646,
"learning_rate": 8.719879415096352e-07,
"loss": 1.4281,
"step": 292
},
{
"epoch": 0.02,
"grad_norm": 3.022154009359811,
"learning_rate": 8.719856816034978e-07,
"loss": 1.435,
"step": 293
},
{
"epoch": 0.02,
"grad_norm": 2.965921869395771,
"learning_rate": 8.719834099615232e-07,
"loss": 1.3766,
"step": 294
},
{
"epoch": 0.02,
"grad_norm": 3.026938014636579,
"learning_rate": 8.719811265837728e-07,
"loss": 1.3612,
"step": 295
},
{
"epoch": 0.02,
"grad_norm": 2.9969914810093115,
"learning_rate": 8.719788314703078e-07,
"loss": 1.3371,
"step": 296
},
{
"epoch": 0.02,
"grad_norm": 2.8906070169866545,
"learning_rate": 8.719765246211902e-07,
"loss": 1.3826,
"step": 297
},
{
"epoch": 0.02,
"grad_norm": 3.0301378229830593,
"learning_rate": 8.71974206036482e-07,
"loss": 1.3937,
"step": 298
},
{
"epoch": 0.02,
"grad_norm": 2.9673359121672145,
"learning_rate": 8.719718757162457e-07,
"loss": 1.3838,
"step": 299
},
{
"epoch": 0.02,
"grad_norm": 2.972360304451488,
"learning_rate": 8.719695336605439e-07,
"loss": 1.4382,
"step": 300
},
{
"epoch": 0.02,
"grad_norm": 2.9493767886841242,
"learning_rate": 8.7196717986944e-07,
"loss": 1.3651,
"step": 301
},
{
"epoch": 0.02,
"grad_norm": 3.031661286320472,
"learning_rate": 8.719648143429969e-07,
"loss": 1.3482,
"step": 302
},
{
"epoch": 0.02,
"grad_norm": 2.920995409830151,
"learning_rate": 8.719624370812787e-07,
"loss": 1.4115,
"step": 303
},
{
"epoch": 0.02,
"grad_norm": 2.920727312220773,
"learning_rate": 8.719600480843491e-07,
"loss": 1.396,
"step": 304
},
{
"epoch": 0.02,
"grad_norm": 2.945684201201641,
"learning_rate": 8.719576473522726e-07,
"loss": 1.3557,
"step": 305
},
{
"epoch": 0.02,
"grad_norm": 3.0651768014119956,
"learning_rate": 8.719552348851139e-07,
"loss": 1.389,
"step": 306
},
{
"epoch": 0.02,
"grad_norm": 3.1250149616393577,
"learning_rate": 8.719528106829378e-07,
"loss": 1.469,
"step": 307
},
{
"epoch": 0.02,
"grad_norm": 2.977539941978143,
"learning_rate": 8.719503747458096e-07,
"loss": 1.3536,
"step": 308
},
{
"epoch": 0.02,
"grad_norm": 3.0745479693463924,
"learning_rate": 8.71947927073795e-07,
"loss": 1.3877,
"step": 309
},
{
"epoch": 0.02,
"grad_norm": 3.2160266553797667,
"learning_rate": 8.719454676669596e-07,
"loss": 1.3988,
"step": 310
},
{
"epoch": 0.02,
"grad_norm": 3.134783336833123,
"learning_rate": 8.719429965253698e-07,
"loss": 1.4104,
"step": 311
},
{
"epoch": 0.02,
"grad_norm": 2.908012731710042,
"learning_rate": 8.719405136490924e-07,
"loss": 1.4186,
"step": 312
},
{
"epoch": 0.02,
"grad_norm": 2.9868078254055934,
"learning_rate": 8.71938019038194e-07,
"loss": 1.2836,
"step": 313
},
{
"epoch": 0.02,
"grad_norm": 9.747982306246115,
"learning_rate": 8.719355126927416e-07,
"loss": 1.3331,
"step": 314
},
{
"epoch": 0.02,
"grad_norm": 2.9503527741257476,
"learning_rate": 8.719329946128029e-07,
"loss": 1.3993,
"step": 315
},
{
"epoch": 0.02,
"grad_norm": 3.000724848973538,
"learning_rate": 8.719304647984458e-07,
"loss": 1.3621,
"step": 316
},
{
"epoch": 0.02,
"grad_norm": 3.0259025600243032,
"learning_rate": 8.719279232497381e-07,
"loss": 1.4128,
"step": 317
},
{
"epoch": 0.02,
"grad_norm": 2.95272448132235,
"learning_rate": 8.719253699667485e-07,
"loss": 1.4239,
"step": 318
},
{
"epoch": 0.02,
"grad_norm": 2.7978451838553986,
"learning_rate": 8.719228049495456e-07,
"loss": 1.3694,
"step": 319
},
{
"epoch": 0.02,
"grad_norm": 2.976459149812301,
"learning_rate": 8.719202281981985e-07,
"loss": 1.3413,
"step": 320
},
{
"epoch": 0.02,
"grad_norm": 2.9711797841726377,
"learning_rate": 8.719176397127765e-07,
"loss": 1.3616,
"step": 321
},
{
"epoch": 0.02,
"grad_norm": 3.0204333297285815,
"learning_rate": 8.719150394933495e-07,
"loss": 1.4126,
"step": 322
},
{
"epoch": 0.02,
"grad_norm": 3.0181766925849556,
"learning_rate": 8.719124275399874e-07,
"loss": 1.4323,
"step": 323
},
{
"epoch": 0.02,
"grad_norm": 2.9167139561107893,
"learning_rate": 8.719098038527604e-07,
"loss": 1.4484,
"step": 324
},
{
"epoch": 0.02,
"grad_norm": 2.9378575085429013,
"learning_rate": 8.719071684317393e-07,
"loss": 1.3775,
"step": 325
},
{
"epoch": 0.02,
"grad_norm": 2.94943978740792,
"learning_rate": 8.719045212769951e-07,
"loss": 1.3897,
"step": 326
},
{
"epoch": 0.02,
"grad_norm": 2.958505276332381,
"learning_rate": 8.719018623885988e-07,
"loss": 1.394,
"step": 327
},
{
"epoch": 0.02,
"grad_norm": 3.145867503995205,
"learning_rate": 8.718991917666222e-07,
"loss": 1.379,
"step": 328
},
{
"epoch": 0.02,
"grad_norm": 2.9734480635815728,
"learning_rate": 8.718965094111372e-07,
"loss": 1.3953,
"step": 329
},
{
"epoch": 0.02,
"grad_norm": 2.9049260797132335,
"learning_rate": 8.71893815322216e-07,
"loss": 1.3827,
"step": 330
},
{
"epoch": 0.02,
"grad_norm": 2.98953226106279,
"learning_rate": 8.718911094999311e-07,
"loss": 1.3862,
"step": 331
},
{
"epoch": 0.02,
"grad_norm": 2.898375435464832,
"learning_rate": 8.718883919443554e-07,
"loss": 1.4134,
"step": 332
},
{
"epoch": 0.02,
"grad_norm": 2.9986915176885027,
"learning_rate": 8.718856626555621e-07,
"loss": 1.391,
"step": 333
},
{
"epoch": 0.02,
"grad_norm": 3.054114308668177,
"learning_rate": 8.718829216336246e-07,
"loss": 1.414,
"step": 334
},
{
"epoch": 0.02,
"grad_norm": 2.966556344517333,
"learning_rate": 8.718801688786166e-07,
"loss": 1.4188,
"step": 335
},
{
"epoch": 0.02,
"grad_norm": 3.1040321426982134,
"learning_rate": 8.718774043906126e-07,
"loss": 1.3538,
"step": 336
},
{
"epoch": 0.02,
"grad_norm": 3.0055709258768832,
"learning_rate": 8.718746281696866e-07,
"loss": 1.4413,
"step": 337
},
{
"epoch": 0.02,
"grad_norm": 2.9449143169829277,
"learning_rate": 8.718718402159136e-07,
"loss": 1.3449,
"step": 338
},
{
"epoch": 0.02,
"grad_norm": 2.9706428099571305,
"learning_rate": 8.718690405293686e-07,
"loss": 1.4158,
"step": 339
},
{
"epoch": 0.02,
"grad_norm": 3.0814964982203112,
"learning_rate": 8.718662291101268e-07,
"loss": 1.3981,
"step": 340
},
{
"epoch": 0.02,
"grad_norm": 3.036501892302947,
"learning_rate": 8.718634059582641e-07,
"loss": 1.4047,
"step": 341
},
{
"epoch": 0.02,
"grad_norm": 3.0667912243139535,
"learning_rate": 8.718605710738567e-07,
"loss": 1.4436,
"step": 342
},
{
"epoch": 0.02,
"grad_norm": 3.096788465549673,
"learning_rate": 8.718577244569806e-07,
"loss": 1.4332,
"step": 343
},
{
"epoch": 0.02,
"grad_norm": 3.031370455846918,
"learning_rate": 8.718548661077125e-07,
"loss": 1.3962,
"step": 344
},
{
"epoch": 0.02,
"grad_norm": 2.9672083000240344,
"learning_rate": 8.718519960261294e-07,
"loss": 1.4205,
"step": 345
},
{
"epoch": 0.02,
"grad_norm": 3.007883947100492,
"learning_rate": 8.718491142123086e-07,
"loss": 1.3446,
"step": 346
},
{
"epoch": 0.02,
"grad_norm": 2.9748300479515253,
"learning_rate": 8.718462206663277e-07,
"loss": 1.3854,
"step": 347
},
{
"epoch": 0.02,
"grad_norm": 3.0397247689440396,
"learning_rate": 8.718433153882645e-07,
"loss": 1.4125,
"step": 348
},
{
"epoch": 0.02,
"grad_norm": 3.001960481680682,
"learning_rate": 8.718403983781974e-07,
"loss": 1.3947,
"step": 349
},
{
"epoch": 0.02,
"grad_norm": 3.0664806351413088,
"learning_rate": 8.718374696362047e-07,
"loss": 1.3624,
"step": 350
},
{
"epoch": 0.02,
"grad_norm": 3.0210504760194175,
"learning_rate": 8.718345291623656e-07,
"loss": 1.4671,
"step": 351
},
{
"epoch": 0.02,
"grad_norm": 2.971388652881855,
"learning_rate": 8.718315769567588e-07,
"loss": 1.3472,
"step": 352
},
{
"epoch": 0.02,
"grad_norm": 2.952431051174851,
"learning_rate": 8.718286130194643e-07,
"loss": 1.3779,
"step": 353
},
{
"epoch": 0.02,
"grad_norm": 3.12691241920494,
"learning_rate": 8.718256373505615e-07,
"loss": 1.4117,
"step": 354
},
{
"epoch": 0.02,
"grad_norm": 3.0148285597590796,
"learning_rate": 8.718226499501307e-07,
"loss": 1.3676,
"step": 355
},
{
"epoch": 0.02,
"grad_norm": 2.98672263886442,
"learning_rate": 8.718196508182523e-07,
"loss": 1.4435,
"step": 356
},
{
"epoch": 0.02,
"grad_norm": 2.968930206941385,
"learning_rate": 8.718166399550071e-07,
"loss": 1.4378,
"step": 357
},
{
"epoch": 0.02,
"grad_norm": 3.0425445228617187,
"learning_rate": 8.718136173604761e-07,
"loss": 1.3597,
"step": 358
},
{
"epoch": 0.02,
"grad_norm": 2.9276879829658107,
"learning_rate": 8.718105830347405e-07,
"loss": 1.3689,
"step": 359
},
{
"epoch": 0.02,
"grad_norm": 2.9317936344250413,
"learning_rate": 8.718075369778825e-07,
"loss": 1.3721,
"step": 360
},
{
"epoch": 0.02,
"grad_norm": 2.9645490048095637,
"learning_rate": 8.718044791899837e-07,
"loss": 1.3987,
"step": 361
},
{
"epoch": 0.02,
"grad_norm": 3.0365564821005977,
"learning_rate": 8.718014096711265e-07,
"loss": 1.3868,
"step": 362
},
{
"epoch": 0.02,
"grad_norm": 2.9682770215203553,
"learning_rate": 8.717983284213936e-07,
"loss": 1.3415,
"step": 363
},
{
"epoch": 0.02,
"grad_norm": 3.134935849661195,
"learning_rate": 8.717952354408679e-07,
"loss": 1.3293,
"step": 364
},
{
"epoch": 0.02,
"grad_norm": 2.8124045479008384,
"learning_rate": 8.717921307296327e-07,
"loss": 1.4101,
"step": 365
},
{
"epoch": 0.02,
"grad_norm": 2.870855459457384,
"learning_rate": 8.717890142877717e-07,
"loss": 1.4129,
"step": 366
},
{
"epoch": 0.02,
"grad_norm": 2.996734500928963,
"learning_rate": 8.717858861153686e-07,
"loss": 1.4188,
"step": 367
},
{
"epoch": 0.02,
"grad_norm": 2.905301300393811,
"learning_rate": 8.717827462125079e-07,
"loss": 1.3503,
"step": 368
},
{
"epoch": 0.02,
"grad_norm": 2.963215198908182,
"learning_rate": 8.717795945792739e-07,
"loss": 1.3539,
"step": 369
},
{
"epoch": 0.02,
"grad_norm": 3.058342559604312,
"learning_rate": 8.717764312157515e-07,
"loss": 1.3911,
"step": 370
},
{
"epoch": 0.02,
"grad_norm": 2.9650165681938128,
"learning_rate": 8.717732561220258e-07,
"loss": 1.4207,
"step": 371
},
{
"epoch": 0.02,
"grad_norm": 2.9793800118049454,
"learning_rate": 8.717700692981826e-07,
"loss": 1.3691,
"step": 372
},
{
"epoch": 0.02,
"grad_norm": 2.9162869272769556,
"learning_rate": 8.717668707443075e-07,
"loss": 1.395,
"step": 373
},
{
"epoch": 0.02,
"grad_norm": 2.9636673086391485,
"learning_rate": 8.717636604604865e-07,
"loss": 1.4023,
"step": 374
},
{
"epoch": 0.02,
"grad_norm": 2.959298293762338,
"learning_rate": 8.717604384468061e-07,
"loss": 1.4328,
"step": 375
},
{
"epoch": 0.02,
"grad_norm": 2.965235704696395,
"learning_rate": 8.717572047033532e-07,
"loss": 1.4354,
"step": 376
},
{
"epoch": 0.02,
"grad_norm": 3.127226354296215,
"learning_rate": 8.717539592302147e-07,
"loss": 1.3904,
"step": 377
},
{
"epoch": 0.02,
"grad_norm": 2.9772007199986428,
"learning_rate": 8.717507020274781e-07,
"loss": 1.3997,
"step": 378
},
{
"epoch": 0.02,
"grad_norm": 3.1654089044177103,
"learning_rate": 8.717474330952311e-07,
"loss": 1.3664,
"step": 379
},
{
"epoch": 0.02,
"grad_norm": 2.897024330171534,
"learning_rate": 8.717441524335616e-07,
"loss": 1.3815,
"step": 380
},
{
"epoch": 0.02,
"grad_norm": 2.9908064118392947,
"learning_rate": 8.717408600425579e-07,
"loss": 1.4008,
"step": 381
},
{
"epoch": 0.02,
"grad_norm": 3.1674379998791475,
"learning_rate": 8.717375559223089e-07,
"loss": 1.4134,
"step": 382
},
{
"epoch": 0.02,
"grad_norm": 2.810265806813017,
"learning_rate": 8.717342400729033e-07,
"loss": 1.4046,
"step": 383
},
{
"epoch": 0.02,
"grad_norm": 2.9789538191272626,
"learning_rate": 8.717309124944306e-07,
"loss": 1.3957,
"step": 384
},
{
"epoch": 0.02,
"grad_norm": 2.995888914357539,
"learning_rate": 8.717275731869801e-07,
"loss": 1.3823,
"step": 385
},
{
"epoch": 0.02,
"grad_norm": 2.94123727534938,
"learning_rate": 8.71724222150642e-07,
"loss": 1.3577,
"step": 386
},
{
"epoch": 0.02,
"grad_norm": 2.907211393678648,
"learning_rate": 8.717208593855062e-07,
"loss": 1.4016,
"step": 387
},
{
"epoch": 0.02,
"grad_norm": 2.9549644446432546,
"learning_rate": 8.717174848916635e-07,
"loss": 1.3554,
"step": 388
},
{
"epoch": 0.02,
"grad_norm": 3.047404295254929,
"learning_rate": 8.717140986692047e-07,
"loss": 1.3977,
"step": 389
},
{
"epoch": 0.02,
"grad_norm": 3.0015613969735218,
"learning_rate": 8.717107007182211e-07,
"loss": 1.4159,
"step": 390
},
{
"epoch": 0.02,
"grad_norm": 2.99276033713888,
"learning_rate": 8.71707291038804e-07,
"loss": 1.4194,
"step": 391
},
{
"epoch": 0.02,
"grad_norm": 2.9613959661694427,
"learning_rate": 8.717038696310452e-07,
"loss": 1.4072,
"step": 392
},
{
"epoch": 0.02,
"grad_norm": 2.984569722219675,
"learning_rate": 8.717004364950369e-07,
"loss": 1.4018,
"step": 393
},
{
"epoch": 0.02,
"grad_norm": 2.927152138759416,
"learning_rate": 8.716969916308715e-07,
"loss": 1.4038,
"step": 394
},
{
"epoch": 0.02,
"grad_norm": 3.0771539333400764,
"learning_rate": 8.716935350386416e-07,
"loss": 1.3754,
"step": 395
},
{
"epoch": 0.02,
"grad_norm": 2.970985940726889,
"learning_rate": 8.716900667184406e-07,
"loss": 1.4458,
"step": 396
},
{
"epoch": 0.02,
"grad_norm": 3.0928065462412633,
"learning_rate": 8.716865866703617e-07,
"loss": 1.371,
"step": 397
},
{
"epoch": 0.02,
"grad_norm": 2.9309896563133697,
"learning_rate": 8.716830948944986e-07,
"loss": 1.3509,
"step": 398
},
{
"epoch": 0.02,
"grad_norm": 3.0455988834094736,
"learning_rate": 8.716795913909452e-07,
"loss": 1.3827,
"step": 399
},
{
"epoch": 0.02,
"grad_norm": 3.061462286190086,
"learning_rate": 8.716760761597961e-07,
"loss": 1.3926,
"step": 400
},
{
"epoch": 0.02,
"grad_norm": 2.9080714574645516,
"learning_rate": 8.716725492011458e-07,
"loss": 1.4101,
"step": 401
},
{
"epoch": 0.02,
"grad_norm": 2.9046604352395207,
"learning_rate": 8.716690105150891e-07,
"loss": 1.335,
"step": 402
},
{
"epoch": 0.02,
"grad_norm": 2.946411297505849,
"learning_rate": 8.716654601017216e-07,
"loss": 1.4109,
"step": 403
},
{
"epoch": 0.02,
"grad_norm": 2.9811491538335915,
"learning_rate": 8.716618979611386e-07,
"loss": 1.4007,
"step": 404
},
{
"epoch": 0.02,
"grad_norm": 2.828391151750033,
"learning_rate": 8.716583240934361e-07,
"loss": 1.4194,
"step": 405
},
{
"epoch": 0.02,
"grad_norm": 2.861846008744537,
"learning_rate": 8.716547384987104e-07,
"loss": 1.3164,
"step": 406
},
{
"epoch": 0.02,
"grad_norm": 2.8751261295501274,
"learning_rate": 8.716511411770581e-07,
"loss": 1.4447,
"step": 407
},
{
"epoch": 0.02,
"grad_norm": 3.085208227054159,
"learning_rate": 8.716475321285758e-07,
"loss": 1.3732,
"step": 408
},
{
"epoch": 0.02,
"grad_norm": 2.926763955103524,
"learning_rate": 8.716439113533609e-07,
"loss": 1.427,
"step": 409
},
{
"epoch": 0.02,
"grad_norm": 3.021150407067775,
"learning_rate": 8.716402788515107e-07,
"loss": 1.4123,
"step": 410
},
{
"epoch": 0.02,
"grad_norm": 3.0835953078832476,
"learning_rate": 8.716366346231232e-07,
"loss": 1.3225,
"step": 411
},
{
"epoch": 0.02,
"grad_norm": 3.084681723058033,
"learning_rate": 8.716329786682964e-07,
"loss": 1.4007,
"step": 412
},
{
"epoch": 0.02,
"grad_norm": 2.9354891039496507,
"learning_rate": 8.716293109871288e-07,
"loss": 1.374,
"step": 413
},
{
"epoch": 0.02,
"grad_norm": 3.072363269165642,
"learning_rate": 8.71625631579719e-07,
"loss": 1.355,
"step": 414
},
{
"epoch": 0.02,
"grad_norm": 2.9000514864569373,
"learning_rate": 8.716219404461663e-07,
"loss": 1.3718,
"step": 415
},
{
"epoch": 0.02,
"grad_norm": 3.0562397719571766,
"learning_rate": 8.716182375865698e-07,
"loss": 1.3814,
"step": 416
},
{
"epoch": 0.02,
"grad_norm": 3.0526646692685717,
"learning_rate": 8.716145230010296e-07,
"loss": 1.3772,
"step": 417
},
{
"epoch": 0.02,
"grad_norm": 3.0966573543083538,
"learning_rate": 8.716107966896452e-07,
"loss": 1.4287,
"step": 418
},
{
"epoch": 0.02,
"grad_norm": 3.0880361698977263,
"learning_rate": 8.716070586525174e-07,
"loss": 1.3751,
"step": 419
},
{
"epoch": 0.02,
"grad_norm": 2.832576232775376,
"learning_rate": 8.716033088897465e-07,
"loss": 1.416,
"step": 420
},
{
"epoch": 0.02,
"grad_norm": 2.965237748755546,
"learning_rate": 8.715995474014337e-07,
"loss": 1.3621,
"step": 421
},
{
"epoch": 0.02,
"grad_norm": 3.0463957419921783,
"learning_rate": 8.7159577418768e-07,
"loss": 1.357,
"step": 422
},
{
"epoch": 0.02,
"grad_norm": 2.9739015917876723,
"learning_rate": 8.715919892485873e-07,
"loss": 1.3873,
"step": 423
},
{
"epoch": 0.02,
"grad_norm": 2.8676852163892037,
"learning_rate": 8.715881925842573e-07,
"loss": 1.4051,
"step": 424
},
{
"epoch": 0.02,
"grad_norm": 3.088087320089484,
"learning_rate": 8.715843841947923e-07,
"loss": 1.3151,
"step": 425
},
{
"epoch": 0.02,
"grad_norm": 3.0270665713855367,
"learning_rate": 8.715805640802949e-07,
"loss": 1.3679,
"step": 426
},
{
"epoch": 0.02,
"grad_norm": 2.9869536543983366,
"learning_rate": 8.715767322408678e-07,
"loss": 1.3528,
"step": 427
},
{
"epoch": 0.02,
"grad_norm": 2.9112845523257675,
"learning_rate": 8.715728886766143e-07,
"loss": 1.442,
"step": 428
},
{
"epoch": 0.02,
"grad_norm": 3.0049960527344544,
"learning_rate": 8.715690333876378e-07,
"loss": 1.3681,
"step": 429
},
{
"epoch": 0.02,
"grad_norm": 2.9921307040597664,
"learning_rate": 8.715651663740421e-07,
"loss": 1.4314,
"step": 430
},
{
"epoch": 0.02,
"grad_norm": 3.0109768676656605,
"learning_rate": 8.715612876359315e-07,
"loss": 1.3847,
"step": 431
},
{
"epoch": 0.02,
"grad_norm": 3.039247694352697,
"learning_rate": 8.715573971734103e-07,
"loss": 1.4317,
"step": 432
},
{
"epoch": 0.02,
"grad_norm": 3.019898918932762,
"learning_rate": 8.71553494986583e-07,
"loss": 1.3623,
"step": 433
},
{
"epoch": 0.02,
"grad_norm": 3.00522079476457,
"learning_rate": 8.71549581075555e-07,
"loss": 1.3884,
"step": 434
},
{
"epoch": 0.02,
"grad_norm": 2.9610972813683794,
"learning_rate": 8.715456554404316e-07,
"loss": 1.3315,
"step": 435
},
{
"epoch": 0.02,
"grad_norm": 2.967170476790787,
"learning_rate": 8.715417180813185e-07,
"loss": 1.4207,
"step": 436
},
{
"epoch": 0.02,
"grad_norm": 2.9522241028635285,
"learning_rate": 8.715377689983216e-07,
"loss": 1.4012,
"step": 437
},
{
"epoch": 0.02,
"grad_norm": 2.928961724018662,
"learning_rate": 8.715338081915475e-07,
"loss": 1.3869,
"step": 438
},
{
"epoch": 0.02,
"grad_norm": 2.952221930130988,
"learning_rate": 8.715298356611025e-07,
"loss": 1.3703,
"step": 439
},
{
"epoch": 0.02,
"grad_norm": 3.0536781428675335,
"learning_rate": 8.715258514070937e-07,
"loss": 1.3682,
"step": 440
},
{
"epoch": 0.02,
"grad_norm": 3.128913913582799,
"learning_rate": 8.715218554296284e-07,
"loss": 1.3435,
"step": 441
},
{
"epoch": 0.02,
"grad_norm": 3.0301910736766318,
"learning_rate": 8.715178477288141e-07,
"loss": 1.3975,
"step": 442
},
{
"epoch": 0.02,
"grad_norm": 2.962217823506329,
"learning_rate": 8.715138283047589e-07,
"loss": 1.3488,
"step": 443
},
{
"epoch": 0.02,
"grad_norm": 3.0039607070469647,
"learning_rate": 8.715097971575708e-07,
"loss": 1.378,
"step": 444
},
{
"epoch": 0.02,
"grad_norm": 3.0602288102993938,
"learning_rate": 8.715057542873585e-07,
"loss": 1.3572,
"step": 445
},
{
"epoch": 0.02,
"grad_norm": 2.969434960629606,
"learning_rate": 8.715016996942307e-07,
"loss": 1.3713,
"step": 446
},
{
"epoch": 0.02,
"grad_norm": 2.9917667276430477,
"learning_rate": 8.714976333782967e-07,
"loss": 1.4607,
"step": 447
},
{
"epoch": 0.02,
"grad_norm": 3.2046134996002436,
"learning_rate": 8.714935553396659e-07,
"loss": 1.3853,
"step": 448
},
{
"epoch": 0.02,
"grad_norm": 3.1097764355868733,
"learning_rate": 8.714894655784481e-07,
"loss": 1.371,
"step": 449
},
{
"epoch": 0.02,
"grad_norm": 2.9715612348659124,
"learning_rate": 8.714853640947534e-07,
"loss": 1.392,
"step": 450
},
{
"epoch": 0.02,
"grad_norm": 2.9321597700386333,
"learning_rate": 8.714812508886925e-07,
"loss": 1.4051,
"step": 451
},
{
"epoch": 0.02,
"grad_norm": 2.949095796893927,
"learning_rate": 8.714771259603758e-07,
"loss": 1.3469,
"step": 452
},
{
"epoch": 0.02,
"grad_norm": 2.9290487074046756,
"learning_rate": 8.714729893099144e-07,
"loss": 1.3629,
"step": 453
},
{
"epoch": 0.02,
"grad_norm": 3.0295532590400196,
"learning_rate": 8.714688409374198e-07,
"loss": 1.3689,
"step": 454
},
{
"epoch": 0.02,
"grad_norm": 2.965897130998238,
"learning_rate": 8.714646808430036e-07,
"loss": 1.3619,
"step": 455
},
{
"epoch": 0.02,
"grad_norm": 3.0935335333535336,
"learning_rate": 8.714605090267779e-07,
"loss": 1.3781,
"step": 456
},
{
"epoch": 0.02,
"grad_norm": 2.986770714363214,
"learning_rate": 8.71456325488855e-07,
"loss": 1.364,
"step": 457
},
{
"epoch": 0.02,
"grad_norm": 2.950553960144967,
"learning_rate": 8.714521302293475e-07,
"loss": 1.4063,
"step": 458
},
{
"epoch": 0.02,
"grad_norm": 3.014174300238565,
"learning_rate": 8.714479232483683e-07,
"loss": 1.3676,
"step": 459
},
{
"epoch": 0.02,
"grad_norm": 3.24671603526281,
"learning_rate": 8.714437045460308e-07,
"loss": 1.3578,
"step": 460
},
{
"epoch": 0.02,
"grad_norm": 3.1623556544347724,
"learning_rate": 8.714394741224484e-07,
"loss": 1.3645,
"step": 461
},
{
"epoch": 0.02,
"grad_norm": 3.0025864567325122,
"learning_rate": 8.714352319777354e-07,
"loss": 1.3607,
"step": 462
},
{
"epoch": 0.02,
"grad_norm": 2.94717894620833,
"learning_rate": 8.714309781120056e-07,
"loss": 1.3498,
"step": 463
},
{
"epoch": 0.02,
"grad_norm": 2.8964926477515935,
"learning_rate": 8.714267125253735e-07,
"loss": 1.3728,
"step": 464
},
{
"epoch": 0.02,
"grad_norm": 3.1656664213505823,
"learning_rate": 8.714224352179544e-07,
"loss": 1.4176,
"step": 465
},
{
"epoch": 0.02,
"grad_norm": 3.020188270295796,
"learning_rate": 8.71418146189863e-07,
"loss": 1.4194,
"step": 466
},
{
"epoch": 0.02,
"grad_norm": 2.978569270321128,
"learning_rate": 8.71413845441215e-07,
"loss": 1.3749,
"step": 467
},
{
"epoch": 0.02,
"grad_norm": 2.9396855197827283,
"learning_rate": 8.714095329721261e-07,
"loss": 1.3795,
"step": 468
},
{
"epoch": 0.02,
"grad_norm": 2.9422669465457387,
"learning_rate": 8.714052087827125e-07,
"loss": 1.3834,
"step": 469
},
{
"epoch": 0.02,
"grad_norm": 3.010282715520298,
"learning_rate": 8.714008728730907e-07,
"loss": 1.3531,
"step": 470
},
{
"epoch": 0.02,
"grad_norm": 2.9717500175741116,
"learning_rate": 8.713965252433773e-07,
"loss": 1.3667,
"step": 471
},
{
"epoch": 0.02,
"grad_norm": 3.052704318228861,
"learning_rate": 8.713921658936892e-07,
"loss": 1.3456,
"step": 472
},
{
"epoch": 0.02,
"grad_norm": 3.528710625831412,
"learning_rate": 8.713877948241442e-07,
"loss": 1.3936,
"step": 473
},
{
"epoch": 0.02,
"grad_norm": 2.9466834730933833,
"learning_rate": 8.713834120348596e-07,
"loss": 1.3217,
"step": 474
},
{
"epoch": 0.02,
"grad_norm": 3.0996541898765226,
"learning_rate": 8.713790175259536e-07,
"loss": 1.3855,
"step": 475
},
{
"epoch": 0.02,
"grad_norm": 2.982252067970699,
"learning_rate": 8.713746112975446e-07,
"loss": 1.384,
"step": 476
},
{
"epoch": 0.02,
"grad_norm": 3.1109077626663844,
"learning_rate": 8.713701933497509e-07,
"loss": 1.3561,
"step": 477
},
{
"epoch": 0.02,
"grad_norm": 3.0394061264038115,
"learning_rate": 8.713657636826918e-07,
"loss": 1.468,
"step": 478
},
{
"epoch": 0.02,
"grad_norm": 3.0969706383479,
"learning_rate": 8.713613222964863e-07,
"loss": 1.3993,
"step": 479
},
{
"epoch": 0.02,
"grad_norm": 3.0348133446475662,
"learning_rate": 8.713568691912542e-07,
"loss": 1.387,
"step": 480
},
{
"epoch": 0.02,
"grad_norm": 3.0273826285615195,
"learning_rate": 8.713524043671153e-07,
"loss": 1.3959,
"step": 481
},
{
"epoch": 0.03,
"grad_norm": 3.141130816615921,
"learning_rate": 8.713479278241898e-07,
"loss": 1.4479,
"step": 482
},
{
"epoch": 0.03,
"grad_norm": 3.0490824173681945,
"learning_rate": 8.713434395625983e-07,
"loss": 1.3583,
"step": 483
},
{
"epoch": 0.03,
"grad_norm": 2.9474474845957404,
"learning_rate": 8.713389395824614e-07,
"loss": 1.3344,
"step": 484
},
{
"epoch": 0.03,
"grad_norm": 2.9486918699061118,
"learning_rate": 8.713344278839005e-07,
"loss": 1.4022,
"step": 485
},
{
"epoch": 0.03,
"grad_norm": 2.9302492343255344,
"learning_rate": 8.71329904467037e-07,
"loss": 1.3344,
"step": 486
},
{
"epoch": 0.03,
"grad_norm": 2.9531126631866753,
"learning_rate": 8.713253693319929e-07,
"loss": 1.3451,
"step": 487
},
{
"epoch": 0.03,
"grad_norm": 2.8555050387157213,
"learning_rate": 8.713208224788899e-07,
"loss": 1.3287,
"step": 488
},
{
"epoch": 0.03,
"grad_norm": 3.0850474903019305,
"learning_rate": 8.713162639078507e-07,
"loss": 1.4153,
"step": 489
},
{
"epoch": 0.03,
"grad_norm": 2.9916349483264004,
"learning_rate": 8.71311693618998e-07,
"loss": 1.4025,
"step": 490
},
{
"epoch": 0.03,
"grad_norm": 3.0448173115664545,
"learning_rate": 8.713071116124549e-07,
"loss": 1.4129,
"step": 491
},
{
"epoch": 0.03,
"grad_norm": 2.9652773442800022,
"learning_rate": 8.713025178883445e-07,
"loss": 1.3688,
"step": 492
},
{
"epoch": 0.03,
"grad_norm": 2.8707509231071127,
"learning_rate": 8.712979124467906e-07,
"loss": 1.3714,
"step": 493
},
{
"epoch": 0.03,
"grad_norm": 2.968740384281211,
"learning_rate": 8.712932952879176e-07,
"loss": 1.4012,
"step": 494
},
{
"epoch": 0.03,
"grad_norm": 2.938361663169202,
"learning_rate": 8.712886664118492e-07,
"loss": 1.406,
"step": 495
},
{
"epoch": 0.03,
"grad_norm": 3.1217977871072775,
"learning_rate": 8.712840258187104e-07,
"loss": 1.3822,
"step": 496
},
{
"epoch": 0.03,
"grad_norm": 2.9353035115177915,
"learning_rate": 8.71279373508626e-07,
"loss": 1.3578,
"step": 497
},
{
"epoch": 0.03,
"grad_norm": 3.014072001165337,
"learning_rate": 8.712747094817213e-07,
"loss": 1.4454,
"step": 498
},
{
"epoch": 0.03,
"grad_norm": 2.969346395207918,
"learning_rate": 8.71270033738122e-07,
"loss": 1.3705,
"step": 499
},
{
"epoch": 0.03,
"grad_norm": 3.050484611788353,
"learning_rate": 8.712653462779539e-07,
"loss": 1.3551,
"step": 500
},
{
"epoch": 0.03,
"grad_norm": 2.940761479764165,
"learning_rate": 8.71260647101343e-07,
"loss": 1.4147,
"step": 501
},
{
"epoch": 0.03,
"grad_norm": 2.9928956127151944,
"learning_rate": 8.712559362084161e-07,
"loss": 1.3604,
"step": 502
},
{
"epoch": 0.03,
"grad_norm": 3.0125472864930805,
"learning_rate": 8.712512135993e-07,
"loss": 1.4183,
"step": 503
},
{
"epoch": 0.03,
"grad_norm": 2.940381713821346,
"learning_rate": 8.712464792741218e-07,
"loss": 1.4414,
"step": 504
},
{
"epoch": 0.03,
"grad_norm": 3.0279511535103336,
"learning_rate": 8.712417332330089e-07,
"loss": 1.3505,
"step": 505
},
{
"epoch": 0.03,
"grad_norm": 2.9922191158837954,
"learning_rate": 8.712369754760892e-07,
"loss": 1.4028,
"step": 506
},
{
"epoch": 0.03,
"grad_norm": 3.0184990677208834,
"learning_rate": 8.712322060034907e-07,
"loss": 1.3465,
"step": 507
},
{
"epoch": 0.03,
"grad_norm": 3.05423073796875,
"learning_rate": 8.712274248153418e-07,
"loss": 1.3416,
"step": 508
},
{
"epoch": 0.03,
"grad_norm": 3.1326601737410025,
"learning_rate": 8.712226319117715e-07,
"loss": 1.3924,
"step": 509
},
{
"epoch": 0.03,
"grad_norm": 2.995785462842688,
"learning_rate": 8.712178272929084e-07,
"loss": 1.3895,
"step": 510
},
{
"epoch": 0.03,
"grad_norm": 2.9760799944825806,
"learning_rate": 8.712130109588823e-07,
"loss": 1.4104,
"step": 511
},
{
"epoch": 0.03,
"grad_norm": 2.9654696172148896,
"learning_rate": 8.712081829098225e-07,
"loss": 1.378,
"step": 512
},
{
"epoch": 0.03,
"grad_norm": 2.9968608921090265,
"learning_rate": 8.712033431458593e-07,
"loss": 1.4264,
"step": 513
},
{
"epoch": 0.03,
"grad_norm": 2.969211803711655,
"learning_rate": 8.711984916671229e-07,
"loss": 1.3607,
"step": 514
},
{
"epoch": 0.03,
"grad_norm": 2.942118444739339,
"learning_rate": 8.711936284737438e-07,
"loss": 1.3899,
"step": 515
},
{
"epoch": 0.03,
"grad_norm": 2.98791974032871,
"learning_rate": 8.711887535658529e-07,
"loss": 1.3459,
"step": 516
},
{
"epoch": 0.03,
"grad_norm": 2.9216444622419537,
"learning_rate": 8.711838669435818e-07,
"loss": 1.4116,
"step": 517
},
{
"epoch": 0.03,
"grad_norm": 2.9987312946412206,
"learning_rate": 8.711789686070618e-07,
"loss": 1.4126,
"step": 518
},
{
"epoch": 0.03,
"grad_norm": 2.9997602427201926,
"learning_rate": 8.711740585564249e-07,
"loss": 1.3392,
"step": 519
},
{
"epoch": 0.03,
"grad_norm": 2.9736815652309683,
"learning_rate": 8.711691367918032e-07,
"loss": 1.3483,
"step": 520
},
{
"epoch": 0.03,
"grad_norm": 3.0854886077878487,
"learning_rate": 8.711642033133292e-07,
"loss": 1.3842,
"step": 521
},
{
"epoch": 0.03,
"grad_norm": 3.061838785099914,
"learning_rate": 8.711592581211358e-07,
"loss": 1.3658,
"step": 522
},
{
"epoch": 0.03,
"grad_norm": 3.023168820723632,
"learning_rate": 8.711543012153561e-07,
"loss": 1.3929,
"step": 523
},
{
"epoch": 0.03,
"grad_norm": 2.970076769155023,
"learning_rate": 8.711493325961236e-07,
"loss": 1.3587,
"step": 524
},
{
"epoch": 0.03,
"grad_norm": 3.002070690376611,
"learning_rate": 8.71144352263572e-07,
"loss": 1.367,
"step": 525
},
{
"epoch": 0.03,
"grad_norm": 2.995354883459173,
"learning_rate": 8.711393602178357e-07,
"loss": 1.4185,
"step": 526
},
{
"epoch": 0.03,
"grad_norm": 24.26995863156332,
"learning_rate": 8.711343564590487e-07,
"loss": 1.3982,
"step": 527
},
{
"epoch": 0.03,
"grad_norm": 3.065387181865347,
"learning_rate": 8.711293409873459e-07,
"loss": 1.3668,
"step": 528
},
{
"epoch": 0.03,
"grad_norm": 2.965733738900826,
"learning_rate": 8.711243138028624e-07,
"loss": 1.3989,
"step": 529
},
{
"epoch": 0.03,
"grad_norm": 2.909922562005101,
"learning_rate": 8.711192749057334e-07,
"loss": 1.3548,
"step": 530
},
{
"epoch": 0.03,
"grad_norm": 2.9485719715840824,
"learning_rate": 8.711142242960946e-07,
"loss": 1.3982,
"step": 531
},
{
"epoch": 0.03,
"grad_norm": 3.0587521051463598,
"learning_rate": 8.711091619740822e-07,
"loss": 1.4208,
"step": 532
},
{
"epoch": 0.03,
"grad_norm": 3.016741791642564,
"learning_rate": 8.711040879398322e-07,
"loss": 1.3776,
"step": 533
},
{
"epoch": 0.03,
"grad_norm": 3.087201787399773,
"learning_rate": 8.710990021934814e-07,
"loss": 1.3509,
"step": 534
},
{
"epoch": 0.03,
"grad_norm": 3.1398002317055873,
"learning_rate": 8.710939047351665e-07,
"loss": 1.4054,
"step": 535
},
{
"epoch": 0.03,
"grad_norm": 3.003141522348306,
"learning_rate": 8.710887955650252e-07,
"loss": 1.3895,
"step": 536
},
{
"epoch": 0.03,
"grad_norm": 3.0461324715788862,
"learning_rate": 8.710836746831946e-07,
"loss": 1.4143,
"step": 537
},
{
"epoch": 0.03,
"grad_norm": 2.956952816424475,
"learning_rate": 8.710785420898127e-07,
"loss": 1.4008,
"step": 538
},
{
"epoch": 0.03,
"grad_norm": 2.8987555743390168,
"learning_rate": 8.710733977850179e-07,
"loss": 1.3823,
"step": 539
},
{
"epoch": 0.03,
"grad_norm": 3.0536968376300284,
"learning_rate": 8.710682417689485e-07,
"loss": 1.4578,
"step": 540
},
{
"epoch": 0.03,
"grad_norm": 2.969066979149947,
"learning_rate": 8.710630740417435e-07,
"loss": 1.428,
"step": 541
},
{
"epoch": 0.03,
"grad_norm": 3.1272335317584905,
"learning_rate": 8.710578946035417e-07,
"loss": 1.3562,
"step": 542
},
{
"epoch": 0.03,
"grad_norm": 2.9753498641377742,
"learning_rate": 8.710527034544828e-07,
"loss": 1.3953,
"step": 543
},
{
"epoch": 0.03,
"grad_norm": 2.943913278878306,
"learning_rate": 8.710475005947067e-07,
"loss": 1.3626,
"step": 544
},
{
"epoch": 0.03,
"grad_norm": 2.9513572324834727,
"learning_rate": 8.710422860243531e-07,
"loss": 1.3461,
"step": 545
},
{
"epoch": 0.03,
"grad_norm": 3.0849634901753284,
"learning_rate": 8.710370597435629e-07,
"loss": 1.3663,
"step": 546
},
{
"epoch": 0.03,
"grad_norm": 3.138354533416878,
"learning_rate": 8.710318217524763e-07,
"loss": 1.3141,
"step": 547
},
{
"epoch": 0.03,
"grad_norm": 2.972312260895404,
"learning_rate": 8.710265720512346e-07,
"loss": 1.3633,
"step": 548
},
{
"epoch": 0.03,
"grad_norm": 3.0591517440498097,
"learning_rate": 8.710213106399791e-07,
"loss": 1.3557,
"step": 549
},
{
"epoch": 0.03,
"grad_norm": 2.8524186182545987,
"learning_rate": 8.710160375188516e-07,
"loss": 1.3676,
"step": 550
},
{
"epoch": 0.03,
"grad_norm": 2.9822890833732956,
"learning_rate": 8.710107526879938e-07,
"loss": 1.4331,
"step": 551
},
{
"epoch": 0.03,
"grad_norm": 2.9176351551822868,
"learning_rate": 8.710054561475481e-07,
"loss": 1.371,
"step": 552
},
{
"epoch": 0.03,
"grad_norm": 2.94388048856757,
"learning_rate": 8.71000147897657e-07,
"loss": 1.393,
"step": 553
},
{
"epoch": 0.03,
"grad_norm": 3.085541520730895,
"learning_rate": 8.709948279384639e-07,
"loss": 1.3937,
"step": 554
},
{
"epoch": 0.03,
"grad_norm": 2.9491352117439757,
"learning_rate": 8.709894962701115e-07,
"loss": 1.3526,
"step": 555
},
{
"epoch": 0.03,
"grad_norm": 2.9933004115427444,
"learning_rate": 8.709841528927436e-07,
"loss": 1.3916,
"step": 556
},
{
"epoch": 0.03,
"grad_norm": 2.959458221957766,
"learning_rate": 8.70978797806504e-07,
"loss": 1.3462,
"step": 557
},
{
"epoch": 0.03,
"grad_norm": 2.963203467802987,
"learning_rate": 8.709734310115368e-07,
"loss": 1.3783,
"step": 558
},
{
"epoch": 0.03,
"grad_norm": 2.9902951999179312,
"learning_rate": 8.709680525079866e-07,
"loss": 1.362,
"step": 559
},
{
"epoch": 0.03,
"grad_norm": 2.948654339418917,
"learning_rate": 8.709626622959983e-07,
"loss": 1.3841,
"step": 560
},
{
"epoch": 0.03,
"grad_norm": 3.001564005849718,
"learning_rate": 8.709572603757169e-07,
"loss": 1.3572,
"step": 561
},
{
"epoch": 0.03,
"grad_norm": 3.0380595324448416,
"learning_rate": 8.709518467472878e-07,
"loss": 1.4171,
"step": 562
},
{
"epoch": 0.03,
"grad_norm": 2.896631790474921,
"learning_rate": 8.709464214108568e-07,
"loss": 1.3448,
"step": 563
},
{
"epoch": 0.03,
"grad_norm": 2.9319635944784475,
"learning_rate": 8.709409843665701e-07,
"loss": 1.3917,
"step": 564
},
{
"epoch": 0.03,
"grad_norm": 2.963100470234962,
"learning_rate": 8.709355356145739e-07,
"loss": 1.3655,
"step": 565
},
{
"epoch": 0.03,
"grad_norm": 3.030553118915969,
"learning_rate": 8.709300751550151e-07,
"loss": 1.3365,
"step": 566
},
{
"epoch": 0.03,
"grad_norm": 3.4928530382016887,
"learning_rate": 8.709246029880405e-07,
"loss": 1.3662,
"step": 567
},
{
"epoch": 0.03,
"grad_norm": 3.030669221337684,
"learning_rate": 8.709191191137976e-07,
"loss": 1.4529,
"step": 568
},
{
"epoch": 0.03,
"grad_norm": 3.155457882933603,
"learning_rate": 8.70913623532434e-07,
"loss": 1.4022,
"step": 569
},
{
"epoch": 0.03,
"grad_norm": 2.9856956704586115,
"learning_rate": 8.709081162440975e-07,
"loss": 1.3989,
"step": 570
},
{
"epoch": 0.03,
"grad_norm": 2.9343339295028015,
"learning_rate": 8.709025972489367e-07,
"loss": 1.4212,
"step": 571
},
{
"epoch": 0.03,
"grad_norm": 3.1952129991940543,
"learning_rate": 8.708970665471e-07,
"loss": 1.4436,
"step": 572
},
{
"epoch": 0.03,
"grad_norm": 3.015106319776789,
"learning_rate": 8.708915241387364e-07,
"loss": 1.4422,
"step": 573
},
{
"epoch": 0.03,
"grad_norm": 3.022781702862628,
"learning_rate": 8.708859700239951e-07,
"loss": 1.3789,
"step": 574
},
{
"epoch": 0.03,
"grad_norm": 2.932454535389442,
"learning_rate": 8.708804042030254e-07,
"loss": 1.3783,
"step": 575
},
{
"epoch": 0.03,
"grad_norm": 3.1075106961875165,
"learning_rate": 8.708748266759774e-07,
"loss": 1.3904,
"step": 576
},
{
"epoch": 0.03,
"grad_norm": 2.8613959925148356,
"learning_rate": 8.708692374430014e-07,
"loss": 1.3701,
"step": 577
},
{
"epoch": 0.03,
"grad_norm": 2.9916282178733447,
"learning_rate": 8.708636365042476e-07,
"loss": 1.3517,
"step": 578
},
{
"epoch": 0.03,
"grad_norm": 2.8578706799379554,
"learning_rate": 8.70858023859867e-07,
"loss": 1.4124,
"step": 579
},
{
"epoch": 0.03,
"grad_norm": 3.0055644119433533,
"learning_rate": 8.708523995100105e-07,
"loss": 1.3869,
"step": 580
},
{
"epoch": 0.03,
"grad_norm": 3.027942807639001,
"learning_rate": 8.708467634548298e-07,
"loss": 1.3703,
"step": 581
},
{
"epoch": 0.03,
"grad_norm": 3.1584010302430836,
"learning_rate": 8.708411156944765e-07,
"loss": 1.3852,
"step": 582
},
{
"epoch": 0.03,
"grad_norm": 3.08434624814754,
"learning_rate": 8.708354562291027e-07,
"loss": 1.4008,
"step": 583
},
{
"epoch": 0.03,
"grad_norm": 2.937949504985518,
"learning_rate": 8.708297850588607e-07,
"loss": 1.4026,
"step": 584
},
{
"epoch": 0.03,
"grad_norm": 3.0080974250898116,
"learning_rate": 8.708241021839032e-07,
"loss": 1.3992,
"step": 585
},
{
"epoch": 0.03,
"grad_norm": 2.986930909094618,
"learning_rate": 8.708184076043833e-07,
"loss": 1.3504,
"step": 586
},
{
"epoch": 0.03,
"grad_norm": 3.1912323292885363,
"learning_rate": 8.708127013204543e-07,
"loss": 1.3919,
"step": 587
},
{
"epoch": 0.03,
"grad_norm": 3.203030731079236,
"learning_rate": 8.708069833322698e-07,
"loss": 1.3601,
"step": 588
},
{
"epoch": 0.03,
"grad_norm": 3.097916109715531,
"learning_rate": 8.708012536399837e-07,
"loss": 1.3619,
"step": 589
},
{
"epoch": 0.03,
"grad_norm": 3.0656903032705327,
"learning_rate": 8.707955122437504e-07,
"loss": 1.3162,
"step": 590
},
{
"epoch": 0.03,
"grad_norm": 3.002875260439292,
"learning_rate": 8.707897591437243e-07,
"loss": 1.389,
"step": 591
},
{
"epoch": 0.03,
"grad_norm": 3.0143271490705077,
"learning_rate": 8.707839943400606e-07,
"loss": 1.3323,
"step": 592
},
{
"epoch": 0.03,
"grad_norm": 3.096313689982386,
"learning_rate": 8.707782178329142e-07,
"loss": 1.3813,
"step": 593
},
{
"epoch": 0.03,
"grad_norm": 2.960471861335116,
"learning_rate": 8.707724296224408e-07,
"loss": 1.3472,
"step": 594
},
{
"epoch": 0.03,
"grad_norm": 3.0072247729027115,
"learning_rate": 8.707666297087963e-07,
"loss": 1.3522,
"step": 595
},
{
"epoch": 0.03,
"grad_norm": 2.949063388714746,
"learning_rate": 8.707608180921366e-07,
"loss": 1.3928,
"step": 596
},
{
"epoch": 0.03,
"grad_norm": 3.1121870350751064,
"learning_rate": 8.707549947726183e-07,
"loss": 1.4399,
"step": 597
},
{
"epoch": 0.03,
"grad_norm": 2.9609559500955003,
"learning_rate": 8.707491597503982e-07,
"loss": 1.3898,
"step": 598
},
{
"epoch": 0.03,
"grad_norm": 2.9585328248619587,
"learning_rate": 8.707433130256336e-07,
"loss": 1.379,
"step": 599
},
{
"epoch": 0.03,
"grad_norm": 2.918687510312111,
"learning_rate": 8.707374545984816e-07,
"loss": 1.4176,
"step": 600
},
{
"epoch": 0.03,
"grad_norm": 3.15488138702356,
"learning_rate": 8.707315844691002e-07,
"loss": 1.3706,
"step": 601
},
{
"epoch": 0.03,
"grad_norm": 2.9610415881399677,
"learning_rate": 8.707257026376471e-07,
"loss": 1.3641,
"step": 602
},
{
"epoch": 0.03,
"grad_norm": 3.0862098441133234,
"learning_rate": 8.707198091042811e-07,
"loss": 1.3893,
"step": 603
},
{
"epoch": 0.03,
"grad_norm": 2.9328888264242106,
"learning_rate": 8.707139038691606e-07,
"loss": 1.333,
"step": 604
},
{
"epoch": 0.03,
"grad_norm": 2.97580891347495,
"learning_rate": 8.707079869324446e-07,
"loss": 1.3607,
"step": 605
},
{
"epoch": 0.03,
"grad_norm": 3.000962296933563,
"learning_rate": 8.707020582942925e-07,
"loss": 1.424,
"step": 606
},
{
"epoch": 0.03,
"grad_norm": 2.984700606909975,
"learning_rate": 8.706961179548639e-07,
"loss": 1.3912,
"step": 607
},
{
"epoch": 0.03,
"grad_norm": 2.998907818017932,
"learning_rate": 8.706901659143189e-07,
"loss": 1.4241,
"step": 608
},
{
"epoch": 0.03,
"grad_norm": 2.95254670370234,
"learning_rate": 8.706842021728173e-07,
"loss": 1.3759,
"step": 609
},
{
"epoch": 0.03,
"grad_norm": 3.134652976665768,
"learning_rate": 8.706782267305202e-07,
"loss": 1.3767,
"step": 610
},
{
"epoch": 0.03,
"grad_norm": 2.972318633221704,
"learning_rate": 8.706722395875881e-07,
"loss": 1.3648,
"step": 611
},
{
"epoch": 0.03,
"grad_norm": 2.8607560362869124,
"learning_rate": 8.706662407441824e-07,
"loss": 1.3946,
"step": 612
},
{
"epoch": 0.03,
"grad_norm": 2.8921206621422653,
"learning_rate": 8.706602302004645e-07,
"loss": 1.4396,
"step": 613
},
{
"epoch": 0.03,
"grad_norm": 2.949264928829496,
"learning_rate": 8.706542079565962e-07,
"loss": 1.3475,
"step": 614
},
{
"epoch": 0.03,
"grad_norm": 2.996499503605724,
"learning_rate": 8.706481740127399e-07,
"loss": 1.37,
"step": 615
},
{
"epoch": 0.03,
"grad_norm": 2.9235932803770868,
"learning_rate": 8.706421283690578e-07,
"loss": 1.2987,
"step": 616
},
{
"epoch": 0.03,
"grad_norm": 3.0655234851680824,
"learning_rate": 8.706360710257128e-07,
"loss": 1.3903,
"step": 617
},
{
"epoch": 0.03,
"grad_norm": 2.968020759665533,
"learning_rate": 8.706300019828679e-07,
"loss": 1.4227,
"step": 618
},
{
"epoch": 0.03,
"grad_norm": 3.116571254583397,
"learning_rate": 8.706239212406866e-07,
"loss": 1.4153,
"step": 619
},
{
"epoch": 0.03,
"grad_norm": 3.0578559650775383,
"learning_rate": 8.706178287993326e-07,
"loss": 1.4168,
"step": 620
},
{
"epoch": 0.03,
"grad_norm": 2.9399476742786907,
"learning_rate": 8.706117246589699e-07,
"loss": 1.3448,
"step": 621
},
{
"epoch": 0.03,
"grad_norm": 3.045715774895227,
"learning_rate": 8.706056088197628e-07,
"loss": 1.4323,
"step": 622
},
{
"epoch": 0.03,
"grad_norm": 2.8892619986413655,
"learning_rate": 8.705994812818759e-07,
"loss": 1.3688,
"step": 623
},
{
"epoch": 0.03,
"grad_norm": 3.0537421060657257,
"learning_rate": 8.705933420454745e-07,
"loss": 1.2805,
"step": 624
},
{
"epoch": 0.03,
"grad_norm": 3.02315805498333,
"learning_rate": 8.705871911107236e-07,
"loss": 1.3664,
"step": 625
},
{
"epoch": 0.03,
"grad_norm": 3.0968669538104705,
"learning_rate": 8.70581028477789e-07,
"loss": 1.4156,
"step": 626
},
{
"epoch": 0.03,
"grad_norm": 2.958411205891844,
"learning_rate": 8.705748541468365e-07,
"loss": 1.3879,
"step": 627
},
{
"epoch": 0.03,
"grad_norm": 3.0644896141269213,
"learning_rate": 8.705686681180324e-07,
"loss": 1.406,
"step": 628
},
{
"epoch": 0.03,
"grad_norm": 2.9355330340666947,
"learning_rate": 8.705624703915431e-07,
"loss": 1.4157,
"step": 629
},
{
"epoch": 0.03,
"grad_norm": 3.0295134625264732,
"learning_rate": 8.705562609675357e-07,
"loss": 1.3595,
"step": 630
}
],
"logging_steps": 1.0,
"max_steps": 19246,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 105,
"total_flos": 164886478848000.0,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}