diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,4431 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.032734074612906575, + "eval_steps": 500, + "global_step": 630, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0, + "grad_norm": 2.967684234392304, + "learning_rate": 8.722043470761813e-09, + "loss": 1.4217, + "step": 1 + }, + { + "epoch": 0.0, + "grad_norm": 3.0271119924702425, + "learning_rate": 1.7444086941523626e-08, + "loss": 1.3854, + "step": 2 + }, + { + "epoch": 0.0, + "grad_norm": 3.0212962641045844, + "learning_rate": 2.6166130412285438e-08, + "loss": 1.421, + "step": 3 + }, + { + "epoch": 0.0, + "grad_norm": 2.969098210999342, + "learning_rate": 3.488817388304725e-08, + "loss": 1.3894, + "step": 4 + }, + { + "epoch": 0.0, + "grad_norm": 2.9600462061407877, + "learning_rate": 4.361021735380907e-08, + "loss": 1.3803, + "step": 5 + }, + { + "epoch": 0.0, + "grad_norm": 3.074991458239833, + "learning_rate": 5.2332260824570876e-08, + "loss": 1.3815, + "step": 6 + }, + { + "epoch": 0.0, + "grad_norm": 3.026445479098331, + "learning_rate": 6.105430429533269e-08, + "loss": 1.3684, + "step": 7 + }, + { + "epoch": 0.0, + "grad_norm": 2.9268992001978744, + "learning_rate": 6.97763477660945e-08, + "loss": 1.4307, + "step": 8 + }, + { + "epoch": 0.0, + "grad_norm": 3.020642302186043, + "learning_rate": 7.849839123685631e-08, + "loss": 1.3318, + "step": 9 + }, + { + "epoch": 0.0, + "grad_norm": 3.0765469986494853, + "learning_rate": 8.722043470761814e-08, + "loss": 1.3942, + "step": 10 + }, + { + "epoch": 0.0, + "grad_norm": 3.081729945762728, + "learning_rate": 9.594247817837994e-08, + "loss": 1.3555, + "step": 11 + }, + { + "epoch": 0.0, + "grad_norm": 3.0021774146670808, + "learning_rate": 1.0466452164914175e-07, + "loss": 1.4366, + "step": 12 + }, + { + "epoch": 0.0, + "grad_norm": 3.059550099782956, + "learning_rate": 1.1338656511990357e-07, + "loss": 1.3999, + "step": 13 + }, + { + "epoch": 0.0, + "grad_norm": 2.9198920843423073, + "learning_rate": 1.2210860859066538e-07, + "loss": 1.3949, + "step": 14 + }, + { + "epoch": 0.0, + "grad_norm": 3.013793892067568, + "learning_rate": 1.308306520614272e-07, + "loss": 1.4062, + "step": 15 + }, + { + "epoch": 0.0, + "grad_norm": 2.9189606946894022, + "learning_rate": 1.39552695532189e-07, + "loss": 1.401, + "step": 16 + }, + { + "epoch": 0.0, + "grad_norm": 2.9989048069702475, + "learning_rate": 1.4827473900295083e-07, + "loss": 1.39, + "step": 17 + }, + { + "epoch": 0.0, + "grad_norm": 2.8873329349365613, + "learning_rate": 1.5699678247371262e-07, + "loss": 1.4061, + "step": 18 + }, + { + "epoch": 0.0, + "grad_norm": 2.8867416403445345, + "learning_rate": 1.6571882594447446e-07, + "loss": 1.4032, + "step": 19 + }, + { + "epoch": 0.0, + "grad_norm": 2.9725069440958127, + "learning_rate": 1.7444086941523627e-07, + "loss": 1.3772, + "step": 20 + }, + { + "epoch": 0.0, + "grad_norm": 3.0049844241627675, + "learning_rate": 1.8316291288599806e-07, + "loss": 1.3907, + "step": 21 + }, + { + "epoch": 0.0, + "grad_norm": 2.878905929699835, + "learning_rate": 1.9188495635675987e-07, + "loss": 1.4029, + "step": 22 + }, + { + "epoch": 0.0, + "grad_norm": 2.8262523355216014, + "learning_rate": 2.006069998275217e-07, + "loss": 1.3488, + "step": 23 + }, + { + "epoch": 0.0, + "grad_norm": 2.913459302454502, + "learning_rate": 2.093290432982835e-07, + "loss": 1.3888, + "step": 24 + }, + { + "epoch": 0.0, + "grad_norm": 2.924777837323026, + "learning_rate": 2.1805108676904532e-07, + "loss": 1.3423, + "step": 25 + }, + { + "epoch": 0.0, + "grad_norm": 3.0238436977422207, + "learning_rate": 2.2677313023980713e-07, + "loss": 1.4083, + "step": 26 + }, + { + "epoch": 0.0, + "grad_norm": 2.9404632380821036, + "learning_rate": 2.3549517371056895e-07, + "loss": 1.3325, + "step": 27 + }, + { + "epoch": 0.0, + "grad_norm": 2.9624670572200222, + "learning_rate": 2.4421721718133076e-07, + "loss": 1.4221, + "step": 28 + }, + { + "epoch": 0.0, + "grad_norm": 2.9307501329723236, + "learning_rate": 2.5293926065209255e-07, + "loss": 1.3827, + "step": 29 + }, + { + "epoch": 0.0, + "grad_norm": 2.8568182093611365, + "learning_rate": 2.616613041228544e-07, + "loss": 1.394, + "step": 30 + }, + { + "epoch": 0.0, + "grad_norm": 3.0056868600386544, + "learning_rate": 2.703833475936162e-07, + "loss": 1.3585, + "step": 31 + }, + { + "epoch": 0.0, + "grad_norm": 2.928843591728527, + "learning_rate": 2.79105391064378e-07, + "loss": 1.3927, + "step": 32 + }, + { + "epoch": 0.0, + "grad_norm": 2.979059280785437, + "learning_rate": 2.878274345351398e-07, + "loss": 1.4622, + "step": 33 + }, + { + "epoch": 0.0, + "grad_norm": 2.948156874136048, + "learning_rate": 2.9654947800590165e-07, + "loss": 1.3583, + "step": 34 + }, + { + "epoch": 0.0, + "grad_norm": 2.883247676830032, + "learning_rate": 3.0527152147666344e-07, + "loss": 1.3533, + "step": 35 + }, + { + "epoch": 0.0, + "grad_norm": 3.058107134456641, + "learning_rate": 3.1399356494742523e-07, + "loss": 1.3904, + "step": 36 + }, + { + "epoch": 0.0, + "grad_norm": 3.0535192707251237, + "learning_rate": 3.2271560841818707e-07, + "loss": 1.3978, + "step": 37 + }, + { + "epoch": 0.0, + "grad_norm": 2.897725634603183, + "learning_rate": 3.314376518889489e-07, + "loss": 1.3422, + "step": 38 + }, + { + "epoch": 0.0, + "grad_norm": 2.869420148217502, + "learning_rate": 3.401596953597107e-07, + "loss": 1.4465, + "step": 39 + }, + { + "epoch": 0.0, + "grad_norm": 2.9815636403050676, + "learning_rate": 3.4888173883047254e-07, + "loss": 1.407, + "step": 40 + }, + { + "epoch": 0.0, + "grad_norm": 2.9486391625075123, + "learning_rate": 3.576037823012343e-07, + "loss": 1.391, + "step": 41 + }, + { + "epoch": 0.0, + "grad_norm": 2.9396534742524283, + "learning_rate": 3.663258257719961e-07, + "loss": 1.4191, + "step": 42 + }, + { + "epoch": 0.0, + "grad_norm": 2.954105365539368, + "learning_rate": 3.7504786924275796e-07, + "loss": 1.4039, + "step": 43 + }, + { + "epoch": 0.0, + "grad_norm": 2.9899210526299482, + "learning_rate": 3.8376991271351975e-07, + "loss": 1.3606, + "step": 44 + }, + { + "epoch": 0.0, + "grad_norm": 2.9715202250605515, + "learning_rate": 3.924919561842816e-07, + "loss": 1.3561, + "step": 45 + }, + { + "epoch": 0.0, + "grad_norm": 2.8726950910953133, + "learning_rate": 4.012139996550434e-07, + "loss": 1.4004, + "step": 46 + }, + { + "epoch": 0.0, + "grad_norm": 2.9356806631744568, + "learning_rate": 4.0993604312580517e-07, + "loss": 1.3626, + "step": 47 + }, + { + "epoch": 0.0, + "grad_norm": 2.828385414972167, + "learning_rate": 4.18658086596567e-07, + "loss": 1.416, + "step": 48 + }, + { + "epoch": 0.0, + "grad_norm": 2.860775941749999, + "learning_rate": 4.273801300673288e-07, + "loss": 1.4232, + "step": 49 + }, + { + "epoch": 0.0, + "grad_norm": 2.908555437519401, + "learning_rate": 4.3610217353809064e-07, + "loss": 1.371, + "step": 50 + }, + { + "epoch": 0.0, + "grad_norm": 3.0110809579629283, + "learning_rate": 4.448242170088525e-07, + "loss": 1.3886, + "step": 51 + }, + { + "epoch": 0.0, + "grad_norm": 2.8879956038976062, + "learning_rate": 4.5354626047961427e-07, + "loss": 1.3677, + "step": 52 + }, + { + "epoch": 0.0, + "grad_norm": 2.9475494022582405, + "learning_rate": 4.622683039503761e-07, + "loss": 1.3596, + "step": 53 + }, + { + "epoch": 0.0, + "grad_norm": 2.9208201499436677, + "learning_rate": 4.709903474211379e-07, + "loss": 1.3795, + "step": 54 + }, + { + "epoch": 0.0, + "grad_norm": 2.9250691956165293, + "learning_rate": 4.797123908918997e-07, + "loss": 1.3521, + "step": 55 + }, + { + "epoch": 0.0, + "grad_norm": 3.0059604294844795, + "learning_rate": 4.884344343626615e-07, + "loss": 1.3872, + "step": 56 + }, + { + "epoch": 0.0, + "grad_norm": 2.9408348770709347, + "learning_rate": 4.971564778334233e-07, + "loss": 1.41, + "step": 57 + }, + { + "epoch": 0.0, + "grad_norm": 2.9941841393676247, + "learning_rate": 5.058785213041851e-07, + "loss": 1.4293, + "step": 58 + }, + { + "epoch": 0.0, + "grad_norm": 2.9287843738148744, + "learning_rate": 5.146005647749469e-07, + "loss": 1.411, + "step": 59 + }, + { + "epoch": 0.0, + "grad_norm": 2.9468093301597533, + "learning_rate": 5.233226082457088e-07, + "loss": 1.3508, + "step": 60 + }, + { + "epoch": 0.0, + "grad_norm": 2.854893968299532, + "learning_rate": 5.320446517164706e-07, + "loss": 1.4042, + "step": 61 + }, + { + "epoch": 0.0, + "grad_norm": 2.9224674242946387, + "learning_rate": 5.407666951872324e-07, + "loss": 1.4182, + "step": 62 + }, + { + "epoch": 0.0, + "grad_norm": 3.025591867750147, + "learning_rate": 5.494887386579943e-07, + "loss": 1.4162, + "step": 63 + }, + { + "epoch": 0.0, + "grad_norm": 2.9351339151935045, + "learning_rate": 5.58210782128756e-07, + "loss": 1.2956, + "step": 64 + }, + { + "epoch": 0.0, + "grad_norm": 2.9740598113404677, + "learning_rate": 5.669328255995178e-07, + "loss": 1.4181, + "step": 65 + }, + { + "epoch": 0.0, + "grad_norm": 2.8698658052840464, + "learning_rate": 5.756548690702796e-07, + "loss": 1.3885, + "step": 66 + }, + { + "epoch": 0.0, + "grad_norm": 2.847745090454755, + "learning_rate": 5.843769125410415e-07, + "loss": 1.4268, + "step": 67 + }, + { + "epoch": 0.0, + "grad_norm": 3.53006793293518, + "learning_rate": 5.930989560118033e-07, + "loss": 1.4369, + "step": 68 + }, + { + "epoch": 0.0, + "grad_norm": 3.02012038605618, + "learning_rate": 6.01820999482565e-07, + "loss": 1.406, + "step": 69 + }, + { + "epoch": 0.0, + "grad_norm": 2.8502665195139074, + "learning_rate": 6.105430429533269e-07, + "loss": 1.4273, + "step": 70 + }, + { + "epoch": 0.0, + "grad_norm": 2.966110454182966, + "learning_rate": 6.192650864240887e-07, + "loss": 1.3822, + "step": 71 + }, + { + "epoch": 0.0, + "grad_norm": 2.9117409152319604, + "learning_rate": 6.279871298948505e-07, + "loss": 1.3758, + "step": 72 + }, + { + "epoch": 0.0, + "grad_norm": 2.9583103749551247, + "learning_rate": 6.367091733656124e-07, + "loss": 1.3821, + "step": 73 + }, + { + "epoch": 0.0, + "grad_norm": 2.9976941225678524, + "learning_rate": 6.454312168363741e-07, + "loss": 1.3818, + "step": 74 + }, + { + "epoch": 0.0, + "grad_norm": 2.9208562033277237, + "learning_rate": 6.541532603071359e-07, + "loss": 1.4249, + "step": 75 + }, + { + "epoch": 0.0, + "grad_norm": 2.9753607287344868, + "learning_rate": 6.628753037778978e-07, + "loss": 1.394, + "step": 76 + }, + { + "epoch": 0.0, + "grad_norm": 2.852059681146882, + "learning_rate": 6.715973472486596e-07, + "loss": 1.3735, + "step": 77 + }, + { + "epoch": 0.0, + "grad_norm": 2.9721541467116075, + "learning_rate": 6.803193907194214e-07, + "loss": 1.3284, + "step": 78 + }, + { + "epoch": 0.0, + "grad_norm": 2.959460844473033, + "learning_rate": 6.890414341901832e-07, + "loss": 1.3998, + "step": 79 + }, + { + "epoch": 0.0, + "grad_norm": 2.8741209861248747, + "learning_rate": 6.977634776609451e-07, + "loss": 1.3696, + "step": 80 + }, + { + "epoch": 0.0, + "grad_norm": 2.916879772765294, + "learning_rate": 7.064855211317069e-07, + "loss": 1.4736, + "step": 81 + }, + { + "epoch": 0.0, + "grad_norm": 3.069300514821944, + "learning_rate": 7.152075646024686e-07, + "loss": 1.3341, + "step": 82 + }, + { + "epoch": 0.0, + "grad_norm": 2.966078167553101, + "learning_rate": 7.239296080732305e-07, + "loss": 1.3705, + "step": 83 + }, + { + "epoch": 0.0, + "grad_norm": 2.891366947828864, + "learning_rate": 7.326516515439922e-07, + "loss": 1.4036, + "step": 84 + }, + { + "epoch": 0.0, + "grad_norm": 3.0015624379292687, + "learning_rate": 7.41373695014754e-07, + "loss": 1.3548, + "step": 85 + }, + { + "epoch": 0.0, + "grad_norm": 2.9762714989519297, + "learning_rate": 7.500957384855159e-07, + "loss": 1.4038, + "step": 86 + }, + { + "epoch": 0.0, + "grad_norm": 3.0417489896152325, + "learning_rate": 7.588177819562777e-07, + "loss": 1.3972, + "step": 87 + }, + { + "epoch": 0.0, + "grad_norm": 2.957367413809754, + "learning_rate": 7.675398254270395e-07, + "loss": 1.376, + "step": 88 + }, + { + "epoch": 0.0, + "grad_norm": 2.9811178726545466, + "learning_rate": 7.762618688978014e-07, + "loss": 1.366, + "step": 89 + }, + { + "epoch": 0.0, + "grad_norm": 2.9882512869731994, + "learning_rate": 7.849839123685632e-07, + "loss": 1.3835, + "step": 90 + }, + { + "epoch": 0.0, + "grad_norm": 2.878782548871192, + "learning_rate": 7.93705955839325e-07, + "loss": 1.3372, + "step": 91 + }, + { + "epoch": 0.0, + "grad_norm": 2.9971417051306344, + "learning_rate": 8.024279993100868e-07, + "loss": 1.4037, + "step": 92 + }, + { + "epoch": 0.0, + "grad_norm": 3.0231933357399066, + "learning_rate": 8.111500427808487e-07, + "loss": 1.3692, + "step": 93 + }, + { + "epoch": 0.0, + "grad_norm": 2.8777305685815024, + "learning_rate": 8.198720862516103e-07, + "loss": 1.3493, + "step": 94 + }, + { + "epoch": 0.0, + "grad_norm": 2.990027107011572, + "learning_rate": 8.285941297223721e-07, + "loss": 1.3586, + "step": 95 + }, + { + "epoch": 0.0, + "grad_norm": 2.818072292309908, + "learning_rate": 8.37316173193134e-07, + "loss": 1.3612, + "step": 96 + }, + { + "epoch": 0.01, + "grad_norm": 3.1271514088993144, + "learning_rate": 8.460382166638958e-07, + "loss": 1.336, + "step": 97 + }, + { + "epoch": 0.01, + "grad_norm": 3.0169265133833396, + "learning_rate": 8.547602601346576e-07, + "loss": 1.3665, + "step": 98 + }, + { + "epoch": 0.01, + "grad_norm": 2.9482660829507994, + "learning_rate": 8.634823036054195e-07, + "loss": 1.3551, + "step": 99 + }, + { + "epoch": 0.01, + "grad_norm": 3.087617940804514, + "learning_rate": 8.722043470761813e-07, + "loss": 1.3659, + "step": 100 + }, + { + "epoch": 0.01, + "grad_norm": 2.894747860556493, + "learning_rate": 8.72204341205319e-07, + "loss": 1.4032, + "step": 101 + }, + { + "epoch": 0.01, + "grad_norm": 2.9567796592966546, + "learning_rate": 8.722043235927325e-07, + "loss": 1.3568, + "step": 102 + }, + { + "epoch": 0.01, + "grad_norm": 3.130564519854559, + "learning_rate": 8.72204294238422e-07, + "loss": 1.3567, + "step": 103 + }, + { + "epoch": 0.01, + "grad_norm": 2.98891915499642, + "learning_rate": 8.722042531423884e-07, + "loss": 1.3865, + "step": 104 + }, + { + "epoch": 0.01, + "grad_norm": 3.036577692929289, + "learning_rate": 8.722042003046327e-07, + "loss": 1.3901, + "step": 105 + }, + { + "epoch": 0.01, + "grad_norm": 3.001761214468399, + "learning_rate": 8.722041357251567e-07, + "loss": 1.4117, + "step": 106 + }, + { + "epoch": 0.01, + "grad_norm": 3.006505553779288, + "learning_rate": 8.722040594039618e-07, + "loss": 1.4083, + "step": 107 + }, + { + "epoch": 0.01, + "grad_norm": 2.970081989296889, + "learning_rate": 8.722039713410501e-07, + "loss": 1.3163, + "step": 108 + }, + { + "epoch": 0.01, + "grad_norm": 2.9452680507051863, + "learning_rate": 8.72203871536424e-07, + "loss": 1.3367, + "step": 109 + }, + { + "epoch": 0.01, + "grad_norm": 3.0057027079906216, + "learning_rate": 8.722037599900863e-07, + "loss": 1.3893, + "step": 110 + }, + { + "epoch": 0.01, + "grad_norm": 2.9731127424958697, + "learning_rate": 8.722036367020397e-07, + "loss": 1.3604, + "step": 111 + }, + { + "epoch": 0.01, + "grad_norm": 2.884218184477182, + "learning_rate": 8.722035016722879e-07, + "loss": 1.4651, + "step": 112 + }, + { + "epoch": 0.01, + "grad_norm": 2.8763866239897644, + "learning_rate": 8.722033549008343e-07, + "loss": 1.3947, + "step": 113 + }, + { + "epoch": 0.01, + "grad_norm": 2.927176134008426, + "learning_rate": 8.722031963876829e-07, + "loss": 1.3667, + "step": 114 + }, + { + "epoch": 0.01, + "grad_norm": 2.9924621440798664, + "learning_rate": 8.72203026132838e-07, + "loss": 1.3632, + "step": 115 + }, + { + "epoch": 0.01, + "grad_norm": 3.0499504528400943, + "learning_rate": 8.72202844136304e-07, + "loss": 1.4176, + "step": 116 + }, + { + "epoch": 0.01, + "grad_norm": 2.9684475725772392, + "learning_rate": 8.722026503980863e-07, + "loss": 1.4146, + "step": 117 + }, + { + "epoch": 0.01, + "grad_norm": 2.9148264494394662, + "learning_rate": 8.722024449181895e-07, + "loss": 1.4205, + "step": 118 + }, + { + "epoch": 0.01, + "grad_norm": 2.947431519938494, + "learning_rate": 8.722022276966194e-07, + "loss": 1.3281, + "step": 119 + }, + { + "epoch": 0.01, + "grad_norm": 2.955626479582277, + "learning_rate": 8.72201998733382e-07, + "loss": 1.3465, + "step": 120 + }, + { + "epoch": 0.01, + "grad_norm": 3.028540598737732, + "learning_rate": 8.722017580284832e-07, + "loss": 1.3472, + "step": 121 + }, + { + "epoch": 0.01, + "grad_norm": 2.9197378030040753, + "learning_rate": 8.722015055819296e-07, + "loss": 1.381, + "step": 122 + }, + { + "epoch": 0.01, + "grad_norm": 3.0574320249257227, + "learning_rate": 8.722012413937282e-07, + "loss": 1.4225, + "step": 123 + }, + { + "epoch": 0.01, + "grad_norm": 2.913538682906452, + "learning_rate": 8.722009654638856e-07, + "loss": 1.3536, + "step": 124 + }, + { + "epoch": 0.01, + "grad_norm": 2.886066575609779, + "learning_rate": 8.722006777924096e-07, + "loss": 1.3736, + "step": 125 + }, + { + "epoch": 0.01, + "grad_norm": 2.957758024407401, + "learning_rate": 8.722003783793081e-07, + "loss": 1.3973, + "step": 126 + }, + { + "epoch": 0.01, + "grad_norm": 2.958995782934072, + "learning_rate": 8.722000672245888e-07, + "loss": 1.3954, + "step": 127 + }, + { + "epoch": 0.01, + "grad_norm": 2.987991679162308, + "learning_rate": 8.721997443282602e-07, + "loss": 1.3757, + "step": 128 + }, + { + "epoch": 0.01, + "grad_norm": 3.0909686056252434, + "learning_rate": 8.721994096903311e-07, + "loss": 1.3462, + "step": 129 + }, + { + "epoch": 0.01, + "grad_norm": 2.9477825764652494, + "learning_rate": 8.721990633108104e-07, + "loss": 1.4295, + "step": 130 + }, + { + "epoch": 0.01, + "grad_norm": 2.947921751933276, + "learning_rate": 8.721987051897074e-07, + "loss": 1.3854, + "step": 131 + }, + { + "epoch": 0.01, + "grad_norm": 2.8440480288328427, + "learning_rate": 8.721983353270319e-07, + "loss": 1.4106, + "step": 132 + }, + { + "epoch": 0.01, + "grad_norm": 2.9792049006251906, + "learning_rate": 8.721979537227935e-07, + "loss": 1.3913, + "step": 133 + }, + { + "epoch": 0.01, + "grad_norm": 2.9238731192373746, + "learning_rate": 8.721975603770031e-07, + "loss": 1.3695, + "step": 134 + }, + { + "epoch": 0.01, + "grad_norm": 3.0773774260583155, + "learning_rate": 8.721971552896706e-07, + "loss": 1.3629, + "step": 135 + }, + { + "epoch": 0.01, + "grad_norm": 2.9679219250269044, + "learning_rate": 8.721967384608074e-07, + "loss": 1.4205, + "step": 136 + }, + { + "epoch": 0.01, + "grad_norm": 3.030565370577699, + "learning_rate": 8.721963098904246e-07, + "loss": 1.4311, + "step": 137 + }, + { + "epoch": 0.01, + "grad_norm": 3.0375097512582503, + "learning_rate": 8.721958695785336e-07, + "loss": 1.4069, + "step": 138 + }, + { + "epoch": 0.01, + "grad_norm": 2.92214551378445, + "learning_rate": 8.721954175251462e-07, + "loss": 1.422, + "step": 139 + }, + { + "epoch": 0.01, + "grad_norm": 3.1283303266405578, + "learning_rate": 8.721949537302749e-07, + "loss": 1.432, + "step": 140 + }, + { + "epoch": 0.01, + "grad_norm": 3.1049340924381705, + "learning_rate": 8.72194478193932e-07, + "loss": 1.3815, + "step": 141 + }, + { + "epoch": 0.01, + "grad_norm": 2.869253522521714, + "learning_rate": 8.721939909161303e-07, + "loss": 1.391, + "step": 142 + }, + { + "epoch": 0.01, + "grad_norm": 2.8752461370622306, + "learning_rate": 8.721934918968828e-07, + "loss": 1.3769, + "step": 143 + }, + { + "epoch": 0.01, + "grad_norm": 2.909751532098998, + "learning_rate": 8.721929811362032e-07, + "loss": 1.3995, + "step": 144 + }, + { + "epoch": 0.01, + "grad_norm": 3.0462324191538572, + "learning_rate": 8.72192458634105e-07, + "loss": 1.3689, + "step": 145 + }, + { + "epoch": 0.01, + "grad_norm": 2.9750194133891363, + "learning_rate": 8.721919243906024e-07, + "loss": 1.3707, + "step": 146 + }, + { + "epoch": 0.01, + "grad_norm": 2.932835851287147, + "learning_rate": 8.721913784057099e-07, + "loss": 1.3676, + "step": 147 + }, + { + "epoch": 0.01, + "grad_norm": 2.9429516387295926, + "learning_rate": 8.721908206794419e-07, + "loss": 1.3731, + "step": 148 + }, + { + "epoch": 0.01, + "grad_norm": 2.994182650351975, + "learning_rate": 8.721902512118136e-07, + "loss": 1.3542, + "step": 149 + }, + { + "epoch": 0.01, + "grad_norm": 2.876833046636617, + "learning_rate": 8.721896700028404e-07, + "loss": 1.4124, + "step": 150 + }, + { + "epoch": 0.01, + "grad_norm": 2.945327518176284, + "learning_rate": 8.721890770525377e-07, + "loss": 1.4137, + "step": 151 + }, + { + "epoch": 0.01, + "grad_norm": 21.047240653653276, + "learning_rate": 8.721884723609218e-07, + "loss": 1.4264, + "step": 152 + }, + { + "epoch": 0.01, + "grad_norm": 3.029952956834553, + "learning_rate": 8.721878559280086e-07, + "loss": 1.4372, + "step": 153 + }, + { + "epoch": 0.01, + "grad_norm": 3.137230528895406, + "learning_rate": 8.721872277538151e-07, + "loss": 1.4019, + "step": 154 + }, + { + "epoch": 0.01, + "grad_norm": 2.9696146910825694, + "learning_rate": 8.72186587838358e-07, + "loss": 1.4515, + "step": 155 + }, + { + "epoch": 0.01, + "grad_norm": 2.980760336638325, + "learning_rate": 8.721859361816546e-07, + "loss": 1.4203, + "step": 156 + }, + { + "epoch": 0.01, + "grad_norm": 3.086006727040003, + "learning_rate": 8.721852727837222e-07, + "loss": 1.3712, + "step": 157 + }, + { + "epoch": 0.01, + "grad_norm": 3.003419403761712, + "learning_rate": 8.72184597644579e-07, + "loss": 1.4107, + "step": 158 + }, + { + "epoch": 0.01, + "grad_norm": 3.0463864266769773, + "learning_rate": 8.72183910764243e-07, + "loss": 1.4082, + "step": 159 + }, + { + "epoch": 0.01, + "grad_norm": 2.950362004991645, + "learning_rate": 8.721832121427326e-07, + "loss": 1.352, + "step": 160 + }, + { + "epoch": 0.01, + "grad_norm": 2.8779668562920815, + "learning_rate": 8.721825017800669e-07, + "loss": 1.4236, + "step": 161 + }, + { + "epoch": 0.01, + "grad_norm": 3.0172810234945455, + "learning_rate": 8.721817796762648e-07, + "loss": 1.3871, + "step": 162 + }, + { + "epoch": 0.01, + "grad_norm": 2.9726094865888224, + "learning_rate": 8.721810458313457e-07, + "loss": 1.349, + "step": 163 + }, + { + "epoch": 0.01, + "grad_norm": 3.0322153773349334, + "learning_rate": 8.721803002453297e-07, + "loss": 1.3935, + "step": 164 + }, + { + "epoch": 0.01, + "grad_norm": 3.0249194383352283, + "learning_rate": 8.721795429182364e-07, + "loss": 1.3849, + "step": 165 + }, + { + "epoch": 0.01, + "grad_norm": 2.9938901642866718, + "learning_rate": 8.721787738500866e-07, + "loss": 1.4267, + "step": 166 + }, + { + "epoch": 0.01, + "grad_norm": 2.923274256584191, + "learning_rate": 8.721779930409007e-07, + "loss": 1.4283, + "step": 167 + }, + { + "epoch": 0.01, + "grad_norm": 2.8466301519873785, + "learning_rate": 8.721772004906999e-07, + "loss": 1.3842, + "step": 168 + }, + { + "epoch": 0.01, + "grad_norm": 3.0052783960700165, + "learning_rate": 8.721763961995056e-07, + "loss": 1.4335, + "step": 169 + }, + { + "epoch": 0.01, + "grad_norm": 3.0935980229307614, + "learning_rate": 8.721755801673391e-07, + "loss": 1.3751, + "step": 170 + }, + { + "epoch": 0.01, + "grad_norm": 3.0201926532008505, + "learning_rate": 8.721747523942229e-07, + "loss": 1.383, + "step": 171 + }, + { + "epoch": 0.01, + "grad_norm": 2.964731945306275, + "learning_rate": 8.721739128801788e-07, + "loss": 1.3359, + "step": 172 + }, + { + "epoch": 0.01, + "grad_norm": 2.928301273992501, + "learning_rate": 8.721730616252297e-07, + "loss": 1.3461, + "step": 173 + }, + { + "epoch": 0.01, + "grad_norm": 2.9759904501938617, + "learning_rate": 8.721721986293985e-07, + "loss": 1.3644, + "step": 174 + }, + { + "epoch": 0.01, + "grad_norm": 3.040389211247673, + "learning_rate": 8.721713238927082e-07, + "loss": 1.4341, + "step": 175 + }, + { + "epoch": 0.01, + "grad_norm": 3.005645814777094, + "learning_rate": 8.721704374151826e-07, + "loss": 1.3967, + "step": 176 + }, + { + "epoch": 0.01, + "grad_norm": 3.1272949994089823, + "learning_rate": 8.721695391968456e-07, + "loss": 1.3796, + "step": 177 + }, + { + "epoch": 0.01, + "grad_norm": 3.0684527006439533, + "learning_rate": 8.721686292377211e-07, + "loss": 1.3905, + "step": 178 + }, + { + "epoch": 0.01, + "grad_norm": 2.9348716760076603, + "learning_rate": 8.721677075378338e-07, + "loss": 1.3905, + "step": 179 + }, + { + "epoch": 0.01, + "grad_norm": 2.9873597802095304, + "learning_rate": 8.721667740972085e-07, + "loss": 1.4103, + "step": 180 + }, + { + "epoch": 0.01, + "grad_norm": 2.979793648840201, + "learning_rate": 8.721658289158703e-07, + "loss": 1.3622, + "step": 181 + }, + { + "epoch": 0.01, + "grad_norm": 2.937468774579186, + "learning_rate": 8.721648719938447e-07, + "loss": 1.414, + "step": 182 + }, + { + "epoch": 0.01, + "grad_norm": 2.879104091071243, + "learning_rate": 8.721639033311573e-07, + "loss": 1.3108, + "step": 183 + }, + { + "epoch": 0.01, + "grad_norm": 3.0663878291218203, + "learning_rate": 8.721629229278344e-07, + "loss": 1.3543, + "step": 184 + }, + { + "epoch": 0.01, + "grad_norm": 2.9407287447315826, + "learning_rate": 8.721619307839025e-07, + "loss": 1.3753, + "step": 185 + }, + { + "epoch": 0.01, + "grad_norm": 2.967538550932994, + "learning_rate": 8.721609268993879e-07, + "loss": 1.3973, + "step": 186 + }, + { + "epoch": 0.01, + "grad_norm": 3.057519293009879, + "learning_rate": 8.721599112743179e-07, + "loss": 1.4036, + "step": 187 + }, + { + "epoch": 0.01, + "grad_norm": 2.936392616519391, + "learning_rate": 8.721588839087197e-07, + "loss": 1.4852, + "step": 188 + }, + { + "epoch": 0.01, + "grad_norm": 2.9670464594249197, + "learning_rate": 8.721578448026212e-07, + "loss": 1.3643, + "step": 189 + }, + { + "epoch": 0.01, + "grad_norm": 3.0273720809079663, + "learning_rate": 8.721567939560502e-07, + "loss": 1.4109, + "step": 190 + }, + { + "epoch": 0.01, + "grad_norm": 3.0651462806238854, + "learning_rate": 8.721557313690349e-07, + "loss": 1.3599, + "step": 191 + }, + { + "epoch": 0.01, + "grad_norm": 2.9487059919929326, + "learning_rate": 8.721546570416042e-07, + "loss": 1.3377, + "step": 192 + }, + { + "epoch": 0.01, + "grad_norm": 3.0998938976371146, + "learning_rate": 8.721535709737867e-07, + "loss": 1.3685, + "step": 193 + }, + { + "epoch": 0.01, + "grad_norm": 2.940826121224176, + "learning_rate": 8.721524731656118e-07, + "loss": 1.4174, + "step": 194 + }, + { + "epoch": 0.01, + "grad_norm": 3.014412763659776, + "learning_rate": 8.721513636171093e-07, + "loss": 1.3758, + "step": 195 + }, + { + "epoch": 0.01, + "grad_norm": 3.0608004542672678, + "learning_rate": 8.721502423283086e-07, + "loss": 1.3716, + "step": 196 + }, + { + "epoch": 0.01, + "grad_norm": 2.94244083669587, + "learning_rate": 8.721491092992403e-07, + "loss": 1.3937, + "step": 197 + }, + { + "epoch": 0.01, + "grad_norm": 2.9911802591493144, + "learning_rate": 8.721479645299345e-07, + "loss": 1.4164, + "step": 198 + }, + { + "epoch": 0.01, + "grad_norm": 3.0344694715702065, + "learning_rate": 8.721468080204223e-07, + "loss": 1.4167, + "step": 199 + }, + { + "epoch": 0.01, + "grad_norm": 2.865667879330454, + "learning_rate": 8.72145639770735e-07, + "loss": 1.4041, + "step": 200 + }, + { + "epoch": 0.01, + "grad_norm": 3.072739743995507, + "learning_rate": 8.721444597809037e-07, + "loss": 1.4133, + "step": 201 + }, + { + "epoch": 0.01, + "grad_norm": 3.0276333439793843, + "learning_rate": 8.721432680509603e-07, + "loss": 1.3605, + "step": 202 + }, + { + "epoch": 0.01, + "grad_norm": 3.0465572110487686, + "learning_rate": 8.721420645809369e-07, + "loss": 1.3134, + "step": 203 + }, + { + "epoch": 0.01, + "grad_norm": 2.9542599421921376, + "learning_rate": 8.721408493708659e-07, + "loss": 1.4148, + "step": 204 + }, + { + "epoch": 0.01, + "grad_norm": 2.90720227559915, + "learning_rate": 8.721396224207801e-07, + "loss": 1.3997, + "step": 205 + }, + { + "epoch": 0.01, + "grad_norm": 3.0246259249156617, + "learning_rate": 8.721383837307123e-07, + "loss": 1.4238, + "step": 206 + }, + { + "epoch": 0.01, + "grad_norm": 2.9975853807488453, + "learning_rate": 8.721371333006962e-07, + "loss": 1.3879, + "step": 207 + }, + { + "epoch": 0.01, + "grad_norm": 3.0021383888830258, + "learning_rate": 8.721358711307651e-07, + "loss": 1.3349, + "step": 208 + }, + { + "epoch": 0.01, + "grad_norm": 3.0201205392426296, + "learning_rate": 8.721345972209533e-07, + "loss": 1.3692, + "step": 209 + }, + { + "epoch": 0.01, + "grad_norm": 2.972139743981842, + "learning_rate": 8.721333115712948e-07, + "loss": 1.3856, + "step": 210 + }, + { + "epoch": 0.01, + "grad_norm": 2.9209251468195276, + "learning_rate": 8.721320141818245e-07, + "loss": 1.3726, + "step": 211 + }, + { + "epoch": 0.01, + "grad_norm": 2.9320196507721277, + "learning_rate": 8.721307050525772e-07, + "loss": 1.4143, + "step": 212 + }, + { + "epoch": 0.01, + "grad_norm": 3.0002549106343337, + "learning_rate": 8.72129384183588e-07, + "loss": 1.3897, + "step": 213 + }, + { + "epoch": 0.01, + "grad_norm": 2.9582570275362206, + "learning_rate": 8.721280515748928e-07, + "loss": 1.3756, + "step": 214 + }, + { + "epoch": 0.01, + "grad_norm": 3.0145408653891526, + "learning_rate": 8.721267072265271e-07, + "loss": 1.3929, + "step": 215 + }, + { + "epoch": 0.01, + "grad_norm": 3.0495215598431553, + "learning_rate": 8.721253511385274e-07, + "loss": 1.4061, + "step": 216 + }, + { + "epoch": 0.01, + "grad_norm": 2.9254365712957613, + "learning_rate": 8.721239833109302e-07, + "loss": 1.3903, + "step": 217 + }, + { + "epoch": 0.01, + "grad_norm": 2.997455071778174, + "learning_rate": 8.72122603743772e-07, + "loss": 1.4246, + "step": 218 + }, + { + "epoch": 0.01, + "grad_norm": 2.9195114563849627, + "learning_rate": 8.721212124370902e-07, + "loss": 1.3968, + "step": 219 + }, + { + "epoch": 0.01, + "grad_norm": 3.0160422542520706, + "learning_rate": 8.721198093909225e-07, + "loss": 1.4347, + "step": 220 + }, + { + "epoch": 0.01, + "grad_norm": 3.01404121750226, + "learning_rate": 8.721183946053062e-07, + "loss": 1.3945, + "step": 221 + }, + { + "epoch": 0.01, + "grad_norm": 3.0611502119276692, + "learning_rate": 8.721169680802796e-07, + "loss": 1.3975, + "step": 222 + }, + { + "epoch": 0.01, + "grad_norm": 3.0684020412598727, + "learning_rate": 8.721155298158811e-07, + "loss": 1.373, + "step": 223 + }, + { + "epoch": 0.01, + "grad_norm": 2.9440206694677027, + "learning_rate": 8.721140798121494e-07, + "loss": 1.3432, + "step": 224 + }, + { + "epoch": 0.01, + "grad_norm": 2.974581850771121, + "learning_rate": 8.721126180691237e-07, + "loss": 1.3095, + "step": 225 + }, + { + "epoch": 0.01, + "grad_norm": 2.9933489105960844, + "learning_rate": 8.721111445868431e-07, + "loss": 1.3885, + "step": 226 + }, + { + "epoch": 0.01, + "grad_norm": 2.9831661987665528, + "learning_rate": 8.721096593653475e-07, + "loss": 1.3126, + "step": 227 + }, + { + "epoch": 0.01, + "grad_norm": 3.100080295310524, + "learning_rate": 8.721081624046766e-07, + "loss": 1.3567, + "step": 228 + }, + { + "epoch": 0.01, + "grad_norm": 2.989885190608965, + "learning_rate": 8.72106653704871e-07, + "loss": 1.3899, + "step": 229 + }, + { + "epoch": 0.01, + "grad_norm": 3.0712176271885023, + "learning_rate": 8.721051332659713e-07, + "loss": 1.4208, + "step": 230 + }, + { + "epoch": 0.01, + "grad_norm": 3.062312373029536, + "learning_rate": 8.721036010880183e-07, + "loss": 1.4147, + "step": 231 + }, + { + "epoch": 0.01, + "grad_norm": 2.9701616634317083, + "learning_rate": 8.721020571710533e-07, + "loss": 1.434, + "step": 232 + }, + { + "epoch": 0.01, + "grad_norm": 2.9572643731393646, + "learning_rate": 8.721005015151179e-07, + "loss": 1.3795, + "step": 233 + }, + { + "epoch": 0.01, + "grad_norm": 2.997406682050713, + "learning_rate": 8.720989341202539e-07, + "loss": 1.4501, + "step": 234 + }, + { + "epoch": 0.01, + "grad_norm": 2.9106126480356522, + "learning_rate": 8.720973549865035e-07, + "loss": 1.3684, + "step": 235 + }, + { + "epoch": 0.01, + "grad_norm": 2.9444070042644817, + "learning_rate": 8.720957641139094e-07, + "loss": 1.4213, + "step": 236 + }, + { + "epoch": 0.01, + "grad_norm": 2.8728037311842822, + "learning_rate": 8.720941615025142e-07, + "loss": 1.3519, + "step": 237 + }, + { + "epoch": 0.01, + "grad_norm": 3.015889105815668, + "learning_rate": 8.720925471523613e-07, + "loss": 1.4162, + "step": 238 + }, + { + "epoch": 0.01, + "grad_norm": 2.9419377055914744, + "learning_rate": 8.72090921063494e-07, + "loss": 1.3357, + "step": 239 + }, + { + "epoch": 0.01, + "grad_norm": 2.89008663153287, + "learning_rate": 8.720892832359559e-07, + "loss": 1.3647, + "step": 240 + }, + { + "epoch": 0.01, + "grad_norm": 2.9219232048658736, + "learning_rate": 8.720876336697914e-07, + "loss": 1.4069, + "step": 241 + }, + { + "epoch": 0.01, + "grad_norm": 2.96537590149616, + "learning_rate": 8.72085972365045e-07, + "loss": 1.4118, + "step": 242 + }, + { + "epoch": 0.01, + "grad_norm": 2.883655732971505, + "learning_rate": 8.720842993217609e-07, + "loss": 1.4136, + "step": 243 + }, + { + "epoch": 0.01, + "grad_norm": 2.9788747864733764, + "learning_rate": 8.720826145399848e-07, + "loss": 1.3976, + "step": 244 + }, + { + "epoch": 0.01, + "grad_norm": 2.875570982035785, + "learning_rate": 8.720809180197616e-07, + "loss": 1.426, + "step": 245 + }, + { + "epoch": 0.01, + "grad_norm": 2.9926412719867304, + "learning_rate": 8.720792097611372e-07, + "loss": 1.3629, + "step": 246 + }, + { + "epoch": 0.01, + "grad_norm": 2.958723584893194, + "learning_rate": 8.720774897641574e-07, + "loss": 1.3918, + "step": 247 + }, + { + "epoch": 0.01, + "grad_norm": 2.968992238648431, + "learning_rate": 8.720757580288688e-07, + "loss": 1.4241, + "step": 248 + }, + { + "epoch": 0.01, + "grad_norm": 2.889688463405204, + "learning_rate": 8.720740145553177e-07, + "loss": 1.4101, + "step": 249 + }, + { + "epoch": 0.01, + "grad_norm": 2.9993006762652312, + "learning_rate": 8.720722593435512e-07, + "loss": 1.3857, + "step": 250 + }, + { + "epoch": 0.01, + "grad_norm": 2.980847240255761, + "learning_rate": 8.720704923936167e-07, + "loss": 1.4077, + "step": 251 + }, + { + "epoch": 0.01, + "grad_norm": 2.884007977441845, + "learning_rate": 8.720687137055615e-07, + "loss": 1.3822, + "step": 252 + }, + { + "epoch": 0.01, + "grad_norm": 2.9646728227580645, + "learning_rate": 8.720669232794336e-07, + "loss": 1.3737, + "step": 253 + }, + { + "epoch": 0.01, + "grad_norm": 2.893382783809774, + "learning_rate": 8.720651211152813e-07, + "loss": 1.3762, + "step": 254 + }, + { + "epoch": 0.01, + "grad_norm": 3.024418707419624, + "learning_rate": 8.72063307213153e-07, + "loss": 1.3546, + "step": 255 + }, + { + "epoch": 0.01, + "grad_norm": 3.0252657870696495, + "learning_rate": 8.720614815730977e-07, + "loss": 1.3661, + "step": 256 + }, + { + "epoch": 0.01, + "grad_norm": 3.0208601885030606, + "learning_rate": 8.720596441951642e-07, + "loss": 1.4182, + "step": 257 + }, + { + "epoch": 0.01, + "grad_norm": 3.0552975630549954, + "learning_rate": 8.720577950794024e-07, + "loss": 1.38, + "step": 258 + }, + { + "epoch": 0.01, + "grad_norm": 2.916749346833794, + "learning_rate": 8.720559342258619e-07, + "loss": 1.4049, + "step": 259 + }, + { + "epoch": 0.01, + "grad_norm": 3.035247531851327, + "learning_rate": 8.720540616345928e-07, + "loss": 1.4256, + "step": 260 + }, + { + "epoch": 0.01, + "grad_norm": 2.9902303644665205, + "learning_rate": 8.720521773056454e-07, + "loss": 1.3356, + "step": 261 + }, + { + "epoch": 0.01, + "grad_norm": 3.119044393884763, + "learning_rate": 8.720502812390706e-07, + "loss": 1.4103, + "step": 262 + }, + { + "epoch": 0.01, + "grad_norm": 2.9552269954583803, + "learning_rate": 8.720483734349194e-07, + "loss": 1.3855, + "step": 263 + }, + { + "epoch": 0.01, + "grad_norm": 3.017213443982555, + "learning_rate": 8.720464538932433e-07, + "loss": 1.3902, + "step": 264 + }, + { + "epoch": 0.01, + "grad_norm": 2.925567645830183, + "learning_rate": 8.720445226140937e-07, + "loss": 1.4519, + "step": 265 + }, + { + "epoch": 0.01, + "grad_norm": 2.983266195022755, + "learning_rate": 8.720425795975228e-07, + "loss": 1.3971, + "step": 266 + }, + { + "epoch": 0.01, + "grad_norm": 2.992218221530755, + "learning_rate": 8.720406248435828e-07, + "loss": 1.4231, + "step": 267 + }, + { + "epoch": 0.01, + "grad_norm": 2.9250759809857882, + "learning_rate": 8.720386583523264e-07, + "loss": 1.3877, + "step": 268 + }, + { + "epoch": 0.01, + "grad_norm": 3.022334104434051, + "learning_rate": 8.720366801238065e-07, + "loss": 1.4133, + "step": 269 + }, + { + "epoch": 0.01, + "grad_norm": 2.90248743689466, + "learning_rate": 8.720346901580765e-07, + "loss": 1.3889, + "step": 270 + }, + { + "epoch": 0.01, + "grad_norm": 3.006981215139682, + "learning_rate": 8.720326884551899e-07, + "loss": 1.3657, + "step": 271 + }, + { + "epoch": 0.01, + "grad_norm": 2.996511837199606, + "learning_rate": 8.720306750152005e-07, + "loss": 1.3918, + "step": 272 + }, + { + "epoch": 0.01, + "grad_norm": 2.9853146230235317, + "learning_rate": 8.720286498381625e-07, + "loss": 1.3983, + "step": 273 + }, + { + "epoch": 0.01, + "grad_norm": 2.936322205558776, + "learning_rate": 8.720266129241307e-07, + "loss": 1.3549, + "step": 274 + }, + { + "epoch": 0.01, + "grad_norm": 3.0336499138604984, + "learning_rate": 8.720245642731596e-07, + "loss": 1.3614, + "step": 275 + }, + { + "epoch": 0.01, + "grad_norm": 2.9761367095195514, + "learning_rate": 8.720225038853046e-07, + "loss": 1.4223, + "step": 276 + }, + { + "epoch": 0.01, + "grad_norm": 2.8383029378391256, + "learning_rate": 8.72020431760621e-07, + "loss": 1.3027, + "step": 277 + }, + { + "epoch": 0.01, + "grad_norm": 3.0012216332284964, + "learning_rate": 8.720183478991647e-07, + "loss": 1.307, + "step": 278 + }, + { + "epoch": 0.01, + "grad_norm": 3.202220665410359, + "learning_rate": 8.720162523009919e-07, + "loss": 1.3495, + "step": 279 + }, + { + "epoch": 0.01, + "grad_norm": 2.926537743004413, + "learning_rate": 8.720141449661587e-07, + "loss": 1.346, + "step": 280 + }, + { + "epoch": 0.01, + "grad_norm": 3.025411005245412, + "learning_rate": 8.720120258947223e-07, + "loss": 1.3581, + "step": 281 + }, + { + "epoch": 0.01, + "grad_norm": 3.010956090798263, + "learning_rate": 8.720098950867392e-07, + "loss": 1.3634, + "step": 282 + }, + { + "epoch": 0.01, + "grad_norm": 3.0383205174055727, + "learning_rate": 8.720077525422671e-07, + "loss": 1.3642, + "step": 283 + }, + { + "epoch": 0.01, + "grad_norm": 2.967895550740301, + "learning_rate": 8.720055982613638e-07, + "loss": 1.3841, + "step": 284 + }, + { + "epoch": 0.01, + "grad_norm": 2.9365373503076246, + "learning_rate": 8.720034322440872e-07, + "loss": 1.3527, + "step": 285 + }, + { + "epoch": 0.01, + "grad_norm": 3.1104370711279214, + "learning_rate": 8.720012544904955e-07, + "loss": 1.3483, + "step": 286 + }, + { + "epoch": 0.01, + "grad_norm": 2.952339289555111, + "learning_rate": 8.719990650006473e-07, + "loss": 1.3956, + "step": 287 + }, + { + "epoch": 0.01, + "grad_norm": 2.942959986729864, + "learning_rate": 8.719968637746018e-07, + "loss": 1.4256, + "step": 288 + }, + { + "epoch": 0.02, + "grad_norm": 3.1019407497257507, + "learning_rate": 8.71994650812418e-07, + "loss": 1.3786, + "step": 289 + }, + { + "epoch": 0.02, + "grad_norm": 3.0501476908245984, + "learning_rate": 8.719924261141557e-07, + "loss": 1.4158, + "step": 290 + }, + { + "epoch": 0.02, + "grad_norm": 2.9781394711393507, + "learning_rate": 8.719901896798748e-07, + "loss": 1.427, + "step": 291 + }, + { + "epoch": 0.02, + "grad_norm": 3.0411329565229646, + "learning_rate": 8.719879415096352e-07, + "loss": 1.4281, + "step": 292 + }, + { + "epoch": 0.02, + "grad_norm": 3.022154009359811, + "learning_rate": 8.719856816034978e-07, + "loss": 1.435, + "step": 293 + }, + { + "epoch": 0.02, + "grad_norm": 2.965921869395771, + "learning_rate": 8.719834099615232e-07, + "loss": 1.3766, + "step": 294 + }, + { + "epoch": 0.02, + "grad_norm": 3.026938014636579, + "learning_rate": 8.719811265837728e-07, + "loss": 1.3612, + "step": 295 + }, + { + "epoch": 0.02, + "grad_norm": 2.9969914810093115, + "learning_rate": 8.719788314703078e-07, + "loss": 1.3371, + "step": 296 + }, + { + "epoch": 0.02, + "grad_norm": 2.8906070169866545, + "learning_rate": 8.719765246211902e-07, + "loss": 1.3826, + "step": 297 + }, + { + "epoch": 0.02, + "grad_norm": 3.0301378229830593, + "learning_rate": 8.71974206036482e-07, + "loss": 1.3937, + "step": 298 + }, + { + "epoch": 0.02, + "grad_norm": 2.9673359121672145, + "learning_rate": 8.719718757162457e-07, + "loss": 1.3838, + "step": 299 + }, + { + "epoch": 0.02, + "grad_norm": 2.972360304451488, + "learning_rate": 8.719695336605439e-07, + "loss": 1.4382, + "step": 300 + }, + { + "epoch": 0.02, + "grad_norm": 2.9493767886841242, + "learning_rate": 8.7196717986944e-07, + "loss": 1.3651, + "step": 301 + }, + { + "epoch": 0.02, + "grad_norm": 3.031661286320472, + "learning_rate": 8.719648143429969e-07, + "loss": 1.3482, + "step": 302 + }, + { + "epoch": 0.02, + "grad_norm": 2.920995409830151, + "learning_rate": 8.719624370812787e-07, + "loss": 1.4115, + "step": 303 + }, + { + "epoch": 0.02, + "grad_norm": 2.920727312220773, + "learning_rate": 8.719600480843491e-07, + "loss": 1.396, + "step": 304 + }, + { + "epoch": 0.02, + "grad_norm": 2.945684201201641, + "learning_rate": 8.719576473522726e-07, + "loss": 1.3557, + "step": 305 + }, + { + "epoch": 0.02, + "grad_norm": 3.0651768014119956, + "learning_rate": 8.719552348851139e-07, + "loss": 1.389, + "step": 306 + }, + { + "epoch": 0.02, + "grad_norm": 3.1250149616393577, + "learning_rate": 8.719528106829378e-07, + "loss": 1.469, + "step": 307 + }, + { + "epoch": 0.02, + "grad_norm": 2.977539941978143, + "learning_rate": 8.719503747458096e-07, + "loss": 1.3536, + "step": 308 + }, + { + "epoch": 0.02, + "grad_norm": 3.0745479693463924, + "learning_rate": 8.71947927073795e-07, + "loss": 1.3877, + "step": 309 + }, + { + "epoch": 0.02, + "grad_norm": 3.2160266553797667, + "learning_rate": 8.719454676669596e-07, + "loss": 1.3988, + "step": 310 + }, + { + "epoch": 0.02, + "grad_norm": 3.134783336833123, + "learning_rate": 8.719429965253698e-07, + "loss": 1.4104, + "step": 311 + }, + { + "epoch": 0.02, + "grad_norm": 2.908012731710042, + "learning_rate": 8.719405136490924e-07, + "loss": 1.4186, + "step": 312 + }, + { + "epoch": 0.02, + "grad_norm": 2.9868078254055934, + "learning_rate": 8.71938019038194e-07, + "loss": 1.2836, + "step": 313 + }, + { + "epoch": 0.02, + "grad_norm": 9.747982306246115, + "learning_rate": 8.719355126927416e-07, + "loss": 1.3331, + "step": 314 + }, + { + "epoch": 0.02, + "grad_norm": 2.9503527741257476, + "learning_rate": 8.719329946128029e-07, + "loss": 1.3993, + "step": 315 + }, + { + "epoch": 0.02, + "grad_norm": 3.000724848973538, + "learning_rate": 8.719304647984458e-07, + "loss": 1.3621, + "step": 316 + }, + { + "epoch": 0.02, + "grad_norm": 3.0259025600243032, + "learning_rate": 8.719279232497381e-07, + "loss": 1.4128, + "step": 317 + }, + { + "epoch": 0.02, + "grad_norm": 2.95272448132235, + "learning_rate": 8.719253699667485e-07, + "loss": 1.4239, + "step": 318 + }, + { + "epoch": 0.02, + "grad_norm": 2.7978451838553986, + "learning_rate": 8.719228049495456e-07, + "loss": 1.3694, + "step": 319 + }, + { + "epoch": 0.02, + "grad_norm": 2.976459149812301, + "learning_rate": 8.719202281981985e-07, + "loss": 1.3413, + "step": 320 + }, + { + "epoch": 0.02, + "grad_norm": 2.9711797841726377, + "learning_rate": 8.719176397127765e-07, + "loss": 1.3616, + "step": 321 + }, + { + "epoch": 0.02, + "grad_norm": 3.0204333297285815, + "learning_rate": 8.719150394933495e-07, + "loss": 1.4126, + "step": 322 + }, + { + "epoch": 0.02, + "grad_norm": 3.0181766925849556, + "learning_rate": 8.719124275399874e-07, + "loss": 1.4323, + "step": 323 + }, + { + "epoch": 0.02, + "grad_norm": 2.9167139561107893, + "learning_rate": 8.719098038527604e-07, + "loss": 1.4484, + "step": 324 + }, + { + "epoch": 0.02, + "grad_norm": 2.9378575085429013, + "learning_rate": 8.719071684317393e-07, + "loss": 1.3775, + "step": 325 + }, + { + "epoch": 0.02, + "grad_norm": 2.94943978740792, + "learning_rate": 8.719045212769951e-07, + "loss": 1.3897, + "step": 326 + }, + { + "epoch": 0.02, + "grad_norm": 2.958505276332381, + "learning_rate": 8.719018623885988e-07, + "loss": 1.394, + "step": 327 + }, + { + "epoch": 0.02, + "grad_norm": 3.145867503995205, + "learning_rate": 8.718991917666222e-07, + "loss": 1.379, + "step": 328 + }, + { + "epoch": 0.02, + "grad_norm": 2.9734480635815728, + "learning_rate": 8.718965094111372e-07, + "loss": 1.3953, + "step": 329 + }, + { + "epoch": 0.02, + "grad_norm": 2.9049260797132335, + "learning_rate": 8.71893815322216e-07, + "loss": 1.3827, + "step": 330 + }, + { + "epoch": 0.02, + "grad_norm": 2.98953226106279, + "learning_rate": 8.718911094999311e-07, + "loss": 1.3862, + "step": 331 + }, + { + "epoch": 0.02, + "grad_norm": 2.898375435464832, + "learning_rate": 8.718883919443554e-07, + "loss": 1.4134, + "step": 332 + }, + { + "epoch": 0.02, + "grad_norm": 2.9986915176885027, + "learning_rate": 8.718856626555621e-07, + "loss": 1.391, + "step": 333 + }, + { + "epoch": 0.02, + "grad_norm": 3.054114308668177, + "learning_rate": 8.718829216336246e-07, + "loss": 1.414, + "step": 334 + }, + { + "epoch": 0.02, + "grad_norm": 2.966556344517333, + "learning_rate": 8.718801688786166e-07, + "loss": 1.4188, + "step": 335 + }, + { + "epoch": 0.02, + "grad_norm": 3.1040321426982134, + "learning_rate": 8.718774043906126e-07, + "loss": 1.3538, + "step": 336 + }, + { + "epoch": 0.02, + "grad_norm": 3.0055709258768832, + "learning_rate": 8.718746281696866e-07, + "loss": 1.4413, + "step": 337 + }, + { + "epoch": 0.02, + "grad_norm": 2.9449143169829277, + "learning_rate": 8.718718402159136e-07, + "loss": 1.3449, + "step": 338 + }, + { + "epoch": 0.02, + "grad_norm": 2.9706428099571305, + "learning_rate": 8.718690405293686e-07, + "loss": 1.4158, + "step": 339 + }, + { + "epoch": 0.02, + "grad_norm": 3.0814964982203112, + "learning_rate": 8.718662291101268e-07, + "loss": 1.3981, + "step": 340 + }, + { + "epoch": 0.02, + "grad_norm": 3.036501892302947, + "learning_rate": 8.718634059582641e-07, + "loss": 1.4047, + "step": 341 + }, + { + "epoch": 0.02, + "grad_norm": 3.0667912243139535, + "learning_rate": 8.718605710738567e-07, + "loss": 1.4436, + "step": 342 + }, + { + "epoch": 0.02, + "grad_norm": 3.096788465549673, + "learning_rate": 8.718577244569806e-07, + "loss": 1.4332, + "step": 343 + }, + { + "epoch": 0.02, + "grad_norm": 3.031370455846918, + "learning_rate": 8.718548661077125e-07, + "loss": 1.3962, + "step": 344 + }, + { + "epoch": 0.02, + "grad_norm": 2.9672083000240344, + "learning_rate": 8.718519960261294e-07, + "loss": 1.4205, + "step": 345 + }, + { + "epoch": 0.02, + "grad_norm": 3.007883947100492, + "learning_rate": 8.718491142123086e-07, + "loss": 1.3446, + "step": 346 + }, + { + "epoch": 0.02, + "grad_norm": 2.9748300479515253, + "learning_rate": 8.718462206663277e-07, + "loss": 1.3854, + "step": 347 + }, + { + "epoch": 0.02, + "grad_norm": 3.0397247689440396, + "learning_rate": 8.718433153882645e-07, + "loss": 1.4125, + "step": 348 + }, + { + "epoch": 0.02, + "grad_norm": 3.001960481680682, + "learning_rate": 8.718403983781974e-07, + "loss": 1.3947, + "step": 349 + }, + { + "epoch": 0.02, + "grad_norm": 3.0664806351413088, + "learning_rate": 8.718374696362047e-07, + "loss": 1.3624, + "step": 350 + }, + { + "epoch": 0.02, + "grad_norm": 3.0210504760194175, + "learning_rate": 8.718345291623656e-07, + "loss": 1.4671, + "step": 351 + }, + { + "epoch": 0.02, + "grad_norm": 2.971388652881855, + "learning_rate": 8.718315769567588e-07, + "loss": 1.3472, + "step": 352 + }, + { + "epoch": 0.02, + "grad_norm": 2.952431051174851, + "learning_rate": 8.718286130194643e-07, + "loss": 1.3779, + "step": 353 + }, + { + "epoch": 0.02, + "grad_norm": 3.12691241920494, + "learning_rate": 8.718256373505615e-07, + "loss": 1.4117, + "step": 354 + }, + { + "epoch": 0.02, + "grad_norm": 3.0148285597590796, + "learning_rate": 8.718226499501307e-07, + "loss": 1.3676, + "step": 355 + }, + { + "epoch": 0.02, + "grad_norm": 2.98672263886442, + "learning_rate": 8.718196508182523e-07, + "loss": 1.4435, + "step": 356 + }, + { + "epoch": 0.02, + "grad_norm": 2.968930206941385, + "learning_rate": 8.718166399550071e-07, + "loss": 1.4378, + "step": 357 + }, + { + "epoch": 0.02, + "grad_norm": 3.0425445228617187, + "learning_rate": 8.718136173604761e-07, + "loss": 1.3597, + "step": 358 + }, + { + "epoch": 0.02, + "grad_norm": 2.9276879829658107, + "learning_rate": 8.718105830347405e-07, + "loss": 1.3689, + "step": 359 + }, + { + "epoch": 0.02, + "grad_norm": 2.9317936344250413, + "learning_rate": 8.718075369778825e-07, + "loss": 1.3721, + "step": 360 + }, + { + "epoch": 0.02, + "grad_norm": 2.9645490048095637, + "learning_rate": 8.718044791899837e-07, + "loss": 1.3987, + "step": 361 + }, + { + "epoch": 0.02, + "grad_norm": 3.0365564821005977, + "learning_rate": 8.718014096711265e-07, + "loss": 1.3868, + "step": 362 + }, + { + "epoch": 0.02, + "grad_norm": 2.9682770215203553, + "learning_rate": 8.717983284213936e-07, + "loss": 1.3415, + "step": 363 + }, + { + "epoch": 0.02, + "grad_norm": 3.134935849661195, + "learning_rate": 8.717952354408679e-07, + "loss": 1.3293, + "step": 364 + }, + { + "epoch": 0.02, + "grad_norm": 2.8124045479008384, + "learning_rate": 8.717921307296327e-07, + "loss": 1.4101, + "step": 365 + }, + { + "epoch": 0.02, + "grad_norm": 2.870855459457384, + "learning_rate": 8.717890142877717e-07, + "loss": 1.4129, + "step": 366 + }, + { + "epoch": 0.02, + "grad_norm": 2.996734500928963, + "learning_rate": 8.717858861153686e-07, + "loss": 1.4188, + "step": 367 + }, + { + "epoch": 0.02, + "grad_norm": 2.905301300393811, + "learning_rate": 8.717827462125079e-07, + "loss": 1.3503, + "step": 368 + }, + { + "epoch": 0.02, + "grad_norm": 2.963215198908182, + "learning_rate": 8.717795945792739e-07, + "loss": 1.3539, + "step": 369 + }, + { + "epoch": 0.02, + "grad_norm": 3.058342559604312, + "learning_rate": 8.717764312157515e-07, + "loss": 1.3911, + "step": 370 + }, + { + "epoch": 0.02, + "grad_norm": 2.9650165681938128, + "learning_rate": 8.717732561220258e-07, + "loss": 1.4207, + "step": 371 + }, + { + "epoch": 0.02, + "grad_norm": 2.9793800118049454, + "learning_rate": 8.717700692981826e-07, + "loss": 1.3691, + "step": 372 + }, + { + "epoch": 0.02, + "grad_norm": 2.9162869272769556, + "learning_rate": 8.717668707443075e-07, + "loss": 1.395, + "step": 373 + }, + { + "epoch": 0.02, + "grad_norm": 2.9636673086391485, + "learning_rate": 8.717636604604865e-07, + "loss": 1.4023, + "step": 374 + }, + { + "epoch": 0.02, + "grad_norm": 2.959298293762338, + "learning_rate": 8.717604384468061e-07, + "loss": 1.4328, + "step": 375 + }, + { + "epoch": 0.02, + "grad_norm": 2.965235704696395, + "learning_rate": 8.717572047033532e-07, + "loss": 1.4354, + "step": 376 + }, + { + "epoch": 0.02, + "grad_norm": 3.127226354296215, + "learning_rate": 8.717539592302147e-07, + "loss": 1.3904, + "step": 377 + }, + { + "epoch": 0.02, + "grad_norm": 2.9772007199986428, + "learning_rate": 8.717507020274781e-07, + "loss": 1.3997, + "step": 378 + }, + { + "epoch": 0.02, + "grad_norm": 3.1654089044177103, + "learning_rate": 8.717474330952311e-07, + "loss": 1.3664, + "step": 379 + }, + { + "epoch": 0.02, + "grad_norm": 2.897024330171534, + "learning_rate": 8.717441524335616e-07, + "loss": 1.3815, + "step": 380 + }, + { + "epoch": 0.02, + "grad_norm": 2.9908064118392947, + "learning_rate": 8.717408600425579e-07, + "loss": 1.4008, + "step": 381 + }, + { + "epoch": 0.02, + "grad_norm": 3.1674379998791475, + "learning_rate": 8.717375559223089e-07, + "loss": 1.4134, + "step": 382 + }, + { + "epoch": 0.02, + "grad_norm": 2.810265806813017, + "learning_rate": 8.717342400729033e-07, + "loss": 1.4046, + "step": 383 + }, + { + "epoch": 0.02, + "grad_norm": 2.9789538191272626, + "learning_rate": 8.717309124944306e-07, + "loss": 1.3957, + "step": 384 + }, + { + "epoch": 0.02, + "grad_norm": 2.995888914357539, + "learning_rate": 8.717275731869801e-07, + "loss": 1.3823, + "step": 385 + }, + { + "epoch": 0.02, + "grad_norm": 2.94123727534938, + "learning_rate": 8.71724222150642e-07, + "loss": 1.3577, + "step": 386 + }, + { + "epoch": 0.02, + "grad_norm": 2.907211393678648, + "learning_rate": 8.717208593855062e-07, + "loss": 1.4016, + "step": 387 + }, + { + "epoch": 0.02, + "grad_norm": 2.9549644446432546, + "learning_rate": 8.717174848916635e-07, + "loss": 1.3554, + "step": 388 + }, + { + "epoch": 0.02, + "grad_norm": 3.047404295254929, + "learning_rate": 8.717140986692047e-07, + "loss": 1.3977, + "step": 389 + }, + { + "epoch": 0.02, + "grad_norm": 3.0015613969735218, + "learning_rate": 8.717107007182211e-07, + "loss": 1.4159, + "step": 390 + }, + { + "epoch": 0.02, + "grad_norm": 2.99276033713888, + "learning_rate": 8.71707291038804e-07, + "loss": 1.4194, + "step": 391 + }, + { + "epoch": 0.02, + "grad_norm": 2.9613959661694427, + "learning_rate": 8.717038696310452e-07, + "loss": 1.4072, + "step": 392 + }, + { + "epoch": 0.02, + "grad_norm": 2.984569722219675, + "learning_rate": 8.717004364950369e-07, + "loss": 1.4018, + "step": 393 + }, + { + "epoch": 0.02, + "grad_norm": 2.927152138759416, + "learning_rate": 8.716969916308715e-07, + "loss": 1.4038, + "step": 394 + }, + { + "epoch": 0.02, + "grad_norm": 3.0771539333400764, + "learning_rate": 8.716935350386416e-07, + "loss": 1.3754, + "step": 395 + }, + { + "epoch": 0.02, + "grad_norm": 2.970985940726889, + "learning_rate": 8.716900667184406e-07, + "loss": 1.4458, + "step": 396 + }, + { + "epoch": 0.02, + "grad_norm": 3.0928065462412633, + "learning_rate": 8.716865866703617e-07, + "loss": 1.371, + "step": 397 + }, + { + "epoch": 0.02, + "grad_norm": 2.9309896563133697, + "learning_rate": 8.716830948944986e-07, + "loss": 1.3509, + "step": 398 + }, + { + "epoch": 0.02, + "grad_norm": 3.0455988834094736, + "learning_rate": 8.716795913909452e-07, + "loss": 1.3827, + "step": 399 + }, + { + "epoch": 0.02, + "grad_norm": 3.061462286190086, + "learning_rate": 8.716760761597961e-07, + "loss": 1.3926, + "step": 400 + }, + { + "epoch": 0.02, + "grad_norm": 2.9080714574645516, + "learning_rate": 8.716725492011458e-07, + "loss": 1.4101, + "step": 401 + }, + { + "epoch": 0.02, + "grad_norm": 2.9046604352395207, + "learning_rate": 8.716690105150891e-07, + "loss": 1.335, + "step": 402 + }, + { + "epoch": 0.02, + "grad_norm": 2.946411297505849, + "learning_rate": 8.716654601017216e-07, + "loss": 1.4109, + "step": 403 + }, + { + "epoch": 0.02, + "grad_norm": 2.9811491538335915, + "learning_rate": 8.716618979611386e-07, + "loss": 1.4007, + "step": 404 + }, + { + "epoch": 0.02, + "grad_norm": 2.828391151750033, + "learning_rate": 8.716583240934361e-07, + "loss": 1.4194, + "step": 405 + }, + { + "epoch": 0.02, + "grad_norm": 2.861846008744537, + "learning_rate": 8.716547384987104e-07, + "loss": 1.3164, + "step": 406 + }, + { + "epoch": 0.02, + "grad_norm": 2.8751261295501274, + "learning_rate": 8.716511411770581e-07, + "loss": 1.4447, + "step": 407 + }, + { + "epoch": 0.02, + "grad_norm": 3.085208227054159, + "learning_rate": 8.716475321285758e-07, + "loss": 1.3732, + "step": 408 + }, + { + "epoch": 0.02, + "grad_norm": 2.926763955103524, + "learning_rate": 8.716439113533609e-07, + "loss": 1.427, + "step": 409 + }, + { + "epoch": 0.02, + "grad_norm": 3.021150407067775, + "learning_rate": 8.716402788515107e-07, + "loss": 1.4123, + "step": 410 + }, + { + "epoch": 0.02, + "grad_norm": 3.0835953078832476, + "learning_rate": 8.716366346231232e-07, + "loss": 1.3225, + "step": 411 + }, + { + "epoch": 0.02, + "grad_norm": 3.084681723058033, + "learning_rate": 8.716329786682964e-07, + "loss": 1.4007, + "step": 412 + }, + { + "epoch": 0.02, + "grad_norm": 2.9354891039496507, + "learning_rate": 8.716293109871288e-07, + "loss": 1.374, + "step": 413 + }, + { + "epoch": 0.02, + "grad_norm": 3.072363269165642, + "learning_rate": 8.71625631579719e-07, + "loss": 1.355, + "step": 414 + }, + { + "epoch": 0.02, + "grad_norm": 2.9000514864569373, + "learning_rate": 8.716219404461663e-07, + "loss": 1.3718, + "step": 415 + }, + { + "epoch": 0.02, + "grad_norm": 3.0562397719571766, + "learning_rate": 8.716182375865698e-07, + "loss": 1.3814, + "step": 416 + }, + { + "epoch": 0.02, + "grad_norm": 3.0526646692685717, + "learning_rate": 8.716145230010296e-07, + "loss": 1.3772, + "step": 417 + }, + { + "epoch": 0.02, + "grad_norm": 3.0966573543083538, + "learning_rate": 8.716107966896452e-07, + "loss": 1.4287, + "step": 418 + }, + { + "epoch": 0.02, + "grad_norm": 3.0880361698977263, + "learning_rate": 8.716070586525174e-07, + "loss": 1.3751, + "step": 419 + }, + { + "epoch": 0.02, + "grad_norm": 2.832576232775376, + "learning_rate": 8.716033088897465e-07, + "loss": 1.416, + "step": 420 + }, + { + "epoch": 0.02, + "grad_norm": 2.965237748755546, + "learning_rate": 8.715995474014337e-07, + "loss": 1.3621, + "step": 421 + }, + { + "epoch": 0.02, + "grad_norm": 3.0463957419921783, + "learning_rate": 8.7159577418768e-07, + "loss": 1.357, + "step": 422 + }, + { + "epoch": 0.02, + "grad_norm": 2.9739015917876723, + "learning_rate": 8.715919892485873e-07, + "loss": 1.3873, + "step": 423 + }, + { + "epoch": 0.02, + "grad_norm": 2.8676852163892037, + "learning_rate": 8.715881925842573e-07, + "loss": 1.4051, + "step": 424 + }, + { + "epoch": 0.02, + "grad_norm": 3.088087320089484, + "learning_rate": 8.715843841947923e-07, + "loss": 1.3151, + "step": 425 + }, + { + "epoch": 0.02, + "grad_norm": 3.0270665713855367, + "learning_rate": 8.715805640802949e-07, + "loss": 1.3679, + "step": 426 + }, + { + "epoch": 0.02, + "grad_norm": 2.9869536543983366, + "learning_rate": 8.715767322408678e-07, + "loss": 1.3528, + "step": 427 + }, + { + "epoch": 0.02, + "grad_norm": 2.9112845523257675, + "learning_rate": 8.715728886766143e-07, + "loss": 1.442, + "step": 428 + }, + { + "epoch": 0.02, + "grad_norm": 3.0049960527344544, + "learning_rate": 8.715690333876378e-07, + "loss": 1.3681, + "step": 429 + }, + { + "epoch": 0.02, + "grad_norm": 2.9921307040597664, + "learning_rate": 8.715651663740421e-07, + "loss": 1.4314, + "step": 430 + }, + { + "epoch": 0.02, + "grad_norm": 3.0109768676656605, + "learning_rate": 8.715612876359315e-07, + "loss": 1.3847, + "step": 431 + }, + { + "epoch": 0.02, + "grad_norm": 3.039247694352697, + "learning_rate": 8.715573971734103e-07, + "loss": 1.4317, + "step": 432 + }, + { + "epoch": 0.02, + "grad_norm": 3.019898918932762, + "learning_rate": 8.71553494986583e-07, + "loss": 1.3623, + "step": 433 + }, + { + "epoch": 0.02, + "grad_norm": 3.00522079476457, + "learning_rate": 8.71549581075555e-07, + "loss": 1.3884, + "step": 434 + }, + { + "epoch": 0.02, + "grad_norm": 2.9610972813683794, + "learning_rate": 8.715456554404316e-07, + "loss": 1.3315, + "step": 435 + }, + { + "epoch": 0.02, + "grad_norm": 2.967170476790787, + "learning_rate": 8.715417180813185e-07, + "loss": 1.4207, + "step": 436 + }, + { + "epoch": 0.02, + "grad_norm": 2.9522241028635285, + "learning_rate": 8.715377689983216e-07, + "loss": 1.4012, + "step": 437 + }, + { + "epoch": 0.02, + "grad_norm": 2.928961724018662, + "learning_rate": 8.715338081915475e-07, + "loss": 1.3869, + "step": 438 + }, + { + "epoch": 0.02, + "grad_norm": 2.952221930130988, + "learning_rate": 8.715298356611025e-07, + "loss": 1.3703, + "step": 439 + }, + { + "epoch": 0.02, + "grad_norm": 3.0536781428675335, + "learning_rate": 8.715258514070937e-07, + "loss": 1.3682, + "step": 440 + }, + { + "epoch": 0.02, + "grad_norm": 3.128913913582799, + "learning_rate": 8.715218554296284e-07, + "loss": 1.3435, + "step": 441 + }, + { + "epoch": 0.02, + "grad_norm": 3.0301910736766318, + "learning_rate": 8.715178477288141e-07, + "loss": 1.3975, + "step": 442 + }, + { + "epoch": 0.02, + "grad_norm": 2.962217823506329, + "learning_rate": 8.715138283047589e-07, + "loss": 1.3488, + "step": 443 + }, + { + "epoch": 0.02, + "grad_norm": 3.0039607070469647, + "learning_rate": 8.715097971575708e-07, + "loss": 1.378, + "step": 444 + }, + { + "epoch": 0.02, + "grad_norm": 3.0602288102993938, + "learning_rate": 8.715057542873585e-07, + "loss": 1.3572, + "step": 445 + }, + { + "epoch": 0.02, + "grad_norm": 2.969434960629606, + "learning_rate": 8.715016996942307e-07, + "loss": 1.3713, + "step": 446 + }, + { + "epoch": 0.02, + "grad_norm": 2.9917667276430477, + "learning_rate": 8.714976333782967e-07, + "loss": 1.4607, + "step": 447 + }, + { + "epoch": 0.02, + "grad_norm": 3.2046134996002436, + "learning_rate": 8.714935553396659e-07, + "loss": 1.3853, + "step": 448 + }, + { + "epoch": 0.02, + "grad_norm": 3.1097764355868733, + "learning_rate": 8.714894655784481e-07, + "loss": 1.371, + "step": 449 + }, + { + "epoch": 0.02, + "grad_norm": 2.9715612348659124, + "learning_rate": 8.714853640947534e-07, + "loss": 1.392, + "step": 450 + }, + { + "epoch": 0.02, + "grad_norm": 2.9321597700386333, + "learning_rate": 8.714812508886925e-07, + "loss": 1.4051, + "step": 451 + }, + { + "epoch": 0.02, + "grad_norm": 2.949095796893927, + "learning_rate": 8.714771259603758e-07, + "loss": 1.3469, + "step": 452 + }, + { + "epoch": 0.02, + "grad_norm": 2.9290487074046756, + "learning_rate": 8.714729893099144e-07, + "loss": 1.3629, + "step": 453 + }, + { + "epoch": 0.02, + "grad_norm": 3.0295532590400196, + "learning_rate": 8.714688409374198e-07, + "loss": 1.3689, + "step": 454 + }, + { + "epoch": 0.02, + "grad_norm": 2.965897130998238, + "learning_rate": 8.714646808430036e-07, + "loss": 1.3619, + "step": 455 + }, + { + "epoch": 0.02, + "grad_norm": 3.0935335333535336, + "learning_rate": 8.714605090267779e-07, + "loss": 1.3781, + "step": 456 + }, + { + "epoch": 0.02, + "grad_norm": 2.986770714363214, + "learning_rate": 8.71456325488855e-07, + "loss": 1.364, + "step": 457 + }, + { + "epoch": 0.02, + "grad_norm": 2.950553960144967, + "learning_rate": 8.714521302293475e-07, + "loss": 1.4063, + "step": 458 + }, + { + "epoch": 0.02, + "grad_norm": 3.014174300238565, + "learning_rate": 8.714479232483683e-07, + "loss": 1.3676, + "step": 459 + }, + { + "epoch": 0.02, + "grad_norm": 3.24671603526281, + "learning_rate": 8.714437045460308e-07, + "loss": 1.3578, + "step": 460 + }, + { + "epoch": 0.02, + "grad_norm": 3.1623556544347724, + "learning_rate": 8.714394741224484e-07, + "loss": 1.3645, + "step": 461 + }, + { + "epoch": 0.02, + "grad_norm": 3.0025864567325122, + "learning_rate": 8.714352319777354e-07, + "loss": 1.3607, + "step": 462 + }, + { + "epoch": 0.02, + "grad_norm": 2.94717894620833, + "learning_rate": 8.714309781120056e-07, + "loss": 1.3498, + "step": 463 + }, + { + "epoch": 0.02, + "grad_norm": 2.8964926477515935, + "learning_rate": 8.714267125253735e-07, + "loss": 1.3728, + "step": 464 + }, + { + "epoch": 0.02, + "grad_norm": 3.1656664213505823, + "learning_rate": 8.714224352179544e-07, + "loss": 1.4176, + "step": 465 + }, + { + "epoch": 0.02, + "grad_norm": 3.020188270295796, + "learning_rate": 8.71418146189863e-07, + "loss": 1.4194, + "step": 466 + }, + { + "epoch": 0.02, + "grad_norm": 2.978569270321128, + "learning_rate": 8.71413845441215e-07, + "loss": 1.3749, + "step": 467 + }, + { + "epoch": 0.02, + "grad_norm": 2.9396855197827283, + "learning_rate": 8.714095329721261e-07, + "loss": 1.3795, + "step": 468 + }, + { + "epoch": 0.02, + "grad_norm": 2.9422669465457387, + "learning_rate": 8.714052087827125e-07, + "loss": 1.3834, + "step": 469 + }, + { + "epoch": 0.02, + "grad_norm": 3.010282715520298, + "learning_rate": 8.714008728730907e-07, + "loss": 1.3531, + "step": 470 + }, + { + "epoch": 0.02, + "grad_norm": 2.9717500175741116, + "learning_rate": 8.713965252433773e-07, + "loss": 1.3667, + "step": 471 + }, + { + "epoch": 0.02, + "grad_norm": 3.052704318228861, + "learning_rate": 8.713921658936892e-07, + "loss": 1.3456, + "step": 472 + }, + { + "epoch": 0.02, + "grad_norm": 3.528710625831412, + "learning_rate": 8.713877948241442e-07, + "loss": 1.3936, + "step": 473 + }, + { + "epoch": 0.02, + "grad_norm": 2.9466834730933833, + "learning_rate": 8.713834120348596e-07, + "loss": 1.3217, + "step": 474 + }, + { + "epoch": 0.02, + "grad_norm": 3.0996541898765226, + "learning_rate": 8.713790175259536e-07, + "loss": 1.3855, + "step": 475 + }, + { + "epoch": 0.02, + "grad_norm": 2.982252067970699, + "learning_rate": 8.713746112975446e-07, + "loss": 1.384, + "step": 476 + }, + { + "epoch": 0.02, + "grad_norm": 3.1109077626663844, + "learning_rate": 8.713701933497509e-07, + "loss": 1.3561, + "step": 477 + }, + { + "epoch": 0.02, + "grad_norm": 3.0394061264038115, + "learning_rate": 8.713657636826918e-07, + "loss": 1.468, + "step": 478 + }, + { + "epoch": 0.02, + "grad_norm": 3.0969706383479, + "learning_rate": 8.713613222964863e-07, + "loss": 1.3993, + "step": 479 + }, + { + "epoch": 0.02, + "grad_norm": 3.0348133446475662, + "learning_rate": 8.713568691912542e-07, + "loss": 1.387, + "step": 480 + }, + { + "epoch": 0.02, + "grad_norm": 3.0273826285615195, + "learning_rate": 8.713524043671153e-07, + "loss": 1.3959, + "step": 481 + }, + { + "epoch": 0.03, + "grad_norm": 3.141130816615921, + "learning_rate": 8.713479278241898e-07, + "loss": 1.4479, + "step": 482 + }, + { + "epoch": 0.03, + "grad_norm": 3.0490824173681945, + "learning_rate": 8.713434395625983e-07, + "loss": 1.3583, + "step": 483 + }, + { + "epoch": 0.03, + "grad_norm": 2.9474474845957404, + "learning_rate": 8.713389395824614e-07, + "loss": 1.3344, + "step": 484 + }, + { + "epoch": 0.03, + "grad_norm": 2.9486918699061118, + "learning_rate": 8.713344278839005e-07, + "loss": 1.4022, + "step": 485 + }, + { + "epoch": 0.03, + "grad_norm": 2.9302492343255344, + "learning_rate": 8.71329904467037e-07, + "loss": 1.3344, + "step": 486 + }, + { + "epoch": 0.03, + "grad_norm": 2.9531126631866753, + "learning_rate": 8.713253693319929e-07, + "loss": 1.3451, + "step": 487 + }, + { + "epoch": 0.03, + "grad_norm": 2.8555050387157213, + "learning_rate": 8.713208224788899e-07, + "loss": 1.3287, + "step": 488 + }, + { + "epoch": 0.03, + "grad_norm": 3.0850474903019305, + "learning_rate": 8.713162639078507e-07, + "loss": 1.4153, + "step": 489 + }, + { + "epoch": 0.03, + "grad_norm": 2.9916349483264004, + "learning_rate": 8.71311693618998e-07, + "loss": 1.4025, + "step": 490 + }, + { + "epoch": 0.03, + "grad_norm": 3.0448173115664545, + "learning_rate": 8.713071116124549e-07, + "loss": 1.4129, + "step": 491 + }, + { + "epoch": 0.03, + "grad_norm": 2.9652773442800022, + "learning_rate": 8.713025178883445e-07, + "loss": 1.3688, + "step": 492 + }, + { + "epoch": 0.03, + "grad_norm": 2.8707509231071127, + "learning_rate": 8.712979124467906e-07, + "loss": 1.3714, + "step": 493 + }, + { + "epoch": 0.03, + "grad_norm": 2.968740384281211, + "learning_rate": 8.712932952879176e-07, + "loss": 1.4012, + "step": 494 + }, + { + "epoch": 0.03, + "grad_norm": 2.938361663169202, + "learning_rate": 8.712886664118492e-07, + "loss": 1.406, + "step": 495 + }, + { + "epoch": 0.03, + "grad_norm": 3.1217977871072775, + "learning_rate": 8.712840258187104e-07, + "loss": 1.3822, + "step": 496 + }, + { + "epoch": 0.03, + "grad_norm": 2.9353035115177915, + "learning_rate": 8.71279373508626e-07, + "loss": 1.3578, + "step": 497 + }, + { + "epoch": 0.03, + "grad_norm": 3.014072001165337, + "learning_rate": 8.712747094817213e-07, + "loss": 1.4454, + "step": 498 + }, + { + "epoch": 0.03, + "grad_norm": 2.969346395207918, + "learning_rate": 8.71270033738122e-07, + "loss": 1.3705, + "step": 499 + }, + { + "epoch": 0.03, + "grad_norm": 3.050484611788353, + "learning_rate": 8.712653462779539e-07, + "loss": 1.3551, + "step": 500 + }, + { + "epoch": 0.03, + "grad_norm": 2.940761479764165, + "learning_rate": 8.71260647101343e-07, + "loss": 1.4147, + "step": 501 + }, + { + "epoch": 0.03, + "grad_norm": 2.9928956127151944, + "learning_rate": 8.712559362084161e-07, + "loss": 1.3604, + "step": 502 + }, + { + "epoch": 0.03, + "grad_norm": 3.0125472864930805, + "learning_rate": 8.712512135993e-07, + "loss": 1.4183, + "step": 503 + }, + { + "epoch": 0.03, + "grad_norm": 2.940381713821346, + "learning_rate": 8.712464792741218e-07, + "loss": 1.4414, + "step": 504 + }, + { + "epoch": 0.03, + "grad_norm": 3.0279511535103336, + "learning_rate": 8.712417332330089e-07, + "loss": 1.3505, + "step": 505 + }, + { + "epoch": 0.03, + "grad_norm": 2.9922191158837954, + "learning_rate": 8.712369754760892e-07, + "loss": 1.4028, + "step": 506 + }, + { + "epoch": 0.03, + "grad_norm": 3.0184990677208834, + "learning_rate": 8.712322060034907e-07, + "loss": 1.3465, + "step": 507 + }, + { + "epoch": 0.03, + "grad_norm": 3.05423073796875, + "learning_rate": 8.712274248153418e-07, + "loss": 1.3416, + "step": 508 + }, + { + "epoch": 0.03, + "grad_norm": 3.1326601737410025, + "learning_rate": 8.712226319117715e-07, + "loss": 1.3924, + "step": 509 + }, + { + "epoch": 0.03, + "grad_norm": 2.995785462842688, + "learning_rate": 8.712178272929084e-07, + "loss": 1.3895, + "step": 510 + }, + { + "epoch": 0.03, + "grad_norm": 2.9760799944825806, + "learning_rate": 8.712130109588823e-07, + "loss": 1.4104, + "step": 511 + }, + { + "epoch": 0.03, + "grad_norm": 2.9654696172148896, + "learning_rate": 8.712081829098225e-07, + "loss": 1.378, + "step": 512 + }, + { + "epoch": 0.03, + "grad_norm": 2.9968608921090265, + "learning_rate": 8.712033431458593e-07, + "loss": 1.4264, + "step": 513 + }, + { + "epoch": 0.03, + "grad_norm": 2.969211803711655, + "learning_rate": 8.711984916671229e-07, + "loss": 1.3607, + "step": 514 + }, + { + "epoch": 0.03, + "grad_norm": 2.942118444739339, + "learning_rate": 8.711936284737438e-07, + "loss": 1.3899, + "step": 515 + }, + { + "epoch": 0.03, + "grad_norm": 2.98791974032871, + "learning_rate": 8.711887535658529e-07, + "loss": 1.3459, + "step": 516 + }, + { + "epoch": 0.03, + "grad_norm": 2.9216444622419537, + "learning_rate": 8.711838669435818e-07, + "loss": 1.4116, + "step": 517 + }, + { + "epoch": 0.03, + "grad_norm": 2.9987312946412206, + "learning_rate": 8.711789686070618e-07, + "loss": 1.4126, + "step": 518 + }, + { + "epoch": 0.03, + "grad_norm": 2.9997602427201926, + "learning_rate": 8.711740585564249e-07, + "loss": 1.3392, + "step": 519 + }, + { + "epoch": 0.03, + "grad_norm": 2.9736815652309683, + "learning_rate": 8.711691367918032e-07, + "loss": 1.3483, + "step": 520 + }, + { + "epoch": 0.03, + "grad_norm": 3.0854886077878487, + "learning_rate": 8.711642033133292e-07, + "loss": 1.3842, + "step": 521 + }, + { + "epoch": 0.03, + "grad_norm": 3.061838785099914, + "learning_rate": 8.711592581211358e-07, + "loss": 1.3658, + "step": 522 + }, + { + "epoch": 0.03, + "grad_norm": 3.023168820723632, + "learning_rate": 8.711543012153561e-07, + "loss": 1.3929, + "step": 523 + }, + { + "epoch": 0.03, + "grad_norm": 2.970076769155023, + "learning_rate": 8.711493325961236e-07, + "loss": 1.3587, + "step": 524 + }, + { + "epoch": 0.03, + "grad_norm": 3.002070690376611, + "learning_rate": 8.71144352263572e-07, + "loss": 1.367, + "step": 525 + }, + { + "epoch": 0.03, + "grad_norm": 2.995354883459173, + "learning_rate": 8.711393602178357e-07, + "loss": 1.4185, + "step": 526 + }, + { + "epoch": 0.03, + "grad_norm": 24.26995863156332, + "learning_rate": 8.711343564590487e-07, + "loss": 1.3982, + "step": 527 + }, + { + "epoch": 0.03, + "grad_norm": 3.065387181865347, + "learning_rate": 8.711293409873459e-07, + "loss": 1.3668, + "step": 528 + }, + { + "epoch": 0.03, + "grad_norm": 2.965733738900826, + "learning_rate": 8.711243138028624e-07, + "loss": 1.3989, + "step": 529 + }, + { + "epoch": 0.03, + "grad_norm": 2.909922562005101, + "learning_rate": 8.711192749057334e-07, + "loss": 1.3548, + "step": 530 + }, + { + "epoch": 0.03, + "grad_norm": 2.9485719715840824, + "learning_rate": 8.711142242960946e-07, + "loss": 1.3982, + "step": 531 + }, + { + "epoch": 0.03, + "grad_norm": 3.0587521051463598, + "learning_rate": 8.711091619740822e-07, + "loss": 1.4208, + "step": 532 + }, + { + "epoch": 0.03, + "grad_norm": 3.016741791642564, + "learning_rate": 8.711040879398322e-07, + "loss": 1.3776, + "step": 533 + }, + { + "epoch": 0.03, + "grad_norm": 3.087201787399773, + "learning_rate": 8.710990021934814e-07, + "loss": 1.3509, + "step": 534 + }, + { + "epoch": 0.03, + "grad_norm": 3.1398002317055873, + "learning_rate": 8.710939047351665e-07, + "loss": 1.4054, + "step": 535 + }, + { + "epoch": 0.03, + "grad_norm": 3.003141522348306, + "learning_rate": 8.710887955650252e-07, + "loss": 1.3895, + "step": 536 + }, + { + "epoch": 0.03, + "grad_norm": 3.0461324715788862, + "learning_rate": 8.710836746831946e-07, + "loss": 1.4143, + "step": 537 + }, + { + "epoch": 0.03, + "grad_norm": 2.956952816424475, + "learning_rate": 8.710785420898127e-07, + "loss": 1.4008, + "step": 538 + }, + { + "epoch": 0.03, + "grad_norm": 2.8987555743390168, + "learning_rate": 8.710733977850179e-07, + "loss": 1.3823, + "step": 539 + }, + { + "epoch": 0.03, + "grad_norm": 3.0536968376300284, + "learning_rate": 8.710682417689485e-07, + "loss": 1.4578, + "step": 540 + }, + { + "epoch": 0.03, + "grad_norm": 2.969066979149947, + "learning_rate": 8.710630740417435e-07, + "loss": 1.428, + "step": 541 + }, + { + "epoch": 0.03, + "grad_norm": 3.1272335317584905, + "learning_rate": 8.710578946035417e-07, + "loss": 1.3562, + "step": 542 + }, + { + "epoch": 0.03, + "grad_norm": 2.9753498641377742, + "learning_rate": 8.710527034544828e-07, + "loss": 1.3953, + "step": 543 + }, + { + "epoch": 0.03, + "grad_norm": 2.943913278878306, + "learning_rate": 8.710475005947067e-07, + "loss": 1.3626, + "step": 544 + }, + { + "epoch": 0.03, + "grad_norm": 2.9513572324834727, + "learning_rate": 8.710422860243531e-07, + "loss": 1.3461, + "step": 545 + }, + { + "epoch": 0.03, + "grad_norm": 3.0849634901753284, + "learning_rate": 8.710370597435629e-07, + "loss": 1.3663, + "step": 546 + }, + { + "epoch": 0.03, + "grad_norm": 3.138354533416878, + "learning_rate": 8.710318217524763e-07, + "loss": 1.3141, + "step": 547 + }, + { + "epoch": 0.03, + "grad_norm": 2.972312260895404, + "learning_rate": 8.710265720512346e-07, + "loss": 1.3633, + "step": 548 + }, + { + "epoch": 0.03, + "grad_norm": 3.0591517440498097, + "learning_rate": 8.710213106399791e-07, + "loss": 1.3557, + "step": 549 + }, + { + "epoch": 0.03, + "grad_norm": 2.8524186182545987, + "learning_rate": 8.710160375188516e-07, + "loss": 1.3676, + "step": 550 + }, + { + "epoch": 0.03, + "grad_norm": 2.9822890833732956, + "learning_rate": 8.710107526879938e-07, + "loss": 1.4331, + "step": 551 + }, + { + "epoch": 0.03, + "grad_norm": 2.9176351551822868, + "learning_rate": 8.710054561475481e-07, + "loss": 1.371, + "step": 552 + }, + { + "epoch": 0.03, + "grad_norm": 2.94388048856757, + "learning_rate": 8.71000147897657e-07, + "loss": 1.393, + "step": 553 + }, + { + "epoch": 0.03, + "grad_norm": 3.085541520730895, + "learning_rate": 8.709948279384639e-07, + "loss": 1.3937, + "step": 554 + }, + { + "epoch": 0.03, + "grad_norm": 2.9491352117439757, + "learning_rate": 8.709894962701115e-07, + "loss": 1.3526, + "step": 555 + }, + { + "epoch": 0.03, + "grad_norm": 2.9933004115427444, + "learning_rate": 8.709841528927436e-07, + "loss": 1.3916, + "step": 556 + }, + { + "epoch": 0.03, + "grad_norm": 2.959458221957766, + "learning_rate": 8.70978797806504e-07, + "loss": 1.3462, + "step": 557 + }, + { + "epoch": 0.03, + "grad_norm": 2.963203467802987, + "learning_rate": 8.709734310115368e-07, + "loss": 1.3783, + "step": 558 + }, + { + "epoch": 0.03, + "grad_norm": 2.9902951999179312, + "learning_rate": 8.709680525079866e-07, + "loss": 1.362, + "step": 559 + }, + { + "epoch": 0.03, + "grad_norm": 2.948654339418917, + "learning_rate": 8.709626622959983e-07, + "loss": 1.3841, + "step": 560 + }, + { + "epoch": 0.03, + "grad_norm": 3.001564005849718, + "learning_rate": 8.709572603757169e-07, + "loss": 1.3572, + "step": 561 + }, + { + "epoch": 0.03, + "grad_norm": 3.0380595324448416, + "learning_rate": 8.709518467472878e-07, + "loss": 1.4171, + "step": 562 + }, + { + "epoch": 0.03, + "grad_norm": 2.896631790474921, + "learning_rate": 8.709464214108568e-07, + "loss": 1.3448, + "step": 563 + }, + { + "epoch": 0.03, + "grad_norm": 2.9319635944784475, + "learning_rate": 8.709409843665701e-07, + "loss": 1.3917, + "step": 564 + }, + { + "epoch": 0.03, + "grad_norm": 2.963100470234962, + "learning_rate": 8.709355356145739e-07, + "loss": 1.3655, + "step": 565 + }, + { + "epoch": 0.03, + "grad_norm": 3.030553118915969, + "learning_rate": 8.709300751550151e-07, + "loss": 1.3365, + "step": 566 + }, + { + "epoch": 0.03, + "grad_norm": 3.4928530382016887, + "learning_rate": 8.709246029880405e-07, + "loss": 1.3662, + "step": 567 + }, + { + "epoch": 0.03, + "grad_norm": 3.030669221337684, + "learning_rate": 8.709191191137976e-07, + "loss": 1.4529, + "step": 568 + }, + { + "epoch": 0.03, + "grad_norm": 3.155457882933603, + "learning_rate": 8.70913623532434e-07, + "loss": 1.4022, + "step": 569 + }, + { + "epoch": 0.03, + "grad_norm": 2.9856956704586115, + "learning_rate": 8.709081162440975e-07, + "loss": 1.3989, + "step": 570 + }, + { + "epoch": 0.03, + "grad_norm": 2.9343339295028015, + "learning_rate": 8.709025972489367e-07, + "loss": 1.4212, + "step": 571 + }, + { + "epoch": 0.03, + "grad_norm": 3.1952129991940543, + "learning_rate": 8.708970665471e-07, + "loss": 1.4436, + "step": 572 + }, + { + "epoch": 0.03, + "grad_norm": 3.015106319776789, + "learning_rate": 8.708915241387364e-07, + "loss": 1.4422, + "step": 573 + }, + { + "epoch": 0.03, + "grad_norm": 3.022781702862628, + "learning_rate": 8.708859700239951e-07, + "loss": 1.3789, + "step": 574 + }, + { + "epoch": 0.03, + "grad_norm": 2.932454535389442, + "learning_rate": 8.708804042030254e-07, + "loss": 1.3783, + "step": 575 + }, + { + "epoch": 0.03, + "grad_norm": 3.1075106961875165, + "learning_rate": 8.708748266759774e-07, + "loss": 1.3904, + "step": 576 + }, + { + "epoch": 0.03, + "grad_norm": 2.8613959925148356, + "learning_rate": 8.708692374430014e-07, + "loss": 1.3701, + "step": 577 + }, + { + "epoch": 0.03, + "grad_norm": 2.9916282178733447, + "learning_rate": 8.708636365042476e-07, + "loss": 1.3517, + "step": 578 + }, + { + "epoch": 0.03, + "grad_norm": 2.8578706799379554, + "learning_rate": 8.70858023859867e-07, + "loss": 1.4124, + "step": 579 + }, + { + "epoch": 0.03, + "grad_norm": 3.0055644119433533, + "learning_rate": 8.708523995100105e-07, + "loss": 1.3869, + "step": 580 + }, + { + "epoch": 0.03, + "grad_norm": 3.027942807639001, + "learning_rate": 8.708467634548298e-07, + "loss": 1.3703, + "step": 581 + }, + { + "epoch": 0.03, + "grad_norm": 3.1584010302430836, + "learning_rate": 8.708411156944765e-07, + "loss": 1.3852, + "step": 582 + }, + { + "epoch": 0.03, + "grad_norm": 3.08434624814754, + "learning_rate": 8.708354562291027e-07, + "loss": 1.4008, + "step": 583 + }, + { + "epoch": 0.03, + "grad_norm": 2.937949504985518, + "learning_rate": 8.708297850588607e-07, + "loss": 1.4026, + "step": 584 + }, + { + "epoch": 0.03, + "grad_norm": 3.0080974250898116, + "learning_rate": 8.708241021839032e-07, + "loss": 1.3992, + "step": 585 + }, + { + "epoch": 0.03, + "grad_norm": 2.986930909094618, + "learning_rate": 8.708184076043833e-07, + "loss": 1.3504, + "step": 586 + }, + { + "epoch": 0.03, + "grad_norm": 3.1912323292885363, + "learning_rate": 8.708127013204543e-07, + "loss": 1.3919, + "step": 587 + }, + { + "epoch": 0.03, + "grad_norm": 3.203030731079236, + "learning_rate": 8.708069833322698e-07, + "loss": 1.3601, + "step": 588 + }, + { + "epoch": 0.03, + "grad_norm": 3.097916109715531, + "learning_rate": 8.708012536399837e-07, + "loss": 1.3619, + "step": 589 + }, + { + "epoch": 0.03, + "grad_norm": 3.0656903032705327, + "learning_rate": 8.707955122437504e-07, + "loss": 1.3162, + "step": 590 + }, + { + "epoch": 0.03, + "grad_norm": 3.002875260439292, + "learning_rate": 8.707897591437243e-07, + "loss": 1.389, + "step": 591 + }, + { + "epoch": 0.03, + "grad_norm": 3.0143271490705077, + "learning_rate": 8.707839943400606e-07, + "loss": 1.3323, + "step": 592 + }, + { + "epoch": 0.03, + "grad_norm": 3.096313689982386, + "learning_rate": 8.707782178329142e-07, + "loss": 1.3813, + "step": 593 + }, + { + "epoch": 0.03, + "grad_norm": 2.960471861335116, + "learning_rate": 8.707724296224408e-07, + "loss": 1.3472, + "step": 594 + }, + { + "epoch": 0.03, + "grad_norm": 3.0072247729027115, + "learning_rate": 8.707666297087963e-07, + "loss": 1.3522, + "step": 595 + }, + { + "epoch": 0.03, + "grad_norm": 2.949063388714746, + "learning_rate": 8.707608180921366e-07, + "loss": 1.3928, + "step": 596 + }, + { + "epoch": 0.03, + "grad_norm": 3.1121870350751064, + "learning_rate": 8.707549947726183e-07, + "loss": 1.4399, + "step": 597 + }, + { + "epoch": 0.03, + "grad_norm": 2.9609559500955003, + "learning_rate": 8.707491597503982e-07, + "loss": 1.3898, + "step": 598 + }, + { + "epoch": 0.03, + "grad_norm": 2.9585328248619587, + "learning_rate": 8.707433130256336e-07, + "loss": 1.379, + "step": 599 + }, + { + "epoch": 0.03, + "grad_norm": 2.918687510312111, + "learning_rate": 8.707374545984816e-07, + "loss": 1.4176, + "step": 600 + }, + { + "epoch": 0.03, + "grad_norm": 3.15488138702356, + "learning_rate": 8.707315844691002e-07, + "loss": 1.3706, + "step": 601 + }, + { + "epoch": 0.03, + "grad_norm": 2.9610415881399677, + "learning_rate": 8.707257026376471e-07, + "loss": 1.3641, + "step": 602 + }, + { + "epoch": 0.03, + "grad_norm": 3.0862098441133234, + "learning_rate": 8.707198091042811e-07, + "loss": 1.3893, + "step": 603 + }, + { + "epoch": 0.03, + "grad_norm": 2.9328888264242106, + "learning_rate": 8.707139038691606e-07, + "loss": 1.333, + "step": 604 + }, + { + "epoch": 0.03, + "grad_norm": 2.97580891347495, + "learning_rate": 8.707079869324446e-07, + "loss": 1.3607, + "step": 605 + }, + { + "epoch": 0.03, + "grad_norm": 3.000962296933563, + "learning_rate": 8.707020582942925e-07, + "loss": 1.424, + "step": 606 + }, + { + "epoch": 0.03, + "grad_norm": 2.984700606909975, + "learning_rate": 8.706961179548639e-07, + "loss": 1.3912, + "step": 607 + }, + { + "epoch": 0.03, + "grad_norm": 2.998907818017932, + "learning_rate": 8.706901659143189e-07, + "loss": 1.4241, + "step": 608 + }, + { + "epoch": 0.03, + "grad_norm": 2.95254670370234, + "learning_rate": 8.706842021728173e-07, + "loss": 1.3759, + "step": 609 + }, + { + "epoch": 0.03, + "grad_norm": 3.134652976665768, + "learning_rate": 8.706782267305202e-07, + "loss": 1.3767, + "step": 610 + }, + { + "epoch": 0.03, + "grad_norm": 2.972318633221704, + "learning_rate": 8.706722395875881e-07, + "loss": 1.3648, + "step": 611 + }, + { + "epoch": 0.03, + "grad_norm": 2.8607560362869124, + "learning_rate": 8.706662407441824e-07, + "loss": 1.3946, + "step": 612 + }, + { + "epoch": 0.03, + "grad_norm": 2.8921206621422653, + "learning_rate": 8.706602302004645e-07, + "loss": 1.4396, + "step": 613 + }, + { + "epoch": 0.03, + "grad_norm": 2.949264928829496, + "learning_rate": 8.706542079565962e-07, + "loss": 1.3475, + "step": 614 + }, + { + "epoch": 0.03, + "grad_norm": 2.996499503605724, + "learning_rate": 8.706481740127399e-07, + "loss": 1.37, + "step": 615 + }, + { + "epoch": 0.03, + "grad_norm": 2.9235932803770868, + "learning_rate": 8.706421283690578e-07, + "loss": 1.2987, + "step": 616 + }, + { + "epoch": 0.03, + "grad_norm": 3.0655234851680824, + "learning_rate": 8.706360710257128e-07, + "loss": 1.3903, + "step": 617 + }, + { + "epoch": 0.03, + "grad_norm": 2.968020759665533, + "learning_rate": 8.706300019828679e-07, + "loss": 1.4227, + "step": 618 + }, + { + "epoch": 0.03, + "grad_norm": 3.116571254583397, + "learning_rate": 8.706239212406866e-07, + "loss": 1.4153, + "step": 619 + }, + { + "epoch": 0.03, + "grad_norm": 3.0578559650775383, + "learning_rate": 8.706178287993326e-07, + "loss": 1.4168, + "step": 620 + }, + { + "epoch": 0.03, + "grad_norm": 2.9399476742786907, + "learning_rate": 8.706117246589699e-07, + "loss": 1.3448, + "step": 621 + }, + { + "epoch": 0.03, + "grad_norm": 3.045715774895227, + "learning_rate": 8.706056088197628e-07, + "loss": 1.4323, + "step": 622 + }, + { + "epoch": 0.03, + "grad_norm": 2.8892619986413655, + "learning_rate": 8.705994812818759e-07, + "loss": 1.3688, + "step": 623 + }, + { + "epoch": 0.03, + "grad_norm": 3.0537421060657257, + "learning_rate": 8.705933420454745e-07, + "loss": 1.2805, + "step": 624 + }, + { + "epoch": 0.03, + "grad_norm": 3.02315805498333, + "learning_rate": 8.705871911107236e-07, + "loss": 1.3664, + "step": 625 + }, + { + "epoch": 0.03, + "grad_norm": 3.0968669538104705, + "learning_rate": 8.70581028477789e-07, + "loss": 1.4156, + "step": 626 + }, + { + "epoch": 0.03, + "grad_norm": 2.958411205891844, + "learning_rate": 8.705748541468365e-07, + "loss": 1.3879, + "step": 627 + }, + { + "epoch": 0.03, + "grad_norm": 3.0644896141269213, + "learning_rate": 8.705686681180324e-07, + "loss": 1.406, + "step": 628 + }, + { + "epoch": 0.03, + "grad_norm": 2.9355330340666947, + "learning_rate": 8.705624703915431e-07, + "loss": 1.4157, + "step": 629 + }, + { + "epoch": 0.03, + "grad_norm": 3.0295134625264732, + "learning_rate": 8.705562609675357e-07, + "loss": 1.3595, + "step": 630 + } + ], + "logging_steps": 1.0, + "max_steps": 19246, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 105, + "total_flos": 164886478848000.0, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +}