lyc123456's picture
Upload 21 files
9e7b707 verified
raw
history blame contribute delete
No virus
67.8 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 34.89855072463768,
"eval_steps": 500,
"global_step": 1505,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.11594202898550725,
"grad_norm": 7.532149791717529,
"learning_rate": 4.999863832700438e-05,
"loss": 3.8116,
"num_input_tokens_seen": 106929,
"step": 5
},
{
"epoch": 0.2318840579710145,
"grad_norm": 4.141716480255127,
"learning_rate": 4.999455345634978e-05,
"loss": 3.6928,
"num_input_tokens_seen": 225964,
"step": 10
},
{
"epoch": 0.34782608695652173,
"grad_norm": 3.470097780227661,
"learning_rate": 4.9987745833016855e-05,
"loss": 3.6227,
"num_input_tokens_seen": 362264,
"step": 15
},
{
"epoch": 0.463768115942029,
"grad_norm": 3.4544646739959717,
"learning_rate": 4.9978216198586135e-05,
"loss": 3.601,
"num_input_tokens_seen": 477807,
"step": 20
},
{
"epoch": 0.5797101449275363,
"grad_norm": 3.249224901199341,
"learning_rate": 4.996596559115731e-05,
"loss": 3.539,
"num_input_tokens_seen": 588900,
"step": 25
},
{
"epoch": 0.6956521739130435,
"grad_norm": 3.395056962966919,
"learning_rate": 4.995099534523607e-05,
"loss": 3.4956,
"num_input_tokens_seen": 706077,
"step": 30
},
{
"epoch": 0.8115942028985508,
"grad_norm": 3.997875213623047,
"learning_rate": 4.9933307091588796e-05,
"loss": 3.5044,
"num_input_tokens_seen": 853504,
"step": 35
},
{
"epoch": 0.927536231884058,
"grad_norm": 3.5168681144714355,
"learning_rate": 4.991290275706486e-05,
"loss": 3.4324,
"num_input_tokens_seen": 990472,
"step": 40
},
{
"epoch": 1.0434782608695652,
"grad_norm": 7.144646167755127,
"learning_rate": 4.988978456438678e-05,
"loss": 3.2542,
"num_input_tokens_seen": 1125870,
"step": 45
},
{
"epoch": 1.1594202898550725,
"grad_norm": 3.257103681564331,
"learning_rate": 4.986395503190805e-05,
"loss": 2.9024,
"num_input_tokens_seen": 1249877,
"step": 50
},
{
"epoch": 1.2753623188405796,
"grad_norm": 3.3208603858947754,
"learning_rate": 4.983541697333881e-05,
"loss": 2.8069,
"num_input_tokens_seen": 1375193,
"step": 55
},
{
"epoch": 1.391304347826087,
"grad_norm": 4.378167629241943,
"learning_rate": 4.980417349743936e-05,
"loss": 2.75,
"num_input_tokens_seen": 1489716,
"step": 60
},
{
"epoch": 1.5072463768115942,
"grad_norm": 4.321849822998047,
"learning_rate": 4.9770228007681494e-05,
"loss": 2.7329,
"num_input_tokens_seen": 1600483,
"step": 65
},
{
"epoch": 1.6231884057971016,
"grad_norm": 3.6365067958831787,
"learning_rate": 4.973358420187776e-05,
"loss": 2.8212,
"num_input_tokens_seen": 1731315,
"step": 70
},
{
"epoch": 1.7391304347826086,
"grad_norm": 3.74035906791687,
"learning_rate": 4.9694246071778604e-05,
"loss": 2.7935,
"num_input_tokens_seen": 1858269,
"step": 75
},
{
"epoch": 1.855072463768116,
"grad_norm": 23.26426124572754,
"learning_rate": 4.9652217902637596e-05,
"loss": 2.7305,
"num_input_tokens_seen": 1984587,
"step": 80
},
{
"epoch": 1.971014492753623,
"grad_norm": 4.870578289031982,
"learning_rate": 4.9607504272744575e-05,
"loss": 2.6482,
"num_input_tokens_seen": 2109391,
"step": 85
},
{
"epoch": 2.0869565217391304,
"grad_norm": 4.419096946716309,
"learning_rate": 4.956011005292692e-05,
"loss": 2.4292,
"num_input_tokens_seen": 2246413,
"step": 90
},
{
"epoch": 2.2028985507246377,
"grad_norm": 18.216915130615234,
"learning_rate": 4.951004040601898e-05,
"loss": 2.1416,
"num_input_tokens_seen": 2386890,
"step": 95
},
{
"epoch": 2.318840579710145,
"grad_norm": 4.4814581871032715,
"learning_rate": 4.945730078629964e-05,
"loss": 2.2847,
"num_input_tokens_seen": 2522302,
"step": 100
},
{
"epoch": 2.4347826086956523,
"grad_norm": 406.9701232910156,
"learning_rate": 4.9401896938898185e-05,
"loss": 2.0944,
"num_input_tokens_seen": 2642208,
"step": 105
},
{
"epoch": 2.550724637681159,
"grad_norm": 3.5890581607818604,
"learning_rate": 4.934383489916843e-05,
"loss": 2.2862,
"num_input_tokens_seen": 2780587,
"step": 110
},
{
"epoch": 2.6666666666666665,
"grad_norm": 5.334541320800781,
"learning_rate": 4.928312099203131e-05,
"loss": 2.105,
"num_input_tokens_seen": 2885320,
"step": 115
},
{
"epoch": 2.782608695652174,
"grad_norm": 5.464664936065674,
"learning_rate": 4.921976183128585e-05,
"loss": 2.0287,
"num_input_tokens_seen": 2996923,
"step": 120
},
{
"epoch": 2.898550724637681,
"grad_norm": 4.113780975341797,
"learning_rate": 4.9153764318888706e-05,
"loss": 2.0162,
"num_input_tokens_seen": 3102391,
"step": 125
},
{
"epoch": 3.0144927536231885,
"grad_norm": 6.009971618652344,
"learning_rate": 4.908513564420231e-05,
"loss": 2.2464,
"num_input_tokens_seen": 3233443,
"step": 130
},
{
"epoch": 3.130434782608696,
"grad_norm": 10.397327423095703,
"learning_rate": 4.90138832832117e-05,
"loss": 1.6561,
"num_input_tokens_seen": 3358966,
"step": 135
},
{
"epoch": 3.246376811594203,
"grad_norm": 4.9139556884765625,
"learning_rate": 4.894001499771015e-05,
"loss": 1.6113,
"num_input_tokens_seen": 3490069,
"step": 140
},
{
"epoch": 3.36231884057971,
"grad_norm": 4.146034240722656,
"learning_rate": 4.886353883445363e-05,
"loss": 1.6235,
"num_input_tokens_seen": 3609842,
"step": 145
},
{
"epoch": 3.4782608695652173,
"grad_norm": 4.301880359649658,
"learning_rate": 4.878446312428424e-05,
"loss": 1.7873,
"num_input_tokens_seen": 3751570,
"step": 150
},
{
"epoch": 3.5942028985507246,
"grad_norm": 3.9485158920288086,
"learning_rate": 4.8702796481222714e-05,
"loss": 1.3723,
"num_input_tokens_seen": 3865303,
"step": 155
},
{
"epoch": 3.710144927536232,
"grad_norm": 4.183668613433838,
"learning_rate": 4.861854780153004e-05,
"loss": 1.6512,
"num_input_tokens_seen": 3991347,
"step": 160
},
{
"epoch": 3.8260869565217392,
"grad_norm": 5.1000471115112305,
"learning_rate": 4.853172626273841e-05,
"loss": 1.5524,
"num_input_tokens_seen": 4113654,
"step": 165
},
{
"epoch": 3.942028985507246,
"grad_norm": 4.142239570617676,
"learning_rate": 4.8442341322651385e-05,
"loss": 1.5954,
"num_input_tokens_seen": 4236348,
"step": 170
},
{
"epoch": 4.057971014492754,
"grad_norm": 3.8976669311523438,
"learning_rate": 4.83504027183137e-05,
"loss": 1.1652,
"num_input_tokens_seen": 4340378,
"step": 175
},
{
"epoch": 4.173913043478261,
"grad_norm": 5.923389911651611,
"learning_rate": 4.825592046495054e-05,
"loss": 1.1995,
"num_input_tokens_seen": 4473601,
"step": 180
},
{
"epoch": 4.2898550724637685,
"grad_norm": 4.220530033111572,
"learning_rate": 4.8158904854876555e-05,
"loss": 0.9431,
"num_input_tokens_seen": 4586911,
"step": 185
},
{
"epoch": 4.405797101449275,
"grad_norm": 5.896139144897461,
"learning_rate": 4.805936645637463e-05,
"loss": 1.1136,
"num_input_tokens_seen": 4702445,
"step": 190
},
{
"epoch": 4.521739130434782,
"grad_norm": 4.467094421386719,
"learning_rate": 4.795731611254473e-05,
"loss": 1.1509,
"num_input_tokens_seen": 4831301,
"step": 195
},
{
"epoch": 4.63768115942029,
"grad_norm": 4.232386112213135,
"learning_rate": 4.785276494012263e-05,
"loss": 0.9962,
"num_input_tokens_seen": 4941656,
"step": 200
},
{
"epoch": 4.753623188405797,
"grad_norm": 4.829892635345459,
"learning_rate": 4.7745724328269e-05,
"loss": 1.2377,
"num_input_tokens_seen": 5088437,
"step": 205
},
{
"epoch": 4.869565217391305,
"grad_norm": 4.1343913078308105,
"learning_rate": 4.763620593732867e-05,
"loss": 1.234,
"num_input_tokens_seen": 5219806,
"step": 210
},
{
"epoch": 4.9855072463768115,
"grad_norm": 4.9217729568481445,
"learning_rate": 4.752422169756048e-05,
"loss": 1.1453,
"num_input_tokens_seen": 5340222,
"step": 215
},
{
"epoch": 5.101449275362318,
"grad_norm": 4.4605865478515625,
"learning_rate": 4.740978380783765e-05,
"loss": 0.9056,
"num_input_tokens_seen": 5476315,
"step": 220
},
{
"epoch": 5.217391304347826,
"grad_norm": 4.396484375,
"learning_rate": 4.7292904734318924e-05,
"loss": 0.7349,
"num_input_tokens_seen": 5589951,
"step": 225
},
{
"epoch": 5.333333333333333,
"grad_norm": 4.053436279296875,
"learning_rate": 4.7173597209090534e-05,
"loss": 0.6968,
"num_input_tokens_seen": 5711449,
"step": 230
},
{
"epoch": 5.449275362318841,
"grad_norm": 5.303736209869385,
"learning_rate": 4.70518742287793e-05,
"loss": 0.851,
"num_input_tokens_seen": 5852650,
"step": 235
},
{
"epoch": 5.565217391304348,
"grad_norm": 3.70810866355896,
"learning_rate": 4.6927749053136866e-05,
"loss": 0.716,
"num_input_tokens_seen": 5972289,
"step": 240
},
{
"epoch": 5.681159420289855,
"grad_norm": 3.9204599857330322,
"learning_rate": 4.6801235203595195e-05,
"loss": 0.6384,
"num_input_tokens_seen": 6088707,
"step": 245
},
{
"epoch": 5.797101449275362,
"grad_norm": 4.06931209564209,
"learning_rate": 4.667234646179368e-05,
"loss": 0.7799,
"num_input_tokens_seen": 6215471,
"step": 250
},
{
"epoch": 5.913043478260869,
"grad_norm": 4.283618450164795,
"learning_rate": 4.654109686807787e-05,
"loss": 0.7923,
"num_input_tokens_seen": 6335935,
"step": 255
},
{
"epoch": 6.028985507246377,
"grad_norm": 4.719886302947998,
"learning_rate": 4.640750071996995e-05,
"loss": 0.7452,
"num_input_tokens_seen": 6463689,
"step": 260
},
{
"epoch": 6.144927536231884,
"grad_norm": 3.8415334224700928,
"learning_rate": 4.6271572570611296e-05,
"loss": 0.4085,
"num_input_tokens_seen": 6576954,
"step": 265
},
{
"epoch": 6.260869565217392,
"grad_norm": 4.19309663772583,
"learning_rate": 4.613332722717714e-05,
"loss": 0.5777,
"num_input_tokens_seen": 6714404,
"step": 270
},
{
"epoch": 6.3768115942028984,
"grad_norm": 5.686235427856445,
"learning_rate": 4.5992779749263546e-05,
"loss": 0.4718,
"num_input_tokens_seen": 6840385,
"step": 275
},
{
"epoch": 6.492753623188406,
"grad_norm": 3.2365808486938477,
"learning_rate": 4.584994544724695e-05,
"loss": 0.3723,
"num_input_tokens_seen": 6954269,
"step": 280
},
{
"epoch": 6.608695652173913,
"grad_norm": 3.530801296234131,
"learning_rate": 4.5704839880616296e-05,
"loss": 0.4453,
"num_input_tokens_seen": 7076143,
"step": 285
},
{
"epoch": 6.72463768115942,
"grad_norm": 3.2134931087493896,
"learning_rate": 4.5557478856278114e-05,
"loss": 0.5742,
"num_input_tokens_seen": 7201833,
"step": 290
},
{
"epoch": 6.840579710144928,
"grad_norm": 6.281985282897949,
"learning_rate": 4.5407878426834596e-05,
"loss": 0.5291,
"num_input_tokens_seen": 7330479,
"step": 295
},
{
"epoch": 6.956521739130435,
"grad_norm": 17.072542190551758,
"learning_rate": 4.5256054888834934e-05,
"loss": 0.4968,
"num_input_tokens_seen": 7449244,
"step": 300
},
{
"epoch": 7.072463768115942,
"grad_norm": 3.737456798553467,
"learning_rate": 4.5102024781000077e-05,
"loss": 0.421,
"num_input_tokens_seen": 7578947,
"step": 305
},
{
"epoch": 7.188405797101449,
"grad_norm": 2.6586523056030273,
"learning_rate": 4.4945804882421086e-05,
"loss": 0.2767,
"num_input_tokens_seen": 7691948,
"step": 310
},
{
"epoch": 7.304347826086957,
"grad_norm": 2.027702808380127,
"learning_rate": 4.478741221073136e-05,
"loss": 0.2922,
"num_input_tokens_seen": 7815786,
"step": 315
},
{
"epoch": 7.420289855072464,
"grad_norm": 3.4651787281036377,
"learning_rate": 4.4626864020252774e-05,
"loss": 0.2768,
"num_input_tokens_seen": 7925106,
"step": 320
},
{
"epoch": 7.536231884057971,
"grad_norm": 4.559577941894531,
"learning_rate": 4.446417780011618e-05,
"loss": 0.3281,
"num_input_tokens_seen": 8057202,
"step": 325
},
{
"epoch": 7.6521739130434785,
"grad_norm": 2.5885751247406006,
"learning_rate": 4.42993712723562e-05,
"loss": 0.3374,
"num_input_tokens_seen": 8187865,
"step": 330
},
{
"epoch": 7.768115942028985,
"grad_norm": 2.792222023010254,
"learning_rate": 4.413246238998069e-05,
"loss": 0.2491,
"num_input_tokens_seen": 8304605,
"step": 335
},
{
"epoch": 7.884057971014493,
"grad_norm": 3.610206127166748,
"learning_rate": 4.3963469335015085e-05,
"loss": 0.3893,
"num_input_tokens_seen": 8437319,
"step": 340
},
{
"epoch": 8.0,
"grad_norm": 4.31643533706665,
"learning_rate": 4.379241051652174e-05,
"loss": 0.3761,
"num_input_tokens_seen": 8573080,
"step": 345
},
{
"epoch": 8.115942028985508,
"grad_norm": 2.2834160327911377,
"learning_rate": 4.361930456859456e-05,
"loss": 0.236,
"num_input_tokens_seen": 8707741,
"step": 350
},
{
"epoch": 8.231884057971014,
"grad_norm": 2.6929121017456055,
"learning_rate": 4.34441703483291e-05,
"loss": 0.1584,
"num_input_tokens_seen": 8825774,
"step": 355
},
{
"epoch": 8.347826086956522,
"grad_norm": 3.8095011711120605,
"learning_rate": 4.326702693376844e-05,
"loss": 0.1481,
"num_input_tokens_seen": 8932249,
"step": 360
},
{
"epoch": 8.46376811594203,
"grad_norm": 2.6493489742279053,
"learning_rate": 4.308789362182492e-05,
"loss": 0.1743,
"num_input_tokens_seen": 9051548,
"step": 365
},
{
"epoch": 8.579710144927537,
"grad_norm": 30.796459197998047,
"learning_rate": 4.290678992617798e-05,
"loss": 0.3162,
"num_input_tokens_seen": 9197232,
"step": 370
},
{
"epoch": 8.695652173913043,
"grad_norm": 3.3164985179901123,
"learning_rate": 4.272373557514858e-05,
"loss": 0.2235,
"num_input_tokens_seen": 9317650,
"step": 375
},
{
"epoch": 8.81159420289855,
"grad_norm": 3.1515417098999023,
"learning_rate": 4.2538750509550054e-05,
"loss": 0.2504,
"num_input_tokens_seen": 9450765,
"step": 380
},
{
"epoch": 8.927536231884059,
"grad_norm": 3.3926901817321777,
"learning_rate": 4.235185488051585e-05,
"loss": 0.2136,
"num_input_tokens_seen": 9582961,
"step": 385
},
{
"epoch": 9.043478260869565,
"grad_norm": 4.670753002166748,
"learning_rate": 4.216306904730447e-05,
"loss": 0.1047,
"num_input_tokens_seen": 9678616,
"step": 390
},
{
"epoch": 9.159420289855072,
"grad_norm": 2.166652202606201,
"learning_rate": 4.1972413575081595e-05,
"loss": 0.1015,
"num_input_tokens_seen": 9788512,
"step": 395
},
{
"epoch": 9.27536231884058,
"grad_norm": 2.1161272525787354,
"learning_rate": 4.177990923267986e-05,
"loss": 0.1505,
"num_input_tokens_seen": 9916229,
"step": 400
},
{
"epoch": 9.391304347826088,
"grad_norm": 2.378105401992798,
"learning_rate": 4.158557699033644e-05,
"loss": 0.1135,
"num_input_tokens_seen": 10042697,
"step": 405
},
{
"epoch": 9.507246376811594,
"grad_norm": 2.5567331314086914,
"learning_rate": 4.138943801740865e-05,
"loss": 0.1832,
"num_input_tokens_seen": 10171849,
"step": 410
},
{
"epoch": 9.623188405797102,
"grad_norm": 2.022610902786255,
"learning_rate": 4.119151368006793e-05,
"loss": 0.1178,
"num_input_tokens_seen": 10281924,
"step": 415
},
{
"epoch": 9.73913043478261,
"grad_norm": 2.5578079223632812,
"learning_rate": 4.099182553897229e-05,
"loss": 0.1426,
"num_input_tokens_seen": 10418758,
"step": 420
},
{
"epoch": 9.855072463768115,
"grad_norm": 2.7287228107452393,
"learning_rate": 4.079039534691767e-05,
"loss": 0.1603,
"num_input_tokens_seen": 10558322,
"step": 425
},
{
"epoch": 9.971014492753623,
"grad_norm": 2.361532688140869,
"learning_rate": 4.058724504646834e-05,
"loss": 0.1548,
"num_input_tokens_seen": 10679536,
"step": 430
},
{
"epoch": 10.08695652173913,
"grad_norm": 1.8757002353668213,
"learning_rate": 4.0382396767566536e-05,
"loss": 0.1407,
"num_input_tokens_seen": 10821076,
"step": 435
},
{
"epoch": 10.202898550724637,
"grad_norm": 2.352725028991699,
"learning_rate": 4.017587282512181e-05,
"loss": 0.0791,
"num_input_tokens_seen": 10949771,
"step": 440
},
{
"epoch": 10.318840579710145,
"grad_norm": 1.7948939800262451,
"learning_rate": 3.9967695716580224e-05,
"loss": 0.0722,
"num_input_tokens_seen": 11072044,
"step": 445
},
{
"epoch": 10.434782608695652,
"grad_norm": 1.954727292060852,
"learning_rate": 3.975788811947351e-05,
"loss": 0.0655,
"num_input_tokens_seen": 11182627,
"step": 450
},
{
"epoch": 10.55072463768116,
"grad_norm": 2.143941640853882,
"learning_rate": 3.954647288894883e-05,
"loss": 0.0723,
"num_input_tokens_seen": 11303028,
"step": 455
},
{
"epoch": 10.666666666666666,
"grad_norm": 2.0527164936065674,
"learning_rate": 3.933347305527898e-05,
"loss": 0.0655,
"num_input_tokens_seen": 11415868,
"step": 460
},
{
"epoch": 10.782608695652174,
"grad_norm": 1.6390535831451416,
"learning_rate": 3.911891182135371e-05,
"loss": 0.1534,
"num_input_tokens_seen": 11555653,
"step": 465
},
{
"epoch": 10.898550724637682,
"grad_norm": 2.3848719596862793,
"learning_rate": 3.8902812560152066e-05,
"loss": 0.0947,
"num_input_tokens_seen": 11681065,
"step": 470
},
{
"epoch": 11.014492753623188,
"grad_norm": 2.2094757556915283,
"learning_rate": 3.868519881219631e-05,
"loss": 0.0868,
"num_input_tokens_seen": 11809957,
"step": 475
},
{
"epoch": 11.130434782608695,
"grad_norm": 4.137216567993164,
"learning_rate": 3.846609428298757e-05,
"loss": 0.0467,
"num_input_tokens_seen": 11937881,
"step": 480
},
{
"epoch": 11.246376811594203,
"grad_norm": 1.6658189296722412,
"learning_rate": 3.824552284042351e-05,
"loss": 0.0521,
"num_input_tokens_seen": 12048905,
"step": 485
},
{
"epoch": 11.36231884057971,
"grad_norm": 1.5732171535491943,
"learning_rate": 3.8023508512198256e-05,
"loss": 0.051,
"num_input_tokens_seen": 12185453,
"step": 490
},
{
"epoch": 11.478260869565217,
"grad_norm": 1.8459701538085938,
"learning_rate": 3.780007548318507e-05,
"loss": 0.0753,
"num_input_tokens_seen": 12310911,
"step": 495
},
{
"epoch": 11.594202898550725,
"grad_norm": 1.4724109172821045,
"learning_rate": 3.7575248092801686e-05,
"loss": 0.0601,
"num_input_tokens_seen": 12439708,
"step": 500
},
{
"epoch": 11.710144927536232,
"grad_norm": 2.4690322875976562,
"learning_rate": 3.734905083235901e-05,
"loss": 0.0533,
"num_input_tokens_seen": 12554467,
"step": 505
},
{
"epoch": 11.826086956521738,
"grad_norm": 2.369218111038208,
"learning_rate": 3.712150834239313e-05,
"loss": 0.064,
"num_input_tokens_seen": 12682329,
"step": 510
},
{
"epoch": 11.942028985507246,
"grad_norm": 1.6901100873947144,
"learning_rate": 3.689264540998116e-05,
"loss": 0.0755,
"num_input_tokens_seen": 12800852,
"step": 515
},
{
"epoch": 12.057971014492754,
"grad_norm": 1.303114414215088,
"learning_rate": 3.66624869660411e-05,
"loss": 0.0553,
"num_input_tokens_seen": 12917527,
"step": 520
},
{
"epoch": 12.173913043478262,
"grad_norm": 1.1986353397369385,
"learning_rate": 3.6431058082615964e-05,
"loss": 0.0355,
"num_input_tokens_seen": 13044774,
"step": 525
},
{
"epoch": 12.289855072463768,
"grad_norm": 1.5653026103973389,
"learning_rate": 3.619838397014263e-05,
"loss": 0.0413,
"num_input_tokens_seen": 13175692,
"step": 530
},
{
"epoch": 12.405797101449275,
"grad_norm": 1.0767664909362793,
"learning_rate": 3.5964489974705553e-05,
"loss": 0.0596,
"num_input_tokens_seen": 13293164,
"step": 535
},
{
"epoch": 12.521739130434783,
"grad_norm": 1.6005312204360962,
"learning_rate": 3.572940157527572e-05,
"loss": 0.0479,
"num_input_tokens_seen": 13417894,
"step": 540
},
{
"epoch": 12.63768115942029,
"grad_norm": 1.627121925354004,
"learning_rate": 3.549314438093515e-05,
"loss": 0.047,
"num_input_tokens_seen": 13551913,
"step": 545
},
{
"epoch": 12.753623188405797,
"grad_norm": 2.239276647567749,
"learning_rate": 3.525574412808717e-05,
"loss": 0.0492,
"num_input_tokens_seen": 13675309,
"step": 550
},
{
"epoch": 12.869565217391305,
"grad_norm": 1.5702998638153076,
"learning_rate": 3.501722667765286e-05,
"loss": 0.0471,
"num_input_tokens_seen": 13797691,
"step": 555
},
{
"epoch": 12.985507246376812,
"grad_norm": 1.8216972351074219,
"learning_rate": 3.47776180122539e-05,
"loss": 0.1041,
"num_input_tokens_seen": 13919770,
"step": 560
},
{
"epoch": 13.101449275362318,
"grad_norm": 0.9026144742965698,
"learning_rate": 3.453694423338225e-05,
"loss": 0.0282,
"num_input_tokens_seen": 14037673,
"step": 565
},
{
"epoch": 13.217391304347826,
"grad_norm": 1.4504765272140503,
"learning_rate": 3.4295231558556715e-05,
"loss": 0.0272,
"num_input_tokens_seen": 14167090,
"step": 570
},
{
"epoch": 13.333333333333334,
"grad_norm": 1.4278969764709473,
"learning_rate": 3.4052506318467084e-05,
"loss": 0.0342,
"num_input_tokens_seen": 14311710,
"step": 575
},
{
"epoch": 13.44927536231884,
"grad_norm": 1.1284997463226318,
"learning_rate": 3.3808794954105716e-05,
"loss": 0.0855,
"num_input_tokens_seen": 14404322,
"step": 580
},
{
"epoch": 13.565217391304348,
"grad_norm": 1.4915614128112793,
"learning_rate": 3.356412401388732e-05,
"loss": 0.0378,
"num_input_tokens_seen": 14530794,
"step": 585
},
{
"epoch": 13.681159420289855,
"grad_norm": 1.372157096862793,
"learning_rate": 3.3318520150756846e-05,
"loss": 0.0457,
"num_input_tokens_seen": 14637342,
"step": 590
},
{
"epoch": 13.797101449275363,
"grad_norm": 1.6492116451263428,
"learning_rate": 3.307201011928616e-05,
"loss": 0.0453,
"num_input_tokens_seen": 14787534,
"step": 595
},
{
"epoch": 13.91304347826087,
"grad_norm": 1.3583859205245972,
"learning_rate": 3.282462077275947e-05,
"loss": 0.0378,
"num_input_tokens_seen": 14909175,
"step": 600
},
{
"epoch": 14.028985507246377,
"grad_norm": 1.0751795768737793,
"learning_rate": 3.257637906024822e-05,
"loss": 0.0296,
"num_input_tokens_seen": 15030530,
"step": 605
},
{
"epoch": 14.144927536231885,
"grad_norm": 1.474602222442627,
"learning_rate": 3.2327312023675287e-05,
"loss": 0.0216,
"num_input_tokens_seen": 15148359,
"step": 610
},
{
"epoch": 14.26086956521739,
"grad_norm": 1.0749961137771606,
"learning_rate": 3.2077446794869295e-05,
"loss": 0.0299,
"num_input_tokens_seen": 15280749,
"step": 615
},
{
"epoch": 14.376811594202898,
"grad_norm": 1.4042794704437256,
"learning_rate": 3.1826810592609036e-05,
"loss": 0.0247,
"num_input_tokens_seen": 15397167,
"step": 620
},
{
"epoch": 14.492753623188406,
"grad_norm": 1.2280118465423584,
"learning_rate": 3.157543071965835e-05,
"loss": 0.0455,
"num_input_tokens_seen": 15522794,
"step": 625
},
{
"epoch": 14.608695652173914,
"grad_norm": 1.2819784879684448,
"learning_rate": 3.132333455979202e-05,
"loss": 0.0262,
"num_input_tokens_seen": 15637987,
"step": 630
},
{
"epoch": 14.72463768115942,
"grad_norm": 1.2691748142242432,
"learning_rate": 3.107054957481271e-05,
"loss": 0.0281,
"num_input_tokens_seen": 15773163,
"step": 635
},
{
"epoch": 14.840579710144928,
"grad_norm": 1.2752504348754883,
"learning_rate": 3.081710330155942e-05,
"loss": 0.0294,
"num_input_tokens_seen": 15892659,
"step": 640
},
{
"epoch": 14.956521739130435,
"grad_norm": 1.3479197025299072,
"learning_rate": 3.056302334890786e-05,
"loss": 0.0291,
"num_input_tokens_seen": 16024576,
"step": 645
},
{
"epoch": 15.072463768115941,
"grad_norm": 1.3151382207870483,
"learning_rate": 3.030833739476285e-05,
"loss": 0.0216,
"num_input_tokens_seen": 16151987,
"step": 650
},
{
"epoch": 15.18840579710145,
"grad_norm": 2.3882877826690674,
"learning_rate": 3.0053073183043256e-05,
"loss": 0.0218,
"num_input_tokens_seen": 16278639,
"step": 655
},
{
"epoch": 15.304347826086957,
"grad_norm": 0.9794278144836426,
"learning_rate": 2.979725852065981e-05,
"loss": 0.0283,
"num_input_tokens_seen": 16414743,
"step": 660
},
{
"epoch": 15.420289855072463,
"grad_norm": 0.8964869976043701,
"learning_rate": 2.954092127448591e-05,
"loss": 0.0259,
"num_input_tokens_seen": 16529298,
"step": 665
},
{
"epoch": 15.53623188405797,
"grad_norm": 1.1441810131072998,
"learning_rate": 2.9284089368322045e-05,
"loss": 0.0716,
"num_input_tokens_seen": 16655909,
"step": 670
},
{
"epoch": 15.652173913043478,
"grad_norm": 1.0959213972091675,
"learning_rate": 2.9026790779853874e-05,
"loss": 0.025,
"num_input_tokens_seen": 16798263,
"step": 675
},
{
"epoch": 15.768115942028986,
"grad_norm": 1.0119343996047974,
"learning_rate": 2.876905353760459e-05,
"loss": 0.0218,
"num_input_tokens_seen": 16916827,
"step": 680
},
{
"epoch": 15.884057971014492,
"grad_norm": 1.1373978853225708,
"learning_rate": 2.8510905717881614e-05,
"loss": 0.0231,
"num_input_tokens_seen": 17040247,
"step": 685
},
{
"epoch": 16.0,
"grad_norm": 1.2512497901916504,
"learning_rate": 2.8252375441718137e-05,
"loss": 0.0228,
"num_input_tokens_seen": 17146160,
"step": 690
},
{
"epoch": 16.115942028985508,
"grad_norm": 0.7410117387771606,
"learning_rate": 2.7993490871809808e-05,
"loss": 0.029,
"num_input_tokens_seen": 17284643,
"step": 695
},
{
"epoch": 16.231884057971016,
"grad_norm": 1.0934263467788696,
"learning_rate": 2.7734280209446865e-05,
"loss": 0.0199,
"num_input_tokens_seen": 17426644,
"step": 700
},
{
"epoch": 16.347826086956523,
"grad_norm": 1.0034395456314087,
"learning_rate": 2.7474771691442018e-05,
"loss": 0.0259,
"num_input_tokens_seen": 17541812,
"step": 705
},
{
"epoch": 16.463768115942027,
"grad_norm": 1.4287781715393066,
"learning_rate": 2.721499358705458e-05,
"loss": 0.021,
"num_input_tokens_seen": 17667755,
"step": 710
},
{
"epoch": 16.579710144927535,
"grad_norm": 1.0989606380462646,
"learning_rate": 2.6954974194910888e-05,
"loss": 0.0199,
"num_input_tokens_seen": 17788162,
"step": 715
},
{
"epoch": 16.695652173913043,
"grad_norm": 0.9687130451202393,
"learning_rate": 2.6694741839921732e-05,
"loss": 0.0189,
"num_input_tokens_seen": 17911718,
"step": 720
},
{
"epoch": 16.81159420289855,
"grad_norm": 1.143617033958435,
"learning_rate": 2.6434324870196748e-05,
"loss": 0.0169,
"num_input_tokens_seen": 18018729,
"step": 725
},
{
"epoch": 16.92753623188406,
"grad_norm": 1.1395140886306763,
"learning_rate": 2.617375165395634e-05,
"loss": 0.0209,
"num_input_tokens_seen": 18139681,
"step": 730
},
{
"epoch": 17.043478260869566,
"grad_norm": 0.881986677646637,
"learning_rate": 2.5913050576441477e-05,
"loss": 0.0201,
"num_input_tokens_seen": 18278544,
"step": 735
},
{
"epoch": 17.159420289855074,
"grad_norm": 0.8654409050941467,
"learning_rate": 2.5652250036821523e-05,
"loss": 0.017,
"num_input_tokens_seen": 18396700,
"step": 740
},
{
"epoch": 17.27536231884058,
"grad_norm": 0.9699842929840088,
"learning_rate": 2.5391378445100644e-05,
"loss": 0.0187,
"num_input_tokens_seen": 18506229,
"step": 745
},
{
"epoch": 17.391304347826086,
"grad_norm": 0.8799194693565369,
"learning_rate": 2.5130464219022992e-05,
"loss": 0.0242,
"num_input_tokens_seen": 18621580,
"step": 750
},
{
"epoch": 17.507246376811594,
"grad_norm": 0.9715821146965027,
"learning_rate": 2.486953578097702e-05,
"loss": 0.0153,
"num_input_tokens_seen": 18748382,
"step": 755
},
{
"epoch": 17.6231884057971,
"grad_norm": 0.8819458484649658,
"learning_rate": 2.4608621554899362e-05,
"loss": 0.0182,
"num_input_tokens_seen": 18884730,
"step": 760
},
{
"epoch": 17.73913043478261,
"grad_norm": 0.8835431933403015,
"learning_rate": 2.4347749963178486e-05,
"loss": 0.0143,
"num_input_tokens_seen": 19003589,
"step": 765
},
{
"epoch": 17.855072463768117,
"grad_norm": 0.780754566192627,
"learning_rate": 2.4086949423558526e-05,
"loss": 0.0164,
"num_input_tokens_seen": 19136411,
"step": 770
},
{
"epoch": 17.971014492753625,
"grad_norm": 0.7591371536254883,
"learning_rate": 2.3826248346043663e-05,
"loss": 0.0157,
"num_input_tokens_seen": 19260436,
"step": 775
},
{
"epoch": 18.08695652173913,
"grad_norm": 0.673797070980072,
"learning_rate": 2.356567512980326e-05,
"loss": 0.0304,
"num_input_tokens_seen": 19388733,
"step": 780
},
{
"epoch": 18.202898550724637,
"grad_norm": 0.4008718729019165,
"learning_rate": 2.3305258160078274e-05,
"loss": 0.009,
"num_input_tokens_seen": 19531204,
"step": 785
},
{
"epoch": 18.318840579710145,
"grad_norm": 0.6676005125045776,
"learning_rate": 2.3045025805089118e-05,
"loss": 0.0105,
"num_input_tokens_seen": 19624608,
"step": 790
},
{
"epoch": 18.434782608695652,
"grad_norm": 0.6956990957260132,
"learning_rate": 2.278500641294543e-05,
"loss": 0.0104,
"num_input_tokens_seen": 19751062,
"step": 795
},
{
"epoch": 18.55072463768116,
"grad_norm": 0.80479896068573,
"learning_rate": 2.252522830855798e-05,
"loss": 0.0103,
"num_input_tokens_seen": 19879837,
"step": 800
},
{
"epoch": 18.666666666666668,
"grad_norm": 0.7206840515136719,
"learning_rate": 2.2265719790553147e-05,
"loss": 0.0107,
"num_input_tokens_seen": 20019385,
"step": 805
},
{
"epoch": 18.782608695652176,
"grad_norm": 0.6994977593421936,
"learning_rate": 2.2006509128190195e-05,
"loss": 0.0269,
"num_input_tokens_seen": 20138003,
"step": 810
},
{
"epoch": 18.89855072463768,
"grad_norm": 0.5642988681793213,
"learning_rate": 2.174762455828187e-05,
"loss": 0.0086,
"num_input_tokens_seen": 20260523,
"step": 815
},
{
"epoch": 19.014492753623188,
"grad_norm": 0.5547834038734436,
"learning_rate": 2.1489094282118395e-05,
"loss": 0.0133,
"num_input_tokens_seen": 20375322,
"step": 820
},
{
"epoch": 19.130434782608695,
"grad_norm": 0.48678871989250183,
"learning_rate": 2.123094646239541e-05,
"loss": 0.0114,
"num_input_tokens_seen": 20477407,
"step": 825
},
{
"epoch": 19.246376811594203,
"grad_norm": 0.4791460633277893,
"learning_rate": 2.0973209220146135e-05,
"loss": 0.007,
"num_input_tokens_seen": 20605728,
"step": 830
},
{
"epoch": 19.36231884057971,
"grad_norm": 1.1198338270187378,
"learning_rate": 2.0715910631677968e-05,
"loss": 0.0088,
"num_input_tokens_seen": 20725799,
"step": 835
},
{
"epoch": 19.47826086956522,
"grad_norm": 0.6645247936248779,
"learning_rate": 2.0459078725514092e-05,
"loss": 0.007,
"num_input_tokens_seen": 20865534,
"step": 840
},
{
"epoch": 19.594202898550726,
"grad_norm": 0.5324479341506958,
"learning_rate": 2.020274147934019e-05,
"loss": 0.0059,
"num_input_tokens_seen": 20977913,
"step": 845
},
{
"epoch": 19.71014492753623,
"grad_norm": 0.6183504462242126,
"learning_rate": 1.9946926816956743e-05,
"loss": 0.0069,
"num_input_tokens_seen": 21102848,
"step": 850
},
{
"epoch": 19.82608695652174,
"grad_norm": 0.6665703058242798,
"learning_rate": 1.9691662605237166e-05,
"loss": 0.008,
"num_input_tokens_seen": 21243679,
"step": 855
},
{
"epoch": 19.942028985507246,
"grad_norm": 0.3298584222793579,
"learning_rate": 1.9436976651092144e-05,
"loss": 0.0127,
"num_input_tokens_seen": 21364202,
"step": 860
},
{
"epoch": 20.057971014492754,
"grad_norm": 0.2818591296672821,
"learning_rate": 1.9182896698440584e-05,
"loss": 0.0059,
"num_input_tokens_seen": 21496089,
"step": 865
},
{
"epoch": 20.17391304347826,
"grad_norm": 0.6906440258026123,
"learning_rate": 1.89294504251873e-05,
"loss": 0.0046,
"num_input_tokens_seen": 21603193,
"step": 870
},
{
"epoch": 20.28985507246377,
"grad_norm": 0.33482542634010315,
"learning_rate": 1.867666544020798e-05,
"loss": 0.0058,
"num_input_tokens_seen": 21742062,
"step": 875
},
{
"epoch": 20.405797101449274,
"grad_norm": 2.443847417831421,
"learning_rate": 1.8424569280341653e-05,
"loss": 0.0082,
"num_input_tokens_seen": 21869307,
"step": 880
},
{
"epoch": 20.52173913043478,
"grad_norm": 0.43886587023735046,
"learning_rate": 1.817318940739098e-05,
"loss": 0.0148,
"num_input_tokens_seen": 21992573,
"step": 885
},
{
"epoch": 20.63768115942029,
"grad_norm": 0.93570876121521,
"learning_rate": 1.7922553205130707e-05,
"loss": 0.0064,
"num_input_tokens_seen": 22101845,
"step": 890
},
{
"epoch": 20.753623188405797,
"grad_norm": 1176.9595947265625,
"learning_rate": 1.767268797632472e-05,
"loss": 0.008,
"num_input_tokens_seen": 22230253,
"step": 895
},
{
"epoch": 20.869565217391305,
"grad_norm": 0.35642215609550476,
"learning_rate": 1.7423620939751788e-05,
"loss": 0.0053,
"num_input_tokens_seen": 22373454,
"step": 900
},
{
"epoch": 20.985507246376812,
"grad_norm": 0.39736178517341614,
"learning_rate": 1.7175379227240523e-05,
"loss": 0.0054,
"num_input_tokens_seen": 22493123,
"step": 905
},
{
"epoch": 21.10144927536232,
"grad_norm": 0.5092463493347168,
"learning_rate": 1.692798988071385e-05,
"loss": 0.0044,
"num_input_tokens_seen": 22629005,
"step": 910
},
{
"epoch": 21.217391304347824,
"grad_norm": 0.26361697912216187,
"learning_rate": 1.6681479849243153e-05,
"loss": 0.0043,
"num_input_tokens_seen": 22752358,
"step": 915
},
{
"epoch": 21.333333333333332,
"grad_norm": 0.19933666288852692,
"learning_rate": 1.6435875986112685e-05,
"loss": 0.0035,
"num_input_tokens_seen": 22880349,
"step": 920
},
{
"epoch": 21.44927536231884,
"grad_norm": 0.22622954845428467,
"learning_rate": 1.6191205045894283e-05,
"loss": 0.0044,
"num_input_tokens_seen": 22987343,
"step": 925
},
{
"epoch": 21.565217391304348,
"grad_norm": 0.30199098587036133,
"learning_rate": 1.594749368153292e-05,
"loss": 0.0178,
"num_input_tokens_seen": 23113462,
"step": 930
},
{
"epoch": 21.681159420289855,
"grad_norm": 0.9627483487129211,
"learning_rate": 1.570476844144329e-05,
"loss": 0.0089,
"num_input_tokens_seen": 23221714,
"step": 935
},
{
"epoch": 21.797101449275363,
"grad_norm": 0.27791452407836914,
"learning_rate": 1.546305576661776e-05,
"loss": 0.004,
"num_input_tokens_seen": 23368857,
"step": 940
},
{
"epoch": 21.91304347826087,
"grad_norm": 0.3269965648651123,
"learning_rate": 1.5222381987746104e-05,
"loss": 0.004,
"num_input_tokens_seen": 23494483,
"step": 945
},
{
"epoch": 22.028985507246375,
"grad_norm": 0.15966826677322388,
"learning_rate": 1.4982773322347144e-05,
"loss": 0.0034,
"num_input_tokens_seen": 23605463,
"step": 950
},
{
"epoch": 22.144927536231883,
"grad_norm": 0.3009255826473236,
"learning_rate": 1.4744255871912823e-05,
"loss": 0.0066,
"num_input_tokens_seen": 23715776,
"step": 955
},
{
"epoch": 22.26086956521739,
"grad_norm": 0.4215935170650482,
"learning_rate": 1.4506855619064846e-05,
"loss": 0.0034,
"num_input_tokens_seen": 23841669,
"step": 960
},
{
"epoch": 22.3768115942029,
"grad_norm": 0.20214155316352844,
"learning_rate": 1.4270598424724292e-05,
"loss": 0.0032,
"num_input_tokens_seen": 23960567,
"step": 965
},
{
"epoch": 22.492753623188406,
"grad_norm": 7.0683207511901855,
"learning_rate": 1.4035510025294462e-05,
"loss": 0.0124,
"num_input_tokens_seen": 24074628,
"step": 970
},
{
"epoch": 22.608695652173914,
"grad_norm": 0.20178793370723724,
"learning_rate": 1.3801616029857378e-05,
"loss": 0.0027,
"num_input_tokens_seen": 24214324,
"step": 975
},
{
"epoch": 22.72463768115942,
"grad_norm": 1.3855236768722534,
"learning_rate": 1.3568941917384036e-05,
"loss": 0.0037,
"num_input_tokens_seen": 24326727,
"step": 980
},
{
"epoch": 22.840579710144926,
"grad_norm": 0.18420317769050598,
"learning_rate": 1.3337513033958904e-05,
"loss": 0.0029,
"num_input_tokens_seen": 24456961,
"step": 985
},
{
"epoch": 22.956521739130434,
"grad_norm": 0.15907694399356842,
"learning_rate": 1.310735459001884e-05,
"loss": 0.0035,
"num_input_tokens_seen": 24606652,
"step": 990
},
{
"epoch": 23.07246376811594,
"grad_norm": 0.2548115849494934,
"learning_rate": 1.2878491657606872e-05,
"loss": 0.002,
"num_input_tokens_seen": 24710410,
"step": 995
},
{
"epoch": 23.18840579710145,
"grad_norm": 0.36587971448898315,
"learning_rate": 1.2650949167640993e-05,
"loss": 0.0023,
"num_input_tokens_seen": 24831908,
"step": 1000
},
{
"epoch": 23.304347826086957,
"grad_norm": 0.13662408292293549,
"learning_rate": 1.2424751907198312e-05,
"loss": 0.0031,
"num_input_tokens_seen": 24951342,
"step": 1005
},
{
"epoch": 23.420289855072465,
"grad_norm": 0.19979843497276306,
"learning_rate": 1.2199924516814939e-05,
"loss": 0.0027,
"num_input_tokens_seen": 25088309,
"step": 1010
},
{
"epoch": 23.536231884057973,
"grad_norm": 0.14170995354652405,
"learning_rate": 1.1976491487801748e-05,
"loss": 0.0124,
"num_input_tokens_seen": 25216080,
"step": 1015
},
{
"epoch": 23.652173913043477,
"grad_norm": 0.06863216310739517,
"learning_rate": 1.1754477159576499e-05,
"loss": 0.0023,
"num_input_tokens_seen": 25326581,
"step": 1020
},
{
"epoch": 23.768115942028984,
"grad_norm": 0.25133436918258667,
"learning_rate": 1.1533905717012428e-05,
"loss": 0.0027,
"num_input_tokens_seen": 25477500,
"step": 1025
},
{
"epoch": 23.884057971014492,
"grad_norm": 0.28348398208618164,
"learning_rate": 1.1314801187803686e-05,
"loss": 0.0041,
"num_input_tokens_seen": 25601354,
"step": 1030
},
{
"epoch": 24.0,
"grad_norm": 0.5024954676628113,
"learning_rate": 1.1097187439847939e-05,
"loss": 0.0021,
"num_input_tokens_seen": 25719240,
"step": 1035
},
{
"epoch": 24.115942028985508,
"grad_norm": 0.1774568408727646,
"learning_rate": 1.088108817864629e-05,
"loss": 0.0039,
"num_input_tokens_seen": 25834910,
"step": 1040
},
{
"epoch": 24.231884057971016,
"grad_norm": 0.08105342090129852,
"learning_rate": 1.0666526944721016e-05,
"loss": 0.0025,
"num_input_tokens_seen": 25974530,
"step": 1045
},
{
"epoch": 24.347826086956523,
"grad_norm": 0.13048779964447021,
"learning_rate": 1.0453527111051184e-05,
"loss": 0.002,
"num_input_tokens_seen": 26104464,
"step": 1050
},
{
"epoch": 24.463768115942027,
"grad_norm": 0.10774020105600357,
"learning_rate": 1.0242111880526495e-05,
"loss": 0.0024,
"num_input_tokens_seen": 26251334,
"step": 1055
},
{
"epoch": 24.579710144927535,
"grad_norm": 0.7494776248931885,
"learning_rate": 1.003230428341979e-05,
"loss": 0.0031,
"num_input_tokens_seen": 26366561,
"step": 1060
},
{
"epoch": 24.695652173913043,
"grad_norm": 0.3580308258533478,
"learning_rate": 9.824127174878195e-06,
"loss": 0.0022,
"num_input_tokens_seen": 26486437,
"step": 1065
},
{
"epoch": 24.81159420289855,
"grad_norm": 0.1473228931427002,
"learning_rate": 9.617603232433475e-06,
"loss": 0.0022,
"num_input_tokens_seen": 26601526,
"step": 1070
},
{
"epoch": 24.92753623188406,
"grad_norm": 0.11716706305742264,
"learning_rate": 9.412754953531663e-06,
"loss": 0.0109,
"num_input_tokens_seen": 26727922,
"step": 1075
},
{
"epoch": 25.043478260869566,
"grad_norm": 0.12043190747499466,
"learning_rate": 9.209604653082326e-06,
"loss": 0.0019,
"num_input_tokens_seen": 26835621,
"step": 1080
},
{
"epoch": 25.159420289855074,
"grad_norm": 0.1277165412902832,
"learning_rate": 9.008174461027724e-06,
"loss": 0.0016,
"num_input_tokens_seen": 26955101,
"step": 1085
},
{
"epoch": 25.27536231884058,
"grad_norm": 0.08892516791820526,
"learning_rate": 8.808486319932083e-06,
"loss": 0.002,
"num_input_tokens_seen": 27077833,
"step": 1090
},
{
"epoch": 25.391304347826086,
"grad_norm": 0.30754807591438293,
"learning_rate": 8.610561982591357e-06,
"loss": 0.0018,
"num_input_tokens_seen": 27192758,
"step": 1095
},
{
"epoch": 25.507246376811594,
"grad_norm": 0.7194050550460815,
"learning_rate": 8.414423009663563e-06,
"loss": 0.0028,
"num_input_tokens_seen": 27324970,
"step": 1100
},
{
"epoch": 25.6231884057971,
"grad_norm": 4777.61328125,
"learning_rate": 8.220090767320137e-06,
"loss": 0.0021,
"num_input_tokens_seen": 27477531,
"step": 1105
},
{
"epoch": 25.73913043478261,
"grad_norm": 2.280327081680298,
"learning_rate": 8.027586424918412e-06,
"loss": 0.0057,
"num_input_tokens_seen": 27592035,
"step": 1110
},
{
"epoch": 25.855072463768117,
"grad_norm": 0.13882993161678314,
"learning_rate": 7.836930952695533e-06,
"loss": 0.0067,
"num_input_tokens_seen": 27712377,
"step": 1115
},
{
"epoch": 25.971014492753625,
"grad_norm": 0.20987676084041595,
"learning_rate": 7.648145119484153e-06,
"loss": 0.002,
"num_input_tokens_seen": 27834613,
"step": 1120
},
{
"epoch": 26.08695652173913,
"grad_norm": 0.09795770049095154,
"learning_rate": 7.461249490449954e-06,
"loss": 0.0021,
"num_input_tokens_seen": 27966996,
"step": 1125
},
{
"epoch": 26.202898550724637,
"grad_norm": 0.14506971836090088,
"learning_rate": 7.276264424851423e-06,
"loss": 0.002,
"num_input_tokens_seen": 28093538,
"step": 1130
},
{
"epoch": 26.318840579710145,
"grad_norm": 0.08091314136981964,
"learning_rate": 7.0932100738220265e-06,
"loss": 0.0017,
"num_input_tokens_seen": 28215579,
"step": 1135
},
{
"epoch": 26.434782608695652,
"grad_norm": 0.22550061345100403,
"learning_rate": 6.912106378175098e-06,
"loss": 0.0014,
"num_input_tokens_seen": 28344144,
"step": 1140
},
{
"epoch": 26.55072463768116,
"grad_norm": 0.23987355828285217,
"learning_rate": 6.732973066231563e-06,
"loss": 0.0022,
"num_input_tokens_seen": 28478650,
"step": 1145
},
{
"epoch": 26.666666666666668,
"grad_norm": 0.1993756741285324,
"learning_rate": 6.555829651670911e-06,
"loss": 0.0023,
"num_input_tokens_seen": 28593004,
"step": 1150
},
{
"epoch": 26.782608695652176,
"grad_norm": 0.7184757590293884,
"learning_rate": 6.380695431405456e-06,
"loss": 0.0028,
"num_input_tokens_seen": 28707392,
"step": 1155
},
{
"epoch": 26.89855072463768,
"grad_norm": 0.06247011199593544,
"learning_rate": 6.207589483478266e-06,
"loss": 0.006,
"num_input_tokens_seen": 28834902,
"step": 1160
},
{
"epoch": 27.014492753623188,
"grad_norm": 0.11046591401100159,
"learning_rate": 6.0365306649849214e-06,
"loss": 0.0045,
"num_input_tokens_seen": 28948812,
"step": 1165
},
{
"epoch": 27.130434782608695,
"grad_norm": 0.12309098988771439,
"learning_rate": 5.867537610019317e-06,
"loss": 0.0019,
"num_input_tokens_seen": 29078309,
"step": 1170
},
{
"epoch": 27.246376811594203,
"grad_norm": 0.11428932845592499,
"learning_rate": 5.700628727643806e-06,
"loss": 0.002,
"num_input_tokens_seen": 29211503,
"step": 1175
},
{
"epoch": 27.36231884057971,
"grad_norm": 0.1093268170952797,
"learning_rate": 5.53582219988382e-06,
"loss": 0.0019,
"num_input_tokens_seen": 29344489,
"step": 1180
},
{
"epoch": 27.47826086956522,
"grad_norm": 0.2166384607553482,
"learning_rate": 5.373135979747227e-06,
"loss": 0.006,
"num_input_tokens_seen": 29464082,
"step": 1185
},
{
"epoch": 27.594202898550726,
"grad_norm": 0.15387850999832153,
"learning_rate": 5.2125877892686496e-06,
"loss": 0.0043,
"num_input_tokens_seen": 29581124,
"step": 1190
},
{
"epoch": 27.71014492753623,
"grad_norm": 0.11962082982063293,
"learning_rate": 5.054195117578914e-06,
"loss": 0.0019,
"num_input_tokens_seen": 29696346,
"step": 1195
},
{
"epoch": 27.82608695652174,
"grad_norm": 0.18724732100963593,
"learning_rate": 4.897975218999926e-06,
"loss": 0.002,
"num_input_tokens_seen": 29815117,
"step": 1200
},
{
"epoch": 27.942028985507246,
"grad_norm": 0.09917350113391876,
"learning_rate": 4.743945111165068e-06,
"loss": 0.0022,
"num_input_tokens_seen": 29939175,
"step": 1205
},
{
"epoch": 28.057971014492754,
"grad_norm": 0.08235369622707367,
"learning_rate": 4.592121573165414e-06,
"loss": 0.0016,
"num_input_tokens_seen": 30079840,
"step": 1210
},
{
"epoch": 28.17391304347826,
"grad_norm": 0.20488996803760529,
"learning_rate": 4.442521143721892e-06,
"loss": 0.0033,
"num_input_tokens_seen": 30192219,
"step": 1215
},
{
"epoch": 28.28985507246377,
"grad_norm": 0.05383768677711487,
"learning_rate": 4.295160119383712e-06,
"loss": 0.0018,
"num_input_tokens_seen": 30330969,
"step": 1220
},
{
"epoch": 28.405797101449274,
"grad_norm": 0.14237363636493683,
"learning_rate": 4.150054552753055e-06,
"loss": 0.0018,
"num_input_tokens_seen": 30453302,
"step": 1225
},
{
"epoch": 28.52173913043478,
"grad_norm": 0.12487669289112091,
"learning_rate": 4.007220250736454e-06,
"loss": 0.0078,
"num_input_tokens_seen": 30568943,
"step": 1230
},
{
"epoch": 28.63768115942029,
"grad_norm": 0.1423855572938919,
"learning_rate": 3.866672772822863e-06,
"loss": 0.0019,
"num_input_tokens_seen": 30696057,
"step": 1235
},
{
"epoch": 28.753623188405797,
"grad_norm": 0.1543101817369461,
"learning_rate": 3.7284274293887115e-06,
"loss": 0.0019,
"num_input_tokens_seen": 30815506,
"step": 1240
},
{
"epoch": 28.869565217391305,
"grad_norm": 0.1402539610862732,
"learning_rate": 3.592499280030057e-06,
"loss": 0.0027,
"num_input_tokens_seen": 30916446,
"step": 1245
},
{
"epoch": 28.985507246376812,
"grad_norm": 0.26191645860671997,
"learning_rate": 3.458903131922134e-06,
"loss": 0.0023,
"num_input_tokens_seen": 31054242,
"step": 1250
},
{
"epoch": 29.10144927536232,
"grad_norm": 0.09874732792377472,
"learning_rate": 3.3276535382063213e-06,
"loss": 0.0029,
"num_input_tokens_seen": 31189078,
"step": 1255
},
{
"epoch": 29.217391304347824,
"grad_norm": 0.11677820980548859,
"learning_rate": 3.198764796404807e-06,
"loss": 0.0018,
"num_input_tokens_seen": 31311374,
"step": 1260
},
{
"epoch": 29.333333333333332,
"grad_norm": 0.05459802597761154,
"learning_rate": 3.0722509468631392e-06,
"loss": 0.0018,
"num_input_tokens_seen": 31444681,
"step": 1265
},
{
"epoch": 29.44927536231884,
"grad_norm": 0.1113714948296547,
"learning_rate": 2.948125771220697e-06,
"loss": 0.0018,
"num_input_tokens_seen": 31567569,
"step": 1270
},
{
"epoch": 29.565217391304348,
"grad_norm": 0.1816156655550003,
"learning_rate": 2.8264027909094715e-06,
"loss": 0.0019,
"num_input_tokens_seen": 31697338,
"step": 1275
},
{
"epoch": 29.681159420289855,
"grad_norm": 0.13639949262142181,
"learning_rate": 2.707095265681081e-06,
"loss": 0.0018,
"num_input_tokens_seen": 31826661,
"step": 1280
},
{
"epoch": 29.797101449275363,
"grad_norm": 0.05292365327477455,
"learning_rate": 2.5902161921623454e-06,
"loss": 0.0023,
"num_input_tokens_seen": 31944680,
"step": 1285
},
{
"epoch": 29.91304347826087,
"grad_norm": 0.16608740389347076,
"learning_rate": 2.475778302439524e-06,
"loss": 0.0078,
"num_input_tokens_seen": 32067106,
"step": 1290
},
{
"epoch": 30.028985507246375,
"grad_norm": 0.09277443587779999,
"learning_rate": 2.3637940626713346e-06,
"loss": 0.0018,
"num_input_tokens_seen": 32184526,
"step": 1295
},
{
"epoch": 30.144927536231883,
"grad_norm": 0.18832191824913025,
"learning_rate": 2.254275671731007e-06,
"loss": 0.0017,
"num_input_tokens_seen": 32309423,
"step": 1300
},
{
"epoch": 30.26086956521739,
"grad_norm": 0.1828456073999405,
"learning_rate": 2.14723505987737e-06,
"loss": 0.0071,
"num_input_tokens_seen": 32429445,
"step": 1305
},
{
"epoch": 30.3768115942029,
"grad_norm": 0.07503814995288849,
"learning_rate": 2.0426838874552714e-06,
"loss": 0.0016,
"num_input_tokens_seen": 32540571,
"step": 1310
},
{
"epoch": 30.492753623188406,
"grad_norm": 0.19047732651233673,
"learning_rate": 1.9406335436253724e-06,
"loss": 0.0018,
"num_input_tokens_seen": 32665528,
"step": 1315
},
{
"epoch": 30.608695652173914,
"grad_norm": 0.17791509628295898,
"learning_rate": 1.8410951451234533e-06,
"loss": 0.0017,
"num_input_tokens_seen": 32800773,
"step": 1320
},
{
"epoch": 30.72463768115942,
"grad_norm": 0.10698456317186356,
"learning_rate": 1.7440795350494588e-06,
"loss": 0.0017,
"num_input_tokens_seen": 32928397,
"step": 1325
},
{
"epoch": 30.840579710144926,
"grad_norm": 0.0963551327586174,
"learning_rate": 1.649597281686302e-06,
"loss": 0.0019,
"num_input_tokens_seen": 33054819,
"step": 1330
},
{
"epoch": 30.956521739130434,
"grad_norm": 0.24703514575958252,
"learning_rate": 1.5576586773486195e-06,
"loss": 0.0018,
"num_input_tokens_seen": 33180616,
"step": 1335
},
{
"epoch": 31.07246376811594,
"grad_norm": 0.12497910857200623,
"learning_rate": 1.4682737372615967e-06,
"loss": 0.0038,
"num_input_tokens_seen": 33298041,
"step": 1340
},
{
"epoch": 31.18840579710145,
"grad_norm": 0.18260960280895233,
"learning_rate": 1.3814521984699596e-06,
"loss": 0.0052,
"num_input_tokens_seen": 33408343,
"step": 1345
},
{
"epoch": 31.304347826086957,
"grad_norm": 0.13422255218029022,
"learning_rate": 1.297203518777293e-06,
"loss": 0.0018,
"num_input_tokens_seen": 33545364,
"step": 1350
},
{
"epoch": 31.420289855072465,
"grad_norm": 0.1285027116537094,
"learning_rate": 1.2155368757157643e-06,
"loss": 0.0019,
"num_input_tokens_seen": 33652900,
"step": 1355
},
{
"epoch": 31.536231884057973,
"grad_norm": 0.12832242250442505,
"learning_rate": 1.1364611655463736e-06,
"loss": 0.0019,
"num_input_tokens_seen": 33768791,
"step": 1360
},
{
"epoch": 31.652173913043477,
"grad_norm": 0.12093157321214676,
"learning_rate": 1.0599850022898539e-06,
"loss": 0.0017,
"num_input_tokens_seen": 33892837,
"step": 1365
},
{
"epoch": 31.768115942028984,
"grad_norm": 0.7227018475532532,
"learning_rate": 9.861167167883046e-07,
"loss": 0.0022,
"num_input_tokens_seen": 34015288,
"step": 1370
},
{
"epoch": 31.884057971014492,
"grad_norm": 2.143653631210327,
"learning_rate": 9.148643557976955e-07,
"loss": 0.0037,
"num_input_tokens_seen": 34154884,
"step": 1375
},
{
"epoch": 32.0,
"grad_norm": 0.17518474161624908,
"learning_rate": 8.462356811112987e-07,
"loss": 0.0019,
"num_input_tokens_seen": 34292320,
"step": 1380
},
{
"epoch": 32.11594202898551,
"grad_norm": 0.1274159997701645,
"learning_rate": 7.802381687141535e-07,
"loss": 0.0017,
"num_input_tokens_seen": 34413850,
"step": 1385
},
{
"epoch": 32.231884057971016,
"grad_norm": 0.11443401873111725,
"learning_rate": 7.168790079686932e-07,
"loss": 0.0018,
"num_input_tokens_seen": 34547127,
"step": 1390
},
{
"epoch": 32.34782608695652,
"grad_norm": 0.08239752799272537,
"learning_rate": 6.561651008315738e-07,
"loss": 0.0035,
"num_input_tokens_seen": 34685112,
"step": 1395
},
{
"epoch": 32.46376811594203,
"grad_norm": 0.7361220717430115,
"learning_rate": 5.981030611018234e-07,
"loss": 0.0063,
"num_input_tokens_seen": 34810484,
"step": 1400
},
{
"epoch": 32.57971014492754,
"grad_norm": 0.20323431491851807,
"learning_rate": 5.426992137003622e-07,
"loss": 0.0018,
"num_input_tokens_seen": 34920531,
"step": 1405
},
{
"epoch": 32.69565217391305,
"grad_norm": 0.11165229231119156,
"learning_rate": 4.899595939810236e-07,
"loss": 0.002,
"num_input_tokens_seen": 35035657,
"step": 1410
},
{
"epoch": 32.81159420289855,
"grad_norm": 0.15023387968540192,
"learning_rate": 4.398899470730827e-07,
"loss": 0.0017,
"num_input_tokens_seen": 35167466,
"step": 1415
},
{
"epoch": 32.927536231884055,
"grad_norm": 0.18479810655117035,
"learning_rate": 3.9249572725543196e-07,
"loss": 0.0016,
"num_input_tokens_seen": 35296818,
"step": 1420
},
{
"epoch": 33.04347826086956,
"grad_norm": 0.08527754247188568,
"learning_rate": 3.477820973624063e-07,
"loss": 0.0015,
"num_input_tokens_seen": 35430399,
"step": 1425
},
{
"epoch": 33.15942028985507,
"grad_norm": 0.16888481378555298,
"learning_rate": 3.0575392822139726e-07,
"loss": 0.0057,
"num_input_tokens_seen": 35551540,
"step": 1430
},
{
"epoch": 33.27536231884058,
"grad_norm": 0.18187086284160614,
"learning_rate": 2.664157981222437e-07,
"loss": 0.0016,
"num_input_tokens_seen": 35676077,
"step": 1435
},
{
"epoch": 33.391304347826086,
"grad_norm": 0.15047162771224976,
"learning_rate": 2.297719923185032e-07,
"loss": 0.0016,
"num_input_tokens_seen": 35785127,
"step": 1440
},
{
"epoch": 33.507246376811594,
"grad_norm": 0.12288761883974075,
"learning_rate": 1.9582650256064205e-07,
"loss": 0.0019,
"num_input_tokens_seen": 35911682,
"step": 1445
},
{
"epoch": 33.6231884057971,
"grad_norm": 0.22509098052978516,
"learning_rate": 1.645830266611914e-07,
"loss": 0.0017,
"num_input_tokens_seen": 36030754,
"step": 1450
},
{
"epoch": 33.73913043478261,
"grad_norm": 2.0408618450164795,
"learning_rate": 1.3604496809195288e-07,
"loss": 0.0042,
"num_input_tokens_seen": 36146749,
"step": 1455
},
{
"epoch": 33.85507246376812,
"grad_norm": 0.10705255717039108,
"learning_rate": 1.1021543561322012e-07,
"loss": 0.0017,
"num_input_tokens_seen": 36278454,
"step": 1460
},
{
"epoch": 33.971014492753625,
"grad_norm": 1876.0384521484375,
"learning_rate": 8.709724293513854e-08,
"loss": 0.0017,
"num_input_tokens_seen": 36408834,
"step": 1465
},
{
"epoch": 34.08695652173913,
"grad_norm": 0.1927630454301834,
"learning_rate": 6.66929084112089e-08,
"loss": 0.0015,
"num_input_tokens_seen": 36550538,
"step": 1470
},
{
"epoch": 34.20289855072464,
"grad_norm": 0.1668202131986618,
"learning_rate": 4.900465476393168e-08,
"loss": 0.0018,
"num_input_tokens_seen": 36647436,
"step": 1475
},
{
"epoch": 34.31884057971015,
"grad_norm": 0.7123565673828125,
"learning_rate": 3.403440884269526e-08,
"loss": 0.0024,
"num_input_tokens_seen": 36785387,
"step": 1480
},
{
"epoch": 34.43478260869565,
"grad_norm": 0.16973845660686493,
"learning_rate": 2.1783801413866046e-08,
"loss": 0.0021,
"num_input_tokens_seen": 36915606,
"step": 1485
},
{
"epoch": 34.55072463768116,
"grad_norm": 2.034724473953247,
"learning_rate": 1.2254166983152737e-08,
"loss": 0.0035,
"num_input_tokens_seen": 37036117,
"step": 1490
},
{
"epoch": 34.666666666666664,
"grad_norm": 0.155415877699852,
"learning_rate": 5.446543650219904e-09,
"loss": 0.0016,
"num_input_tokens_seen": 37165587,
"step": 1495
},
{
"epoch": 34.78260869565217,
"grad_norm": 0.10199662297964096,
"learning_rate": 1.3616729956228425e-09,
"loss": 0.0015,
"num_input_tokens_seen": 37290827,
"step": 1500
},
{
"epoch": 34.89855072463768,
"grad_norm": 0.14740267395973206,
"learning_rate": 0.0,
"loss": 0.0053,
"num_input_tokens_seen": 37412688,
"step": 1505
},
{
"epoch": 34.89855072463768,
"num_input_tokens_seen": 37412688,
"step": 1505,
"total_flos": 8.033958240027034e+16,
"train_loss": 0.3889684765070578,
"train_runtime": 37510.9602,
"train_samples_per_second": 0.322,
"train_steps_per_second": 0.04
}
],
"logging_steps": 5,
"max_steps": 1505,
"num_input_tokens_seen": 37412688,
"num_train_epochs": 35,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 8.033958240027034e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}