gemma-7b-sft-full-openhermes-v0 / trainer_state.json
lewtun's picture
lewtun HF staff
Model save
34ed1ed verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.9946018893387314,
"eval_steps": 500,
"global_step": 1480,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0,
"grad_norm": 5172.492790419565,
"learning_rate": 6.756756756756757e-08,
"loss": 16.2669,
"step": 1
},
{
"epoch": 0.01,
"grad_norm": 4320.729018206391,
"learning_rate": 3.378378378378379e-07,
"loss": 15.6969,
"step": 5
},
{
"epoch": 0.03,
"grad_norm": 488.5525501221014,
"learning_rate": 6.756756756756758e-07,
"loss": 12.0521,
"step": 10
},
{
"epoch": 0.04,
"grad_norm": 289.24969349599496,
"learning_rate": 1.0135135135135136e-06,
"loss": 9.2401,
"step": 15
},
{
"epoch": 0.05,
"grad_norm": 126.31621322294927,
"learning_rate": 1.3513513513513515e-06,
"loss": 7.9461,
"step": 20
},
{
"epoch": 0.07,
"grad_norm": 71.89469484569835,
"learning_rate": 1.6891891891891894e-06,
"loss": 7.2699,
"step": 25
},
{
"epoch": 0.08,
"grad_norm": 103.51185464247453,
"learning_rate": 2.0270270270270273e-06,
"loss": 6.9743,
"step": 30
},
{
"epoch": 0.09,
"grad_norm": 47.053277376605855,
"learning_rate": 2.364864864864865e-06,
"loss": 6.7672,
"step": 35
},
{
"epoch": 0.11,
"grad_norm": 49.644931428662844,
"learning_rate": 2.702702702702703e-06,
"loss": 6.5877,
"step": 40
},
{
"epoch": 0.12,
"grad_norm": 35.00554720125077,
"learning_rate": 3.040540540540541e-06,
"loss": 6.4428,
"step": 45
},
{
"epoch": 0.13,
"grad_norm": 33.14232577992853,
"learning_rate": 3.3783783783783788e-06,
"loss": 6.2824,
"step": 50
},
{
"epoch": 0.15,
"grad_norm": 35.40029424621722,
"learning_rate": 3.7162162162162162e-06,
"loss": 6.0897,
"step": 55
},
{
"epoch": 0.16,
"grad_norm": 40.23393814067933,
"learning_rate": 4.0540540540540545e-06,
"loss": 5.8782,
"step": 60
},
{
"epoch": 0.18,
"grad_norm": 55.427875975850924,
"learning_rate": 4.391891891891892e-06,
"loss": 5.5613,
"step": 65
},
{
"epoch": 0.19,
"grad_norm": 96.7762735033588,
"learning_rate": 4.72972972972973e-06,
"loss": 4.8791,
"step": 70
},
{
"epoch": 0.2,
"grad_norm": 88.58451501801524,
"learning_rate": 5.067567567567568e-06,
"loss": 3.2214,
"step": 75
},
{
"epoch": 0.22,
"grad_norm": 22.221611879190654,
"learning_rate": 5.405405405405406e-06,
"loss": 1.7005,
"step": 80
},
{
"epoch": 0.23,
"grad_norm": 14.404618337150714,
"learning_rate": 5.743243243243244e-06,
"loss": 1.55,
"step": 85
},
{
"epoch": 0.24,
"grad_norm": 14.657001400244646,
"learning_rate": 6.081081081081082e-06,
"loss": 1.5228,
"step": 90
},
{
"epoch": 0.26,
"grad_norm": 6.295090886362698,
"learning_rate": 6.41891891891892e-06,
"loss": 1.4585,
"step": 95
},
{
"epoch": 0.27,
"grad_norm": 5.24494216356606,
"learning_rate": 6.7567567567567575e-06,
"loss": 1.4326,
"step": 100
},
{
"epoch": 0.28,
"grad_norm": 6.440473193782044,
"learning_rate": 7.0945945945945946e-06,
"loss": 1.3865,
"step": 105
},
{
"epoch": 0.3,
"grad_norm": 4.8147339321014,
"learning_rate": 7.4324324324324324e-06,
"loss": 1.3655,
"step": 110
},
{
"epoch": 0.31,
"grad_norm": 5.442428248209423,
"learning_rate": 7.77027027027027e-06,
"loss": 1.3394,
"step": 115
},
{
"epoch": 0.32,
"grad_norm": 5.78770666444895,
"learning_rate": 8.108108108108109e-06,
"loss": 1.3354,
"step": 120
},
{
"epoch": 0.34,
"grad_norm": 7.28821915344424,
"learning_rate": 8.445945945945948e-06,
"loss": 1.3065,
"step": 125
},
{
"epoch": 0.35,
"grad_norm": 6.492851696765312,
"learning_rate": 8.783783783783785e-06,
"loss": 1.2902,
"step": 130
},
{
"epoch": 0.36,
"grad_norm": 3.1773499503060454,
"learning_rate": 9.121621621621622e-06,
"loss": 1.2769,
"step": 135
},
{
"epoch": 0.38,
"grad_norm": 2.9874866899468238,
"learning_rate": 9.45945945945946e-06,
"loss": 1.266,
"step": 140
},
{
"epoch": 0.39,
"grad_norm": 9.94355578144089,
"learning_rate": 9.797297297297298e-06,
"loss": 1.2624,
"step": 145
},
{
"epoch": 0.4,
"grad_norm": 3.0193035270709077,
"learning_rate": 9.999944372378571e-06,
"loss": 1.2452,
"step": 150
},
{
"epoch": 0.42,
"grad_norm": 5.139515176735739,
"learning_rate": 9.999318575852451e-06,
"loss": 1.2341,
"step": 155
},
{
"epoch": 0.43,
"grad_norm": 3.156285780849088,
"learning_rate": 9.99799753559161e-06,
"loss": 1.2106,
"step": 160
},
{
"epoch": 0.45,
"grad_norm": 5.924005437849429,
"learning_rate": 9.995981435310078e-06,
"loss": 1.221,
"step": 165
},
{
"epoch": 0.46,
"grad_norm": 4.165437348359594,
"learning_rate": 9.993270555382283e-06,
"loss": 1.2105,
"step": 170
},
{
"epoch": 0.47,
"grad_norm": 2.342058479511419,
"learning_rate": 9.989865272804064e-06,
"loss": 1.2067,
"step": 175
},
{
"epoch": 0.49,
"grad_norm": 2.6456490809423467,
"learning_rate": 9.985766061140233e-06,
"loss": 1.196,
"step": 180
},
{
"epoch": 0.5,
"grad_norm": 6.465974100807255,
"learning_rate": 9.980973490458728e-06,
"loss": 1.1881,
"step": 185
},
{
"epoch": 0.51,
"grad_norm": 2.808324774416455,
"learning_rate": 9.97548822725133e-06,
"loss": 1.1852,
"step": 190
},
{
"epoch": 0.53,
"grad_norm": 2.834201027736285,
"learning_rate": 9.969311034340977e-06,
"loss": 1.1839,
"step": 195
},
{
"epoch": 0.54,
"grad_norm": 4.502666083698819,
"learning_rate": 9.962442770775675e-06,
"loss": 1.1648,
"step": 200
},
{
"epoch": 0.55,
"grad_norm": 5.780648711057674,
"learning_rate": 9.954884391709043e-06,
"loss": 1.1619,
"step": 205
},
{
"epoch": 0.57,
"grad_norm": 3.498462904619314,
"learning_rate": 9.946636948267468e-06,
"loss": 1.1483,
"step": 210
},
{
"epoch": 0.58,
"grad_norm": 3.0642292894829213,
"learning_rate": 9.937701587403941e-06,
"loss": 1.1412,
"step": 215
},
{
"epoch": 0.59,
"grad_norm": 4.727096769429053,
"learning_rate": 9.928079551738542e-06,
"loss": 1.1446,
"step": 220
},
{
"epoch": 0.61,
"grad_norm": 2.206091891548497,
"learning_rate": 9.91777217938564e-06,
"loss": 1.1354,
"step": 225
},
{
"epoch": 0.62,
"grad_norm": 3.242676582016566,
"learning_rate": 9.906780903767799e-06,
"loss": 1.137,
"step": 230
},
{
"epoch": 0.63,
"grad_norm": 2.5806553543973396,
"learning_rate": 9.895107253416434e-06,
"loss": 1.134,
"step": 235
},
{
"epoch": 0.65,
"grad_norm": 2.305243444378998,
"learning_rate": 9.882752851759247e-06,
"loss": 1.1411,
"step": 240
},
{
"epoch": 0.66,
"grad_norm": 3.911161316221469,
"learning_rate": 9.869719416894462e-06,
"loss": 1.1144,
"step": 245
},
{
"epoch": 0.67,
"grad_norm": 4.2780174060875025,
"learning_rate": 9.856008761351882e-06,
"loss": 1.109,
"step": 250
},
{
"epoch": 0.69,
"grad_norm": 2.3947970947073425,
"learning_rate": 9.841622791840839e-06,
"loss": 1.1143,
"step": 255
},
{
"epoch": 0.7,
"grad_norm": 5.2640331483926595,
"learning_rate": 9.826563508985017e-06,
"loss": 1.1158,
"step": 260
},
{
"epoch": 0.72,
"grad_norm": 2.8813671261206033,
"learning_rate": 9.810833007044247e-06,
"loss": 1.1038,
"step": 265
},
{
"epoch": 0.73,
"grad_norm": 1.7947375217746502,
"learning_rate": 9.794433473623249e-06,
"loss": 1.1063,
"step": 270
},
{
"epoch": 0.74,
"grad_norm": 2.239529292108957,
"learning_rate": 9.777367189367412e-06,
"loss": 1.1034,
"step": 275
},
{
"epoch": 0.76,
"grad_norm": 2.8141790841216405,
"learning_rate": 9.759636527645633e-06,
"loss": 1.094,
"step": 280
},
{
"epoch": 0.77,
"grad_norm": 3.3580672092709722,
"learning_rate": 9.74124395422025e-06,
"loss": 1.0989,
"step": 285
},
{
"epoch": 0.78,
"grad_norm": 3.116597962727994,
"learning_rate": 9.722192026904145e-06,
"loss": 1.0927,
"step": 290
},
{
"epoch": 0.8,
"grad_norm": 8.108195152123194,
"learning_rate": 9.702483395205023e-06,
"loss": 1.0785,
"step": 295
},
{
"epoch": 0.81,
"grad_norm": 2.4053331358570564,
"learning_rate": 9.682120799956961e-06,
"loss": 1.0939,
"step": 300
},
{
"epoch": 0.82,
"grad_norm": 3.036700462595253,
"learning_rate": 9.661107072939244e-06,
"loss": 1.0957,
"step": 305
},
{
"epoch": 0.84,
"grad_norm": 1.9808850854861142,
"learning_rate": 9.639445136482549e-06,
"loss": 1.0718,
"step": 310
},
{
"epoch": 0.85,
"grad_norm": 1.7199945370090353,
"learning_rate": 9.61713800306255e-06,
"loss": 1.0908,
"step": 315
},
{
"epoch": 0.86,
"grad_norm": 3.416121355561205,
"learning_rate": 9.594188774880981e-06,
"loss": 1.0683,
"step": 320
},
{
"epoch": 0.88,
"grad_norm": 2.6221732481646085,
"learning_rate": 9.570600643434217e-06,
"loss": 1.0812,
"step": 325
},
{
"epoch": 0.89,
"grad_norm": 3.0081014994296655,
"learning_rate": 9.546376889069442e-06,
"loss": 1.0672,
"step": 330
},
{
"epoch": 0.9,
"grad_norm": 1.928715018400264,
"learning_rate": 9.521520880528453e-06,
"loss": 1.0696,
"step": 335
},
{
"epoch": 0.92,
"grad_norm": 3.510557690197284,
"learning_rate": 9.496036074479184e-06,
"loss": 1.0748,
"step": 340
},
{
"epoch": 0.93,
"grad_norm": 2.1531622758856206,
"learning_rate": 9.46992601503499e-06,
"loss": 1.072,
"step": 345
},
{
"epoch": 0.94,
"grad_norm": 1.5600473945363453,
"learning_rate": 9.44319433326178e-06,
"loss": 1.0674,
"step": 350
},
{
"epoch": 0.96,
"grad_norm": 1.9316399234342516,
"learning_rate": 9.415844746673047e-06,
"loss": 1.0715,
"step": 355
},
{
"epoch": 0.97,
"grad_norm": 3.1641780363359624,
"learning_rate": 9.387881058712888e-06,
"loss": 1.0607,
"step": 360
},
{
"epoch": 0.99,
"grad_norm": 2.5389167608685477,
"learning_rate": 9.359307158227067e-06,
"loss": 1.0571,
"step": 365
},
{
"epoch": 1.0,
"grad_norm": 3.209876430466923,
"learning_rate": 9.330127018922195e-06,
"loss": 1.0656,
"step": 370
},
{
"epoch": 1.01,
"grad_norm": 2.0112001074770856,
"learning_rate": 9.300344698813124e-06,
"loss": 1.0473,
"step": 375
},
{
"epoch": 1.03,
"grad_norm": 2.3621990368530037,
"learning_rate": 9.269964339658605e-06,
"loss": 1.0479,
"step": 380
},
{
"epoch": 1.04,
"grad_norm": 1.9555447312938368,
"learning_rate": 9.238990166385304e-06,
"loss": 1.0481,
"step": 385
},
{
"epoch": 1.05,
"grad_norm": 1.9979169106404795,
"learning_rate": 9.207426486500252e-06,
"loss": 1.0489,
"step": 390
},
{
"epoch": 1.07,
"grad_norm": 2.7432607054958145,
"learning_rate": 9.175277689491804e-06,
"loss": 1.0553,
"step": 395
},
{
"epoch": 1.08,
"grad_norm": 1.9273849661076683,
"learning_rate": 9.142548246219212e-06,
"loss": 1.0398,
"step": 400
},
{
"epoch": 1.09,
"grad_norm": 2.0451128390125235,
"learning_rate": 9.109242708290864e-06,
"loss": 1.0417,
"step": 405
},
{
"epoch": 1.11,
"grad_norm": 4.496881794147535,
"learning_rate": 9.075365707431311e-06,
"loss": 1.0348,
"step": 410
},
{
"epoch": 1.12,
"grad_norm": 3.7112676473960327,
"learning_rate": 9.040921954837139e-06,
"loss": 1.0475,
"step": 415
},
{
"epoch": 1.13,
"grad_norm": 1.7843942716305177,
"learning_rate": 9.005916240521788e-06,
"loss": 1.0435,
"step": 420
},
{
"epoch": 1.15,
"grad_norm": 3.688231246483607,
"learning_rate": 8.97035343264943e-06,
"loss": 1.0316,
"step": 425
},
{
"epoch": 1.16,
"grad_norm": 3.0133406140771672,
"learning_rate": 8.93423847685795e-06,
"loss": 1.0363,
"step": 430
},
{
"epoch": 1.17,
"grad_norm": 3.65260559397669,
"learning_rate": 8.89757639557118e-06,
"loss": 1.0295,
"step": 435
},
{
"epoch": 1.19,
"grad_norm": 2.0183294533927083,
"learning_rate": 8.860372287300432e-06,
"loss": 1.033,
"step": 440
},
{
"epoch": 1.2,
"grad_norm": 1.8479490054275187,
"learning_rate": 8.822631325935463e-06,
"loss": 1.0472,
"step": 445
},
{
"epoch": 1.21,
"grad_norm": 2.114757094329874,
"learning_rate": 8.78435876002496e-06,
"loss": 1.0295,
"step": 450
},
{
"epoch": 1.23,
"grad_norm": 1.9082626009926926,
"learning_rate": 8.745559912046625e-06,
"loss": 1.033,
"step": 455
},
{
"epoch": 1.24,
"grad_norm": 1.730783726030946,
"learning_rate": 8.706240177667003e-06,
"loss": 1.0385,
"step": 460
},
{
"epoch": 1.26,
"grad_norm": 2.5320093097552614,
"learning_rate": 8.666405024991105e-06,
"loss": 1.0375,
"step": 465
},
{
"epoch": 1.27,
"grad_norm": 3.400617294472977,
"learning_rate": 8.626059993801988e-06,
"loss": 1.0253,
"step": 470
},
{
"epoch": 1.28,
"grad_norm": 2.7269138208449157,
"learning_rate": 8.585210694790333e-06,
"loss": 1.0191,
"step": 475
},
{
"epoch": 1.3,
"grad_norm": 1.943069131491427,
"learning_rate": 8.543862808774193e-06,
"loss": 1.021,
"step": 480
},
{
"epoch": 1.31,
"grad_norm": 1.8339225500743817,
"learning_rate": 8.502022085908963e-06,
"loss": 1.0164,
"step": 485
},
{
"epoch": 1.32,
"grad_norm": 2.137449239722819,
"learning_rate": 8.459694344887732e-06,
"loss": 1.0367,
"step": 490
},
{
"epoch": 1.34,
"grad_norm": 2.581812285540177,
"learning_rate": 8.416885472132077e-06,
"loss": 1.0283,
"step": 495
},
{
"epoch": 1.35,
"grad_norm": 1.8160848127506812,
"learning_rate": 8.373601420973464e-06,
"loss": 1.017,
"step": 500
},
{
"epoch": 1.36,
"grad_norm": 2.346923224689188,
"learning_rate": 8.329848210825322e-06,
"loss": 1.0237,
"step": 505
},
{
"epoch": 1.38,
"grad_norm": 2.024470398972999,
"learning_rate": 8.285631926345943e-06,
"loss": 1.028,
"step": 510
},
{
"epoch": 1.39,
"grad_norm": 2.382418457888275,
"learning_rate": 8.240958716592304e-06,
"loss": 1.0103,
"step": 515
},
{
"epoch": 1.4,
"grad_norm": 1.754061139243149,
"learning_rate": 8.195834794164925e-06,
"loss": 1.0179,
"step": 520
},
{
"epoch": 1.42,
"grad_norm": 1.6884398821609177,
"learning_rate": 8.150266434343904e-06,
"loss": 1.0218,
"step": 525
},
{
"epoch": 1.43,
"grad_norm": 1.8408143701714872,
"learning_rate": 8.104259974216219e-06,
"loss": 1.0109,
"step": 530
},
{
"epoch": 1.44,
"grad_norm": 1.8898424458167393,
"learning_rate": 8.057821811794457e-06,
"loss": 1.0113,
"step": 535
},
{
"epoch": 1.46,
"grad_norm": 2.0040005987182923,
"learning_rate": 8.010958405127048e-06,
"loss": 1.0205,
"step": 540
},
{
"epoch": 1.47,
"grad_norm": 1.8031516449231006,
"learning_rate": 7.963676271400158e-06,
"loss": 1.0102,
"step": 545
},
{
"epoch": 1.48,
"grad_norm": 2.0776879759963798,
"learning_rate": 7.915981986031367e-06,
"loss": 1.0089,
"step": 550
},
{
"epoch": 1.5,
"grad_norm": 2.4545184199390695,
"learning_rate": 7.86788218175523e-06,
"loss": 1.021,
"step": 555
},
{
"epoch": 1.51,
"grad_norm": 1.9664285212166526,
"learning_rate": 7.819383547700889e-06,
"loss": 1.008,
"step": 560
},
{
"epoch": 1.52,
"grad_norm": 2.7035372351508404,
"learning_rate": 7.770492828461824e-06,
"loss": 1.013,
"step": 565
},
{
"epoch": 1.54,
"grad_norm": 2.737641626508961,
"learning_rate": 7.721216823157896e-06,
"loss": 1.0125,
"step": 570
},
{
"epoch": 1.55,
"grad_norm": 1.7709332048302728,
"learning_rate": 7.671562384489819e-06,
"loss": 1.0197,
"step": 575
},
{
"epoch": 1.57,
"grad_norm": 1.592514417019854,
"learning_rate": 7.621536417786159e-06,
"loss": 1.006,
"step": 580
},
{
"epoch": 1.58,
"grad_norm": 1.8116371870435586,
"learning_rate": 7.571145880043036e-06,
"loss": 1.0021,
"step": 585
},
{
"epoch": 1.59,
"grad_norm": 2.742484137764987,
"learning_rate": 7.520397778956623e-06,
"loss": 0.9942,
"step": 590
},
{
"epoch": 1.61,
"grad_norm": 1.8390385024848412,
"learning_rate": 7.469299171948608e-06,
"loss": 1.0085,
"step": 595
},
{
"epoch": 1.62,
"grad_norm": 1.8627501795637935,
"learning_rate": 7.417857165184723e-06,
"loss": 1.0065,
"step": 600
},
{
"epoch": 1.63,
"grad_norm": 1.8662148392820241,
"learning_rate": 7.366078912586523e-06,
"loss": 1.0069,
"step": 605
},
{
"epoch": 1.65,
"grad_norm": 1.9898176799731553,
"learning_rate": 7.313971614836496e-06,
"loss": 1.0075,
"step": 610
},
{
"epoch": 1.66,
"grad_norm": 1.6374351854744644,
"learning_rate": 7.261542518376677e-06,
"loss": 0.9954,
"step": 615
},
{
"epoch": 1.67,
"grad_norm": 2.3713553804160554,
"learning_rate": 7.208798914400916e-06,
"loss": 1.0041,
"step": 620
},
{
"epoch": 1.69,
"grad_norm": 2.43390000834423,
"learning_rate": 7.155748137840892e-06,
"loss": 1.0043,
"step": 625
},
{
"epoch": 1.7,
"grad_norm": 1.9046613363144396,
"learning_rate": 7.102397566346073e-06,
"loss": 0.9984,
"step": 630
},
{
"epoch": 1.71,
"grad_norm": 1.667190267378518,
"learning_rate": 7.048754619257716e-06,
"loss": 1.0029,
"step": 635
},
{
"epoch": 1.73,
"grad_norm": 1.638081070696857,
"learning_rate": 6.994826756577082e-06,
"loss": 1.0032,
"step": 640
},
{
"epoch": 1.74,
"grad_norm": 2.2506476935835407,
"learning_rate": 6.940621477927988e-06,
"loss": 1.0023,
"step": 645
},
{
"epoch": 1.75,
"grad_norm": 2.0559944356090116,
"learning_rate": 6.88614632151385e-06,
"loss": 1.0036,
"step": 650
},
{
"epoch": 1.77,
"grad_norm": 1.7722705407021138,
"learning_rate": 6.831408863069364e-06,
"loss": 0.9981,
"step": 655
},
{
"epoch": 1.78,
"grad_norm": 1.510560677423063,
"learning_rate": 6.7764167148069695e-06,
"loss": 0.9845,
"step": 660
},
{
"epoch": 1.79,
"grad_norm": 1.6159682830961077,
"learning_rate": 6.721177524358226e-06,
"loss": 1.0014,
"step": 665
},
{
"epoch": 1.81,
"grad_norm": 2.547134856734226,
"learning_rate": 6.665698973710289e-06,
"loss": 0.9955,
"step": 670
},
{
"epoch": 1.82,
"grad_norm": 2.6937900779118524,
"learning_rate": 6.609988778137582e-06,
"loss": 0.9856,
"step": 675
},
{
"epoch": 1.84,
"grad_norm": 2.0660116621981937,
"learning_rate": 6.554054685128857e-06,
"loss": 0.987,
"step": 680
},
{
"epoch": 1.85,
"grad_norm": 1.7256065453185816,
"learning_rate": 6.497904473309766e-06,
"loss": 0.9907,
"step": 685
},
{
"epoch": 1.86,
"grad_norm": 1.9848736194116567,
"learning_rate": 6.44154595136111e-06,
"loss": 0.9938,
"step": 690
},
{
"epoch": 1.88,
"grad_norm": 2.226796297825019,
"learning_rate": 6.384986956932897e-06,
"loss": 0.9891,
"step": 695
},
{
"epoch": 1.89,
"grad_norm": 1.7756206272082837,
"learning_rate": 6.328235355554382e-06,
"loss": 0.989,
"step": 700
},
{
"epoch": 1.9,
"grad_norm": 2.055029436503266,
"learning_rate": 6.271299039540228e-06,
"loss": 0.9854,
"step": 705
},
{
"epoch": 1.92,
"grad_norm": 2.1572868944232733,
"learning_rate": 6.214185926892936e-06,
"loss": 0.991,
"step": 710
},
{
"epoch": 1.93,
"grad_norm": 1.705088008506106,
"learning_rate": 6.156903960201709e-06,
"loss": 0.99,
"step": 715
},
{
"epoch": 1.94,
"grad_norm": 2.394966448163764,
"learning_rate": 6.099461105537889e-06,
"loss": 0.9866,
"step": 720
},
{
"epoch": 1.96,
"grad_norm": 2.016768798949277,
"learning_rate": 6.041865351347146e-06,
"loss": 0.9919,
"step": 725
},
{
"epoch": 1.97,
"grad_norm": 1.8476093711911672,
"learning_rate": 5.984124707338528e-06,
"loss": 0.9932,
"step": 730
},
{
"epoch": 1.98,
"grad_norm": 1.9896940215311267,
"learning_rate": 5.926247203370583e-06,
"loss": 0.9958,
"step": 735
},
{
"epoch": 2.0,
"grad_norm": 2.5258042306430637,
"learning_rate": 5.8682408883346535e-06,
"loss": 0.9915,
"step": 740
},
{
"epoch": 2.01,
"grad_norm": 2.300171514270174,
"learning_rate": 5.810113829035544e-06,
"loss": 0.9681,
"step": 745
},
{
"epoch": 2.02,
"grad_norm": 1.9721303641863295,
"learning_rate": 5.751874109069685e-06,
"loss": 0.9597,
"step": 750
},
{
"epoch": 2.04,
"grad_norm": 2.2390559024844086,
"learning_rate": 5.693529827700967e-06,
"loss": 0.9578,
"step": 755
},
{
"epoch": 2.05,
"grad_norm": 2.0408626045391167,
"learning_rate": 5.635089098734394e-06,
"loss": 0.9759,
"step": 760
},
{
"epoch": 2.06,
"grad_norm": 2.3474388556665353,
"learning_rate": 5.576560049387717e-06,
"loss": 0.9692,
"step": 765
},
{
"epoch": 2.08,
"grad_norm": 1.577362362698819,
"learning_rate": 5.517950819161196e-06,
"loss": 0.9693,
"step": 770
},
{
"epoch": 2.09,
"grad_norm": 1.798565847037349,
"learning_rate": 5.459269558705667e-06,
"loss": 0.9583,
"step": 775
},
{
"epoch": 2.11,
"grad_norm": 1.9304589279981041,
"learning_rate": 5.400524428689035e-06,
"loss": 0.9717,
"step": 780
},
{
"epoch": 2.12,
"grad_norm": 1.721891020322892,
"learning_rate": 5.341723598661409e-06,
"loss": 0.9639,
"step": 785
},
{
"epoch": 2.13,
"grad_norm": 1.7487840479558292,
"learning_rate": 5.282875245918963e-06,
"loss": 0.9705,
"step": 790
},
{
"epoch": 2.15,
"grad_norm": 1.8716448470255413,
"learning_rate": 5.2239875543667465e-06,
"loss": 0.9714,
"step": 795
},
{
"epoch": 2.16,
"grad_norm": 1.5692767824863278,
"learning_rate": 5.165068713380568e-06,
"loss": 0.9728,
"step": 800
},
{
"epoch": 2.17,
"grad_norm": 1.6666960593346472,
"learning_rate": 5.106126916668118e-06,
"loss": 0.9634,
"step": 805
},
{
"epoch": 2.19,
"grad_norm": 1.6602656394819693,
"learning_rate": 5.047170361129484e-06,
"loss": 0.9665,
"step": 810
},
{
"epoch": 2.2,
"grad_norm": 1.5606922426383971,
"learning_rate": 4.988207245717232e-06,
"loss": 0.9615,
"step": 815
},
{
"epoch": 2.21,
"grad_norm": 1.8903502303369755,
"learning_rate": 4.929245770296191e-06,
"loss": 0.9517,
"step": 820
},
{
"epoch": 2.23,
"grad_norm": 2.0904161072997076,
"learning_rate": 4.870294134503123e-06,
"loss": 0.9578,
"step": 825
},
{
"epoch": 2.24,
"grad_norm": 1.8320825036562747,
"learning_rate": 4.811360536606416e-06,
"loss": 0.9695,
"step": 830
},
{
"epoch": 2.25,
"grad_norm": 1.9133665529378556,
"learning_rate": 4.752453172365966e-06,
"loss": 0.9585,
"step": 835
},
{
"epoch": 2.27,
"grad_norm": 1.8758941343800197,
"learning_rate": 4.69358023389342e-06,
"loss": 0.9608,
"step": 840
},
{
"epoch": 2.28,
"grad_norm": 2.133289277786595,
"learning_rate": 4.634749908512907e-06,
"loss": 0.9619,
"step": 845
},
{
"epoch": 2.29,
"grad_norm": 1.8131410749776957,
"learning_rate": 4.575970377622456e-06,
"loss": 0.9623,
"step": 850
},
{
"epoch": 2.31,
"grad_norm": 2.0361433234498514,
"learning_rate": 4.517249815556219e-06,
"loss": 0.9572,
"step": 855
},
{
"epoch": 2.32,
"grad_norm": 1.8450224454341402,
"learning_rate": 4.458596388447691e-06,
"loss": 0.9588,
"step": 860
},
{
"epoch": 2.33,
"grad_norm": 1.5855628731202065,
"learning_rate": 4.400018253094065e-06,
"loss": 0.9544,
"step": 865
},
{
"epoch": 2.35,
"grad_norm": 1.7015227109739095,
"learning_rate": 4.341523555821882e-06,
"loss": 0.9629,
"step": 870
},
{
"epoch": 2.36,
"grad_norm": 1.533925044445312,
"learning_rate": 4.283120431354137e-06,
"loss": 0.9571,
"step": 875
},
{
"epoch": 2.38,
"grad_norm": 1.5138597264361546,
"learning_rate": 4.224817001679011e-06,
"loss": 0.9642,
"step": 880
},
{
"epoch": 2.39,
"grad_norm": 2.3034815379141795,
"learning_rate": 4.1666213749203545e-06,
"loss": 0.9562,
"step": 885
},
{
"epoch": 2.4,
"grad_norm": 1.6982866609702905,
"learning_rate": 4.1085416442101205e-06,
"loss": 0.9628,
"step": 890
},
{
"epoch": 2.42,
"grad_norm": 1.8780386107869884,
"learning_rate": 4.050585886562858e-06,
"loss": 0.9549,
"step": 895
},
{
"epoch": 2.43,
"grad_norm": 1.683670163601615,
"learning_rate": 3.992762161752474e-06,
"loss": 0.9615,
"step": 900
},
{
"epoch": 2.44,
"grad_norm": 2.001515746557216,
"learning_rate": 3.935078511191368e-06,
"loss": 0.9617,
"step": 905
},
{
"epoch": 2.46,
"grad_norm": 2.4485180814449925,
"learning_rate": 3.877542956812137e-06,
"loss": 0.9599,
"step": 910
},
{
"epoch": 2.47,
"grad_norm": 2.2535920157325373,
"learning_rate": 3.820163499951984e-06,
"loss": 0.9485,
"step": 915
},
{
"epoch": 2.48,
"grad_norm": 1.6920809388339335,
"learning_rate": 3.7629481202399886e-06,
"loss": 0.9464,
"step": 920
},
{
"epoch": 2.5,
"grad_norm": 1.9251860711568731,
"learning_rate": 3.705904774487396e-06,
"loss": 0.9565,
"step": 925
},
{
"epoch": 2.51,
"grad_norm": 1.6963234663767837,
"learning_rate": 3.64904139558109e-06,
"loss": 0.9657,
"step": 930
},
{
"epoch": 2.52,
"grad_norm": 1.723192642702987,
"learning_rate": 3.5923658913803726e-06,
"loss": 0.9545,
"step": 935
},
{
"epoch": 2.54,
"grad_norm": 1.857223252079437,
"learning_rate": 3.5358861436172487e-06,
"loss": 0.951,
"step": 940
},
{
"epoch": 2.55,
"grad_norm": 1.9463652794821473,
"learning_rate": 3.47961000680032e-06,
"loss": 0.9545,
"step": 945
},
{
"epoch": 2.56,
"grad_norm": 1.6275095264739814,
"learning_rate": 3.4235453071224882e-06,
"loss": 0.956,
"step": 950
},
{
"epoch": 2.58,
"grad_norm": 2.132166681197278,
"learning_rate": 3.3676998413725726e-06,
"loss": 0.9597,
"step": 955
},
{
"epoch": 2.59,
"grad_norm": 2.389342012426533,
"learning_rate": 3.3120813758510385e-06,
"loss": 0.9444,
"step": 960
},
{
"epoch": 2.6,
"grad_norm": 2.1360938817924424,
"learning_rate": 3.2566976452899507e-06,
"loss": 0.9488,
"step": 965
},
{
"epoch": 2.62,
"grad_norm": 2.1039499268212323,
"learning_rate": 3.2015563517773214e-06,
"loss": 0.953,
"step": 970
},
{
"epoch": 2.63,
"grad_norm": 2.310285226796488,
"learning_rate": 3.1466651636860025e-06,
"loss": 0.9617,
"step": 975
},
{
"epoch": 2.65,
"grad_norm": 2.0128464455024817,
"learning_rate": 3.0920317146072577e-06,
"loss": 0.9516,
"step": 980
},
{
"epoch": 2.66,
"grad_norm": 1.4735182975439758,
"learning_rate": 3.0376636022891813e-06,
"loss": 0.9575,
"step": 985
},
{
"epoch": 2.67,
"grad_norm": 2.318196631109318,
"learning_rate": 2.983568387580094e-06,
"loss": 0.9536,
"step": 990
},
{
"epoch": 2.69,
"grad_norm": 1.616843986566008,
"learning_rate": 2.9297535933770732e-06,
"loss": 0.9474,
"step": 995
},
{
"epoch": 2.7,
"grad_norm": 1.7693581907423008,
"learning_rate": 2.8762267035797607e-06,
"loss": 0.95,
"step": 1000
},
{
"epoch": 2.71,
"grad_norm": 2.122336403201617,
"learning_rate": 2.822995162049599e-06,
"loss": 0.9505,
"step": 1005
},
{
"epoch": 2.73,
"grad_norm": 1.5679869228768337,
"learning_rate": 2.7700663715746213e-06,
"loss": 0.9474,
"step": 1010
},
{
"epoch": 2.74,
"grad_norm": 1.5424846799708574,
"learning_rate": 2.7174476928399685e-06,
"loss": 0.9483,
"step": 1015
},
{
"epoch": 2.75,
"grad_norm": 1.654933401000696,
"learning_rate": 2.66514644340426e-06,
"loss": 0.9521,
"step": 1020
},
{
"epoch": 2.77,
"grad_norm": 1.5993372025284307,
"learning_rate": 2.613169896681949e-06,
"loss": 0.9597,
"step": 1025
},
{
"epoch": 2.78,
"grad_norm": 1.780666444469016,
"learning_rate": 2.5615252809318287e-06,
"loss": 0.9488,
"step": 1030
},
{
"epoch": 2.79,
"grad_norm": 1.5279733243149005,
"learning_rate": 2.5102197782518145e-06,
"loss": 0.9438,
"step": 1035
},
{
"epoch": 2.81,
"grad_norm": 1.8194536821858915,
"learning_rate": 2.4592605235801544e-06,
"loss": 0.943,
"step": 1040
},
{
"epoch": 2.82,
"grad_norm": 1.6480818333921619,
"learning_rate": 2.4086546037031734e-06,
"loss": 0.9437,
"step": 1045
},
{
"epoch": 2.83,
"grad_norm": 1.9968939308919122,
"learning_rate": 2.3584090562697427e-06,
"loss": 0.949,
"step": 1050
},
{
"epoch": 2.85,
"grad_norm": 1.712960437774201,
"learning_rate": 2.3085308688125695e-06,
"loss": 0.9522,
"step": 1055
},
{
"epoch": 2.86,
"grad_norm": 1.7812514943208981,
"learning_rate": 2.2590269777764516e-06,
"loss": 0.9503,
"step": 1060
},
{
"epoch": 2.87,
"grad_norm": 1.4783532304720237,
"learning_rate": 2.2099042675536437e-06,
"loss": 0.9565,
"step": 1065
},
{
"epoch": 2.89,
"grad_norm": 1.4892687284353463,
"learning_rate": 2.161169569526461e-06,
"loss": 0.9481,
"step": 1070
},
{
"epoch": 2.9,
"grad_norm": 1.5738512766968247,
"learning_rate": 2.1128296611172593e-06,
"loss": 0.9483,
"step": 1075
},
{
"epoch": 2.91,
"grad_norm": 1.7245957459593824,
"learning_rate": 2.0648912648459072e-06,
"loss": 0.9424,
"step": 1080
},
{
"epoch": 2.93,
"grad_norm": 1.6116002624575987,
"learning_rate": 2.0173610473949048e-06,
"loss": 0.9462,
"step": 1085
},
{
"epoch": 2.94,
"grad_norm": 1.5813885175238174,
"learning_rate": 1.9702456186822595e-06,
"loss": 0.9443,
"step": 1090
},
{
"epoch": 2.96,
"grad_norm": 1.6220303849517923,
"learning_rate": 1.9235515309422685e-06,
"loss": 0.9428,
"step": 1095
},
{
"epoch": 2.97,
"grad_norm": 1.847653265677833,
"learning_rate": 1.8772852778143064e-06,
"loss": 0.9431,
"step": 1100
},
{
"epoch": 2.98,
"grad_norm": 1.7766441613041704,
"learning_rate": 1.831453293439771e-06,
"loss": 0.9371,
"step": 1105
},
{
"epoch": 3.0,
"grad_norm": 1.7572717020990576,
"learning_rate": 1.7860619515673034e-06,
"loss": 0.9455,
"step": 1110
},
{
"epoch": 3.01,
"grad_norm": 2.0366564271261396,
"learning_rate": 1.7411175646664103e-06,
"loss": 0.9415,
"step": 1115
},
{
"epoch": 3.02,
"grad_norm": 1.641890714375309,
"learning_rate": 1.6966263830495939e-06,
"loss": 0.9245,
"step": 1120
},
{
"epoch": 3.04,
"grad_norm": 1.7389532361568043,
"learning_rate": 1.6525945940031407e-06,
"loss": 0.9387,
"step": 1125
},
{
"epoch": 3.05,
"grad_norm": 1.6927577102388072,
"learning_rate": 1.6090283209266682e-06,
"loss": 0.9358,
"step": 1130
},
{
"epoch": 3.06,
"grad_norm": 1.6810193706370176,
"learning_rate": 1.5659336224815642e-06,
"loss": 0.9302,
"step": 1135
},
{
"epoch": 3.08,
"grad_norm": 1.5836446992301734,
"learning_rate": 1.5233164917484117e-06,
"loss": 0.9339,
"step": 1140
},
{
"epoch": 3.09,
"grad_norm": 1.4925374056585887,
"learning_rate": 1.4811828553935498e-06,
"loss": 0.9359,
"step": 1145
},
{
"epoch": 3.1,
"grad_norm": 1.562133820438609,
"learning_rate": 1.439538572844873e-06,
"loss": 0.9265,
"step": 1150
},
{
"epoch": 3.12,
"grad_norm": 1.508715392161331,
"learning_rate": 1.3983894354769616e-06,
"loss": 0.94,
"step": 1155
},
{
"epoch": 3.13,
"grad_norm": 1.4833918032553421,
"learning_rate": 1.3577411658056965e-06,
"loss": 0.9359,
"step": 1160
},
{
"epoch": 3.14,
"grad_norm": 1.7381479633485923,
"learning_rate": 1.3175994166924394e-06,
"loss": 0.9292,
"step": 1165
},
{
"epoch": 3.16,
"grad_norm": 2.4773729170453276,
"learning_rate": 1.2779697705579058e-06,
"loss": 0.935,
"step": 1170
},
{
"epoch": 3.17,
"grad_norm": 1.6423953500818698,
"learning_rate": 1.2388577386058248e-06,
"loss": 0.9281,
"step": 1175
},
{
"epoch": 3.18,
"grad_norm": 1.5747180314126341,
"learning_rate": 1.2002687600565138e-06,
"loss": 0.9251,
"step": 1180
},
{
"epoch": 3.2,
"grad_norm": 1.4796953478538386,
"learning_rate": 1.1622082013904535e-06,
"loss": 0.9309,
"step": 1185
},
{
"epoch": 3.21,
"grad_norm": 1.9087071208773028,
"learning_rate": 1.1246813556019925e-06,
"loss": 0.9302,
"step": 1190
},
{
"epoch": 3.23,
"grad_norm": 1.441416955211069,
"learning_rate": 1.0876934414632523e-06,
"loss": 0.9222,
"step": 1195
},
{
"epoch": 3.24,
"grad_norm": 1.4528421452118758,
"learning_rate": 1.0512496027983715e-06,
"loss": 0.9358,
"step": 1200
},
{
"epoch": 3.25,
"grad_norm": 1.4391309054269417,
"learning_rate": 1.0153549077681617e-06,
"loss": 0.9294,
"step": 1205
},
{
"epoch": 3.27,
"grad_norm": 1.5484810522299401,
"learning_rate": 9.80014348165298e-07,
"loss": 0.9258,
"step": 1210
},
{
"epoch": 3.28,
"grad_norm": 1.5093687112295802,
"learning_rate": 9.452328387201104e-07,
"loss": 0.9294,
"step": 1215
},
{
"epoch": 3.29,
"grad_norm": 1.4811632926742155,
"learning_rate": 9.110152164171127e-07,
"loss": 0.9291,
"step": 1220
},
{
"epoch": 3.31,
"grad_norm": 1.6851965695398232,
"learning_rate": 8.773662398223276e-07,
"loss": 0.9315,
"step": 1225
},
{
"epoch": 3.32,
"grad_norm": 1.5680181843321963,
"learning_rate": 8.44290588421533e-07,
"loss": 0.9245,
"step": 1230
},
{
"epoch": 3.33,
"grad_norm": 1.501501151688511,
"learning_rate": 8.117928619694848e-07,
"loss": 0.9312,
"step": 1235
},
{
"epoch": 3.35,
"grad_norm": 1.5069447398439229,
"learning_rate": 7.798775798502484e-07,
"loss": 0.9256,
"step": 1240
},
{
"epoch": 3.36,
"grad_norm": 1.4179041700918098,
"learning_rate": 7.485491804486972e-07,
"loss": 0.9209,
"step": 1245
},
{
"epoch": 3.37,
"grad_norm": 1.388023240525021,
"learning_rate": 7.178120205332717e-07,
"loss": 0.9311,
"step": 1250
},
{
"epoch": 3.39,
"grad_norm": 1.6540883448168693,
"learning_rate": 6.876703746500984e-07,
"loss": 0.9398,
"step": 1255
},
{
"epoch": 3.4,
"grad_norm": 1.3848004961206983,
"learning_rate": 6.581284345285371e-07,
"loss": 0.9316,
"step": 1260
},
{
"epoch": 3.41,
"grad_norm": 1.3665324107772734,
"learning_rate": 6.291903084982481e-07,
"loss": 0.9301,
"step": 1265
},
{
"epoch": 3.43,
"grad_norm": 1.4952001545175686,
"learning_rate": 6.008600209178539e-07,
"loss": 0.9375,
"step": 1270
},
{
"epoch": 3.44,
"grad_norm": 1.4437175986820594,
"learning_rate": 5.7314151161528e-07,
"loss": 0.9324,
"step": 1275
},
{
"epoch": 3.45,
"grad_norm": 1.3632138123352606,
"learning_rate": 5.460386353398583e-07,
"loss": 0.9355,
"step": 1280
},
{
"epoch": 3.47,
"grad_norm": 1.4355913142058574,
"learning_rate": 5.195551612262478e-07,
"loss": 0.9369,
"step": 1285
},
{
"epoch": 3.48,
"grad_norm": 1.3593932016675367,
"learning_rate": 4.936947722702762e-07,
"loss": 0.9248,
"step": 1290
},
{
"epoch": 3.5,
"grad_norm": 1.5313367509871634,
"learning_rate": 4.6846106481675035e-07,
"loss": 0.9313,
"step": 1295
},
{
"epoch": 3.51,
"grad_norm": 1.5679931446767343,
"learning_rate": 4.43857548059321e-07,
"loss": 0.9241,
"step": 1300
},
{
"epoch": 3.52,
"grad_norm": 1.4231649052268165,
"learning_rate": 4.198876435524718e-07,
"loss": 0.9288,
"step": 1305
},
{
"epoch": 3.54,
"grad_norm": 1.5147664043720925,
"learning_rate": 3.9655468473568436e-07,
"loss": 0.937,
"step": 1310
},
{
"epoch": 3.55,
"grad_norm": 1.505384912577037,
"learning_rate": 3.7386191646987094e-07,
"loss": 0.9362,
"step": 1315
},
{
"epoch": 3.56,
"grad_norm": 1.7241739930760707,
"learning_rate": 3.51812494586114e-07,
"loss": 0.9209,
"step": 1320
},
{
"epoch": 3.58,
"grad_norm": 1.498479993175917,
"learning_rate": 3.3040948544679817e-07,
"loss": 0.9323,
"step": 1325
},
{
"epoch": 3.59,
"grad_norm": 1.464257082461974,
"learning_rate": 3.096558655191706e-07,
"loss": 0.9365,
"step": 1330
},
{
"epoch": 3.6,
"grad_norm": 1.4384497143831931,
"learning_rate": 2.895545209614176e-07,
"loss": 0.939,
"step": 1335
},
{
"epoch": 3.62,
"grad_norm": 1.4541548130541968,
"learning_rate": 2.701082472212879e-07,
"loss": 0.9277,
"step": 1340
},
{
"epoch": 3.63,
"grad_norm": 1.5175629961859558,
"learning_rate": 2.5131974864734063e-07,
"loss": 0.9279,
"step": 1345
},
{
"epoch": 3.64,
"grad_norm": 1.4611673173663144,
"learning_rate": 2.331916381128535e-07,
"loss": 0.9268,
"step": 1350
},
{
"epoch": 3.66,
"grad_norm": 1.3907174934591902,
"learning_rate": 2.157264366524603e-07,
"loss": 0.9297,
"step": 1355
},
{
"epoch": 3.67,
"grad_norm": 1.6602489752653944,
"learning_rate": 1.989265731115525e-07,
"loss": 0.9279,
"step": 1360
},
{
"epoch": 3.68,
"grad_norm": 1.4335900513136002,
"learning_rate": 1.827943838085111e-07,
"loss": 0.9261,
"step": 1365
},
{
"epoch": 3.7,
"grad_norm": 1.4382877939912813,
"learning_rate": 1.6733211220979316e-07,
"loss": 0.9286,
"step": 1370
},
{
"epoch": 3.71,
"grad_norm": 1.5442925793363935,
"learning_rate": 1.5254190861794415e-07,
"loss": 0.931,
"step": 1375
},
{
"epoch": 3.72,
"grad_norm": 1.474749185950876,
"learning_rate": 1.3842582987255494e-07,
"loss": 0.9237,
"step": 1380
},
{
"epoch": 3.74,
"grad_norm": 1.3949775304121632,
"learning_rate": 1.2498583906422779e-07,
"loss": 0.9264,
"step": 1385
},
{
"epoch": 3.75,
"grad_norm": 1.4864944753529736,
"learning_rate": 1.1222380526156929e-07,
"loss": 0.9337,
"step": 1390
},
{
"epoch": 3.77,
"grad_norm": 1.4352950913827418,
"learning_rate": 1.0014150325126315e-07,
"loss": 0.9296,
"step": 1395
},
{
"epoch": 3.78,
"grad_norm": 1.54647479187705,
"learning_rate": 8.874061329125939e-08,
"loss": 0.9284,
"step": 1400
},
{
"epoch": 3.79,
"grad_norm": 1.5426799677559493,
"learning_rate": 7.802272087709951e-08,
"loss": 0.9267,
"step": 1405
},
{
"epoch": 3.81,
"grad_norm": 1.4345720592945528,
"learning_rate": 6.798931652142738e-08,
"loss": 0.9412,
"step": 1410
},
{
"epoch": 3.82,
"grad_norm": 1.4666166802740588,
"learning_rate": 5.864179554670724e-08,
"loss": 0.9274,
"step": 1415
},
{
"epoch": 3.83,
"grad_norm": 1.37454276267336,
"learning_rate": 4.998145789118114e-08,
"loss": 0.9347,
"step": 1420
},
{
"epoch": 3.85,
"grad_norm": 1.3537939594957207,
"learning_rate": 4.2009507928084116e-08,
"loss": 0.9374,
"step": 1425
},
{
"epoch": 3.86,
"grad_norm": 1.5201092358761166,
"learning_rate": 3.4727054298161475e-08,
"loss": 0.9278,
"step": 1430
},
{
"epoch": 3.87,
"grad_norm": 1.3548157698669074,
"learning_rate": 2.8135109755487723e-08,
"loss": 0.9229,
"step": 1435
},
{
"epoch": 3.89,
"grad_norm": 1.413302347919726,
"learning_rate": 2.223459102662695e-08,
"loss": 0.9341,
"step": 1440
},
{
"epoch": 3.9,
"grad_norm": 1.3765482397349185,
"learning_rate": 1.7026318683147082e-08,
"loss": 0.9289,
"step": 1445
},
{
"epoch": 3.91,
"grad_norm": 1.3538302103790776,
"learning_rate": 1.2511017027501682e-08,
"loss": 0.9233,
"step": 1450
},
{
"epoch": 3.93,
"grad_norm": 1.466960745073582,
"learning_rate": 8.689313992306104e-09,
"loss": 0.9303,
"step": 1455
},
{
"epoch": 3.94,
"grad_norm": 1.3714566713014886,
"learning_rate": 5.561741053010661e-09,
"loss": 0.936,
"step": 1460
},
{
"epoch": 3.95,
"grad_norm": 1.4859645171436624,
"learning_rate": 3.1287331539903155e-09,
"loss": 0.9247,
"step": 1465
},
{
"epoch": 3.97,
"grad_norm": 1.386772570990769,
"learning_rate": 1.3906286480563913e-09,
"loss": 0.9195,
"step": 1470
},
{
"epoch": 3.98,
"grad_norm": 1.3873824123701393,
"learning_rate": 3.4766924940476954e-10,
"loss": 0.9374,
"step": 1475
},
{
"epoch": 3.99,
"grad_norm": 1.492034850579766,
"learning_rate": 0.0,
"loss": 0.9213,
"step": 1480
},
{
"epoch": 3.99,
"step": 1480,
"total_flos": 1630317777518592.0,
"train_loss": 1.3370767467730753,
"train_runtime": 16871.6289,
"train_samples_per_second": 44.965,
"train_steps_per_second": 0.088
}
],
"logging_steps": 5,
"max_steps": 1480,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 500,
"total_flos": 1630317777518592.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}