labor-llama3-8b-Instruct-20241002 / trainer_state.json
clinno's picture
Upload folder using huggingface_hub
2b16540 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 11.989100817438691,
"eval_steps": 1000,
"global_step": 3300,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.03633060853769301,
"grad_norm": 5.34375,
"learning_rate": 3.0303030303030305e-07,
"loss": 2.0607,
"step": 10
},
{
"epoch": 0.07266121707538602,
"grad_norm": 5.90625,
"learning_rate": 6.060606060606061e-07,
"loss": 2.033,
"step": 20
},
{
"epoch": 0.10899182561307902,
"grad_norm": 5.46875,
"learning_rate": 9.090909090909091e-07,
"loss": 2.0622,
"step": 30
},
{
"epoch": 0.14532243415077203,
"grad_norm": 4.96875,
"learning_rate": 1.2121212121212122e-06,
"loss": 1.9896,
"step": 40
},
{
"epoch": 0.18165304268846502,
"grad_norm": 15.625,
"learning_rate": 1.5151515151515152e-06,
"loss": 1.9998,
"step": 50
},
{
"epoch": 0.21798365122615804,
"grad_norm": 18.0,
"learning_rate": 1.8181818181818183e-06,
"loss": 1.9608,
"step": 60
},
{
"epoch": 0.254314259763851,
"grad_norm": 15.625,
"learning_rate": 2.1212121212121216e-06,
"loss": 1.9272,
"step": 70
},
{
"epoch": 0.29064486830154407,
"grad_norm": 15.5625,
"learning_rate": 2.4242424242424244e-06,
"loss": 1.9031,
"step": 80
},
{
"epoch": 0.32697547683923706,
"grad_norm": 14.1875,
"learning_rate": 2.7272727272727272e-06,
"loss": 1.8012,
"step": 90
},
{
"epoch": 0.36330608537693004,
"grad_norm": 3.421875,
"learning_rate": 3.0303030303030305e-06,
"loss": 1.7756,
"step": 100
},
{
"epoch": 0.3996366939146231,
"grad_norm": 3.09375,
"learning_rate": 3.3333333333333333e-06,
"loss": 1.6836,
"step": 110
},
{
"epoch": 0.4359673024523161,
"grad_norm": 2.46875,
"learning_rate": 3.6363636363636366e-06,
"loss": 1.7013,
"step": 120
},
{
"epoch": 0.47229791099000906,
"grad_norm": 2.859375,
"learning_rate": 3.93939393939394e-06,
"loss": 1.6328,
"step": 130
},
{
"epoch": 0.508628519527702,
"grad_norm": 2.6875,
"learning_rate": 4.242424242424243e-06,
"loss": 1.6888,
"step": 140
},
{
"epoch": 0.5449591280653951,
"grad_norm": 2.828125,
"learning_rate": 4.5454545454545455e-06,
"loss": 1.6583,
"step": 150
},
{
"epoch": 0.5812897366030881,
"grad_norm": 2.0625,
"learning_rate": 4.848484848484849e-06,
"loss": 1.5485,
"step": 160
},
{
"epoch": 0.6176203451407811,
"grad_norm": 2.1875,
"learning_rate": 5.151515151515152e-06,
"loss": 1.5464,
"step": 170
},
{
"epoch": 0.6539509536784741,
"grad_norm": 2.25,
"learning_rate": 5.4545454545454545e-06,
"loss": 1.5506,
"step": 180
},
{
"epoch": 0.6902815622161671,
"grad_norm": 2.4375,
"learning_rate": 5.7575757575757586e-06,
"loss": 1.5301,
"step": 190
},
{
"epoch": 0.7266121707538601,
"grad_norm": 2.4375,
"learning_rate": 6.060606060606061e-06,
"loss": 1.5116,
"step": 200
},
{
"epoch": 0.7629427792915532,
"grad_norm": 2.609375,
"learning_rate": 6.363636363636364e-06,
"loss": 1.4874,
"step": 210
},
{
"epoch": 0.7992733878292462,
"grad_norm": 2.3125,
"learning_rate": 6.666666666666667e-06,
"loss": 1.5056,
"step": 220
},
{
"epoch": 0.8356039963669392,
"grad_norm": 2.21875,
"learning_rate": 6.969696969696971e-06,
"loss": 1.5036,
"step": 230
},
{
"epoch": 0.8719346049046321,
"grad_norm": 2.359375,
"learning_rate": 7.272727272727273e-06,
"loss": 1.4625,
"step": 240
},
{
"epoch": 0.9082652134423251,
"grad_norm": 1.921875,
"learning_rate": 7.5757575757575764e-06,
"loss": 1.4439,
"step": 250
},
{
"epoch": 0.9445958219800181,
"grad_norm": 2.265625,
"learning_rate": 7.87878787878788e-06,
"loss": 1.4025,
"step": 260
},
{
"epoch": 0.9809264305177112,
"grad_norm": 2.609375,
"learning_rate": 8.181818181818183e-06,
"loss": 1.4267,
"step": 270
},
{
"epoch": 1.017257039055404,
"grad_norm": 1.84375,
"learning_rate": 8.484848484848486e-06,
"loss": 1.3484,
"step": 280
},
{
"epoch": 1.0535876475930972,
"grad_norm": 2.0,
"learning_rate": 8.787878787878788e-06,
"loss": 1.3809,
"step": 290
},
{
"epoch": 1.0899182561307903,
"grad_norm": 2.1875,
"learning_rate": 9.090909090909091e-06,
"loss": 1.3901,
"step": 300
},
{
"epoch": 1.1262488646684832,
"grad_norm": 1.8984375,
"learning_rate": 9.393939393939396e-06,
"loss": 1.3203,
"step": 310
},
{
"epoch": 1.1625794732061763,
"grad_norm": 1.9296875,
"learning_rate": 9.696969696969698e-06,
"loss": 1.3295,
"step": 320
},
{
"epoch": 1.1989100817438691,
"grad_norm": 1.9140625,
"learning_rate": 1e-05,
"loss": 1.3936,
"step": 330
},
{
"epoch": 1.2352406902815622,
"grad_norm": 2.5,
"learning_rate": 9.999720280459576e-06,
"loss": 1.3483,
"step": 340
},
{
"epoch": 1.2715712988192553,
"grad_norm": 1.7890625,
"learning_rate": 9.99888115313551e-06,
"loss": 1.3384,
"step": 350
},
{
"epoch": 1.3079019073569482,
"grad_norm": 1.921875,
"learning_rate": 9.997482711915926e-06,
"loss": 1.2888,
"step": 360
},
{
"epoch": 1.344232515894641,
"grad_norm": 1.9453125,
"learning_rate": 9.99552511326936e-06,
"loss": 1.3144,
"step": 370
},
{
"epoch": 1.3805631244323342,
"grad_norm": 1.796875,
"learning_rate": 9.993008576227248e-06,
"loss": 1.2905,
"step": 380
},
{
"epoch": 1.4168937329700273,
"grad_norm": 1.8203125,
"learning_rate": 9.989933382359423e-06,
"loss": 1.2816,
"step": 390
},
{
"epoch": 1.4532243415077202,
"grad_norm": 1.6953125,
"learning_rate": 9.986299875742612e-06,
"loss": 1.3315,
"step": 400
},
{
"epoch": 1.4895549500454133,
"grad_norm": 1.7734375,
"learning_rate": 9.982108462921938e-06,
"loss": 1.2984,
"step": 410
},
{
"epoch": 1.5258855585831061,
"grad_norm": 2.765625,
"learning_rate": 9.977359612865424e-06,
"loss": 1.2677,
"step": 420
},
{
"epoch": 1.5622161671207992,
"grad_norm": 1.859375,
"learning_rate": 9.972053856911534e-06,
"loss": 1.3098,
"step": 430
},
{
"epoch": 1.5985467756584923,
"grad_norm": 1.84375,
"learning_rate": 9.966191788709716e-06,
"loss": 1.3503,
"step": 440
},
{
"epoch": 1.6348773841961854,
"grad_norm": 2.609375,
"learning_rate": 9.959774064153977e-06,
"loss": 1.3301,
"step": 450
},
{
"epoch": 1.6712079927338783,
"grad_norm": 1.8359375,
"learning_rate": 9.952801401309504e-06,
"loss": 1.2454,
"step": 460
},
{
"epoch": 1.7075386012715712,
"grad_norm": 1.8515625,
"learning_rate": 9.945274580332316e-06,
"loss": 1.239,
"step": 470
},
{
"epoch": 1.7438692098092643,
"grad_norm": 1.8046875,
"learning_rate": 9.937194443381972e-06,
"loss": 1.2823,
"step": 480
},
{
"epoch": 1.7801998183469574,
"grad_norm": 2.203125,
"learning_rate": 9.928561894527354e-06,
"loss": 1.2774,
"step": 490
},
{
"epoch": 1.8165304268846503,
"grad_norm": 2.15625,
"learning_rate": 9.919377899645497e-06,
"loss": 1.2215,
"step": 500
},
{
"epoch": 1.8528610354223434,
"grad_norm": 1.671875,
"learning_rate": 9.909643486313533e-06,
"loss": 1.2509,
"step": 510
},
{
"epoch": 1.8891916439600362,
"grad_norm": 1.8828125,
"learning_rate": 9.899359743693715e-06,
"loss": 1.2656,
"step": 520
},
{
"epoch": 1.9255222524977293,
"grad_norm": 1.6328125,
"learning_rate": 9.888527822411543e-06,
"loss": 1.186,
"step": 530
},
{
"epoch": 1.9618528610354224,
"grad_norm": 1.6953125,
"learning_rate": 9.877148934427037e-06,
"loss": 1.2468,
"step": 540
},
{
"epoch": 1.9981834695731153,
"grad_norm": 1.59375,
"learning_rate": 9.86522435289912e-06,
"loss": 1.2471,
"step": 550
},
{
"epoch": 2.034514078110808,
"grad_norm": 1.9140625,
"learning_rate": 9.85275541204318e-06,
"loss": 1.1688,
"step": 560
},
{
"epoch": 2.0708446866485013,
"grad_norm": 2.046875,
"learning_rate": 9.839743506981783e-06,
"loss": 1.1216,
"step": 570
},
{
"epoch": 2.1071752951861944,
"grad_norm": 1.6328125,
"learning_rate": 9.826190093588564e-06,
"loss": 1.1333,
"step": 580
},
{
"epoch": 2.1435059037238875,
"grad_norm": 1.6953125,
"learning_rate": 9.812096688325354e-06,
"loss": 1.1148,
"step": 590
},
{
"epoch": 2.1798365122615806,
"grad_norm": 1.796875,
"learning_rate": 9.797464868072489e-06,
"loss": 1.1331,
"step": 600
},
{
"epoch": 2.2161671207992732,
"grad_norm": 1.890625,
"learning_rate": 9.78229626995238e-06,
"loss": 1.1931,
"step": 610
},
{
"epoch": 2.2524977293369663,
"grad_norm": 1.578125,
"learning_rate": 9.766592591146353e-06,
"loss": 1.146,
"step": 620
},
{
"epoch": 2.2888283378746594,
"grad_norm": 1.7890625,
"learning_rate": 9.750355588704728e-06,
"loss": 1.1397,
"step": 630
},
{
"epoch": 2.3251589464123525,
"grad_norm": 1.8359375,
"learning_rate": 9.733587079350254e-06,
"loss": 1.0941,
"step": 640
},
{
"epoch": 2.3614895549500456,
"grad_norm": 1.7265625,
"learning_rate": 9.716288939274818e-06,
"loss": 1.101,
"step": 650
},
{
"epoch": 2.3978201634877383,
"grad_norm": 2.703125,
"learning_rate": 9.698463103929542e-06,
"loss": 1.0632,
"step": 660
},
{
"epoch": 2.4341507720254314,
"grad_norm": 2.5625,
"learning_rate": 9.680111567808212e-06,
"loss": 1.1315,
"step": 670
},
{
"epoch": 2.4704813805631245,
"grad_norm": 1.890625,
"learning_rate": 9.66123638422413e-06,
"loss": 1.0989,
"step": 680
},
{
"epoch": 2.5068119891008176,
"grad_norm": 1.8828125,
"learning_rate": 9.641839665080363e-06,
"loss": 1.0829,
"step": 690
},
{
"epoch": 2.5431425976385107,
"grad_norm": 1.9140625,
"learning_rate": 9.621923580633462e-06,
"loss": 1.1155,
"step": 700
},
{
"epoch": 2.5794732061762033,
"grad_norm": 1.5546875,
"learning_rate": 9.601490359250616e-06,
"loss": 1.1248,
"step": 710
},
{
"epoch": 2.6158038147138964,
"grad_norm": 1.7890625,
"learning_rate": 9.580542287160348e-06,
"loss": 1.1391,
"step": 720
},
{
"epoch": 2.6521344232515895,
"grad_norm": 1.734375,
"learning_rate": 9.559081708196696e-06,
"loss": 1.0538,
"step": 730
},
{
"epoch": 2.688465031789282,
"grad_norm": 1.953125,
"learning_rate": 9.537111023536973e-06,
"loss": 1.055,
"step": 740
},
{
"epoch": 2.7247956403269757,
"grad_norm": 1.6328125,
"learning_rate": 9.514632691433108e-06,
"loss": 1.1769,
"step": 750
},
{
"epoch": 2.7611262488646684,
"grad_norm": 2.609375,
"learning_rate": 9.491649226936586e-06,
"loss": 1.0949,
"step": 760
},
{
"epoch": 2.7974568574023615,
"grad_norm": 1.9609375,
"learning_rate": 9.468163201617063e-06,
"loss": 1.1308,
"step": 770
},
{
"epoch": 2.8337874659400546,
"grad_norm": 1.953125,
"learning_rate": 9.444177243274619e-06,
"loss": 1.1283,
"step": 780
},
{
"epoch": 2.8701180744777472,
"grad_norm": 1.703125,
"learning_rate": 9.419694035645753e-06,
"loss": 1.0907,
"step": 790
},
{
"epoch": 2.9064486830154403,
"grad_norm": 1.5,
"learning_rate": 9.394716318103098e-06,
"loss": 1.1043,
"step": 800
},
{
"epoch": 2.9427792915531334,
"grad_norm": 2.3125,
"learning_rate": 9.369246885348926e-06,
"loss": 1.0676,
"step": 810
},
{
"epoch": 2.9791099000908265,
"grad_norm": 1.65625,
"learning_rate": 9.343288587102444e-06,
"loss": 1.0642,
"step": 820
},
{
"epoch": 3.0154405086285196,
"grad_norm": 1.5234375,
"learning_rate": 9.316844327780955e-06,
"loss": 1.0437,
"step": 830
},
{
"epoch": 3.0517711171662127,
"grad_norm": 1.6953125,
"learning_rate": 9.289917066174887e-06,
"loss": 0.9874,
"step": 840
},
{
"epoch": 3.0881017257039054,
"grad_norm": 1.578125,
"learning_rate": 9.262509815116732e-06,
"loss": 0.9582,
"step": 850
},
{
"epoch": 3.1244323342415985,
"grad_norm": 1.8046875,
"learning_rate": 9.234625641143962e-06,
"loss": 0.9471,
"step": 860
},
{
"epoch": 3.1607629427792916,
"grad_norm": 2.015625,
"learning_rate": 9.206267664155906e-06,
"loss": 0.9424,
"step": 870
},
{
"epoch": 3.1970935513169847,
"grad_norm": 1.625,
"learning_rate": 9.177439057064684e-06,
"loss": 0.972,
"step": 880
},
{
"epoch": 3.2334241598546773,
"grad_norm": 1.4921875,
"learning_rate": 9.148143045440181e-06,
"loss": 0.9729,
"step": 890
},
{
"epoch": 3.2697547683923704,
"grad_norm": 1.8203125,
"learning_rate": 9.118382907149164e-06,
"loss": 0.9558,
"step": 900
},
{
"epoch": 3.3060853769300635,
"grad_norm": 1.453125,
"learning_rate": 9.088161971988517e-06,
"loss": 0.9518,
"step": 910
},
{
"epoch": 3.3424159854677566,
"grad_norm": 1.3671875,
"learning_rate": 9.057483621312671e-06,
"loss": 0.9694,
"step": 920
},
{
"epoch": 3.3787465940054497,
"grad_norm": 1.15625,
"learning_rate": 9.026351287655294e-06,
"loss": 0.9457,
"step": 930
},
{
"epoch": 3.4150772025431424,
"grad_norm": 1.3359375,
"learning_rate": 8.994768454345207e-06,
"loss": 0.9585,
"step": 940
},
{
"epoch": 3.4514078110808355,
"grad_norm": 1.3203125,
"learning_rate": 8.96273865511666e-06,
"loss": 0.9479,
"step": 950
},
{
"epoch": 3.4877384196185286,
"grad_norm": 1.3203125,
"learning_rate": 8.930265473713939e-06,
"loss": 0.9105,
"step": 960
},
{
"epoch": 3.5240690281562217,
"grad_norm": 1.3828125,
"learning_rate": 8.897352543490396e-06,
"loss": 0.9795,
"step": 970
},
{
"epoch": 3.560399636693915,
"grad_norm": 1.21875,
"learning_rate": 8.864003547001916e-06,
"loss": 0.9403,
"step": 980
},
{
"epoch": 3.5967302452316074,
"grad_norm": 1.3203125,
"learning_rate": 8.83022221559489e-06,
"loss": 0.925,
"step": 990
},
{
"epoch": 3.6330608537693005,
"grad_norm": 1.25,
"learning_rate": 8.796012328988716e-06,
"loss": 0.9225,
"step": 1000
},
{
"epoch": 3.6330608537693005,
"eval_loss": 1.1292678117752075,
"eval_runtime": 10.2075,
"eval_samples_per_second": 24.002,
"eval_steps_per_second": 24.002,
"step": 1000
},
{
"epoch": 3.6693914623069936,
"grad_norm": 1.171875,
"learning_rate": 8.7613777148529e-06,
"loss": 0.9778,
"step": 1010
},
{
"epoch": 3.7057220708446867,
"grad_norm": 1.234375,
"learning_rate": 8.726322248378775e-06,
"loss": 0.8997,
"step": 1020
},
{
"epoch": 3.74205267938238,
"grad_norm": 1.1328125,
"learning_rate": 8.690849851845933e-06,
"loss": 0.9874,
"step": 1030
},
{
"epoch": 3.7783832879200725,
"grad_norm": 1.453125,
"learning_rate": 8.65496449418336e-06,
"loss": 0.9101,
"step": 1040
},
{
"epoch": 3.8147138964577656,
"grad_norm": 1.3515625,
"learning_rate": 8.61867019052535e-06,
"loss": 0.9101,
"step": 1050
},
{
"epoch": 3.8510445049954587,
"grad_norm": 1.25,
"learning_rate": 8.581971001762287e-06,
"loss": 0.9549,
"step": 1060
},
{
"epoch": 3.887375113533152,
"grad_norm": 1.375,
"learning_rate": 8.54487103408625e-06,
"loss": 0.9486,
"step": 1070
},
{
"epoch": 3.923705722070845,
"grad_norm": 1.1953125,
"learning_rate": 8.507374438531606e-06,
"loss": 0.9476,
"step": 1080
},
{
"epoch": 3.9600363306085375,
"grad_norm": 1.2421875,
"learning_rate": 8.469485410510545e-06,
"loss": 0.9283,
"step": 1090
},
{
"epoch": 3.9963669391462306,
"grad_norm": 1.2109375,
"learning_rate": 8.43120818934367e-06,
"loss": 0.9616,
"step": 1100
},
{
"epoch": 4.032697547683924,
"grad_norm": 1.5,
"learning_rate": 8.392547057785662e-06,
"loss": 0.8816,
"step": 1110
},
{
"epoch": 4.069028156221616,
"grad_norm": 1.0546875,
"learning_rate": 8.353506341546106e-06,
"loss": 0.8704,
"step": 1120
},
{
"epoch": 4.10535876475931,
"grad_norm": 1.0859375,
"learning_rate": 8.314090408805481e-06,
"loss": 0.8515,
"step": 1130
},
{
"epoch": 4.141689373297003,
"grad_norm": 1.125,
"learning_rate": 8.274303669726427e-06,
"loss": 0.7986,
"step": 1140
},
{
"epoch": 4.178019981834696,
"grad_norm": 0.9765625,
"learning_rate": 8.234150575960288e-06,
"loss": 0.859,
"step": 1150
},
{
"epoch": 4.214350590372389,
"grad_norm": 1.0625,
"learning_rate": 8.193635620149041e-06,
"loss": 0.8537,
"step": 1160
},
{
"epoch": 4.2506811989100814,
"grad_norm": 1.0625,
"learning_rate": 8.152763335422612e-06,
"loss": 0.8513,
"step": 1170
},
{
"epoch": 4.287011807447775,
"grad_norm": 1.0234375,
"learning_rate": 8.111538294891684e-06,
"loss": 0.8318,
"step": 1180
},
{
"epoch": 4.323342415985468,
"grad_norm": 1.03125,
"learning_rate": 8.06996511113601e-06,
"loss": 0.846,
"step": 1190
},
{
"epoch": 4.359673024523161,
"grad_norm": 1.0,
"learning_rate": 8.028048435688333e-06,
"loss": 0.8667,
"step": 1200
},
{
"epoch": 4.396003633060854,
"grad_norm": 0.9921875,
"learning_rate": 7.985792958513932e-06,
"loss": 0.8715,
"step": 1210
},
{
"epoch": 4.4323342415985465,
"grad_norm": 0.9296875,
"learning_rate": 7.943203407485864e-06,
"loss": 0.848,
"step": 1220
},
{
"epoch": 4.46866485013624,
"grad_norm": 0.94921875,
"learning_rate": 7.900284547855992e-06,
"loss": 0.8859,
"step": 1230
},
{
"epoch": 4.504995458673933,
"grad_norm": 0.98046875,
"learning_rate": 7.857041181721788e-06,
"loss": 0.8575,
"step": 1240
},
{
"epoch": 4.541326067211626,
"grad_norm": 0.94140625,
"learning_rate": 7.813478147489052e-06,
"loss": 0.8496,
"step": 1250
},
{
"epoch": 4.577656675749319,
"grad_norm": 0.92578125,
"learning_rate": 7.769600319330553e-06,
"loss": 0.8166,
"step": 1260
},
{
"epoch": 4.6139872842870115,
"grad_norm": 1.15625,
"learning_rate": 7.725412606640658e-06,
"loss": 0.8191,
"step": 1270
},
{
"epoch": 4.650317892824705,
"grad_norm": 1.046875,
"learning_rate": 7.680919953486047e-06,
"loss": 0.833,
"step": 1280
},
{
"epoch": 4.686648501362398,
"grad_norm": 1.0390625,
"learning_rate": 7.636127338052513e-06,
"loss": 0.839,
"step": 1290
},
{
"epoch": 4.722979109900091,
"grad_norm": 1.2421875,
"learning_rate": 7.5910397720879785e-06,
"loss": 0.807,
"step": 1300
},
{
"epoch": 4.759309718437784,
"grad_norm": 0.984375,
"learning_rate": 7.545662300341736e-06,
"loss": 0.8215,
"step": 1310
},
{
"epoch": 4.795640326975477,
"grad_norm": 1.03125,
"learning_rate": 7.500000000000001e-06,
"loss": 0.8584,
"step": 1320
},
{
"epoch": 4.83197093551317,
"grad_norm": 1.0546875,
"learning_rate": 7.454057980117842e-06,
"loss": 0.7955,
"step": 1330
},
{
"epoch": 4.868301544050863,
"grad_norm": 1.140625,
"learning_rate": 7.407841381047533e-06,
"loss": 0.8482,
"step": 1340
},
{
"epoch": 4.904632152588556,
"grad_norm": 1.234375,
"learning_rate": 7.361355373863415e-06,
"loss": 0.8457,
"step": 1350
},
{
"epoch": 4.940962761126249,
"grad_norm": 0.91796875,
"learning_rate": 7.314605159783313e-06,
"loss": 0.8705,
"step": 1360
},
{
"epoch": 4.977293369663942,
"grad_norm": 1.0234375,
"learning_rate": 7.2675959695865896e-06,
"loss": 0.829,
"step": 1370
},
{
"epoch": 5.013623978201635,
"grad_norm": 0.9375,
"learning_rate": 7.2203330630288714e-06,
"loss": 0.8391,
"step": 1380
},
{
"epoch": 5.049954586739328,
"grad_norm": 0.9921875,
"learning_rate": 7.172821728253563e-06,
"loss": 0.7915,
"step": 1390
},
{
"epoch": 5.0862851952770205,
"grad_norm": 1.1171875,
"learning_rate": 7.1250672812001505e-06,
"loss": 0.8199,
"step": 1400
},
{
"epoch": 5.122615803814714,
"grad_norm": 1.0,
"learning_rate": 7.0770750650094335e-06,
"loss": 0.8065,
"step": 1410
},
{
"epoch": 5.158946412352407,
"grad_norm": 1.2578125,
"learning_rate": 7.02885044942567e-06,
"loss": 0.8098,
"step": 1420
},
{
"epoch": 5.1952770208901,
"grad_norm": 1.0625,
"learning_rate": 6.980398830195785e-06,
"loss": 0.7675,
"step": 1430
},
{
"epoch": 5.231607629427793,
"grad_norm": 1.125,
"learning_rate": 6.931725628465643e-06,
"loss": 0.7747,
"step": 1440
},
{
"epoch": 5.2679382379654855,
"grad_norm": 1.375,
"learning_rate": 6.882836290173493e-06,
"loss": 0.7623,
"step": 1450
},
{
"epoch": 5.304268846503179,
"grad_norm": 1.2109375,
"learning_rate": 6.833736285440632e-06,
"loss": 0.8113,
"step": 1460
},
{
"epoch": 5.340599455040872,
"grad_norm": 1.3359375,
"learning_rate": 6.78443110795936e-06,
"loss": 0.8411,
"step": 1470
},
{
"epoch": 5.376930063578565,
"grad_norm": 1.1171875,
"learning_rate": 6.734926274378313e-06,
"loss": 0.7592,
"step": 1480
},
{
"epoch": 5.413260672116258,
"grad_norm": 1.234375,
"learning_rate": 6.685227323685209e-06,
"loss": 0.7776,
"step": 1490
},
{
"epoch": 5.449591280653951,
"grad_norm": 1.46875,
"learning_rate": 6.635339816587109e-06,
"loss": 0.7727,
"step": 1500
},
{
"epoch": 5.485921889191644,
"grad_norm": 1.40625,
"learning_rate": 6.5852693348882345e-06,
"loss": 0.7539,
"step": 1510
},
{
"epoch": 5.522252497729337,
"grad_norm": 1.4609375,
"learning_rate": 6.535021480865439e-06,
"loss": 0.8292,
"step": 1520
},
{
"epoch": 5.55858310626703,
"grad_norm": 1.8359375,
"learning_rate": 6.484601876641375e-06,
"loss": 0.7463,
"step": 1530
},
{
"epoch": 5.594913714804723,
"grad_norm": 1.3515625,
"learning_rate": 6.434016163555452e-06,
"loss": 0.7744,
"step": 1540
},
{
"epoch": 5.631244323342416,
"grad_norm": 2.578125,
"learning_rate": 6.383270001532636e-06,
"loss": 0.7868,
"step": 1550
},
{
"epoch": 5.667574931880109,
"grad_norm": 2.65625,
"learning_rate": 6.332369068450175e-06,
"loss": 0.7485,
"step": 1560
},
{
"epoch": 5.703905540417802,
"grad_norm": 2.34375,
"learning_rate": 6.2813190595023135e-06,
"loss": 0.7783,
"step": 1570
},
{
"epoch": 5.740236148955495,
"grad_norm": 2.40625,
"learning_rate": 6.230125686563068e-06,
"loss": 0.7769,
"step": 1580
},
{
"epoch": 5.776566757493188,
"grad_norm": 2.390625,
"learning_rate": 6.178794677547138e-06,
"loss": 0.7977,
"step": 1590
},
{
"epoch": 5.812897366030881,
"grad_norm": 3.65625,
"learning_rate": 6.127331775769023e-06,
"loss": 0.8101,
"step": 1600
},
{
"epoch": 5.849227974568574,
"grad_norm": 5.4375,
"learning_rate": 6.07574273930042e-06,
"loss": 0.8397,
"step": 1610
},
{
"epoch": 5.885558583106267,
"grad_norm": 4.78125,
"learning_rate": 6.024033340325954e-06,
"loss": 0.8191,
"step": 1620
},
{
"epoch": 5.9218891916439595,
"grad_norm": 4.625,
"learning_rate": 5.972209364497355e-06,
"loss": 0.7843,
"step": 1630
},
{
"epoch": 5.958219800181653,
"grad_norm": 5.4375,
"learning_rate": 5.920276610286102e-06,
"loss": 0.8099,
"step": 1640
},
{
"epoch": 5.994550408719346,
"grad_norm": 10.875,
"learning_rate": 5.8682408883346535e-06,
"loss": 0.8025,
"step": 1650
},
{
"epoch": 6.030881017257039,
"grad_norm": 10.25,
"learning_rate": 5.816108020806297e-06,
"loss": 0.7426,
"step": 1660
},
{
"epoch": 6.067211625794732,
"grad_norm": 9.8125,
"learning_rate": 5.763883840733736e-06,
"loss": 0.7439,
"step": 1670
},
{
"epoch": 6.1035422343324255,
"grad_norm": 8.4375,
"learning_rate": 5.711574191366427e-06,
"loss": 0.7617,
"step": 1680
},
{
"epoch": 6.139872842870118,
"grad_norm": 8.4375,
"learning_rate": 5.659184925516802e-06,
"loss": 0.7333,
"step": 1690
},
{
"epoch": 6.176203451407811,
"grad_norm": 2.453125,
"learning_rate": 5.60672190490541e-06,
"loss": 0.7747,
"step": 1700
},
{
"epoch": 6.212534059945504,
"grad_norm": 2.5625,
"learning_rate": 5.5541909995050554e-06,
"loss": 0.7469,
"step": 1710
},
{
"epoch": 6.248864668483197,
"grad_norm": 3.015625,
"learning_rate": 5.5015980868840254e-06,
"loss": 0.7507,
"step": 1720
},
{
"epoch": 6.28519527702089,
"grad_norm": 2.546875,
"learning_rate": 5.448949051548459e-06,
"loss": 0.7774,
"step": 1730
},
{
"epoch": 6.321525885558583,
"grad_norm": 2.921875,
"learning_rate": 5.396249784283943e-06,
"loss": 0.773,
"step": 1740
},
{
"epoch": 6.357856494096276,
"grad_norm": 2.21875,
"learning_rate": 5.343506181496405e-06,
"loss": 0.7333,
"step": 1750
},
{
"epoch": 6.394187102633969,
"grad_norm": 2.28125,
"learning_rate": 5.290724144552379e-06,
"loss": 0.7352,
"step": 1760
},
{
"epoch": 6.430517711171662,
"grad_norm": 2.3125,
"learning_rate": 5.237909579118713e-06,
"loss": 0.68,
"step": 1770
},
{
"epoch": 6.466848319709355,
"grad_norm": 2.296875,
"learning_rate": 5.185068394501791e-06,
"loss": 0.7279,
"step": 1780
},
{
"epoch": 6.503178928247048,
"grad_norm": 2.53125,
"learning_rate": 5.132206502986368e-06,
"loss": 0.7547,
"step": 1790
},
{
"epoch": 6.539509536784741,
"grad_norm": 2.421875,
"learning_rate": 5.07932981917404e-06,
"loss": 0.7015,
"step": 1800
},
{
"epoch": 6.575840145322434,
"grad_norm": 2.4375,
"learning_rate": 5.026444259321489e-06,
"loss": 0.7614,
"step": 1810
},
{
"epoch": 6.612170753860127,
"grad_norm": 2.171875,
"learning_rate": 4.973555740678512e-06,
"loss": 0.7953,
"step": 1820
},
{
"epoch": 6.64850136239782,
"grad_norm": 2.40625,
"learning_rate": 4.9206701808259605e-06,
"loss": 0.736,
"step": 1830
},
{
"epoch": 6.684831970935513,
"grad_norm": 2.28125,
"learning_rate": 4.867793497013634e-06,
"loss": 0.7502,
"step": 1840
},
{
"epoch": 6.721162579473206,
"grad_norm": 2.390625,
"learning_rate": 4.81493160549821e-06,
"loss": 0.7496,
"step": 1850
},
{
"epoch": 6.7574931880108995,
"grad_norm": 2.359375,
"learning_rate": 4.762090420881289e-06,
"loss": 0.7888,
"step": 1860
},
{
"epoch": 6.793823796548592,
"grad_norm": 2.078125,
"learning_rate": 4.7092758554476215e-06,
"loss": 0.7238,
"step": 1870
},
{
"epoch": 6.830154405086285,
"grad_norm": 2.1875,
"learning_rate": 4.6564938185035954e-06,
"loss": 0.7484,
"step": 1880
},
{
"epoch": 6.866485013623978,
"grad_norm": 2.0625,
"learning_rate": 4.603750215716057e-06,
"loss": 0.7835,
"step": 1890
},
{
"epoch": 6.902815622161671,
"grad_norm": 2.53125,
"learning_rate": 4.551050948451542e-06,
"loss": 0.6878,
"step": 1900
},
{
"epoch": 6.9391462306993645,
"grad_norm": 1.953125,
"learning_rate": 4.498401913115975e-06,
"loss": 0.7528,
"step": 1910
},
{
"epoch": 6.975476839237057,
"grad_norm": 2.0625,
"learning_rate": 4.445809000494945e-06,
"loss": 0.7486,
"step": 1920
},
{
"epoch": 7.01180744777475,
"grad_norm": 2.15625,
"learning_rate": 4.393278095094591e-06,
"loss": 0.7948,
"step": 1930
},
{
"epoch": 7.048138056312443,
"grad_norm": 2.03125,
"learning_rate": 4.340815074483199e-06,
"loss": 0.6872,
"step": 1940
},
{
"epoch": 7.084468664850136,
"grad_norm": 1.953125,
"learning_rate": 4.2884258086335755e-06,
"loss": 0.7289,
"step": 1950
},
{
"epoch": 7.12079927338783,
"grad_norm": 2.15625,
"learning_rate": 4.2361161592662655e-06,
"loss": 0.612,
"step": 1960
},
{
"epoch": 7.157129881925522,
"grad_norm": 2.1875,
"learning_rate": 4.183891979193703e-06,
"loss": 0.7135,
"step": 1970
},
{
"epoch": 7.193460490463215,
"grad_norm": 2.28125,
"learning_rate": 4.131759111665349e-06,
"loss": 0.6932,
"step": 1980
},
{
"epoch": 7.229791099000908,
"grad_norm": 1.9453125,
"learning_rate": 4.079723389713899e-06,
"loss": 0.705,
"step": 1990
},
{
"epoch": 7.266121707538601,
"grad_norm": 2.09375,
"learning_rate": 4.027790635502646e-06,
"loss": 0.7145,
"step": 2000
},
{
"epoch": 7.266121707538601,
"eval_loss": 1.1055254936218262,
"eval_runtime": 10.2576,
"eval_samples_per_second": 23.885,
"eval_steps_per_second": 23.885,
"step": 2000
},
{
"epoch": 7.302452316076295,
"grad_norm": 1.921875,
"learning_rate": 3.975966659674048e-06,
"loss": 0.6724,
"step": 2010
},
{
"epoch": 7.338782924613987,
"grad_norm": 2.3125,
"learning_rate": 3.924257260699583e-06,
"loss": 0.6468,
"step": 2020
},
{
"epoch": 7.37511353315168,
"grad_norm": 2.140625,
"learning_rate": 3.872668224230979e-06,
"loss": 0.6792,
"step": 2030
},
{
"epoch": 7.4114441416893735,
"grad_norm": 2.265625,
"learning_rate": 3.821205322452863e-06,
"loss": 0.723,
"step": 2040
},
{
"epoch": 7.447774750227066,
"grad_norm": 1.984375,
"learning_rate": 3.769874313436933e-06,
"loss": 0.6608,
"step": 2050
},
{
"epoch": 7.48410535876476,
"grad_norm": 2.21875,
"learning_rate": 3.7186809404976877e-06,
"loss": 0.6679,
"step": 2060
},
{
"epoch": 7.520435967302452,
"grad_norm": 2.078125,
"learning_rate": 3.667630931549826e-06,
"loss": 0.6944,
"step": 2070
},
{
"epoch": 7.556766575840145,
"grad_norm": 2.15625,
"learning_rate": 3.6167299984673655e-06,
"loss": 0.7172,
"step": 2080
},
{
"epoch": 7.5930971843778385,
"grad_norm": 2.265625,
"learning_rate": 3.5659838364445505e-06,
"loss": 0.7029,
"step": 2090
},
{
"epoch": 7.629427792915531,
"grad_norm": 2.078125,
"learning_rate": 3.5153981233586277e-06,
"loss": 0.6549,
"step": 2100
},
{
"epoch": 7.665758401453225,
"grad_norm": 2.1875,
"learning_rate": 3.4649785191345613e-06,
"loss": 0.7342,
"step": 2110
},
{
"epoch": 7.702089009990917,
"grad_norm": 2.09375,
"learning_rate": 3.4147306651117663e-06,
"loss": 0.6246,
"step": 2120
},
{
"epoch": 7.73841961852861,
"grad_norm": 1.90625,
"learning_rate": 3.3646601834128924e-06,
"loss": 0.6587,
"step": 2130
},
{
"epoch": 7.774750227066304,
"grad_norm": 1.953125,
"learning_rate": 3.3147726763147913e-06,
"loss": 0.6682,
"step": 2140
},
{
"epoch": 7.811080835603996,
"grad_norm": 2.421875,
"learning_rate": 3.2650737256216885e-06,
"loss": 0.6854,
"step": 2150
},
{
"epoch": 7.84741144414169,
"grad_norm": 2.1875,
"learning_rate": 3.2155688920406415e-06,
"loss": 0.6908,
"step": 2160
},
{
"epoch": 7.883742052679382,
"grad_norm": 2.046875,
"learning_rate": 3.16626371455937e-06,
"loss": 0.6875,
"step": 2170
},
{
"epoch": 7.920072661217075,
"grad_norm": 2.28125,
"learning_rate": 3.1171637098265063e-06,
"loss": 0.6732,
"step": 2180
},
{
"epoch": 7.956403269754769,
"grad_norm": 1.9765625,
"learning_rate": 3.0682743715343565e-06,
"loss": 0.742,
"step": 2190
},
{
"epoch": 7.992733878292461,
"grad_norm": 2.796875,
"learning_rate": 3.019601169804216e-06,
"loss": 0.6791,
"step": 2200
},
{
"epoch": 8.029064486830155,
"grad_norm": 1.796875,
"learning_rate": 2.9711495505743317e-06,
"loss": 0.654,
"step": 2210
},
{
"epoch": 8.065395095367847,
"grad_norm": 2.0,
"learning_rate": 2.9229249349905686e-06,
"loss": 0.705,
"step": 2220
},
{
"epoch": 8.10172570390554,
"grad_norm": 2.0,
"learning_rate": 2.8749327187998516e-06,
"loss": 0.6067,
"step": 2230
},
{
"epoch": 8.138056312443233,
"grad_norm": 2.03125,
"learning_rate": 2.8271782717464413e-06,
"loss": 0.6043,
"step": 2240
},
{
"epoch": 8.174386920980927,
"grad_norm": 1.6875,
"learning_rate": 2.7796669369711294e-06,
"loss": 0.6292,
"step": 2250
},
{
"epoch": 8.21071752951862,
"grad_norm": 2.0625,
"learning_rate": 2.7324040304134125e-06,
"loss": 0.6041,
"step": 2260
},
{
"epoch": 8.247048138056313,
"grad_norm": 2.296875,
"learning_rate": 2.685394840216688e-06,
"loss": 0.6222,
"step": 2270
},
{
"epoch": 8.283378746594005,
"grad_norm": 2.046875,
"learning_rate": 2.6386446261365874e-06,
"loss": 0.6586,
"step": 2280
},
{
"epoch": 8.319709355131698,
"grad_norm": 2.015625,
"learning_rate": 2.5921586189524694e-06,
"loss": 0.6612,
"step": 2290
},
{
"epoch": 8.356039963669392,
"grad_norm": 2.0,
"learning_rate": 2.5459420198821604e-06,
"loss": 0.6693,
"step": 2300
},
{
"epoch": 8.392370572207085,
"grad_norm": 2.125,
"learning_rate": 2.5000000000000015e-06,
"loss": 0.6244,
"step": 2310
},
{
"epoch": 8.428701180744778,
"grad_norm": 2.09375,
"learning_rate": 2.454337699658267e-06,
"loss": 0.6312,
"step": 2320
},
{
"epoch": 8.46503178928247,
"grad_norm": 2.09375,
"learning_rate": 2.4089602279120224e-06,
"loss": 0.6729,
"step": 2330
},
{
"epoch": 8.501362397820163,
"grad_norm": 1.8125,
"learning_rate": 2.363872661947488e-06,
"loss": 0.6276,
"step": 2340
},
{
"epoch": 8.537693006357856,
"grad_norm": 1.90625,
"learning_rate": 2.319080046513954e-06,
"loss": 0.6292,
"step": 2350
},
{
"epoch": 8.57402361489555,
"grad_norm": 2.125,
"learning_rate": 2.274587393359342e-06,
"loss": 0.6593,
"step": 2360
},
{
"epoch": 8.610354223433243,
"grad_norm": 1.875,
"learning_rate": 2.230399680669449e-06,
"loss": 0.6302,
"step": 2370
},
{
"epoch": 8.646684831970935,
"grad_norm": 1.796875,
"learning_rate": 2.1865218525109496e-06,
"loss": 0.6481,
"step": 2380
},
{
"epoch": 8.683015440508628,
"grad_norm": 2.15625,
"learning_rate": 2.1429588182782147e-06,
"loss": 0.6175,
"step": 2390
},
{
"epoch": 8.719346049046322,
"grad_norm": 1.953125,
"learning_rate": 2.09971545214401e-06,
"loss": 0.6144,
"step": 2400
},
{
"epoch": 8.755676657584015,
"grad_norm": 1.6875,
"learning_rate": 2.0567965925141366e-06,
"loss": 0.6572,
"step": 2410
},
{
"epoch": 8.792007266121708,
"grad_norm": 1.890625,
"learning_rate": 2.0142070414860704e-06,
"loss": 0.654,
"step": 2420
},
{
"epoch": 8.8283378746594,
"grad_norm": 1.8359375,
"learning_rate": 1.971951564311668e-06,
"loss": 0.621,
"step": 2430
},
{
"epoch": 8.864668483197093,
"grad_norm": 1.78125,
"learning_rate": 1.9300348888639915e-06,
"loss": 0.62,
"step": 2440
},
{
"epoch": 8.900999091734786,
"grad_norm": 1.6640625,
"learning_rate": 1.8884617051083183e-06,
"loss": 0.5858,
"step": 2450
},
{
"epoch": 8.93732970027248,
"grad_norm": 1.578125,
"learning_rate": 1.8472366645773892e-06,
"loss": 0.6467,
"step": 2460
},
{
"epoch": 8.973660308810173,
"grad_norm": 1.921875,
"learning_rate": 1.8063643798509594e-06,
"loss": 0.6138,
"step": 2470
},
{
"epoch": 9.009990917347865,
"grad_norm": 1.65625,
"learning_rate": 1.7658494240397127e-06,
"loss": 0.6388,
"step": 2480
},
{
"epoch": 9.046321525885558,
"grad_norm": 1.4453125,
"learning_rate": 1.7256963302735752e-06,
"loss": 0.5733,
"step": 2490
},
{
"epoch": 9.082652134423252,
"grad_norm": 1.421875,
"learning_rate": 1.68590959119452e-06,
"loss": 0.6066,
"step": 2500
},
{
"epoch": 9.118982742960945,
"grad_norm": 1.2890625,
"learning_rate": 1.646493658453896e-06,
"loss": 0.5849,
"step": 2510
},
{
"epoch": 9.155313351498638,
"grad_norm": 1.34375,
"learning_rate": 1.6074529422143398e-06,
"loss": 0.6136,
"step": 2520
},
{
"epoch": 9.19164396003633,
"grad_norm": 1.34375,
"learning_rate": 1.5687918106563326e-06,
"loss": 0.5891,
"step": 2530
},
{
"epoch": 9.227974568574023,
"grad_norm": 1.5234375,
"learning_rate": 1.5305145894894547e-06,
"loss": 0.6065,
"step": 2540
},
{
"epoch": 9.264305177111716,
"grad_norm": 1.4453125,
"learning_rate": 1.4926255614683931e-06,
"loss": 0.5854,
"step": 2550
},
{
"epoch": 9.30063578564941,
"grad_norm": 1.5703125,
"learning_rate": 1.4551289659137497e-06,
"loss": 0.6023,
"step": 2560
},
{
"epoch": 9.336966394187103,
"grad_norm": 1.4296875,
"learning_rate": 1.4180289982377138e-06,
"loss": 0.6052,
"step": 2570
},
{
"epoch": 9.373297002724795,
"grad_norm": 1.4453125,
"learning_rate": 1.3813298094746491e-06,
"loss": 0.5806,
"step": 2580
},
{
"epoch": 9.409627611262488,
"grad_norm": 1.421875,
"learning_rate": 1.345035505816642e-06,
"loss": 0.6058,
"step": 2590
},
{
"epoch": 9.44595821980018,
"grad_norm": 1.3671875,
"learning_rate": 1.3091501481540676e-06,
"loss": 0.6339,
"step": 2600
},
{
"epoch": 9.482288828337875,
"grad_norm": 1.2734375,
"learning_rate": 1.2736777516212267e-06,
"loss": 0.6058,
"step": 2610
},
{
"epoch": 9.518619436875568,
"grad_norm": 1.546875,
"learning_rate": 1.238622285147103e-06,
"loss": 0.6186,
"step": 2620
},
{
"epoch": 9.55495004541326,
"grad_norm": 1.3203125,
"learning_rate": 1.2039876710112847e-06,
"loss": 0.5913,
"step": 2630
},
{
"epoch": 9.591280653950953,
"grad_norm": 1.359375,
"learning_rate": 1.1697777844051105e-06,
"loss": 0.663,
"step": 2640
},
{
"epoch": 9.627611262488646,
"grad_norm": 1.2109375,
"learning_rate": 1.135996452998085e-06,
"loss": 0.6278,
"step": 2650
},
{
"epoch": 9.66394187102634,
"grad_norm": 1.390625,
"learning_rate": 1.1026474565096068e-06,
"loss": 0.6074,
"step": 2660
},
{
"epoch": 9.700272479564033,
"grad_norm": 1.4609375,
"learning_rate": 1.0697345262860638e-06,
"loss": 0.6545,
"step": 2670
},
{
"epoch": 9.736603088101726,
"grad_norm": 1.421875,
"learning_rate": 1.0372613448833429e-06,
"loss": 0.6141,
"step": 2680
},
{
"epoch": 9.772933696639418,
"grad_norm": 1.3125,
"learning_rate": 1.0052315456547934e-06,
"loss": 0.5699,
"step": 2690
},
{
"epoch": 9.809264305177111,
"grad_norm": 1.296875,
"learning_rate": 9.73648712344707e-07,
"loss": 0.579,
"step": 2700
},
{
"epoch": 9.845594913714805,
"grad_norm": 1.2421875,
"learning_rate": 9.425163786873292e-07,
"loss": 0.61,
"step": 2710
},
{
"epoch": 9.881925522252498,
"grad_norm": 1.296875,
"learning_rate": 9.118380280114858e-07,
"loss": 0.6106,
"step": 2720
},
{
"epoch": 9.91825613079019,
"grad_norm": 1.296875,
"learning_rate": 8.816170928508367e-07,
"loss": 0.6066,
"step": 2730
},
{
"epoch": 9.954586739327883,
"grad_norm": 1.0703125,
"learning_rate": 8.518569545598198e-07,
"loss": 0.6094,
"step": 2740
},
{
"epoch": 9.990917347865576,
"grad_norm": 1.1640625,
"learning_rate": 8.225609429353187e-07,
"loss": 0.6383,
"step": 2750
},
{
"epoch": 10.02724795640327,
"grad_norm": 1.203125,
"learning_rate": 7.937323358440935e-07,
"loss": 0.6017,
"step": 2760
},
{
"epoch": 10.063578564940963,
"grad_norm": 1.03125,
"learning_rate": 7.653743588560387e-07,
"loss": 0.6005,
"step": 2770
},
{
"epoch": 10.099909173478656,
"grad_norm": 1.03125,
"learning_rate": 7.374901848832683e-07,
"loss": 0.647,
"step": 2780
},
{
"epoch": 10.136239782016348,
"grad_norm": 1.0625,
"learning_rate": 7.100829338251147e-07,
"loss": 0.5894,
"step": 2790
},
{
"epoch": 10.172570390554041,
"grad_norm": 0.953125,
"learning_rate": 6.831556722190453e-07,
"loss": 0.6075,
"step": 2800
},
{
"epoch": 10.208900999091735,
"grad_norm": 0.97265625,
"learning_rate": 6.567114128975571e-07,
"loss": 0.5531,
"step": 2810
},
{
"epoch": 10.245231607629428,
"grad_norm": 1.1171875,
"learning_rate": 6.307531146510754e-07,
"loss": 0.6052,
"step": 2820
},
{
"epoch": 10.28156221616712,
"grad_norm": 1.0234375,
"learning_rate": 6.052836818969027e-07,
"loss": 0.5833,
"step": 2830
},
{
"epoch": 10.317892824704813,
"grad_norm": 1.1484375,
"learning_rate": 5.803059643542491e-07,
"loss": 0.5934,
"step": 2840
},
{
"epoch": 10.354223433242506,
"grad_norm": 0.9609375,
"learning_rate": 5.558227567253832e-07,
"loss": 0.5931,
"step": 2850
},
{
"epoch": 10.3905540417802,
"grad_norm": 1.15625,
"learning_rate": 5.318367983829393e-07,
"loss": 0.6253,
"step": 2860
},
{
"epoch": 10.426884650317893,
"grad_norm": 1.0390625,
"learning_rate": 5.083507730634152e-07,
"loss": 0.61,
"step": 2870
},
{
"epoch": 10.463215258855586,
"grad_norm": 1.0234375,
"learning_rate": 4.853673085668947e-07,
"loss": 0.6329,
"step": 2880
},
{
"epoch": 10.499545867393278,
"grad_norm": 0.96484375,
"learning_rate": 4.628889764630279e-07,
"loss": 0.5733,
"step": 2890
},
{
"epoch": 10.535876475930971,
"grad_norm": 1.0,
"learning_rate": 4.4091829180330503e-07,
"loss": 0.5922,
"step": 2900
},
{
"epoch": 10.572207084468666,
"grad_norm": 1.0078125,
"learning_rate": 4.194577128396521e-07,
"loss": 0.5843,
"step": 2910
},
{
"epoch": 10.608537693006358,
"grad_norm": 1.0546875,
"learning_rate": 3.985096407493838e-07,
"loss": 0.6028,
"step": 2920
},
{
"epoch": 10.64486830154405,
"grad_norm": 0.984375,
"learning_rate": 3.7807641936653984e-07,
"loss": 0.5767,
"step": 2930
},
{
"epoch": 10.681198910081743,
"grad_norm": 1.0625,
"learning_rate": 3.581603349196372e-07,
"loss": 0.6062,
"step": 2940
},
{
"epoch": 10.717529518619436,
"grad_norm": 1.0,
"learning_rate": 3.3876361577587115e-07,
"loss": 0.5978,
"step": 2950
},
{
"epoch": 10.75386012715713,
"grad_norm": 1.046875,
"learning_rate": 3.1988843219178776e-07,
"loss": 0.5984,
"step": 2960
},
{
"epoch": 10.790190735694823,
"grad_norm": 1.1953125,
"learning_rate": 3.015368960704584e-07,
"loss": 0.5761,
"step": 2970
},
{
"epoch": 10.826521344232516,
"grad_norm": 1.0078125,
"learning_rate": 2.8371106072518194e-07,
"loss": 0.5988,
"step": 2980
},
{
"epoch": 10.862851952770209,
"grad_norm": 1.015625,
"learning_rate": 2.664129206497479e-07,
"loss": 0.593,
"step": 2990
},
{
"epoch": 10.899182561307901,
"grad_norm": 1.125,
"learning_rate": 2.4964441129527337e-07,
"loss": 0.6401,
"step": 3000
},
{
"epoch": 10.899182561307901,
"eval_loss": 1.12338387966156,
"eval_runtime": 10.3307,
"eval_samples_per_second": 23.716,
"eval_steps_per_second": 23.716,
"step": 3000
},
{
"epoch": 10.935513169845596,
"grad_norm": 0.921875,
"learning_rate": 2.3340740885364922e-07,
"loss": 0.5798,
"step": 3010
},
{
"epoch": 10.971843778383288,
"grad_norm": 1.0546875,
"learning_rate": 2.1770373004762035e-07,
"loss": 0.5684,
"step": 3020
},
{
"epoch": 11.008174386920981,
"grad_norm": 1.171875,
"learning_rate": 2.0253513192751374e-07,
"loss": 0.619,
"step": 3030
},
{
"epoch": 11.044504995458674,
"grad_norm": 1.125,
"learning_rate": 1.8790331167464758e-07,
"loss": 0.5739,
"step": 3040
},
{
"epoch": 11.080835603996366,
"grad_norm": 1.140625,
"learning_rate": 1.738099064114368e-07,
"loss": 0.6039,
"step": 3050
},
{
"epoch": 11.11716621253406,
"grad_norm": 1.3359375,
"learning_rate": 1.6025649301821877e-07,
"loss": 0.6057,
"step": 3060
},
{
"epoch": 11.153496821071753,
"grad_norm": 1.2109375,
"learning_rate": 1.4724458795681962e-07,
"loss": 0.5966,
"step": 3070
},
{
"epoch": 11.189827429609446,
"grad_norm": 1.109375,
"learning_rate": 1.3477564710088097e-07,
"loss": 0.5987,
"step": 3080
},
{
"epoch": 11.226158038147139,
"grad_norm": 1.1796875,
"learning_rate": 1.2285106557296479e-07,
"loss": 0.614,
"step": 3090
},
{
"epoch": 11.262488646684831,
"grad_norm": 1.25,
"learning_rate": 1.1147217758845752e-07,
"loss": 0.5854,
"step": 3100
},
{
"epoch": 11.298819255222526,
"grad_norm": 1.2734375,
"learning_rate": 1.0064025630628583e-07,
"loss": 0.5866,
"step": 3110
},
{
"epoch": 11.335149863760218,
"grad_norm": 1.25,
"learning_rate": 9.035651368646647e-08,
"loss": 0.6412,
"step": 3120
},
{
"epoch": 11.371480472297911,
"grad_norm": 1.203125,
"learning_rate": 8.06221003545038e-08,
"loss": 0.5935,
"step": 3130
},
{
"epoch": 11.407811080835604,
"grad_norm": 1.3125,
"learning_rate": 7.143810547264762e-08,
"loss": 0.6097,
"step": 3140
},
{
"epoch": 11.444141689373296,
"grad_norm": 2.1875,
"learning_rate": 6.280555661802857e-08,
"loss": 0.5737,
"step": 3150
},
{
"epoch": 11.48047229791099,
"grad_norm": 2.09375,
"learning_rate": 5.472541966768552e-08,
"loss": 0.5908,
"step": 3160
},
{
"epoch": 11.516802906448683,
"grad_norm": 2.078125,
"learning_rate": 4.719859869049659e-08,
"loss": 0.5884,
"step": 3170
},
{
"epoch": 11.553133514986376,
"grad_norm": 2.125,
"learning_rate": 4.02259358460233e-08,
"loss": 0.6072,
"step": 3180
},
{
"epoch": 11.589464123524069,
"grad_norm": 2.1875,
"learning_rate": 3.3808211290284886e-08,
"loss": 0.6049,
"step": 3190
},
{
"epoch": 11.625794732061761,
"grad_norm": 3.734375,
"learning_rate": 2.7946143088466437e-08,
"loss": 0.6021,
"step": 3200
},
{
"epoch": 11.662125340599456,
"grad_norm": 4.34375,
"learning_rate": 2.264038713457706e-08,
"loss": 0.6059,
"step": 3210
},
{
"epoch": 11.698455949137148,
"grad_norm": 3.9375,
"learning_rate": 1.789153707806357e-08,
"loss": 0.601,
"step": 3220
},
{
"epoch": 11.734786557674841,
"grad_norm": 4.0625,
"learning_rate": 1.3700124257388092e-08,
"loss": 0.5657,
"step": 3230
},
{
"epoch": 11.771117166212534,
"grad_norm": 4.25,
"learning_rate": 1.006661764057837e-08,
"loss": 0.6255,
"step": 3240
},
{
"epoch": 11.807447774750226,
"grad_norm": 7.65625,
"learning_rate": 6.991423772753636e-09,
"loss": 0.6209,
"step": 3250
},
{
"epoch": 11.84377838328792,
"grad_norm": 7.90625,
"learning_rate": 4.474886730641004e-09,
"loss": 0.6111,
"step": 3260
},
{
"epoch": 11.880108991825614,
"grad_norm": 8.5625,
"learning_rate": 2.5172880840745873e-09,
"loss": 0.6328,
"step": 3270
},
{
"epoch": 11.916439600363306,
"grad_norm": 8.375,
"learning_rate": 1.118846864490708e-09,
"loss": 0.5511,
"step": 3280
},
{
"epoch": 11.952770208900999,
"grad_norm": 8.4375,
"learning_rate": 2.797195404247166e-10,
"loss": 0.5795,
"step": 3290
},
{
"epoch": 11.989100817438691,
"grad_norm": 2.59375,
"learning_rate": 0.0,
"loss": 0.5714,
"step": 3300
},
{
"epoch": 11.989100817438691,
"step": 3300,
"total_flos": 3.144153968877896e+17,
"train_loss": 0.880180758562955,
"train_runtime": 3585.8754,
"train_samples_per_second": 7.369,
"train_steps_per_second": 0.92
}
],
"logging_steps": 10,
"max_steps": 3300,
"num_input_tokens_seen": 0,
"num_train_epochs": 12,
"save_steps": 0,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 3.144153968877896e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}