hp_ablations_mistral_bsz256 / trainer_state.json
sedrickkeh's picture
End of training
1874780 verified
raw
history blame
48.1 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.9984810126582278,
"eval_steps": 500,
"global_step": 2961,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.010126582278481013,
"grad_norm": 2.9319232619463147,
"learning_rate": 5e-06,
"loss": 0.754,
"step": 10
},
{
"epoch": 0.020253164556962026,
"grad_norm": 2.3527842138237443,
"learning_rate": 5e-06,
"loss": 0.6565,
"step": 20
},
{
"epoch": 0.030379746835443037,
"grad_norm": 2.218097297827581,
"learning_rate": 5e-06,
"loss": 0.6343,
"step": 30
},
{
"epoch": 0.04050632911392405,
"grad_norm": 1.728885416020883,
"learning_rate": 5e-06,
"loss": 0.6266,
"step": 40
},
{
"epoch": 0.05063291139240506,
"grad_norm": 1.7600150721761239,
"learning_rate": 5e-06,
"loss": 0.6225,
"step": 50
},
{
"epoch": 0.060759493670886074,
"grad_norm": 2.036486573113631,
"learning_rate": 5e-06,
"loss": 0.6185,
"step": 60
},
{
"epoch": 0.07088607594936709,
"grad_norm": 1.8472990337897077,
"learning_rate": 5e-06,
"loss": 0.6137,
"step": 70
},
{
"epoch": 0.0810126582278481,
"grad_norm": 2.0268825324116926,
"learning_rate": 5e-06,
"loss": 0.6035,
"step": 80
},
{
"epoch": 0.09113924050632911,
"grad_norm": 1.7502041363801772,
"learning_rate": 5e-06,
"loss": 0.6063,
"step": 90
},
{
"epoch": 0.10126582278481013,
"grad_norm": 1.7706135514845762,
"learning_rate": 5e-06,
"loss": 0.5954,
"step": 100
},
{
"epoch": 0.11139240506329114,
"grad_norm": 1.6647485123629138,
"learning_rate": 5e-06,
"loss": 0.5933,
"step": 110
},
{
"epoch": 0.12151898734177215,
"grad_norm": 2.299944971561371,
"learning_rate": 5e-06,
"loss": 0.598,
"step": 120
},
{
"epoch": 0.13164556962025317,
"grad_norm": 1.4323042297049071,
"learning_rate": 5e-06,
"loss": 0.597,
"step": 130
},
{
"epoch": 0.14177215189873418,
"grad_norm": 1.5735470897186044,
"learning_rate": 5e-06,
"loss": 0.5859,
"step": 140
},
{
"epoch": 0.1518987341772152,
"grad_norm": 1.4396445107803788,
"learning_rate": 5e-06,
"loss": 0.5833,
"step": 150
},
{
"epoch": 0.1620253164556962,
"grad_norm": 1.6048875360767698,
"learning_rate": 5e-06,
"loss": 0.5926,
"step": 160
},
{
"epoch": 0.17215189873417722,
"grad_norm": 1.7629611019629856,
"learning_rate": 5e-06,
"loss": 0.5878,
"step": 170
},
{
"epoch": 0.18227848101265823,
"grad_norm": 1.692513854306383,
"learning_rate": 5e-06,
"loss": 0.5875,
"step": 180
},
{
"epoch": 0.19240506329113924,
"grad_norm": 2.4645583630800614,
"learning_rate": 5e-06,
"loss": 0.5908,
"step": 190
},
{
"epoch": 0.20253164556962025,
"grad_norm": 1.6304964629686653,
"learning_rate": 5e-06,
"loss": 0.5887,
"step": 200
},
{
"epoch": 0.21265822784810126,
"grad_norm": 2.117270092101592,
"learning_rate": 5e-06,
"loss": 0.5872,
"step": 210
},
{
"epoch": 0.22278481012658227,
"grad_norm": 1.5677168838868496,
"learning_rate": 5e-06,
"loss": 0.575,
"step": 220
},
{
"epoch": 0.23291139240506328,
"grad_norm": 1.6515647911636342,
"learning_rate": 5e-06,
"loss": 0.5747,
"step": 230
},
{
"epoch": 0.2430379746835443,
"grad_norm": 1.4834458457301771,
"learning_rate": 5e-06,
"loss": 0.5834,
"step": 240
},
{
"epoch": 0.25316455696202533,
"grad_norm": 2.090691823913493,
"learning_rate": 5e-06,
"loss": 0.5793,
"step": 250
},
{
"epoch": 0.26329113924050634,
"grad_norm": 1.4520944372541345,
"learning_rate": 5e-06,
"loss": 0.5779,
"step": 260
},
{
"epoch": 0.27341772151898736,
"grad_norm": 1.7905780272726646,
"learning_rate": 5e-06,
"loss": 0.5637,
"step": 270
},
{
"epoch": 0.28354430379746837,
"grad_norm": 1.6753642665165924,
"learning_rate": 5e-06,
"loss": 0.5745,
"step": 280
},
{
"epoch": 0.2936708860759494,
"grad_norm": 1.8117100564747939,
"learning_rate": 5e-06,
"loss": 0.577,
"step": 290
},
{
"epoch": 0.3037974683544304,
"grad_norm": 1.5160896257954382,
"learning_rate": 5e-06,
"loss": 0.5821,
"step": 300
},
{
"epoch": 0.3139240506329114,
"grad_norm": 1.5338602444115943,
"learning_rate": 5e-06,
"loss": 0.574,
"step": 310
},
{
"epoch": 0.3240506329113924,
"grad_norm": 1.5768816348730115,
"learning_rate": 5e-06,
"loss": 0.5831,
"step": 320
},
{
"epoch": 0.3341772151898734,
"grad_norm": 1.5251244093370149,
"learning_rate": 5e-06,
"loss": 0.5793,
"step": 330
},
{
"epoch": 0.34430379746835443,
"grad_norm": 1.5422210280401378,
"learning_rate": 5e-06,
"loss": 0.579,
"step": 340
},
{
"epoch": 0.35443037974683544,
"grad_norm": 1.4422891794630028,
"learning_rate": 5e-06,
"loss": 0.5794,
"step": 350
},
{
"epoch": 0.36455696202531646,
"grad_norm": 1.778098375508798,
"learning_rate": 5e-06,
"loss": 0.5702,
"step": 360
},
{
"epoch": 0.37468354430379747,
"grad_norm": 1.4276847831429058,
"learning_rate": 5e-06,
"loss": 0.5762,
"step": 370
},
{
"epoch": 0.3848101265822785,
"grad_norm": 1.8350692685441172,
"learning_rate": 5e-06,
"loss": 0.5742,
"step": 380
},
{
"epoch": 0.3949367088607595,
"grad_norm": 1.6654466906631886,
"learning_rate": 5e-06,
"loss": 0.57,
"step": 390
},
{
"epoch": 0.4050632911392405,
"grad_norm": 1.5625914451532021,
"learning_rate": 5e-06,
"loss": 0.5688,
"step": 400
},
{
"epoch": 0.4151898734177215,
"grad_norm": 1.5962831101550994,
"learning_rate": 5e-06,
"loss": 0.5567,
"step": 410
},
{
"epoch": 0.4253164556962025,
"grad_norm": 1.4805817323596187,
"learning_rate": 5e-06,
"loss": 0.5723,
"step": 420
},
{
"epoch": 0.43544303797468353,
"grad_norm": 1.4164796466458769,
"learning_rate": 5e-06,
"loss": 0.5712,
"step": 430
},
{
"epoch": 0.44556962025316454,
"grad_norm": 1.5450127362729664,
"learning_rate": 5e-06,
"loss": 0.5704,
"step": 440
},
{
"epoch": 0.45569620253164556,
"grad_norm": 1.6570421627781717,
"learning_rate": 5e-06,
"loss": 0.5729,
"step": 450
},
{
"epoch": 0.46582278481012657,
"grad_norm": 1.623306340780734,
"learning_rate": 5e-06,
"loss": 0.5673,
"step": 460
},
{
"epoch": 0.4759493670886076,
"grad_norm": 1.747763610168544,
"learning_rate": 5e-06,
"loss": 0.5609,
"step": 470
},
{
"epoch": 0.4860759493670886,
"grad_norm": 1.295527401866004,
"learning_rate": 5e-06,
"loss": 0.5699,
"step": 480
},
{
"epoch": 0.4962025316455696,
"grad_norm": 1.3529876566902441,
"learning_rate": 5e-06,
"loss": 0.5783,
"step": 490
},
{
"epoch": 0.5063291139240507,
"grad_norm": 1.4541590535564763,
"learning_rate": 5e-06,
"loss": 0.5606,
"step": 500
},
{
"epoch": 0.5164556962025316,
"grad_norm": 1.5391153152763637,
"learning_rate": 5e-06,
"loss": 0.5598,
"step": 510
},
{
"epoch": 0.5265822784810127,
"grad_norm": 1.3816263219270335,
"learning_rate": 5e-06,
"loss": 0.552,
"step": 520
},
{
"epoch": 0.5367088607594936,
"grad_norm": 1.407148137331511,
"learning_rate": 5e-06,
"loss": 0.561,
"step": 530
},
{
"epoch": 0.5468354430379747,
"grad_norm": 1.4939662242734717,
"learning_rate": 5e-06,
"loss": 0.5691,
"step": 540
},
{
"epoch": 0.5569620253164557,
"grad_norm": 1.4769024813528058,
"learning_rate": 5e-06,
"loss": 0.5628,
"step": 550
},
{
"epoch": 0.5670886075949367,
"grad_norm": 1.5234796484078978,
"learning_rate": 5e-06,
"loss": 0.5571,
"step": 560
},
{
"epoch": 0.5772151898734177,
"grad_norm": 1.38748512162588,
"learning_rate": 5e-06,
"loss": 0.5588,
"step": 570
},
{
"epoch": 0.5873417721518988,
"grad_norm": 1.6616670500045834,
"learning_rate": 5e-06,
"loss": 0.5629,
"step": 580
},
{
"epoch": 0.5974683544303797,
"grad_norm": 1.3613770956091802,
"learning_rate": 5e-06,
"loss": 0.5597,
"step": 590
},
{
"epoch": 0.6075949367088608,
"grad_norm": 1.9645875965732293,
"learning_rate": 5e-06,
"loss": 0.5614,
"step": 600
},
{
"epoch": 0.6177215189873417,
"grad_norm": 1.3122962556953037,
"learning_rate": 5e-06,
"loss": 0.57,
"step": 610
},
{
"epoch": 0.6278481012658228,
"grad_norm": 1.5024516266743335,
"learning_rate": 5e-06,
"loss": 0.5629,
"step": 620
},
{
"epoch": 0.6379746835443038,
"grad_norm": 1.3599801460050238,
"learning_rate": 5e-06,
"loss": 0.5679,
"step": 630
},
{
"epoch": 0.6481012658227848,
"grad_norm": 1.3697118174446543,
"learning_rate": 5e-06,
"loss": 0.5567,
"step": 640
},
{
"epoch": 0.6582278481012658,
"grad_norm": 1.447467777446668,
"learning_rate": 5e-06,
"loss": 0.5614,
"step": 650
},
{
"epoch": 0.6683544303797468,
"grad_norm": 1.4332946464750023,
"learning_rate": 5e-06,
"loss": 0.5629,
"step": 660
},
{
"epoch": 0.6784810126582278,
"grad_norm": 1.4245518863907969,
"learning_rate": 5e-06,
"loss": 0.5607,
"step": 670
},
{
"epoch": 0.6886075949367089,
"grad_norm": 1.392960913673911,
"learning_rate": 5e-06,
"loss": 0.5631,
"step": 680
},
{
"epoch": 0.6987341772151898,
"grad_norm": 1.3851040024314174,
"learning_rate": 5e-06,
"loss": 0.5547,
"step": 690
},
{
"epoch": 0.7088607594936709,
"grad_norm": 1.5665731498555628,
"learning_rate": 5e-06,
"loss": 0.5729,
"step": 700
},
{
"epoch": 0.7189873417721518,
"grad_norm": 1.3541133459928674,
"learning_rate": 5e-06,
"loss": 0.567,
"step": 710
},
{
"epoch": 0.7291139240506329,
"grad_norm": 1.3050422174455243,
"learning_rate": 5e-06,
"loss": 0.5649,
"step": 720
},
{
"epoch": 0.739240506329114,
"grad_norm": 1.4474203168962876,
"learning_rate": 5e-06,
"loss": 0.5587,
"step": 730
},
{
"epoch": 0.7493670886075949,
"grad_norm": 1.2849691486389958,
"learning_rate": 5e-06,
"loss": 0.5584,
"step": 740
},
{
"epoch": 0.759493670886076,
"grad_norm": 1.4374182399089128,
"learning_rate": 5e-06,
"loss": 0.5485,
"step": 750
},
{
"epoch": 0.769620253164557,
"grad_norm": 1.2784166011759992,
"learning_rate": 5e-06,
"loss": 0.5619,
"step": 760
},
{
"epoch": 0.779746835443038,
"grad_norm": 1.338741085455322,
"learning_rate": 5e-06,
"loss": 0.5625,
"step": 770
},
{
"epoch": 0.789873417721519,
"grad_norm": 1.3830345955439087,
"learning_rate": 5e-06,
"loss": 0.5532,
"step": 780
},
{
"epoch": 0.8,
"grad_norm": 1.2560229071250701,
"learning_rate": 5e-06,
"loss": 0.557,
"step": 790
},
{
"epoch": 0.810126582278481,
"grad_norm": 1.3094948207554686,
"learning_rate": 5e-06,
"loss": 0.5606,
"step": 800
},
{
"epoch": 0.8202531645569621,
"grad_norm": 1.2980406537156193,
"learning_rate": 5e-06,
"loss": 0.5435,
"step": 810
},
{
"epoch": 0.830379746835443,
"grad_norm": 1.3336678028790407,
"learning_rate": 5e-06,
"loss": 0.5644,
"step": 820
},
{
"epoch": 0.8405063291139241,
"grad_norm": 1.4362113888678505,
"learning_rate": 5e-06,
"loss": 0.5615,
"step": 830
},
{
"epoch": 0.850632911392405,
"grad_norm": 1.5496595209680522,
"learning_rate": 5e-06,
"loss": 0.5619,
"step": 840
},
{
"epoch": 0.8607594936708861,
"grad_norm": 1.4648845925364151,
"learning_rate": 5e-06,
"loss": 0.5556,
"step": 850
},
{
"epoch": 0.8708860759493671,
"grad_norm": 1.3830147030582034,
"learning_rate": 5e-06,
"loss": 0.5534,
"step": 860
},
{
"epoch": 0.8810126582278481,
"grad_norm": 1.3802375047462647,
"learning_rate": 5e-06,
"loss": 0.5654,
"step": 870
},
{
"epoch": 0.8911392405063291,
"grad_norm": 1.665356350061181,
"learning_rate": 5e-06,
"loss": 0.5542,
"step": 880
},
{
"epoch": 0.9012658227848102,
"grad_norm": 1.5275969827966436,
"learning_rate": 5e-06,
"loss": 0.5492,
"step": 890
},
{
"epoch": 0.9113924050632911,
"grad_norm": 1.5257937868842386,
"learning_rate": 5e-06,
"loss": 0.5509,
"step": 900
},
{
"epoch": 0.9215189873417722,
"grad_norm": 1.3059469433496236,
"learning_rate": 5e-06,
"loss": 0.5563,
"step": 910
},
{
"epoch": 0.9316455696202531,
"grad_norm": 1.3687394115237168,
"learning_rate": 5e-06,
"loss": 0.5527,
"step": 920
},
{
"epoch": 0.9417721518987342,
"grad_norm": 1.3468586235595001,
"learning_rate": 5e-06,
"loss": 0.5594,
"step": 930
},
{
"epoch": 0.9518987341772152,
"grad_norm": 1.4377274430310882,
"learning_rate": 5e-06,
"loss": 0.5512,
"step": 940
},
{
"epoch": 0.9620253164556962,
"grad_norm": 1.2680685522134791,
"learning_rate": 5e-06,
"loss": 0.5486,
"step": 950
},
{
"epoch": 0.9721518987341772,
"grad_norm": 1.2807954706357672,
"learning_rate": 5e-06,
"loss": 0.5566,
"step": 960
},
{
"epoch": 0.9822784810126582,
"grad_norm": 1.3423829666439555,
"learning_rate": 5e-06,
"loss": 0.5485,
"step": 970
},
{
"epoch": 0.9924050632911392,
"grad_norm": 1.4283880670670543,
"learning_rate": 5e-06,
"loss": 0.5515,
"step": 980
},
{
"epoch": 0.999493670886076,
"eval_loss": 0.13834980130195618,
"eval_runtime": 507.9163,
"eval_samples_per_second": 26.193,
"eval_steps_per_second": 0.41,
"step": 987
},
{
"epoch": 1.0025316455696203,
"grad_norm": 1.6161728864680243,
"learning_rate": 5e-06,
"loss": 0.5308,
"step": 990
},
{
"epoch": 1.0126582278481013,
"grad_norm": 1.5737881952906483,
"learning_rate": 5e-06,
"loss": 0.4633,
"step": 1000
},
{
"epoch": 1.0227848101265822,
"grad_norm": 1.5616941170427905,
"learning_rate": 5e-06,
"loss": 0.4634,
"step": 1010
},
{
"epoch": 1.0329113924050632,
"grad_norm": 1.5346519713859017,
"learning_rate": 5e-06,
"loss": 0.4435,
"step": 1020
},
{
"epoch": 1.0430379746835443,
"grad_norm": 1.3492138788127321,
"learning_rate": 5e-06,
"loss": 0.4503,
"step": 1030
},
{
"epoch": 1.0531645569620254,
"grad_norm": 1.4763337790140822,
"learning_rate": 5e-06,
"loss": 0.4506,
"step": 1040
},
{
"epoch": 1.0632911392405062,
"grad_norm": 1.4225225244349122,
"learning_rate": 5e-06,
"loss": 0.4591,
"step": 1050
},
{
"epoch": 1.0734177215189873,
"grad_norm": 1.377686472346139,
"learning_rate": 5e-06,
"loss": 0.4498,
"step": 1060
},
{
"epoch": 1.0835443037974684,
"grad_norm": 1.524899956605274,
"learning_rate": 5e-06,
"loss": 0.4521,
"step": 1070
},
{
"epoch": 1.0936708860759494,
"grad_norm": 1.4555671547809872,
"learning_rate": 5e-06,
"loss": 0.4537,
"step": 1080
},
{
"epoch": 1.1037974683544305,
"grad_norm": 1.352652609202294,
"learning_rate": 5e-06,
"loss": 0.4591,
"step": 1090
},
{
"epoch": 1.1139240506329113,
"grad_norm": 1.389780851883711,
"learning_rate": 5e-06,
"loss": 0.451,
"step": 1100
},
{
"epoch": 1.1240506329113924,
"grad_norm": 1.5493151434769157,
"learning_rate": 5e-06,
"loss": 0.4545,
"step": 1110
},
{
"epoch": 1.1341772151898735,
"grad_norm": 1.4921163467682323,
"learning_rate": 5e-06,
"loss": 0.4511,
"step": 1120
},
{
"epoch": 1.1443037974683543,
"grad_norm": 1.3424106662166821,
"learning_rate": 5e-06,
"loss": 0.4597,
"step": 1130
},
{
"epoch": 1.1544303797468354,
"grad_norm": 1.4205445966678738,
"learning_rate": 5e-06,
"loss": 0.4587,
"step": 1140
},
{
"epoch": 1.1645569620253164,
"grad_norm": 1.3812137606947525,
"learning_rate": 5e-06,
"loss": 0.457,
"step": 1150
},
{
"epoch": 1.1746835443037975,
"grad_norm": 1.504094206656827,
"learning_rate": 5e-06,
"loss": 0.4582,
"step": 1160
},
{
"epoch": 1.1848101265822786,
"grad_norm": 1.4384269712621547,
"learning_rate": 5e-06,
"loss": 0.4478,
"step": 1170
},
{
"epoch": 1.1949367088607594,
"grad_norm": 1.4604436825013358,
"learning_rate": 5e-06,
"loss": 0.4561,
"step": 1180
},
{
"epoch": 1.2050632911392405,
"grad_norm": 1.484357252258588,
"learning_rate": 5e-06,
"loss": 0.463,
"step": 1190
},
{
"epoch": 1.2151898734177216,
"grad_norm": 1.3507024353159274,
"learning_rate": 5e-06,
"loss": 0.4557,
"step": 1200
},
{
"epoch": 1.2253164556962026,
"grad_norm": 1.5356768758566504,
"learning_rate": 5e-06,
"loss": 0.4616,
"step": 1210
},
{
"epoch": 1.2354430379746835,
"grad_norm": 1.507777364560312,
"learning_rate": 5e-06,
"loss": 0.4492,
"step": 1220
},
{
"epoch": 1.2455696202531645,
"grad_norm": 1.402742621286163,
"learning_rate": 5e-06,
"loss": 0.4623,
"step": 1230
},
{
"epoch": 1.2556962025316456,
"grad_norm": 1.4914361590080172,
"learning_rate": 5e-06,
"loss": 0.4576,
"step": 1240
},
{
"epoch": 1.2658227848101267,
"grad_norm": 1.4449809880641054,
"learning_rate": 5e-06,
"loss": 0.4581,
"step": 1250
},
{
"epoch": 1.2759493670886077,
"grad_norm": 1.401311614162818,
"learning_rate": 5e-06,
"loss": 0.4613,
"step": 1260
},
{
"epoch": 1.2860759493670886,
"grad_norm": 1.4035116916924748,
"learning_rate": 5e-06,
"loss": 0.4546,
"step": 1270
},
{
"epoch": 1.2962025316455696,
"grad_norm": 1.5360414127344846,
"learning_rate": 5e-06,
"loss": 0.4638,
"step": 1280
},
{
"epoch": 1.3063291139240507,
"grad_norm": 1.4721064580495236,
"learning_rate": 5e-06,
"loss": 0.4605,
"step": 1290
},
{
"epoch": 1.3164556962025316,
"grad_norm": 1.5018662441760837,
"learning_rate": 5e-06,
"loss": 0.4656,
"step": 1300
},
{
"epoch": 1.3265822784810126,
"grad_norm": 1.4057750381381058,
"learning_rate": 5e-06,
"loss": 0.4546,
"step": 1310
},
{
"epoch": 1.3367088607594937,
"grad_norm": 1.472547127671551,
"learning_rate": 5e-06,
"loss": 0.4509,
"step": 1320
},
{
"epoch": 1.3468354430379748,
"grad_norm": 1.378222631334251,
"learning_rate": 5e-06,
"loss": 0.4592,
"step": 1330
},
{
"epoch": 1.3569620253164558,
"grad_norm": 1.477645235480499,
"learning_rate": 5e-06,
"loss": 0.4731,
"step": 1340
},
{
"epoch": 1.3670886075949367,
"grad_norm": 1.4190131074856087,
"learning_rate": 5e-06,
"loss": 0.4614,
"step": 1350
},
{
"epoch": 1.3772151898734177,
"grad_norm": 1.598773468818255,
"learning_rate": 5e-06,
"loss": 0.4642,
"step": 1360
},
{
"epoch": 1.3873417721518988,
"grad_norm": 1.401701470107724,
"learning_rate": 5e-06,
"loss": 0.4599,
"step": 1370
},
{
"epoch": 1.3974683544303796,
"grad_norm": 1.5834771319486265,
"learning_rate": 5e-06,
"loss": 0.4635,
"step": 1380
},
{
"epoch": 1.4075949367088607,
"grad_norm": 1.398082340799174,
"learning_rate": 5e-06,
"loss": 0.4581,
"step": 1390
},
{
"epoch": 1.4177215189873418,
"grad_norm": 1.4006340071029044,
"learning_rate": 5e-06,
"loss": 0.4602,
"step": 1400
},
{
"epoch": 1.4278481012658228,
"grad_norm": 1.3336328229398797,
"learning_rate": 5e-06,
"loss": 0.4602,
"step": 1410
},
{
"epoch": 1.437974683544304,
"grad_norm": 1.3492817883075563,
"learning_rate": 5e-06,
"loss": 0.4687,
"step": 1420
},
{
"epoch": 1.4481012658227848,
"grad_norm": 1.342687219512621,
"learning_rate": 5e-06,
"loss": 0.4651,
"step": 1430
},
{
"epoch": 1.4582278481012658,
"grad_norm": 1.4950318280118493,
"learning_rate": 5e-06,
"loss": 0.4643,
"step": 1440
},
{
"epoch": 1.4683544303797469,
"grad_norm": 1.5341419482481558,
"learning_rate": 5e-06,
"loss": 0.467,
"step": 1450
},
{
"epoch": 1.4784810126582277,
"grad_norm": 1.3873503711966886,
"learning_rate": 5e-06,
"loss": 0.4692,
"step": 1460
},
{
"epoch": 1.4886075949367088,
"grad_norm": 1.3838670814377412,
"learning_rate": 5e-06,
"loss": 0.4606,
"step": 1470
},
{
"epoch": 1.4987341772151899,
"grad_norm": 1.435159074036311,
"learning_rate": 5e-06,
"loss": 0.4607,
"step": 1480
},
{
"epoch": 1.508860759493671,
"grad_norm": 1.416449038632833,
"learning_rate": 5e-06,
"loss": 0.4603,
"step": 1490
},
{
"epoch": 1.518987341772152,
"grad_norm": 1.349053423409347,
"learning_rate": 5e-06,
"loss": 0.4671,
"step": 1500
},
{
"epoch": 1.529113924050633,
"grad_norm": 1.5418750945784405,
"learning_rate": 5e-06,
"loss": 0.4642,
"step": 1510
},
{
"epoch": 1.539240506329114,
"grad_norm": 1.5350888472278188,
"learning_rate": 5e-06,
"loss": 0.4654,
"step": 1520
},
{
"epoch": 1.549367088607595,
"grad_norm": 1.5918459039624637,
"learning_rate": 5e-06,
"loss": 0.4623,
"step": 1530
},
{
"epoch": 1.5594936708860758,
"grad_norm": 1.4044783253886717,
"learning_rate": 5e-06,
"loss": 0.4644,
"step": 1540
},
{
"epoch": 1.5696202531645569,
"grad_norm": 1.4733622287724428,
"learning_rate": 5e-06,
"loss": 0.4699,
"step": 1550
},
{
"epoch": 1.579746835443038,
"grad_norm": 1.4190534275279456,
"learning_rate": 5e-06,
"loss": 0.4585,
"step": 1560
},
{
"epoch": 1.589873417721519,
"grad_norm": 1.588541120324257,
"learning_rate": 5e-06,
"loss": 0.4674,
"step": 1570
},
{
"epoch": 1.6,
"grad_norm": 1.4489487698353771,
"learning_rate": 5e-06,
"loss": 0.4659,
"step": 1580
},
{
"epoch": 1.6101265822784812,
"grad_norm": 1.3499923545579962,
"learning_rate": 5e-06,
"loss": 0.4582,
"step": 1590
},
{
"epoch": 1.620253164556962,
"grad_norm": 1.3640430679010667,
"learning_rate": 5e-06,
"loss": 0.4676,
"step": 1600
},
{
"epoch": 1.630379746835443,
"grad_norm": 1.3690059243296218,
"learning_rate": 5e-06,
"loss": 0.4612,
"step": 1610
},
{
"epoch": 1.640506329113924,
"grad_norm": 1.2447374210211029,
"learning_rate": 5e-06,
"loss": 0.4506,
"step": 1620
},
{
"epoch": 1.650632911392405,
"grad_norm": 1.351219366349358,
"learning_rate": 5e-06,
"loss": 0.4611,
"step": 1630
},
{
"epoch": 1.660759493670886,
"grad_norm": 1.335122455907824,
"learning_rate": 5e-06,
"loss": 0.4699,
"step": 1640
},
{
"epoch": 1.6708860759493671,
"grad_norm": 1.5625860035007904,
"learning_rate": 5e-06,
"loss": 0.4612,
"step": 1650
},
{
"epoch": 1.6810126582278482,
"grad_norm": 1.6296205380869357,
"learning_rate": 5e-06,
"loss": 0.4623,
"step": 1660
},
{
"epoch": 1.6911392405063292,
"grad_norm": 1.4330065379895607,
"learning_rate": 5e-06,
"loss": 0.4625,
"step": 1670
},
{
"epoch": 1.70126582278481,
"grad_norm": 1.3278761181861125,
"learning_rate": 5e-06,
"loss": 0.4627,
"step": 1680
},
{
"epoch": 1.7113924050632912,
"grad_norm": 1.4304366765861536,
"learning_rate": 5e-06,
"loss": 0.4646,
"step": 1690
},
{
"epoch": 1.721518987341772,
"grad_norm": 1.3460103891901807,
"learning_rate": 5e-06,
"loss": 0.4648,
"step": 1700
},
{
"epoch": 1.731645569620253,
"grad_norm": 1.4336438757742112,
"learning_rate": 5e-06,
"loss": 0.4708,
"step": 1710
},
{
"epoch": 1.7417721518987341,
"grad_norm": 1.333163056745564,
"learning_rate": 5e-06,
"loss": 0.4591,
"step": 1720
},
{
"epoch": 1.7518987341772152,
"grad_norm": 1.3819280069426765,
"learning_rate": 5e-06,
"loss": 0.4651,
"step": 1730
},
{
"epoch": 1.7620253164556963,
"grad_norm": 1.413028622360716,
"learning_rate": 5e-06,
"loss": 0.4714,
"step": 1740
},
{
"epoch": 1.7721518987341773,
"grad_norm": 1.3989242299354576,
"learning_rate": 5e-06,
"loss": 0.466,
"step": 1750
},
{
"epoch": 1.7822784810126582,
"grad_norm": 1.3712554248183217,
"learning_rate": 5e-06,
"loss": 0.4727,
"step": 1760
},
{
"epoch": 1.7924050632911392,
"grad_norm": 1.4780769754046013,
"learning_rate": 5e-06,
"loss": 0.4679,
"step": 1770
},
{
"epoch": 1.80253164556962,
"grad_norm": 1.4002831190634755,
"learning_rate": 5e-06,
"loss": 0.4647,
"step": 1780
},
{
"epoch": 1.8126582278481012,
"grad_norm": 1.3944675393169206,
"learning_rate": 5e-06,
"loss": 0.4644,
"step": 1790
},
{
"epoch": 1.8227848101265822,
"grad_norm": 1.4867966880006709,
"learning_rate": 5e-06,
"loss": 0.4675,
"step": 1800
},
{
"epoch": 1.8329113924050633,
"grad_norm": 1.404496416486814,
"learning_rate": 5e-06,
"loss": 0.4624,
"step": 1810
},
{
"epoch": 1.8430379746835444,
"grad_norm": 1.5305845929183253,
"learning_rate": 5e-06,
"loss": 0.4739,
"step": 1820
},
{
"epoch": 1.8531645569620254,
"grad_norm": 1.3084464311737185,
"learning_rate": 5e-06,
"loss": 0.4655,
"step": 1830
},
{
"epoch": 1.8632911392405065,
"grad_norm": 1.3059438675535409,
"learning_rate": 5e-06,
"loss": 0.4678,
"step": 1840
},
{
"epoch": 1.8734177215189873,
"grad_norm": 1.3867484618996055,
"learning_rate": 5e-06,
"loss": 0.4609,
"step": 1850
},
{
"epoch": 1.8835443037974684,
"grad_norm": 1.3420942943215763,
"learning_rate": 5e-06,
"loss": 0.4663,
"step": 1860
},
{
"epoch": 1.8936708860759492,
"grad_norm": 1.421266278043212,
"learning_rate": 5e-06,
"loss": 0.4665,
"step": 1870
},
{
"epoch": 1.9037974683544303,
"grad_norm": 1.36634296470369,
"learning_rate": 5e-06,
"loss": 0.4676,
"step": 1880
},
{
"epoch": 1.9139240506329114,
"grad_norm": 1.2659671091408768,
"learning_rate": 5e-06,
"loss": 0.4669,
"step": 1890
},
{
"epoch": 1.9240506329113924,
"grad_norm": 1.373264332933901,
"learning_rate": 5e-06,
"loss": 0.4701,
"step": 1900
},
{
"epoch": 1.9341772151898735,
"grad_norm": 1.3568414058185374,
"learning_rate": 5e-06,
"loss": 0.4693,
"step": 1910
},
{
"epoch": 1.9443037974683546,
"grad_norm": 1.3940261082697252,
"learning_rate": 5e-06,
"loss": 0.4763,
"step": 1920
},
{
"epoch": 1.9544303797468354,
"grad_norm": 1.3879188735035681,
"learning_rate": 5e-06,
"loss": 0.4756,
"step": 1930
},
{
"epoch": 1.9645569620253165,
"grad_norm": 1.363003506285662,
"learning_rate": 5e-06,
"loss": 0.4667,
"step": 1940
},
{
"epoch": 1.9746835443037973,
"grad_norm": 1.4533412320271968,
"learning_rate": 5e-06,
"loss": 0.469,
"step": 1950
},
{
"epoch": 1.9848101265822784,
"grad_norm": 1.4479310447515883,
"learning_rate": 5e-06,
"loss": 0.4748,
"step": 1960
},
{
"epoch": 1.9949367088607595,
"grad_norm": 1.2966648783230335,
"learning_rate": 5e-06,
"loss": 0.468,
"step": 1970
},
{
"epoch": 2.0,
"eval_loss": 0.1398227959871292,
"eval_runtime": 506.8788,
"eval_samples_per_second": 26.247,
"eval_steps_per_second": 0.41,
"step": 1975
},
{
"epoch": 2.0050632911392405,
"grad_norm": 3.06600016953371,
"learning_rate": 5e-06,
"loss": 0.4085,
"step": 1980
},
{
"epoch": 2.0151898734177216,
"grad_norm": 2.0901467389116335,
"learning_rate": 5e-06,
"loss": 0.3545,
"step": 1990
},
{
"epoch": 2.0253164556962027,
"grad_norm": 1.7198026240231687,
"learning_rate": 5e-06,
"loss": 0.3486,
"step": 2000
},
{
"epoch": 2.0354430379746837,
"grad_norm": 1.6596559930998485,
"learning_rate": 5e-06,
"loss": 0.3485,
"step": 2010
},
{
"epoch": 2.0455696202531644,
"grad_norm": 1.6010412767976676,
"learning_rate": 5e-06,
"loss": 0.3469,
"step": 2020
},
{
"epoch": 2.0556962025316454,
"grad_norm": 1.619456570416184,
"learning_rate": 5e-06,
"loss": 0.3412,
"step": 2030
},
{
"epoch": 2.0658227848101265,
"grad_norm": 1.5477510460994646,
"learning_rate": 5e-06,
"loss": 0.3499,
"step": 2040
},
{
"epoch": 2.0759493670886076,
"grad_norm": 1.6995317518377548,
"learning_rate": 5e-06,
"loss": 0.3431,
"step": 2050
},
{
"epoch": 2.0860759493670886,
"grad_norm": 1.6672377260042517,
"learning_rate": 5e-06,
"loss": 0.3509,
"step": 2060
},
{
"epoch": 2.0962025316455697,
"grad_norm": 1.5137523020681678,
"learning_rate": 5e-06,
"loss": 0.3485,
"step": 2070
},
{
"epoch": 2.1063291139240508,
"grad_norm": 1.5728338684227074,
"learning_rate": 5e-06,
"loss": 0.3414,
"step": 2080
},
{
"epoch": 2.116455696202532,
"grad_norm": 1.7215854826795536,
"learning_rate": 5e-06,
"loss": 0.356,
"step": 2090
},
{
"epoch": 2.1265822784810124,
"grad_norm": 1.547269577634038,
"learning_rate": 5e-06,
"loss": 0.351,
"step": 2100
},
{
"epoch": 2.1367088607594935,
"grad_norm": 1.7846477440752144,
"learning_rate": 5e-06,
"loss": 0.3475,
"step": 2110
},
{
"epoch": 2.1468354430379746,
"grad_norm": 1.6766174751768645,
"learning_rate": 5e-06,
"loss": 0.3507,
"step": 2120
},
{
"epoch": 2.1569620253164556,
"grad_norm": 1.6129490070505184,
"learning_rate": 5e-06,
"loss": 0.3544,
"step": 2130
},
{
"epoch": 2.1670886075949367,
"grad_norm": 1.6006266055236678,
"learning_rate": 5e-06,
"loss": 0.3514,
"step": 2140
},
{
"epoch": 2.1772151898734178,
"grad_norm": 1.5978325737155568,
"learning_rate": 5e-06,
"loss": 0.3525,
"step": 2150
},
{
"epoch": 2.187341772151899,
"grad_norm": 1.5951449058616392,
"learning_rate": 5e-06,
"loss": 0.3507,
"step": 2160
},
{
"epoch": 2.19746835443038,
"grad_norm": 1.5513789073675182,
"learning_rate": 5e-06,
"loss": 0.3496,
"step": 2170
},
{
"epoch": 2.207594936708861,
"grad_norm": 1.5229143655144641,
"learning_rate": 5e-06,
"loss": 0.3494,
"step": 2180
},
{
"epoch": 2.2177215189873416,
"grad_norm": 1.660242717310412,
"learning_rate": 5e-06,
"loss": 0.3553,
"step": 2190
},
{
"epoch": 2.2278481012658227,
"grad_norm": 1.6739776038113652,
"learning_rate": 5e-06,
"loss": 0.3607,
"step": 2200
},
{
"epoch": 2.2379746835443037,
"grad_norm": 1.7005249152138244,
"learning_rate": 5e-06,
"loss": 0.3486,
"step": 2210
},
{
"epoch": 2.248101265822785,
"grad_norm": 1.6489218772469403,
"learning_rate": 5e-06,
"loss": 0.3592,
"step": 2220
},
{
"epoch": 2.258227848101266,
"grad_norm": 1.6209760997530658,
"learning_rate": 5e-06,
"loss": 0.3584,
"step": 2230
},
{
"epoch": 2.268354430379747,
"grad_norm": 1.6153685359109242,
"learning_rate": 5e-06,
"loss": 0.3554,
"step": 2240
},
{
"epoch": 2.278481012658228,
"grad_norm": 1.6268217139529644,
"learning_rate": 5e-06,
"loss": 0.3542,
"step": 2250
},
{
"epoch": 2.2886075949367086,
"grad_norm": 1.7752820541706038,
"learning_rate": 5e-06,
"loss": 0.3621,
"step": 2260
},
{
"epoch": 2.2987341772151897,
"grad_norm": 1.5709969247849427,
"learning_rate": 5e-06,
"loss": 0.3524,
"step": 2270
},
{
"epoch": 2.3088607594936708,
"grad_norm": 1.5534897904208944,
"learning_rate": 5e-06,
"loss": 0.3563,
"step": 2280
},
{
"epoch": 2.318987341772152,
"grad_norm": 1.6608945730503153,
"learning_rate": 5e-06,
"loss": 0.3565,
"step": 2290
},
{
"epoch": 2.329113924050633,
"grad_norm": 1.5801622846665264,
"learning_rate": 5e-06,
"loss": 0.3622,
"step": 2300
},
{
"epoch": 2.339240506329114,
"grad_norm": 1.5320500459364543,
"learning_rate": 5e-06,
"loss": 0.3589,
"step": 2310
},
{
"epoch": 2.349367088607595,
"grad_norm": 1.581287461651533,
"learning_rate": 5e-06,
"loss": 0.3613,
"step": 2320
},
{
"epoch": 2.359493670886076,
"grad_norm": 1.6038211850984352,
"learning_rate": 5e-06,
"loss": 0.3617,
"step": 2330
},
{
"epoch": 2.369620253164557,
"grad_norm": 1.625966530104118,
"learning_rate": 5e-06,
"loss": 0.3598,
"step": 2340
},
{
"epoch": 2.379746835443038,
"grad_norm": 1.7645641137037487,
"learning_rate": 5e-06,
"loss": 0.362,
"step": 2350
},
{
"epoch": 2.389873417721519,
"grad_norm": 1.6940524616595496,
"learning_rate": 5e-06,
"loss": 0.361,
"step": 2360
},
{
"epoch": 2.4,
"grad_norm": 1.5384052283713547,
"learning_rate": 5e-06,
"loss": 0.3639,
"step": 2370
},
{
"epoch": 2.410126582278481,
"grad_norm": 1.6483060393926068,
"learning_rate": 5e-06,
"loss": 0.3586,
"step": 2380
},
{
"epoch": 2.420253164556962,
"grad_norm": 1.6074760148392138,
"learning_rate": 5e-06,
"loss": 0.3648,
"step": 2390
},
{
"epoch": 2.430379746835443,
"grad_norm": 1.7080201956192906,
"learning_rate": 5e-06,
"loss": 0.3554,
"step": 2400
},
{
"epoch": 2.440506329113924,
"grad_norm": 1.625757072296565,
"learning_rate": 5e-06,
"loss": 0.3531,
"step": 2410
},
{
"epoch": 2.4506329113924052,
"grad_norm": 1.6544192748138193,
"learning_rate": 5e-06,
"loss": 0.3659,
"step": 2420
},
{
"epoch": 2.460759493670886,
"grad_norm": 1.6016056620993897,
"learning_rate": 5e-06,
"loss": 0.3618,
"step": 2430
},
{
"epoch": 2.470886075949367,
"grad_norm": 1.6466444516635712,
"learning_rate": 5e-06,
"loss": 0.3651,
"step": 2440
},
{
"epoch": 2.481012658227848,
"grad_norm": 1.6213376274576505,
"learning_rate": 5e-06,
"loss": 0.3611,
"step": 2450
},
{
"epoch": 2.491139240506329,
"grad_norm": 1.61939669911579,
"learning_rate": 5e-06,
"loss": 0.3655,
"step": 2460
},
{
"epoch": 2.50126582278481,
"grad_norm": 1.7185252999562315,
"learning_rate": 5e-06,
"loss": 0.3717,
"step": 2470
},
{
"epoch": 2.511392405063291,
"grad_norm": 1.5701430968600476,
"learning_rate": 5e-06,
"loss": 0.3616,
"step": 2480
},
{
"epoch": 2.5215189873417723,
"grad_norm": 1.6525891145092144,
"learning_rate": 5e-06,
"loss": 0.3618,
"step": 2490
},
{
"epoch": 2.5316455696202533,
"grad_norm": 1.6331648861432553,
"learning_rate": 5e-06,
"loss": 0.3608,
"step": 2500
},
{
"epoch": 2.5417721518987344,
"grad_norm": 1.5954836812756767,
"learning_rate": 5e-06,
"loss": 0.3564,
"step": 2510
},
{
"epoch": 2.5518987341772155,
"grad_norm": 1.6221389574344411,
"learning_rate": 5e-06,
"loss": 0.3603,
"step": 2520
},
{
"epoch": 2.562025316455696,
"grad_norm": 1.645557544125129,
"learning_rate": 5e-06,
"loss": 0.3633,
"step": 2530
},
{
"epoch": 2.572151898734177,
"grad_norm": 1.7704063332654312,
"learning_rate": 5e-06,
"loss": 0.3658,
"step": 2540
},
{
"epoch": 2.5822784810126582,
"grad_norm": 1.6779345444764124,
"learning_rate": 5e-06,
"loss": 0.3706,
"step": 2550
},
{
"epoch": 2.5924050632911393,
"grad_norm": 1.6024752118154975,
"learning_rate": 5e-06,
"loss": 0.3618,
"step": 2560
},
{
"epoch": 2.6025316455696204,
"grad_norm": 1.5688766989229237,
"learning_rate": 5e-06,
"loss": 0.3719,
"step": 2570
},
{
"epoch": 2.6126582278481014,
"grad_norm": 1.5972321010318007,
"learning_rate": 5e-06,
"loss": 0.3695,
"step": 2580
},
{
"epoch": 2.622784810126582,
"grad_norm": 1.7926779657765413,
"learning_rate": 5e-06,
"loss": 0.3671,
"step": 2590
},
{
"epoch": 2.632911392405063,
"grad_norm": 1.6497318559260037,
"learning_rate": 5e-06,
"loss": 0.3648,
"step": 2600
},
{
"epoch": 2.643037974683544,
"grad_norm": 1.825449432151336,
"learning_rate": 5e-06,
"loss": 0.3675,
"step": 2610
},
{
"epoch": 2.6531645569620252,
"grad_norm": 1.6048797855449353,
"learning_rate": 5e-06,
"loss": 0.3638,
"step": 2620
},
{
"epoch": 2.6632911392405063,
"grad_norm": 1.6207591062208428,
"learning_rate": 5e-06,
"loss": 0.3683,
"step": 2630
},
{
"epoch": 2.6734177215189874,
"grad_norm": 1.7272301549499494,
"learning_rate": 5e-06,
"loss": 0.3685,
"step": 2640
},
{
"epoch": 2.6835443037974684,
"grad_norm": 1.6364841258418936,
"learning_rate": 5e-06,
"loss": 0.3684,
"step": 2650
},
{
"epoch": 2.6936708860759495,
"grad_norm": 1.6937643730038003,
"learning_rate": 5e-06,
"loss": 0.3723,
"step": 2660
},
{
"epoch": 2.7037974683544306,
"grad_norm": 1.5774730979594618,
"learning_rate": 5e-06,
"loss": 0.3671,
"step": 2670
},
{
"epoch": 2.7139240506329116,
"grad_norm": 1.6768910733845062,
"learning_rate": 5e-06,
"loss": 0.3715,
"step": 2680
},
{
"epoch": 2.7240506329113923,
"grad_norm": 1.6389744367173145,
"learning_rate": 5e-06,
"loss": 0.3674,
"step": 2690
},
{
"epoch": 2.7341772151898733,
"grad_norm": 1.7387883655123013,
"learning_rate": 5e-06,
"loss": 0.3701,
"step": 2700
},
{
"epoch": 2.7443037974683544,
"grad_norm": 1.622855925296899,
"learning_rate": 5e-06,
"loss": 0.3692,
"step": 2710
},
{
"epoch": 2.7544303797468355,
"grad_norm": 1.6049744247340423,
"learning_rate": 5e-06,
"loss": 0.3605,
"step": 2720
},
{
"epoch": 2.7645569620253165,
"grad_norm": 1.674299981616671,
"learning_rate": 5e-06,
"loss": 0.3681,
"step": 2730
},
{
"epoch": 2.7746835443037976,
"grad_norm": 1.6563350827590924,
"learning_rate": 5e-06,
"loss": 0.3732,
"step": 2740
},
{
"epoch": 2.7848101265822782,
"grad_norm": 1.6562436415268564,
"learning_rate": 5e-06,
"loss": 0.3666,
"step": 2750
},
{
"epoch": 2.7949367088607593,
"grad_norm": 1.5565903819361853,
"learning_rate": 5e-06,
"loss": 0.3706,
"step": 2760
},
{
"epoch": 2.8050632911392404,
"grad_norm": 1.7142515863322454,
"learning_rate": 5e-06,
"loss": 0.3679,
"step": 2770
},
{
"epoch": 2.8151898734177214,
"grad_norm": 1.5712425643932983,
"learning_rate": 5e-06,
"loss": 0.3761,
"step": 2780
},
{
"epoch": 2.8253164556962025,
"grad_norm": 1.597329501161643,
"learning_rate": 5e-06,
"loss": 0.3697,
"step": 2790
},
{
"epoch": 2.8354430379746836,
"grad_norm": 1.6642023307988005,
"learning_rate": 5e-06,
"loss": 0.3613,
"step": 2800
},
{
"epoch": 2.8455696202531646,
"grad_norm": 1.6302362931142675,
"learning_rate": 5e-06,
"loss": 0.3773,
"step": 2810
},
{
"epoch": 2.8556962025316457,
"grad_norm": 1.594346958058013,
"learning_rate": 5e-06,
"loss": 0.3699,
"step": 2820
},
{
"epoch": 2.8658227848101268,
"grad_norm": 1.5803480355180004,
"learning_rate": 5e-06,
"loss": 0.3711,
"step": 2830
},
{
"epoch": 2.875949367088608,
"grad_norm": 1.4800494871531567,
"learning_rate": 5e-06,
"loss": 0.3691,
"step": 2840
},
{
"epoch": 2.8860759493670884,
"grad_norm": 1.6631915526393122,
"learning_rate": 5e-06,
"loss": 0.3747,
"step": 2850
},
{
"epoch": 2.8962025316455695,
"grad_norm": 1.630862108030936,
"learning_rate": 5e-06,
"loss": 0.3674,
"step": 2860
},
{
"epoch": 2.9063291139240506,
"grad_norm": 1.6893754841333202,
"learning_rate": 5e-06,
"loss": 0.373,
"step": 2870
},
{
"epoch": 2.9164556962025316,
"grad_norm": 1.705060151220597,
"learning_rate": 5e-06,
"loss": 0.3723,
"step": 2880
},
{
"epoch": 2.9265822784810127,
"grad_norm": 1.6188859279801948,
"learning_rate": 5e-06,
"loss": 0.3796,
"step": 2890
},
{
"epoch": 2.9367088607594938,
"grad_norm": 1.629353248790061,
"learning_rate": 5e-06,
"loss": 0.3828,
"step": 2900
},
{
"epoch": 2.946835443037975,
"grad_norm": 1.7509445495897462,
"learning_rate": 5e-06,
"loss": 0.3713,
"step": 2910
},
{
"epoch": 2.9569620253164555,
"grad_norm": 1.6650857655447995,
"learning_rate": 5e-06,
"loss": 0.369,
"step": 2920
},
{
"epoch": 2.9670886075949365,
"grad_norm": 1.8285248313648312,
"learning_rate": 5e-06,
"loss": 0.3721,
"step": 2930
},
{
"epoch": 2.9772151898734176,
"grad_norm": 1.6243458834916522,
"learning_rate": 5e-06,
"loss": 0.3759,
"step": 2940
},
{
"epoch": 2.9873417721518987,
"grad_norm": 1.6911778320262214,
"learning_rate": 5e-06,
"loss": 0.3746,
"step": 2950
},
{
"epoch": 2.9974683544303797,
"grad_norm": 1.7509767039817499,
"learning_rate": 5e-06,
"loss": 0.3711,
"step": 2960
},
{
"epoch": 2.9984810126582278,
"eval_loss": 0.15120381116867065,
"eval_runtime": 508.8119,
"eval_samples_per_second": 26.147,
"eval_steps_per_second": 0.409,
"step": 2961
},
{
"epoch": 2.9984810126582278,
"step": 2961,
"total_flos": 2479683262218240.0,
"train_loss": 0.46617310421968466,
"train_runtime": 84499.4991,
"train_samples_per_second": 8.974,
"train_steps_per_second": 0.035
}
],
"logging_steps": 10,
"max_steps": 2961,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2479683262218240.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}