gemma7b-coding-gpt4o-100k / trainer_state.json
chansung's picture
Model save
eb4b3e1 verified
raw
history blame contribute delete
No virus
157 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 9.989373007438894,
"eval_steps": 500,
"global_step": 4700,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0021253985122210413,
"grad_norm": 512.0,
"learning_rate": 4.2553191489361704e-07,
"loss": 23.7791,
"step": 1
},
{
"epoch": 0.010626992561105207,
"grad_norm": 488.0,
"learning_rate": 2.1276595744680853e-06,
"loss": 23.5159,
"step": 5
},
{
"epoch": 0.021253985122210415,
"grad_norm": 244.0,
"learning_rate": 4.255319148936171e-06,
"loss": 20.5987,
"step": 10
},
{
"epoch": 0.031880977683315624,
"grad_norm": 83.5,
"learning_rate": 6.3829787234042555e-06,
"loss": 17.7875,
"step": 15
},
{
"epoch": 0.04250797024442083,
"grad_norm": 45.25,
"learning_rate": 8.510638297872341e-06,
"loss": 15.0704,
"step": 20
},
{
"epoch": 0.053134962805526036,
"grad_norm": 19.125,
"learning_rate": 1.0638297872340426e-05,
"loss": 14.1038,
"step": 25
},
{
"epoch": 0.06376195536663125,
"grad_norm": 9.5,
"learning_rate": 1.2765957446808511e-05,
"loss": 13.5893,
"step": 30
},
{
"epoch": 0.07438894792773645,
"grad_norm": 5.96875,
"learning_rate": 1.4893617021276596e-05,
"loss": 12.8509,
"step": 35
},
{
"epoch": 0.08501594048884166,
"grad_norm": 6.34375,
"learning_rate": 1.7021276595744682e-05,
"loss": 12.4395,
"step": 40
},
{
"epoch": 0.09564293304994687,
"grad_norm": 8.875,
"learning_rate": 1.9148936170212766e-05,
"loss": 12.1862,
"step": 45
},
{
"epoch": 0.10626992561105207,
"grad_norm": 14.5625,
"learning_rate": 2.1276595744680852e-05,
"loss": 11.1895,
"step": 50
},
{
"epoch": 0.11689691817215728,
"grad_norm": 29.5,
"learning_rate": 2.340425531914894e-05,
"loss": 9.9832,
"step": 55
},
{
"epoch": 0.1275239107332625,
"grad_norm": 46.75,
"learning_rate": 2.5531914893617022e-05,
"loss": 6.8574,
"step": 60
},
{
"epoch": 0.1381509032943677,
"grad_norm": 9.8125,
"learning_rate": 2.765957446808511e-05,
"loss": 2.6513,
"step": 65
},
{
"epoch": 0.1487778958554729,
"grad_norm": 2.9375,
"learning_rate": 2.9787234042553192e-05,
"loss": 1.4605,
"step": 70
},
{
"epoch": 0.1594048884165781,
"grad_norm": 3.453125,
"learning_rate": 3.191489361702128e-05,
"loss": 1.1919,
"step": 75
},
{
"epoch": 0.17003188097768332,
"grad_norm": 2.015625,
"learning_rate": 3.4042553191489365e-05,
"loss": 1.0938,
"step": 80
},
{
"epoch": 0.1806588735387885,
"grad_norm": 2.265625,
"learning_rate": 3.617021276595745e-05,
"loss": 0.9763,
"step": 85
},
{
"epoch": 0.19128586609989373,
"grad_norm": 4.96875,
"learning_rate": 3.829787234042553e-05,
"loss": 0.9307,
"step": 90
},
{
"epoch": 0.20191285866099895,
"grad_norm": 1.5390625,
"learning_rate": 4.0425531914893614e-05,
"loss": 0.866,
"step": 95
},
{
"epoch": 0.21253985122210414,
"grad_norm": 4.40625,
"learning_rate": 4.2553191489361704e-05,
"loss": 0.8456,
"step": 100
},
{
"epoch": 0.22316684378320936,
"grad_norm": 2.6875,
"learning_rate": 4.468085106382979e-05,
"loss": 0.821,
"step": 105
},
{
"epoch": 0.23379383634431455,
"grad_norm": 2.75,
"learning_rate": 4.680851063829788e-05,
"loss": 0.7823,
"step": 110
},
{
"epoch": 0.24442082890541977,
"grad_norm": 2.953125,
"learning_rate": 4.893617021276596e-05,
"loss": 0.7831,
"step": 115
},
{
"epoch": 0.255047821466525,
"grad_norm": 3.90625,
"learning_rate": 5.1063829787234044e-05,
"loss": 0.7535,
"step": 120
},
{
"epoch": 0.26567481402763016,
"grad_norm": 4.3125,
"learning_rate": 5.319148936170213e-05,
"loss": 0.714,
"step": 125
},
{
"epoch": 0.2763018065887354,
"grad_norm": 7.09375,
"learning_rate": 5.531914893617022e-05,
"loss": 0.7084,
"step": 130
},
{
"epoch": 0.2869287991498406,
"grad_norm": 4.15625,
"learning_rate": 5.744680851063831e-05,
"loss": 0.7074,
"step": 135
},
{
"epoch": 0.2975557917109458,
"grad_norm": 3.203125,
"learning_rate": 5.9574468085106384e-05,
"loss": 0.6913,
"step": 140
},
{
"epoch": 0.30818278427205104,
"grad_norm": 1.9296875,
"learning_rate": 6.170212765957447e-05,
"loss": 0.6714,
"step": 145
},
{
"epoch": 0.3188097768331562,
"grad_norm": 29.25,
"learning_rate": 6.382978723404256e-05,
"loss": 0.6761,
"step": 150
},
{
"epoch": 0.3294367693942614,
"grad_norm": 2.921875,
"learning_rate": 6.595744680851063e-05,
"loss": 0.6739,
"step": 155
},
{
"epoch": 0.34006376195536664,
"grad_norm": 3.921875,
"learning_rate": 6.808510638297873e-05,
"loss": 0.6572,
"step": 160
},
{
"epoch": 0.35069075451647186,
"grad_norm": 8.25,
"learning_rate": 7.021276595744681e-05,
"loss": 0.644,
"step": 165
},
{
"epoch": 0.361317747077577,
"grad_norm": 11.125,
"learning_rate": 7.23404255319149e-05,
"loss": 0.6699,
"step": 170
},
{
"epoch": 0.37194473963868224,
"grad_norm": 2.828125,
"learning_rate": 7.446808510638298e-05,
"loss": 0.6613,
"step": 175
},
{
"epoch": 0.38257173219978746,
"grad_norm": 4.84375,
"learning_rate": 7.659574468085106e-05,
"loss": 0.6376,
"step": 180
},
{
"epoch": 0.3931987247608927,
"grad_norm": 3.796875,
"learning_rate": 7.872340425531916e-05,
"loss": 0.6444,
"step": 185
},
{
"epoch": 0.4038257173219979,
"grad_norm": 2.078125,
"learning_rate": 8.085106382978723e-05,
"loss": 0.6349,
"step": 190
},
{
"epoch": 0.41445270988310307,
"grad_norm": 3.53125,
"learning_rate": 8.297872340425533e-05,
"loss": 0.644,
"step": 195
},
{
"epoch": 0.4250797024442083,
"grad_norm": 1.9921875,
"learning_rate": 8.510638297872341e-05,
"loss": 0.6224,
"step": 200
},
{
"epoch": 0.4357066950053135,
"grad_norm": 3.1875,
"learning_rate": 8.723404255319149e-05,
"loss": 0.6213,
"step": 205
},
{
"epoch": 0.4463336875664187,
"grad_norm": 6.375,
"learning_rate": 8.936170212765958e-05,
"loss": 0.6095,
"step": 210
},
{
"epoch": 0.4569606801275239,
"grad_norm": 4.28125,
"learning_rate": 9.148936170212766e-05,
"loss": 0.6226,
"step": 215
},
{
"epoch": 0.4675876726886291,
"grad_norm": 2.671875,
"learning_rate": 9.361702127659576e-05,
"loss": 0.6042,
"step": 220
},
{
"epoch": 0.4782146652497343,
"grad_norm": 0.88671875,
"learning_rate": 9.574468085106384e-05,
"loss": 0.6009,
"step": 225
},
{
"epoch": 0.48884165781083955,
"grad_norm": 1.2265625,
"learning_rate": 9.787234042553192e-05,
"loss": 0.5855,
"step": 230
},
{
"epoch": 0.49946865037194477,
"grad_norm": 1.8671875,
"learning_rate": 0.0001,
"loss": 0.5729,
"step": 235
},
{
"epoch": 0.51009564293305,
"grad_norm": 2.171875,
"learning_rate": 0.00010212765957446809,
"loss": 0.5751,
"step": 240
},
{
"epoch": 0.5207226354941552,
"grad_norm": 13.5625,
"learning_rate": 0.00010425531914893618,
"loss": 0.5832,
"step": 245
},
{
"epoch": 0.5313496280552603,
"grad_norm": 1.0859375,
"learning_rate": 0.00010638297872340425,
"loss": 0.5858,
"step": 250
},
{
"epoch": 0.5419766206163655,
"grad_norm": 14.5,
"learning_rate": 0.00010851063829787234,
"loss": 0.5908,
"step": 255
},
{
"epoch": 0.5526036131774708,
"grad_norm": 1.5546875,
"learning_rate": 0.00011063829787234043,
"loss": 0.5924,
"step": 260
},
{
"epoch": 0.563230605738576,
"grad_norm": 1.7578125,
"learning_rate": 0.00011276595744680852,
"loss": 0.5704,
"step": 265
},
{
"epoch": 0.5738575982996812,
"grad_norm": 0.7734375,
"learning_rate": 0.00011489361702127661,
"loss": 0.5695,
"step": 270
},
{
"epoch": 0.5844845908607864,
"grad_norm": 2.25,
"learning_rate": 0.00011702127659574468,
"loss": 0.5654,
"step": 275
},
{
"epoch": 0.5951115834218916,
"grad_norm": 2.703125,
"learning_rate": 0.00011914893617021277,
"loss": 0.5816,
"step": 280
},
{
"epoch": 0.6057385759829969,
"grad_norm": 0.95703125,
"learning_rate": 0.00012127659574468086,
"loss": 0.5627,
"step": 285
},
{
"epoch": 0.6163655685441021,
"grad_norm": 2.703125,
"learning_rate": 0.00012340425531914893,
"loss": 0.5557,
"step": 290
},
{
"epoch": 0.6269925611052072,
"grad_norm": 1.8046875,
"learning_rate": 0.00012553191489361702,
"loss": 0.5571,
"step": 295
},
{
"epoch": 0.6376195536663124,
"grad_norm": 4.46875,
"learning_rate": 0.00012765957446808513,
"loss": 0.544,
"step": 300
},
{
"epoch": 0.6482465462274176,
"grad_norm": 3.421875,
"learning_rate": 0.00012978723404255318,
"loss": 0.5593,
"step": 305
},
{
"epoch": 0.6588735387885228,
"grad_norm": 3.5625,
"learning_rate": 0.00013191489361702127,
"loss": 0.5513,
"step": 310
},
{
"epoch": 0.6695005313496281,
"grad_norm": 9.0,
"learning_rate": 0.00013404255319148938,
"loss": 0.5681,
"step": 315
},
{
"epoch": 0.6801275239107333,
"grad_norm": 1.03125,
"learning_rate": 0.00013617021276595746,
"loss": 0.554,
"step": 320
},
{
"epoch": 0.6907545164718385,
"grad_norm": 1.484375,
"learning_rate": 0.00013829787234042554,
"loss": 0.5383,
"step": 325
},
{
"epoch": 0.7013815090329437,
"grad_norm": 1.5234375,
"learning_rate": 0.00014042553191489363,
"loss": 0.5381,
"step": 330
},
{
"epoch": 0.7120085015940489,
"grad_norm": 1.125,
"learning_rate": 0.0001425531914893617,
"loss": 0.5544,
"step": 335
},
{
"epoch": 0.722635494155154,
"grad_norm": 0.96484375,
"learning_rate": 0.0001446808510638298,
"loss": 0.539,
"step": 340
},
{
"epoch": 0.7332624867162593,
"grad_norm": 2.890625,
"learning_rate": 0.00014680851063829788,
"loss": 0.542,
"step": 345
},
{
"epoch": 0.7438894792773645,
"grad_norm": 0.83984375,
"learning_rate": 0.00014893617021276596,
"loss": 0.531,
"step": 350
},
{
"epoch": 0.7545164718384697,
"grad_norm": 3.015625,
"learning_rate": 0.00015106382978723407,
"loss": 0.5358,
"step": 355
},
{
"epoch": 0.7651434643995749,
"grad_norm": 0.9296875,
"learning_rate": 0.00015319148936170213,
"loss": 0.5366,
"step": 360
},
{
"epoch": 0.7757704569606801,
"grad_norm": 2.71875,
"learning_rate": 0.0001553191489361702,
"loss": 0.5359,
"step": 365
},
{
"epoch": 0.7863974495217854,
"grad_norm": 1.8984375,
"learning_rate": 0.00015744680851063832,
"loss": 0.5506,
"step": 370
},
{
"epoch": 0.7970244420828906,
"grad_norm": 2.484375,
"learning_rate": 0.00015957446808510637,
"loss": 0.5286,
"step": 375
},
{
"epoch": 0.8076514346439958,
"grad_norm": 0.72265625,
"learning_rate": 0.00016170212765957446,
"loss": 0.5319,
"step": 380
},
{
"epoch": 0.8182784272051009,
"grad_norm": 0.5859375,
"learning_rate": 0.00016382978723404257,
"loss": 0.5312,
"step": 385
},
{
"epoch": 0.8289054197662061,
"grad_norm": 1.234375,
"learning_rate": 0.00016595744680851065,
"loss": 0.5284,
"step": 390
},
{
"epoch": 0.8395324123273114,
"grad_norm": 2.859375,
"learning_rate": 0.00016808510638297873,
"loss": 0.5448,
"step": 395
},
{
"epoch": 0.8501594048884166,
"grad_norm": 1.90625,
"learning_rate": 0.00017021276595744682,
"loss": 0.5292,
"step": 400
},
{
"epoch": 0.8607863974495218,
"grad_norm": 1.0234375,
"learning_rate": 0.0001723404255319149,
"loss": 0.5246,
"step": 405
},
{
"epoch": 0.871413390010627,
"grad_norm": 0.87109375,
"learning_rate": 0.00017446808510638298,
"loss": 0.5353,
"step": 410
},
{
"epoch": 0.8820403825717322,
"grad_norm": 21.0,
"learning_rate": 0.00017659574468085107,
"loss": 0.5399,
"step": 415
},
{
"epoch": 0.8926673751328374,
"grad_norm": 1.296875,
"learning_rate": 0.00017872340425531915,
"loss": 0.5405,
"step": 420
},
{
"epoch": 0.9032943676939427,
"grad_norm": 0.58984375,
"learning_rate": 0.00018085106382978726,
"loss": 0.5233,
"step": 425
},
{
"epoch": 0.9139213602550478,
"grad_norm": 2.21875,
"learning_rate": 0.00018297872340425532,
"loss": 0.5225,
"step": 430
},
{
"epoch": 0.924548352816153,
"grad_norm": 0.6328125,
"learning_rate": 0.0001851063829787234,
"loss": 0.532,
"step": 435
},
{
"epoch": 0.9351753453772582,
"grad_norm": 0.87109375,
"learning_rate": 0.0001872340425531915,
"loss": 0.5353,
"step": 440
},
{
"epoch": 0.9458023379383634,
"grad_norm": 1.8671875,
"learning_rate": 0.00018936170212765957,
"loss": 0.5365,
"step": 445
},
{
"epoch": 0.9564293304994687,
"grad_norm": 0.74609375,
"learning_rate": 0.00019148936170212768,
"loss": 0.5395,
"step": 450
},
{
"epoch": 0.9670563230605739,
"grad_norm": 2.421875,
"learning_rate": 0.00019361702127659576,
"loss": 0.5309,
"step": 455
},
{
"epoch": 0.9776833156216791,
"grad_norm": 4.96875,
"learning_rate": 0.00019574468085106384,
"loss": 0.5281,
"step": 460
},
{
"epoch": 0.9883103081827843,
"grad_norm": 1.0234375,
"learning_rate": 0.00019787234042553193,
"loss": 0.5182,
"step": 465
},
{
"epoch": 0.9989373007438895,
"grad_norm": 1.7109375,
"learning_rate": 0.0002,
"loss": 0.5262,
"step": 470
},
{
"epoch": 0.9989373007438895,
"eval_loss": 1.3223822116851807,
"eval_runtime": 0.8107,
"eval_samples_per_second": 6.167,
"eval_steps_per_second": 1.233,
"step": 470
},
{
"epoch": 1.0095642933049946,
"grad_norm": 1.484375,
"learning_rate": 0.00019999931050912237,
"loss": 0.5078,
"step": 475
},
{
"epoch": 1.0201912858661,
"grad_norm": 1.125,
"learning_rate": 0.00019999724204599747,
"loss": 0.4809,
"step": 480
},
{
"epoch": 1.030818278427205,
"grad_norm": 2.703125,
"learning_rate": 0.00019999379463914898,
"loss": 0.4983,
"step": 485
},
{
"epoch": 1.0414452709883104,
"grad_norm": 1.90625,
"learning_rate": 0.00019998896833611603,
"loss": 0.4957,
"step": 490
},
{
"epoch": 1.0520722635494155,
"grad_norm": 0.6640625,
"learning_rate": 0.00019998276320345247,
"loss": 0.4901,
"step": 495
},
{
"epoch": 1.0626992561105206,
"grad_norm": 2.171875,
"learning_rate": 0.0001999751793267259,
"loss": 0.4931,
"step": 500
},
{
"epoch": 1.073326248671626,
"grad_norm": 1.40625,
"learning_rate": 0.00019996621681051669,
"loss": 0.4916,
"step": 505
},
{
"epoch": 1.083953241232731,
"grad_norm": 2.75,
"learning_rate": 0.0001999558757784162,
"loss": 0.501,
"step": 510
},
{
"epoch": 1.0945802337938364,
"grad_norm": 1.390625,
"learning_rate": 0.00019994415637302547,
"loss": 0.5011,
"step": 515
},
{
"epoch": 1.1052072263549415,
"grad_norm": 2.625,
"learning_rate": 0.0001999310587559529,
"loss": 0.4966,
"step": 520
},
{
"epoch": 1.1158342189160468,
"grad_norm": 2.703125,
"learning_rate": 0.00019991658310781224,
"loss": 0.5037,
"step": 525
},
{
"epoch": 1.126461211477152,
"grad_norm": 1.0078125,
"learning_rate": 0.00019990072962822007,
"loss": 0.5134,
"step": 530
},
{
"epoch": 1.1370882040382573,
"grad_norm": 0.5703125,
"learning_rate": 0.00019988349853579295,
"loss": 0.4997,
"step": 535
},
{
"epoch": 1.1477151965993624,
"grad_norm": 1.0390625,
"learning_rate": 0.00019986489006814452,
"loss": 0.4965,
"step": 540
},
{
"epoch": 1.1583421891604675,
"grad_norm": 7.1875,
"learning_rate": 0.00019984490448188218,
"loss": 0.4843,
"step": 545
},
{
"epoch": 1.1689691817215728,
"grad_norm": 2.546875,
"learning_rate": 0.00019982354205260347,
"loss": 0.5033,
"step": 550
},
{
"epoch": 1.179596174282678,
"grad_norm": 0.443359375,
"learning_rate": 0.0001998008030748924,
"loss": 0.4976,
"step": 555
},
{
"epoch": 1.1902231668437833,
"grad_norm": 0.58984375,
"learning_rate": 0.00019977668786231534,
"loss": 0.497,
"step": 560
},
{
"epoch": 1.2008501594048884,
"grad_norm": 0.79296875,
"learning_rate": 0.00019975119674741664,
"loss": 0.4973,
"step": 565
},
{
"epoch": 1.2114771519659937,
"grad_norm": 1.171875,
"learning_rate": 0.00019972433008171416,
"loss": 0.5,
"step": 570
},
{
"epoch": 1.2221041445270988,
"grad_norm": 1.609375,
"learning_rate": 0.00019969608823569433,
"loss": 0.4925,
"step": 575
},
{
"epoch": 1.2327311370882041,
"grad_norm": 1.6875,
"learning_rate": 0.00019966647159880703,
"loss": 0.4848,
"step": 580
},
{
"epoch": 1.2433581296493093,
"grad_norm": 2.734375,
"learning_rate": 0.00019963548057946024,
"loss": 0.5033,
"step": 585
},
{
"epoch": 1.2539851222104144,
"grad_norm": 0.5390625,
"learning_rate": 0.00019960311560501454,
"loss": 0.4925,
"step": 590
},
{
"epoch": 1.2646121147715197,
"grad_norm": 1.3125,
"learning_rate": 0.000199569377121777,
"loss": 0.487,
"step": 595
},
{
"epoch": 1.2752391073326248,
"grad_norm": 0.44921875,
"learning_rate": 0.0001995342655949951,
"loss": 0.497,
"step": 600
},
{
"epoch": 1.2858660998937301,
"grad_norm": 0.57421875,
"learning_rate": 0.00019949778150885042,
"loss": 0.4877,
"step": 605
},
{
"epoch": 1.2964930924548352,
"grad_norm": 0.56640625,
"learning_rate": 0.00019945992536645187,
"loss": 0.4942,
"step": 610
},
{
"epoch": 1.3071200850159406,
"grad_norm": 0.55859375,
"learning_rate": 0.00019942069768982872,
"loss": 0.4881,
"step": 615
},
{
"epoch": 1.3177470775770457,
"grad_norm": 0.5390625,
"learning_rate": 0.0001993800990199235,
"loss": 0.4959,
"step": 620
},
{
"epoch": 1.328374070138151,
"grad_norm": 0.61328125,
"learning_rate": 0.0001993381299165844,
"loss": 0.4883,
"step": 625
},
{
"epoch": 1.3390010626992561,
"grad_norm": 0.546875,
"learning_rate": 0.0001992947909585578,
"loss": 0.4861,
"step": 630
},
{
"epoch": 1.3496280552603612,
"grad_norm": 0.74609375,
"learning_rate": 0.00019925008274347995,
"loss": 0.4868,
"step": 635
},
{
"epoch": 1.3602550478214666,
"grad_norm": 0.66796875,
"learning_rate": 0.000199204005887869,
"loss": 0.4979,
"step": 640
},
{
"epoch": 1.3708820403825717,
"grad_norm": 1.0703125,
"learning_rate": 0.00019915656102711634,
"loss": 0.4885,
"step": 645
},
{
"epoch": 1.381509032943677,
"grad_norm": 0.55078125,
"learning_rate": 0.000199107748815478,
"loss": 0.4928,
"step": 650
},
{
"epoch": 1.392136025504782,
"grad_norm": 1.46875,
"learning_rate": 0.00019905756992606548,
"loss": 0.4879,
"step": 655
},
{
"epoch": 1.4027630180658874,
"grad_norm": 0.70703125,
"learning_rate": 0.00019900602505083648,
"loss": 0.4854,
"step": 660
},
{
"epoch": 1.4133900106269925,
"grad_norm": 5.0,
"learning_rate": 0.00019895311490058542,
"loss": 0.4932,
"step": 665
},
{
"epoch": 1.4240170031880979,
"grad_norm": 0.5390625,
"learning_rate": 0.0001988988402049336,
"loss": 0.4911,
"step": 670
},
{
"epoch": 1.434643995749203,
"grad_norm": 0.490234375,
"learning_rate": 0.00019884320171231925,
"loss": 0.4967,
"step": 675
},
{
"epoch": 1.445270988310308,
"grad_norm": 0.859375,
"learning_rate": 0.00019878620018998696,
"loss": 0.4893,
"step": 680
},
{
"epoch": 1.4558979808714134,
"grad_norm": 1.09375,
"learning_rate": 0.00019872783642397733,
"loss": 0.4849,
"step": 685
},
{
"epoch": 1.4665249734325185,
"grad_norm": 0.431640625,
"learning_rate": 0.00019866811121911607,
"loss": 0.485,
"step": 690
},
{
"epoch": 1.4771519659936239,
"grad_norm": 0.984375,
"learning_rate": 0.00019860702539900287,
"loss": 0.4993,
"step": 695
},
{
"epoch": 1.487778958554729,
"grad_norm": 3.546875,
"learning_rate": 0.000198544579806,
"loss": 0.4858,
"step": 700
},
{
"epoch": 1.4984059511158343,
"grad_norm": 0.8359375,
"learning_rate": 0.00019848077530122083,
"loss": 0.4781,
"step": 705
},
{
"epoch": 1.5090329436769394,
"grad_norm": 1.109375,
"learning_rate": 0.0001984156127645178,
"loss": 0.4771,
"step": 710
},
{
"epoch": 1.5196599362380447,
"grad_norm": 1.734375,
"learning_rate": 0.00019834909309447045,
"loss": 0.4913,
"step": 715
},
{
"epoch": 1.5302869287991498,
"grad_norm": 0.62890625,
"learning_rate": 0.00019828121720837286,
"loss": 0.4957,
"step": 720
},
{
"epoch": 1.540913921360255,
"grad_norm": 0.63671875,
"learning_rate": 0.00019821198604222113,
"loss": 0.4829,
"step": 725
},
{
"epoch": 1.5515409139213603,
"grad_norm": 1.9140625,
"learning_rate": 0.00019814140055070042,
"loss": 0.4901,
"step": 730
},
{
"epoch": 1.5621679064824656,
"grad_norm": 0.6640625,
"learning_rate": 0.00019806946170717175,
"loss": 0.4813,
"step": 735
},
{
"epoch": 1.5727948990435707,
"grad_norm": 0.408203125,
"learning_rate": 0.0001979961705036587,
"loss": 0.472,
"step": 740
},
{
"epoch": 1.5834218916046758,
"grad_norm": 0.484375,
"learning_rate": 0.00019792152795083351,
"loss": 0.4895,
"step": 745
},
{
"epoch": 1.594048884165781,
"grad_norm": 1.21875,
"learning_rate": 0.00019784553507800349,
"loss": 0.4911,
"step": 750
},
{
"epoch": 1.6046758767268863,
"grad_norm": 0.7421875,
"learning_rate": 0.00019776819293309633,
"loss": 0.4801,
"step": 755
},
{
"epoch": 1.6153028692879916,
"grad_norm": 0.98046875,
"learning_rate": 0.00019768950258264623,
"loss": 0.488,
"step": 760
},
{
"epoch": 1.6259298618490967,
"grad_norm": 0.63671875,
"learning_rate": 0.00019760946511177872,
"loss": 0.4795,
"step": 765
},
{
"epoch": 1.6365568544102018,
"grad_norm": 0.94140625,
"learning_rate": 0.0001975280816241959,
"loss": 0.4797,
"step": 770
},
{
"epoch": 1.6471838469713072,
"grad_norm": 5.125,
"learning_rate": 0.00019744535324216127,
"loss": 0.4913,
"step": 775
},
{
"epoch": 1.6578108395324125,
"grad_norm": 0.7109375,
"learning_rate": 0.00019736128110648407,
"loss": 0.4751,
"step": 780
},
{
"epoch": 1.6684378320935176,
"grad_norm": 6.5,
"learning_rate": 0.00019727586637650373,
"loss": 0.4956,
"step": 785
},
{
"epoch": 1.6790648246546227,
"grad_norm": 1.1640625,
"learning_rate": 0.0001971891102300738,
"loss": 0.4918,
"step": 790
},
{
"epoch": 1.6896918172157278,
"grad_norm": 0.53125,
"learning_rate": 0.0001971010138635457,
"loss": 0.4869,
"step": 795
},
{
"epoch": 1.7003188097768331,
"grad_norm": 0.451171875,
"learning_rate": 0.00019701157849175228,
"loss": 0.4823,
"step": 800
},
{
"epoch": 1.7109458023379385,
"grad_norm": 0.6484375,
"learning_rate": 0.00019692080534799096,
"loss": 0.4922,
"step": 805
},
{
"epoch": 1.7215727948990436,
"grad_norm": 0.80078125,
"learning_rate": 0.00019682869568400684,
"loss": 0.4918,
"step": 810
},
{
"epoch": 1.7321997874601487,
"grad_norm": 0.5,
"learning_rate": 0.0001967352507699754,
"loss": 0.4834,
"step": 815
},
{
"epoch": 1.742826780021254,
"grad_norm": 6.21875,
"learning_rate": 0.00019664047189448493,
"loss": 0.4807,
"step": 820
},
{
"epoch": 1.7534537725823593,
"grad_norm": 0.78515625,
"learning_rate": 0.0001965443603645189,
"loss": 0.486,
"step": 825
},
{
"epoch": 1.7640807651434645,
"grad_norm": 1.3046875,
"learning_rate": 0.00019644691750543767,
"loss": 0.5035,
"step": 830
},
{
"epoch": 1.7747077577045696,
"grad_norm": 1.40625,
"learning_rate": 0.00019634814466096056,
"loss": 0.4976,
"step": 835
},
{
"epoch": 1.7853347502656747,
"grad_norm": 0.78515625,
"learning_rate": 0.00019624804319314705,
"loss": 0.5016,
"step": 840
},
{
"epoch": 1.79596174282678,
"grad_norm": 1.125,
"learning_rate": 0.0001961466144823781,
"loss": 0.4952,
"step": 845
},
{
"epoch": 1.8065887353878853,
"grad_norm": 0.5546875,
"learning_rate": 0.00019604385992733715,
"loss": 0.4849,
"step": 850
},
{
"epoch": 1.8172157279489904,
"grad_norm": 0.61328125,
"learning_rate": 0.00019593978094499076,
"loss": 0.4792,
"step": 855
},
{
"epoch": 1.8278427205100956,
"grad_norm": 0.494140625,
"learning_rate": 0.00019583437897056915,
"loss": 0.4858,
"step": 860
},
{
"epoch": 1.8384697130712009,
"grad_norm": 0.85546875,
"learning_rate": 0.00019572765545754626,
"loss": 0.4905,
"step": 865
},
{
"epoch": 1.8490967056323062,
"grad_norm": 1.3359375,
"learning_rate": 0.00019561961187761985,
"loss": 0.4786,
"step": 870
},
{
"epoch": 1.8597236981934113,
"grad_norm": 0.5078125,
"learning_rate": 0.00019551024972069126,
"loss": 0.4694,
"step": 875
},
{
"epoch": 1.8703506907545164,
"grad_norm": 0.91796875,
"learning_rate": 0.00019539957049484458,
"loss": 0.474,
"step": 880
},
{
"epoch": 1.8809776833156215,
"grad_norm": 1.5,
"learning_rate": 0.00019528757572632622,
"loss": 0.4803,
"step": 885
},
{
"epoch": 1.8916046758767269,
"grad_norm": 0.890625,
"learning_rate": 0.00019517426695952358,
"loss": 0.4798,
"step": 890
},
{
"epoch": 1.9022316684378322,
"grad_norm": 0.57421875,
"learning_rate": 0.00019505964575694385,
"loss": 0.4779,
"step": 895
},
{
"epoch": 1.9128586609989373,
"grad_norm": 0.9296875,
"learning_rate": 0.0001949437136991925,
"loss": 0.4796,
"step": 900
},
{
"epoch": 1.9234856535600424,
"grad_norm": 0.6484375,
"learning_rate": 0.00019482647238495152,
"loss": 0.4808,
"step": 905
},
{
"epoch": 1.9341126461211477,
"grad_norm": 0.69140625,
"learning_rate": 0.00019470792343095718,
"loss": 0.4837,
"step": 910
},
{
"epoch": 1.944739638682253,
"grad_norm": 0.4921875,
"learning_rate": 0.000194588068471978,
"loss": 0.4861,
"step": 915
},
{
"epoch": 1.9553666312433582,
"grad_norm": 0.5859375,
"learning_rate": 0.0001944669091607919,
"loss": 0.4632,
"step": 920
},
{
"epoch": 1.9659936238044633,
"grad_norm": 0.76171875,
"learning_rate": 0.00019434444716816374,
"loss": 0.4767,
"step": 925
},
{
"epoch": 1.9766206163655684,
"grad_norm": 0.47265625,
"learning_rate": 0.00019422068418282202,
"loss": 0.465,
"step": 930
},
{
"epoch": 1.9872476089266737,
"grad_norm": 0.63671875,
"learning_rate": 0.00019409562191143577,
"loss": 0.4613,
"step": 935
},
{
"epoch": 1.997874601487779,
"grad_norm": 0.484375,
"learning_rate": 0.00019396926207859084,
"loss": 0.4826,
"step": 940
},
{
"epoch": 2.0,
"eval_loss": 1.3435465097427368,
"eval_runtime": 0.7971,
"eval_samples_per_second": 6.273,
"eval_steps_per_second": 1.255,
"step": 941
},
{
"epoch": 2.008501594048884,
"grad_norm": 0.734375,
"learning_rate": 0.00019384160642676635,
"loss": 0.4356,
"step": 945
},
{
"epoch": 2.0191285866099893,
"grad_norm": 0.5625,
"learning_rate": 0.00019371265671631037,
"loss": 0.4104,
"step": 950
},
{
"epoch": 2.0297555791710944,
"grad_norm": 0.484375,
"learning_rate": 0.0001935824147254159,
"loss": 0.4108,
"step": 955
},
{
"epoch": 2.0403825717322,
"grad_norm": 0.59765625,
"learning_rate": 0.00019345088225009626,
"loss": 0.4187,
"step": 960
},
{
"epoch": 2.051009564293305,
"grad_norm": 0.470703125,
"learning_rate": 0.00019331806110416027,
"loss": 0.4235,
"step": 965
},
{
"epoch": 2.06163655685441,
"grad_norm": 0.42578125,
"learning_rate": 0.0001931839531191873,
"loss": 0.4266,
"step": 970
},
{
"epoch": 2.0722635494155153,
"grad_norm": 0.7421875,
"learning_rate": 0.000193048560144502,
"loss": 0.4157,
"step": 975
},
{
"epoch": 2.082890541976621,
"grad_norm": 0.84765625,
"learning_rate": 0.00019291188404714878,
"loss": 0.4255,
"step": 980
},
{
"epoch": 2.093517534537726,
"grad_norm": 0.7421875,
"learning_rate": 0.00019277392671186608,
"loss": 0.4309,
"step": 985
},
{
"epoch": 2.104144527098831,
"grad_norm": 1.0234375,
"learning_rate": 0.0001926346900410604,
"loss": 0.4288,
"step": 990
},
{
"epoch": 2.114771519659936,
"grad_norm": 1.8359375,
"learning_rate": 0.00019249417595478002,
"loss": 0.4208,
"step": 995
},
{
"epoch": 2.1253985122210413,
"grad_norm": 0.427734375,
"learning_rate": 0.00019235238639068856,
"loss": 0.4249,
"step": 1000
},
{
"epoch": 2.136025504782147,
"grad_norm": 0.61328125,
"learning_rate": 0.00019220932330403823,
"loss": 0.427,
"step": 1005
},
{
"epoch": 2.146652497343252,
"grad_norm": 0.6875,
"learning_rate": 0.00019206498866764288,
"loss": 0.4276,
"step": 1010
},
{
"epoch": 2.157279489904357,
"grad_norm": 0.50390625,
"learning_rate": 0.00019191938447185084,
"loss": 0.4201,
"step": 1015
},
{
"epoch": 2.167906482465462,
"grad_norm": 1.6328125,
"learning_rate": 0.0001917725127245174,
"loss": 0.4299,
"step": 1020
},
{
"epoch": 2.1785334750265677,
"grad_norm": 0.5078125,
"learning_rate": 0.00019162437545097719,
"loss": 0.4268,
"step": 1025
},
{
"epoch": 2.189160467587673,
"grad_norm": 1.2578125,
"learning_rate": 0.0001914749746940161,
"loss": 0.4241,
"step": 1030
},
{
"epoch": 2.199787460148778,
"grad_norm": 0.69140625,
"learning_rate": 0.00019132431251384335,
"loss": 0.4297,
"step": 1035
},
{
"epoch": 2.210414452709883,
"grad_norm": 0.423828125,
"learning_rate": 0.00019117239098806295,
"loss": 0.4275,
"step": 1040
},
{
"epoch": 2.221041445270988,
"grad_norm": 0.55859375,
"learning_rate": 0.000191019212211645,
"loss": 0.4293,
"step": 1045
},
{
"epoch": 2.2316684378320937,
"grad_norm": 0.427734375,
"learning_rate": 0.00019086477829689685,
"loss": 0.421,
"step": 1050
},
{
"epoch": 2.242295430393199,
"grad_norm": 0.419921875,
"learning_rate": 0.00019070909137343408,
"loss": 0.4312,
"step": 1055
},
{
"epoch": 2.252922422954304,
"grad_norm": 0.42578125,
"learning_rate": 0.0001905521535881509,
"loss": 0.4299,
"step": 1060
},
{
"epoch": 2.263549415515409,
"grad_norm": 0.5859375,
"learning_rate": 0.00019039396710519076,
"loss": 0.4291,
"step": 1065
},
{
"epoch": 2.2741764080765146,
"grad_norm": 0.404296875,
"learning_rate": 0.00019023453410591635,
"loss": 0.4236,
"step": 1070
},
{
"epoch": 2.2848034006376197,
"grad_norm": 0.51171875,
"learning_rate": 0.00019007385678887975,
"loss": 0.4201,
"step": 1075
},
{
"epoch": 2.2954303931987248,
"grad_norm": 0.47265625,
"learning_rate": 0.00018991193736979175,
"loss": 0.4257,
"step": 1080
},
{
"epoch": 2.30605738575983,
"grad_norm": 0.439453125,
"learning_rate": 0.0001897487780814916,
"loss": 0.4219,
"step": 1085
},
{
"epoch": 2.316684378320935,
"grad_norm": 0.50390625,
"learning_rate": 0.00018958438117391618,
"loss": 0.4212,
"step": 1090
},
{
"epoch": 2.3273113708820405,
"grad_norm": 0.4765625,
"learning_rate": 0.00018941874891406882,
"loss": 0.4351,
"step": 1095
},
{
"epoch": 2.3379383634431457,
"grad_norm": 0.474609375,
"learning_rate": 0.00018925188358598813,
"loss": 0.4141,
"step": 1100
},
{
"epoch": 2.3485653560042508,
"grad_norm": 1.8515625,
"learning_rate": 0.0001890837874907166,
"loss": 0.4243,
"step": 1105
},
{
"epoch": 2.359192348565356,
"grad_norm": 0.66015625,
"learning_rate": 0.00018891446294626866,
"loss": 0.4408,
"step": 1110
},
{
"epoch": 2.369819341126461,
"grad_norm": 0.53515625,
"learning_rate": 0.00018874391228759893,
"loss": 0.4277,
"step": 1115
},
{
"epoch": 2.3804463336875665,
"grad_norm": 0.546875,
"learning_rate": 0.00018857213786656985,
"loss": 0.4279,
"step": 1120
},
{
"epoch": 2.3910733262486716,
"grad_norm": 0.77734375,
"learning_rate": 0.00018839914205191936,
"loss": 0.4259,
"step": 1125
},
{
"epoch": 2.4017003188097767,
"grad_norm": 0.79296875,
"learning_rate": 0.0001882249272292282,
"loss": 0.4287,
"step": 1130
},
{
"epoch": 2.412327311370882,
"grad_norm": 0.76953125,
"learning_rate": 0.00018804949580088692,
"loss": 0.4394,
"step": 1135
},
{
"epoch": 2.4229543039319874,
"grad_norm": 0.435546875,
"learning_rate": 0.00018787285018606297,
"loss": 0.4268,
"step": 1140
},
{
"epoch": 2.4335812964930925,
"grad_norm": 0.462890625,
"learning_rate": 0.00018769499282066717,
"loss": 0.4271,
"step": 1145
},
{
"epoch": 2.4442082890541976,
"grad_norm": 0.6484375,
"learning_rate": 0.00018751592615732005,
"loss": 0.4338,
"step": 1150
},
{
"epoch": 2.4548352816153027,
"grad_norm": 0.89453125,
"learning_rate": 0.0001873356526653183,
"loss": 0.4306,
"step": 1155
},
{
"epoch": 2.4654622741764083,
"grad_norm": 1.09375,
"learning_rate": 0.0001871541748306005,
"loss": 0.4374,
"step": 1160
},
{
"epoch": 2.4760892667375134,
"grad_norm": 0.44921875,
"learning_rate": 0.00018697149515571284,
"loss": 0.429,
"step": 1165
},
{
"epoch": 2.4867162592986185,
"grad_norm": 1.3203125,
"learning_rate": 0.00018678761615977468,
"loss": 0.4351,
"step": 1170
},
{
"epoch": 2.4973432518597236,
"grad_norm": 0.47265625,
"learning_rate": 0.00018660254037844388,
"loss": 0.4264,
"step": 1175
},
{
"epoch": 2.5079702444208287,
"grad_norm": 0.58203125,
"learning_rate": 0.00018641627036388169,
"loss": 0.431,
"step": 1180
},
{
"epoch": 2.5185972369819343,
"grad_norm": 0.41015625,
"learning_rate": 0.00018622880868471756,
"loss": 0.4265,
"step": 1185
},
{
"epoch": 2.5292242295430394,
"grad_norm": 0.400390625,
"learning_rate": 0.00018604015792601396,
"loss": 0.4259,
"step": 1190
},
{
"epoch": 2.5398512221041445,
"grad_norm": 7.6875,
"learning_rate": 0.00018585032068923032,
"loss": 0.4309,
"step": 1195
},
{
"epoch": 2.5504782146652496,
"grad_norm": 0.4296875,
"learning_rate": 0.00018565929959218758,
"loss": 0.4254,
"step": 1200
},
{
"epoch": 2.5611052072263547,
"grad_norm": 0.42578125,
"learning_rate": 0.00018546709726903178,
"loss": 0.4253,
"step": 1205
},
{
"epoch": 2.5717321997874603,
"grad_norm": 1.0625,
"learning_rate": 0.0001852737163701979,
"loss": 0.4403,
"step": 1210
},
{
"epoch": 2.5823591923485654,
"grad_norm": 0.640625,
"learning_rate": 0.00018507915956237326,
"loss": 0.4326,
"step": 1215
},
{
"epoch": 2.5929861849096705,
"grad_norm": 2.25,
"learning_rate": 0.00018488342952846073,
"loss": 0.437,
"step": 1220
},
{
"epoch": 2.603613177470776,
"grad_norm": 0.90625,
"learning_rate": 0.00018468652896754177,
"loss": 0.4344,
"step": 1225
},
{
"epoch": 2.614240170031881,
"grad_norm": 0.87890625,
"learning_rate": 0.0001844884605948392,
"loss": 0.4387,
"step": 1230
},
{
"epoch": 2.6248671625929862,
"grad_norm": 0.66796875,
"learning_rate": 0.0001842892271416797,
"loss": 0.4371,
"step": 1235
},
{
"epoch": 2.6354941551540914,
"grad_norm": 0.515625,
"learning_rate": 0.00018408883135545632,
"loss": 0.4345,
"step": 1240
},
{
"epoch": 2.6461211477151965,
"grad_norm": 0.8515625,
"learning_rate": 0.00018388727599959033,
"loss": 0.4405,
"step": 1245
},
{
"epoch": 2.656748140276302,
"grad_norm": 1.078125,
"learning_rate": 0.00018368456385349334,
"loss": 0.4369,
"step": 1250
},
{
"epoch": 2.667375132837407,
"grad_norm": 0.50390625,
"learning_rate": 0.0001834806977125288,
"loss": 0.4354,
"step": 1255
},
{
"epoch": 2.6780021253985122,
"grad_norm": 0.41796875,
"learning_rate": 0.0001832756803879737,
"loss": 0.4297,
"step": 1260
},
{
"epoch": 2.6886291179596173,
"grad_norm": 0.578125,
"learning_rate": 0.00018306951470697946,
"loss": 0.4268,
"step": 1265
},
{
"epoch": 2.6992561105207225,
"grad_norm": 0.44140625,
"learning_rate": 0.0001828622035125332,
"loss": 0.4299,
"step": 1270
},
{
"epoch": 2.709883103081828,
"grad_norm": 0.8046875,
"learning_rate": 0.0001826537496634186,
"loss": 0.4362,
"step": 1275
},
{
"epoch": 2.720510095642933,
"grad_norm": 1.484375,
"learning_rate": 0.00018244415603417603,
"loss": 0.4364,
"step": 1280
},
{
"epoch": 2.731137088204038,
"grad_norm": 0.953125,
"learning_rate": 0.0001822334255150635,
"loss": 0.439,
"step": 1285
},
{
"epoch": 2.7417640807651433,
"grad_norm": 0.921875,
"learning_rate": 0.00018202156101201645,
"loss": 0.4342,
"step": 1290
},
{
"epoch": 2.7523910733262484,
"grad_norm": 0.55859375,
"learning_rate": 0.0001818085654466076,
"loss": 0.4308,
"step": 1295
},
{
"epoch": 2.763018065887354,
"grad_norm": 0.494140625,
"learning_rate": 0.00018159444175600703,
"loss": 0.4306,
"step": 1300
},
{
"epoch": 2.773645058448459,
"grad_norm": 1.125,
"learning_rate": 0.00018137919289294135,
"loss": 0.4342,
"step": 1305
},
{
"epoch": 2.784272051009564,
"grad_norm": 0.53125,
"learning_rate": 0.00018116282182565311,
"loss": 0.428,
"step": 1310
},
{
"epoch": 2.7948990435706698,
"grad_norm": 0.50390625,
"learning_rate": 0.00018094533153785984,
"loss": 0.4221,
"step": 1315
},
{
"epoch": 2.805526036131775,
"grad_norm": 1.5859375,
"learning_rate": 0.00018072672502871296,
"loss": 0.433,
"step": 1320
},
{
"epoch": 2.81615302869288,
"grad_norm": 0.44140625,
"learning_rate": 0.0001805070053127563,
"loss": 0.4306,
"step": 1325
},
{
"epoch": 2.826780021253985,
"grad_norm": 0.6171875,
"learning_rate": 0.00018028617541988472,
"loss": 0.4217,
"step": 1330
},
{
"epoch": 2.83740701381509,
"grad_norm": 0.49609375,
"learning_rate": 0.00018006423839530205,
"loss": 0.426,
"step": 1335
},
{
"epoch": 2.8480340063761957,
"grad_norm": 0.6484375,
"learning_rate": 0.00017984119729947944,
"loss": 0.4402,
"step": 1340
},
{
"epoch": 2.858660998937301,
"grad_norm": 0.4609375,
"learning_rate": 0.00017961705520811293,
"loss": 0.4306,
"step": 1345
},
{
"epoch": 2.869287991498406,
"grad_norm": 0.375,
"learning_rate": 0.000179391815212081,
"loss": 0.434,
"step": 1350
},
{
"epoch": 2.879914984059511,
"grad_norm": 0.70703125,
"learning_rate": 0.00017916548041740213,
"loss": 0.4295,
"step": 1355
},
{
"epoch": 2.890541976620616,
"grad_norm": 0.72265625,
"learning_rate": 0.0001789380539451919,
"loss": 0.4373,
"step": 1360
},
{
"epoch": 2.9011689691817217,
"grad_norm": 0.4296875,
"learning_rate": 0.00017870953893161975,
"loss": 0.4287,
"step": 1365
},
{
"epoch": 2.911795961742827,
"grad_norm": 0.5625,
"learning_rate": 0.0001784799385278661,
"loss": 0.4358,
"step": 1370
},
{
"epoch": 2.922422954303932,
"grad_norm": 0.419921875,
"learning_rate": 0.00017824925590007864,
"loss": 0.4304,
"step": 1375
},
{
"epoch": 2.933049946865037,
"grad_norm": 0.98046875,
"learning_rate": 0.0001780174942293287,
"loss": 0.4296,
"step": 1380
},
{
"epoch": 2.943676939426142,
"grad_norm": 0.6796875,
"learning_rate": 0.00017778465671156743,
"loss": 0.4326,
"step": 1385
},
{
"epoch": 2.9543039319872477,
"grad_norm": 0.486328125,
"learning_rate": 0.00017755074655758174,
"loss": 0.4348,
"step": 1390
},
{
"epoch": 2.964930924548353,
"grad_norm": 0.61328125,
"learning_rate": 0.0001773157669929499,
"loss": 0.4345,
"step": 1395
},
{
"epoch": 2.975557917109458,
"grad_norm": 0.51171875,
"learning_rate": 0.00017707972125799735,
"loss": 0.4259,
"step": 1400
},
{
"epoch": 2.9861849096705635,
"grad_norm": 0.5859375,
"learning_rate": 0.00017684261260775163,
"loss": 0.4306,
"step": 1405
},
{
"epoch": 2.9968119022316686,
"grad_norm": 0.60546875,
"learning_rate": 0.0001766044443118978,
"loss": 0.4369,
"step": 1410
},
{
"epoch": 2.9989373007438895,
"eval_loss": 1.4787013530731201,
"eval_runtime": 0.8123,
"eval_samples_per_second": 6.155,
"eval_steps_per_second": 1.231,
"step": 1411
},
{
"epoch": 3.0074388947927737,
"grad_norm": 0.54296875,
"learning_rate": 0.00017636521965473323,
"loss": 0.3835,
"step": 1415
},
{
"epoch": 3.018065887353879,
"grad_norm": 0.455078125,
"learning_rate": 0.0001761249419351222,
"loss": 0.3594,
"step": 1420
},
{
"epoch": 3.028692879914984,
"grad_norm": 0.478515625,
"learning_rate": 0.00017588361446645073,
"loss": 0.3596,
"step": 1425
},
{
"epoch": 3.0393198724760895,
"grad_norm": 0.51171875,
"learning_rate": 0.00017564124057658056,
"loss": 0.3504,
"step": 1430
},
{
"epoch": 3.0499468650371946,
"grad_norm": 0.4375,
"learning_rate": 0.00017539782360780334,
"loss": 0.3575,
"step": 1435
},
{
"epoch": 3.0605738575982997,
"grad_norm": 0.419921875,
"learning_rate": 0.00017515336691679477,
"loss": 0.3619,
"step": 1440
},
{
"epoch": 3.071200850159405,
"grad_norm": 0.41015625,
"learning_rate": 0.0001749078738745679,
"loss": 0.3678,
"step": 1445
},
{
"epoch": 3.08182784272051,
"grad_norm": 0.57421875,
"learning_rate": 0.0001746613478664271,
"loss": 0.3551,
"step": 1450
},
{
"epoch": 3.0924548352816155,
"grad_norm": 0.75,
"learning_rate": 0.00017441379229192098,
"loss": 0.3668,
"step": 1455
},
{
"epoch": 3.1030818278427206,
"grad_norm": 0.498046875,
"learning_rate": 0.00017416521056479577,
"loss": 0.3664,
"step": 1460
},
{
"epoch": 3.1137088204038257,
"grad_norm": 1.6640625,
"learning_rate": 0.0001739156061129481,
"loss": 0.3709,
"step": 1465
},
{
"epoch": 3.124335812964931,
"grad_norm": 0.8125,
"learning_rate": 0.0001736649823783779,
"loss": 0.3612,
"step": 1470
},
{
"epoch": 3.134962805526036,
"grad_norm": 0.64453125,
"learning_rate": 0.00017341334281714064,
"loss": 0.3619,
"step": 1475
},
{
"epoch": 3.1455897980871415,
"grad_norm": 1.0859375,
"learning_rate": 0.00017316069089930007,
"loss": 0.37,
"step": 1480
},
{
"epoch": 3.1562167906482466,
"grad_norm": 0.5859375,
"learning_rate": 0.00017290703010887994,
"loss": 0.3717,
"step": 1485
},
{
"epoch": 3.1668437832093517,
"grad_norm": 0.53515625,
"learning_rate": 0.00017265236394381633,
"loss": 0.3749,
"step": 1490
},
{
"epoch": 3.177470775770457,
"grad_norm": 0.5,
"learning_rate": 0.00017239669591590916,
"loss": 0.3754,
"step": 1495
},
{
"epoch": 3.1880977683315623,
"grad_norm": 0.55859375,
"learning_rate": 0.00017214002955077393,
"loss": 0.3726,
"step": 1500
},
{
"epoch": 3.1987247608926674,
"grad_norm": 0.53125,
"learning_rate": 0.00017188236838779295,
"loss": 0.3726,
"step": 1505
},
{
"epoch": 3.2093517534537725,
"grad_norm": 0.7265625,
"learning_rate": 0.00017162371598006666,
"loss": 0.3798,
"step": 1510
},
{
"epoch": 3.2199787460148777,
"grad_norm": 0.94140625,
"learning_rate": 0.00017136407589436457,
"loss": 0.3767,
"step": 1515
},
{
"epoch": 3.230605738575983,
"grad_norm": 0.5703125,
"learning_rate": 0.0001711034517110761,
"loss": 0.3763,
"step": 1520
},
{
"epoch": 3.2412327311370883,
"grad_norm": 0.53125,
"learning_rate": 0.00017084184702416115,
"loss": 0.3703,
"step": 1525
},
{
"epoch": 3.2518597236981934,
"grad_norm": 1.0234375,
"learning_rate": 0.0001705792654411007,
"loss": 0.378,
"step": 1530
},
{
"epoch": 3.2624867162592985,
"grad_norm": 0.5234375,
"learning_rate": 0.00017031571058284678,
"loss": 0.3818,
"step": 1535
},
{
"epoch": 3.2731137088204036,
"grad_norm": 0.55859375,
"learning_rate": 0.00017005118608377288,
"loss": 0.3709,
"step": 1540
},
{
"epoch": 3.283740701381509,
"grad_norm": 0.59375,
"learning_rate": 0.00016978569559162357,
"loss": 0.3761,
"step": 1545
},
{
"epoch": 3.2943676939426143,
"grad_norm": 0.44921875,
"learning_rate": 0.00016951924276746425,
"loss": 0.3786,
"step": 1550
},
{
"epoch": 3.3049946865037194,
"grad_norm": 0.451171875,
"learning_rate": 0.00016925183128563078,
"loss": 0.3661,
"step": 1555
},
{
"epoch": 3.3156216790648245,
"grad_norm": 0.490234375,
"learning_rate": 0.00016898346483367867,
"loss": 0.3721,
"step": 1560
},
{
"epoch": 3.3262486716259296,
"grad_norm": 0.4375,
"learning_rate": 0.0001687141471123324,
"loss": 0.3752,
"step": 1565
},
{
"epoch": 3.336875664187035,
"grad_norm": 0.4921875,
"learning_rate": 0.00016844388183543418,
"loss": 0.3784,
"step": 1570
},
{
"epoch": 3.3475026567481403,
"grad_norm": 0.44921875,
"learning_rate": 0.00016817267272989286,
"loss": 0.3807,
"step": 1575
},
{
"epoch": 3.3581296493092454,
"grad_norm": 0.55859375,
"learning_rate": 0.00016790052353563253,
"loss": 0.3887,
"step": 1580
},
{
"epoch": 3.3687566418703505,
"grad_norm": 0.6484375,
"learning_rate": 0.0001676274380055409,
"loss": 0.3831,
"step": 1585
},
{
"epoch": 3.379383634431456,
"grad_norm": 0.45703125,
"learning_rate": 0.00016735341990541764,
"loss": 0.3772,
"step": 1590
},
{
"epoch": 3.390010626992561,
"grad_norm": 0.56640625,
"learning_rate": 0.00016707847301392236,
"loss": 0.3799,
"step": 1595
},
{
"epoch": 3.4006376195536663,
"grad_norm": 0.431640625,
"learning_rate": 0.0001668026011225225,
"loss": 0.3854,
"step": 1600
},
{
"epoch": 3.4112646121147714,
"grad_norm": 0.439453125,
"learning_rate": 0.00016652580803544112,
"loss": 0.3725,
"step": 1605
},
{
"epoch": 3.421891604675877,
"grad_norm": 0.439453125,
"learning_rate": 0.00016624809756960444,
"loss": 0.3785,
"step": 1610
},
{
"epoch": 3.432518597236982,
"grad_norm": 0.490234375,
"learning_rate": 0.00016596947355458904,
"loss": 0.3803,
"step": 1615
},
{
"epoch": 3.443145589798087,
"grad_norm": 0.640625,
"learning_rate": 0.0001656899398325693,
"loss": 0.3893,
"step": 1620
},
{
"epoch": 3.4537725823591923,
"grad_norm": 0.53515625,
"learning_rate": 0.00016540950025826422,
"loss": 0.3776,
"step": 1625
},
{
"epoch": 3.4643995749202974,
"grad_norm": 0.494140625,
"learning_rate": 0.0001651281586988844,
"loss": 0.3802,
"step": 1630
},
{
"epoch": 3.475026567481403,
"grad_norm": 0.73046875,
"learning_rate": 0.00016484591903407857,
"loss": 0.3765,
"step": 1635
},
{
"epoch": 3.485653560042508,
"grad_norm": 0.57421875,
"learning_rate": 0.00016456278515588024,
"loss": 0.3843,
"step": 1640
},
{
"epoch": 3.496280552603613,
"grad_norm": 0.59375,
"learning_rate": 0.00016427876096865394,
"loss": 0.3748,
"step": 1645
},
{
"epoch": 3.5069075451647183,
"grad_norm": 0.4609375,
"learning_rate": 0.00016399385038904138,
"loss": 0.3785,
"step": 1650
},
{
"epoch": 3.5175345377258234,
"grad_norm": 0.466796875,
"learning_rate": 0.00016370805734590747,
"loss": 0.3879,
"step": 1655
},
{
"epoch": 3.528161530286929,
"grad_norm": 0.4609375,
"learning_rate": 0.00016342138578028613,
"loss": 0.382,
"step": 1660
},
{
"epoch": 3.538788522848034,
"grad_norm": 0.423828125,
"learning_rate": 0.00016313383964532596,
"loss": 0.3915,
"step": 1665
},
{
"epoch": 3.549415515409139,
"grad_norm": 0.42578125,
"learning_rate": 0.00016284542290623567,
"loss": 0.3767,
"step": 1670
},
{
"epoch": 3.5600425079702447,
"grad_norm": 0.443359375,
"learning_rate": 0.0001625561395402295,
"loss": 0.373,
"step": 1675
},
{
"epoch": 3.57066950053135,
"grad_norm": 0.4453125,
"learning_rate": 0.00016226599353647228,
"loss": 0.3811,
"step": 1680
},
{
"epoch": 3.581296493092455,
"grad_norm": 0.43359375,
"learning_rate": 0.00016197498889602448,
"loss": 0.3842,
"step": 1685
},
{
"epoch": 3.59192348565356,
"grad_norm": 0.59765625,
"learning_rate": 0.00016168312963178697,
"loss": 0.3864,
"step": 1690
},
{
"epoch": 3.602550478214665,
"grad_norm": 1.2578125,
"learning_rate": 0.00016139041976844583,
"loss": 0.3835,
"step": 1695
},
{
"epoch": 3.6131774707757707,
"grad_norm": 0.578125,
"learning_rate": 0.00016109686334241655,
"loss": 0.3801,
"step": 1700
},
{
"epoch": 3.623804463336876,
"grad_norm": 0.5078125,
"learning_rate": 0.00016080246440178874,
"loss": 0.382,
"step": 1705
},
{
"epoch": 3.634431455897981,
"grad_norm": 0.66796875,
"learning_rate": 0.00016050722700627012,
"loss": 0.3923,
"step": 1710
},
{
"epoch": 3.645058448459086,
"grad_norm": 0.75390625,
"learning_rate": 0.00016021115522713047,
"loss": 0.3836,
"step": 1715
},
{
"epoch": 3.655685441020191,
"grad_norm": 0.77734375,
"learning_rate": 0.0001599142531471456,
"loss": 0.3801,
"step": 1720
},
{
"epoch": 3.6663124335812967,
"grad_norm": 0.61328125,
"learning_rate": 0.00015961652486054103,
"loss": 0.3871,
"step": 1725
},
{
"epoch": 3.6769394261424018,
"grad_norm": 0.51171875,
"learning_rate": 0.00015931797447293552,
"loss": 0.3843,
"step": 1730
},
{
"epoch": 3.687566418703507,
"grad_norm": 0.4765625,
"learning_rate": 0.00015901860610128448,
"loss": 0.3846,
"step": 1735
},
{
"epoch": 3.698193411264612,
"grad_norm": 0.50390625,
"learning_rate": 0.00015871842387382305,
"loss": 0.3788,
"step": 1740
},
{
"epoch": 3.708820403825717,
"grad_norm": 0.66015625,
"learning_rate": 0.00015841743193000944,
"loss": 0.3899,
"step": 1745
},
{
"epoch": 3.7194473963868226,
"grad_norm": 0.50390625,
"learning_rate": 0.00015811563442046767,
"loss": 0.3863,
"step": 1750
},
{
"epoch": 3.7300743889479278,
"grad_norm": 0.396484375,
"learning_rate": 0.00015781303550693024,
"loss": 0.3822,
"step": 1755
},
{
"epoch": 3.740701381509033,
"grad_norm": 0.5859375,
"learning_rate": 0.00015750963936218105,
"loss": 0.3821,
"step": 1760
},
{
"epoch": 3.7513283740701384,
"grad_norm": 0.70703125,
"learning_rate": 0.00015720545016999752,
"loss": 0.3809,
"step": 1765
},
{
"epoch": 3.761955366631243,
"grad_norm": 0.515625,
"learning_rate": 0.00015690047212509316,
"loss": 0.3809,
"step": 1770
},
{
"epoch": 3.7725823591923486,
"grad_norm": 0.5859375,
"learning_rate": 0.00015659470943305955,
"loss": 0.3815,
"step": 1775
},
{
"epoch": 3.7832093517534537,
"grad_norm": 0.443359375,
"learning_rate": 0.00015628816631030836,
"loss": 0.3808,
"step": 1780
},
{
"epoch": 3.793836344314559,
"grad_norm": 0.44140625,
"learning_rate": 0.00015598084698401342,
"loss": 0.38,
"step": 1785
},
{
"epoch": 3.8044633368756644,
"grad_norm": 0.53125,
"learning_rate": 0.00015567275569205218,
"loss": 0.3851,
"step": 1790
},
{
"epoch": 3.8150903294367695,
"grad_norm": 0.44140625,
"learning_rate": 0.00015536389668294724,
"loss": 0.3917,
"step": 1795
},
{
"epoch": 3.8257173219978746,
"grad_norm": 0.40625,
"learning_rate": 0.00015505427421580808,
"loss": 0.3799,
"step": 1800
},
{
"epoch": 3.8363443145589797,
"grad_norm": 0.45703125,
"learning_rate": 0.000154743892560272,
"loss": 0.3846,
"step": 1805
},
{
"epoch": 3.846971307120085,
"grad_norm": 0.46484375,
"learning_rate": 0.00015443275599644538,
"loss": 0.3865,
"step": 1810
},
{
"epoch": 3.8575982996811904,
"grad_norm": 0.486328125,
"learning_rate": 0.0001541208688148447,
"loss": 0.3802,
"step": 1815
},
{
"epoch": 3.8682252922422955,
"grad_norm": 0.462890625,
"learning_rate": 0.00015380823531633729,
"loss": 0.3823,
"step": 1820
},
{
"epoch": 3.8788522848034006,
"grad_norm": 0.61328125,
"learning_rate": 0.00015349485981208202,
"loss": 0.3866,
"step": 1825
},
{
"epoch": 3.8894792773645057,
"grad_norm": 0.490234375,
"learning_rate": 0.00015318074662346994,
"loss": 0.3758,
"step": 1830
},
{
"epoch": 3.900106269925611,
"grad_norm": 0.5625,
"learning_rate": 0.00015286590008206465,
"loss": 0.3858,
"step": 1835
},
{
"epoch": 3.9107332624867164,
"grad_norm": 0.65625,
"learning_rate": 0.00015255032452954245,
"loss": 0.3867,
"step": 1840
},
{
"epoch": 3.9213602550478215,
"grad_norm": 0.42578125,
"learning_rate": 0.00015223402431763269,
"loss": 0.3893,
"step": 1845
},
{
"epoch": 3.9319872476089266,
"grad_norm": 0.58984375,
"learning_rate": 0.00015191700380805752,
"loss": 0.3791,
"step": 1850
},
{
"epoch": 3.942614240170032,
"grad_norm": 0.51953125,
"learning_rate": 0.00015159926737247202,
"loss": 0.3932,
"step": 1855
},
{
"epoch": 3.953241232731137,
"grad_norm": 0.69921875,
"learning_rate": 0.00015128081939240357,
"loss": 0.3794,
"step": 1860
},
{
"epoch": 3.9638682252922424,
"grad_norm": 0.54296875,
"learning_rate": 0.00015096166425919175,
"loss": 0.3857,
"step": 1865
},
{
"epoch": 3.9744952178533475,
"grad_norm": 0.412109375,
"learning_rate": 0.00015064180637392764,
"loss": 0.3887,
"step": 1870
},
{
"epoch": 3.9851222104144526,
"grad_norm": 0.51171875,
"learning_rate": 0.0001503212501473931,
"loss": 0.3886,
"step": 1875
},
{
"epoch": 3.995749202975558,
"grad_norm": 0.5078125,
"learning_rate": 0.00015000000000000001,
"loss": 0.3819,
"step": 1880
},
{
"epoch": 4.0,
"eval_loss": 1.7431671619415283,
"eval_runtime": 0.7941,
"eval_samples_per_second": 6.296,
"eval_steps_per_second": 1.259,
"step": 1882
},
{
"epoch": 4.006376195536663,
"grad_norm": 0.47265625,
"learning_rate": 0.0001496780603617293,
"loss": 0.3464,
"step": 1885
},
{
"epoch": 4.017003188097768,
"grad_norm": 0.46484375,
"learning_rate": 0.00014935543567206984,
"loss": 0.3125,
"step": 1890
},
{
"epoch": 4.027630180658874,
"grad_norm": 0.5546875,
"learning_rate": 0.00014903213037995724,
"loss": 0.3159,
"step": 1895
},
{
"epoch": 4.038257173219979,
"grad_norm": 0.478515625,
"learning_rate": 0.00014870814894371245,
"loss": 0.3227,
"step": 1900
},
{
"epoch": 4.048884165781084,
"grad_norm": 0.470703125,
"learning_rate": 0.00014838349583098045,
"loss": 0.3122,
"step": 1905
},
{
"epoch": 4.059511158342189,
"grad_norm": 0.43359375,
"learning_rate": 0.00014805817551866838,
"loss": 0.3194,
"step": 1910
},
{
"epoch": 4.070138150903294,
"grad_norm": 0.458984375,
"learning_rate": 0.00014773219249288402,
"loss": 0.3123,
"step": 1915
},
{
"epoch": 4.0807651434644,
"grad_norm": 0.4765625,
"learning_rate": 0.00014740555124887375,
"loss": 0.3254,
"step": 1920
},
{
"epoch": 4.0913921360255046,
"grad_norm": 0.478515625,
"learning_rate": 0.00014707825629096084,
"loss": 0.3178,
"step": 1925
},
{
"epoch": 4.10201912858661,
"grad_norm": 0.609375,
"learning_rate": 0.00014675031213248296,
"loss": 0.3214,
"step": 1930
},
{
"epoch": 4.112646121147715,
"grad_norm": 0.470703125,
"learning_rate": 0.00014642172329573026,
"loss": 0.321,
"step": 1935
},
{
"epoch": 4.12327311370882,
"grad_norm": 0.474609375,
"learning_rate": 0.00014609249431188278,
"loss": 0.3227,
"step": 1940
},
{
"epoch": 4.133900106269926,
"grad_norm": 0.470703125,
"learning_rate": 0.00014576262972094828,
"loss": 0.3277,
"step": 1945
},
{
"epoch": 4.1445270988310305,
"grad_norm": 0.578125,
"learning_rate": 0.0001454321340716992,
"loss": 0.3216,
"step": 1950
},
{
"epoch": 4.155154091392136,
"grad_norm": 0.4453125,
"learning_rate": 0.00014510101192161018,
"loss": 0.3196,
"step": 1955
},
{
"epoch": 4.165781083953242,
"grad_norm": 0.57421875,
"learning_rate": 0.00014476926783679538,
"loss": 0.3284,
"step": 1960
},
{
"epoch": 4.176408076514346,
"grad_norm": 0.46875,
"learning_rate": 0.00014443690639194515,
"loss": 0.3244,
"step": 1965
},
{
"epoch": 4.187035069075452,
"grad_norm": 0.80078125,
"learning_rate": 0.00014410393217026318,
"loss": 0.3205,
"step": 1970
},
{
"epoch": 4.1976620616365565,
"grad_norm": 0.71875,
"learning_rate": 0.0001437703497634032,
"loss": 0.3216,
"step": 1975
},
{
"epoch": 4.208289054197662,
"grad_norm": 0.76171875,
"learning_rate": 0.00014343616377140582,
"loss": 0.3178,
"step": 1980
},
{
"epoch": 4.218916046758768,
"grad_norm": 0.86328125,
"learning_rate": 0.00014310137880263482,
"loss": 0.3219,
"step": 1985
},
{
"epoch": 4.229543039319872,
"grad_norm": 0.5546875,
"learning_rate": 0.00014276599947371388,
"loss": 0.3262,
"step": 1990
},
{
"epoch": 4.240170031880978,
"grad_norm": 0.51171875,
"learning_rate": 0.00014243003040946274,
"loss": 0.3287,
"step": 1995
},
{
"epoch": 4.2507970244420825,
"grad_norm": 0.57421875,
"learning_rate": 0.0001420934762428335,
"loss": 0.328,
"step": 2000
},
{
"epoch": 4.261424017003188,
"grad_norm": 0.64453125,
"learning_rate": 0.00014175634161484675,
"loss": 0.3334,
"step": 2005
},
{
"epoch": 4.272051009564294,
"grad_norm": 0.515625,
"learning_rate": 0.00014141863117452745,
"loss": 0.3255,
"step": 2010
},
{
"epoch": 4.282678002125398,
"grad_norm": 0.419921875,
"learning_rate": 0.00014108034957884094,
"loss": 0.3226,
"step": 2015
},
{
"epoch": 4.293304994686504,
"grad_norm": 0.48046875,
"learning_rate": 0.0001407415014926288,
"loss": 0.3314,
"step": 2020
},
{
"epoch": 4.3039319872476085,
"grad_norm": 0.67578125,
"learning_rate": 0.00014040209158854426,
"loss": 0.3334,
"step": 2025
},
{
"epoch": 4.314558979808714,
"grad_norm": 0.48828125,
"learning_rate": 0.00014006212454698797,
"loss": 0.3349,
"step": 2030
},
{
"epoch": 4.32518597236982,
"grad_norm": 0.455078125,
"learning_rate": 0.00013972160505604342,
"loss": 0.3301,
"step": 2035
},
{
"epoch": 4.335812964930924,
"grad_norm": 0.6171875,
"learning_rate": 0.00013938053781141222,
"loss": 0.325,
"step": 2040
},
{
"epoch": 4.34643995749203,
"grad_norm": 0.5,
"learning_rate": 0.00013903892751634947,
"loss": 0.3293,
"step": 2045
},
{
"epoch": 4.357066950053135,
"grad_norm": 0.6484375,
"learning_rate": 0.00013869677888159887,
"loss": 0.3285,
"step": 2050
},
{
"epoch": 4.36769394261424,
"grad_norm": 0.71875,
"learning_rate": 0.00013835409662532762,
"loss": 0.3271,
"step": 2055
},
{
"epoch": 4.378320935175346,
"grad_norm": 0.515625,
"learning_rate": 0.00013801088547306148,
"loss": 0.3296,
"step": 2060
},
{
"epoch": 4.38894792773645,
"grad_norm": 0.6015625,
"learning_rate": 0.0001376671501576197,
"loss": 0.3292,
"step": 2065
},
{
"epoch": 4.399574920297556,
"grad_norm": 0.53515625,
"learning_rate": 0.00013732289541904948,
"loss": 0.3324,
"step": 2070
},
{
"epoch": 4.410201912858661,
"grad_norm": 0.55078125,
"learning_rate": 0.00013697812600456093,
"loss": 0.3326,
"step": 2075
},
{
"epoch": 4.420828905419766,
"grad_norm": 0.462890625,
"learning_rate": 0.00013663284666846134,
"loss": 0.3329,
"step": 2080
},
{
"epoch": 4.431455897980872,
"grad_norm": 0.65234375,
"learning_rate": 0.00013628706217208976,
"loss": 0.3305,
"step": 2085
},
{
"epoch": 4.442082890541976,
"grad_norm": 0.4765625,
"learning_rate": 0.00013594077728375128,
"loss": 0.3305,
"step": 2090
},
{
"epoch": 4.452709883103082,
"grad_norm": 0.451171875,
"learning_rate": 0.0001355939967786514,
"loss": 0.3327,
"step": 2095
},
{
"epoch": 4.463336875664187,
"grad_norm": 0.427734375,
"learning_rate": 0.00013524672543882996,
"loss": 0.3329,
"step": 2100
},
{
"epoch": 4.473963868225292,
"grad_norm": 0.63671875,
"learning_rate": 0.00013489896805309542,
"loss": 0.3376,
"step": 2105
},
{
"epoch": 4.484590860786398,
"grad_norm": 0.63671875,
"learning_rate": 0.00013455072941695863,
"loss": 0.329,
"step": 2110
},
{
"epoch": 4.495217853347503,
"grad_norm": 0.53125,
"learning_rate": 0.00013420201433256689,
"loss": 0.333,
"step": 2115
},
{
"epoch": 4.505844845908608,
"grad_norm": 0.470703125,
"learning_rate": 0.00013385282760863758,
"loss": 0.328,
"step": 2120
},
{
"epoch": 4.516471838469713,
"grad_norm": 0.439453125,
"learning_rate": 0.00013350317406039187,
"loss": 0.3321,
"step": 2125
},
{
"epoch": 4.527098831030818,
"grad_norm": 0.48046875,
"learning_rate": 0.00013315305850948846,
"loss": 0.3325,
"step": 2130
},
{
"epoch": 4.537725823591924,
"grad_norm": 0.48046875,
"learning_rate": 0.0001328024857839569,
"loss": 0.3324,
"step": 2135
},
{
"epoch": 4.548352816153029,
"grad_norm": 0.447265625,
"learning_rate": 0.00013245146071813114,
"loss": 0.3351,
"step": 2140
},
{
"epoch": 4.558979808714134,
"grad_norm": 0.41796875,
"learning_rate": 0.00013209998815258273,
"loss": 0.3264,
"step": 2145
},
{
"epoch": 4.569606801275239,
"grad_norm": 0.6640625,
"learning_rate": 0.00013174807293405428,
"loss": 0.3303,
"step": 2150
},
{
"epoch": 4.580233793836344,
"grad_norm": 0.5625,
"learning_rate": 0.00013139571991539238,
"loss": 0.3308,
"step": 2155
},
{
"epoch": 4.5908607863974495,
"grad_norm": 0.494140625,
"learning_rate": 0.00013104293395548098,
"loss": 0.3314,
"step": 2160
},
{
"epoch": 4.601487778958555,
"grad_norm": 0.45703125,
"learning_rate": 0.000130689719919174,
"loss": 0.3328,
"step": 2165
},
{
"epoch": 4.61211477151966,
"grad_norm": 0.453125,
"learning_rate": 0.00013033608267722858,
"loss": 0.3337,
"step": 2170
},
{
"epoch": 4.622741764080765,
"grad_norm": 0.54296875,
"learning_rate": 0.0001299820271062378,
"loss": 0.3343,
"step": 2175
},
{
"epoch": 4.63336875664187,
"grad_norm": 0.63671875,
"learning_rate": 0.00012962755808856342,
"loss": 0.3317,
"step": 2180
},
{
"epoch": 4.6439957492029755,
"grad_norm": 0.478515625,
"learning_rate": 0.00012927268051226844,
"loss": 0.3311,
"step": 2185
},
{
"epoch": 4.654622741764081,
"grad_norm": 0.55078125,
"learning_rate": 0.0001289173992710499,
"loss": 0.3347,
"step": 2190
},
{
"epoch": 4.665249734325186,
"grad_norm": 0.5859375,
"learning_rate": 0.00012856171926417133,
"loss": 0.3353,
"step": 2195
},
{
"epoch": 4.675876726886291,
"grad_norm": 0.59375,
"learning_rate": 0.00012820564539639512,
"loss": 0.3393,
"step": 2200
},
{
"epoch": 4.686503719447396,
"grad_norm": 0.73046875,
"learning_rate": 0.00012784918257791495,
"loss": 0.3373,
"step": 2205
},
{
"epoch": 4.6971307120085015,
"grad_norm": 0.7265625,
"learning_rate": 0.00012749233572428804,
"loss": 0.3271,
"step": 2210
},
{
"epoch": 4.707757704569607,
"grad_norm": 0.77734375,
"learning_rate": 0.00012713510975636741,
"loss": 0.3363,
"step": 2215
},
{
"epoch": 4.718384697130712,
"grad_norm": 0.67578125,
"learning_rate": 0.00012677750960023396,
"loss": 0.3355,
"step": 2220
},
{
"epoch": 4.729011689691817,
"grad_norm": 0.5703125,
"learning_rate": 0.00012641954018712863,
"loss": 0.3378,
"step": 2225
},
{
"epoch": 4.739638682252922,
"grad_norm": 0.484375,
"learning_rate": 0.0001260612064533843,
"loss": 0.3354,
"step": 2230
},
{
"epoch": 4.7502656748140275,
"grad_norm": 0.5,
"learning_rate": 0.0001257025133403577,
"loss": 0.3363,
"step": 2235
},
{
"epoch": 4.760892667375133,
"grad_norm": 0.57421875,
"learning_rate": 0.0001253434657943616,
"loss": 0.3296,
"step": 2240
},
{
"epoch": 4.771519659936238,
"grad_norm": 0.5546875,
"learning_rate": 0.00012498406876659598,
"loss": 0.3381,
"step": 2245
},
{
"epoch": 4.782146652497343,
"grad_norm": 0.5234375,
"learning_rate": 0.0001246243272130804,
"loss": 0.3222,
"step": 2250
},
{
"epoch": 4.792773645058449,
"grad_norm": 0.474609375,
"learning_rate": 0.00012426424609458518,
"loss": 0.3348,
"step": 2255
},
{
"epoch": 4.8034006376195535,
"grad_norm": 0.431640625,
"learning_rate": 0.00012390383037656327,
"loss": 0.3321,
"step": 2260
},
{
"epoch": 4.814027630180659,
"grad_norm": 0.43359375,
"learning_rate": 0.00012354308502908164,
"loss": 0.3349,
"step": 2265
},
{
"epoch": 4.824654622741764,
"grad_norm": 0.453125,
"learning_rate": 0.00012318201502675285,
"loss": 0.3269,
"step": 2270
},
{
"epoch": 4.835281615302869,
"grad_norm": 0.4453125,
"learning_rate": 0.00012282062534866632,
"loss": 0.3379,
"step": 2275
},
{
"epoch": 4.845908607863975,
"grad_norm": 0.423828125,
"learning_rate": 0.00012245892097831982,
"loss": 0.3316,
"step": 2280
},
{
"epoch": 4.8565356004250795,
"grad_norm": 0.4609375,
"learning_rate": 0.0001220969069035506,
"loss": 0.3411,
"step": 2285
},
{
"epoch": 4.867162592986185,
"grad_norm": 0.5546875,
"learning_rate": 0.0001217345881164667,
"loss": 0.3362,
"step": 2290
},
{
"epoch": 4.877789585547291,
"grad_norm": 0.46484375,
"learning_rate": 0.00012137196961337811,
"loss": 0.3316,
"step": 2295
},
{
"epoch": 4.888416578108395,
"grad_norm": 0.546875,
"learning_rate": 0.00012100905639472779,
"loss": 0.3309,
"step": 2300
},
{
"epoch": 4.899043570669501,
"grad_norm": 0.50390625,
"learning_rate": 0.00012064585346502286,
"loss": 0.3341,
"step": 2305
},
{
"epoch": 4.9096705632306055,
"grad_norm": 0.53515625,
"learning_rate": 0.00012028236583276542,
"loss": 0.3362,
"step": 2310
},
{
"epoch": 4.920297555791711,
"grad_norm": 0.734375,
"learning_rate": 0.0001199185985103836,
"loss": 0.3397,
"step": 2315
},
{
"epoch": 4.930924548352817,
"grad_norm": 0.49609375,
"learning_rate": 0.00011955455651416246,
"loss": 0.3363,
"step": 2320
},
{
"epoch": 4.941551540913921,
"grad_norm": 0.48828125,
"learning_rate": 0.00011919024486417463,
"loss": 0.3335,
"step": 2325
},
{
"epoch": 4.952178533475027,
"grad_norm": 0.5078125,
"learning_rate": 0.00011882566858421135,
"loss": 0.3311,
"step": 2330
},
{
"epoch": 4.9628055260361315,
"grad_norm": 0.54296875,
"learning_rate": 0.000118460832701713,
"loss": 0.3346,
"step": 2335
},
{
"epoch": 4.973432518597237,
"grad_norm": 0.54296875,
"learning_rate": 0.00011809574224769981,
"loss": 0.3334,
"step": 2340
},
{
"epoch": 4.984059511158343,
"grad_norm": 0.462890625,
"learning_rate": 0.00011773040225670256,
"loss": 0.3309,
"step": 2345
},
{
"epoch": 4.994686503719447,
"grad_norm": 0.490234375,
"learning_rate": 0.00011736481776669306,
"loss": 0.3345,
"step": 2350
},
{
"epoch": 4.99893730074389,
"eval_loss": 2.1233744621276855,
"eval_runtime": 0.8188,
"eval_samples_per_second": 6.106,
"eval_steps_per_second": 1.221,
"step": 2352
},
{
"epoch": 5.005313496280553,
"grad_norm": 0.435546875,
"learning_rate": 0.0001169989938190147,
"loss": 0.3151,
"step": 2355
},
{
"epoch": 5.015940488841657,
"grad_norm": 0.55859375,
"learning_rate": 0.00011663293545831302,
"loss": 0.2706,
"step": 2360
},
{
"epoch": 5.026567481402763,
"grad_norm": 0.5390625,
"learning_rate": 0.000116266647732466,
"loss": 0.276,
"step": 2365
},
{
"epoch": 5.0371944739638685,
"grad_norm": 0.55078125,
"learning_rate": 0.00011590013569251457,
"loss": 0.2787,
"step": 2370
},
{
"epoch": 5.047821466524973,
"grad_norm": 0.56640625,
"learning_rate": 0.00011553340439259286,
"loss": 0.2753,
"step": 2375
},
{
"epoch": 5.058448459086079,
"grad_norm": 0.466796875,
"learning_rate": 0.0001151664588898586,
"loss": 0.2707,
"step": 2380
},
{
"epoch": 5.069075451647183,
"grad_norm": 0.52734375,
"learning_rate": 0.00011479930424442335,
"loss": 0.2703,
"step": 2385
},
{
"epoch": 5.079702444208289,
"grad_norm": 0.53125,
"learning_rate": 0.00011443194551928266,
"loss": 0.2715,
"step": 2390
},
{
"epoch": 5.0903294367693945,
"grad_norm": 0.6328125,
"learning_rate": 0.00011406438778024635,
"loss": 0.2739,
"step": 2395
},
{
"epoch": 5.100956429330499,
"grad_norm": 0.5625,
"learning_rate": 0.00011369663609586854,
"loss": 0.278,
"step": 2400
},
{
"epoch": 5.111583421891605,
"grad_norm": 0.52734375,
"learning_rate": 0.0001133286955373779,
"loss": 0.2809,
"step": 2405
},
{
"epoch": 5.12221041445271,
"grad_norm": 0.57421875,
"learning_rate": 0.00011296057117860759,
"loss": 0.2706,
"step": 2410
},
{
"epoch": 5.132837407013815,
"grad_norm": 0.54296875,
"learning_rate": 0.00011259226809592534,
"loss": 0.2765,
"step": 2415
},
{
"epoch": 5.1434643995749205,
"grad_norm": 0.46484375,
"learning_rate": 0.00011222379136816345,
"loss": 0.2717,
"step": 2420
},
{
"epoch": 5.154091392136025,
"grad_norm": 0.515625,
"learning_rate": 0.00011185514607654881,
"loss": 0.2796,
"step": 2425
},
{
"epoch": 5.164718384697131,
"grad_norm": 0.51953125,
"learning_rate": 0.00011148633730463273,
"loss": 0.2801,
"step": 2430
},
{
"epoch": 5.175345377258236,
"grad_norm": 0.5078125,
"learning_rate": 0.00011111737013822088,
"loss": 0.2745,
"step": 2435
},
{
"epoch": 5.185972369819341,
"grad_norm": 0.51171875,
"learning_rate": 0.00011074824966530312,
"loss": 0.2786,
"step": 2440
},
{
"epoch": 5.1965993623804465,
"grad_norm": 0.55078125,
"learning_rate": 0.00011037898097598352,
"loss": 0.2728,
"step": 2445
},
{
"epoch": 5.207226354941551,
"grad_norm": 0.5625,
"learning_rate": 0.00011000956916240985,
"loss": 0.2827,
"step": 2450
},
{
"epoch": 5.217853347502657,
"grad_norm": 0.59765625,
"learning_rate": 0.00010964001931870365,
"loss": 0.2783,
"step": 2455
},
{
"epoch": 5.228480340063762,
"grad_norm": 0.51171875,
"learning_rate": 0.00010927033654088983,
"loss": 0.2847,
"step": 2460
},
{
"epoch": 5.239107332624867,
"grad_norm": 0.78125,
"learning_rate": 0.0001089005259268265,
"loss": 0.2833,
"step": 2465
},
{
"epoch": 5.2497343251859725,
"grad_norm": 0.6953125,
"learning_rate": 0.00010853059257613448,
"loss": 0.2819,
"step": 2470
},
{
"epoch": 5.260361317747078,
"grad_norm": 0.5859375,
"learning_rate": 0.00010816054159012723,
"loss": 0.2766,
"step": 2475
},
{
"epoch": 5.270988310308183,
"grad_norm": 0.53125,
"learning_rate": 0.00010779037807174033,
"loss": 0.2767,
"step": 2480
},
{
"epoch": 5.281615302869288,
"grad_norm": 0.48828125,
"learning_rate": 0.00010742010712546116,
"loss": 0.2737,
"step": 2485
},
{
"epoch": 5.292242295430393,
"grad_norm": 0.47265625,
"learning_rate": 0.00010704973385725851,
"loss": 0.2799,
"step": 2490
},
{
"epoch": 5.3028692879914985,
"grad_norm": 0.58984375,
"learning_rate": 0.00010667926337451217,
"loss": 0.2815,
"step": 2495
},
{
"epoch": 5.313496280552604,
"grad_norm": 0.48046875,
"learning_rate": 0.00010630870078594249,
"loss": 0.2796,
"step": 2500
},
{
"epoch": 5.324123273113709,
"grad_norm": 0.5703125,
"learning_rate": 0.00010593805120154001,
"loss": 0.2764,
"step": 2505
},
{
"epoch": 5.334750265674814,
"grad_norm": 0.48828125,
"learning_rate": 0.00010556731973249485,
"loss": 0.279,
"step": 2510
},
{
"epoch": 5.345377258235919,
"grad_norm": 0.5625,
"learning_rate": 0.00010519651149112631,
"loss": 0.2817,
"step": 2515
},
{
"epoch": 5.3560042507970245,
"grad_norm": 0.578125,
"learning_rate": 0.00010482563159081238,
"loss": 0.283,
"step": 2520
},
{
"epoch": 5.36663124335813,
"grad_norm": 0.5390625,
"learning_rate": 0.00010445468514591925,
"loss": 0.2815,
"step": 2525
},
{
"epoch": 5.377258235919235,
"grad_norm": 0.50390625,
"learning_rate": 0.00010408367727173067,
"loss": 0.2793,
"step": 2530
},
{
"epoch": 5.38788522848034,
"grad_norm": 0.5078125,
"learning_rate": 0.0001037126130843776,
"loss": 0.2839,
"step": 2535
},
{
"epoch": 5.398512221041445,
"grad_norm": 0.46875,
"learning_rate": 0.00010334149770076747,
"loss": 0.2843,
"step": 2540
},
{
"epoch": 5.4091392136025505,
"grad_norm": 0.494140625,
"learning_rate": 0.00010297033623851369,
"loss": 0.2864,
"step": 2545
},
{
"epoch": 5.419766206163656,
"grad_norm": 0.609375,
"learning_rate": 0.0001025991338158651,
"loss": 0.2883,
"step": 2550
},
{
"epoch": 5.430393198724761,
"grad_norm": 0.609375,
"learning_rate": 0.0001022278955516354,
"loss": 0.2919,
"step": 2555
},
{
"epoch": 5.441020191285866,
"grad_norm": 0.55859375,
"learning_rate": 0.00010185662656513251,
"loss": 0.28,
"step": 2560
},
{
"epoch": 5.451647183846971,
"grad_norm": 0.51171875,
"learning_rate": 0.00010148533197608803,
"loss": 0.2815,
"step": 2565
},
{
"epoch": 5.462274176408076,
"grad_norm": 0.5546875,
"learning_rate": 0.00010111401690458654,
"loss": 0.2858,
"step": 2570
},
{
"epoch": 5.472901168969182,
"grad_norm": 0.474609375,
"learning_rate": 0.00010074268647099525,
"loss": 0.2824,
"step": 2575
},
{
"epoch": 5.483528161530287,
"grad_norm": 0.58203125,
"learning_rate": 0.00010037134579589302,
"loss": 0.2863,
"step": 2580
},
{
"epoch": 5.494155154091392,
"grad_norm": 0.58203125,
"learning_rate": 0.0001,
"loss": 0.2755,
"step": 2585
},
{
"epoch": 5.504782146652497,
"grad_norm": 0.59375,
"learning_rate": 9.962865420410701e-05,
"loss": 0.2812,
"step": 2590
},
{
"epoch": 5.515409139213602,
"grad_norm": 0.72265625,
"learning_rate": 9.925731352900478e-05,
"loss": 0.2781,
"step": 2595
},
{
"epoch": 5.526036131774708,
"grad_norm": 0.73828125,
"learning_rate": 9.888598309541347e-05,
"loss": 0.2864,
"step": 2600
},
{
"epoch": 5.536663124335813,
"grad_norm": 0.5234375,
"learning_rate": 9.851466802391201e-05,
"loss": 0.2813,
"step": 2605
},
{
"epoch": 5.547290116896918,
"grad_norm": 0.60546875,
"learning_rate": 9.814337343486754e-05,
"loss": 0.2801,
"step": 2610
},
{
"epoch": 5.557917109458024,
"grad_norm": 0.546875,
"learning_rate": 9.777210444836463e-05,
"loss": 0.281,
"step": 2615
},
{
"epoch": 5.568544102019128,
"grad_norm": 0.51953125,
"learning_rate": 9.740086618413495e-05,
"loss": 0.2817,
"step": 2620
},
{
"epoch": 5.579171094580234,
"grad_norm": 0.578125,
"learning_rate": 9.702966376148635e-05,
"loss": 0.2859,
"step": 2625
},
{
"epoch": 5.589798087141339,
"grad_norm": 0.52734375,
"learning_rate": 9.665850229923258e-05,
"loss": 0.2814,
"step": 2630
},
{
"epoch": 5.600425079702444,
"grad_norm": 0.609375,
"learning_rate": 9.62873869156224e-05,
"loss": 0.2827,
"step": 2635
},
{
"epoch": 5.61105207226355,
"grad_norm": 0.48046875,
"learning_rate": 9.591632272826934e-05,
"loss": 0.2831,
"step": 2640
},
{
"epoch": 5.621679064824654,
"grad_norm": 0.609375,
"learning_rate": 9.554531485408078e-05,
"loss": 0.2791,
"step": 2645
},
{
"epoch": 5.63230605738576,
"grad_norm": 0.51953125,
"learning_rate": 9.517436840918766e-05,
"loss": 0.2829,
"step": 2650
},
{
"epoch": 5.6429330499468655,
"grad_norm": 0.5390625,
"learning_rate": 9.480348850887373e-05,
"loss": 0.2826,
"step": 2655
},
{
"epoch": 5.65356004250797,
"grad_norm": 0.490234375,
"learning_rate": 9.44326802675052e-05,
"loss": 0.2827,
"step": 2660
},
{
"epoch": 5.664187035069076,
"grad_norm": 0.46875,
"learning_rate": 9.406194879846e-05,
"loss": 0.2857,
"step": 2665
},
{
"epoch": 5.67481402763018,
"grad_norm": 0.484375,
"learning_rate": 9.369129921405754e-05,
"loss": 0.2818,
"step": 2670
},
{
"epoch": 5.685441020191286,
"grad_norm": 0.45703125,
"learning_rate": 9.332073662548784e-05,
"loss": 0.278,
"step": 2675
},
{
"epoch": 5.6960680127523915,
"grad_norm": 0.5234375,
"learning_rate": 9.295026614274152e-05,
"loss": 0.2815,
"step": 2680
},
{
"epoch": 5.706695005313496,
"grad_norm": 0.5546875,
"learning_rate": 9.257989287453883e-05,
"loss": 0.2832,
"step": 2685
},
{
"epoch": 5.717321997874602,
"grad_norm": 0.59375,
"learning_rate": 9.220962192825968e-05,
"loss": 0.285,
"step": 2690
},
{
"epoch": 5.727948990435706,
"grad_norm": 0.45703125,
"learning_rate": 9.183945840987276e-05,
"loss": 0.283,
"step": 2695
},
{
"epoch": 5.738575982996812,
"grad_norm": 0.4921875,
"learning_rate": 9.146940742386553e-05,
"loss": 0.2869,
"step": 2700
},
{
"epoch": 5.7492029755579175,
"grad_norm": 0.466796875,
"learning_rate": 9.109947407317352e-05,
"loss": 0.2876,
"step": 2705
},
{
"epoch": 5.759829968119022,
"grad_norm": 0.51953125,
"learning_rate": 9.072966345911019e-05,
"loss": 0.2879,
"step": 2710
},
{
"epoch": 5.770456960680128,
"grad_norm": 0.5,
"learning_rate": 9.035998068129636e-05,
"loss": 0.2773,
"step": 2715
},
{
"epoch": 5.781083953241232,
"grad_norm": 0.5390625,
"learning_rate": 8.999043083759017e-05,
"loss": 0.2863,
"step": 2720
},
{
"epoch": 5.791710945802338,
"grad_norm": 0.47265625,
"learning_rate": 8.962101902401648e-05,
"loss": 0.2813,
"step": 2725
},
{
"epoch": 5.8023379383634435,
"grad_norm": 0.54296875,
"learning_rate": 8.925175033469688e-05,
"loss": 0.2772,
"step": 2730
},
{
"epoch": 5.812964930924548,
"grad_norm": 0.5234375,
"learning_rate": 8.888262986177913e-05,
"loss": 0.2842,
"step": 2735
},
{
"epoch": 5.823591923485654,
"grad_norm": 0.53125,
"learning_rate": 8.851366269536729e-05,
"loss": 0.2806,
"step": 2740
},
{
"epoch": 5.834218916046758,
"grad_norm": 0.51953125,
"learning_rate": 8.814485392345118e-05,
"loss": 0.2754,
"step": 2745
},
{
"epoch": 5.844845908607864,
"grad_norm": 0.482421875,
"learning_rate": 8.777620863183657e-05,
"loss": 0.2852,
"step": 2750
},
{
"epoch": 5.8554729011689695,
"grad_norm": 0.50390625,
"learning_rate": 8.740773190407471e-05,
"loss": 0.2782,
"step": 2755
},
{
"epoch": 5.866099893730074,
"grad_norm": 0.486328125,
"learning_rate": 8.703942882139245e-05,
"loss": 0.2887,
"step": 2760
},
{
"epoch": 5.87672688629118,
"grad_norm": 0.46875,
"learning_rate": 8.667130446262214e-05,
"loss": 0.2812,
"step": 2765
},
{
"epoch": 5.887353878852284,
"grad_norm": 0.53515625,
"learning_rate": 8.630336390413147e-05,
"loss": 0.2809,
"step": 2770
},
{
"epoch": 5.89798087141339,
"grad_norm": 0.498046875,
"learning_rate": 8.59356122197537e-05,
"loss": 0.2816,
"step": 2775
},
{
"epoch": 5.9086078639744954,
"grad_norm": 0.59375,
"learning_rate": 8.556805448071735e-05,
"loss": 0.2788,
"step": 2780
},
{
"epoch": 5.9192348565356,
"grad_norm": 0.515625,
"learning_rate": 8.520069575557667e-05,
"loss": 0.2773,
"step": 2785
},
{
"epoch": 5.929861849096706,
"grad_norm": 0.53125,
"learning_rate": 8.483354111014141e-05,
"loss": 0.2882,
"step": 2790
},
{
"epoch": 5.940488841657811,
"grad_norm": 0.50390625,
"learning_rate": 8.446659560740717e-05,
"loss": 0.285,
"step": 2795
},
{
"epoch": 5.951115834218916,
"grad_norm": 0.515625,
"learning_rate": 8.409986430748545e-05,
"loss": 0.2788,
"step": 2800
},
{
"epoch": 5.961742826780021,
"grad_norm": 0.50390625,
"learning_rate": 8.373335226753404e-05,
"loss": 0.2799,
"step": 2805
},
{
"epoch": 5.972369819341126,
"grad_norm": 0.478515625,
"learning_rate": 8.336706454168701e-05,
"loss": 0.2835,
"step": 2810
},
{
"epoch": 5.982996811902232,
"grad_norm": 0.5234375,
"learning_rate": 8.300100618098534e-05,
"loss": 0.2865,
"step": 2815
},
{
"epoch": 5.993623804463337,
"grad_norm": 0.53515625,
"learning_rate": 8.263518223330697e-05,
"loss": 0.2875,
"step": 2820
},
{
"epoch": 6.0,
"eval_loss": 2.5845518112182617,
"eval_runtime": 0.7933,
"eval_samples_per_second": 6.303,
"eval_steps_per_second": 1.261,
"step": 2823
},
{
"epoch": 6.004250797024442,
"grad_norm": 0.3984375,
"learning_rate": 8.226959774329747e-05,
"loss": 0.2603,
"step": 2825
},
{
"epoch": 6.014877789585547,
"grad_norm": 0.578125,
"learning_rate": 8.190425775230021e-05,
"loss": 0.2316,
"step": 2830
},
{
"epoch": 6.025504782146652,
"grad_norm": 0.470703125,
"learning_rate": 8.153916729828703e-05,
"loss": 0.2244,
"step": 2835
},
{
"epoch": 6.036131774707758,
"grad_norm": 0.470703125,
"learning_rate": 8.117433141578866e-05,
"loss": 0.2276,
"step": 2840
},
{
"epoch": 6.046758767268863,
"grad_norm": 0.484375,
"learning_rate": 8.080975513582539e-05,
"loss": 0.2338,
"step": 2845
},
{
"epoch": 6.057385759829968,
"grad_norm": 0.51171875,
"learning_rate": 8.044544348583755e-05,
"loss": 0.2334,
"step": 2850
},
{
"epoch": 6.068012752391073,
"grad_norm": 0.51953125,
"learning_rate": 8.008140148961641e-05,
"loss": 0.2319,
"step": 2855
},
{
"epoch": 6.078639744952179,
"grad_norm": 0.53125,
"learning_rate": 7.971763416723459e-05,
"loss": 0.2326,
"step": 2860
},
{
"epoch": 6.089266737513284,
"grad_norm": 0.494140625,
"learning_rate": 7.935414653497715e-05,
"loss": 0.2305,
"step": 2865
},
{
"epoch": 6.099893730074389,
"grad_norm": 0.46484375,
"learning_rate": 7.89909436052722e-05,
"loss": 0.2297,
"step": 2870
},
{
"epoch": 6.110520722635494,
"grad_norm": 0.482421875,
"learning_rate": 7.862803038662191e-05,
"loss": 0.2251,
"step": 2875
},
{
"epoch": 6.121147715196599,
"grad_norm": 0.515625,
"learning_rate": 7.826541188353329e-05,
"loss": 0.2321,
"step": 2880
},
{
"epoch": 6.131774707757705,
"grad_norm": 0.52734375,
"learning_rate": 7.790309309644942e-05,
"loss": 0.2324,
"step": 2885
},
{
"epoch": 6.14240170031881,
"grad_norm": 0.5078125,
"learning_rate": 7.754107902168019e-05,
"loss": 0.2305,
"step": 2890
},
{
"epoch": 6.153028692879915,
"grad_norm": 0.52734375,
"learning_rate": 7.717937465133371e-05,
"loss": 0.2258,
"step": 2895
},
{
"epoch": 6.16365568544102,
"grad_norm": 0.5078125,
"learning_rate": 7.681798497324716e-05,
"loss": 0.2375,
"step": 2900
},
{
"epoch": 6.174282678002125,
"grad_norm": 0.50390625,
"learning_rate": 7.645691497091838e-05,
"loss": 0.2351,
"step": 2905
},
{
"epoch": 6.184909670563231,
"grad_norm": 0.52734375,
"learning_rate": 7.609616962343675e-05,
"loss": 0.2359,
"step": 2910
},
{
"epoch": 6.195536663124336,
"grad_norm": 0.5234375,
"learning_rate": 7.573575390541485e-05,
"loss": 0.2396,
"step": 2915
},
{
"epoch": 6.206163655685441,
"grad_norm": 0.5390625,
"learning_rate": 7.537567278691964e-05,
"loss": 0.2316,
"step": 2920
},
{
"epoch": 6.216790648246546,
"grad_norm": 0.486328125,
"learning_rate": 7.501593123340403e-05,
"loss": 0.2356,
"step": 2925
},
{
"epoch": 6.227417640807651,
"grad_norm": 0.52734375,
"learning_rate": 7.465653420563845e-05,
"loss": 0.2373,
"step": 2930
},
{
"epoch": 6.238044633368757,
"grad_norm": 0.53515625,
"learning_rate": 7.42974866596423e-05,
"loss": 0.2349,
"step": 2935
},
{
"epoch": 6.248671625929862,
"grad_norm": 0.5625,
"learning_rate": 7.393879354661577e-05,
"loss": 0.2326,
"step": 2940
},
{
"epoch": 6.259298618490967,
"grad_norm": 0.60546875,
"learning_rate": 7.358045981287141e-05,
"loss": 0.2366,
"step": 2945
},
{
"epoch": 6.269925611052072,
"grad_norm": 0.5390625,
"learning_rate": 7.322249039976608e-05,
"loss": 0.2372,
"step": 2950
},
{
"epoch": 6.280552603613177,
"grad_norm": 0.4765625,
"learning_rate": 7.286489024363261e-05,
"loss": 0.2319,
"step": 2955
},
{
"epoch": 6.291179596174283,
"grad_norm": 0.546875,
"learning_rate": 7.2507664275712e-05,
"loss": 0.2341,
"step": 2960
},
{
"epoch": 6.301806588735388,
"grad_norm": 0.5703125,
"learning_rate": 7.215081742208508e-05,
"loss": 0.2334,
"step": 2965
},
{
"epoch": 6.312433581296493,
"grad_norm": 0.515625,
"learning_rate": 7.179435460360491e-05,
"loss": 0.2305,
"step": 2970
},
{
"epoch": 6.323060573857599,
"grad_norm": 0.61328125,
"learning_rate": 7.14382807358287e-05,
"loss": 0.2351,
"step": 2975
},
{
"epoch": 6.333687566418703,
"grad_norm": 0.51953125,
"learning_rate": 7.108260072895013e-05,
"loss": 0.2312,
"step": 2980
},
{
"epoch": 6.344314558979809,
"grad_norm": 0.52734375,
"learning_rate": 7.072731948773159e-05,
"loss": 0.2309,
"step": 2985
},
{
"epoch": 6.354941551540914,
"grad_norm": 0.54296875,
"learning_rate": 7.037244191143661e-05,
"loss": 0.2392,
"step": 2990
},
{
"epoch": 6.365568544102019,
"grad_norm": 0.56640625,
"learning_rate": 7.001797289376218e-05,
"loss": 0.2363,
"step": 2995
},
{
"epoch": 6.376195536663125,
"grad_norm": 0.55859375,
"learning_rate": 6.966391732277143e-05,
"loss": 0.2342,
"step": 3000
},
{
"epoch": 6.386822529224229,
"grad_norm": 0.55859375,
"learning_rate": 6.931028008082602e-05,
"loss": 0.2329,
"step": 3005
},
{
"epoch": 6.397449521785335,
"grad_norm": 0.55078125,
"learning_rate": 6.895706604451905e-05,
"loss": 0.2377,
"step": 3010
},
{
"epoch": 6.4080765143464395,
"grad_norm": 0.53515625,
"learning_rate": 6.860428008460762e-05,
"loss": 0.2402,
"step": 3015
},
{
"epoch": 6.418703506907545,
"grad_norm": 0.5703125,
"learning_rate": 6.825192706594575e-05,
"loss": 0.2366,
"step": 3020
},
{
"epoch": 6.429330499468651,
"grad_norm": 0.6015625,
"learning_rate": 6.790001184741728e-05,
"loss": 0.2331,
"step": 3025
},
{
"epoch": 6.439957492029755,
"grad_norm": 0.56640625,
"learning_rate": 6.75485392818689e-05,
"loss": 0.2337,
"step": 3030
},
{
"epoch": 6.450584484590861,
"grad_norm": 0.53515625,
"learning_rate": 6.719751421604309e-05,
"loss": 0.2357,
"step": 3035
},
{
"epoch": 6.461211477151966,
"grad_norm": 0.5,
"learning_rate": 6.684694149051156e-05,
"loss": 0.2395,
"step": 3040
},
{
"epoch": 6.471838469713071,
"grad_norm": 0.54296875,
"learning_rate": 6.649682593960814e-05,
"loss": 0.2258,
"step": 3045
},
{
"epoch": 6.482465462274177,
"grad_norm": 0.46484375,
"learning_rate": 6.614717239136246e-05,
"loss": 0.234,
"step": 3050
},
{
"epoch": 6.493092454835281,
"grad_norm": 0.5234375,
"learning_rate": 6.579798566743314e-05,
"loss": 0.2417,
"step": 3055
},
{
"epoch": 6.503719447396387,
"grad_norm": 0.66796875,
"learning_rate": 6.54492705830414e-05,
"loss": 0.2319,
"step": 3060
},
{
"epoch": 6.514346439957492,
"grad_norm": 0.58203125,
"learning_rate": 6.51010319469046e-05,
"loss": 0.2325,
"step": 3065
},
{
"epoch": 6.524973432518597,
"grad_norm": 0.578125,
"learning_rate": 6.475327456117005e-05,
"loss": 0.2353,
"step": 3070
},
{
"epoch": 6.535600425079703,
"grad_norm": 0.490234375,
"learning_rate": 6.440600322134864e-05,
"loss": 0.2345,
"step": 3075
},
{
"epoch": 6.546227417640807,
"grad_norm": 0.52734375,
"learning_rate": 6.405922271624874e-05,
"loss": 0.2327,
"step": 3080
},
{
"epoch": 6.556854410201913,
"grad_norm": 0.54296875,
"learning_rate": 6.371293782791028e-05,
"loss": 0.2356,
"step": 3085
},
{
"epoch": 6.567481402763018,
"grad_norm": 0.49609375,
"learning_rate": 6.336715333153869e-05,
"loss": 0.2336,
"step": 3090
},
{
"epoch": 6.578108395324123,
"grad_norm": 0.51953125,
"learning_rate": 6.302187399543911e-05,
"loss": 0.2365,
"step": 3095
},
{
"epoch": 6.588735387885229,
"grad_norm": 0.46875,
"learning_rate": 6.267710458095053e-05,
"loss": 0.2377,
"step": 3100
},
{
"epoch": 6.599362380446333,
"grad_norm": 0.53515625,
"learning_rate": 6.233284984238035e-05,
"loss": 0.2297,
"step": 3105
},
{
"epoch": 6.609989373007439,
"grad_norm": 0.53515625,
"learning_rate": 6.198911452693853e-05,
"loss": 0.2347,
"step": 3110
},
{
"epoch": 6.620616365568544,
"grad_norm": 0.51953125,
"learning_rate": 6.164590337467243e-05,
"loss": 0.2363,
"step": 3115
},
{
"epoch": 6.631243358129649,
"grad_norm": 0.5078125,
"learning_rate": 6.130322111840114e-05,
"loss": 0.2339,
"step": 3120
},
{
"epoch": 6.641870350690755,
"grad_norm": 0.57421875,
"learning_rate": 6.0961072483650526e-05,
"loss": 0.2328,
"step": 3125
},
{
"epoch": 6.652497343251859,
"grad_norm": 0.578125,
"learning_rate": 6.0619462188587793e-05,
"loss": 0.234,
"step": 3130
},
{
"epoch": 6.663124335812965,
"grad_norm": 0.5390625,
"learning_rate": 6.027839494395664e-05,
"loss": 0.2397,
"step": 3135
},
{
"epoch": 6.67375132837407,
"grad_norm": 0.50390625,
"learning_rate": 5.993787545301204e-05,
"loss": 0.2379,
"step": 3140
},
{
"epoch": 6.684378320935175,
"grad_norm": 0.50390625,
"learning_rate": 5.959790841145577e-05,
"loss": 0.236,
"step": 3145
},
{
"epoch": 6.695005313496281,
"grad_norm": 0.53125,
"learning_rate": 5.9258498507371194e-05,
"loss": 0.2393,
"step": 3150
},
{
"epoch": 6.705632306057386,
"grad_norm": 0.50390625,
"learning_rate": 5.891965042115907e-05,
"loss": 0.2399,
"step": 3155
},
{
"epoch": 6.716259298618491,
"grad_norm": 0.59765625,
"learning_rate": 5.8581368825472585e-05,
"loss": 0.2261,
"step": 3160
},
{
"epoch": 6.726886291179596,
"grad_norm": 0.51953125,
"learning_rate": 5.8243658385153285e-05,
"loss": 0.2379,
"step": 3165
},
{
"epoch": 6.737513283740701,
"grad_norm": 0.55859375,
"learning_rate": 5.790652375716652e-05,
"loss": 0.2327,
"step": 3170
},
{
"epoch": 6.748140276301807,
"grad_norm": 0.58984375,
"learning_rate": 5.7569969590537284e-05,
"loss": 0.2299,
"step": 3175
},
{
"epoch": 6.758767268862912,
"grad_norm": 0.490234375,
"learning_rate": 5.7234000526286156e-05,
"loss": 0.2373,
"step": 3180
},
{
"epoch": 6.769394261424017,
"grad_norm": 0.55859375,
"learning_rate": 5.689862119736522e-05,
"loss": 0.2406,
"step": 3185
},
{
"epoch": 6.780021253985122,
"grad_norm": 0.55078125,
"learning_rate": 5.656383622859418e-05,
"loss": 0.2394,
"step": 3190
},
{
"epoch": 6.790648246546228,
"grad_norm": 0.5625,
"learning_rate": 5.622965023659683e-05,
"loss": 0.2378,
"step": 3195
},
{
"epoch": 6.801275239107333,
"grad_norm": 0.5703125,
"learning_rate": 5.589606782973683e-05,
"loss": 0.2342,
"step": 3200
},
{
"epoch": 6.811902231668438,
"grad_norm": 0.51171875,
"learning_rate": 5.55630936080549e-05,
"loss": 0.2344,
"step": 3205
},
{
"epoch": 6.822529224229543,
"grad_norm": 0.53125,
"learning_rate": 5.5230732163204615e-05,
"loss": 0.2344,
"step": 3210
},
{
"epoch": 6.833156216790648,
"grad_norm": 0.5390625,
"learning_rate": 5.48989880783898e-05,
"loss": 0.2352,
"step": 3215
},
{
"epoch": 6.843783209351754,
"grad_norm": 0.51953125,
"learning_rate": 5.456786592830083e-05,
"loss": 0.2337,
"step": 3220
},
{
"epoch": 6.8544102019128585,
"grad_norm": 0.50390625,
"learning_rate": 5.423737027905173e-05,
"loss": 0.2326,
"step": 3225
},
{
"epoch": 6.865037194473964,
"grad_norm": 0.50390625,
"learning_rate": 5.39075056881172e-05,
"loss": 0.2348,
"step": 3230
},
{
"epoch": 6.875664187035069,
"grad_norm": 0.54296875,
"learning_rate": 5.357827670426977e-05,
"loss": 0.2361,
"step": 3235
},
{
"epoch": 6.886291179596174,
"grad_norm": 0.4921875,
"learning_rate": 5.3249687867517095e-05,
"loss": 0.2298,
"step": 3240
},
{
"epoch": 6.89691817215728,
"grad_norm": 0.53515625,
"learning_rate": 5.292174370903919e-05,
"loss": 0.2344,
"step": 3245
},
{
"epoch": 6.9075451647183845,
"grad_norm": 0.51953125,
"learning_rate": 5.259444875112624e-05,
"loss": 0.2366,
"step": 3250
},
{
"epoch": 6.91817215727949,
"grad_norm": 0.8203125,
"learning_rate": 5.226780750711602e-05,
"loss": 0.2343,
"step": 3255
},
{
"epoch": 6.928799149840595,
"grad_norm": 0.53125,
"learning_rate": 5.1941824481331626e-05,
"loss": 0.236,
"step": 3260
},
{
"epoch": 6.9394261424017,
"grad_norm": 0.498046875,
"learning_rate": 5.1616504169019564e-05,
"loss": 0.2271,
"step": 3265
},
{
"epoch": 6.950053134962806,
"grad_norm": 0.50390625,
"learning_rate": 5.129185105628756e-05,
"loss": 0.2338,
"step": 3270
},
{
"epoch": 6.9606801275239105,
"grad_norm": 0.515625,
"learning_rate": 5.0967869620042794e-05,
"loss": 0.2311,
"step": 3275
},
{
"epoch": 6.971307120085016,
"grad_norm": 0.51171875,
"learning_rate": 5.064456432793019e-05,
"loss": 0.23,
"step": 3280
},
{
"epoch": 6.981934112646121,
"grad_norm": 0.59765625,
"learning_rate": 5.032193963827073e-05,
"loss": 0.2347,
"step": 3285
},
{
"epoch": 6.992561105207226,
"grad_norm": 0.56640625,
"learning_rate": 5.000000000000002e-05,
"loss": 0.2319,
"step": 3290
},
{
"epoch": 6.99893730074389,
"eval_loss": 3.1057090759277344,
"eval_runtime": 0.816,
"eval_samples_per_second": 6.127,
"eval_steps_per_second": 1.225,
"step": 3293
},
{
"epoch": 7.003188097768332,
"grad_norm": 0.40234375,
"learning_rate": 4.96787498526069e-05,
"loss": 0.2177,
"step": 3295
},
{
"epoch": 7.0138150903294365,
"grad_norm": 0.458984375,
"learning_rate": 4.93581936260724e-05,
"loss": 0.1956,
"step": 3300
},
{
"epoch": 7.024442082890542,
"grad_norm": 0.56640625,
"learning_rate": 4.903833574080825e-05,
"loss": 0.1954,
"step": 3305
},
{
"epoch": 7.035069075451648,
"grad_norm": 0.494140625,
"learning_rate": 4.8719180607596484e-05,
"loss": 0.1939,
"step": 3310
},
{
"epoch": 7.045696068012752,
"grad_norm": 0.4453125,
"learning_rate": 4.8400732627528e-05,
"loss": 0.1996,
"step": 3315
},
{
"epoch": 7.056323060573858,
"grad_norm": 0.55859375,
"learning_rate": 4.808299619194251e-05,
"loss": 0.1946,
"step": 3320
},
{
"epoch": 7.0669500531349625,
"grad_norm": 0.54296875,
"learning_rate": 4.776597568236731e-05,
"loss": 0.1914,
"step": 3325
},
{
"epoch": 7.077577045696068,
"grad_norm": 0.4609375,
"learning_rate": 4.744967547045754e-05,
"loss": 0.1963,
"step": 3330
},
{
"epoch": 7.088204038257174,
"grad_norm": 0.4921875,
"learning_rate": 4.713409991793536e-05,
"loss": 0.2024,
"step": 3335
},
{
"epoch": 7.098831030818278,
"grad_norm": 0.5625,
"learning_rate": 4.681925337653006e-05,
"loss": 0.1974,
"step": 3340
},
{
"epoch": 7.109458023379384,
"grad_norm": 0.53515625,
"learning_rate": 4.650514018791799e-05,
"loss": 0.1954,
"step": 3345
},
{
"epoch": 7.1200850159404885,
"grad_norm": 0.53125,
"learning_rate": 4.6191764683662744e-05,
"loss": 0.2007,
"step": 3350
},
{
"epoch": 7.130712008501594,
"grad_norm": 0.478515625,
"learning_rate": 4.587913118515532e-05,
"loss": 0.1959,
"step": 3355
},
{
"epoch": 7.1413390010627,
"grad_norm": 0.5078125,
"learning_rate": 4.5567244003554645e-05,
"loss": 0.1994,
"step": 3360
},
{
"epoch": 7.151965993623804,
"grad_norm": 0.51953125,
"learning_rate": 4.5256107439728035e-05,
"loss": 0.1979,
"step": 3365
},
{
"epoch": 7.16259298618491,
"grad_norm": 0.53125,
"learning_rate": 4.494572578419194e-05,
"loss": 0.1951,
"step": 3370
},
{
"epoch": 7.1732199787460145,
"grad_norm": 0.474609375,
"learning_rate": 4.463610331705273e-05,
"loss": 0.1996,
"step": 3375
},
{
"epoch": 7.18384697130712,
"grad_norm": 0.474609375,
"learning_rate": 4.432724430794786e-05,
"loss": 0.1989,
"step": 3380
},
{
"epoch": 7.194473963868226,
"grad_norm": 0.51953125,
"learning_rate": 4.4019153015986546e-05,
"loss": 0.2009,
"step": 3385
},
{
"epoch": 7.20510095642933,
"grad_norm": 0.53125,
"learning_rate": 4.371183368969165e-05,
"loss": 0.1926,
"step": 3390
},
{
"epoch": 7.215727948990436,
"grad_norm": 0.51171875,
"learning_rate": 4.340529056694047e-05,
"loss": 0.1997,
"step": 3395
},
{
"epoch": 7.226354941551541,
"grad_norm": 0.4765625,
"learning_rate": 4.309952787490689e-05,
"loss": 0.1969,
"step": 3400
},
{
"epoch": 7.236981934112646,
"grad_norm": 0.48046875,
"learning_rate": 4.279454983000251e-05,
"loss": 0.1974,
"step": 3405
},
{
"epoch": 7.247608926673752,
"grad_norm": 0.63671875,
"learning_rate": 4.249036063781896e-05,
"loss": 0.1969,
"step": 3410
},
{
"epoch": 7.258235919234856,
"grad_norm": 0.5234375,
"learning_rate": 4.2186964493069794e-05,
"loss": 0.1953,
"step": 3415
},
{
"epoch": 7.268862911795962,
"grad_norm": 0.54296875,
"learning_rate": 4.1884365579532346e-05,
"loss": 0.1973,
"step": 3420
},
{
"epoch": 7.279489904357067,
"grad_norm": 0.62890625,
"learning_rate": 4.158256806999059e-05,
"loss": 0.1938,
"step": 3425
},
{
"epoch": 7.290116896918172,
"grad_norm": 0.51953125,
"learning_rate": 4.128157612617696e-05,
"loss": 0.1964,
"step": 3430
},
{
"epoch": 7.3007438894792775,
"grad_norm": 0.53125,
"learning_rate": 4.0981393898715545e-05,
"loss": 0.202,
"step": 3435
},
{
"epoch": 7.311370882040382,
"grad_norm": 0.494140625,
"learning_rate": 4.0682025527064486e-05,
"loss": 0.1964,
"step": 3440
},
{
"epoch": 7.321997874601488,
"grad_norm": 0.53515625,
"learning_rate": 4.038347513945898e-05,
"loss": 0.1999,
"step": 3445
},
{
"epoch": 7.332624867162593,
"grad_norm": 0.4921875,
"learning_rate": 4.008574685285442e-05,
"loss": 0.1938,
"step": 3450
},
{
"epoch": 7.343251859723698,
"grad_norm": 0.94921875,
"learning_rate": 3.978884477286956e-05,
"loss": 0.1941,
"step": 3455
},
{
"epoch": 7.3538788522848035,
"grad_norm": 0.5859375,
"learning_rate": 3.94927729937299e-05,
"loss": 0.1996,
"step": 3460
},
{
"epoch": 7.364505844845908,
"grad_norm": 0.578125,
"learning_rate": 3.9197535598211274e-05,
"loss": 0.2,
"step": 3465
},
{
"epoch": 7.375132837407014,
"grad_norm": 0.50390625,
"learning_rate": 3.890313665758348e-05,
"loss": 0.1937,
"step": 3470
},
{
"epoch": 7.385759829968119,
"grad_norm": 0.5390625,
"learning_rate": 3.860958023155422e-05,
"loss": 0.1991,
"step": 3475
},
{
"epoch": 7.396386822529224,
"grad_norm": 0.5078125,
"learning_rate": 3.8316870368213e-05,
"loss": 0.1979,
"step": 3480
},
{
"epoch": 7.4070138150903295,
"grad_norm": 0.546875,
"learning_rate": 3.802501110397553e-05,
"loss": 0.2021,
"step": 3485
},
{
"epoch": 7.417640807651434,
"grad_norm": 0.54296875,
"learning_rate": 3.773400646352769e-05,
"loss": 0.2034,
"step": 3490
},
{
"epoch": 7.42826780021254,
"grad_norm": 0.50390625,
"learning_rate": 3.744386045977052e-05,
"loss": 0.2033,
"step": 3495
},
{
"epoch": 7.438894792773645,
"grad_norm": 0.5,
"learning_rate": 3.7154577093764334e-05,
"loss": 0.2012,
"step": 3500
},
{
"epoch": 7.44952178533475,
"grad_norm": 0.53515625,
"learning_rate": 3.686616035467408e-05,
"loss": 0.1952,
"step": 3505
},
{
"epoch": 7.4601487778958555,
"grad_norm": 0.57421875,
"learning_rate": 3.657861421971388e-05,
"loss": 0.2011,
"step": 3510
},
{
"epoch": 7.470775770456961,
"grad_norm": 0.490234375,
"learning_rate": 3.629194265409255e-05,
"loss": 0.2033,
"step": 3515
},
{
"epoch": 7.481402763018066,
"grad_norm": 0.546875,
"learning_rate": 3.6006149610958625e-05,
"loss": 0.2003,
"step": 3520
},
{
"epoch": 7.492029755579171,
"grad_norm": 0.56640625,
"learning_rate": 3.5721239031346066e-05,
"loss": 0.1988,
"step": 3525
},
{
"epoch": 7.502656748140276,
"grad_norm": 0.484375,
"learning_rate": 3.543721484411976e-05,
"loss": 0.1903,
"step": 3530
},
{
"epoch": 7.5132837407013815,
"grad_norm": 0.53125,
"learning_rate": 3.515408096592144e-05,
"loss": 0.1991,
"step": 3535
},
{
"epoch": 7.523910733262487,
"grad_norm": 0.49609375,
"learning_rate": 3.487184130111562e-05,
"loss": 0.1967,
"step": 3540
},
{
"epoch": 7.534537725823592,
"grad_norm": 0.54296875,
"learning_rate": 3.459049974173579e-05,
"loss": 0.1983,
"step": 3545
},
{
"epoch": 7.545164718384697,
"grad_norm": 0.53125,
"learning_rate": 3.4310060167430725e-05,
"loss": 0.1998,
"step": 3550
},
{
"epoch": 7.555791710945803,
"grad_norm": 0.5390625,
"learning_rate": 3.4030526445410984e-05,
"loss": 0.198,
"step": 3555
},
{
"epoch": 7.5664187035069075,
"grad_norm": 0.49609375,
"learning_rate": 3.375190243039556e-05,
"loss": 0.1961,
"step": 3560
},
{
"epoch": 7.577045696068013,
"grad_norm": 0.478515625,
"learning_rate": 3.3474191964558885e-05,
"loss": 0.1974,
"step": 3565
},
{
"epoch": 7.587672688629118,
"grad_norm": 0.5234375,
"learning_rate": 3.319739887747752e-05,
"loss": 0.195,
"step": 3570
},
{
"epoch": 7.598299681190223,
"grad_norm": 0.51171875,
"learning_rate": 3.292152698607768e-05,
"loss": 0.1992,
"step": 3575
},
{
"epoch": 7.608926673751329,
"grad_norm": 0.50390625,
"learning_rate": 3.264658009458239e-05,
"loss": 0.1996,
"step": 3580
},
{
"epoch": 7.6195536663124335,
"grad_norm": 0.5,
"learning_rate": 3.2372561994459136e-05,
"loss": 0.1966,
"step": 3585
},
{
"epoch": 7.630180658873539,
"grad_norm": 0.5078125,
"learning_rate": 3.209947646436752e-05,
"loss": 0.1941,
"step": 3590
},
{
"epoch": 7.640807651434644,
"grad_norm": 0.54296875,
"learning_rate": 3.182732727010715e-05,
"loss": 0.1959,
"step": 3595
},
{
"epoch": 7.651434643995749,
"grad_norm": 0.53515625,
"learning_rate": 3.155611816456586e-05,
"loss": 0.2002,
"step": 3600
},
{
"epoch": 7.662061636556855,
"grad_norm": 0.51171875,
"learning_rate": 3.12858528876676e-05,
"loss": 0.196,
"step": 3605
},
{
"epoch": 7.6726886291179595,
"grad_norm": 0.53125,
"learning_rate": 3.1016535166321356e-05,
"loss": 0.1988,
"step": 3610
},
{
"epoch": 7.683315621679065,
"grad_norm": 0.51171875,
"learning_rate": 3.074816871436924e-05,
"loss": 0.2016,
"step": 3615
},
{
"epoch": 7.69394261424017,
"grad_norm": 0.54296875,
"learning_rate": 3.0480757232535772e-05,
"loss": 0.1973,
"step": 3620
},
{
"epoch": 7.704569606801275,
"grad_norm": 0.5234375,
"learning_rate": 3.021430440837646e-05,
"loss": 0.1945,
"step": 3625
},
{
"epoch": 7.715196599362381,
"grad_norm": 0.5234375,
"learning_rate": 2.9948813916227115e-05,
"loss": 0.1985,
"step": 3630
},
{
"epoch": 7.7258235919234854,
"grad_norm": 0.5859375,
"learning_rate": 2.968428941715321e-05,
"loss": 0.1963,
"step": 3635
},
{
"epoch": 7.736450584484591,
"grad_norm": 0.50390625,
"learning_rate": 2.9420734558899322e-05,
"loss": 0.2008,
"step": 3640
},
{
"epoch": 7.747077577045696,
"grad_norm": 0.5234375,
"learning_rate": 2.915815297583886e-05,
"loss": 0.1983,
"step": 3645
},
{
"epoch": 7.757704569606801,
"grad_norm": 0.51953125,
"learning_rate": 2.889654828892393e-05,
"loss": 0.2,
"step": 3650
},
{
"epoch": 7.768331562167907,
"grad_norm": 0.515625,
"learning_rate": 2.8635924105635458e-05,
"loss": 0.1978,
"step": 3655
},
{
"epoch": 7.778958554729011,
"grad_norm": 0.546875,
"learning_rate": 2.8376284019933373e-05,
"loss": 0.1944,
"step": 3660
},
{
"epoch": 7.789585547290117,
"grad_norm": 0.474609375,
"learning_rate": 2.8117631612207084e-05,
"loss": 0.1963,
"step": 3665
},
{
"epoch": 7.800212539851222,
"grad_norm": 0.53515625,
"learning_rate": 2.7859970449226104e-05,
"loss": 0.1961,
"step": 3670
},
{
"epoch": 7.810839532412327,
"grad_norm": 0.54296875,
"learning_rate": 2.760330408409083e-05,
"loss": 0.1974,
"step": 3675
},
{
"epoch": 7.821466524973433,
"grad_norm": 0.51171875,
"learning_rate": 2.73476360561837e-05,
"loss": 0.1977,
"step": 3680
},
{
"epoch": 7.832093517534537,
"grad_norm": 0.55078125,
"learning_rate": 2.7092969891120058e-05,
"loss": 0.1956,
"step": 3685
},
{
"epoch": 7.842720510095643,
"grad_norm": 0.51171875,
"learning_rate": 2.6839309100699973e-05,
"loss": 0.1905,
"step": 3690
},
{
"epoch": 7.8533475026567485,
"grad_norm": 0.5234375,
"learning_rate": 2.6586657182859343e-05,
"loss": 0.1984,
"step": 3695
},
{
"epoch": 7.863974495217853,
"grad_norm": 0.5625,
"learning_rate": 2.6335017621622116e-05,
"loss": 0.1949,
"step": 3700
},
{
"epoch": 7.874601487778959,
"grad_norm": 0.5703125,
"learning_rate": 2.6084393887051884e-05,
"loss": 0.196,
"step": 3705
},
{
"epoch": 7.885228480340063,
"grad_norm": 0.5234375,
"learning_rate": 2.5834789435204243e-05,
"loss": 0.1922,
"step": 3710
},
{
"epoch": 7.895855472901169,
"grad_norm": 0.478515625,
"learning_rate": 2.5586207708079034e-05,
"loss": 0.1921,
"step": 3715
},
{
"epoch": 7.9064824654622745,
"grad_norm": 0.55078125,
"learning_rate": 2.5338652133572915e-05,
"loss": 0.1969,
"step": 3720
},
{
"epoch": 7.917109458023379,
"grad_norm": 0.54296875,
"learning_rate": 2.5092126125432136e-05,
"loss": 0.2008,
"step": 3725
},
{
"epoch": 7.927736450584485,
"grad_norm": 0.54296875,
"learning_rate": 2.4846633083205263e-05,
"loss": 0.1963,
"step": 3730
},
{
"epoch": 7.93836344314559,
"grad_norm": 0.49609375,
"learning_rate": 2.4602176392196662e-05,
"loss": 0.1944,
"step": 3735
},
{
"epoch": 7.948990435706695,
"grad_norm": 0.51953125,
"learning_rate": 2.4358759423419474e-05,
"loss": 0.1954,
"step": 3740
},
{
"epoch": 7.9596174282678005,
"grad_norm": 0.5,
"learning_rate": 2.411638553354928e-05,
"loss": 0.1979,
"step": 3745
},
{
"epoch": 7.970244420828905,
"grad_norm": 0.50390625,
"learning_rate": 2.3875058064877807e-05,
"loss": 0.195,
"step": 3750
},
{
"epoch": 7.980871413390011,
"grad_norm": 0.515625,
"learning_rate": 2.3634780345266806e-05,
"loss": 0.1936,
"step": 3755
},
{
"epoch": 7.991498405951116,
"grad_norm": 0.5,
"learning_rate": 2.339555568810221e-05,
"loss": 0.1968,
"step": 3760
},
{
"epoch": 8.0,
"eval_loss": 3.6609108448028564,
"eval_runtime": 0.7997,
"eval_samples_per_second": 6.252,
"eval_steps_per_second": 1.25,
"step": 3764
},
{
"epoch": 8.002125398512222,
"grad_norm": 0.431640625,
"learning_rate": 2.3157387392248385e-05,
"loss": 0.1921,
"step": 3765
},
{
"epoch": 8.012752391073326,
"grad_norm": 0.41015625,
"learning_rate": 2.2920278742002676e-05,
"loss": 0.1828,
"step": 3770
},
{
"epoch": 8.023379383634431,
"grad_norm": 0.453125,
"learning_rate": 2.2684233007050115e-05,
"loss": 0.1811,
"step": 3775
},
{
"epoch": 8.034006376195537,
"grad_norm": 0.515625,
"learning_rate": 2.244925344241828e-05,
"loss": 0.1809,
"step": 3780
},
{
"epoch": 8.044633368756642,
"grad_norm": 0.4765625,
"learning_rate": 2.221534328843261e-05,
"loss": 0.1758,
"step": 3785
},
{
"epoch": 8.055260361317748,
"grad_norm": 0.51171875,
"learning_rate": 2.1982505770671303e-05,
"loss": 0.1764,
"step": 3790
},
{
"epoch": 8.065887353878852,
"grad_norm": 0.51953125,
"learning_rate": 2.1750744099921396e-05,
"loss": 0.1779,
"step": 3795
},
{
"epoch": 8.076514346439957,
"grad_norm": 0.478515625,
"learning_rate": 2.1520061472133902e-05,
"loss": 0.1798,
"step": 3800
},
{
"epoch": 8.087141339001063,
"grad_norm": 0.51953125,
"learning_rate": 2.1290461068380298e-05,
"loss": 0.1756,
"step": 3805
},
{
"epoch": 8.097768331562168,
"grad_norm": 0.4921875,
"learning_rate": 2.1061946054808146e-05,
"loss": 0.1804,
"step": 3810
},
{
"epoch": 8.108395324123274,
"grad_norm": 0.51953125,
"learning_rate": 2.0834519582597878e-05,
"loss": 0.1782,
"step": 3815
},
{
"epoch": 8.119022316684378,
"grad_norm": 0.490234375,
"learning_rate": 2.0608184787919026e-05,
"loss": 0.1804,
"step": 3820
},
{
"epoch": 8.129649309245483,
"grad_norm": 0.51171875,
"learning_rate": 2.0382944791887106e-05,
"loss": 0.1795,
"step": 3825
},
{
"epoch": 8.140276301806589,
"grad_norm": 0.490234375,
"learning_rate": 2.0158802700520574e-05,
"loss": 0.1815,
"step": 3830
},
{
"epoch": 8.150903294367694,
"grad_norm": 0.4921875,
"learning_rate": 1.9935761604697976e-05,
"loss": 0.1793,
"step": 3835
},
{
"epoch": 8.1615302869288,
"grad_norm": 0.4765625,
"learning_rate": 1.9713824580115335e-05,
"loss": 0.1785,
"step": 3840
},
{
"epoch": 8.172157279489904,
"grad_norm": 0.486328125,
"learning_rate": 1.9492994687243714e-05,
"loss": 0.1784,
"step": 3845
},
{
"epoch": 8.182784272051009,
"grad_norm": 0.48046875,
"learning_rate": 1.927327497128706e-05,
"loss": 0.177,
"step": 3850
},
{
"epoch": 8.193411264612115,
"grad_norm": 0.4921875,
"learning_rate": 1.9054668462140168e-05,
"loss": 0.1804,
"step": 3855
},
{
"epoch": 8.20403825717322,
"grad_norm": 0.515625,
"learning_rate": 1.883717817434688e-05,
"loss": 0.1753,
"step": 3860
},
{
"epoch": 8.214665249734326,
"grad_norm": 0.484375,
"learning_rate": 1.8620807107058668e-05,
"loss": 0.1785,
"step": 3865
},
{
"epoch": 8.22529224229543,
"grad_norm": 0.46875,
"learning_rate": 1.840555824399296e-05,
"loss": 0.1775,
"step": 3870
},
{
"epoch": 8.235919234856535,
"grad_norm": 0.578125,
"learning_rate": 1.8191434553392428e-05,
"loss": 0.18,
"step": 3875
},
{
"epoch": 8.24654622741764,
"grad_norm": 0.515625,
"learning_rate": 1.797843898798358e-05,
"loss": 0.1806,
"step": 3880
},
{
"epoch": 8.257173219978746,
"grad_norm": 0.51171875,
"learning_rate": 1.7766574484936482e-05,
"loss": 0.1837,
"step": 3885
},
{
"epoch": 8.267800212539852,
"grad_norm": 0.494140625,
"learning_rate": 1.7555843965823992e-05,
"loss": 0.1817,
"step": 3890
},
{
"epoch": 8.278427205100957,
"grad_norm": 0.5390625,
"learning_rate": 1.7346250336581427e-05,
"loss": 0.1806,
"step": 3895
},
{
"epoch": 8.289054197662061,
"grad_norm": 0.486328125,
"learning_rate": 1.7137796487466797e-05,
"loss": 0.1791,
"step": 3900
},
{
"epoch": 8.299681190223167,
"grad_norm": 0.498046875,
"learning_rate": 1.693048529302056e-05,
"loss": 0.1788,
"step": 3905
},
{
"epoch": 8.310308182784272,
"grad_norm": 0.46484375,
"learning_rate": 1.672431961202635e-05,
"loss": 0.1809,
"step": 3910
},
{
"epoch": 8.320935175345378,
"grad_norm": 0.484375,
"learning_rate": 1.6519302287471207e-05,
"loss": 0.1774,
"step": 3915
},
{
"epoch": 8.331562167906483,
"grad_norm": 0.5,
"learning_rate": 1.6315436146506703e-05,
"loss": 0.1737,
"step": 3920
},
{
"epoch": 8.342189160467587,
"grad_norm": 0.50390625,
"learning_rate": 1.6112724000409694e-05,
"loss": 0.1813,
"step": 3925
},
{
"epoch": 8.352816153028693,
"grad_norm": 0.4921875,
"learning_rate": 1.5911168644543707e-05,
"loss": 0.175,
"step": 3930
},
{
"epoch": 8.363443145589798,
"grad_norm": 0.5546875,
"learning_rate": 1.57107728583203e-05,
"loss": 0.1838,
"step": 3935
},
{
"epoch": 8.374070138150904,
"grad_norm": 0.51953125,
"learning_rate": 1.5511539405160825e-05,
"loss": 0.1783,
"step": 3940
},
{
"epoch": 8.38469713071201,
"grad_norm": 0.5234375,
"learning_rate": 1.5313471032458247e-05,
"loss": 0.1763,
"step": 3945
},
{
"epoch": 8.395324123273113,
"grad_norm": 0.51953125,
"learning_rate": 1.5116570471539293e-05,
"loss": 0.1758,
"step": 3950
},
{
"epoch": 8.405951115834219,
"grad_norm": 0.50390625,
"learning_rate": 1.4920840437626759e-05,
"loss": 0.1829,
"step": 3955
},
{
"epoch": 8.416578108395324,
"grad_norm": 0.53125,
"learning_rate": 1.4726283629802107e-05,
"loss": 0.1779,
"step": 3960
},
{
"epoch": 8.42720510095643,
"grad_norm": 0.53125,
"learning_rate": 1.4532902730968212e-05,
"loss": 0.1792,
"step": 3965
},
{
"epoch": 8.437832093517535,
"grad_norm": 0.48828125,
"learning_rate": 1.4340700407812435e-05,
"loss": 0.1813,
"step": 3970
},
{
"epoch": 8.448459086078639,
"grad_norm": 0.50390625,
"learning_rate": 1.414967931076967e-05,
"loss": 0.1785,
"step": 3975
},
{
"epoch": 8.459086078639745,
"grad_norm": 0.5625,
"learning_rate": 1.3959842073986085e-05,
"loss": 0.1746,
"step": 3980
},
{
"epoch": 8.46971307120085,
"grad_norm": 0.62109375,
"learning_rate": 1.3771191315282427e-05,
"loss": 0.1832,
"step": 3985
},
{
"epoch": 8.480340063761956,
"grad_norm": 0.490234375,
"learning_rate": 1.3583729636118358e-05,
"loss": 0.1843,
"step": 3990
},
{
"epoch": 8.490967056323061,
"grad_norm": 0.515625,
"learning_rate": 1.339745962155613e-05,
"loss": 0.1801,
"step": 3995
},
{
"epoch": 8.501594048884165,
"grad_norm": 0.54296875,
"learning_rate": 1.3212383840225329e-05,
"loss": 0.18,
"step": 4000
},
{
"epoch": 8.51222104144527,
"grad_norm": 0.55078125,
"learning_rate": 1.3028504844287182e-05,
"loss": 0.18,
"step": 4005
},
{
"epoch": 8.522848034006376,
"grad_norm": 0.51171875,
"learning_rate": 1.2845825169399507e-05,
"loss": 0.1778,
"step": 4010
},
{
"epoch": 8.533475026567482,
"grad_norm": 0.46875,
"learning_rate": 1.2664347334681681e-05,
"loss": 0.1849,
"step": 4015
},
{
"epoch": 8.544102019128587,
"grad_norm": 0.46875,
"learning_rate": 1.2484073842679944e-05,
"loss": 0.1799,
"step": 4020
},
{
"epoch": 8.554729011689691,
"grad_norm": 0.478515625,
"learning_rate": 1.230500717933285e-05,
"loss": 0.1802,
"step": 4025
},
{
"epoch": 8.565356004250797,
"grad_norm": 0.51953125,
"learning_rate": 1.2127149813937022e-05,
"loss": 0.1776,
"step": 4030
},
{
"epoch": 8.575982996811902,
"grad_norm": 0.52734375,
"learning_rate": 1.1950504199113088e-05,
"loss": 0.1794,
"step": 4035
},
{
"epoch": 8.586609989373008,
"grad_norm": 0.478515625,
"learning_rate": 1.1775072770771834e-05,
"loss": 0.1757,
"step": 4040
},
{
"epoch": 8.597236981934113,
"grad_norm": 0.5078125,
"learning_rate": 1.1600857948080624e-05,
"loss": 0.1786,
"step": 4045
},
{
"epoch": 8.607863974495217,
"grad_norm": 0.50390625,
"learning_rate": 1.1427862133430156e-05,
"loss": 0.1838,
"step": 4050
},
{
"epoch": 8.618490967056323,
"grad_norm": 0.5234375,
"learning_rate": 1.1256087712401087e-05,
"loss": 0.1755,
"step": 4055
},
{
"epoch": 8.629117959617428,
"grad_norm": 0.478515625,
"learning_rate": 1.1085537053731354e-05,
"loss": 0.1777,
"step": 4060
},
{
"epoch": 8.639744952178534,
"grad_norm": 0.5234375,
"learning_rate": 1.091621250928343e-05,
"loss": 0.1799,
"step": 4065
},
{
"epoch": 8.65037194473964,
"grad_norm": 0.51953125,
"learning_rate": 1.0748116414011888e-05,
"loss": 0.1774,
"step": 4070
},
{
"epoch": 8.660998937300743,
"grad_norm": 0.5,
"learning_rate": 1.058125108593122e-05,
"loss": 0.1844,
"step": 4075
},
{
"epoch": 8.671625929861849,
"grad_norm": 0.50390625,
"learning_rate": 1.0415618826083828e-05,
"loss": 0.1803,
"step": 4080
},
{
"epoch": 8.682252922422954,
"grad_norm": 0.546875,
"learning_rate": 1.0251221918508425e-05,
"loss": 0.1787,
"step": 4085
},
{
"epoch": 8.69287991498406,
"grad_norm": 0.5546875,
"learning_rate": 1.0088062630208273e-05,
"loss": 0.1804,
"step": 4090
},
{
"epoch": 8.703506907545165,
"grad_norm": 0.50390625,
"learning_rate": 9.926143211120286e-06,
"loss": 0.1868,
"step": 4095
},
{
"epoch": 8.71413390010627,
"grad_norm": 0.5,
"learning_rate": 9.765465894083636e-06,
"loss": 0.1831,
"step": 4100
},
{
"epoch": 8.724760892667375,
"grad_norm": 0.458984375,
"learning_rate": 9.606032894809258e-06,
"loss": 0.1774,
"step": 4105
},
{
"epoch": 8.73538788522848,
"grad_norm": 0.515625,
"learning_rate": 9.447846411849115e-06,
"loss": 0.1775,
"step": 4110
},
{
"epoch": 8.746014877789586,
"grad_norm": 0.51953125,
"learning_rate": 9.29090862656593e-06,
"loss": 0.1796,
"step": 4115
},
{
"epoch": 8.756641870350691,
"grad_norm": 0.5,
"learning_rate": 9.135221703103136e-06,
"loss": 0.1798,
"step": 4120
},
{
"epoch": 8.767268862911797,
"grad_norm": 0.5,
"learning_rate": 8.980787788355016e-06,
"loss": 0.179,
"step": 4125
},
{
"epoch": 8.7778958554729,
"grad_norm": 0.46484375,
"learning_rate": 8.827609011937066e-06,
"loss": 0.1802,
"step": 4130
},
{
"epoch": 8.788522848034006,
"grad_norm": 0.51171875,
"learning_rate": 8.675687486156659e-06,
"loss": 0.184,
"step": 4135
},
{
"epoch": 8.799149840595112,
"grad_norm": 0.60546875,
"learning_rate": 8.525025305983936e-06,
"loss": 0.173,
"step": 4140
},
{
"epoch": 8.809776833156217,
"grad_norm": 0.53125,
"learning_rate": 8.375624549022854e-06,
"loss": 0.179,
"step": 4145
},
{
"epoch": 8.820403825717323,
"grad_norm": 0.56640625,
"learning_rate": 8.227487275482592e-06,
"loss": 0.1812,
"step": 4150
},
{
"epoch": 8.831030818278427,
"grad_norm": 0.494140625,
"learning_rate": 8.08061552814917e-06,
"loss": 0.183,
"step": 4155
},
{
"epoch": 8.841657810839532,
"grad_norm": 0.51953125,
"learning_rate": 7.935011332357112e-06,
"loss": 0.1786,
"step": 4160
},
{
"epoch": 8.852284803400638,
"grad_norm": 0.474609375,
"learning_rate": 7.790676695961796e-06,
"loss": 0.178,
"step": 4165
},
{
"epoch": 8.862911795961743,
"grad_norm": 0.5,
"learning_rate": 7.647613609311455e-06,
"loss": 0.1809,
"step": 4170
},
{
"epoch": 8.873538788522849,
"grad_norm": 0.55078125,
"learning_rate": 7.505824045220011e-06,
"loss": 0.1756,
"step": 4175
},
{
"epoch": 8.884165781083952,
"grad_norm": 0.5390625,
"learning_rate": 7.365309958939615e-06,
"loss": 0.175,
"step": 4180
},
{
"epoch": 8.894792773645058,
"grad_norm": 0.5078125,
"learning_rate": 7.226073288133939e-06,
"loss": 0.1792,
"step": 4185
},
{
"epoch": 8.905419766206164,
"grad_norm": 0.486328125,
"learning_rate": 7.088115952851238e-06,
"loss": 0.1733,
"step": 4190
},
{
"epoch": 8.91604675876727,
"grad_norm": 0.49609375,
"learning_rate": 6.951439855498021e-06,
"loss": 0.1792,
"step": 4195
},
{
"epoch": 8.926673751328375,
"grad_norm": 0.498046875,
"learning_rate": 6.81604688081271e-06,
"loss": 0.1803,
"step": 4200
},
{
"epoch": 8.937300743889478,
"grad_norm": 0.45703125,
"learning_rate": 6.681938895839746e-06,
"loss": 0.177,
"step": 4205
},
{
"epoch": 8.947927736450584,
"grad_norm": 0.486328125,
"learning_rate": 6.549117749903755e-06,
"loss": 0.1791,
"step": 4210
},
{
"epoch": 8.95855472901169,
"grad_norm": 0.498046875,
"learning_rate": 6.417585274584115e-06,
"loss": 0.1778,
"step": 4215
},
{
"epoch": 8.969181721572795,
"grad_norm": 0.494140625,
"learning_rate": 6.287343283689661e-06,
"loss": 0.176,
"step": 4220
},
{
"epoch": 8.9798087141339,
"grad_norm": 0.5234375,
"learning_rate": 6.158393573233678e-06,
"loss": 0.1759,
"step": 4225
},
{
"epoch": 8.990435706695006,
"grad_norm": 0.5234375,
"learning_rate": 6.030737921409169e-06,
"loss": 0.1809,
"step": 4230
},
{
"epoch": 8.99893730074389,
"eval_loss": 3.939990520477295,
"eval_runtime": 0.8191,
"eval_samples_per_second": 6.105,
"eval_steps_per_second": 1.221,
"step": 4234
},
{
"epoch": 9.00106269925611,
"grad_norm": 0.47265625,
"learning_rate": 5.904378088564255e-06,
"loss": 0.1819,
"step": 4235
},
{
"epoch": 9.011689691817216,
"grad_norm": 0.5,
"learning_rate": 5.779315817178e-06,
"loss": 0.1763,
"step": 4240
},
{
"epoch": 9.022316684378321,
"grad_norm": 0.4921875,
"learning_rate": 5.655552831836297e-06,
"loss": 0.1767,
"step": 4245
},
{
"epoch": 9.032943676939427,
"grad_norm": 0.46875,
"learning_rate": 5.533090839208133e-06,
"loss": 0.1744,
"step": 4250
},
{
"epoch": 9.043570669500532,
"grad_norm": 0.490234375,
"learning_rate": 5.4119315280220405e-06,
"loss": 0.1777,
"step": 4255
},
{
"epoch": 9.054197662061636,
"grad_norm": 0.44921875,
"learning_rate": 5.292076569042826e-06,
"loss": 0.1744,
"step": 4260
},
{
"epoch": 9.064824654622742,
"grad_norm": 0.451171875,
"learning_rate": 5.1735276150484905e-06,
"loss": 0.1792,
"step": 4265
},
{
"epoch": 9.075451647183847,
"grad_norm": 0.50390625,
"learning_rate": 5.056286300807511e-06,
"loss": 0.1751,
"step": 4270
},
{
"epoch": 9.086078639744953,
"grad_norm": 0.47265625,
"learning_rate": 4.940354243056178e-06,
"loss": 0.174,
"step": 4275
},
{
"epoch": 9.096705632306058,
"grad_norm": 0.458984375,
"learning_rate": 4.825733040476465e-06,
"loss": 0.1732,
"step": 4280
},
{
"epoch": 9.107332624867162,
"grad_norm": 0.51953125,
"learning_rate": 4.712424273673788e-06,
"loss": 0.1767,
"step": 4285
},
{
"epoch": 9.117959617428268,
"grad_norm": 0.44921875,
"learning_rate": 4.600429505155424e-06,
"loss": 0.1726,
"step": 4290
},
{
"epoch": 9.128586609989373,
"grad_norm": 0.4765625,
"learning_rate": 4.489750279308757e-06,
"loss": 0.174,
"step": 4295
},
{
"epoch": 9.139213602550479,
"grad_norm": 0.484375,
"learning_rate": 4.380388122380141e-06,
"loss": 0.1787,
"step": 4300
},
{
"epoch": 9.149840595111584,
"grad_norm": 0.515625,
"learning_rate": 4.2723445424537544e-06,
"loss": 0.1802,
"step": 4305
},
{
"epoch": 9.160467587672688,
"grad_norm": 0.5,
"learning_rate": 4.165621029430855e-06,
"loss": 0.1786,
"step": 4310
},
{
"epoch": 9.171094580233794,
"grad_norm": 0.462890625,
"learning_rate": 4.060219055009229e-06,
"loss": 0.1752,
"step": 4315
},
{
"epoch": 9.181721572794899,
"grad_norm": 0.48828125,
"learning_rate": 3.9561400726628505e-06,
"loss": 0.1793,
"step": 4320
},
{
"epoch": 9.192348565356005,
"grad_norm": 0.56640625,
"learning_rate": 3.85338551762191e-06,
"loss": 0.175,
"step": 4325
},
{
"epoch": 9.20297555791711,
"grad_norm": 0.462890625,
"learning_rate": 3.7519568068529855e-06,
"loss": 0.1728,
"step": 4330
},
{
"epoch": 9.213602550478214,
"grad_norm": 0.478515625,
"learning_rate": 3.651855339039456e-06,
"loss": 0.1698,
"step": 4335
},
{
"epoch": 9.22422954303932,
"grad_norm": 0.69921875,
"learning_rate": 3.5530824945623542e-06,
"loss": 0.1765,
"step": 4340
},
{
"epoch": 9.234856535600425,
"grad_norm": 0.4765625,
"learning_rate": 3.4556396354811226e-06,
"loss": 0.1762,
"step": 4345
},
{
"epoch": 9.24548352816153,
"grad_norm": 0.490234375,
"learning_rate": 3.359528105515064e-06,
"loss": 0.1781,
"step": 4350
},
{
"epoch": 9.256110520722636,
"grad_norm": 0.49609375,
"learning_rate": 3.2647492300245994e-06,
"loss": 0.1754,
"step": 4355
},
{
"epoch": 9.26673751328374,
"grad_norm": 0.50390625,
"learning_rate": 3.1713043159931734e-06,
"loss": 0.1767,
"step": 4360
},
{
"epoch": 9.277364505844846,
"grad_norm": 0.515625,
"learning_rate": 3.079194652009054e-06,
"loss": 0.1708,
"step": 4365
},
{
"epoch": 9.287991498405951,
"grad_norm": 0.48046875,
"learning_rate": 2.9884215082477408e-06,
"loss": 0.1723,
"step": 4370
},
{
"epoch": 9.298618490967057,
"grad_norm": 0.51171875,
"learning_rate": 2.898986136454318e-06,
"loss": 0.1801,
"step": 4375
},
{
"epoch": 9.309245483528162,
"grad_norm": 0.494140625,
"learning_rate": 2.810889769926217e-06,
"loss": 0.1774,
"step": 4380
},
{
"epoch": 9.319872476089266,
"grad_norm": 0.447265625,
"learning_rate": 2.7241336234962944e-06,
"loss": 0.1722,
"step": 4385
},
{
"epoch": 9.330499468650371,
"grad_norm": 0.515625,
"learning_rate": 2.6387188935159456e-06,
"loss": 0.1779,
"step": 4390
},
{
"epoch": 9.341126461211477,
"grad_norm": 0.46875,
"learning_rate": 2.5546467578387635e-06,
"loss": 0.1743,
"step": 4395
},
{
"epoch": 9.351753453772583,
"grad_norm": 0.4921875,
"learning_rate": 2.471918375804105e-06,
"loss": 0.1733,
"step": 4400
},
{
"epoch": 9.362380446333688,
"grad_norm": 0.4921875,
"learning_rate": 2.3905348882212964e-06,
"loss": 0.179,
"step": 4405
},
{
"epoch": 9.373007438894792,
"grad_norm": 0.49609375,
"learning_rate": 2.3104974173537743e-06,
"loss": 0.1742,
"step": 4410
},
{
"epoch": 9.383634431455897,
"grad_norm": 0.484375,
"learning_rate": 2.2318070669036685e-06,
"loss": 0.1743,
"step": 4415
},
{
"epoch": 9.394261424017003,
"grad_norm": 0.4921875,
"learning_rate": 2.1544649219965575e-06,
"loss": 0.1734,
"step": 4420
},
{
"epoch": 9.404888416578109,
"grad_norm": 0.46875,
"learning_rate": 2.07847204916648e-06,
"loss": 0.1729,
"step": 4425
},
{
"epoch": 9.415515409139214,
"grad_norm": 0.546875,
"learning_rate": 2.003829496341325e-06,
"loss": 0.1762,
"step": 4430
},
{
"epoch": 9.426142401700318,
"grad_norm": 0.51953125,
"learning_rate": 1.9305382928282546e-06,
"loss": 0.1798,
"step": 4435
},
{
"epoch": 9.436769394261423,
"grad_norm": 0.51953125,
"learning_rate": 1.8585994492995916e-06,
"loss": 0.1757,
"step": 4440
},
{
"epoch": 9.447396386822529,
"grad_norm": 0.5234375,
"learning_rate": 1.788013957778878e-06,
"loss": 0.1765,
"step": 4445
},
{
"epoch": 9.458023379383635,
"grad_norm": 0.5,
"learning_rate": 1.7187827916271382e-06,
"loss": 0.1739,
"step": 4450
},
{
"epoch": 9.46865037194474,
"grad_norm": 0.45703125,
"learning_rate": 1.65090690552957e-06,
"loss": 0.1732,
"step": 4455
},
{
"epoch": 9.479277364505846,
"grad_norm": 0.498046875,
"learning_rate": 1.5843872354822097e-06,
"loss": 0.173,
"step": 4460
},
{
"epoch": 9.48990435706695,
"grad_norm": 0.5078125,
"learning_rate": 1.5192246987791981e-06,
"loss": 0.177,
"step": 4465
},
{
"epoch": 9.500531349628055,
"grad_norm": 0.49609375,
"learning_rate": 1.4554201940000123e-06,
"loss": 0.177,
"step": 4470
},
{
"epoch": 9.51115834218916,
"grad_norm": 0.5,
"learning_rate": 1.3929746009971433e-06,
"loss": 0.179,
"step": 4475
},
{
"epoch": 9.521785334750266,
"grad_norm": 0.5234375,
"learning_rate": 1.3318887808839274e-06,
"loss": 0.1714,
"step": 4480
},
{
"epoch": 9.532412327311372,
"grad_norm": 0.484375,
"learning_rate": 1.272163576022667e-06,
"loss": 0.1729,
"step": 4485
},
{
"epoch": 9.543039319872475,
"grad_norm": 0.46484375,
"learning_rate": 1.21379981001305e-06,
"loss": 0.1733,
"step": 4490
},
{
"epoch": 9.553666312433581,
"grad_norm": 0.5078125,
"learning_rate": 1.1567982876807603e-06,
"loss": 0.175,
"step": 4495
},
{
"epoch": 9.564293304994687,
"grad_norm": 0.474609375,
"learning_rate": 1.1011597950663865e-06,
"loss": 0.1816,
"step": 4500
},
{
"epoch": 9.574920297555792,
"grad_norm": 0.490234375,
"learning_rate": 1.0468850994145963e-06,
"loss": 0.1777,
"step": 4505
},
{
"epoch": 9.585547290116898,
"grad_norm": 0.5,
"learning_rate": 9.939749491635341e-07,
"loss": 0.1776,
"step": 4510
},
{
"epoch": 9.596174282678001,
"grad_norm": 0.5,
"learning_rate": 9.424300739345304e-07,
"loss": 0.1727,
"step": 4515
},
{
"epoch": 9.606801275239107,
"grad_norm": 0.478515625,
"learning_rate": 8.922511845219971e-07,
"loss": 0.1734,
"step": 4520
},
{
"epoch": 9.617428267800213,
"grad_norm": 0.49609375,
"learning_rate": 8.434389728836589e-07,
"loss": 0.1806,
"step": 4525
},
{
"epoch": 9.628055260361318,
"grad_norm": 0.51953125,
"learning_rate": 7.959941121310266e-07,
"loss": 0.1729,
"step": 4530
},
{
"epoch": 9.638682252922424,
"grad_norm": 0.486328125,
"learning_rate": 7.499172565200718e-07,
"loss": 0.1729,
"step": 4535
},
{
"epoch": 9.649309245483527,
"grad_norm": 0.455078125,
"learning_rate": 7.052090414422119e-07,
"loss": 0.1821,
"step": 4540
},
{
"epoch": 9.659936238044633,
"grad_norm": 0.47265625,
"learning_rate": 6.618700834155945e-07,
"loss": 0.1825,
"step": 4545
},
{
"epoch": 9.670563230605739,
"grad_norm": 0.51171875,
"learning_rate": 6.199009800765265e-07,
"loss": 0.1709,
"step": 4550
},
{
"epoch": 9.681190223166844,
"grad_norm": 0.5,
"learning_rate": 5.793023101712924e-07,
"loss": 0.1771,
"step": 4555
},
{
"epoch": 9.69181721572795,
"grad_norm": 0.5390625,
"learning_rate": 5.400746335481488e-07,
"loss": 0.1771,
"step": 4560
},
{
"epoch": 9.702444208289053,
"grad_norm": 0.52734375,
"learning_rate": 5.022184911495864e-07,
"loss": 0.1807,
"step": 4565
},
{
"epoch": 9.713071200850159,
"grad_norm": 0.474609375,
"learning_rate": 4.6573440500492504e-07,
"loss": 0.1677,
"step": 4570
},
{
"epoch": 9.723698193411265,
"grad_norm": 0.5,
"learning_rate": 4.306228782230304e-07,
"loss": 0.1755,
"step": 4575
},
{
"epoch": 9.73432518597237,
"grad_norm": 0.494140625,
"learning_rate": 3.96884394985475e-07,
"loss": 0.1741,
"step": 4580
},
{
"epoch": 9.744952178533476,
"grad_norm": 0.466796875,
"learning_rate": 3.6451942053975507e-07,
"loss": 0.1708,
"step": 4585
},
{
"epoch": 9.755579171094581,
"grad_norm": 0.490234375,
"learning_rate": 3.335284011929951e-07,
"loss": 0.1732,
"step": 4590
},
{
"epoch": 9.766206163655685,
"grad_norm": 0.48828125,
"learning_rate": 3.0391176430567547e-07,
"loss": 0.1757,
"step": 4595
},
{
"epoch": 9.77683315621679,
"grad_norm": 0.50390625,
"learning_rate": 2.756699182858369e-07,
"loss": 0.1786,
"step": 4600
},
{
"epoch": 9.787460148777896,
"grad_norm": 0.48828125,
"learning_rate": 2.488032525833628e-07,
"loss": 0.1746,
"step": 4605
},
{
"epoch": 9.798087141339002,
"grad_norm": 0.494140625,
"learning_rate": 2.2331213768468363e-07,
"loss": 0.1767,
"step": 4610
},
{
"epoch": 9.808714133900107,
"grad_norm": 0.484375,
"learning_rate": 1.9919692510762533e-07,
"loss": 0.1795,
"step": 4615
},
{
"epoch": 9.819341126461211,
"grad_norm": 0.5078125,
"learning_rate": 1.7645794739654665e-07,
"loss": 0.1764,
"step": 4620
},
{
"epoch": 9.829968119022316,
"grad_norm": 0.5234375,
"learning_rate": 1.5509551811783152e-07,
"loss": 0.1781,
"step": 4625
},
{
"epoch": 9.840595111583422,
"grad_norm": 0.52734375,
"learning_rate": 1.351099318554705e-07,
"loss": 0.1755,
"step": 4630
},
{
"epoch": 9.851222104144528,
"grad_norm": 0.498046875,
"learning_rate": 1.1650146420704166e-07,
"loss": 0.1758,
"step": 4635
},
{
"epoch": 9.861849096705633,
"grad_norm": 0.515625,
"learning_rate": 9.927037177993592e-08,
"loss": 0.1774,
"step": 4640
},
{
"epoch": 9.872476089266737,
"grad_norm": 0.5234375,
"learning_rate": 8.341689218775984e-08,
"loss": 0.1827,
"step": 4645
},
{
"epoch": 9.883103081827842,
"grad_norm": 0.49609375,
"learning_rate": 6.894124404711599e-08,
"loss": 0.1717,
"step": 4650
},
{
"epoch": 9.893730074388948,
"grad_norm": 0.4609375,
"learning_rate": 5.584362697453882e-08,
"loss": 0.1773,
"step": 4655
},
{
"epoch": 9.904357066950054,
"grad_norm": 0.494140625,
"learning_rate": 4.4124221583785595e-08,
"loss": 0.1738,
"step": 4660
},
{
"epoch": 9.91498405951116,
"grad_norm": 0.49609375,
"learning_rate": 3.378318948332737e-08,
"loss": 0.1792,
"step": 4665
},
{
"epoch": 9.925611052072263,
"grad_norm": 0.5859375,
"learning_rate": 2.482067327409521e-08,
"loss": 0.173,
"step": 4670
},
{
"epoch": 9.936238044633368,
"grad_norm": 0.455078125,
"learning_rate": 1.7236796547559497e-08,
"loss": 0.1719,
"step": 4675
},
{
"epoch": 9.946865037194474,
"grad_norm": 0.451171875,
"learning_rate": 1.103166388398691e-08,
"loss": 0.1741,
"step": 4680
},
{
"epoch": 9.95749202975558,
"grad_norm": 0.486328125,
"learning_rate": 6.205360851041508e-09,
"loss": 0.1773,
"step": 4685
},
{
"epoch": 9.968119022316685,
"grad_norm": 0.490234375,
"learning_rate": 2.7579540025524097e-09,
"loss": 0.1762,
"step": 4690
},
{
"epoch": 9.978746014877789,
"grad_norm": 0.44140625,
"learning_rate": 6.894908776255982e-10,
"loss": 0.1745,
"step": 4695
},
{
"epoch": 9.989373007438894,
"grad_norm": 0.51953125,
"learning_rate": 0.0,
"loss": 0.1757,
"step": 4700
},
{
"epoch": 9.989373007438894,
"eval_loss": 3.9657981395721436,
"eval_runtime": 0.7855,
"eval_samples_per_second": 6.366,
"eval_steps_per_second": 1.273,
"step": 4700
},
{
"epoch": 9.989373007438894,
"step": 4700,
"total_flos": 7.255013733871649e+18,
"train_loss": 0.5076825852977469,
"train_runtime": 24662.4426,
"train_samples_per_second": 6.102,
"train_steps_per_second": 0.191
}
],
"logging_steps": 5,
"max_steps": 4700,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 100,
"total_flos": 7.255013733871649e+18,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}