MiniLMv2-L6-H384_R-fineweb-100k / trainer_state.json
pszemraj's picture
End of training
4cefba7
raw
history blame contribute delete
No virus
112 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.9998993609419817,
"eval_steps": 150,
"global_step": 2484,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0040255623207366776,
"grad_norm": 38.58929443359375,
"learning_rate": 4.000000000000001e-06,
"loss": 9.0093,
"num_input_tokens_seen": 327680,
"step": 5
},
{
"epoch": 0.008051124641473355,
"grad_norm": 16.55693817138672,
"learning_rate": 8.000000000000001e-06,
"loss": 8.4074,
"num_input_tokens_seen": 655360,
"step": 10
},
{
"epoch": 0.012076686962210034,
"grad_norm": 10.222895622253418,
"learning_rate": 1.2e-05,
"loss": 7.9014,
"num_input_tokens_seen": 983040,
"step": 15
},
{
"epoch": 0.01610224928294671,
"grad_norm": 8.999139785766602,
"learning_rate": 1.6000000000000003e-05,
"loss": 7.6741,
"num_input_tokens_seen": 1310720,
"step": 20
},
{
"epoch": 0.02012781160368339,
"grad_norm": 8.544699668884277,
"learning_rate": 2e-05,
"loss": 7.3022,
"num_input_tokens_seen": 1638400,
"step": 25
},
{
"epoch": 0.024153373924420067,
"grad_norm": 7.598087310791016,
"learning_rate": 2.4e-05,
"loss": 6.7376,
"num_input_tokens_seen": 1966080,
"step": 30
},
{
"epoch": 0.028178936245156744,
"grad_norm": 5.1022233963012695,
"learning_rate": 2.8e-05,
"loss": 6.2707,
"num_input_tokens_seen": 2293760,
"step": 35
},
{
"epoch": 0.03220449856589342,
"grad_norm": 3.5978548526763916,
"learning_rate": 3.2000000000000005e-05,
"loss": 5.9112,
"num_input_tokens_seen": 2621440,
"step": 40
},
{
"epoch": 0.036230060886630104,
"grad_norm": 2.6466448307037354,
"learning_rate": 3.6e-05,
"loss": 5.6431,
"num_input_tokens_seen": 2949120,
"step": 45
},
{
"epoch": 0.04025562320736678,
"grad_norm": 1.9538609981536865,
"learning_rate": 4e-05,
"loss": 5.4321,
"num_input_tokens_seen": 3276800,
"step": 50
},
{
"epoch": 0.04428118552810346,
"grad_norm": 1.7290977239608765,
"learning_rate": 4.4000000000000006e-05,
"loss": 5.2876,
"num_input_tokens_seen": 3604480,
"step": 55
},
{
"epoch": 0.048306747848840134,
"grad_norm": 1.647463321685791,
"learning_rate": 4.8e-05,
"loss": 5.1749,
"num_input_tokens_seen": 3932160,
"step": 60
},
{
"epoch": 0.05233231016957681,
"grad_norm": 1.367295742034912,
"learning_rate": 5.2000000000000004e-05,
"loss": 5.0826,
"num_input_tokens_seen": 4259840,
"step": 65
},
{
"epoch": 0.05635787249031349,
"grad_norm": 1.4710743427276611,
"learning_rate": 5.6e-05,
"loss": 5.0053,
"num_input_tokens_seen": 4587520,
"step": 70
},
{
"epoch": 0.06038343481105017,
"grad_norm": 1.501865029335022,
"learning_rate": 6.000000000000001e-05,
"loss": 4.8943,
"num_input_tokens_seen": 4915200,
"step": 75
},
{
"epoch": 0.06440899713178684,
"grad_norm": 1.3167399168014526,
"learning_rate": 6.400000000000001e-05,
"loss": 4.9062,
"num_input_tokens_seen": 5242880,
"step": 80
},
{
"epoch": 0.06843455945252352,
"grad_norm": 1.6458464860916138,
"learning_rate": 6.8e-05,
"loss": 4.8432,
"num_input_tokens_seen": 5570560,
"step": 85
},
{
"epoch": 0.07246012177326021,
"grad_norm": 1.4131243228912354,
"learning_rate": 7.2e-05,
"loss": 4.8507,
"num_input_tokens_seen": 5898240,
"step": 90
},
{
"epoch": 0.07648568409399688,
"grad_norm": 1.6352394819259644,
"learning_rate": 7.6e-05,
"loss": 4.8214,
"num_input_tokens_seen": 6225920,
"step": 95
},
{
"epoch": 0.08051124641473356,
"grad_norm": 1.3868008852005005,
"learning_rate": 8e-05,
"loss": 4.7863,
"num_input_tokens_seen": 6553600,
"step": 100
},
{
"epoch": 0.08453680873547023,
"grad_norm": 1.5053304433822632,
"learning_rate": 7.807200583588265e-05,
"loss": 4.7478,
"num_input_tokens_seen": 6881280,
"step": 105
},
{
"epoch": 0.08856237105620691,
"grad_norm": 1.6993811130523682,
"learning_rate": 7.627700713964738e-05,
"loss": 4.7303,
"num_input_tokens_seen": 7208960,
"step": 110
},
{
"epoch": 0.0925879333769436,
"grad_norm": 1.4282785654067993,
"learning_rate": 7.460038465922511e-05,
"loss": 4.7332,
"num_input_tokens_seen": 7536640,
"step": 115
},
{
"epoch": 0.09661349569768027,
"grad_norm": 1.6601049900054932,
"learning_rate": 7.302967433402216e-05,
"loss": 4.7639,
"num_input_tokens_seen": 7864320,
"step": 120
},
{
"epoch": 0.10063905801841695,
"grad_norm": 1.4113882780075073,
"learning_rate": 7.155417527999328e-05,
"loss": 4.7138,
"num_input_tokens_seen": 8192000,
"step": 125
},
{
"epoch": 0.10466462033915362,
"grad_norm": 1.5578097105026245,
"learning_rate": 7.016464154456235e-05,
"loss": 4.7051,
"num_input_tokens_seen": 8519680,
"step": 130
},
{
"epoch": 0.1086901826598903,
"grad_norm": 1.4274314641952515,
"learning_rate": 6.885303726590964e-05,
"loss": 4.736,
"num_input_tokens_seen": 8847360,
"step": 135
},
{
"epoch": 0.11271574498062698,
"grad_norm": 1.354647159576416,
"learning_rate": 6.761234037828134e-05,
"loss": 4.6927,
"num_input_tokens_seen": 9175040,
"step": 140
},
{
"epoch": 0.11674130730136366,
"grad_norm": 1.402042031288147,
"learning_rate": 6.643638388299198e-05,
"loss": 4.703,
"num_input_tokens_seen": 9502720,
"step": 145
},
{
"epoch": 0.12076686962210034,
"grad_norm": 1.4121787548065186,
"learning_rate": 6.531972647421809e-05,
"loss": 4.6583,
"num_input_tokens_seen": 9830400,
"step": 150
},
{
"epoch": 0.12076686962210034,
"eval_accuracy": 0.3405695459290188,
"eval_loss": 4.505166053771973,
"eval_runtime": 6.5101,
"eval_samples_per_second": 46.082,
"eval_steps_per_second": 5.837,
"num_input_tokens_seen": 9830400,
"step": 150
},
{
"epoch": 0.12479243194283701,
"grad_norm": 1.5172020196914673,
"learning_rate": 6.425754631219992e-05,
"loss": 4.6715,
"num_input_tokens_seen": 10158080,
"step": 155
},
{
"epoch": 0.12881799426357368,
"grad_norm": 1.4244070053100586,
"learning_rate": 6.324555320336759e-05,
"loss": 4.6351,
"num_input_tokens_seen": 10485760,
"step": 160
},
{
"epoch": 0.13284355658431038,
"grad_norm": 1.352230191230774,
"learning_rate": 6.227991553292185e-05,
"loss": 4.6386,
"num_input_tokens_seen": 10813440,
"step": 165
},
{
"epoch": 0.13686911890504705,
"grad_norm": 1.3570430278778076,
"learning_rate": 6.135719910778964e-05,
"loss": 4.6563,
"num_input_tokens_seen": 11141120,
"step": 170
},
{
"epoch": 0.14089468122578372,
"grad_norm": 1.3318084478378296,
"learning_rate": 6.047431568147636e-05,
"loss": 4.6472,
"num_input_tokens_seen": 11468800,
"step": 175
},
{
"epoch": 0.14492024354652042,
"grad_norm": 1.4813034534454346,
"learning_rate": 5.96284793999944e-05,
"loss": 4.6493,
"num_input_tokens_seen": 11796480,
"step": 180
},
{
"epoch": 0.1489458058672571,
"grad_norm": 1.367981195449829,
"learning_rate": 5.881716976750463e-05,
"loss": 4.5858,
"num_input_tokens_seen": 12124160,
"step": 185
},
{
"epoch": 0.15297136818799376,
"grad_norm": 1.3320348262786865,
"learning_rate": 5.803810000880094e-05,
"loss": 4.5985,
"num_input_tokens_seen": 12451840,
"step": 190
},
{
"epoch": 0.15699693050873043,
"grad_norm": 1.233577847480774,
"learning_rate": 5.7289189923154636e-05,
"loss": 4.6574,
"num_input_tokens_seen": 12779520,
"step": 195
},
{
"epoch": 0.16102249282946712,
"grad_norm": 1.2315726280212402,
"learning_rate": 5.6568542494923805e-05,
"loss": 4.6412,
"num_input_tokens_seen": 13107200,
"step": 200
},
{
"epoch": 0.1650480551502038,
"grad_norm": 1.3596357107162476,
"learning_rate": 5.5874423661566265e-05,
"loss": 4.6032,
"num_input_tokens_seen": 13434880,
"step": 205
},
{
"epoch": 0.16907361747094046,
"grad_norm": 1.3829398155212402,
"learning_rate": 5.5205244747388335e-05,
"loss": 4.5954,
"num_input_tokens_seen": 13762560,
"step": 210
},
{
"epoch": 0.17309917979167716,
"grad_norm": 1.4083398580551147,
"learning_rate": 5.455954715763789e-05,
"loss": 4.5888,
"num_input_tokens_seen": 14090240,
"step": 215
},
{
"epoch": 0.17712474211241383,
"grad_norm": 1.467612862586975,
"learning_rate": 5.393598899705937e-05,
"loss": 4.6193,
"num_input_tokens_seen": 14417920,
"step": 220
},
{
"epoch": 0.1811503044331505,
"grad_norm": 1.2575774192810059,
"learning_rate": 5.333333333333333e-05,
"loss": 4.5579,
"num_input_tokens_seen": 14745600,
"step": 225
},
{
"epoch": 0.1851758667538872,
"grad_norm": 1.4239579439163208,
"learning_rate": 5.2750437871662975e-05,
"loss": 4.5607,
"num_input_tokens_seen": 15073280,
"step": 230
},
{
"epoch": 0.18920142907462387,
"grad_norm": 1.2363418340682983,
"learning_rate": 5.218624584427538e-05,
"loss": 4.5617,
"num_input_tokens_seen": 15400960,
"step": 235
},
{
"epoch": 0.19322699139536054,
"grad_norm": 1.242336392402649,
"learning_rate": 5.163977794943223e-05,
"loss": 4.5878,
"num_input_tokens_seen": 15728640,
"step": 240
},
{
"epoch": 0.1972525537160972,
"grad_norm": 1.2200363874435425,
"learning_rate": 5.11101251999952e-05,
"loss": 4.5518,
"num_input_tokens_seen": 16056320,
"step": 245
},
{
"epoch": 0.2012781160368339,
"grad_norm": 1.1616241931915283,
"learning_rate": 5.0596442562694074e-05,
"loss": 4.5944,
"num_input_tokens_seen": 16384000,
"step": 250
},
{
"epoch": 0.20530367835757057,
"grad_norm": 1.3502942323684692,
"learning_rate": 5.009794328681197e-05,
"loss": 4.5356,
"num_input_tokens_seen": 16711680,
"step": 255
},
{
"epoch": 0.20932924067830724,
"grad_norm": 1.1606285572052002,
"learning_rate": 4.961389383568338e-05,
"loss": 4.5341,
"num_input_tokens_seen": 17039360,
"step": 260
},
{
"epoch": 0.21335480299904394,
"grad_norm": 1.4353660345077515,
"learning_rate": 4.9143609346716104e-05,
"loss": 4.5766,
"num_input_tokens_seen": 17367040,
"step": 265
},
{
"epoch": 0.2173803653197806,
"grad_norm": 1.2260538339614868,
"learning_rate": 4.8686449556014764e-05,
"loss": 4.5123,
"num_input_tokens_seen": 17694720,
"step": 270
},
{
"epoch": 0.22140592764051728,
"grad_norm": 1.2337218523025513,
"learning_rate": 4.8241815132442184e-05,
"loss": 4.4989,
"num_input_tokens_seen": 18022400,
"step": 275
},
{
"epoch": 0.22543148996125395,
"grad_norm": 1.2576812505722046,
"learning_rate": 4.7809144373375745e-05,
"loss": 4.5362,
"num_input_tokens_seen": 18350080,
"step": 280
},
{
"epoch": 0.22945705228199065,
"grad_norm": 1.3166711330413818,
"learning_rate": 4.738791022072739e-05,
"loss": 4.5318,
"num_input_tokens_seen": 18677760,
"step": 285
},
{
"epoch": 0.23348261460272732,
"grad_norm": 1.353780746459961,
"learning_rate": 4.6977617561176284e-05,
"loss": 4.5475,
"num_input_tokens_seen": 19005440,
"step": 290
},
{
"epoch": 0.237508176923464,
"grad_norm": 1.2797174453735352,
"learning_rate": 4.657780077916657e-05,
"loss": 4.4992,
"num_input_tokens_seen": 19333120,
"step": 295
},
{
"epoch": 0.24153373924420068,
"grad_norm": 1.433968424797058,
"learning_rate": 4.618802153517007e-05,
"loss": 4.5365,
"num_input_tokens_seen": 19660800,
"step": 300
},
{
"epoch": 0.24153373924420068,
"eval_accuracy": 0.35246274229682784,
"eval_loss": 4.371213436126709,
"eval_runtime": 6.282,
"eval_samples_per_second": 47.755,
"eval_steps_per_second": 6.049,
"num_input_tokens_seen": 19660800,
"step": 300
},
{
"epoch": 0.24555930156493735,
"grad_norm": 1.3408712148666382,
"learning_rate": 4.580786674510946e-05,
"loss": 4.4919,
"num_input_tokens_seen": 19988480,
"step": 305
},
{
"epoch": 0.24958486388567402,
"grad_norm": 1.3724451065063477,
"learning_rate": 4.543694673976519e-05,
"loss": 4.498,
"num_input_tokens_seen": 20316160,
"step": 310
},
{
"epoch": 0.2536104262064107,
"grad_norm": 1.3110082149505615,
"learning_rate": 4.507489358552088e-05,
"loss": 4.4623,
"num_input_tokens_seen": 20643840,
"step": 315
},
{
"epoch": 0.25763598852714736,
"grad_norm": 1.2995046377182007,
"learning_rate": 4.47213595499958e-05,
"loss": 4.5249,
"num_input_tokens_seen": 20971520,
"step": 320
},
{
"epoch": 0.26166155084788406,
"grad_norm": 1.2765010595321655,
"learning_rate": 4.4376015698018335e-05,
"loss": 4.5525,
"num_input_tokens_seen": 21299200,
"step": 325
},
{
"epoch": 0.26568711316862076,
"grad_norm": 1.268876552581787,
"learning_rate": 4.403855060505443e-05,
"loss": 4.4876,
"num_input_tokens_seen": 21626880,
"step": 330
},
{
"epoch": 0.2697126754893574,
"grad_norm": 1.4470906257629395,
"learning_rate": 4.3708669176652244e-05,
"loss": 4.4901,
"num_input_tokens_seen": 21954560,
"step": 335
},
{
"epoch": 0.2737382378100941,
"grad_norm": 1.3335119485855103,
"learning_rate": 4.3386091563731235e-05,
"loss": 4.4488,
"num_input_tokens_seen": 22282240,
"step": 340
},
{
"epoch": 0.2777638001308308,
"grad_norm": 1.4562543630599976,
"learning_rate": 4.3070552164653247e-05,
"loss": 4.4705,
"num_input_tokens_seen": 22609920,
"step": 345
},
{
"epoch": 0.28178936245156744,
"grad_norm": 1.4459863901138306,
"learning_rate": 4.276179870598791e-05,
"loss": 4.4942,
"num_input_tokens_seen": 22937600,
"step": 350
},
{
"epoch": 0.28581492477230414,
"grad_norm": 1.193866491317749,
"learning_rate": 4.2459591394742024e-05,
"loss": 4.5002,
"num_input_tokens_seen": 23265280,
"step": 355
},
{
"epoch": 0.28984048709304083,
"grad_norm": 1.4731342792510986,
"learning_rate": 4.2163702135578394e-05,
"loss": 4.4794,
"num_input_tokens_seen": 23592960,
"step": 360
},
{
"epoch": 0.2938660494137775,
"grad_norm": 1.1683562994003296,
"learning_rate": 4.1873913807217094e-05,
"loss": 4.4578,
"num_input_tokens_seen": 23920640,
"step": 365
},
{
"epoch": 0.2978916117345142,
"grad_norm": 1.2372169494628906,
"learning_rate": 4.159001959280291e-05,
"loss": 4.5118,
"num_input_tokens_seen": 24248320,
"step": 370
},
{
"epoch": 0.30191717405525087,
"grad_norm": 1.4430092573165894,
"learning_rate": 4.131182235954578e-05,
"loss": 4.4861,
"num_input_tokens_seen": 24576000,
"step": 375
},
{
"epoch": 0.3059427363759875,
"grad_norm": 1.3162908554077148,
"learning_rate": 4.103913408340617e-05,
"loss": 4.4968,
"num_input_tokens_seen": 24903680,
"step": 380
},
{
"epoch": 0.3099682986967242,
"grad_norm": 1.238364815711975,
"learning_rate": 4.077177531501004e-05,
"loss": 4.4465,
"num_input_tokens_seen": 25231360,
"step": 385
},
{
"epoch": 0.31399386101746085,
"grad_norm": 1.2819691896438599,
"learning_rate": 4.050957468334667e-05,
"loss": 4.4656,
"num_input_tokens_seen": 25559040,
"step": 390
},
{
"epoch": 0.31801942333819755,
"grad_norm": 1.2802215814590454,
"learning_rate": 4.025236843413021e-05,
"loss": 4.4934,
"num_input_tokens_seen": 25886720,
"step": 395
},
{
"epoch": 0.32204498565893425,
"grad_norm": 1.232289433479309,
"learning_rate": 4e-05,
"loss": 4.4965,
"num_input_tokens_seen": 26214400,
"step": 400
},
{
"epoch": 0.3260705479796709,
"grad_norm": 1.3022078275680542,
"learning_rate": 3.975231959999627e-05,
"loss": 4.4332,
"num_input_tokens_seen": 26542080,
"step": 405
},
{
"epoch": 0.3300961103004076,
"grad_norm": 1.1387150287628174,
"learning_rate": 3.950918386598359e-05,
"loss": 4.4345,
"num_input_tokens_seen": 26869760,
"step": 410
},
{
"epoch": 0.3341216726211443,
"grad_norm": 1.1623753309249878,
"learning_rate": 3.9270455493905284e-05,
"loss": 4.4572,
"num_input_tokens_seen": 27197440,
"step": 415
},
{
"epoch": 0.3381472349418809,
"grad_norm": 1.1756770610809326,
"learning_rate": 3.9036002917941324e-05,
"loss": 4.4603,
"num_input_tokens_seen": 27525120,
"step": 420
},
{
"epoch": 0.3421727972626176,
"grad_norm": 1.243285059928894,
"learning_rate": 3.8805700005813275e-05,
"loss": 4.4576,
"num_input_tokens_seen": 27852800,
"step": 425
},
{
"epoch": 0.3461983595833543,
"grad_norm": 1.4163405895233154,
"learning_rate": 3.8579425773632976e-05,
"loss": 4.4693,
"num_input_tokens_seen": 28180480,
"step": 430
},
{
"epoch": 0.35022392190409096,
"grad_norm": 1.2970571517944336,
"learning_rate": 3.835706411883074e-05,
"loss": 4.4731,
"num_input_tokens_seen": 28508160,
"step": 435
},
{
"epoch": 0.35424948422482766,
"grad_norm": 1.2691751718521118,
"learning_rate": 3.813850356982369e-05,
"loss": 4.4548,
"num_input_tokens_seen": 28835840,
"step": 440
},
{
"epoch": 0.35827504654556436,
"grad_norm": 1.2060904502868652,
"learning_rate": 3.792363705119819e-05,
"loss": 4.5089,
"num_input_tokens_seen": 29163520,
"step": 445
},
{
"epoch": 0.362300608866301,
"grad_norm": 1.1918907165527344,
"learning_rate": 3.771236166328254e-05,
"loss": 4.4621,
"num_input_tokens_seen": 29491200,
"step": 450
},
{
"epoch": 0.362300608866301,
"eval_accuracy": 0.3575167675445771,
"eval_loss": 4.2810378074646,
"eval_runtime": 6.4142,
"eval_samples_per_second": 46.771,
"eval_steps_per_second": 5.924,
"num_input_tokens_seen": 29491200,
"step": 450
},
{
"epoch": 0.3663261711870377,
"grad_norm": 1.257439136505127,
"learning_rate": 3.750457847507964e-05,
"loss": 4.4083,
"num_input_tokens_seen": 29818880,
"step": 455
},
{
"epoch": 0.3703517335077744,
"grad_norm": 1.2655524015426636,
"learning_rate": 3.7300192329612555e-05,
"loss": 4.428,
"num_input_tokens_seen": 30146560,
"step": 460
},
{
"epoch": 0.37437729582851104,
"grad_norm": 1.3658162355422974,
"learning_rate": 3.709911166081346e-05,
"loss": 4.4843,
"num_input_tokens_seen": 30474240,
"step": 465
},
{
"epoch": 0.37840285814924773,
"grad_norm": 1.32870352268219,
"learning_rate": 3.6901248321155405e-05,
"loss": 4.449,
"num_input_tokens_seen": 30801920,
"step": 470
},
{
"epoch": 0.3824284204699844,
"grad_norm": 1.3358460664749146,
"learning_rate": 3.670651741928988e-05,
"loss": 4.435,
"num_input_tokens_seen": 31129600,
"step": 475
},
{
"epoch": 0.3864539827907211,
"grad_norm": 1.1790028810501099,
"learning_rate": 3.651483716701108e-05,
"loss": 4.426,
"num_input_tokens_seen": 31457280,
"step": 480
},
{
"epoch": 0.39047954511145777,
"grad_norm": 1.2213348150253296,
"learning_rate": 3.6326128734919986e-05,
"loss": 4.4451,
"num_input_tokens_seen": 31784960,
"step": 485
},
{
"epoch": 0.3945051074321944,
"grad_norm": 1.4075285196304321,
"learning_rate": 3.6140316116210055e-05,
"loss": 4.4582,
"num_input_tokens_seen": 32112640,
"step": 490
},
{
"epoch": 0.3985306697529311,
"grad_norm": 1.3073724508285522,
"learning_rate": 3.595732599803958e-05,
"loss": 4.4541,
"num_input_tokens_seen": 32440320,
"step": 495
},
{
"epoch": 0.4025562320736678,
"grad_norm": 1.3178701400756836,
"learning_rate": 3.577708763999664e-05,
"loss": 4.4466,
"num_input_tokens_seen": 32768000,
"step": 500
},
{
"epoch": 0.40658179439440445,
"grad_norm": 1.2765254974365234,
"learning_rate": 3.5599532759198786e-05,
"loss": 4.4159,
"num_input_tokens_seen": 33095680,
"step": 505
},
{
"epoch": 0.41060735671514115,
"grad_norm": 1.2380224466323853,
"learning_rate": 3.542459542160382e-05,
"loss": 4.4387,
"num_input_tokens_seen": 33423360,
"step": 510
},
{
"epoch": 0.41463291903587784,
"grad_norm": 1.4157094955444336,
"learning_rate": 3.525221193913854e-05,
"loss": 4.4529,
"num_input_tokens_seen": 33751040,
"step": 515
},
{
"epoch": 0.4186584813566145,
"grad_norm": 1.3274729251861572,
"learning_rate": 3.5082320772281174e-05,
"loss": 4.3776,
"num_input_tokens_seen": 34078720,
"step": 520
},
{
"epoch": 0.4226840436773512,
"grad_norm": 1.2400481700897217,
"learning_rate": 3.4914862437758787e-05,
"loss": 4.4103,
"num_input_tokens_seen": 34406400,
"step": 525
},
{
"epoch": 0.4267096059980879,
"grad_norm": 1.2797058820724487,
"learning_rate": 3.4749779421045556e-05,
"loss": 4.4005,
"num_input_tokens_seen": 34734080,
"step": 530
},
{
"epoch": 0.4307351683188245,
"grad_norm": 1.4183690547943115,
"learning_rate": 3.458701609336936e-05,
"loss": 4.4181,
"num_input_tokens_seen": 35061760,
"step": 535
},
{
"epoch": 0.4347607306395612,
"grad_norm": 1.2432236671447754,
"learning_rate": 3.442651863295482e-05,
"loss": 4.4037,
"num_input_tokens_seen": 35389440,
"step": 540
},
{
"epoch": 0.4387862929602979,
"grad_norm": 1.19386887550354,
"learning_rate": 3.426823495024955e-05,
"loss": 4.4211,
"num_input_tokens_seen": 35717120,
"step": 545
},
{
"epoch": 0.44281185528103456,
"grad_norm": 1.24849534034729,
"learning_rate": 3.4112114616897666e-05,
"loss": 4.3953,
"num_input_tokens_seen": 36044800,
"step": 550
},
{
"epoch": 0.44683741760177126,
"grad_norm": 1.230302333831787,
"learning_rate": 3.395810879824072e-05,
"loss": 4.4187,
"num_input_tokens_seen": 36372480,
"step": 555
},
{
"epoch": 0.4508629799225079,
"grad_norm": 1.293820858001709,
"learning_rate": 3.380617018914067e-05,
"loss": 4.3944,
"num_input_tokens_seen": 36700160,
"step": 560
},
{
"epoch": 0.4548885422432446,
"grad_norm": 1.3907512426376343,
"learning_rate": 3.36562529529337e-05,
"loss": 4.4285,
"num_input_tokens_seen": 37027840,
"step": 565
},
{
"epoch": 0.4589141045639813,
"grad_norm": 1.4436863660812378,
"learning_rate": 3.350831266333564e-05,
"loss": 4.4322,
"num_input_tokens_seen": 37355520,
"step": 570
},
{
"epoch": 0.46293966688471794,
"grad_norm": 1.3175897598266602,
"learning_rate": 3.336230624913197e-05,
"loss": 4.4196,
"num_input_tokens_seen": 37683200,
"step": 575
},
{
"epoch": 0.46696522920545463,
"grad_norm": 1.1326998472213745,
"learning_rate": 3.321819194149599e-05,
"loss": 4.4478,
"num_input_tokens_seen": 38010880,
"step": 580
},
{
"epoch": 0.47099079152619133,
"grad_norm": 1.3267349004745483,
"learning_rate": 3.307592922378893e-05,
"loss": 4.4068,
"num_input_tokens_seen": 38338560,
"step": 585
},
{
"epoch": 0.475016353846928,
"grad_norm": 1.2397940158843994,
"learning_rate": 3.2935478783704735e-05,
"loss": 4.3565,
"num_input_tokens_seen": 38666240,
"step": 590
},
{
"epoch": 0.47904191616766467,
"grad_norm": 1.290412187576294,
"learning_rate": 3.279680246763151e-05,
"loss": 4.3957,
"num_input_tokens_seen": 38993920,
"step": 595
},
{
"epoch": 0.48306747848840137,
"grad_norm": 1.3627963066101074,
"learning_rate": 3.2659863237109046e-05,
"loss": 4.4116,
"num_input_tokens_seen": 39321600,
"step": 600
},
{
"epoch": 0.48306747848840137,
"eval_accuracy": 0.36151503242287286,
"eval_loss": 4.246631622314453,
"eval_runtime": 6.3292,
"eval_samples_per_second": 47.4,
"eval_steps_per_second": 6.004,
"num_input_tokens_seen": 39321600,
"step": 600
},
{
"epoch": 0.487093040809138,
"grad_norm": 1.2708163261413574,
"learning_rate": 3.2524625127269674e-05,
"loss": 4.3632,
"num_input_tokens_seen": 39649280,
"step": 605
},
{
"epoch": 0.4911186031298747,
"grad_norm": 1.1186234951019287,
"learning_rate": 3.2391053207156646e-05,
"loss": 4.3974,
"num_input_tokens_seen": 39976960,
"step": 610
},
{
"epoch": 0.4951441654506114,
"grad_norm": 1.2380205392837524,
"learning_rate": 3.2259113541820474e-05,
"loss": 4.3891,
"num_input_tokens_seen": 40304640,
"step": 615
},
{
"epoch": 0.49916972777134805,
"grad_norm": 1.220374584197998,
"learning_rate": 3.212877315609996e-05,
"loss": 4.3911,
"num_input_tokens_seen": 40632320,
"step": 620
},
{
"epoch": 0.5031952900920847,
"grad_norm": 1.273556113243103,
"learning_rate": 3.2000000000000005e-05,
"loss": 4.3715,
"num_input_tokens_seen": 40960000,
"step": 625
},
{
"epoch": 0.5072208524128214,
"grad_norm": 1.2880346775054932,
"learning_rate": 3.187276291558383e-05,
"loss": 4.4068,
"num_input_tokens_seen": 41287680,
"step": 630
},
{
"epoch": 0.5112464147335581,
"grad_norm": 1.3127074241638184,
"learning_rate": 3.1747031605301803e-05,
"loss": 4.3636,
"num_input_tokens_seen": 41615360,
"step": 635
},
{
"epoch": 0.5152719770542947,
"grad_norm": 1.2499847412109375,
"learning_rate": 3.1622776601683795e-05,
"loss": 4.384,
"num_input_tokens_seen": 41943040,
"step": 640
},
{
"epoch": 0.5192975393750314,
"grad_norm": 1.28691828250885,
"learning_rate": 3.1499969238326315e-05,
"loss": 4.3927,
"num_input_tokens_seen": 42270720,
"step": 645
},
{
"epoch": 0.5233231016957681,
"grad_norm": 1.221285104751587,
"learning_rate": 3.137858162210945e-05,
"loss": 4.3894,
"num_input_tokens_seen": 42598400,
"step": 650
},
{
"epoch": 0.5273486640165048,
"grad_norm": 1.2766237258911133,
"learning_rate": 3.125858660658254e-05,
"loss": 4.3689,
"num_input_tokens_seen": 42926080,
"step": 655
},
{
"epoch": 0.5313742263372415,
"grad_norm": 1.3662595748901367,
"learning_rate": 3.113995776646092e-05,
"loss": 4.3842,
"num_input_tokens_seen": 43253760,
"step": 660
},
{
"epoch": 0.5353997886579782,
"grad_norm": 1.264195203781128,
"learning_rate": 3.102266937317925e-05,
"loss": 4.3665,
"num_input_tokens_seen": 43581440,
"step": 665
},
{
"epoch": 0.5394253509787148,
"grad_norm": 1.4342658519744873,
"learning_rate": 3.0906696371450235e-05,
"loss": 4.3987,
"num_input_tokens_seen": 43909120,
"step": 670
},
{
"epoch": 0.5434509132994515,
"grad_norm": 1.2556202411651611,
"learning_rate": 3.0792014356780046e-05,
"loss": 4.3837,
"num_input_tokens_seen": 44236800,
"step": 675
},
{
"epoch": 0.5474764756201882,
"grad_norm": 1.3614763021469116,
"learning_rate": 3.067859955389482e-05,
"loss": 4.339,
"num_input_tokens_seen": 44564480,
"step": 680
},
{
"epoch": 0.5515020379409249,
"grad_norm": 1.3664342164993286,
"learning_rate": 3.0566428796034804e-05,
"loss": 4.3543,
"num_input_tokens_seen": 44892160,
"step": 685
},
{
"epoch": 0.5555276002616616,
"grad_norm": 1.2612766027450562,
"learning_rate": 3.045547950507524e-05,
"loss": 4.3698,
"num_input_tokens_seen": 45219840,
"step": 690
},
{
"epoch": 0.5595531625823982,
"grad_norm": 1.3563127517700195,
"learning_rate": 3.0345729672435268e-05,
"loss": 4.3561,
"num_input_tokens_seen": 45547520,
"step": 695
},
{
"epoch": 0.5635787249031349,
"grad_norm": 1.15888249874115,
"learning_rate": 3.023715784073818e-05,
"loss": 4.4154,
"num_input_tokens_seen": 45875200,
"step": 700
},
{
"epoch": 0.5676042872238716,
"grad_norm": 1.2462992668151855,
"learning_rate": 3.0129743086188382e-05,
"loss": 4.4011,
"num_input_tokens_seen": 46202880,
"step": 705
},
{
"epoch": 0.5716298495446083,
"grad_norm": 1.2652802467346191,
"learning_rate": 3.0023465001632065e-05,
"loss": 4.4053,
"num_input_tokens_seen": 46530560,
"step": 710
},
{
"epoch": 0.575655411865345,
"grad_norm": 1.4011398553848267,
"learning_rate": 2.991830368027063e-05,
"loss": 4.3539,
"num_input_tokens_seen": 46858240,
"step": 715
},
{
"epoch": 0.5796809741860817,
"grad_norm": 1.248789668083191,
"learning_rate": 2.98142396999972e-05,
"loss": 4.3758,
"num_input_tokens_seen": 47185920,
"step": 720
},
{
"epoch": 0.5837065365068183,
"grad_norm": 1.218556523323059,
"learning_rate": 2.97112541083283e-05,
"loss": 4.3695,
"num_input_tokens_seen": 47513600,
"step": 725
},
{
"epoch": 0.587732098827555,
"grad_norm": 1.28030526638031,
"learning_rate": 2.960932840790421e-05,
"loss": 4.3676,
"num_input_tokens_seen": 47841280,
"step": 730
},
{
"epoch": 0.5917576611482916,
"grad_norm": 1.1834255456924438,
"learning_rate": 2.95084445425327e-05,
"loss": 4.3445,
"num_input_tokens_seen": 48168960,
"step": 735
},
{
"epoch": 0.5957832234690283,
"grad_norm": 1.324885606765747,
"learning_rate": 2.9408584883752315e-05,
"loss": 4.3495,
"num_input_tokens_seen": 48496640,
"step": 740
},
{
"epoch": 0.599808785789765,
"grad_norm": 1.3806536197662354,
"learning_rate": 2.930973221789247e-05,
"loss": 4.3369,
"num_input_tokens_seen": 48824320,
"step": 745
},
{
"epoch": 0.6038343481105017,
"grad_norm": 1.239935278892517,
"learning_rate": 2.9211869733608858e-05,
"loss": 4.3487,
"num_input_tokens_seen": 49152000,
"step": 750
},
{
"epoch": 0.6038343481105017,
"eval_accuracy": 0.3661367484834038,
"eval_loss": 4.179458141326904,
"eval_runtime": 6.3307,
"eval_samples_per_second": 47.388,
"eval_steps_per_second": 6.003,
"num_input_tokens_seen": 49152000,
"step": 750
},
{
"epoch": 0.6078599104312383,
"grad_norm": 1.294979214668274,
"learning_rate": 2.911498100987356e-05,
"loss": 4.3559,
"num_input_tokens_seen": 49479680,
"step": 755
},
{
"epoch": 0.611885472751975,
"grad_norm": 1.2415525913238525,
"learning_rate": 2.901905000440047e-05,
"loss": 4.3862,
"num_input_tokens_seen": 49807360,
"step": 760
},
{
"epoch": 0.6159110350727117,
"grad_norm": 1.2934507131576538,
"learning_rate": 2.8924061042487487e-05,
"loss": 4.3778,
"num_input_tokens_seen": 50135040,
"step": 765
},
{
"epoch": 0.6199365973934484,
"grad_norm": 1.3969470262527466,
"learning_rate": 2.8829998806257887e-05,
"loss": 4.3623,
"num_input_tokens_seen": 50462720,
"step": 770
},
{
"epoch": 0.6239621597141851,
"grad_norm": 1.3450722694396973,
"learning_rate": 2.873684832428399e-05,
"loss": 4.3819,
"num_input_tokens_seen": 50790400,
"step": 775
},
{
"epoch": 0.6279877220349217,
"grad_norm": 1.4024951457977295,
"learning_rate": 2.8644594961577318e-05,
"loss": 4.3577,
"num_input_tokens_seen": 51118080,
"step": 780
},
{
"epoch": 0.6320132843556584,
"grad_norm": 1.4493414163589478,
"learning_rate": 2.855322440992999e-05,
"loss": 4.3402,
"num_input_tokens_seen": 51445760,
"step": 785
},
{
"epoch": 0.6360388466763951,
"grad_norm": 1.12810480594635,
"learning_rate": 2.8462722678592803e-05,
"loss": 4.3095,
"num_input_tokens_seen": 51773440,
"step": 790
},
{
"epoch": 0.6400644089971318,
"grad_norm": 1.2812013626098633,
"learning_rate": 2.837307608527635e-05,
"loss": 4.3664,
"num_input_tokens_seen": 52101120,
"step": 795
},
{
"epoch": 0.6440899713178685,
"grad_norm": 1.269395112991333,
"learning_rate": 2.8284271247461902e-05,
"loss": 4.3699,
"num_input_tokens_seen": 52428800,
"step": 800
},
{
"epoch": 0.6481155336386052,
"grad_norm": 1.2164851427078247,
"learning_rate": 2.819629507400966e-05,
"loss": 4.3331,
"num_input_tokens_seen": 52756480,
"step": 805
},
{
"epoch": 0.6521410959593418,
"grad_norm": 1.249258279800415,
"learning_rate": 2.8109134757052262e-05,
"loss": 4.3384,
"num_input_tokens_seen": 53084160,
"step": 810
},
{
"epoch": 0.6561666582800785,
"grad_norm": 1.2432383298873901,
"learning_rate": 2.8022777764162353e-05,
"loss": 4.3628,
"num_input_tokens_seen": 53411840,
"step": 815
},
{
"epoch": 0.6601922206008152,
"grad_norm": 1.2008548974990845,
"learning_rate": 2.7937211830783133e-05,
"loss": 4.3751,
"num_input_tokens_seen": 53739520,
"step": 820
},
{
"epoch": 0.6642177829215519,
"grad_norm": 1.3261739015579224,
"learning_rate": 2.785242495291166e-05,
"loss": 4.3405,
"num_input_tokens_seen": 54067200,
"step": 825
},
{
"epoch": 0.6682433452422886,
"grad_norm": 1.3359473943710327,
"learning_rate": 2.7768405380024934e-05,
"loss": 4.3384,
"num_input_tokens_seen": 54394880,
"step": 830
},
{
"epoch": 0.6722689075630253,
"grad_norm": 1.3322173357009888,
"learning_rate": 2.7685141608239315e-05,
"loss": 4.3532,
"num_input_tokens_seen": 54722560,
"step": 835
},
{
"epoch": 0.6762944698837619,
"grad_norm": 1.2270052433013916,
"learning_rate": 2.7602622373694168e-05,
"loss": 4.3314,
"num_input_tokens_seen": 55050240,
"step": 840
},
{
"epoch": 0.6803200322044985,
"grad_norm": 1.2989898920059204,
"learning_rate": 2.752083664615126e-05,
"loss": 4.3486,
"num_input_tokens_seen": 55377920,
"step": 845
},
{
"epoch": 0.6843455945252352,
"grad_norm": 1.2281562089920044,
"learning_rate": 2.7439773622801412e-05,
"loss": 4.3269,
"num_input_tokens_seen": 55705600,
"step": 850
},
{
"epoch": 0.6883711568459719,
"grad_norm": 1.3172951936721802,
"learning_rate": 2.7359422722270782e-05,
"loss": 4.3333,
"num_input_tokens_seen": 56033280,
"step": 855
},
{
"epoch": 0.6923967191667086,
"grad_norm": 1.3179869651794434,
"learning_rate": 2.7279773578818943e-05,
"loss": 4.3657,
"num_input_tokens_seen": 56360960,
"step": 860
},
{
"epoch": 0.6964222814874452,
"grad_norm": 1.277402400970459,
"learning_rate": 2.720081603672184e-05,
"loss": 4.3581,
"num_input_tokens_seen": 56688640,
"step": 865
},
{
"epoch": 0.7004478438081819,
"grad_norm": 1.2060281038284302,
"learning_rate": 2.712254014483242e-05,
"loss": 4.372,
"num_input_tokens_seen": 57016320,
"step": 870
},
{
"epoch": 0.7044734061289186,
"grad_norm": 1.3300620317459106,
"learning_rate": 2.704493615131253e-05,
"loss": 4.3463,
"num_input_tokens_seen": 57344000,
"step": 875
},
{
"epoch": 0.7084989684496553,
"grad_norm": 1.2079811096191406,
"learning_rate": 2.6967994498529685e-05,
"loss": 4.3365,
"num_input_tokens_seen": 57671680,
"step": 880
},
{
"epoch": 0.712524530770392,
"grad_norm": 1.2995448112487793,
"learning_rate": 2.6891705818112575e-05,
"loss": 4.3474,
"num_input_tokens_seen": 57999360,
"step": 885
},
{
"epoch": 0.7165500930911287,
"grad_norm": 1.3099631071090698,
"learning_rate": 2.681606092615964e-05,
"loss": 4.3505,
"num_input_tokens_seen": 58327040,
"step": 890
},
{
"epoch": 0.7205756554118653,
"grad_norm": 1.1766512393951416,
"learning_rate": 2.6741050818595026e-05,
"loss": 4.3314,
"num_input_tokens_seen": 58654720,
"step": 895
},
{
"epoch": 0.724601217732602,
"grad_norm": 1.2225860357284546,
"learning_rate": 2.6666666666666667e-05,
"loss": 4.338,
"num_input_tokens_seen": 58982400,
"step": 900
},
{
"epoch": 0.724601217732602,
"eval_accuracy": 0.3663363114127162,
"eval_loss": 4.187371730804443,
"eval_runtime": 6.3051,
"eval_samples_per_second": 47.581,
"eval_steps_per_second": 6.027,
"num_input_tokens_seen": 58982400,
"step": 900
},
{
"epoch": 0.7286267800533387,
"grad_norm": 1.1610972881317139,
"learning_rate": 2.6592899812581347e-05,
"loss": 4.3445,
"num_input_tokens_seen": 59310080,
"step": 905
},
{
"epoch": 0.7326523423740754,
"grad_norm": 1.3581031560897827,
"learning_rate": 2.6519741765271837e-05,
"loss": 4.3591,
"num_input_tokens_seen": 59637760,
"step": 910
},
{
"epoch": 0.7366779046948121,
"grad_norm": 1.255743145942688,
"learning_rate": 2.6447184196291452e-05,
"loss": 4.342,
"num_input_tokens_seen": 59965440,
"step": 915
},
{
"epoch": 0.7407034670155488,
"grad_norm": 1.4238628149032593,
"learning_rate": 2.6375218935831487e-05,
"loss": 4.3019,
"num_input_tokens_seen": 60293120,
"step": 920
},
{
"epoch": 0.7447290293362854,
"grad_norm": 1.2544608116149902,
"learning_rate": 2.6303837968857168e-05,
"loss": 4.3089,
"num_input_tokens_seen": 60620800,
"step": 925
},
{
"epoch": 0.7487545916570221,
"grad_norm": 1.1759463548660278,
"learning_rate": 2.623303343135812e-05,
"loss": 4.3121,
"num_input_tokens_seen": 60948480,
"step": 930
},
{
"epoch": 0.7527801539777588,
"grad_norm": 1.3340330123901367,
"learning_rate": 2.616279760670915e-05,
"loss": 4.3074,
"num_input_tokens_seen": 61276160,
"step": 935
},
{
"epoch": 0.7568057162984955,
"grad_norm": 1.2744321823120117,
"learning_rate": 2.609312292213769e-05,
"loss": 4.3436,
"num_input_tokens_seen": 61603840,
"step": 940
},
{
"epoch": 0.7608312786192322,
"grad_norm": 1.1901994943618774,
"learning_rate": 2.602400194529422e-05,
"loss": 4.2908,
"num_input_tokens_seen": 61931520,
"step": 945
},
{
"epoch": 0.7648568409399688,
"grad_norm": 1.202046513557434,
"learning_rate": 2.595542738092201e-05,
"loss": 4.3347,
"num_input_tokens_seen": 62259200,
"step": 950
},
{
"epoch": 0.7688824032607054,
"grad_norm": 1.2437232732772827,
"learning_rate": 2.5887392067622943e-05,
"loss": 4.3434,
"num_input_tokens_seen": 62586880,
"step": 955
},
{
"epoch": 0.7729079655814421,
"grad_norm": 1.2383480072021484,
"learning_rate": 2.5819888974716114e-05,
"loss": 4.3144,
"num_input_tokens_seen": 62914560,
"step": 960
},
{
"epoch": 0.7769335279021788,
"grad_norm": 1.2294889688491821,
"learning_rate": 2.575291119918602e-05,
"loss": 4.2971,
"num_input_tokens_seen": 63242240,
"step": 965
},
{
"epoch": 0.7809590902229155,
"grad_norm": 1.235829472541809,
"learning_rate": 2.5686451962717426e-05,
"loss": 4.3006,
"num_input_tokens_seen": 63569920,
"step": 970
},
{
"epoch": 0.7849846525436522,
"grad_norm": 1.2483339309692383,
"learning_rate": 2.5620504608813944e-05,
"loss": 4.2846,
"num_input_tokens_seen": 63897600,
"step": 975
},
{
"epoch": 0.7890102148643888,
"grad_norm": 1.2668383121490479,
"learning_rate": 2.55550625999976e-05,
"loss": 4.3256,
"num_input_tokens_seen": 64225280,
"step": 980
},
{
"epoch": 0.7930357771851255,
"grad_norm": 1.2814356088638306,
"learning_rate": 2.549011951508665e-05,
"loss": 4.2869,
"num_input_tokens_seen": 64552960,
"step": 985
},
{
"epoch": 0.7970613395058622,
"grad_norm": 1.3956332206726074,
"learning_rate": 2.5425669046549127e-05,
"loss": 4.3438,
"num_input_tokens_seen": 64880640,
"step": 990
},
{
"epoch": 0.8010869018265989,
"grad_norm": 1.3514597415924072,
"learning_rate": 2.536170499792966e-05,
"loss": 4.3132,
"num_input_tokens_seen": 65208320,
"step": 995
},
{
"epoch": 0.8051124641473356,
"grad_norm": 1.198144793510437,
"learning_rate": 2.5298221281347037e-05,
"loss": 4.291,
"num_input_tokens_seen": 65536000,
"step": 1000
},
{
"epoch": 0.8091380264680723,
"grad_norm": 1.2150394916534424,
"learning_rate": 2.523521191506047e-05,
"loss": 4.3226,
"num_input_tokens_seen": 65863680,
"step": 1005
},
{
"epoch": 0.8131635887888089,
"grad_norm": 1.3794989585876465,
"learning_rate": 2.5172671021102106e-05,
"loss": 4.331,
"num_input_tokens_seen": 66191360,
"step": 1010
},
{
"epoch": 0.8171891511095456,
"grad_norm": 1.2272028923034668,
"learning_rate": 2.511059282297378e-05,
"loss": 4.3232,
"num_input_tokens_seen": 66519040,
"step": 1015
},
{
"epoch": 0.8212147134302823,
"grad_norm": 1.275801420211792,
"learning_rate": 2.5048971643405984e-05,
"loss": 4.3274,
"num_input_tokens_seen": 66846720,
"step": 1020
},
{
"epoch": 0.825240275751019,
"grad_norm": 1.3816701173782349,
"learning_rate": 2.4987801902176975e-05,
"loss": 4.3331,
"num_input_tokens_seen": 67174400,
"step": 1025
},
{
"epoch": 0.8292658380717557,
"grad_norm": 1.2978571653366089,
"learning_rate": 2.4927078113990235e-05,
"loss": 4.2978,
"num_input_tokens_seen": 67502080,
"step": 1030
},
{
"epoch": 0.8332914003924923,
"grad_norm": 1.3196457624435425,
"learning_rate": 2.486679488640837e-05,
"loss": 4.2904,
"num_input_tokens_seen": 67829760,
"step": 1035
},
{
"epoch": 0.837316962713229,
"grad_norm": 1.3695876598358154,
"learning_rate": 2.480694691784169e-05,
"loss": 4.3261,
"num_input_tokens_seen": 68157440,
"step": 1040
},
{
"epoch": 0.8413425250339657,
"grad_norm": 1.3481625318527222,
"learning_rate": 2.474752899558984e-05,
"loss": 4.2861,
"num_input_tokens_seen": 68485120,
"step": 1045
},
{
"epoch": 0.8453680873547024,
"grad_norm": 1.1765189170837402,
"learning_rate": 2.4688535993934706e-05,
"loss": 4.342,
"num_input_tokens_seen": 68812800,
"step": 1050
},
{
"epoch": 0.8453680873547024,
"eval_accuracy": 0.3695067117398011,
"eval_loss": 4.1475090980529785,
"eval_runtime": 6.6416,
"eval_samples_per_second": 45.17,
"eval_steps_per_second": 5.722,
"num_input_tokens_seen": 68812800,
"step": 1050
},
{
"epoch": 0.8493936496754391,
"grad_norm": 1.3343596458435059,
"learning_rate": 2.4629962872283168e-05,
"loss": 4.2779,
"num_input_tokens_seen": 69140480,
"step": 1055
},
{
"epoch": 0.8534192119961758,
"grad_norm": 1.31355619430542,
"learning_rate": 2.4571804673358052e-05,
"loss": 4.2672,
"num_input_tokens_seen": 69468160,
"step": 1060
},
{
"epoch": 0.8574447743169124,
"grad_norm": 1.2236064672470093,
"learning_rate": 2.4514056521435823e-05,
"loss": 4.3232,
"num_input_tokens_seen": 69795840,
"step": 1065
},
{
"epoch": 0.861470336637649,
"grad_norm": 1.2435745000839233,
"learning_rate": 2.4456713620629726e-05,
"loss": 4.3067,
"num_input_tokens_seen": 70123520,
"step": 1070
},
{
"epoch": 0.8654958989583857,
"grad_norm": 1.2577111721038818,
"learning_rate": 2.439977125321675e-05,
"loss": 4.3227,
"num_input_tokens_seen": 70451200,
"step": 1075
},
{
"epoch": 0.8695214612791224,
"grad_norm": 1.1730388402938843,
"learning_rate": 2.4343224778007382e-05,
"loss": 4.2894,
"num_input_tokens_seen": 70778880,
"step": 1080
},
{
"epoch": 0.8735470235998591,
"grad_norm": 1.2375553846359253,
"learning_rate": 2.4287069628756655e-05,
"loss": 4.3004,
"num_input_tokens_seen": 71106560,
"step": 1085
},
{
"epoch": 0.8775725859205958,
"grad_norm": 1.3142080307006836,
"learning_rate": 2.4231301312615306e-05,
"loss": 4.3036,
"num_input_tokens_seen": 71434240,
"step": 1090
},
{
"epoch": 0.8815981482413324,
"grad_norm": 1.218998670578003,
"learning_rate": 2.4175915408619977e-05,
"loss": 4.3097,
"num_input_tokens_seen": 71761920,
"step": 1095
},
{
"epoch": 0.8856237105620691,
"grad_norm": 1.312283992767334,
"learning_rate": 2.4120907566221092e-05,
"loss": 4.2724,
"num_input_tokens_seen": 72089600,
"step": 1100
},
{
"epoch": 0.8896492728828058,
"grad_norm": 1.273376703262329,
"learning_rate": 2.4066273503847476e-05,
"loss": 4.2896,
"num_input_tokens_seen": 72417280,
"step": 1105
},
{
"epoch": 0.8936748352035425,
"grad_norm": 1.2696881294250488,
"learning_rate": 2.4012009007506573e-05,
"loss": 4.3627,
"num_input_tokens_seen": 72744960,
"step": 1110
},
{
"epoch": 0.8977003975242792,
"grad_norm": 1.3189187049865723,
"learning_rate": 2.3958109929419195e-05,
"loss": 4.2899,
"num_input_tokens_seen": 73072640,
"step": 1115
},
{
"epoch": 0.9017259598450158,
"grad_norm": 1.1770976781845093,
"learning_rate": 2.3904572186687872e-05,
"loss": 4.2978,
"num_input_tokens_seen": 73400320,
"step": 1120
},
{
"epoch": 0.9057515221657525,
"grad_norm": 1.1813838481903076,
"learning_rate": 2.385139175999776e-05,
"loss": 4.2919,
"num_input_tokens_seen": 73728000,
"step": 1125
},
{
"epoch": 0.9097770844864892,
"grad_norm": 1.2608633041381836,
"learning_rate": 2.3798564692349184e-05,
"loss": 4.3118,
"num_input_tokens_seen": 74055680,
"step": 1130
},
{
"epoch": 0.9138026468072259,
"grad_norm": 1.3048661947250366,
"learning_rate": 2.3746087087820993e-05,
"loss": 4.286,
"num_input_tokens_seen": 74383360,
"step": 1135
},
{
"epoch": 0.9178282091279626,
"grad_norm": 1.1029573678970337,
"learning_rate": 2.3693955110363694e-05,
"loss": 4.2935,
"num_input_tokens_seen": 74711040,
"step": 1140
},
{
"epoch": 0.9218537714486993,
"grad_norm": 1.2217323780059814,
"learning_rate": 2.3642164982621672e-05,
"loss": 4.3049,
"num_input_tokens_seen": 75038720,
"step": 1145
},
{
"epoch": 0.9258793337694359,
"grad_norm": 1.2029818296432495,
"learning_rate": 2.3590712984783544e-05,
"loss": 4.3105,
"num_input_tokens_seen": 75366400,
"step": 1150
},
{
"epoch": 0.9299048960901726,
"grad_norm": 1.3691645860671997,
"learning_rate": 2.353959545345999e-05,
"loss": 4.2643,
"num_input_tokens_seen": 75694080,
"step": 1155
},
{
"epoch": 0.9339304584109093,
"grad_norm": 1.441178321838379,
"learning_rate": 2.3488808780588142e-05,
"loss": 4.2654,
"num_input_tokens_seen": 76021760,
"step": 1160
},
{
"epoch": 0.937956020731646,
"grad_norm": 1.2421964406967163,
"learning_rate": 2.3438349412361906e-05,
"loss": 4.3056,
"num_input_tokens_seen": 76349440,
"step": 1165
},
{
"epoch": 0.9419815830523827,
"grad_norm": 1.349487543106079,
"learning_rate": 2.338821384818745e-05,
"loss": 4.281,
"num_input_tokens_seen": 76677120,
"step": 1170
},
{
"epoch": 0.9460071453731194,
"grad_norm": 1.3027498722076416,
"learning_rate": 2.3338398639663132e-05,
"loss": 4.2827,
"num_input_tokens_seen": 77004800,
"step": 1175
},
{
"epoch": 0.950032707693856,
"grad_norm": 1.313684344291687,
"learning_rate": 2.3288900389583283e-05,
"loss": 4.3015,
"num_input_tokens_seen": 77332480,
"step": 1180
},
{
"epoch": 0.9540582700145926,
"grad_norm": 1.1965482234954834,
"learning_rate": 2.3239715750965074e-05,
"loss": 4.2841,
"num_input_tokens_seen": 77660160,
"step": 1185
},
{
"epoch": 0.9580838323353293,
"grad_norm": 1.2446861267089844,
"learning_rate": 2.3190841426097943e-05,
"loss": 4.3269,
"num_input_tokens_seen": 77987840,
"step": 1190
},
{
"epoch": 0.962109394656066,
"grad_norm": 1.218593955039978,
"learning_rate": 2.314227416561487e-05,
"loss": 4.3048,
"num_input_tokens_seen": 78315520,
"step": 1195
},
{
"epoch": 0.9661349569768027,
"grad_norm": 1.1979824304580688,
"learning_rate": 2.3094010767585035e-05,
"loss": 4.268,
"num_input_tokens_seen": 78643200,
"step": 1200
},
{
"epoch": 0.9661349569768027,
"eval_accuracy": 0.37141276761892367,
"eval_loss": 4.121544361114502,
"eval_runtime": 6.4083,
"eval_samples_per_second": 46.814,
"eval_steps_per_second": 5.93,
"num_input_tokens_seen": 78643200,
"step": 1200
},
{
"epoch": 0.9701605192975393,
"grad_norm": 1.2462401390075684,
"learning_rate": 2.3046048076627097e-05,
"loss": 4.2736,
"num_input_tokens_seen": 78970880,
"step": 1205
},
{
"epoch": 0.974186081618276,
"grad_norm": 1.3557130098342896,
"learning_rate": 2.2998382983042763e-05,
"loss": 4.2592,
"num_input_tokens_seen": 79298560,
"step": 1210
},
{
"epoch": 0.9782116439390127,
"grad_norm": 1.342148780822754,
"learning_rate": 2.2951012421969877e-05,
"loss": 4.2947,
"num_input_tokens_seen": 79626240,
"step": 1215
},
{
"epoch": 0.9822372062597494,
"grad_norm": 1.269107699394226,
"learning_rate": 2.290393337255473e-05,
"loss": 4.2789,
"num_input_tokens_seen": 79953920,
"step": 1220
},
{
"epoch": 0.9862627685804861,
"grad_norm": 1.189437747001648,
"learning_rate": 2.2857142857142858e-05,
"loss": 4.2793,
"num_input_tokens_seen": 80281600,
"step": 1225
},
{
"epoch": 0.9902883309012228,
"grad_norm": 1.217842698097229,
"learning_rate": 2.2810637940488043e-05,
"loss": 4.2606,
"num_input_tokens_seen": 80609280,
"step": 1230
},
{
"epoch": 0.9943138932219594,
"grad_norm": 1.1911648511886597,
"learning_rate": 2.2764415728978892e-05,
"loss": 4.2786,
"num_input_tokens_seen": 80936960,
"step": 1235
},
{
"epoch": 0.9983394555426961,
"grad_norm": 1.2993587255477905,
"learning_rate": 2.2718473369882594e-05,
"loss": 4.2791,
"num_input_tokens_seen": 81264640,
"step": 1240
},
{
"epoch": 1.002365017863433,
"grad_norm": 1.2911006212234497,
"learning_rate": 2.2672808050605434e-05,
"loss": 4.2619,
"num_input_tokens_seen": 81591296,
"step": 1245
},
{
"epoch": 1.0063905801841695,
"grad_norm": 1.4278241395950317,
"learning_rate": 2.2627416997969522e-05,
"loss": 4.2724,
"num_input_tokens_seen": 81918976,
"step": 1250
},
{
"epoch": 1.010416142504906,
"grad_norm": 1.3427342176437378,
"learning_rate": 2.2582297477505467e-05,
"loss": 4.2481,
"num_input_tokens_seen": 82246656,
"step": 1255
},
{
"epoch": 1.0144417048256429,
"grad_norm": 1.337169885635376,
"learning_rate": 2.253744679276044e-05,
"loss": 4.291,
"num_input_tokens_seen": 82574336,
"step": 1260
},
{
"epoch": 1.0184672671463795,
"grad_norm": 1.2553439140319824,
"learning_rate": 2.249286228462133e-05,
"loss": 4.2706,
"num_input_tokens_seen": 82902016,
"step": 1265
},
{
"epoch": 1.0224928294671163,
"grad_norm": 1.2947965860366821,
"learning_rate": 2.244854133065255e-05,
"loss": 4.21,
"num_input_tokens_seen": 83229696,
"step": 1270
},
{
"epoch": 1.0265183917878529,
"grad_norm": 1.327486515045166,
"learning_rate": 2.2404481344448157e-05,
"loss": 4.2847,
"num_input_tokens_seen": 83557376,
"step": 1275
},
{
"epoch": 1.0305439541085895,
"grad_norm": 1.2696667909622192,
"learning_rate": 2.23606797749979e-05,
"loss": 4.2864,
"num_input_tokens_seen": 83885056,
"step": 1280
},
{
"epoch": 1.0345695164293263,
"grad_norm": 1.3726356029510498,
"learning_rate": 2.2317134106066828e-05,
"loss": 4.265,
"num_input_tokens_seen": 84212736,
"step": 1285
},
{
"epoch": 1.0385950787500629,
"grad_norm": 1.250604271888733,
"learning_rate": 2.2273841855588186e-05,
"loss": 4.2894,
"num_input_tokens_seen": 84540416,
"step": 1290
},
{
"epoch": 1.0426206410707997,
"grad_norm": 1.3346271514892578,
"learning_rate": 2.223080057506914e-05,
"loss": 4.2751,
"num_input_tokens_seen": 84868096,
"step": 1295
},
{
"epoch": 1.0466462033915362,
"grad_norm": 1.272316575050354,
"learning_rate": 2.2188007849009167e-05,
"loss": 4.2783,
"num_input_tokens_seen": 85195776,
"step": 1300
},
{
"epoch": 1.0506717657122728,
"grad_norm": 1.37924063205719,
"learning_rate": 2.214546129433066e-05,
"loss": 4.2534,
"num_input_tokens_seen": 85523456,
"step": 1305
},
{
"epoch": 1.0546973280330096,
"grad_norm": 1.4074842929840088,
"learning_rate": 2.2103158559821507e-05,
"loss": 4.2809,
"num_input_tokens_seen": 85851136,
"step": 1310
},
{
"epoch": 1.0587228903537462,
"grad_norm": 1.3106921911239624,
"learning_rate": 2.206109732558935e-05,
"loss": 4.2831,
"num_input_tokens_seen": 86178816,
"step": 1315
},
{
"epoch": 1.062748452674483,
"grad_norm": 1.1723382472991943,
"learning_rate": 2.2019275302527215e-05,
"loss": 4.2617,
"num_input_tokens_seen": 86506496,
"step": 1320
},
{
"epoch": 1.0667740149952196,
"grad_norm": 1.2937371730804443,
"learning_rate": 2.1977690231790248e-05,
"loss": 4.2897,
"num_input_tokens_seen": 86834176,
"step": 1325
},
{
"epoch": 1.0707995773159564,
"grad_norm": 1.2420538663864136,
"learning_rate": 2.193633988428327e-05,
"loss": 4.2304,
"num_input_tokens_seen": 87161856,
"step": 1330
},
{
"epoch": 1.074825139636693,
"grad_norm": 1.2572234869003296,
"learning_rate": 2.1895222060158936e-05,
"loss": 4.2599,
"num_input_tokens_seen": 87489536,
"step": 1335
},
{
"epoch": 1.0788507019574296,
"grad_norm": 1.302157998085022,
"learning_rate": 2.1854334588326122e-05,
"loss": 4.2394,
"num_input_tokens_seen": 87817216,
"step": 1340
},
{
"epoch": 1.0828762642781664,
"grad_norm": 1.2681870460510254,
"learning_rate": 2.1813675325968476e-05,
"loss": 4.3021,
"num_input_tokens_seen": 88144896,
"step": 1345
},
{
"epoch": 1.086901826598903,
"grad_norm": 1.1405740976333618,
"learning_rate": 2.1773242158072697e-05,
"loss": 4.2185,
"num_input_tokens_seen": 88472576,
"step": 1350
},
{
"epoch": 1.086901826598903,
"eval_accuracy": 0.37249442300490127,
"eval_loss": 4.103240489959717,
"eval_runtime": 6.6226,
"eval_samples_per_second": 45.3,
"eval_steps_per_second": 5.738,
"num_input_tokens_seen": 88472576,
"step": 1350
},
{
"epoch": 1.0909273889196398,
"grad_norm": 1.144243836402893,
"learning_rate": 2.1733032996966454e-05,
"loss": 4.2665,
"num_input_tokens_seen": 88800256,
"step": 1355
},
{
"epoch": 1.0949529512403764,
"grad_norm": 1.158898115158081,
"learning_rate": 2.1693045781865617e-05,
"loss": 4.2608,
"num_input_tokens_seen": 89127936,
"step": 1360
},
{
"epoch": 1.098978513561113,
"grad_norm": 1.2325701713562012,
"learning_rate": 2.165327847843067e-05,
"loss": 4.2529,
"num_input_tokens_seen": 89455616,
"step": 1365
},
{
"epoch": 1.1030040758818498,
"grad_norm": 1.3270164728164673,
"learning_rate": 2.161372907833197e-05,
"loss": 4.2576,
"num_input_tokens_seen": 89783296,
"step": 1370
},
{
"epoch": 1.1070296382025864,
"grad_norm": 1.2680999040603638,
"learning_rate": 2.157439559882375e-05,
"loss": 4.2411,
"num_input_tokens_seen": 90110976,
"step": 1375
},
{
"epoch": 1.1110552005233232,
"grad_norm": 1.37815260887146,
"learning_rate": 2.1535276082326623e-05,
"loss": 4.3014,
"num_input_tokens_seen": 90438656,
"step": 1380
},
{
"epoch": 1.1150807628440598,
"grad_norm": 1.165544867515564,
"learning_rate": 2.149636859601836e-05,
"loss": 4.2405,
"num_input_tokens_seen": 90766336,
"step": 1385
},
{
"epoch": 1.1191063251647964,
"grad_norm": 1.184525966644287,
"learning_rate": 2.1457671231432803e-05,
"loss": 4.2646,
"num_input_tokens_seen": 91094016,
"step": 1390
},
{
"epoch": 1.1231318874855332,
"grad_norm": 1.216493010520935,
"learning_rate": 2.1419182104066638e-05,
"loss": 4.2586,
"num_input_tokens_seen": 91421696,
"step": 1395
},
{
"epoch": 1.1271574498062698,
"grad_norm": 1.4095522165298462,
"learning_rate": 2.1380899352993955e-05,
"loss": 4.2576,
"num_input_tokens_seen": 91749376,
"step": 1400
},
{
"epoch": 1.1311830121270066,
"grad_norm": 1.1573916673660278,
"learning_rate": 2.134282114048833e-05,
"loss": 4.2463,
"num_input_tokens_seen": 92077056,
"step": 1405
},
{
"epoch": 1.1352085744477431,
"grad_norm": 1.213179111480713,
"learning_rate": 2.13049456516523e-05,
"loss": 4.2402,
"num_input_tokens_seen": 92404736,
"step": 1410
},
{
"epoch": 1.1392341367684797,
"grad_norm": 1.197209119796753,
"learning_rate": 2.1267271094054026e-05,
"loss": 4.237,
"num_input_tokens_seen": 92732416,
"step": 1415
},
{
"epoch": 1.1432596990892165,
"grad_norm": 1.2352854013442993,
"learning_rate": 2.1229795697371012e-05,
"loss": 4.248,
"num_input_tokens_seen": 93060096,
"step": 1420
},
{
"epoch": 1.1472852614099531,
"grad_norm": 1.196413516998291,
"learning_rate": 2.1192517713040704e-05,
"loss": 4.272,
"num_input_tokens_seen": 93387776,
"step": 1425
},
{
"epoch": 1.15131082373069,
"grad_norm": 1.2881050109863281,
"learning_rate": 2.11554354139178e-05,
"loss": 4.258,
"num_input_tokens_seen": 93715456,
"step": 1430
},
{
"epoch": 1.1553363860514265,
"grad_norm": 1.2045120000839233,
"learning_rate": 2.1118547093938186e-05,
"loss": 4.2621,
"num_input_tokens_seen": 94043136,
"step": 1435
},
{
"epoch": 1.1593619483721633,
"grad_norm": 1.1793156862258911,
"learning_rate": 2.1081851067789197e-05,
"loss": 4.2386,
"num_input_tokens_seen": 94370816,
"step": 1440
},
{
"epoch": 1.1633875106929,
"grad_norm": 1.3340944051742554,
"learning_rate": 2.1045345670586257e-05,
"loss": 4.2837,
"num_input_tokens_seen": 94698496,
"step": 1445
},
{
"epoch": 1.1674130730136365,
"grad_norm": 1.2929213047027588,
"learning_rate": 2.100902925755561e-05,
"loss": 4.2334,
"num_input_tokens_seen": 95026176,
"step": 1450
},
{
"epoch": 1.1714386353343733,
"grad_norm": 1.3204522132873535,
"learning_rate": 2.0972900203723056e-05,
"loss": 4.2522,
"num_input_tokens_seen": 95353856,
"step": 1455
},
{
"epoch": 1.17546419765511,
"grad_norm": 1.305235743522644,
"learning_rate": 2.0936956903608547e-05,
"loss": 4.2534,
"num_input_tokens_seen": 95681536,
"step": 1460
},
{
"epoch": 1.1794897599758467,
"grad_norm": 1.2224024534225464,
"learning_rate": 2.0901197770926567e-05,
"loss": 4.2439,
"num_input_tokens_seen": 96009216,
"step": 1465
},
{
"epoch": 1.1835153222965833,
"grad_norm": 1.249813199043274,
"learning_rate": 2.0865621238292045e-05,
"loss": 4.2009,
"num_input_tokens_seen": 96336896,
"step": 1470
},
{
"epoch": 1.18754088461732,
"grad_norm": 1.2258191108703613,
"learning_rate": 2.0830225756931824e-05,
"loss": 4.2517,
"num_input_tokens_seen": 96664576,
"step": 1475
},
{
"epoch": 1.1915664469380567,
"grad_norm": 1.2485374212265015,
"learning_rate": 2.0795009796401456e-05,
"loss": 4.2254,
"num_input_tokens_seen": 96992256,
"step": 1480
},
{
"epoch": 1.1955920092587933,
"grad_norm": 1.2434406280517578,
"learning_rate": 2.0759971844307282e-05,
"loss": 4.208,
"num_input_tokens_seen": 97319936,
"step": 1485
},
{
"epoch": 1.19961757157953,
"grad_norm": 1.212769865989685,
"learning_rate": 2.072511040603359e-05,
"loss": 4.2253,
"num_input_tokens_seen": 97647616,
"step": 1490
},
{
"epoch": 1.2036431339002667,
"grad_norm": 1.1651945114135742,
"learning_rate": 2.0690424004474868e-05,
"loss": 4.2694,
"num_input_tokens_seen": 97975296,
"step": 1495
},
{
"epoch": 1.2076686962210035,
"grad_norm": 1.2737419605255127,
"learning_rate": 2.065591117977289e-05,
"loss": 4.2645,
"num_input_tokens_seen": 98302976,
"step": 1500
},
{
"epoch": 1.2076686962210035,
"eval_accuracy": 0.37566890863567187,
"eval_loss": 4.08589506149292,
"eval_runtime": 6.6303,
"eval_samples_per_second": 45.247,
"eval_steps_per_second": 5.731,
"num_input_tokens_seen": 98302976,
"step": 1500
},
{
"epoch": 1.21169425854174,
"grad_norm": 1.2633168697357178,
"learning_rate": 2.0621570489058684e-05,
"loss": 4.2746,
"num_input_tokens_seen": 98630656,
"step": 1505
},
{
"epoch": 1.2157198208624767,
"grad_norm": 1.2094192504882812,
"learning_rate": 2.0587400506199153e-05,
"loss": 4.2431,
"num_input_tokens_seen": 98958336,
"step": 1510
},
{
"epoch": 1.2197453831832135,
"grad_norm": 1.2098276615142822,
"learning_rate": 2.0553399821548317e-05,
"loss": 4.2635,
"num_input_tokens_seen": 99286016,
"step": 1515
},
{
"epoch": 1.22377094550395,
"grad_norm": 1.1902521848678589,
"learning_rate": 2.0519567041703087e-05,
"loss": 4.283,
"num_input_tokens_seen": 99613696,
"step": 1520
},
{
"epoch": 1.2277965078246869,
"grad_norm": 1.2712699174880981,
"learning_rate": 2.048590078926336e-05,
"loss": 4.2737,
"num_input_tokens_seen": 99941376,
"step": 1525
},
{
"epoch": 1.2318220701454234,
"grad_norm": 1.2726564407348633,
"learning_rate": 2.0452399702596545e-05,
"loss": 4.2715,
"num_input_tokens_seen": 100269056,
"step": 1530
},
{
"epoch": 1.23584763246616,
"grad_norm": 1.2632936239242554,
"learning_rate": 2.0419062435606238e-05,
"loss": 4.2266,
"num_input_tokens_seen": 100596736,
"step": 1535
},
{
"epoch": 1.2398731947868968,
"grad_norm": 1.470897912979126,
"learning_rate": 2.038588765750502e-05,
"loss": 4.235,
"num_input_tokens_seen": 100924416,
"step": 1540
},
{
"epoch": 1.2438987571076334,
"grad_norm": 1.3083152770996094,
"learning_rate": 2.035287405259138e-05,
"loss": 4.233,
"num_input_tokens_seen": 101252096,
"step": 1545
},
{
"epoch": 1.2479243194283702,
"grad_norm": 1.2998907566070557,
"learning_rate": 2.032002032003048e-05,
"loss": 4.22,
"num_input_tokens_seen": 101579776,
"step": 1550
},
{
"epoch": 1.2519498817491068,
"grad_norm": 1.3198977708816528,
"learning_rate": 2.0287325173638952e-05,
"loss": 4.236,
"num_input_tokens_seen": 101907456,
"step": 1555
},
{
"epoch": 1.2559754440698434,
"grad_norm": 1.3605294227600098,
"learning_rate": 2.0254787341673334e-05,
"loss": 4.2318,
"num_input_tokens_seen": 102235136,
"step": 1560
},
{
"epoch": 1.2600010063905802,
"grad_norm": 1.2605648040771484,
"learning_rate": 2.0222405566622346e-05,
"loss": 4.232,
"num_input_tokens_seen": 102562816,
"step": 1565
},
{
"epoch": 1.2640265687113168,
"grad_norm": 1.2441262006759644,
"learning_rate": 2.019017860500275e-05,
"loss": 4.1917,
"num_input_tokens_seen": 102890496,
"step": 1570
},
{
"epoch": 1.2680521310320536,
"grad_norm": 1.3246742486953735,
"learning_rate": 2.0158105227158783e-05,
"loss": 4.2272,
"num_input_tokens_seen": 103218176,
"step": 1575
},
{
"epoch": 1.2720776933527902,
"grad_norm": 1.2322101593017578,
"learning_rate": 2.0126184217065105e-05,
"loss": 4.1989,
"num_input_tokens_seen": 103545856,
"step": 1580
},
{
"epoch": 1.2761032556735268,
"grad_norm": 1.2290526628494263,
"learning_rate": 2.0094414372133134e-05,
"loss": 4.2389,
"num_input_tokens_seen": 103873536,
"step": 1585
},
{
"epoch": 1.2801288179942636,
"grad_norm": 1.3795405626296997,
"learning_rate": 2.006279450302077e-05,
"loss": 4.1988,
"num_input_tokens_seen": 104201216,
"step": 1590
},
{
"epoch": 1.2841543803150002,
"grad_norm": 1.3456906080245972,
"learning_rate": 2.003132343344538e-05,
"loss": 4.2586,
"num_input_tokens_seen": 104528896,
"step": 1595
},
{
"epoch": 1.288179942635737,
"grad_norm": 1.2667040824890137,
"learning_rate": 2e-05,
"loss": 4.2505,
"num_input_tokens_seen": 104856576,
"step": 1600
},
{
"epoch": 1.2922055049564736,
"grad_norm": 1.2646143436431885,
"learning_rate": 1.9968823051972716e-05,
"loss": 4.2537,
"num_input_tokens_seen": 105184256,
"step": 1605
},
{
"epoch": 1.2962310672772102,
"grad_norm": 1.2952028512954712,
"learning_rate": 1.9937791451169073e-05,
"loss": 4.2439,
"num_input_tokens_seen": 105511936,
"step": 1610
},
{
"epoch": 1.300256629597947,
"grad_norm": 1.346574068069458,
"learning_rate": 1.9906904071737584e-05,
"loss": 4.24,
"num_input_tokens_seen": 105839616,
"step": 1615
},
{
"epoch": 1.3042821919186838,
"grad_norm": 1.2765649557113647,
"learning_rate": 1.9876159799998135e-05,
"loss": 4.2229,
"num_input_tokens_seen": 106167296,
"step": 1620
},
{
"epoch": 1.3083077542394204,
"grad_norm": 1.3729643821716309,
"learning_rate": 1.9845557534273358e-05,
"loss": 4.201,
"num_input_tokens_seen": 106494976,
"step": 1625
},
{
"epoch": 1.312333316560157,
"grad_norm": 1.2387956380844116,
"learning_rate": 1.98150961847228e-05,
"loss": 4.2235,
"num_input_tokens_seen": 106822656,
"step": 1630
},
{
"epoch": 1.3163588788808938,
"grad_norm": 1.283973217010498,
"learning_rate": 1.978477467317992e-05,
"loss": 4.2616,
"num_input_tokens_seen": 107150336,
"step": 1635
},
{
"epoch": 1.3203844412016303,
"grad_norm": 1.3440768718719482,
"learning_rate": 1.9754591932991796e-05,
"loss": 4.2215,
"num_input_tokens_seen": 107478016,
"step": 1640
},
{
"epoch": 1.3244100035223672,
"grad_norm": 1.292891263961792,
"learning_rate": 1.9724546908861517e-05,
"loss": 4.1927,
"num_input_tokens_seen": 107805696,
"step": 1645
},
{
"epoch": 1.3284355658431037,
"grad_norm": 1.3360971212387085,
"learning_rate": 1.9694638556693238e-05,
"loss": 4.2542,
"num_input_tokens_seen": 108133376,
"step": 1650
},
{
"epoch": 1.3284355658431037,
"eval_accuracy": 0.37502241786907964,
"eval_loss": 4.072964668273926,
"eval_runtime": 6.6283,
"eval_samples_per_second": 45.26,
"eval_steps_per_second": 5.733,
"num_input_tokens_seen": 108133376,
"step": 1650
},
{
"epoch": 1.3324611281638403,
"grad_norm": 1.2728638648986816,
"learning_rate": 1.9664865843439752e-05,
"loss": 4.2287,
"num_input_tokens_seen": 108461056,
"step": 1655
},
{
"epoch": 1.3364866904845771,
"grad_norm": 1.242002248764038,
"learning_rate": 1.9635227746952642e-05,
"loss": 4.2355,
"num_input_tokens_seen": 108788736,
"step": 1660
},
{
"epoch": 1.3405122528053137,
"grad_norm": 1.2821120023727417,
"learning_rate": 1.9605723255834874e-05,
"loss": 4.2256,
"num_input_tokens_seen": 109116416,
"step": 1665
},
{
"epoch": 1.3445378151260505,
"grad_norm": 1.380469560623169,
"learning_rate": 1.9576351369295855e-05,
"loss": 4.2284,
"num_input_tokens_seen": 109444096,
"step": 1670
},
{
"epoch": 1.3485633774467871,
"grad_norm": 1.241410493850708,
"learning_rate": 1.9547111097008838e-05,
"loss": 4.2122,
"num_input_tokens_seen": 109771776,
"step": 1675
},
{
"epoch": 1.3525889397675237,
"grad_norm": 1.2036563158035278,
"learning_rate": 1.9518001458970662e-05,
"loss": 4.2312,
"num_input_tokens_seen": 110099456,
"step": 1680
},
{
"epoch": 1.3566145020882605,
"grad_norm": 1.3382368087768555,
"learning_rate": 1.9489021485363838e-05,
"loss": 4.2168,
"num_input_tokens_seen": 110427136,
"step": 1685
},
{
"epoch": 1.360640064408997,
"grad_norm": 1.2150115966796875,
"learning_rate": 1.9460170216420796e-05,
"loss": 4.2322,
"num_input_tokens_seen": 110754816,
"step": 1690
},
{
"epoch": 1.364665626729734,
"grad_norm": 1.2400802373886108,
"learning_rate": 1.9431446702290413e-05,
"loss": 4.2018,
"num_input_tokens_seen": 111082496,
"step": 1695
},
{
"epoch": 1.3686911890504705,
"grad_norm": 1.3871190547943115,
"learning_rate": 1.9402850002906638e-05,
"loss": 4.2346,
"num_input_tokens_seen": 111410176,
"step": 1700
},
{
"epoch": 1.372716751371207,
"grad_norm": 1.2957957983016968,
"learning_rate": 1.9374379187859312e-05,
"loss": 4.2078,
"num_input_tokens_seen": 111737856,
"step": 1705
},
{
"epoch": 1.3767423136919439,
"grad_norm": 1.2237967252731323,
"learning_rate": 1.934603333626698e-05,
"loss": 4.2472,
"num_input_tokens_seen": 112065536,
"step": 1710
},
{
"epoch": 1.3807678760126805,
"grad_norm": 1.2709431648254395,
"learning_rate": 1.9317811536651808e-05,
"loss": 4.2679,
"num_input_tokens_seen": 112393216,
"step": 1715
},
{
"epoch": 1.3847934383334173,
"grad_norm": 1.3072236776351929,
"learning_rate": 1.9289712886816488e-05,
"loss": 4.2672,
"num_input_tokens_seen": 112720896,
"step": 1720
},
{
"epoch": 1.3888190006541539,
"grad_norm": 1.2216880321502686,
"learning_rate": 1.9261736493723075e-05,
"loss": 4.2124,
"num_input_tokens_seen": 113048576,
"step": 1725
},
{
"epoch": 1.3928445629748905,
"grad_norm": 1.2216253280639648,
"learning_rate": 1.9233881473373802e-05,
"loss": 4.2168,
"num_input_tokens_seen": 113376256,
"step": 1730
},
{
"epoch": 1.3968701252956273,
"grad_norm": 1.15086030960083,
"learning_rate": 1.9206146950693745e-05,
"loss": 4.2097,
"num_input_tokens_seen": 113703936,
"step": 1735
},
{
"epoch": 1.4008956876163638,
"grad_norm": 1.2962745428085327,
"learning_rate": 1.917853205941537e-05,
"loss": 4.2328,
"num_input_tokens_seen": 114031616,
"step": 1740
},
{
"epoch": 1.4049212499371007,
"grad_norm": 1.3710294961929321,
"learning_rate": 1.915103594196486e-05,
"loss": 4.2205,
"num_input_tokens_seen": 114359296,
"step": 1745
},
{
"epoch": 1.4089468122578372,
"grad_norm": 1.290256142616272,
"learning_rate": 1.91236577493503e-05,
"loss": 4.2239,
"num_input_tokens_seen": 114686976,
"step": 1750
},
{
"epoch": 1.4129723745785738,
"grad_norm": 1.1622158288955688,
"learning_rate": 1.9096396641051548e-05,
"loss": 4.2077,
"num_input_tokens_seen": 115014656,
"step": 1755
},
{
"epoch": 1.4169979368993106,
"grad_norm": 1.2251081466674805,
"learning_rate": 1.9069251784911845e-05,
"loss": 4.2479,
"num_input_tokens_seen": 115342336,
"step": 1760
},
{
"epoch": 1.4210234992200472,
"grad_norm": 1.2181967496871948,
"learning_rate": 1.9042222357031166e-05,
"loss": 4.2252,
"num_input_tokens_seen": 115670016,
"step": 1765
},
{
"epoch": 1.425049061540784,
"grad_norm": 1.3037039041519165,
"learning_rate": 1.9015307541661134e-05,
"loss": 4.22,
"num_input_tokens_seen": 115997696,
"step": 1770
},
{
"epoch": 1.4290746238615206,
"grad_norm": 1.2290278673171997,
"learning_rate": 1.8988506531101655e-05,
"loss": 4.22,
"num_input_tokens_seen": 116325376,
"step": 1775
},
{
"epoch": 1.4331001861822572,
"grad_norm": 1.287455677986145,
"learning_rate": 1.8961818525599093e-05,
"loss": 4.2533,
"num_input_tokens_seen": 116653056,
"step": 1780
},
{
"epoch": 1.437125748502994,
"grad_norm": 1.1899845600128174,
"learning_rate": 1.8935242733246034e-05,
"loss": 4.2289,
"num_input_tokens_seen": 116980736,
"step": 1785
},
{
"epoch": 1.4411513108237308,
"grad_norm": 1.2166035175323486,
"learning_rate": 1.8908778369882623e-05,
"loss": 4.218,
"num_input_tokens_seen": 117308416,
"step": 1790
},
{
"epoch": 1.4451768731444674,
"grad_norm": 1.2712666988372803,
"learning_rate": 1.888242465899932e-05,
"loss": 4.1944,
"num_input_tokens_seen": 117636096,
"step": 1795
},
{
"epoch": 1.449202435465204,
"grad_norm": 1.1924896240234375,
"learning_rate": 1.885618083164127e-05,
"loss": 4.2614,
"num_input_tokens_seen": 117963776,
"step": 1800
},
{
"epoch": 1.449202435465204,
"eval_accuracy": 0.37485385814497274,
"eval_loss": 4.068163871765137,
"eval_runtime": 7.4672,
"eval_samples_per_second": 40.176,
"eval_steps_per_second": 5.089,
"num_input_tokens_seen": 117963776,
"step": 1800
},
{
"epoch": 1.4532279977859408,
"grad_norm": 1.2685041427612305,
"learning_rate": 1.883004612631402e-05,
"loss": 4.236,
"num_input_tokens_seen": 118291456,
"step": 1805
},
{
"epoch": 1.4572535601066774,
"grad_norm": 1.1977604627609253,
"learning_rate": 1.880401978889074e-05,
"loss": 4.258,
"num_input_tokens_seen": 118619136,
"step": 1810
},
{
"epoch": 1.4612791224274142,
"grad_norm": 1.337408423423767,
"learning_rate": 1.877810107252081e-05,
"loss": 4.2549,
"num_input_tokens_seen": 118946816,
"step": 1815
},
{
"epoch": 1.4653046847481508,
"grad_norm": 1.348046898841858,
"learning_rate": 1.875228923753982e-05,
"loss": 4.2513,
"num_input_tokens_seen": 119274496,
"step": 1820
},
{
"epoch": 1.4693302470688874,
"grad_norm": 1.2587336301803589,
"learning_rate": 1.8726583551380893e-05,
"loss": 4.2251,
"num_input_tokens_seen": 119602176,
"step": 1825
},
{
"epoch": 1.4733558093896242,
"grad_norm": 1.3538405895233154,
"learning_rate": 1.8700983288487377e-05,
"loss": 4.2219,
"num_input_tokens_seen": 119929856,
"step": 1830
},
{
"epoch": 1.4773813717103608,
"grad_norm": 1.2679195404052734,
"learning_rate": 1.8675487730226835e-05,
"loss": 4.2026,
"num_input_tokens_seen": 120257536,
"step": 1835
},
{
"epoch": 1.4814069340310976,
"grad_norm": 1.2712976932525635,
"learning_rate": 1.8650096164806278e-05,
"loss": 4.2277,
"num_input_tokens_seen": 120585216,
"step": 1840
},
{
"epoch": 1.4854324963518342,
"grad_norm": 1.2575690746307373,
"learning_rate": 1.862480788718875e-05,
"loss": 4.2329,
"num_input_tokens_seen": 120912896,
"step": 1845
},
{
"epoch": 1.4894580586725708,
"grad_norm": 1.2137782573699951,
"learning_rate": 1.8599622199011086e-05,
"loss": 4.2076,
"num_input_tokens_seen": 121240576,
"step": 1850
},
{
"epoch": 1.4934836209933076,
"grad_norm": 1.2346251010894775,
"learning_rate": 1.8574538408502883e-05,
"loss": 4.2357,
"num_input_tokens_seen": 121568256,
"step": 1855
},
{
"epoch": 1.4975091833140441,
"grad_norm": 1.2532144784927368,
"learning_rate": 1.854955583040673e-05,
"loss": 4.2616,
"num_input_tokens_seen": 121895936,
"step": 1860
},
{
"epoch": 1.501534745634781,
"grad_norm": 1.1912769079208374,
"learning_rate": 1.8524673785899573e-05,
"loss": 4.1788,
"num_input_tokens_seen": 122223616,
"step": 1865
},
{
"epoch": 1.5055603079555175,
"grad_norm": 1.2742645740509033,
"learning_rate": 1.849989160251521e-05,
"loss": 4.2119,
"num_input_tokens_seen": 122551296,
"step": 1870
},
{
"epoch": 1.5095858702762541,
"grad_norm": 1.264236569404602,
"learning_rate": 1.847520861406802e-05,
"loss": 4.2219,
"num_input_tokens_seen": 122878976,
"step": 1875
},
{
"epoch": 1.513611432596991,
"grad_norm": 1.2975112199783325,
"learning_rate": 1.8450624160577702e-05,
"loss": 4.2227,
"num_input_tokens_seen": 123206656,
"step": 1880
},
{
"epoch": 1.5176369949177275,
"grad_norm": 1.231823205947876,
"learning_rate": 1.842613758819515e-05,
"loss": 4.2185,
"num_input_tokens_seen": 123534336,
"step": 1885
},
{
"epoch": 1.5216625572384643,
"grad_norm": 1.3271822929382324,
"learning_rate": 1.8401748249129447e-05,
"loss": 4.2174,
"num_input_tokens_seen": 123862016,
"step": 1890
},
{
"epoch": 1.525688119559201,
"grad_norm": 1.2804478406906128,
"learning_rate": 1.8377455501575864e-05,
"loss": 4.1906,
"num_input_tokens_seen": 124189696,
"step": 1895
},
{
"epoch": 1.5297136818799375,
"grad_norm": 1.2299883365631104,
"learning_rate": 1.835325870964494e-05,
"loss": 4.1957,
"num_input_tokens_seen": 124517376,
"step": 1900
},
{
"epoch": 1.5337392442006743,
"grad_norm": 1.2697046995162964,
"learning_rate": 1.8329157243292555e-05,
"loss": 4.2115,
"num_input_tokens_seen": 124845056,
"step": 1905
},
{
"epoch": 1.5377648065214111,
"grad_norm": 1.20834219455719,
"learning_rate": 1.8305150478251023e-05,
"loss": 4.2224,
"num_input_tokens_seen": 125172736,
"step": 1910
},
{
"epoch": 1.5417903688421477,
"grad_norm": 1.2465049028396606,
"learning_rate": 1.8281237795961206e-05,
"loss": 4.1757,
"num_input_tokens_seen": 125500416,
"step": 1915
},
{
"epoch": 1.5458159311628843,
"grad_norm": 1.2020412683486938,
"learning_rate": 1.825741858350554e-05,
"loss": 4.2047,
"num_input_tokens_seen": 125828096,
"step": 1920
},
{
"epoch": 1.5498414934836209,
"grad_norm": 1.384621024131775,
"learning_rate": 1.8233692233542074e-05,
"loss": 4.2553,
"num_input_tokens_seen": 126155776,
"step": 1925
},
{
"epoch": 1.5538670558043577,
"grad_norm": 1.212430477142334,
"learning_rate": 1.8210058144239417e-05,
"loss": 4.1952,
"num_input_tokens_seen": 126483456,
"step": 1930
},
{
"epoch": 1.5578926181250945,
"grad_norm": 1.2624080181121826,
"learning_rate": 1.8186515719212627e-05,
"loss": 4.2208,
"num_input_tokens_seen": 126811136,
"step": 1935
},
{
"epoch": 1.561918180445831,
"grad_norm": 1.2795370817184448,
"learning_rate": 1.8163064367459993e-05,
"loss": 4.1835,
"num_input_tokens_seen": 127138816,
"step": 1940
},
{
"epoch": 1.5659437427665677,
"grad_norm": 1.3003804683685303,
"learning_rate": 1.813970350330073e-05,
"loss": 4.2235,
"num_input_tokens_seen": 127466496,
"step": 1945
},
{
"epoch": 1.5699693050873043,
"grad_norm": 1.328814148902893,
"learning_rate": 1.8116432546313533e-05,
"loss": 4.1928,
"num_input_tokens_seen": 127794176,
"step": 1950
},
{
"epoch": 1.5699693050873043,
"eval_accuracy": 0.37578634243079373,
"eval_loss": 4.0595831871032715,
"eval_runtime": 7.0878,
"eval_samples_per_second": 42.326,
"eval_steps_per_second": 5.361,
"num_input_tokens_seen": 127794176,
"step": 1950
},
{
"epoch": 1.573994867408041,
"grad_norm": 1.258589506149292,
"learning_rate": 1.8093250921276027e-05,
"loss": 4.2336,
"num_input_tokens_seen": 128121856,
"step": 1955
},
{
"epoch": 1.5780204297287779,
"grad_norm": 1.226361632347107,
"learning_rate": 1.8070158058105027e-05,
"loss": 4.1945,
"num_input_tokens_seen": 128449536,
"step": 1960
},
{
"epoch": 1.5820459920495145,
"grad_norm": 1.2249149084091187,
"learning_rate": 1.8047153391797662e-05,
"loss": 4.1916,
"num_input_tokens_seen": 128777216,
"step": 1965
},
{
"epoch": 1.586071554370251,
"grad_norm": 1.3121646642684937,
"learning_rate": 1.8024236362373317e-05,
"loss": 4.1833,
"num_input_tokens_seen": 129104896,
"step": 1970
},
{
"epoch": 1.5900971166909876,
"grad_norm": 1.2635945081710815,
"learning_rate": 1.8001406414816385e-05,
"loss": 4.208,
"num_input_tokens_seen": 129432576,
"step": 1975
},
{
"epoch": 1.5941226790117244,
"grad_norm": 1.2306846380233765,
"learning_rate": 1.797866299901979e-05,
"loss": 4.1832,
"num_input_tokens_seen": 129760256,
"step": 1980
},
{
"epoch": 1.5981482413324613,
"grad_norm": 1.2278416156768799,
"learning_rate": 1.7956005569729342e-05,
"loss": 4.1765,
"num_input_tokens_seen": 130087936,
"step": 1985
},
{
"epoch": 1.6021738036531978,
"grad_norm": 1.2951785326004028,
"learning_rate": 1.7933433586488815e-05,
"loss": 4.2123,
"num_input_tokens_seen": 130415616,
"step": 1990
},
{
"epoch": 1.6061993659739344,
"grad_norm": 1.341288447380066,
"learning_rate": 1.79109465135858e-05,
"loss": 4.1865,
"num_input_tokens_seen": 130743296,
"step": 1995
},
{
"epoch": 1.610224928294671,
"grad_norm": 1.269637942314148,
"learning_rate": 1.788854381999832e-05,
"loss": 4.2004,
"num_input_tokens_seen": 131070976,
"step": 2000
},
{
"epoch": 1.6142504906154078,
"grad_norm": 1.3138514757156372,
"learning_rate": 1.7866224979342134e-05,
"loss": 4.2172,
"num_input_tokens_seen": 131398656,
"step": 2005
},
{
"epoch": 1.6182760529361446,
"grad_norm": 1.1714235544204712,
"learning_rate": 1.7843989469818822e-05,
"loss": 4.1903,
"num_input_tokens_seen": 131726336,
"step": 2010
},
{
"epoch": 1.6223016152568812,
"grad_norm": 1.3786474466323853,
"learning_rate": 1.7821836774164523e-05,
"loss": 4.2,
"num_input_tokens_seen": 132054016,
"step": 2015
},
{
"epoch": 1.6263271775776178,
"grad_norm": 1.3901410102844238,
"learning_rate": 1.7799766379599393e-05,
"loss": 4.1928,
"num_input_tokens_seen": 132381696,
"step": 2020
},
{
"epoch": 1.6303527398983544,
"grad_norm": 1.3432561159133911,
"learning_rate": 1.7777777777777777e-05,
"loss": 4.1983,
"num_input_tokens_seen": 132709376,
"step": 2025
},
{
"epoch": 1.6343783022190912,
"grad_norm": 1.2629408836364746,
"learning_rate": 1.7755870464739014e-05,
"loss": 4.2136,
"num_input_tokens_seen": 133037056,
"step": 2030
},
{
"epoch": 1.638403864539828,
"grad_norm": 1.2436974048614502,
"learning_rate": 1.7734043940858906e-05,
"loss": 4.1695,
"num_input_tokens_seen": 133364736,
"step": 2035
},
{
"epoch": 1.6424294268605646,
"grad_norm": 1.2254369258880615,
"learning_rate": 1.771229771080191e-05,
"loss": 4.1636,
"num_input_tokens_seen": 133692416,
"step": 2040
},
{
"epoch": 1.6464549891813012,
"grad_norm": 1.3272947072982788,
"learning_rate": 1.769063128347386e-05,
"loss": 4.1979,
"num_input_tokens_seen": 134020096,
"step": 2045
},
{
"epoch": 1.650480551502038,
"grad_norm": 1.2953760623931885,
"learning_rate": 1.7669044171975447e-05,
"loss": 4.2026,
"num_input_tokens_seen": 134347776,
"step": 2050
},
{
"epoch": 1.6545061138227746,
"grad_norm": 1.2235223054885864,
"learning_rate": 1.764753589355622e-05,
"loss": 4.1674,
"num_input_tokens_seen": 134675456,
"step": 2055
},
{
"epoch": 1.6585316761435114,
"grad_norm": 1.1912940740585327,
"learning_rate": 1.762610596956927e-05,
"loss": 4.1811,
"num_input_tokens_seen": 135003136,
"step": 2060
},
{
"epoch": 1.662557238464248,
"grad_norm": 1.3255517482757568,
"learning_rate": 1.7604753925426462e-05,
"loss": 4.2094,
"num_input_tokens_seen": 135330816,
"step": 2065
},
{
"epoch": 1.6665828007849846,
"grad_norm": 1.2693239450454712,
"learning_rate": 1.758347929055432e-05,
"loss": 4.2113,
"num_input_tokens_seen": 135658496,
"step": 2070
},
{
"epoch": 1.6706083631057214,
"grad_norm": 1.2487210035324097,
"learning_rate": 1.7562281598350458e-05,
"loss": 4.2341,
"num_input_tokens_seen": 135986176,
"step": 2075
},
{
"epoch": 1.6746339254264582,
"grad_norm": 1.3119174242019653,
"learning_rate": 1.7541160386140587e-05,
"loss": 4.1846,
"num_input_tokens_seen": 136313856,
"step": 2080
},
{
"epoch": 1.6786594877471948,
"grad_norm": 1.2278319597244263,
"learning_rate": 1.7520115195136116e-05,
"loss": 4.1637,
"num_input_tokens_seen": 136641536,
"step": 2085
},
{
"epoch": 1.6826850500679313,
"grad_norm": 1.2323991060256958,
"learning_rate": 1.7499145570392288e-05,
"loss": 4.1943,
"num_input_tokens_seen": 136969216,
"step": 2090
},
{
"epoch": 1.686710612388668,
"grad_norm": 1.3337785005569458,
"learning_rate": 1.747825106076687e-05,
"loss": 4.2297,
"num_input_tokens_seen": 137296896,
"step": 2095
},
{
"epoch": 1.6907361747094047,
"grad_norm": 1.1806713342666626,
"learning_rate": 1.7457431218879393e-05,
"loss": 4.1971,
"num_input_tokens_seen": 137624576,
"step": 2100
},
{
"epoch": 1.6907361747094047,
"eval_accuracy": 0.3776899715899814,
"eval_loss": 4.050466537475586,
"eval_runtime": 6.6452,
"eval_samples_per_second": 45.145,
"eval_steps_per_second": 5.718,
"num_input_tokens_seen": 137624576,
"step": 2100
},
{
"epoch": 1.6947617370301415,
"grad_norm": 1.1816240549087524,
"learning_rate": 1.7436685601070914e-05,
"loss": 4.1944,
"num_input_tokens_seen": 137952256,
"step": 2105
},
{
"epoch": 1.6987872993508781,
"grad_norm": 1.258262276649475,
"learning_rate": 1.7416013767364324e-05,
"loss": 4.1668,
"num_input_tokens_seen": 138279936,
"step": 2110
},
{
"epoch": 1.7028128616716147,
"grad_norm": 1.2995336055755615,
"learning_rate": 1.7395415281425125e-05,
"loss": 4.2091,
"num_input_tokens_seen": 138607616,
"step": 2115
},
{
"epoch": 1.7068384239923513,
"grad_norm": 1.341255784034729,
"learning_rate": 1.7374889710522778e-05,
"loss": 4.1956,
"num_input_tokens_seen": 138935296,
"step": 2120
},
{
"epoch": 1.7108639863130881,
"grad_norm": 1.3196258544921875,
"learning_rate": 1.7354436625492496e-05,
"loss": 4.1887,
"num_input_tokens_seen": 139262976,
"step": 2125
},
{
"epoch": 1.714889548633825,
"grad_norm": 1.3612850904464722,
"learning_rate": 1.7334055600697583e-05,
"loss": 4.1805,
"num_input_tokens_seen": 139590656,
"step": 2130
},
{
"epoch": 1.7189151109545615,
"grad_norm": 1.259021282196045,
"learning_rate": 1.73137462139922e-05,
"loss": 4.2013,
"num_input_tokens_seen": 139918336,
"step": 2135
},
{
"epoch": 1.722940673275298,
"grad_norm": 1.226407766342163,
"learning_rate": 1.729350804668468e-05,
"loss": 4.1963,
"num_input_tokens_seen": 140246016,
"step": 2140
},
{
"epoch": 1.7269662355960347,
"grad_norm": 1.278818130493164,
"learning_rate": 1.727334068350122e-05,
"loss": 4.1719,
"num_input_tokens_seen": 140573696,
"step": 2145
},
{
"epoch": 1.7309917979167715,
"grad_norm": 1.2456446886062622,
"learning_rate": 1.7253243712550147e-05,
"loss": 4.2157,
"num_input_tokens_seen": 140901376,
"step": 2150
},
{
"epoch": 1.7350173602375083,
"grad_norm": 1.249311089515686,
"learning_rate": 1.723321672528655e-05,
"loss": 4.1854,
"num_input_tokens_seen": 141229056,
"step": 2155
},
{
"epoch": 1.7390429225582449,
"grad_norm": 1.3078937530517578,
"learning_rate": 1.721325931647741e-05,
"loss": 4.1911,
"num_input_tokens_seen": 141556736,
"step": 2160
},
{
"epoch": 1.7430684848789815,
"grad_norm": 1.2260226011276245,
"learning_rate": 1.7193371084167158e-05,
"loss": 4.1945,
"num_input_tokens_seen": 141884416,
"step": 2165
},
{
"epoch": 1.747094047199718,
"grad_norm": 1.2438477277755737,
"learning_rate": 1.7173551629643676e-05,
"loss": 4.1877,
"num_input_tokens_seen": 142212096,
"step": 2170
},
{
"epoch": 1.7511196095204549,
"grad_norm": 1.3562901020050049,
"learning_rate": 1.7153800557404717e-05,
"loss": 4.1662,
"num_input_tokens_seen": 142539776,
"step": 2175
},
{
"epoch": 1.7551451718411917,
"grad_norm": 1.2909834384918213,
"learning_rate": 1.7134117475124774e-05,
"loss": 4.1718,
"num_input_tokens_seen": 142867456,
"step": 2180
},
{
"epoch": 1.7591707341619283,
"grad_norm": 1.3816372156143188,
"learning_rate": 1.7114501993622322e-05,
"loss": 4.187,
"num_input_tokens_seen": 143195136,
"step": 2185
},
{
"epoch": 1.7631962964826648,
"grad_norm": 1.1970911026000977,
"learning_rate": 1.7094953726827533e-05,
"loss": 4.2293,
"num_input_tokens_seen": 143522816,
"step": 2190
},
{
"epoch": 1.7672218588034014,
"grad_norm": 1.300299882888794,
"learning_rate": 1.707547229175031e-05,
"loss": 4.1583,
"num_input_tokens_seen": 143850496,
"step": 2195
},
{
"epoch": 1.7712474211241382,
"grad_norm": 1.3158056735992432,
"learning_rate": 1.7056057308448833e-05,
"loss": 4.1986,
"num_input_tokens_seen": 144178176,
"step": 2200
},
{
"epoch": 1.775272983444875,
"grad_norm": 1.3262073993682861,
"learning_rate": 1.7036708399998398e-05,
"loss": 4.1911,
"num_input_tokens_seen": 144505856,
"step": 2205
},
{
"epoch": 1.7792985457656116,
"grad_norm": 1.2565948963165283,
"learning_rate": 1.7017425192460684e-05,
"loss": 4.1982,
"num_input_tokens_seen": 144833536,
"step": 2210
},
{
"epoch": 1.7833241080863482,
"grad_norm": 1.3026916980743408,
"learning_rate": 1.699820731485341e-05,
"loss": 4.1539,
"num_input_tokens_seen": 145161216,
"step": 2215
},
{
"epoch": 1.787349670407085,
"grad_norm": 1.1916109323501587,
"learning_rate": 1.697905439912036e-05,
"loss": 4.1723,
"num_input_tokens_seen": 145488896,
"step": 2220
},
{
"epoch": 1.7913752327278216,
"grad_norm": 1.2661770582199097,
"learning_rate": 1.6959966080101762e-05,
"loss": 4.1575,
"num_input_tokens_seen": 145816576,
"step": 2225
},
{
"epoch": 1.7954007950485584,
"grad_norm": 1.1987600326538086,
"learning_rate": 1.694094199550507e-05,
"loss": 4.1784,
"num_input_tokens_seen": 146144256,
"step": 2230
},
{
"epoch": 1.799426357369295,
"grad_norm": 1.2095016241073608,
"learning_rate": 1.6921981785876064e-05,
"loss": 4.1995,
"num_input_tokens_seen": 146471936,
"step": 2235
},
{
"epoch": 1.8034519196900316,
"grad_norm": 1.220253586769104,
"learning_rate": 1.6903085094570334e-05,
"loss": 4.1876,
"num_input_tokens_seen": 146799616,
"step": 2240
},
{
"epoch": 1.8074774820107684,
"grad_norm": 1.3598766326904297,
"learning_rate": 1.6884251567725093e-05,
"loss": 4.1952,
"num_input_tokens_seen": 147127296,
"step": 2245
},
{
"epoch": 1.8115030443315052,
"grad_norm": 1.2614084482192993,
"learning_rate": 1.6865480854231357e-05,
"loss": 4.1966,
"num_input_tokens_seen": 147454976,
"step": 2250
},
{
"epoch": 1.8115030443315052,
"eval_accuracy": 0.3787434709282334,
"eval_loss": 4.016251564025879,
"eval_runtime": 6.8672,
"eval_samples_per_second": 43.686,
"eval_steps_per_second": 5.534,
"num_input_tokens_seen": 147454976,
"step": 2250
},
{
"epoch": 1.8155286066522418,
"grad_norm": 1.1849647760391235,
"learning_rate": 1.6846772605706453e-05,
"loss": 4.1795,
"num_input_tokens_seen": 147782656,
"step": 2255
},
{
"epoch": 1.8195541689729784,
"grad_norm": 1.217572569847107,
"learning_rate": 1.682812647646685e-05,
"loss": 4.2173,
"num_input_tokens_seen": 148110336,
"step": 2260
},
{
"epoch": 1.823579731293715,
"grad_norm": 1.3434370756149292,
"learning_rate": 1.6809542123501345e-05,
"loss": 4.1826,
"num_input_tokens_seen": 148438016,
"step": 2265
},
{
"epoch": 1.8276052936144518,
"grad_norm": 1.2767517566680908,
"learning_rate": 1.6791019206444543e-05,
"loss": 4.1785,
"num_input_tokens_seen": 148765696,
"step": 2270
},
{
"epoch": 1.8316308559351886,
"grad_norm": 1.1992509365081787,
"learning_rate": 1.6772557387550694e-05,
"loss": 4.1884,
"num_input_tokens_seen": 149093376,
"step": 2275
},
{
"epoch": 1.8356564182559252,
"grad_norm": 1.1988239288330078,
"learning_rate": 1.675415633166782e-05,
"loss": 4.1726,
"num_input_tokens_seen": 149421056,
"step": 2280
},
{
"epoch": 1.8396819805766618,
"grad_norm": 1.3448524475097656,
"learning_rate": 1.673581570621216e-05,
"loss": 4.1467,
"num_input_tokens_seen": 149748736,
"step": 2285
},
{
"epoch": 1.8437075428973984,
"grad_norm": 1.2856345176696777,
"learning_rate": 1.6717535181142915e-05,
"loss": 4.1613,
"num_input_tokens_seen": 150076416,
"step": 2290
},
{
"epoch": 1.8477331052181352,
"grad_norm": 1.357901692390442,
"learning_rate": 1.669931442893732e-05,
"loss": 4.1587,
"num_input_tokens_seen": 150404096,
"step": 2295
},
{
"epoch": 1.851758667538872,
"grad_norm": 1.295551061630249,
"learning_rate": 1.6681153124565983e-05,
"loss": 4.1588,
"num_input_tokens_seen": 150731776,
"step": 2300
},
{
"epoch": 1.8557842298596086,
"grad_norm": 1.2811824083328247,
"learning_rate": 1.6663050945468536e-05,
"loss": 4.1839,
"num_input_tokens_seen": 151059456,
"step": 2305
},
{
"epoch": 1.8598097921803451,
"grad_norm": 1.3775153160095215,
"learning_rate": 1.6645007571529582e-05,
"loss": 4.2016,
"num_input_tokens_seen": 151387136,
"step": 2310
},
{
"epoch": 1.8638353545010817,
"grad_norm": 1.3014804124832153,
"learning_rate": 1.6627022685054907e-05,
"loss": 4.2082,
"num_input_tokens_seen": 151714816,
"step": 2315
},
{
"epoch": 1.8678609168218185,
"grad_norm": 1.2765071392059326,
"learning_rate": 1.6609095970747994e-05,
"loss": 4.1453,
"num_input_tokens_seen": 152042496,
"step": 2320
},
{
"epoch": 1.8718864791425553,
"grad_norm": 1.4745502471923828,
"learning_rate": 1.6591227115686806e-05,
"loss": 4.1856,
"num_input_tokens_seen": 152370176,
"step": 2325
},
{
"epoch": 1.875912041463292,
"grad_norm": 1.2726967334747314,
"learning_rate": 1.6573415809300835e-05,
"loss": 4.1977,
"num_input_tokens_seen": 152697856,
"step": 2330
},
{
"epoch": 1.8799376037840285,
"grad_norm": 1.286971092224121,
"learning_rate": 1.6555661743348445e-05,
"loss": 4.1962,
"num_input_tokens_seen": 153025536,
"step": 2335
},
{
"epoch": 1.883963166104765,
"grad_norm": 1.2794855833053589,
"learning_rate": 1.6537964611894465e-05,
"loss": 4.1606,
"num_input_tokens_seen": 153353216,
"step": 2340
},
{
"epoch": 1.887988728425502,
"grad_norm": 1.2264660596847534,
"learning_rate": 1.652032411128802e-05,
"loss": 4.1724,
"num_input_tokens_seen": 153680896,
"step": 2345
},
{
"epoch": 1.8920142907462387,
"grad_norm": 1.3459347486495972,
"learning_rate": 1.6502739940140696e-05,
"loss": 4.1663,
"num_input_tokens_seen": 154008576,
"step": 2350
},
{
"epoch": 1.8960398530669753,
"grad_norm": 1.284248948097229,
"learning_rate": 1.6485211799304873e-05,
"loss": 4.1927,
"num_input_tokens_seen": 154336256,
"step": 2355
},
{
"epoch": 1.900065415387712,
"grad_norm": 1.273288369178772,
"learning_rate": 1.6467739391852368e-05,
"loss": 4.1442,
"num_input_tokens_seen": 154663936,
"step": 2360
},
{
"epoch": 1.9040909777084485,
"grad_norm": 1.3548243045806885,
"learning_rate": 1.6450322423053303e-05,
"loss": 4.153,
"num_input_tokens_seen": 154991616,
"step": 2365
},
{
"epoch": 1.9081165400291853,
"grad_norm": 1.2603306770324707,
"learning_rate": 1.6432960600355222e-05,
"loss": 4.1805,
"num_input_tokens_seen": 155319296,
"step": 2370
},
{
"epoch": 1.912142102349922,
"grad_norm": 1.3440006971359253,
"learning_rate": 1.6415653633362467e-05,
"loss": 4.1954,
"num_input_tokens_seen": 155646976,
"step": 2375
},
{
"epoch": 1.9161676646706587,
"grad_norm": 1.385733723640442,
"learning_rate": 1.6398401233815756e-05,
"loss": 4.1677,
"num_input_tokens_seen": 155974656,
"step": 2380
},
{
"epoch": 1.9201932269913953,
"grad_norm": 1.408675193786621,
"learning_rate": 1.638120311557203e-05,
"loss": 4.1724,
"num_input_tokens_seen": 156302336,
"step": 2385
},
{
"epoch": 1.924218789312132,
"grad_norm": 1.3172813653945923,
"learning_rate": 1.6364058994584528e-05,
"loss": 4.1658,
"num_input_tokens_seen": 156630016,
"step": 2390
},
{
"epoch": 1.9282443516328687,
"grad_norm": 1.4357519149780273,
"learning_rate": 1.6346968588883048e-05,
"loss": 4.1538,
"num_input_tokens_seen": 156957696,
"step": 2395
},
{
"epoch": 1.9322699139536055,
"grad_norm": 1.3051438331604004,
"learning_rate": 1.6329931618554523e-05,
"loss": 4.16,
"num_input_tokens_seen": 157285376,
"step": 2400
},
{
"epoch": 1.9322699139536055,
"eval_accuracy": 0.3773649363114314,
"eval_loss": 4.035182952880859,
"eval_runtime": 7.0048,
"eval_samples_per_second": 42.828,
"eval_steps_per_second": 5.425,
"num_input_tokens_seen": 157285376,
"step": 2400
},
{
"epoch": 1.936295476274342,
"grad_norm": 1.229457139968872,
"learning_rate": 1.6312947805723712e-05,
"loss": 4.1306,
"num_input_tokens_seen": 157613056,
"step": 2405
},
{
"epoch": 1.9403210385950787,
"grad_norm": 1.3241009712219238,
"learning_rate": 1.629601687453421e-05,
"loss": 4.1803,
"num_input_tokens_seen": 157940736,
"step": 2410
},
{
"epoch": 1.9443466009158155,
"grad_norm": 1.462347149848938,
"learning_rate": 1.6279138551129596e-05,
"loss": 4.1708,
"num_input_tokens_seen": 158268416,
"step": 2415
},
{
"epoch": 1.9483721632365523,
"grad_norm": 1.261893630027771,
"learning_rate": 1.6262312563634837e-05,
"loss": 4.1739,
"num_input_tokens_seen": 158596096,
"step": 2420
},
{
"epoch": 1.9523977255572889,
"grad_norm": 1.2686680555343628,
"learning_rate": 1.624553864213791e-05,
"loss": 4.1601,
"num_input_tokens_seen": 158923776,
"step": 2425
},
{
"epoch": 1.9564232878780254,
"grad_norm": 1.291801929473877,
"learning_rate": 1.622881651867159e-05,
"loss": 4.1971,
"num_input_tokens_seen": 159251456,
"step": 2430
},
{
"epoch": 1.960448850198762,
"grad_norm": 1.2599908113479614,
"learning_rate": 1.6212145927195505e-05,
"loss": 4.1876,
"num_input_tokens_seen": 159579136,
"step": 2435
},
{
"epoch": 1.9644744125194988,
"grad_norm": 1.2517198324203491,
"learning_rate": 1.6195526603578323e-05,
"loss": 4.2083,
"num_input_tokens_seen": 159906816,
"step": 2440
},
{
"epoch": 1.9684999748402356,
"grad_norm": 1.407639741897583,
"learning_rate": 1.6178958285580194e-05,
"loss": 4.1763,
"num_input_tokens_seen": 160234496,
"step": 2445
},
{
"epoch": 1.9725255371609722,
"grad_norm": 1.3766093254089355,
"learning_rate": 1.6162440712835372e-05,
"loss": 4.1656,
"num_input_tokens_seen": 160562176,
"step": 2450
},
{
"epoch": 1.9765510994817088,
"grad_norm": 1.201467514038086,
"learning_rate": 1.6145973626835027e-05,
"loss": 4.1317,
"num_input_tokens_seen": 160889856,
"step": 2455
},
{
"epoch": 1.9805766618024454,
"grad_norm": 1.245469093322754,
"learning_rate": 1.6129556770910237e-05,
"loss": 4.188,
"num_input_tokens_seen": 161217536,
"step": 2460
},
{
"epoch": 1.9846022241231822,
"grad_norm": 1.360424518585205,
"learning_rate": 1.611318989021522e-05,
"loss": 4.1817,
"num_input_tokens_seen": 161545216,
"step": 2465
},
{
"epoch": 1.988627786443919,
"grad_norm": 1.2917958498001099,
"learning_rate": 1.6096872731710677e-05,
"loss": 4.1912,
"num_input_tokens_seen": 161872896,
"step": 2470
},
{
"epoch": 1.9926533487646556,
"grad_norm": 1.311262607574463,
"learning_rate": 1.6080605044147395e-05,
"loss": 4.1897,
"num_input_tokens_seen": 162200576,
"step": 2475
},
{
"epoch": 1.9966789110853922,
"grad_norm": 1.233081340789795,
"learning_rate": 1.606438657804998e-05,
"loss": 4.1536,
"num_input_tokens_seen": 162528256,
"step": 2480
},
{
"epoch": 1.9998993609419817,
"num_input_tokens_seen": 162790400,
"step": 2484,
"total_flos": 1.059416318592e+16,
"train_loss": 4.377665276304727,
"train_runtime": 2972.3443,
"train_samples_per_second": 106.974,
"train_steps_per_second": 0.836,
"train_tokens_per_second": 54770.764
}
],
"logging_steps": 5,
"max_steps": 2484,
"num_input_tokens_seen": 162790400,
"num_train_epochs": 2,
"save_steps": 200,
"total_flos": 1.059416318592e+16,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}