distilbert_base_train_book_v2 / trainer_state.json
gokulsrinivasagan's picture
End of training
d13dda2 verified
{
"best_metric": 1.2004581689834595,
"best_model_checkpoint": "distilbert_base_train_book_v2/checkpoint-350000",
"epoch": 25.0,
"eval_steps": 10000,
"global_step": 355850,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0351271603203597,
"grad_norm": 1.9653347730636597,
"learning_rate": 5e-06,
"loss": 8.8875,
"step": 500
},
{
"epoch": 0.0702543206407194,
"grad_norm": 0.7303386926651001,
"learning_rate": 1e-05,
"loss": 6.5828,
"step": 1000
},
{
"epoch": 0.1053814809610791,
"grad_norm": 0.6638388633728027,
"learning_rate": 1.5e-05,
"loss": 6.0119,
"step": 1500
},
{
"epoch": 0.1405086412814388,
"grad_norm": 0.693476140499115,
"learning_rate": 2e-05,
"loss": 5.8991,
"step": 2000
},
{
"epoch": 0.1756358016017985,
"grad_norm": 0.6988317370414734,
"learning_rate": 2.5e-05,
"loss": 5.8401,
"step": 2500
},
{
"epoch": 0.2107629619221582,
"grad_norm": 0.5958765149116516,
"learning_rate": 3e-05,
"loss": 5.7955,
"step": 3000
},
{
"epoch": 0.2458901222425179,
"grad_norm": 0.4958369731903076,
"learning_rate": 3.5e-05,
"loss": 5.7665,
"step": 3500
},
{
"epoch": 0.2810172825628776,
"grad_norm": 0.4899258613586426,
"learning_rate": 4e-05,
"loss": 5.7418,
"step": 4000
},
{
"epoch": 0.31614444288323734,
"grad_norm": 0.5407448410987854,
"learning_rate": 4.5e-05,
"loss": 5.7212,
"step": 4500
},
{
"epoch": 0.351271603203597,
"grad_norm": 0.5828524231910706,
"learning_rate": 5e-05,
"loss": 5.7049,
"step": 5000
},
{
"epoch": 0.38639876352395675,
"grad_norm": 0.5220353007316589,
"learning_rate": 5.500000000000001e-05,
"loss": 5.6856,
"step": 5500
},
{
"epoch": 0.4215259238443164,
"grad_norm": 0.4556717574596405,
"learning_rate": 6e-05,
"loss": 5.6762,
"step": 6000
},
{
"epoch": 0.45665308416467615,
"grad_norm": 0.49513307213783264,
"learning_rate": 6.500000000000001e-05,
"loss": 5.6635,
"step": 6500
},
{
"epoch": 0.4917802444850358,
"grad_norm": 0.45069605112075806,
"learning_rate": 7e-05,
"loss": 5.6566,
"step": 7000
},
{
"epoch": 0.5269074048053956,
"grad_norm": 0.5079524517059326,
"learning_rate": 7.500000000000001e-05,
"loss": 5.6468,
"step": 7500
},
{
"epoch": 0.5620345651257552,
"grad_norm": 0.5765870809555054,
"learning_rate": 8e-05,
"loss": 5.6365,
"step": 8000
},
{
"epoch": 0.5971617254461149,
"grad_norm": 0.47896265983581543,
"learning_rate": 8.5e-05,
"loss": 5.627,
"step": 8500
},
{
"epoch": 0.6322888857664747,
"grad_norm": 0.6472557783126831,
"learning_rate": 9e-05,
"loss": 5.6161,
"step": 9000
},
{
"epoch": 0.6674160460868344,
"grad_norm": 0.43458178639411926,
"learning_rate": 9.5e-05,
"loss": 5.6111,
"step": 9500
},
{
"epoch": 0.702543206407194,
"grad_norm": 0.4662564694881439,
"learning_rate": 0.0001,
"loss": 5.604,
"step": 10000
},
{
"epoch": 0.702543206407194,
"eval_accuracy": 0.16497299638082222,
"eval_loss": 5.45039176940918,
"eval_runtime": 143.2482,
"eval_samples_per_second": 842.635,
"eval_steps_per_second": 5.271,
"step": 10000
},
{
"epoch": 0.7376703667275537,
"grad_norm": 0.5193693041801453,
"learning_rate": 9.985542865404077e-05,
"loss": 5.5963,
"step": 10500
},
{
"epoch": 0.7727975270479135,
"grad_norm": 0.41712281107902527,
"learning_rate": 9.971085730808155e-05,
"loss": 5.5891,
"step": 11000
},
{
"epoch": 0.8079246873682732,
"grad_norm": 0.4692609906196594,
"learning_rate": 9.956628596212231e-05,
"loss": 5.5854,
"step": 11500
},
{
"epoch": 0.8430518476886328,
"grad_norm": 0.4751366376876831,
"learning_rate": 9.942171461616309e-05,
"loss": 5.5763,
"step": 12000
},
{
"epoch": 0.8781790080089925,
"grad_norm": 0.4744866192340851,
"learning_rate": 9.927714327020385e-05,
"loss": 5.5709,
"step": 12500
},
{
"epoch": 0.9133061683293523,
"grad_norm": 0.4810524582862854,
"learning_rate": 9.913257192424462e-05,
"loss": 5.5608,
"step": 13000
},
{
"epoch": 0.948433328649712,
"grad_norm": 0.5423319935798645,
"learning_rate": 9.89880005782854e-05,
"loss": 5.5541,
"step": 13500
},
{
"epoch": 0.9835604889700716,
"grad_norm": 0.4694032371044159,
"learning_rate": 9.884342923232616e-05,
"loss": 5.5498,
"step": 14000
},
{
"epoch": 1.0186876492904313,
"grad_norm": 0.44750478863716125,
"learning_rate": 9.869885788636692e-05,
"loss": 5.5414,
"step": 14500
},
{
"epoch": 1.053814809610791,
"grad_norm": 0.4432881772518158,
"learning_rate": 9.855428654040769e-05,
"loss": 5.5374,
"step": 15000
},
{
"epoch": 1.0889419699311507,
"grad_norm": 0.4728132486343384,
"learning_rate": 9.840971519444846e-05,
"loss": 5.5297,
"step": 15500
},
{
"epoch": 1.1240691302515105,
"grad_norm": 0.48674553632736206,
"learning_rate": 9.826514384848924e-05,
"loss": 5.5254,
"step": 16000
},
{
"epoch": 1.1591962905718702,
"grad_norm": 0.5000836849212646,
"learning_rate": 9.812057250253e-05,
"loss": 5.52,
"step": 16500
},
{
"epoch": 1.1943234508922298,
"grad_norm": 0.43097734451293945,
"learning_rate": 9.797600115657077e-05,
"loss": 5.5166,
"step": 17000
},
{
"epoch": 1.2294506112125896,
"grad_norm": 0.4628123342990875,
"learning_rate": 9.783142981061155e-05,
"loss": 5.5085,
"step": 17500
},
{
"epoch": 1.2645777715329491,
"grad_norm": 0.46364983916282654,
"learning_rate": 9.768685846465231e-05,
"loss": 5.507,
"step": 18000
},
{
"epoch": 1.299704931853309,
"grad_norm": 0.6745333671569824,
"learning_rate": 9.754228711869307e-05,
"loss": 5.4862,
"step": 18500
},
{
"epoch": 1.3348320921736687,
"grad_norm": 0.9819153547286987,
"learning_rate": 9.739771577273384e-05,
"loss": 5.3901,
"step": 19000
},
{
"epoch": 1.3699592524940285,
"grad_norm": 2.0434741973876953,
"learning_rate": 9.725314442677463e-05,
"loss": 5.1978,
"step": 19500
},
{
"epoch": 1.405086412814388,
"grad_norm": 2.197896957397461,
"learning_rate": 9.71085730808154e-05,
"loss": 4.6179,
"step": 20000
},
{
"epoch": 1.405086412814388,
"eval_accuracy": 0.3758464833297413,
"eval_loss": 3.827747344970703,
"eval_runtime": 146.1003,
"eval_samples_per_second": 826.186,
"eval_steps_per_second": 5.168,
"step": 20000
},
{
"epoch": 1.4402135731347478,
"grad_norm": 1.750232219696045,
"learning_rate": 9.696400173485616e-05,
"loss": 4.0652,
"step": 20500
},
{
"epoch": 1.4753407334551074,
"grad_norm": 1.4147313833236694,
"learning_rate": 9.681943038889692e-05,
"loss": 3.6557,
"step": 21000
},
{
"epoch": 1.5104678937754672,
"grad_norm": 1.1749171018600464,
"learning_rate": 9.66748590429377e-05,
"loss": 3.3421,
"step": 21500
},
{
"epoch": 1.545595054095827,
"grad_norm": 1.1297008991241455,
"learning_rate": 9.653028769697846e-05,
"loss": 3.1555,
"step": 22000
},
{
"epoch": 1.5807222144161865,
"grad_norm": 1.0993692874908447,
"learning_rate": 9.638571635101923e-05,
"loss": 3.0326,
"step": 22500
},
{
"epoch": 1.6158493747365463,
"grad_norm": 1.1023796796798706,
"learning_rate": 9.624114500506e-05,
"loss": 2.9348,
"step": 23000
},
{
"epoch": 1.650976535056906,
"grad_norm": 1.036067008972168,
"learning_rate": 9.609657365910077e-05,
"loss": 2.8555,
"step": 23500
},
{
"epoch": 1.6861036953772657,
"grad_norm": 1.0167666673660278,
"learning_rate": 9.595200231314155e-05,
"loss": 2.7872,
"step": 24000
},
{
"epoch": 1.7212308556976255,
"grad_norm": 1.0188016891479492,
"learning_rate": 9.580743096718231e-05,
"loss": 2.7329,
"step": 24500
},
{
"epoch": 1.7563580160179852,
"grad_norm": 1.0286072492599487,
"learning_rate": 9.566285962122307e-05,
"loss": 2.6825,
"step": 25000
},
{
"epoch": 1.7914851763383448,
"grad_norm": 0.9929981827735901,
"learning_rate": 9.551828827526385e-05,
"loss": 2.6381,
"step": 25500
},
{
"epoch": 1.8266123366587044,
"grad_norm": 0.9288387894630432,
"learning_rate": 9.537371692930462e-05,
"loss": 2.5937,
"step": 26000
},
{
"epoch": 1.8617394969790642,
"grad_norm": 0.9797139763832092,
"learning_rate": 9.522914558334538e-05,
"loss": 2.5524,
"step": 26500
},
{
"epoch": 1.896866657299424,
"grad_norm": 0.9662619233131409,
"learning_rate": 9.508457423738616e-05,
"loss": 2.519,
"step": 27000
},
{
"epoch": 1.9319938176197837,
"grad_norm": 0.9594979286193848,
"learning_rate": 9.494000289142692e-05,
"loss": 2.4854,
"step": 27500
},
{
"epoch": 1.9671209779401433,
"grad_norm": 0.9515316486358643,
"learning_rate": 9.47954315454677e-05,
"loss": 2.4515,
"step": 28000
},
{
"epoch": 2.002248138260503,
"grad_norm": 0.9741261601448059,
"learning_rate": 9.465086019950846e-05,
"loss": 2.4314,
"step": 28500
},
{
"epoch": 2.0373752985808626,
"grad_norm": 0.9476631283760071,
"learning_rate": 9.450628885354923e-05,
"loss": 2.3975,
"step": 29000
},
{
"epoch": 2.0725024589012224,
"grad_norm": 0.9616317749023438,
"learning_rate": 9.436171750758999e-05,
"loss": 2.3707,
"step": 29500
},
{
"epoch": 2.107629619221582,
"grad_norm": 0.9059496521949768,
"learning_rate": 9.421714616163077e-05,
"loss": 2.353,
"step": 30000
},
{
"epoch": 2.107629619221582,
"eval_accuracy": 0.5933188090938791,
"eval_loss": 2.040786027908325,
"eval_runtime": 146.0422,
"eval_samples_per_second": 826.515,
"eval_steps_per_second": 5.17,
"step": 30000
},
{
"epoch": 2.142756779541942,
"grad_norm": 0.9514384865760803,
"learning_rate": 9.407257481567155e-05,
"loss": 2.3312,
"step": 30500
},
{
"epoch": 2.1778839398623013,
"grad_norm": 0.9169853925704956,
"learning_rate": 9.392800346971231e-05,
"loss": 2.3091,
"step": 31000
},
{
"epoch": 2.213011100182661,
"grad_norm": 0.9202592968940735,
"learning_rate": 9.378343212375307e-05,
"loss": 2.2937,
"step": 31500
},
{
"epoch": 2.248138260503021,
"grad_norm": 0.9011738300323486,
"learning_rate": 9.363886077779385e-05,
"loss": 2.2759,
"step": 32000
},
{
"epoch": 2.2832654208233807,
"grad_norm": 0.8867260217666626,
"learning_rate": 9.349428943183462e-05,
"loss": 2.2593,
"step": 32500
},
{
"epoch": 2.3183925811437405,
"grad_norm": 0.9071447253227234,
"learning_rate": 9.334971808587538e-05,
"loss": 2.2402,
"step": 33000
},
{
"epoch": 2.3535197414641003,
"grad_norm": 0.9013161063194275,
"learning_rate": 9.320514673991614e-05,
"loss": 2.2255,
"step": 33500
},
{
"epoch": 2.3886469017844596,
"grad_norm": 0.8916395306587219,
"learning_rate": 9.306057539395692e-05,
"loss": 2.2123,
"step": 34000
},
{
"epoch": 2.4237740621048194,
"grad_norm": 0.8660832047462463,
"learning_rate": 9.29160040479977e-05,
"loss": 2.1955,
"step": 34500
},
{
"epoch": 2.458901222425179,
"grad_norm": 0.8876006603240967,
"learning_rate": 9.277143270203846e-05,
"loss": 2.1837,
"step": 35000
},
{
"epoch": 2.494028382745539,
"grad_norm": 0.8954652547836304,
"learning_rate": 9.262686135607923e-05,
"loss": 2.1732,
"step": 35500
},
{
"epoch": 2.5291555430658983,
"grad_norm": 0.9105496406555176,
"learning_rate": 9.248229001012e-05,
"loss": 2.1599,
"step": 36000
},
{
"epoch": 2.564282703386258,
"grad_norm": 0.8876945972442627,
"learning_rate": 9.233771866416077e-05,
"loss": 2.1472,
"step": 36500
},
{
"epoch": 2.599409863706618,
"grad_norm": 0.8663467168807983,
"learning_rate": 9.219314731820153e-05,
"loss": 2.1345,
"step": 37000
},
{
"epoch": 2.6345370240269776,
"grad_norm": 0.9176456928253174,
"learning_rate": 9.20485759722423e-05,
"loss": 2.1258,
"step": 37500
},
{
"epoch": 2.6696641843473374,
"grad_norm": 0.9134395718574524,
"learning_rate": 9.190400462628307e-05,
"loss": 2.1125,
"step": 38000
},
{
"epoch": 2.704791344667697,
"grad_norm": 0.8617592453956604,
"learning_rate": 9.175943328032385e-05,
"loss": 2.1041,
"step": 38500
},
{
"epoch": 2.739918504988057,
"grad_norm": 0.8791117668151855,
"learning_rate": 9.161486193436461e-05,
"loss": 2.0931,
"step": 39000
},
{
"epoch": 2.7750456653084163,
"grad_norm": 0.8784831762313843,
"learning_rate": 9.147029058840538e-05,
"loss": 2.0854,
"step": 39500
},
{
"epoch": 2.810172825628776,
"grad_norm": 0.8679407238960266,
"learning_rate": 9.132571924244616e-05,
"loss": 2.0739,
"step": 40000
},
{
"epoch": 2.810172825628776,
"eval_accuracy": 0.6315697031283101,
"eval_loss": 1.7932097911834717,
"eval_runtime": 144.409,
"eval_samples_per_second": 835.862,
"eval_steps_per_second": 5.228,
"step": 40000
},
{
"epoch": 2.845299985949136,
"grad_norm": 0.8974806070327759,
"learning_rate": 9.118114789648692e-05,
"loss": 2.0627,
"step": 40500
},
{
"epoch": 2.8804271462694957,
"grad_norm": 0.8850300312042236,
"learning_rate": 9.103657655052768e-05,
"loss": 2.0544,
"step": 41000
},
{
"epoch": 2.915554306589855,
"grad_norm": 0.8605426549911499,
"learning_rate": 9.089200520456845e-05,
"loss": 2.0464,
"step": 41500
},
{
"epoch": 2.950681466910215,
"grad_norm": 0.8664257526397705,
"learning_rate": 9.074743385860923e-05,
"loss": 2.0404,
"step": 42000
},
{
"epoch": 2.9858086272305746,
"grad_norm": 0.8507780432701111,
"learning_rate": 9.060286251265e-05,
"loss": 2.0325,
"step": 42500
},
{
"epoch": 3.0209357875509344,
"grad_norm": 0.8595208525657654,
"learning_rate": 9.045829116669077e-05,
"loss": 2.018,
"step": 43000
},
{
"epoch": 3.056062947871294,
"grad_norm": 0.8866372108459473,
"learning_rate": 9.031371982073153e-05,
"loss": 2.0135,
"step": 43500
},
{
"epoch": 3.091190108191654,
"grad_norm": 0.8739855289459229,
"learning_rate": 9.016914847477231e-05,
"loss": 2.0101,
"step": 44000
},
{
"epoch": 3.1263172685120133,
"grad_norm": 0.8769083619117737,
"learning_rate": 9.002457712881307e-05,
"loss": 2.0,
"step": 44500
},
{
"epoch": 3.161444428832373,
"grad_norm": 0.8757010102272034,
"learning_rate": 8.988000578285384e-05,
"loss": 1.9949,
"step": 45000
},
{
"epoch": 3.196571589152733,
"grad_norm": 0.8755298256874084,
"learning_rate": 8.973543443689461e-05,
"loss": 1.9849,
"step": 45500
},
{
"epoch": 3.2316987494730927,
"grad_norm": 0.8611502647399902,
"learning_rate": 8.959086309093538e-05,
"loss": 1.9821,
"step": 46000
},
{
"epoch": 3.2668259097934524,
"grad_norm": 0.8552160859107971,
"learning_rate": 8.944629174497616e-05,
"loss": 1.9738,
"step": 46500
},
{
"epoch": 3.301953070113812,
"grad_norm": 0.8645509481430054,
"learning_rate": 8.930172039901692e-05,
"loss": 1.9678,
"step": 47000
},
{
"epoch": 3.3370802304341716,
"grad_norm": 0.8320329189300537,
"learning_rate": 8.915714905305768e-05,
"loss": 1.9614,
"step": 47500
},
{
"epoch": 3.3722073907545314,
"grad_norm": 0.8820323944091797,
"learning_rate": 8.901257770709845e-05,
"loss": 1.9578,
"step": 48000
},
{
"epoch": 3.407334551074891,
"grad_norm": 0.847457230091095,
"learning_rate": 8.886800636113923e-05,
"loss": 1.9494,
"step": 48500
},
{
"epoch": 3.442461711395251,
"grad_norm": 0.858734667301178,
"learning_rate": 8.872343501518e-05,
"loss": 1.942,
"step": 49000
},
{
"epoch": 3.4775888717156107,
"grad_norm": 0.8552329540252686,
"learning_rate": 8.857886366922077e-05,
"loss": 1.9367,
"step": 49500
},
{
"epoch": 3.51271603203597,
"grad_norm": 0.8536255359649658,
"learning_rate": 8.843429232326153e-05,
"loss": 1.9345,
"step": 50000
},
{
"epoch": 3.51271603203597,
"eval_accuracy": 0.652718528822783,
"eval_loss": 1.6618030071258545,
"eval_runtime": 145.7638,
"eval_samples_per_second": 828.093,
"eval_steps_per_second": 5.18,
"step": 50000
},
{
"epoch": 3.54784319235633,
"grad_norm": 0.8829108476638794,
"learning_rate": 8.828972097730231e-05,
"loss": 1.9303,
"step": 50500
},
{
"epoch": 3.5829703526766896,
"grad_norm": 0.8787323236465454,
"learning_rate": 8.814514963134307e-05,
"loss": 1.9231,
"step": 51000
},
{
"epoch": 3.6180975129970494,
"grad_norm": 0.8392846584320068,
"learning_rate": 8.800057828538384e-05,
"loss": 1.9205,
"step": 51500
},
{
"epoch": 3.6532246733174087,
"grad_norm": 0.8578284382820129,
"learning_rate": 8.78560069394246e-05,
"loss": 1.9147,
"step": 52000
},
{
"epoch": 3.6883518336377685,
"grad_norm": 0.8548659086227417,
"learning_rate": 8.771143559346538e-05,
"loss": 1.9079,
"step": 52500
},
{
"epoch": 3.7234789939581283,
"grad_norm": 0.8397179245948792,
"learning_rate": 8.756686424750616e-05,
"loss": 1.9064,
"step": 53000
},
{
"epoch": 3.758606154278488,
"grad_norm": 0.886605978012085,
"learning_rate": 8.742229290154692e-05,
"loss": 1.9006,
"step": 53500
},
{
"epoch": 3.793733314598848,
"grad_norm": 0.8749421834945679,
"learning_rate": 8.727772155558768e-05,
"loss": 1.894,
"step": 54000
},
{
"epoch": 3.8288604749192077,
"grad_norm": 0.8433464765548706,
"learning_rate": 8.713315020962846e-05,
"loss": 1.8917,
"step": 54500
},
{
"epoch": 3.8639876352395675,
"grad_norm": 0.8542115092277527,
"learning_rate": 8.698857886366923e-05,
"loss": 1.8846,
"step": 55000
},
{
"epoch": 3.899114795559927,
"grad_norm": 0.8477907776832581,
"learning_rate": 8.684400751770999e-05,
"loss": 1.88,
"step": 55500
},
{
"epoch": 3.9342419558802866,
"grad_norm": 0.8515335321426392,
"learning_rate": 8.669943617175075e-05,
"loss": 1.8768,
"step": 56000
},
{
"epoch": 3.9693691162006464,
"grad_norm": 0.8547754287719727,
"learning_rate": 8.655486482579153e-05,
"loss": 1.8742,
"step": 56500
},
{
"epoch": 4.004496276521006,
"grad_norm": 0.8512959480285645,
"learning_rate": 8.641029347983231e-05,
"loss": 1.868,
"step": 57000
},
{
"epoch": 4.0396234368413655,
"grad_norm": 0.8635633587837219,
"learning_rate": 8.626572213387307e-05,
"loss": 1.8603,
"step": 57500
},
{
"epoch": 4.074750597161725,
"grad_norm": 0.8565309643745422,
"learning_rate": 8.612115078791384e-05,
"loss": 1.858,
"step": 58000
},
{
"epoch": 4.109877757482085,
"grad_norm": 0.8473580479621887,
"learning_rate": 8.597657944195461e-05,
"loss": 1.8543,
"step": 58500
},
{
"epoch": 4.145004917802445,
"grad_norm": 0.8602383732795715,
"learning_rate": 8.583200809599538e-05,
"loss": 1.8505,
"step": 59000
},
{
"epoch": 4.180132078122805,
"grad_norm": 0.8420585989952087,
"learning_rate": 8.568743675003614e-05,
"loss": 1.8507,
"step": 59500
},
{
"epoch": 4.215259238443164,
"grad_norm": 0.8763858079910278,
"learning_rate": 8.55428654040769e-05,
"loss": 1.844,
"step": 60000
},
{
"epoch": 4.215259238443164,
"eval_accuracy": 0.6653352124865598,
"eval_loss": 1.5828626155853271,
"eval_runtime": 145.5436,
"eval_samples_per_second": 829.346,
"eval_steps_per_second": 5.187,
"step": 60000
},
{
"epoch": 4.250386398763524,
"grad_norm": 0.8648141026496887,
"learning_rate": 8.539829405811768e-05,
"loss": 1.8439,
"step": 60500
},
{
"epoch": 4.285513559083884,
"grad_norm": 0.8564465045928955,
"learning_rate": 8.525372271215846e-05,
"loss": 1.8372,
"step": 61000
},
{
"epoch": 4.320640719404244,
"grad_norm": 0.8347111344337463,
"learning_rate": 8.510915136619923e-05,
"loss": 1.8322,
"step": 61500
},
{
"epoch": 4.355767879724603,
"grad_norm": 0.8819388747215271,
"learning_rate": 8.496458002023999e-05,
"loss": 1.8299,
"step": 62000
},
{
"epoch": 4.3908950400449624,
"grad_norm": 0.8561258316040039,
"learning_rate": 8.482000867428075e-05,
"loss": 1.8319,
"step": 62500
},
{
"epoch": 4.426022200365322,
"grad_norm": 0.8755387663841248,
"learning_rate": 8.467543732832153e-05,
"loss": 1.827,
"step": 63000
},
{
"epoch": 4.461149360685682,
"grad_norm": 0.8702017068862915,
"learning_rate": 8.45308659823623e-05,
"loss": 1.8213,
"step": 63500
},
{
"epoch": 4.496276521006042,
"grad_norm": 0.8973068594932556,
"learning_rate": 8.438629463640307e-05,
"loss": 1.819,
"step": 64000
},
{
"epoch": 4.531403681326402,
"grad_norm": 0.8871073126792908,
"learning_rate": 8.424172329044384e-05,
"loss": 1.8166,
"step": 64500
},
{
"epoch": 4.566530841646761,
"grad_norm": 0.8684224486351013,
"learning_rate": 8.409715194448461e-05,
"loss": 1.8129,
"step": 65000
},
{
"epoch": 4.601658001967121,
"grad_norm": 0.8582248687744141,
"learning_rate": 8.395258059852538e-05,
"loss": 1.8113,
"step": 65500
},
{
"epoch": 4.636785162287481,
"grad_norm": 0.843012809753418,
"learning_rate": 8.380800925256614e-05,
"loss": 1.807,
"step": 66000
},
{
"epoch": 4.671912322607841,
"grad_norm": 0.8517187237739563,
"learning_rate": 8.36634379066069e-05,
"loss": 1.8064,
"step": 66500
},
{
"epoch": 4.7070394829282005,
"grad_norm": 0.8724709153175354,
"learning_rate": 8.351886656064768e-05,
"loss": 1.8003,
"step": 67000
},
{
"epoch": 4.742166643248559,
"grad_norm": 0.8867350220680237,
"learning_rate": 8.337429521468846e-05,
"loss": 1.799,
"step": 67500
},
{
"epoch": 4.777293803568919,
"grad_norm": 0.8441566824913025,
"learning_rate": 8.322972386872923e-05,
"loss": 1.7979,
"step": 68000
},
{
"epoch": 4.812420963889279,
"grad_norm": 0.8734049201011658,
"learning_rate": 8.308515252276999e-05,
"loss": 1.7957,
"step": 68500
},
{
"epoch": 4.847548124209639,
"grad_norm": 0.8671762347221375,
"learning_rate": 8.294058117681077e-05,
"loss": 1.789,
"step": 69000
},
{
"epoch": 4.8826752845299985,
"grad_norm": 0.8785460591316223,
"learning_rate": 8.279600983085153e-05,
"loss": 1.7892,
"step": 69500
},
{
"epoch": 4.917802444850358,
"grad_norm": 0.8827086687088013,
"learning_rate": 8.26514384848923e-05,
"loss": 1.7874,
"step": 70000
},
{
"epoch": 4.917802444850358,
"eval_accuracy": 0.6750020166957967,
"eval_loss": 1.5247992277145386,
"eval_runtime": 144.6205,
"eval_samples_per_second": 834.639,
"eval_steps_per_second": 5.221,
"step": 70000
},
{
"epoch": 4.952929605170718,
"grad_norm": 0.8440523743629456,
"learning_rate": 8.250686713893306e-05,
"loss": 1.7851,
"step": 70500
},
{
"epoch": 4.988056765491078,
"grad_norm": 0.8763037919998169,
"learning_rate": 8.236229579297384e-05,
"loss": 1.7801,
"step": 71000
},
{
"epoch": 5.023183925811438,
"grad_norm": 0.88693767786026,
"learning_rate": 8.221772444701461e-05,
"loss": 1.7789,
"step": 71500
},
{
"epoch": 5.0583110861317975,
"grad_norm": 0.866883397102356,
"learning_rate": 8.207315310105538e-05,
"loss": 1.7714,
"step": 72000
},
{
"epoch": 5.093438246452157,
"grad_norm": 0.8699513077735901,
"learning_rate": 8.192858175509614e-05,
"loss": 1.7682,
"step": 72500
},
{
"epoch": 5.128565406772516,
"grad_norm": 0.8703712821006775,
"learning_rate": 8.178401040913692e-05,
"loss": 1.7662,
"step": 73000
},
{
"epoch": 5.163692567092876,
"grad_norm": 0.854120135307312,
"learning_rate": 8.163943906317768e-05,
"loss": 1.7647,
"step": 73500
},
{
"epoch": 5.198819727413236,
"grad_norm": 0.8597338795661926,
"learning_rate": 8.149486771721845e-05,
"loss": 1.7613,
"step": 74000
},
{
"epoch": 5.2339468877335955,
"grad_norm": 0.8737318515777588,
"learning_rate": 8.135029637125921e-05,
"loss": 1.7627,
"step": 74500
},
{
"epoch": 5.269074048053955,
"grad_norm": 0.8847633600234985,
"learning_rate": 8.120572502529999e-05,
"loss": 1.7594,
"step": 75000
},
{
"epoch": 5.304201208374315,
"grad_norm": 0.8676943778991699,
"learning_rate": 8.106115367934077e-05,
"loss": 1.7574,
"step": 75500
},
{
"epoch": 5.339328368694675,
"grad_norm": 0.8717545866966248,
"learning_rate": 8.091658233338153e-05,
"loss": 1.7551,
"step": 76000
},
{
"epoch": 5.374455529015035,
"grad_norm": 0.8879154920578003,
"learning_rate": 8.07720109874223e-05,
"loss": 1.752,
"step": 76500
},
{
"epoch": 5.409582689335394,
"grad_norm": 0.861165463924408,
"learning_rate": 8.062743964146306e-05,
"loss": 1.748,
"step": 77000
},
{
"epoch": 5.444709849655754,
"grad_norm": 0.885162889957428,
"learning_rate": 8.048286829550384e-05,
"loss": 1.7444,
"step": 77500
},
{
"epoch": 5.479837009976113,
"grad_norm": 0.8831413388252258,
"learning_rate": 8.03382969495446e-05,
"loss": 1.7457,
"step": 78000
},
{
"epoch": 5.514964170296473,
"grad_norm": 0.8693949580192566,
"learning_rate": 8.019372560358536e-05,
"loss": 1.7439,
"step": 78500
},
{
"epoch": 5.550091330616833,
"grad_norm": 0.8950614333152771,
"learning_rate": 8.004915425762614e-05,
"loss": 1.7414,
"step": 79000
},
{
"epoch": 5.5852184909371925,
"grad_norm": 0.8772277235984802,
"learning_rate": 7.990458291166692e-05,
"loss": 1.7397,
"step": 79500
},
{
"epoch": 5.620345651257552,
"grad_norm": 0.8824251890182495,
"learning_rate": 7.976001156570768e-05,
"loss": 1.737,
"step": 80000
},
{
"epoch": 5.620345651257552,
"eval_accuracy": 0.6819064890501978,
"eval_loss": 1.4823895692825317,
"eval_runtime": 145.2164,
"eval_samples_per_second": 831.215,
"eval_steps_per_second": 5.199,
"step": 80000
},
{
"epoch": 5.655472811577912,
"grad_norm": 0.9014139175415039,
"learning_rate": 7.961544021974845e-05,
"loss": 1.7339,
"step": 80500
},
{
"epoch": 5.690599971898272,
"grad_norm": 0.8738614320755005,
"learning_rate": 7.947086887378921e-05,
"loss": 1.7395,
"step": 81000
},
{
"epoch": 5.725727132218632,
"grad_norm": 0.8764089941978455,
"learning_rate": 7.932629752782999e-05,
"loss": 1.7336,
"step": 81500
},
{
"epoch": 5.760854292538991,
"grad_norm": 0.9202150106430054,
"learning_rate": 7.918172618187075e-05,
"loss": 1.7291,
"step": 82000
},
{
"epoch": 5.795981452859351,
"grad_norm": 0.8928285837173462,
"learning_rate": 7.903715483591153e-05,
"loss": 1.7302,
"step": 82500
},
{
"epoch": 5.83110861317971,
"grad_norm": 0.8655526041984558,
"learning_rate": 7.88925834899523e-05,
"loss": 1.7287,
"step": 83000
},
{
"epoch": 5.866235773500071,
"grad_norm": 0.8782069087028503,
"learning_rate": 7.874801214399307e-05,
"loss": 1.7261,
"step": 83500
},
{
"epoch": 5.90136293382043,
"grad_norm": 0.8890528678894043,
"learning_rate": 7.860344079803384e-05,
"loss": 1.7237,
"step": 84000
},
{
"epoch": 5.936490094140789,
"grad_norm": 0.8649280667304993,
"learning_rate": 7.84588694520746e-05,
"loss": 1.7213,
"step": 84500
},
{
"epoch": 5.971617254461149,
"grad_norm": 0.9018586277961731,
"learning_rate": 7.831429810611536e-05,
"loss": 1.7194,
"step": 85000
},
{
"epoch": 6.006744414781509,
"grad_norm": 0.8934980034828186,
"learning_rate": 7.816972676015614e-05,
"loss": 1.7184,
"step": 85500
},
{
"epoch": 6.041871575101869,
"grad_norm": 0.8839065432548523,
"learning_rate": 7.802515541419692e-05,
"loss": 1.7127,
"step": 86000
},
{
"epoch": 6.076998735422229,
"grad_norm": 0.8806888461112976,
"learning_rate": 7.788058406823768e-05,
"loss": 1.7144,
"step": 86500
},
{
"epoch": 6.112125895742588,
"grad_norm": 0.8924936652183533,
"learning_rate": 7.773601272227845e-05,
"loss": 1.7084,
"step": 87000
},
{
"epoch": 6.147253056062948,
"grad_norm": 0.9053957462310791,
"learning_rate": 7.759144137631922e-05,
"loss": 1.7076,
"step": 87500
},
{
"epoch": 6.182380216383308,
"grad_norm": 0.8794803023338318,
"learning_rate": 7.744687003035999e-05,
"loss": 1.7069,
"step": 88000
},
{
"epoch": 6.217507376703667,
"grad_norm": 0.9028088450431824,
"learning_rate": 7.730229868440075e-05,
"loss": 1.7067,
"step": 88500
},
{
"epoch": 6.252634537024027,
"grad_norm": 0.8775391578674316,
"learning_rate": 7.715772733844152e-05,
"loss": 1.7021,
"step": 89000
},
{
"epoch": 6.287761697344386,
"grad_norm": 0.9188543558120728,
"learning_rate": 7.70131559924823e-05,
"loss": 1.7019,
"step": 89500
},
{
"epoch": 6.322888857664746,
"grad_norm": 0.8761645555496216,
"learning_rate": 7.686858464652307e-05,
"loss": 1.7017,
"step": 90000
},
{
"epoch": 6.322888857664746,
"eval_accuracy": 0.6875799717584923,
"eval_loss": 1.45062255859375,
"eval_runtime": 146.6571,
"eval_samples_per_second": 823.049,
"eval_steps_per_second": 5.148,
"step": 90000
},
{
"epoch": 6.358016017985106,
"grad_norm": 0.8915168642997742,
"learning_rate": 7.672401330056384e-05,
"loss": 1.7011,
"step": 90500
},
{
"epoch": 6.393143178305466,
"grad_norm": 0.8964366912841797,
"learning_rate": 7.65794419546046e-05,
"loss": 1.6985,
"step": 91000
},
{
"epoch": 6.4282703386258255,
"grad_norm": 0.8956195712089539,
"learning_rate": 7.643487060864536e-05,
"loss": 1.6956,
"step": 91500
},
{
"epoch": 6.463397498946185,
"grad_norm": 0.9271071553230286,
"learning_rate": 7.629029926268614e-05,
"loss": 1.695,
"step": 92000
},
{
"epoch": 6.498524659266545,
"grad_norm": 0.882736086845398,
"learning_rate": 7.61457279167269e-05,
"loss": 1.6941,
"step": 92500
},
{
"epoch": 6.533651819586905,
"grad_norm": 0.866625964641571,
"learning_rate": 7.600115657076767e-05,
"loss": 1.6942,
"step": 93000
},
{
"epoch": 6.568778979907265,
"grad_norm": 0.9276591539382935,
"learning_rate": 7.585658522480845e-05,
"loss": 1.6915,
"step": 93500
},
{
"epoch": 6.603906140227624,
"grad_norm": 0.8902279138565063,
"learning_rate": 7.571201387884922e-05,
"loss": 1.6874,
"step": 94000
},
{
"epoch": 6.639033300547983,
"grad_norm": 0.8612838983535767,
"learning_rate": 7.556744253288999e-05,
"loss": 1.6864,
"step": 94500
},
{
"epoch": 6.674160460868343,
"grad_norm": 0.8922297358512878,
"learning_rate": 7.542287118693075e-05,
"loss": 1.6851,
"step": 95000
},
{
"epoch": 6.709287621188703,
"grad_norm": 0.9003462791442871,
"learning_rate": 7.527829984097152e-05,
"loss": 1.6843,
"step": 95500
},
{
"epoch": 6.744414781509063,
"grad_norm": 0.8967103958129883,
"learning_rate": 7.51337284950123e-05,
"loss": 1.6888,
"step": 96000
},
{
"epoch": 6.7795419418294225,
"grad_norm": 0.8979239463806152,
"learning_rate": 7.498915714905306e-05,
"loss": 1.6813,
"step": 96500
},
{
"epoch": 6.814669102149782,
"grad_norm": 0.8968941569328308,
"learning_rate": 7.484458580309382e-05,
"loss": 1.6822,
"step": 97000
},
{
"epoch": 6.849796262470142,
"grad_norm": 0.8832077383995056,
"learning_rate": 7.47000144571346e-05,
"loss": 1.676,
"step": 97500
},
{
"epoch": 6.884923422790502,
"grad_norm": 0.8957480192184448,
"learning_rate": 7.455544311117538e-05,
"loss": 1.6772,
"step": 98000
},
{
"epoch": 6.920050583110862,
"grad_norm": 0.911115825176239,
"learning_rate": 7.441087176521614e-05,
"loss": 1.674,
"step": 98500
},
{
"epoch": 6.955177743431221,
"grad_norm": 0.9081796407699585,
"learning_rate": 7.42663004192569e-05,
"loss": 1.6746,
"step": 99000
},
{
"epoch": 6.99030490375158,
"grad_norm": 0.9057807922363281,
"learning_rate": 7.412172907329767e-05,
"loss": 1.6702,
"step": 99500
},
{
"epoch": 7.02543206407194,
"grad_norm": 0.9082474708557129,
"learning_rate": 7.397715772733845e-05,
"loss": 1.6703,
"step": 100000
},
{
"epoch": 7.02543206407194,
"eval_accuracy": 0.6921466981210505,
"eval_loss": 1.420421838760376,
"eval_runtime": 146.0221,
"eval_samples_per_second": 826.628,
"eval_steps_per_second": 5.17,
"step": 100000
},
{
"epoch": 7.0605592243923,
"grad_norm": 0.9067661166191101,
"learning_rate": 7.383258638137921e-05,
"loss": 1.6691,
"step": 100500
},
{
"epoch": 7.09568638471266,
"grad_norm": 0.9043187499046326,
"learning_rate": 7.368801503541999e-05,
"loss": 1.669,
"step": 101000
},
{
"epoch": 7.1308135450330195,
"grad_norm": 0.8964481949806213,
"learning_rate": 7.354344368946075e-05,
"loss": 1.667,
"step": 101500
},
{
"epoch": 7.165940705353379,
"grad_norm": 0.9258145689964294,
"learning_rate": 7.339887234350153e-05,
"loss": 1.6653,
"step": 102000
},
{
"epoch": 7.201067865673739,
"grad_norm": 0.8769683241844177,
"learning_rate": 7.32543009975423e-05,
"loss": 1.664,
"step": 102500
},
{
"epoch": 7.236195025994099,
"grad_norm": 0.910929262638092,
"learning_rate": 7.310972965158306e-05,
"loss": 1.6624,
"step": 103000
},
{
"epoch": 7.271322186314459,
"grad_norm": 0.8838810324668884,
"learning_rate": 7.296515830562382e-05,
"loss": 1.6628,
"step": 103500
},
{
"epoch": 7.306449346634818,
"grad_norm": 0.8857362270355225,
"learning_rate": 7.28205869596646e-05,
"loss": 1.658,
"step": 104000
},
{
"epoch": 7.341576506955178,
"grad_norm": 0.8846690654754639,
"learning_rate": 7.267601561370536e-05,
"loss": 1.6543,
"step": 104500
},
{
"epoch": 7.376703667275537,
"grad_norm": 0.8960644006729126,
"learning_rate": 7.253144426774614e-05,
"loss": 1.6574,
"step": 105000
},
{
"epoch": 7.411830827595897,
"grad_norm": 0.9093925356864929,
"learning_rate": 7.23868729217869e-05,
"loss": 1.6602,
"step": 105500
},
{
"epoch": 7.446957987916257,
"grad_norm": 0.8878596425056458,
"learning_rate": 7.224230157582768e-05,
"loss": 1.6583,
"step": 106000
},
{
"epoch": 7.482085148236616,
"grad_norm": 0.8982939124107361,
"learning_rate": 7.209773022986845e-05,
"loss": 1.6542,
"step": 106500
},
{
"epoch": 7.517212308556976,
"grad_norm": 0.931877613067627,
"learning_rate": 7.195315888390921e-05,
"loss": 1.6513,
"step": 107000
},
{
"epoch": 7.552339468877336,
"grad_norm": 0.9102097749710083,
"learning_rate": 7.180858753794997e-05,
"loss": 1.6505,
"step": 107500
},
{
"epoch": 7.587466629197696,
"grad_norm": 0.8833571076393127,
"learning_rate": 7.166401619199075e-05,
"loss": 1.6506,
"step": 108000
},
{
"epoch": 7.6225937895180556,
"grad_norm": 0.8974440693855286,
"learning_rate": 7.151944484603153e-05,
"loss": 1.6529,
"step": 108500
},
{
"epoch": 7.657720949838415,
"grad_norm": 0.8859068155288696,
"learning_rate": 7.13748735000723e-05,
"loss": 1.6483,
"step": 109000
},
{
"epoch": 7.692848110158775,
"grad_norm": 0.9290302991867065,
"learning_rate": 7.123030215411306e-05,
"loss": 1.649,
"step": 109500
},
{
"epoch": 7.727975270479135,
"grad_norm": 0.912346601486206,
"learning_rate": 7.108573080815382e-05,
"loss": 1.6497,
"step": 110000
},
{
"epoch": 7.727975270479135,
"eval_accuracy": 0.696107465901776,
"eval_loss": 1.3987531661987305,
"eval_runtime": 145.5941,
"eval_samples_per_second": 829.058,
"eval_steps_per_second": 5.186,
"step": 110000
},
{
"epoch": 7.763102430799494,
"grad_norm": 0.9191212058067322,
"learning_rate": 7.09411594621946e-05,
"loss": 1.6456,
"step": 110500
},
{
"epoch": 7.798229591119854,
"grad_norm": 0.9030919075012207,
"learning_rate": 7.079658811623536e-05,
"loss": 1.6428,
"step": 111000
},
{
"epoch": 7.833356751440213,
"grad_norm": 0.9317150115966797,
"learning_rate": 7.065201677027613e-05,
"loss": 1.6428,
"step": 111500
},
{
"epoch": 7.868483911760573,
"grad_norm": 0.9374363422393799,
"learning_rate": 7.05074454243169e-05,
"loss": 1.6444,
"step": 112000
},
{
"epoch": 7.903611072080933,
"grad_norm": 0.8871532678604126,
"learning_rate": 7.036287407835768e-05,
"loss": 1.6428,
"step": 112500
},
{
"epoch": 7.938738232401293,
"grad_norm": 0.947810709476471,
"learning_rate": 7.021830273239845e-05,
"loss": 1.6366,
"step": 113000
},
{
"epoch": 7.9738653927216525,
"grad_norm": 0.8954718708992004,
"learning_rate": 7.007373138643921e-05,
"loss": 1.6379,
"step": 113500
},
{
"epoch": 8.008992553042011,
"grad_norm": 0.8930262327194214,
"learning_rate": 6.992916004047997e-05,
"loss": 1.6387,
"step": 114000
},
{
"epoch": 8.044119713362372,
"grad_norm": 0.9441146850585938,
"learning_rate": 6.978458869452075e-05,
"loss": 1.6332,
"step": 114500
},
{
"epoch": 8.079246873682731,
"grad_norm": 0.9339563846588135,
"learning_rate": 6.964001734856152e-05,
"loss": 1.6326,
"step": 115000
},
{
"epoch": 8.114374034003092,
"grad_norm": 0.917434811592102,
"learning_rate": 6.949544600260228e-05,
"loss": 1.6311,
"step": 115500
},
{
"epoch": 8.14950119432345,
"grad_norm": 0.9269524216651917,
"learning_rate": 6.935087465664306e-05,
"loss": 1.6327,
"step": 116000
},
{
"epoch": 8.184628354643811,
"grad_norm": 0.9198341965675354,
"learning_rate": 6.920630331068383e-05,
"loss": 1.6323,
"step": 116500
},
{
"epoch": 8.21975551496417,
"grad_norm": 0.9190070629119873,
"learning_rate": 6.90617319647246e-05,
"loss": 1.6286,
"step": 117000
},
{
"epoch": 8.25488267528453,
"grad_norm": 0.9343723654747009,
"learning_rate": 6.891716061876536e-05,
"loss": 1.6291,
"step": 117500
},
{
"epoch": 8.29000983560489,
"grad_norm": 0.9313236474990845,
"learning_rate": 6.877258927280613e-05,
"loss": 1.6253,
"step": 118000
},
{
"epoch": 8.325136995925249,
"grad_norm": 0.9170568585395813,
"learning_rate": 6.86280179268469e-05,
"loss": 1.6244,
"step": 118500
},
{
"epoch": 8.36026415624561,
"grad_norm": 0.9194092154502869,
"learning_rate": 6.848344658088767e-05,
"loss": 1.6251,
"step": 119000
},
{
"epoch": 8.395391316565968,
"grad_norm": 0.903550922870636,
"learning_rate": 6.833887523492845e-05,
"loss": 1.6245,
"step": 119500
},
{
"epoch": 8.430518476886329,
"grad_norm": 0.9155687689781189,
"learning_rate": 6.819430388896921e-05,
"loss": 1.6245,
"step": 120000
},
{
"epoch": 8.430518476886329,
"eval_accuracy": 0.699622936687572,
"eval_loss": 1.376625895500183,
"eval_runtime": 146.4286,
"eval_samples_per_second": 824.334,
"eval_steps_per_second": 5.156,
"step": 120000
},
{
"epoch": 8.465645637206688,
"grad_norm": 0.9170736074447632,
"learning_rate": 6.804973254300999e-05,
"loss": 1.6241,
"step": 120500
},
{
"epoch": 8.500772797527048,
"grad_norm": 0.918855607509613,
"learning_rate": 6.790516119705075e-05,
"loss": 1.6238,
"step": 121000
},
{
"epoch": 8.535899957847407,
"grad_norm": 0.9043271541595459,
"learning_rate": 6.776058985109152e-05,
"loss": 1.6253,
"step": 121500
},
{
"epoch": 8.571027118167768,
"grad_norm": 0.9378098845481873,
"learning_rate": 6.761601850513228e-05,
"loss": 1.6207,
"step": 122000
},
{
"epoch": 8.606154278488127,
"grad_norm": 0.9023655652999878,
"learning_rate": 6.747144715917306e-05,
"loss": 1.6192,
"step": 122500
},
{
"epoch": 8.641281438808488,
"grad_norm": 0.90032958984375,
"learning_rate": 6.732687581321382e-05,
"loss": 1.6188,
"step": 123000
},
{
"epoch": 8.676408599128846,
"grad_norm": 0.9296461939811707,
"learning_rate": 6.71823044672546e-05,
"loss": 1.6147,
"step": 123500
},
{
"epoch": 8.711535759449205,
"grad_norm": 0.917468249797821,
"learning_rate": 6.703773312129536e-05,
"loss": 1.614,
"step": 124000
},
{
"epoch": 8.746662919769566,
"grad_norm": 0.9283198714256287,
"learning_rate": 6.689316177533613e-05,
"loss": 1.6088,
"step": 124500
},
{
"epoch": 8.781790080089925,
"grad_norm": 0.9109067916870117,
"learning_rate": 6.67485904293769e-05,
"loss": 1.6148,
"step": 125000
},
{
"epoch": 8.816917240410286,
"grad_norm": 0.9425482153892517,
"learning_rate": 6.660401908341767e-05,
"loss": 1.6125,
"step": 125500
},
{
"epoch": 8.852044400730644,
"grad_norm": 0.9108548164367676,
"learning_rate": 6.645944773745843e-05,
"loss": 1.6118,
"step": 126000
},
{
"epoch": 8.887171561051005,
"grad_norm": 0.9401180744171143,
"learning_rate": 6.631487639149921e-05,
"loss": 1.6121,
"step": 126500
},
{
"epoch": 8.922298721371364,
"grad_norm": 0.9434899091720581,
"learning_rate": 6.617030504553999e-05,
"loss": 1.6099,
"step": 127000
},
{
"epoch": 8.957425881691725,
"grad_norm": 0.9705188870429993,
"learning_rate": 6.602573369958075e-05,
"loss": 1.6143,
"step": 127500
},
{
"epoch": 8.992553042012084,
"grad_norm": 0.9429416060447693,
"learning_rate": 6.588116235362152e-05,
"loss": 1.6131,
"step": 128000
},
{
"epoch": 9.027680202332444,
"grad_norm": 0.9325101971626282,
"learning_rate": 6.573659100766228e-05,
"loss": 1.6054,
"step": 128500
},
{
"epoch": 9.062807362652803,
"grad_norm": 0.9009438157081604,
"learning_rate": 6.559201966170306e-05,
"loss": 1.6049,
"step": 129000
},
{
"epoch": 9.097934522973162,
"grad_norm": 0.9559112191200256,
"learning_rate": 6.544744831574382e-05,
"loss": 1.6065,
"step": 129500
},
{
"epoch": 9.133061683293523,
"grad_norm": 0.9075998067855835,
"learning_rate": 6.530287696978458e-05,
"loss": 1.6015,
"step": 130000
},
{
"epoch": 9.133061683293523,
"eval_accuracy": 0.701941754455614,
"eval_loss": 1.362774133682251,
"eval_runtime": 146.6899,
"eval_samples_per_second": 822.865,
"eval_steps_per_second": 5.147,
"step": 130000
},
{
"epoch": 9.168188843613882,
"grad_norm": 0.9487557411193848,
"learning_rate": 6.515830562382536e-05,
"loss": 1.603,
"step": 130500
},
{
"epoch": 9.203316003934242,
"grad_norm": 0.9400453567504883,
"learning_rate": 6.501373427786614e-05,
"loss": 1.6042,
"step": 131000
},
{
"epoch": 9.238443164254601,
"grad_norm": 0.9347612261772156,
"learning_rate": 6.48691629319069e-05,
"loss": 1.6033,
"step": 131500
},
{
"epoch": 9.273570324574962,
"grad_norm": 0.9069722890853882,
"learning_rate": 6.472459158594767e-05,
"loss": 1.6038,
"step": 132000
},
{
"epoch": 9.30869748489532,
"grad_norm": 0.9513418674468994,
"learning_rate": 6.458002023998843e-05,
"loss": 1.6023,
"step": 132500
},
{
"epoch": 9.343824645215681,
"grad_norm": 0.9263851642608643,
"learning_rate": 6.443544889402921e-05,
"loss": 1.5997,
"step": 133000
},
{
"epoch": 9.37895180553604,
"grad_norm": 0.9571240544319153,
"learning_rate": 6.429087754806997e-05,
"loss": 1.5987,
"step": 133500
},
{
"epoch": 9.414078965856401,
"grad_norm": 0.9037933349609375,
"learning_rate": 6.414630620211074e-05,
"loss": 1.5979,
"step": 134000
},
{
"epoch": 9.44920612617676,
"grad_norm": 0.9799155592918396,
"learning_rate": 6.400173485615151e-05,
"loss": 1.5952,
"step": 134500
},
{
"epoch": 9.484333286497119,
"grad_norm": 0.9275627732276917,
"learning_rate": 6.385716351019229e-05,
"loss": 1.595,
"step": 135000
},
{
"epoch": 9.51946044681748,
"grad_norm": 0.9277747273445129,
"learning_rate": 6.371259216423306e-05,
"loss": 1.5961,
"step": 135500
},
{
"epoch": 9.554587607137838,
"grad_norm": 0.9405021071434021,
"learning_rate": 6.356802081827382e-05,
"loss": 1.5955,
"step": 136000
},
{
"epoch": 9.589714767458199,
"grad_norm": 0.9006831049919128,
"learning_rate": 6.342344947231458e-05,
"loss": 1.5967,
"step": 136500
},
{
"epoch": 9.624841927778558,
"grad_norm": 0.9479206800460815,
"learning_rate": 6.327887812635536e-05,
"loss": 1.5923,
"step": 137000
},
{
"epoch": 9.659969088098919,
"grad_norm": 0.9280245304107666,
"learning_rate": 6.313430678039613e-05,
"loss": 1.5932,
"step": 137500
},
{
"epoch": 9.695096248419278,
"grad_norm": 0.9124751091003418,
"learning_rate": 6.29897354344369e-05,
"loss": 1.5882,
"step": 138000
},
{
"epoch": 9.730223408739638,
"grad_norm": 0.9291515946388245,
"learning_rate": 6.284516408847767e-05,
"loss": 1.5948,
"step": 138500
},
{
"epoch": 9.765350569059997,
"grad_norm": 0.9335107207298279,
"learning_rate": 6.270059274251843e-05,
"loss": 1.5896,
"step": 139000
},
{
"epoch": 9.800477729380358,
"grad_norm": 0.9176722764968872,
"learning_rate": 6.255602139655921e-05,
"loss": 1.5901,
"step": 139500
},
{
"epoch": 9.835604889700717,
"grad_norm": 0.9138041138648987,
"learning_rate": 6.241145005059997e-05,
"loss": 1.5882,
"step": 140000
},
{
"epoch": 9.835604889700717,
"eval_accuracy": 0.7051619721997767,
"eval_loss": 1.3451346158981323,
"eval_runtime": 145.0544,
"eval_samples_per_second": 832.143,
"eval_steps_per_second": 5.205,
"step": 140000
},
{
"epoch": 9.870732050021076,
"grad_norm": 0.9771543741226196,
"learning_rate": 6.226687870464074e-05,
"loss": 1.5882,
"step": 140500
},
{
"epoch": 9.905859210341436,
"grad_norm": 0.9482930302619934,
"learning_rate": 6.212230735868151e-05,
"loss": 1.5866,
"step": 141000
},
{
"epoch": 9.940986370661795,
"grad_norm": 0.9545276761054993,
"learning_rate": 6.197773601272228e-05,
"loss": 1.5868,
"step": 141500
},
{
"epoch": 9.976113530982156,
"grad_norm": 0.962986171245575,
"learning_rate": 6.183316466676306e-05,
"loss": 1.5833,
"step": 142000
},
{
"epoch": 10.011240691302515,
"grad_norm": 0.9559778571128845,
"learning_rate": 6.168859332080382e-05,
"loss": 1.5858,
"step": 142500
},
{
"epoch": 10.046367851622875,
"grad_norm": 0.9056123495101929,
"learning_rate": 6.154402197484458e-05,
"loss": 1.5855,
"step": 143000
},
{
"epoch": 10.081495011943234,
"grad_norm": 0.9433762431144714,
"learning_rate": 6.139945062888536e-05,
"loss": 1.581,
"step": 143500
},
{
"epoch": 10.116622172263595,
"grad_norm": 0.9369404315948486,
"learning_rate": 6.125487928292613e-05,
"loss": 1.5795,
"step": 144000
},
{
"epoch": 10.151749332583954,
"grad_norm": 1.009413480758667,
"learning_rate": 6.111030793696689e-05,
"loss": 1.5793,
"step": 144500
},
{
"epoch": 10.186876492904315,
"grad_norm": 0.9554888606071472,
"learning_rate": 6.096573659100766e-05,
"loss": 1.5797,
"step": 145000
},
{
"epoch": 10.222003653224673,
"grad_norm": 0.9255361557006836,
"learning_rate": 6.082116524504844e-05,
"loss": 1.5809,
"step": 145500
},
{
"epoch": 10.257130813545032,
"grad_norm": 0.9112059473991394,
"learning_rate": 6.067659389908921e-05,
"loss": 1.5813,
"step": 146000
},
{
"epoch": 10.292257973865393,
"grad_norm": 0.9442374110221863,
"learning_rate": 6.053202255312997e-05,
"loss": 1.5757,
"step": 146500
},
{
"epoch": 10.327385134185752,
"grad_norm": 0.9561610221862793,
"learning_rate": 6.0387451207170744e-05,
"loss": 1.5765,
"step": 147000
},
{
"epoch": 10.362512294506113,
"grad_norm": 0.9480382800102234,
"learning_rate": 6.024287986121151e-05,
"loss": 1.5769,
"step": 147500
},
{
"epoch": 10.397639454826471,
"grad_norm": 0.9375914931297302,
"learning_rate": 6.009830851525228e-05,
"loss": 1.5744,
"step": 148000
},
{
"epoch": 10.432766615146832,
"grad_norm": 0.9849138259887695,
"learning_rate": 5.995373716929304e-05,
"loss": 1.5778,
"step": 148500
},
{
"epoch": 10.467893775467191,
"grad_norm": 0.9578582048416138,
"learning_rate": 5.980916582333381e-05,
"loss": 1.5729,
"step": 149000
},
{
"epoch": 10.503020935787552,
"grad_norm": 0.9367204308509827,
"learning_rate": 5.966459447737459e-05,
"loss": 1.5714,
"step": 149500
},
{
"epoch": 10.53814809610791,
"grad_norm": 0.9462071061134338,
"learning_rate": 5.952002313141536e-05,
"loss": 1.5738,
"step": 150000
},
{
"epoch": 10.53814809610791,
"eval_accuracy": 0.7075515620989917,
"eval_loss": 1.331017255783081,
"eval_runtime": 146.1625,
"eval_samples_per_second": 825.834,
"eval_steps_per_second": 5.165,
"step": 150000
},
{
"epoch": 10.573275256428271,
"grad_norm": 0.9778929948806763,
"learning_rate": 5.9375451785456126e-05,
"loss": 1.5759,
"step": 150500
},
{
"epoch": 10.60840241674863,
"grad_norm": 0.9483457803726196,
"learning_rate": 5.9230880439496896e-05,
"loss": 1.5711,
"step": 151000
},
{
"epoch": 10.643529577068989,
"grad_norm": 0.9661710262298584,
"learning_rate": 5.908630909353766e-05,
"loss": 1.5713,
"step": 151500
},
{
"epoch": 10.67865673738935,
"grad_norm": 0.926715612411499,
"learning_rate": 5.894173774757843e-05,
"loss": 1.5706,
"step": 152000
},
{
"epoch": 10.713783897709709,
"grad_norm": 0.9615681767463684,
"learning_rate": 5.8797166401619195e-05,
"loss": 1.5685,
"step": 152500
},
{
"epoch": 10.74891105803007,
"grad_norm": 0.9441356658935547,
"learning_rate": 5.865259505565997e-05,
"loss": 1.5689,
"step": 153000
},
{
"epoch": 10.784038218350428,
"grad_norm": 0.9553456902503967,
"learning_rate": 5.8508023709700743e-05,
"loss": 1.5708,
"step": 153500
},
{
"epoch": 10.819165378670789,
"grad_norm": 0.9203099012374878,
"learning_rate": 5.8363452363741514e-05,
"loss": 1.5691,
"step": 154000
},
{
"epoch": 10.854292538991148,
"grad_norm": 0.9613307118415833,
"learning_rate": 5.821888101778228e-05,
"loss": 1.5691,
"step": 154500
},
{
"epoch": 10.889419699311508,
"grad_norm": 0.93576580286026,
"learning_rate": 5.807430967182305e-05,
"loss": 1.5683,
"step": 155000
},
{
"epoch": 10.924546859631867,
"grad_norm": 0.959653913974762,
"learning_rate": 5.792973832586381e-05,
"loss": 1.5685,
"step": 155500
},
{
"epoch": 10.959674019952226,
"grad_norm": 0.9653539657592773,
"learning_rate": 5.7785166979904584e-05,
"loss": 1.5662,
"step": 156000
},
{
"epoch": 10.994801180272587,
"grad_norm": 0.9286296367645264,
"learning_rate": 5.764059563394536e-05,
"loss": 1.5656,
"step": 156500
},
{
"epoch": 11.029928340592946,
"grad_norm": 0.9834336042404175,
"learning_rate": 5.7496024287986125e-05,
"loss": 1.5588,
"step": 157000
},
{
"epoch": 11.065055500913306,
"grad_norm": 0.9424375295639038,
"learning_rate": 5.7351452942026896e-05,
"loss": 1.562,
"step": 157500
},
{
"epoch": 11.100182661233665,
"grad_norm": 0.9713384509086609,
"learning_rate": 5.720688159606766e-05,
"loss": 1.562,
"step": 158000
},
{
"epoch": 11.135309821554026,
"grad_norm": 0.9827730655670166,
"learning_rate": 5.706231025010843e-05,
"loss": 1.5595,
"step": 158500
},
{
"epoch": 11.170436981874385,
"grad_norm": 0.969085693359375,
"learning_rate": 5.6917738904149195e-05,
"loss": 1.5606,
"step": 159000
},
{
"epoch": 11.205564142194746,
"grad_norm": 0.9821890592575073,
"learning_rate": 5.6773167558189966e-05,
"loss": 1.5607,
"step": 159500
},
{
"epoch": 11.240691302515105,
"grad_norm": 0.9560614824295044,
"learning_rate": 5.6628596212230736e-05,
"loss": 1.563,
"step": 160000
},
{
"epoch": 11.240691302515105,
"eval_accuracy": 0.7090796716396742,
"eval_loss": 1.3214398622512817,
"eval_runtime": 147.4505,
"eval_samples_per_second": 818.621,
"eval_steps_per_second": 5.12,
"step": 160000
},
{
"epoch": 11.275818462835465,
"grad_norm": 0.9682839512825012,
"learning_rate": 5.6484024866271514e-05,
"loss": 1.558,
"step": 160500
},
{
"epoch": 11.310945623155824,
"grad_norm": 0.958651065826416,
"learning_rate": 5.633945352031228e-05,
"loss": 1.5618,
"step": 161000
},
{
"epoch": 11.346072783476183,
"grad_norm": 0.9705203771591187,
"learning_rate": 5.619488217435305e-05,
"loss": 1.5576,
"step": 161500
},
{
"epoch": 11.381199943796544,
"grad_norm": 0.9865082502365112,
"learning_rate": 5.605031082839381e-05,
"loss": 1.5583,
"step": 162000
},
{
"epoch": 11.416327104116903,
"grad_norm": 0.9645934700965881,
"learning_rate": 5.5905739482434584e-05,
"loss": 1.5551,
"step": 162500
},
{
"epoch": 11.451454264437263,
"grad_norm": 0.9389051795005798,
"learning_rate": 5.576116813647535e-05,
"loss": 1.5521,
"step": 163000
},
{
"epoch": 11.486581424757622,
"grad_norm": 0.9918972849845886,
"learning_rate": 5.561659679051612e-05,
"loss": 1.5551,
"step": 163500
},
{
"epoch": 11.521708585077983,
"grad_norm": 0.96946120262146,
"learning_rate": 5.5472025444556896e-05,
"loss": 1.5554,
"step": 164000
},
{
"epoch": 11.556835745398342,
"grad_norm": 0.9327898621559143,
"learning_rate": 5.532745409859767e-05,
"loss": 1.5536,
"step": 164500
},
{
"epoch": 11.591962905718702,
"grad_norm": 0.9520469307899475,
"learning_rate": 5.518288275263843e-05,
"loss": 1.554,
"step": 165000
},
{
"epoch": 11.627090066039061,
"grad_norm": 0.9468493461608887,
"learning_rate": 5.50383114066792e-05,
"loss": 1.5541,
"step": 165500
},
{
"epoch": 11.662217226359422,
"grad_norm": 0.9519227147102356,
"learning_rate": 5.4893740060719965e-05,
"loss": 1.551,
"step": 166000
},
{
"epoch": 11.69734438667978,
"grad_norm": 0.9576557278633118,
"learning_rate": 5.4749168714760736e-05,
"loss": 1.5519,
"step": 166500
},
{
"epoch": 11.73247154700014,
"grad_norm": 0.9346196055412292,
"learning_rate": 5.46045973688015e-05,
"loss": 1.549,
"step": 167000
},
{
"epoch": 11.7675987073205,
"grad_norm": 0.964485764503479,
"learning_rate": 5.446002602284227e-05,
"loss": 1.5494,
"step": 167500
},
{
"epoch": 11.80272586764086,
"grad_norm": 0.9761068820953369,
"learning_rate": 5.431545467688305e-05,
"loss": 1.551,
"step": 168000
},
{
"epoch": 11.83785302796122,
"grad_norm": 0.9633313417434692,
"learning_rate": 5.417088333092382e-05,
"loss": 1.5475,
"step": 168500
},
{
"epoch": 11.872980188281579,
"grad_norm": 0.9491254091262817,
"learning_rate": 5.402631198496458e-05,
"loss": 1.5468,
"step": 169000
},
{
"epoch": 11.90810734860194,
"grad_norm": 0.9572640061378479,
"learning_rate": 5.3881740639005354e-05,
"loss": 1.5463,
"step": 169500
},
{
"epoch": 11.943234508922298,
"grad_norm": 0.9447980523109436,
"learning_rate": 5.373716929304612e-05,
"loss": 1.5473,
"step": 170000
},
{
"epoch": 11.943234508922298,
"eval_accuracy": 0.7113065905004013,
"eval_loss": 1.308741807937622,
"eval_runtime": 147.0083,
"eval_samples_per_second": 821.083,
"eval_steps_per_second": 5.136,
"step": 170000
},
{
"epoch": 11.978361669242659,
"grad_norm": 0.9711387753486633,
"learning_rate": 5.359259794708689e-05,
"loss": 1.5469,
"step": 170500
},
{
"epoch": 12.013488829563018,
"grad_norm": 0.9664705395698547,
"learning_rate": 5.344802660112765e-05,
"loss": 1.547,
"step": 171000
},
{
"epoch": 12.048615989883379,
"grad_norm": 0.9575341939926147,
"learning_rate": 5.330345525516843e-05,
"loss": 1.5439,
"step": 171500
},
{
"epoch": 12.083743150203738,
"grad_norm": 0.9776381850242615,
"learning_rate": 5.31588839092092e-05,
"loss": 1.5416,
"step": 172000
},
{
"epoch": 12.118870310524096,
"grad_norm": 0.9849351048469543,
"learning_rate": 5.3014312563249965e-05,
"loss": 1.544,
"step": 172500
},
{
"epoch": 12.153997470844457,
"grad_norm": 1.0078338384628296,
"learning_rate": 5.2869741217290736e-05,
"loss": 1.5463,
"step": 173000
},
{
"epoch": 12.189124631164816,
"grad_norm": 0.9730892777442932,
"learning_rate": 5.272516987133151e-05,
"loss": 1.5424,
"step": 173500
},
{
"epoch": 12.224251791485177,
"grad_norm": 1.0085258483886719,
"learning_rate": 5.258059852537227e-05,
"loss": 1.5384,
"step": 174000
},
{
"epoch": 12.259378951805536,
"grad_norm": 0.9556441307067871,
"learning_rate": 5.243602717941304e-05,
"loss": 1.5427,
"step": 174500
},
{
"epoch": 12.294506112125896,
"grad_norm": 0.9524454474449158,
"learning_rate": 5.229145583345382e-05,
"loss": 1.5373,
"step": 175000
},
{
"epoch": 12.329633272446255,
"grad_norm": 0.9806123375892639,
"learning_rate": 5.214688448749458e-05,
"loss": 1.5425,
"step": 175500
},
{
"epoch": 12.364760432766616,
"grad_norm": 0.9411491751670837,
"learning_rate": 5.2002313141535354e-05,
"loss": 1.5394,
"step": 176000
},
{
"epoch": 12.399887593086975,
"grad_norm": 1.0032098293304443,
"learning_rate": 5.185774179557612e-05,
"loss": 1.5388,
"step": 176500
},
{
"epoch": 12.435014753407334,
"grad_norm": 0.9663287401199341,
"learning_rate": 5.171317044961689e-05,
"loss": 1.5389,
"step": 177000
},
{
"epoch": 12.470141913727694,
"grad_norm": 0.9497302770614624,
"learning_rate": 5.156859910365765e-05,
"loss": 1.537,
"step": 177500
},
{
"epoch": 12.505269074048053,
"grad_norm": 0.9728620648384094,
"learning_rate": 5.1424027757698423e-05,
"loss": 1.5397,
"step": 178000
},
{
"epoch": 12.540396234368414,
"grad_norm": 0.9789900779724121,
"learning_rate": 5.127945641173919e-05,
"loss": 1.5405,
"step": 178500
},
{
"epoch": 12.575523394688773,
"grad_norm": 0.9774718284606934,
"learning_rate": 5.113488506577997e-05,
"loss": 1.5368,
"step": 179000
},
{
"epoch": 12.610650555009133,
"grad_norm": 0.9627470970153809,
"learning_rate": 5.0990313719820736e-05,
"loss": 1.5365,
"step": 179500
},
{
"epoch": 12.645777715329492,
"grad_norm": 0.9805966019630432,
"learning_rate": 5.0845742373861507e-05,
"loss": 1.5364,
"step": 180000
},
{
"epoch": 12.645777715329492,
"eval_accuracy": 0.7135348973126568,
"eval_loss": 1.2944310903549194,
"eval_runtime": 147.0212,
"eval_samples_per_second": 821.011,
"eval_steps_per_second": 5.135,
"step": 180000
},
{
"epoch": 12.680904875649853,
"grad_norm": 1.0078229904174805,
"learning_rate": 5.070117102790227e-05,
"loss": 1.5344,
"step": 180500
},
{
"epoch": 12.716032035970212,
"grad_norm": 1.0073761940002441,
"learning_rate": 5.055659968194304e-05,
"loss": 1.5357,
"step": 181000
},
{
"epoch": 12.751159196290573,
"grad_norm": 0.9848378300666809,
"learning_rate": 5.0412028335983805e-05,
"loss": 1.5352,
"step": 181500
},
{
"epoch": 12.786286356610931,
"grad_norm": 0.9649278521537781,
"learning_rate": 5.0267456990024576e-05,
"loss": 1.5356,
"step": 182000
},
{
"epoch": 12.82141351693129,
"grad_norm": 0.9573400616645813,
"learning_rate": 5.0122885644065354e-05,
"loss": 1.5363,
"step": 182500
},
{
"epoch": 12.856540677251651,
"grad_norm": 0.9978417754173279,
"learning_rate": 4.997831429810612e-05,
"loss": 1.5338,
"step": 183000
},
{
"epoch": 12.89166783757201,
"grad_norm": 0.99289870262146,
"learning_rate": 4.983374295214689e-05,
"loss": 1.5328,
"step": 183500
},
{
"epoch": 12.92679499789237,
"grad_norm": 0.974976122379303,
"learning_rate": 4.968917160618766e-05,
"loss": 1.5298,
"step": 184000
},
{
"epoch": 12.96192215821273,
"grad_norm": 0.9644068479537964,
"learning_rate": 4.954460026022842e-05,
"loss": 1.5341,
"step": 184500
},
{
"epoch": 12.99704931853309,
"grad_norm": 0.9787358641624451,
"learning_rate": 4.9400028914269194e-05,
"loss": 1.532,
"step": 185000
},
{
"epoch": 13.032176478853449,
"grad_norm": 0.9854114651679993,
"learning_rate": 4.9255457568309965e-05,
"loss": 1.5277,
"step": 185500
},
{
"epoch": 13.06730363917381,
"grad_norm": 0.9944697618484497,
"learning_rate": 4.9110886222350736e-05,
"loss": 1.5284,
"step": 186000
},
{
"epoch": 13.102430799494169,
"grad_norm": 0.9880611300468445,
"learning_rate": 4.89663148763915e-05,
"loss": 1.5284,
"step": 186500
},
{
"epoch": 13.13755795981453,
"grad_norm": 0.9830440282821655,
"learning_rate": 4.882174353043227e-05,
"loss": 1.5251,
"step": 187000
},
{
"epoch": 13.172685120134888,
"grad_norm": 0.992473304271698,
"learning_rate": 4.867717218447304e-05,
"loss": 1.5256,
"step": 187500
},
{
"epoch": 13.207812280455247,
"grad_norm": 0.948597252368927,
"learning_rate": 4.853260083851381e-05,
"loss": 1.5271,
"step": 188000
},
{
"epoch": 13.242939440775608,
"grad_norm": 0.9829763174057007,
"learning_rate": 4.8388029492554576e-05,
"loss": 1.5259,
"step": 188500
},
{
"epoch": 13.278066601095967,
"grad_norm": 0.9461808800697327,
"learning_rate": 4.824345814659535e-05,
"loss": 1.525,
"step": 189000
},
{
"epoch": 13.313193761416327,
"grad_norm": 0.9874758720397949,
"learning_rate": 4.809888680063612e-05,
"loss": 1.5286,
"step": 189500
},
{
"epoch": 13.348320921736686,
"grad_norm": 0.9801004528999329,
"learning_rate": 4.795431545467688e-05,
"loss": 1.5257,
"step": 190000
},
{
"epoch": 13.348320921736686,
"eval_accuracy": 0.7146395658479376,
"eval_loss": 1.2904942035675049,
"eval_runtime": 148.7015,
"eval_samples_per_second": 811.734,
"eval_steps_per_second": 5.077,
"step": 190000
},
{
"epoch": 13.383448082057047,
"grad_norm": 1.0048809051513672,
"learning_rate": 4.780974410871765e-05,
"loss": 1.5222,
"step": 190500
},
{
"epoch": 13.418575242377406,
"grad_norm": 0.9752638339996338,
"learning_rate": 4.766517276275842e-05,
"loss": 1.5242,
"step": 191000
},
{
"epoch": 13.453702402697767,
"grad_norm": 0.9492470622062683,
"learning_rate": 4.7520601416799194e-05,
"loss": 1.5219,
"step": 191500
},
{
"epoch": 13.488829563018125,
"grad_norm": 0.973146915435791,
"learning_rate": 4.737603007083996e-05,
"loss": 1.5234,
"step": 192000
},
{
"epoch": 13.523956723338486,
"grad_norm": 0.9545627236366272,
"learning_rate": 4.723145872488073e-05,
"loss": 1.5252,
"step": 192500
},
{
"epoch": 13.559083883658845,
"grad_norm": 0.9910150170326233,
"learning_rate": 4.70868873789215e-05,
"loss": 1.5229,
"step": 193000
},
{
"epoch": 13.594211043979204,
"grad_norm": 1.0169930458068848,
"learning_rate": 4.694231603296227e-05,
"loss": 1.5233,
"step": 193500
},
{
"epoch": 13.629338204299565,
"grad_norm": 1.0121934413909912,
"learning_rate": 4.6797744687003034e-05,
"loss": 1.5201,
"step": 194000
},
{
"epoch": 13.664465364619923,
"grad_norm": 0.9987900853157043,
"learning_rate": 4.665317334104381e-05,
"loss": 1.5187,
"step": 194500
},
{
"epoch": 13.699592524940284,
"grad_norm": 1.0241477489471436,
"learning_rate": 4.6508601995084576e-05,
"loss": 1.5206,
"step": 195000
},
{
"epoch": 13.734719685260643,
"grad_norm": 0.9895103573799133,
"learning_rate": 4.6364030649125346e-05,
"loss": 1.5201,
"step": 195500
},
{
"epoch": 13.769846845581004,
"grad_norm": 0.9821479916572571,
"learning_rate": 4.621945930316611e-05,
"loss": 1.5201,
"step": 196000
},
{
"epoch": 13.804974005901363,
"grad_norm": 1.005090594291687,
"learning_rate": 4.607488795720689e-05,
"loss": 1.5182,
"step": 196500
},
{
"epoch": 13.840101166221723,
"grad_norm": 0.9969967603683472,
"learning_rate": 4.593031661124765e-05,
"loss": 1.5195,
"step": 197000
},
{
"epoch": 13.875228326542082,
"grad_norm": 0.9817072153091431,
"learning_rate": 4.578574526528842e-05,
"loss": 1.5175,
"step": 197500
},
{
"epoch": 13.910355486862443,
"grad_norm": 0.9905675053596497,
"learning_rate": 4.564117391932919e-05,
"loss": 1.5182,
"step": 198000
},
{
"epoch": 13.945482647182802,
"grad_norm": 0.9947062134742737,
"learning_rate": 4.5496602573369964e-05,
"loss": 1.5162,
"step": 198500
},
{
"epoch": 13.98060980750316,
"grad_norm": 0.97999107837677,
"learning_rate": 4.535203122741073e-05,
"loss": 1.5216,
"step": 199000
},
{
"epoch": 14.015736967823521,
"grad_norm": 0.9755523204803467,
"learning_rate": 4.52074598814515e-05,
"loss": 1.5161,
"step": 199500
},
{
"epoch": 14.05086412814388,
"grad_norm": 0.9752248525619507,
"learning_rate": 4.506288853549227e-05,
"loss": 1.5164,
"step": 200000
},
{
"epoch": 14.05086412814388,
"eval_accuracy": 0.716148456990617,
"eval_loss": 1.2788549661636353,
"eval_runtime": 145.5914,
"eval_samples_per_second": 829.074,
"eval_steps_per_second": 5.186,
"step": 200000
},
{
"epoch": 14.08599128846424,
"grad_norm": 0.984459638595581,
"learning_rate": 4.491831718953304e-05,
"loss": 1.5128,
"step": 200500
},
{
"epoch": 14.1211184487846,
"grad_norm": 0.9917986392974854,
"learning_rate": 4.4773745843573805e-05,
"loss": 1.5142,
"step": 201000
},
{
"epoch": 14.15624560910496,
"grad_norm": 0.9588636159896851,
"learning_rate": 4.4629174497614575e-05,
"loss": 1.5122,
"step": 201500
},
{
"epoch": 14.19137276942532,
"grad_norm": 0.9958001971244812,
"learning_rate": 4.4484603151655346e-05,
"loss": 1.5105,
"step": 202000
},
{
"epoch": 14.22649992974568,
"grad_norm": 1.012403130531311,
"learning_rate": 4.434003180569612e-05,
"loss": 1.5129,
"step": 202500
},
{
"epoch": 14.261627090066039,
"grad_norm": 1.0305534601211548,
"learning_rate": 4.419546045973688e-05,
"loss": 1.5133,
"step": 203000
},
{
"epoch": 14.2967542503864,
"grad_norm": 0.9992120265960693,
"learning_rate": 4.405088911377765e-05,
"loss": 1.5114,
"step": 203500
},
{
"epoch": 14.331881410706758,
"grad_norm": 1.008920669555664,
"learning_rate": 4.390631776781842e-05,
"loss": 1.5139,
"step": 204000
},
{
"epoch": 14.367008571027117,
"grad_norm": 1.0051392316818237,
"learning_rate": 4.3761746421859187e-05,
"loss": 1.5112,
"step": 204500
},
{
"epoch": 14.402135731347478,
"grad_norm": 0.9850152730941772,
"learning_rate": 4.361717507589996e-05,
"loss": 1.5111,
"step": 205000
},
{
"epoch": 14.437262891667837,
"grad_norm": 0.9894864559173584,
"learning_rate": 4.347260372994073e-05,
"loss": 1.5119,
"step": 205500
},
{
"epoch": 14.472390051988198,
"grad_norm": 1.0255316495895386,
"learning_rate": 4.33280323839815e-05,
"loss": 1.51,
"step": 206000
},
{
"epoch": 14.507517212308557,
"grad_norm": 1.044765830039978,
"learning_rate": 4.318346103802226e-05,
"loss": 1.5082,
"step": 206500
},
{
"epoch": 14.542644372628917,
"grad_norm": 1.0084095001220703,
"learning_rate": 4.3038889692063034e-05,
"loss": 1.5073,
"step": 207000
},
{
"epoch": 14.577771532949276,
"grad_norm": 0.9992097616195679,
"learning_rate": 4.2894318346103804e-05,
"loss": 1.5085,
"step": 207500
},
{
"epoch": 14.612898693269637,
"grad_norm": 1.016776204109192,
"learning_rate": 4.2749747000144575e-05,
"loss": 1.5092,
"step": 208000
},
{
"epoch": 14.648025853589996,
"grad_norm": 1.0093616247177124,
"learning_rate": 4.260517565418534e-05,
"loss": 1.5074,
"step": 208500
},
{
"epoch": 14.683153013910356,
"grad_norm": 1.0291818380355835,
"learning_rate": 4.246060430822611e-05,
"loss": 1.5079,
"step": 209000
},
{
"epoch": 14.718280174230715,
"grad_norm": 0.9756091833114624,
"learning_rate": 4.231603296226688e-05,
"loss": 1.5071,
"step": 209500
},
{
"epoch": 14.753407334551074,
"grad_norm": 1.0051275491714478,
"learning_rate": 4.217146161630765e-05,
"loss": 1.5071,
"step": 210000
},
{
"epoch": 14.753407334551074,
"eval_accuracy": 0.7175581891473035,
"eval_loss": 1.2701668739318848,
"eval_runtime": 145.5115,
"eval_samples_per_second": 829.529,
"eval_steps_per_second": 5.189,
"step": 210000
},
{
"epoch": 14.788534494871435,
"grad_norm": 1.011240005493164,
"learning_rate": 4.2026890270348416e-05,
"loss": 1.5067,
"step": 210500
},
{
"epoch": 14.823661655191794,
"grad_norm": 1.0316245555877686,
"learning_rate": 4.1882318924389186e-05,
"loss": 1.5059,
"step": 211000
},
{
"epoch": 14.858788815512154,
"grad_norm": 0.9921131730079651,
"learning_rate": 4.173774757842996e-05,
"loss": 1.5065,
"step": 211500
},
{
"epoch": 14.893915975832513,
"grad_norm": 1.0233694314956665,
"learning_rate": 4.159317623247073e-05,
"loss": 1.5039,
"step": 212000
},
{
"epoch": 14.929043136152874,
"grad_norm": 1.0409730672836304,
"learning_rate": 4.144860488651149e-05,
"loss": 1.507,
"step": 212500
},
{
"epoch": 14.964170296473233,
"grad_norm": 1.0045446157455444,
"learning_rate": 4.130403354055227e-05,
"loss": 1.5061,
"step": 213000
},
{
"epoch": 14.999297456793593,
"grad_norm": 0.9876782298088074,
"learning_rate": 4.1159462194593033e-05,
"loss": 1.5067,
"step": 213500
},
{
"epoch": 15.034424617113952,
"grad_norm": 1.0133213996887207,
"learning_rate": 4.1014890848633804e-05,
"loss": 1.5002,
"step": 214000
},
{
"epoch": 15.069551777434313,
"grad_norm": 0.9735503792762756,
"learning_rate": 4.087031950267457e-05,
"loss": 1.5016,
"step": 214500
},
{
"epoch": 15.104678937754672,
"grad_norm": 0.985035240650177,
"learning_rate": 4.0725748156715346e-05,
"loss": 1.5021,
"step": 215000
},
{
"epoch": 15.13980609807503,
"grad_norm": 1.0019500255584717,
"learning_rate": 4.058117681075611e-05,
"loss": 1.4975,
"step": 215500
},
{
"epoch": 15.174933258395392,
"grad_norm": 1.0004901885986328,
"learning_rate": 4.043660546479688e-05,
"loss": 1.5017,
"step": 216000
},
{
"epoch": 15.21006041871575,
"grad_norm": 0.9935667514801025,
"learning_rate": 4.0292034118837645e-05,
"loss": 1.5015,
"step": 216500
},
{
"epoch": 15.245187579036111,
"grad_norm": 1.0111830234527588,
"learning_rate": 4.014746277287842e-05,
"loss": 1.5004,
"step": 217000
},
{
"epoch": 15.28031473935647,
"grad_norm": 1.0003619194030762,
"learning_rate": 4.0002891426919186e-05,
"loss": 1.4991,
"step": 217500
},
{
"epoch": 15.31544189967683,
"grad_norm": 1.0029890537261963,
"learning_rate": 3.985832008095996e-05,
"loss": 1.4992,
"step": 218000
},
{
"epoch": 15.35056905999719,
"grad_norm": 1.0191558599472046,
"learning_rate": 3.971374873500072e-05,
"loss": 1.5008,
"step": 218500
},
{
"epoch": 15.38569622031755,
"grad_norm": 1.01618492603302,
"learning_rate": 3.95691773890415e-05,
"loss": 1.4989,
"step": 219000
},
{
"epoch": 15.42082338063791,
"grad_norm": 1.020931601524353,
"learning_rate": 3.942460604308226e-05,
"loss": 1.4985,
"step": 219500
},
{
"epoch": 15.45595054095827,
"grad_norm": 1.0255357027053833,
"learning_rate": 3.928003469712303e-05,
"loss": 1.4972,
"step": 220000
},
{
"epoch": 15.45595054095827,
"eval_accuracy": 0.7193424504315581,
"eval_loss": 1.2618342638015747,
"eval_runtime": 145.885,
"eval_samples_per_second": 827.405,
"eval_steps_per_second": 5.175,
"step": 220000
},
{
"epoch": 15.491077701278629,
"grad_norm": 1.0047690868377686,
"learning_rate": 3.9135463351163804e-05,
"loss": 1.4989,
"step": 220500
},
{
"epoch": 15.526204861598988,
"grad_norm": 1.040840983390808,
"learning_rate": 3.899089200520457e-05,
"loss": 1.5001,
"step": 221000
},
{
"epoch": 15.561332021919348,
"grad_norm": 0.9770060777664185,
"learning_rate": 3.884632065924534e-05,
"loss": 1.4976,
"step": 221500
},
{
"epoch": 15.596459182239707,
"grad_norm": 0.9793146848678589,
"learning_rate": 3.870174931328611e-05,
"loss": 1.4986,
"step": 222000
},
{
"epoch": 15.631586342560068,
"grad_norm": 0.9713142514228821,
"learning_rate": 3.855717796732688e-05,
"loss": 1.4947,
"step": 222500
},
{
"epoch": 15.666713502880427,
"grad_norm": 1.0131899118423462,
"learning_rate": 3.8412606621367644e-05,
"loss": 1.4965,
"step": 223000
},
{
"epoch": 15.701840663200787,
"grad_norm": 1.0238277912139893,
"learning_rate": 3.8268035275408415e-05,
"loss": 1.4961,
"step": 223500
},
{
"epoch": 15.736967823521146,
"grad_norm": 1.0393719673156738,
"learning_rate": 3.812346392944918e-05,
"loss": 1.4944,
"step": 224000
},
{
"epoch": 15.772094983841507,
"grad_norm": 0.993442952632904,
"learning_rate": 3.797889258348996e-05,
"loss": 1.4932,
"step": 224500
},
{
"epoch": 15.807222144161866,
"grad_norm": 1.010707974433899,
"learning_rate": 3.783432123753072e-05,
"loss": 1.4919,
"step": 225000
},
{
"epoch": 15.842349304482227,
"grad_norm": 1.0285263061523438,
"learning_rate": 3.768974989157149e-05,
"loss": 1.4922,
"step": 225500
},
{
"epoch": 15.877476464802585,
"grad_norm": 0.9992517232894897,
"learning_rate": 3.754517854561226e-05,
"loss": 1.496,
"step": 226000
},
{
"epoch": 15.912603625122944,
"grad_norm": 1.0487091541290283,
"learning_rate": 3.740060719965303e-05,
"loss": 1.496,
"step": 226500
},
{
"epoch": 15.947730785443305,
"grad_norm": 0.9960684776306152,
"learning_rate": 3.72560358536938e-05,
"loss": 1.4916,
"step": 227000
},
{
"epoch": 15.982857945763664,
"grad_norm": 1.0461766719818115,
"learning_rate": 3.711146450773457e-05,
"loss": 1.4945,
"step": 227500
},
{
"epoch": 16.017985106084023,
"grad_norm": 1.058475375175476,
"learning_rate": 3.696689316177534e-05,
"loss": 1.4899,
"step": 228000
},
{
"epoch": 16.053112266404383,
"grad_norm": 1.0232276916503906,
"learning_rate": 3.682232181581611e-05,
"loss": 1.4884,
"step": 228500
},
{
"epoch": 16.088239426724744,
"grad_norm": 1.0691484212875366,
"learning_rate": 3.667775046985687e-05,
"loss": 1.4886,
"step": 229000
},
{
"epoch": 16.123366587045105,
"grad_norm": 1.003165602684021,
"learning_rate": 3.6533179123897644e-05,
"loss": 1.4894,
"step": 229500
},
{
"epoch": 16.158493747365462,
"grad_norm": 1.010514736175537,
"learning_rate": 3.6388607777938415e-05,
"loss": 1.4915,
"step": 230000
},
{
"epoch": 16.158493747365462,
"eval_accuracy": 0.7201344726261764,
"eval_loss": 1.2572591304779053,
"eval_runtime": 146.9469,
"eval_samples_per_second": 821.426,
"eval_steps_per_second": 5.138,
"step": 230000
},
{
"epoch": 16.193620907685823,
"grad_norm": 1.0277786254882812,
"learning_rate": 3.6244036431979186e-05,
"loss": 1.4867,
"step": 230500
},
{
"epoch": 16.228748068006183,
"grad_norm": 1.0123547315597534,
"learning_rate": 3.609946508601995e-05,
"loss": 1.4901,
"step": 231000
},
{
"epoch": 16.26387522832654,
"grad_norm": 1.0431201457977295,
"learning_rate": 3.595489374006073e-05,
"loss": 1.485,
"step": 231500
},
{
"epoch": 16.2990023886469,
"grad_norm": 0.9797715544700623,
"learning_rate": 3.581032239410149e-05,
"loss": 1.4898,
"step": 232000
},
{
"epoch": 16.33412954896726,
"grad_norm": 1.0554850101470947,
"learning_rate": 3.566575104814226e-05,
"loss": 1.4867,
"step": 232500
},
{
"epoch": 16.369256709287622,
"grad_norm": 1.0041229724884033,
"learning_rate": 3.5521179702183026e-05,
"loss": 1.4876,
"step": 233000
},
{
"epoch": 16.40438386960798,
"grad_norm": 1.004367470741272,
"learning_rate": 3.5376608356223804e-05,
"loss": 1.4866,
"step": 233500
},
{
"epoch": 16.43951102992834,
"grad_norm": 0.9991381764411926,
"learning_rate": 3.523203701026457e-05,
"loss": 1.4866,
"step": 234000
},
{
"epoch": 16.4746381902487,
"grad_norm": 1.0062518119812012,
"learning_rate": 3.508746566430534e-05,
"loss": 1.4869,
"step": 234500
},
{
"epoch": 16.50976535056906,
"grad_norm": 1.0116221904754639,
"learning_rate": 3.49428943183461e-05,
"loss": 1.4867,
"step": 235000
},
{
"epoch": 16.54489251088942,
"grad_norm": 1.0057295560836792,
"learning_rate": 3.479832297238687e-05,
"loss": 1.4836,
"step": 235500
},
{
"epoch": 16.58001967120978,
"grad_norm": 0.9946209788322449,
"learning_rate": 3.4653751626427644e-05,
"loss": 1.4871,
"step": 236000
},
{
"epoch": 16.61514683153014,
"grad_norm": 1.052956223487854,
"learning_rate": 3.4509180280468415e-05,
"loss": 1.4836,
"step": 236500
},
{
"epoch": 16.650273991850497,
"grad_norm": 1.0504807233810425,
"learning_rate": 3.436460893450918e-05,
"loss": 1.4898,
"step": 237000
},
{
"epoch": 16.685401152170858,
"grad_norm": 1.0513533353805542,
"learning_rate": 3.422003758854995e-05,
"loss": 1.4857,
"step": 237500
},
{
"epoch": 16.72052831249122,
"grad_norm": 1.0254963636398315,
"learning_rate": 3.407546624259072e-05,
"loss": 1.484,
"step": 238000
},
{
"epoch": 16.75565547281158,
"grad_norm": 1.0302786827087402,
"learning_rate": 3.393089489663149e-05,
"loss": 1.4862,
"step": 238500
},
{
"epoch": 16.790782633131936,
"grad_norm": 1.0228885412216187,
"learning_rate": 3.378632355067226e-05,
"loss": 1.4836,
"step": 239000
},
{
"epoch": 16.825909793452297,
"grad_norm": 1.018209457397461,
"learning_rate": 3.3641752204713026e-05,
"loss": 1.4827,
"step": 239500
},
{
"epoch": 16.861036953772658,
"grad_norm": 0.989748477935791,
"learning_rate": 3.3497180858753797e-05,
"loss": 1.4824,
"step": 240000
},
{
"epoch": 16.861036953772658,
"eval_accuracy": 0.7210892050293545,
"eval_loss": 1.2514859437942505,
"eval_runtime": 145.8765,
"eval_samples_per_second": 827.453,
"eval_steps_per_second": 5.176,
"step": 240000
},
{
"epoch": 16.89616411409302,
"grad_norm": 1.0294135808944702,
"learning_rate": 3.335260951279456e-05,
"loss": 1.4826,
"step": 240500
},
{
"epoch": 16.931291274413375,
"grad_norm": 1.06317138671875,
"learning_rate": 3.320803816683534e-05,
"loss": 1.4851,
"step": 241000
},
{
"epoch": 16.966418434733736,
"grad_norm": 1.072068452835083,
"learning_rate": 3.30634668208761e-05,
"loss": 1.4808,
"step": 241500
},
{
"epoch": 17.001545595054097,
"grad_norm": 1.0590496063232422,
"learning_rate": 3.291889547491687e-05,
"loss": 1.4849,
"step": 242000
},
{
"epoch": 17.036672755374454,
"grad_norm": 1.0390037298202515,
"learning_rate": 3.277432412895764e-05,
"loss": 1.4826,
"step": 242500
},
{
"epoch": 17.071799915694815,
"grad_norm": 1.0264649391174316,
"learning_rate": 3.2629752782998414e-05,
"loss": 1.4784,
"step": 243000
},
{
"epoch": 17.106927076015175,
"grad_norm": 1.0615930557250977,
"learning_rate": 3.248518143703918e-05,
"loss": 1.4799,
"step": 243500
},
{
"epoch": 17.142054236335536,
"grad_norm": 1.0263277292251587,
"learning_rate": 3.234061009107995e-05,
"loss": 1.4773,
"step": 244000
},
{
"epoch": 17.177181396655893,
"grad_norm": 1.016964077949524,
"learning_rate": 3.219603874512072e-05,
"loss": 1.4776,
"step": 244500
},
{
"epoch": 17.212308556976254,
"grad_norm": 1.0330182313919067,
"learning_rate": 3.205146739916149e-05,
"loss": 1.48,
"step": 245000
},
{
"epoch": 17.247435717296614,
"grad_norm": 1.0279370546340942,
"learning_rate": 3.1906896053202255e-05,
"loss": 1.4767,
"step": 245500
},
{
"epoch": 17.282562877616975,
"grad_norm": 1.053545594215393,
"learning_rate": 3.1762324707243026e-05,
"loss": 1.4807,
"step": 246000
},
{
"epoch": 17.317690037937332,
"grad_norm": 1.0184204578399658,
"learning_rate": 3.1617753361283796e-05,
"loss": 1.4776,
"step": 246500
},
{
"epoch": 17.352817198257693,
"grad_norm": 1.0180692672729492,
"learning_rate": 3.147318201532457e-05,
"loss": 1.4765,
"step": 247000
},
{
"epoch": 17.387944358578054,
"grad_norm": 1.03151535987854,
"learning_rate": 3.132861066936533e-05,
"loss": 1.4784,
"step": 247500
},
{
"epoch": 17.42307151889841,
"grad_norm": 1.0360081195831299,
"learning_rate": 3.11840393234061e-05,
"loss": 1.4725,
"step": 248000
},
{
"epoch": 17.45819867921877,
"grad_norm": 1.0606461763381958,
"learning_rate": 3.103946797744687e-05,
"loss": 1.4753,
"step": 248500
},
{
"epoch": 17.493325839539132,
"grad_norm": 1.0544891357421875,
"learning_rate": 3.0894896631487643e-05,
"loss": 1.474,
"step": 249000
},
{
"epoch": 17.528452999859493,
"grad_norm": 1.0830744504928589,
"learning_rate": 3.075032528552841e-05,
"loss": 1.4728,
"step": 249500
},
{
"epoch": 17.56358016017985,
"grad_norm": 1.0082223415374756,
"learning_rate": 3.060575393956918e-05,
"loss": 1.4748,
"step": 250000
},
{
"epoch": 17.56358016017985,
"eval_accuracy": 0.7222705826647003,
"eval_loss": 1.245027780532837,
"eval_runtime": 145.2394,
"eval_samples_per_second": 831.083,
"eval_steps_per_second": 5.198,
"step": 250000
},
{
"epoch": 17.59870732050021,
"grad_norm": 1.0504813194274902,
"learning_rate": 3.046118259360995e-05,
"loss": 1.4783,
"step": 250500
},
{
"epoch": 17.63383448082057,
"grad_norm": 1.0163605213165283,
"learning_rate": 3.0316611247650716e-05,
"loss": 1.4719,
"step": 251000
},
{
"epoch": 17.668961641140932,
"grad_norm": 1.0471776723861694,
"learning_rate": 3.0172039901691484e-05,
"loss": 1.4745,
"step": 251500
},
{
"epoch": 17.70408880146129,
"grad_norm": 1.050743818283081,
"learning_rate": 3.0027468555732258e-05,
"loss": 1.4754,
"step": 252000
},
{
"epoch": 17.73921596178165,
"grad_norm": 1.0446287393569946,
"learning_rate": 2.9882897209773025e-05,
"loss": 1.4742,
"step": 252500
},
{
"epoch": 17.77434312210201,
"grad_norm": 1.0279436111450195,
"learning_rate": 2.9738325863813793e-05,
"loss": 1.4728,
"step": 253000
},
{
"epoch": 17.809470282422367,
"grad_norm": 1.0320532321929932,
"learning_rate": 2.959375451785456e-05,
"loss": 1.4751,
"step": 253500
},
{
"epoch": 17.844597442742728,
"grad_norm": 1.0268456935882568,
"learning_rate": 2.9449183171895334e-05,
"loss": 1.472,
"step": 254000
},
{
"epoch": 17.87972460306309,
"grad_norm": 1.0411797761917114,
"learning_rate": 2.93046118259361e-05,
"loss": 1.4757,
"step": 254500
},
{
"epoch": 17.91485176338345,
"grad_norm": 1.0576931238174438,
"learning_rate": 2.916004047997687e-05,
"loss": 1.4738,
"step": 255000
},
{
"epoch": 17.949978923703807,
"grad_norm": 1.0698353052139282,
"learning_rate": 2.9015469134017636e-05,
"loss": 1.4749,
"step": 255500
},
{
"epoch": 17.985106084024167,
"grad_norm": 1.0517598390579224,
"learning_rate": 2.887089778805841e-05,
"loss": 1.4719,
"step": 256000
},
{
"epoch": 18.020233244344528,
"grad_norm": 1.030639886856079,
"learning_rate": 2.8726326442099178e-05,
"loss": 1.4722,
"step": 256500
},
{
"epoch": 18.05536040466489,
"grad_norm": 1.038255214691162,
"learning_rate": 2.8581755096139945e-05,
"loss": 1.4713,
"step": 257000
},
{
"epoch": 18.090487564985246,
"grad_norm": 1.0348975658416748,
"learning_rate": 2.843718375018072e-05,
"loss": 1.4678,
"step": 257500
},
{
"epoch": 18.125614725305606,
"grad_norm": 1.0541919469833374,
"learning_rate": 2.8292612404221487e-05,
"loss": 1.4711,
"step": 258000
},
{
"epoch": 18.160741885625967,
"grad_norm": 1.0226538181304932,
"learning_rate": 2.8148041058262254e-05,
"loss": 1.4693,
"step": 258500
},
{
"epoch": 18.195869045946324,
"grad_norm": 1.0200276374816895,
"learning_rate": 2.800346971230302e-05,
"loss": 1.4705,
"step": 259000
},
{
"epoch": 18.230996206266685,
"grad_norm": 1.0429800748825073,
"learning_rate": 2.7858898366343792e-05,
"loss": 1.4704,
"step": 259500
},
{
"epoch": 18.266123366587045,
"grad_norm": 1.0760138034820557,
"learning_rate": 2.7714327020384563e-05,
"loss": 1.4686,
"step": 260000
},
{
"epoch": 18.266123366587045,
"eval_accuracy": 0.7233516160401716,
"eval_loss": 1.238907814025879,
"eval_runtime": 142.7876,
"eval_samples_per_second": 845.353,
"eval_steps_per_second": 5.288,
"step": 260000
},
{
"epoch": 18.301250526907406,
"grad_norm": 1.0434221029281616,
"learning_rate": 2.756975567442533e-05,
"loss": 1.468,
"step": 260500
},
{
"epoch": 18.336377687227763,
"grad_norm": 1.0621415376663208,
"learning_rate": 2.7425184328466098e-05,
"loss": 1.4665,
"step": 261000
},
{
"epoch": 18.371504847548124,
"grad_norm": 1.0416877269744873,
"learning_rate": 2.728061298250687e-05,
"loss": 1.4703,
"step": 261500
},
{
"epoch": 18.406632007868485,
"grad_norm": 1.0364477634429932,
"learning_rate": 2.7136041636547636e-05,
"loss": 1.4679,
"step": 262000
},
{
"epoch": 18.441759168188845,
"grad_norm": 1.0509207248687744,
"learning_rate": 2.6991470290588407e-05,
"loss": 1.4645,
"step": 262500
},
{
"epoch": 18.476886328509202,
"grad_norm": 1.0550258159637451,
"learning_rate": 2.6846898944629178e-05,
"loss": 1.465,
"step": 263000
},
{
"epoch": 18.512013488829563,
"grad_norm": 1.045780897140503,
"learning_rate": 2.6702327598669945e-05,
"loss": 1.4645,
"step": 263500
},
{
"epoch": 18.547140649149924,
"grad_norm": 1.033547043800354,
"learning_rate": 2.6557756252710713e-05,
"loss": 1.464,
"step": 264000
},
{
"epoch": 18.58226780947028,
"grad_norm": 1.0473381280899048,
"learning_rate": 2.641318490675148e-05,
"loss": 1.4647,
"step": 264500
},
{
"epoch": 18.61739496979064,
"grad_norm": 1.0651229619979858,
"learning_rate": 2.6268613560792254e-05,
"loss": 1.4683,
"step": 265000
},
{
"epoch": 18.652522130111002,
"grad_norm": 1.0227559804916382,
"learning_rate": 2.612404221483302e-05,
"loss": 1.4653,
"step": 265500
},
{
"epoch": 18.687649290431363,
"grad_norm": 1.0518014430999756,
"learning_rate": 2.597947086887379e-05,
"loss": 1.4634,
"step": 266000
},
{
"epoch": 18.72277645075172,
"grad_norm": 1.0625571012496948,
"learning_rate": 2.5834899522914556e-05,
"loss": 1.4663,
"step": 266500
},
{
"epoch": 18.75790361107208,
"grad_norm": 1.0320320129394531,
"learning_rate": 2.569032817695533e-05,
"loss": 1.4658,
"step": 267000
},
{
"epoch": 18.79303077139244,
"grad_norm": 1.0507465600967407,
"learning_rate": 2.5545756830996098e-05,
"loss": 1.4667,
"step": 267500
},
{
"epoch": 18.828157931712802,
"grad_norm": 1.0624668598175049,
"learning_rate": 2.5401185485036865e-05,
"loss": 1.465,
"step": 268000
},
{
"epoch": 18.86328509203316,
"grad_norm": 1.1198619604110718,
"learning_rate": 2.5256614139077633e-05,
"loss": 1.4639,
"step": 268500
},
{
"epoch": 18.89841225235352,
"grad_norm": 1.0421173572540283,
"learning_rate": 2.5112042793118407e-05,
"loss": 1.4641,
"step": 269000
},
{
"epoch": 18.93353941267388,
"grad_norm": 1.0857394933700562,
"learning_rate": 2.4967471447159174e-05,
"loss": 1.4633,
"step": 269500
},
{
"epoch": 18.968666572994238,
"grad_norm": 1.0285334587097168,
"learning_rate": 2.4822900101199945e-05,
"loss": 1.4649,
"step": 270000
},
{
"epoch": 18.968666572994238,
"eval_accuracy": 0.7242654194082216,
"eval_loss": 1.2332816123962402,
"eval_runtime": 141.3716,
"eval_samples_per_second": 853.821,
"eval_steps_per_second": 5.341,
"step": 270000
},
{
"epoch": 19.0037937333146,
"grad_norm": 1.0661959648132324,
"learning_rate": 2.4678328755240712e-05,
"loss": 1.464,
"step": 270500
},
{
"epoch": 19.03892089363496,
"grad_norm": 1.0679501295089722,
"learning_rate": 2.4533757409281483e-05,
"loss": 1.4604,
"step": 271000
},
{
"epoch": 19.07404805395532,
"grad_norm": 1.0748372077941895,
"learning_rate": 2.438918606332225e-05,
"loss": 1.4611,
"step": 271500
},
{
"epoch": 19.109175214275677,
"grad_norm": 1.0354524850845337,
"learning_rate": 2.424461471736302e-05,
"loss": 1.4604,
"step": 272000
},
{
"epoch": 19.144302374596037,
"grad_norm": 1.0497136116027832,
"learning_rate": 2.410004337140379e-05,
"loss": 1.4616,
"step": 272500
},
{
"epoch": 19.179429534916398,
"grad_norm": 1.0558326244354248,
"learning_rate": 2.395547202544456e-05,
"loss": 1.4611,
"step": 273000
},
{
"epoch": 19.21455669523676,
"grad_norm": 1.0470489263534546,
"learning_rate": 2.3810900679485327e-05,
"loss": 1.4575,
"step": 273500
},
{
"epoch": 19.249683855557116,
"grad_norm": 1.0182609558105469,
"learning_rate": 2.3666329333526098e-05,
"loss": 1.4594,
"step": 274000
},
{
"epoch": 19.284811015877477,
"grad_norm": 1.0594130754470825,
"learning_rate": 2.3521757987566865e-05,
"loss": 1.4582,
"step": 274500
},
{
"epoch": 19.319938176197837,
"grad_norm": 1.0967395305633545,
"learning_rate": 2.3377186641607636e-05,
"loss": 1.4607,
"step": 275000
},
{
"epoch": 19.355065336518194,
"grad_norm": 1.0475187301635742,
"learning_rate": 2.3232615295648403e-05,
"loss": 1.4617,
"step": 275500
},
{
"epoch": 19.390192496838555,
"grad_norm": 1.0457967519760132,
"learning_rate": 2.3088043949689174e-05,
"loss": 1.4569,
"step": 276000
},
{
"epoch": 19.425319657158916,
"grad_norm": 1.0595502853393555,
"learning_rate": 2.2943472603729945e-05,
"loss": 1.4609,
"step": 276500
},
{
"epoch": 19.460446817479276,
"grad_norm": 1.0646470785140991,
"learning_rate": 2.2798901257770712e-05,
"loss": 1.4572,
"step": 277000
},
{
"epoch": 19.495573977799634,
"grad_norm": 1.0434619188308716,
"learning_rate": 2.265432991181148e-05,
"loss": 1.4597,
"step": 277500
},
{
"epoch": 19.530701138119994,
"grad_norm": 1.039019227027893,
"learning_rate": 2.250975856585225e-05,
"loss": 1.4576,
"step": 278000
},
{
"epoch": 19.565828298440355,
"grad_norm": 1.0945369005203247,
"learning_rate": 2.2365187219893018e-05,
"loss": 1.459,
"step": 278500
},
{
"epoch": 19.600955458760716,
"grad_norm": 1.0497277975082397,
"learning_rate": 2.2220615873933785e-05,
"loss": 1.4555,
"step": 279000
},
{
"epoch": 19.636082619081073,
"grad_norm": 1.0733200311660767,
"learning_rate": 2.2076044527974556e-05,
"loss": 1.4586,
"step": 279500
},
{
"epoch": 19.671209779401433,
"grad_norm": 1.0626968145370483,
"learning_rate": 2.1931473182015323e-05,
"loss": 1.4566,
"step": 280000
},
{
"epoch": 19.671209779401433,
"eval_accuracy": 0.7253028435871818,
"eval_loss": 1.2285022735595703,
"eval_runtime": 141.7716,
"eval_samples_per_second": 851.412,
"eval_steps_per_second": 5.325,
"step": 280000
},
{
"epoch": 19.706336939721794,
"grad_norm": 1.0224528312683105,
"learning_rate": 2.1786901836056094e-05,
"loss": 1.4583,
"step": 280500
},
{
"epoch": 19.74146410004215,
"grad_norm": 1.0466312170028687,
"learning_rate": 2.164233049009686e-05,
"loss": 1.457,
"step": 281000
},
{
"epoch": 19.776591260362512,
"grad_norm": 1.0637938976287842,
"learning_rate": 2.1497759144137632e-05,
"loss": 1.4577,
"step": 281500
},
{
"epoch": 19.811718420682872,
"grad_norm": 1.0693141222000122,
"learning_rate": 2.1353187798178403e-05,
"loss": 1.4566,
"step": 282000
},
{
"epoch": 19.846845581003233,
"grad_norm": 1.0720393657684326,
"learning_rate": 2.120861645221917e-05,
"loss": 1.456,
"step": 282500
},
{
"epoch": 19.88197274132359,
"grad_norm": 1.071040153503418,
"learning_rate": 2.106404510625994e-05,
"loss": 1.4546,
"step": 283000
},
{
"epoch": 19.91709990164395,
"grad_norm": 1.0675835609436035,
"learning_rate": 2.091947376030071e-05,
"loss": 1.4582,
"step": 283500
},
{
"epoch": 19.95222706196431,
"grad_norm": 1.0603220462799072,
"learning_rate": 2.077490241434148e-05,
"loss": 1.4547,
"step": 284000
},
{
"epoch": 19.987354222284672,
"grad_norm": 1.0802963972091675,
"learning_rate": 2.0630331068382247e-05,
"loss": 1.4546,
"step": 284500
},
{
"epoch": 20.02248138260503,
"grad_norm": 1.0588197708129883,
"learning_rate": 2.0485759722423017e-05,
"loss": 1.4575,
"step": 285000
},
{
"epoch": 20.05760854292539,
"grad_norm": 1.0602234601974487,
"learning_rate": 2.0341188376463785e-05,
"loss": 1.4521,
"step": 285500
},
{
"epoch": 20.09273570324575,
"grad_norm": 1.1232300996780396,
"learning_rate": 2.0196617030504556e-05,
"loss": 1.4517,
"step": 286000
},
{
"epoch": 20.127862863566108,
"grad_norm": 1.0660282373428345,
"learning_rate": 2.0052045684545323e-05,
"loss": 1.4523,
"step": 286500
},
{
"epoch": 20.16299002388647,
"grad_norm": 1.0167709589004517,
"learning_rate": 1.9907474338586094e-05,
"loss": 1.4519,
"step": 287000
},
{
"epoch": 20.19811718420683,
"grad_norm": 1.047339677810669,
"learning_rate": 1.976290299262686e-05,
"loss": 1.4502,
"step": 287500
},
{
"epoch": 20.23324434452719,
"grad_norm": 1.0285905599594116,
"learning_rate": 1.9618331646667632e-05,
"loss": 1.4514,
"step": 288000
},
{
"epoch": 20.268371504847547,
"grad_norm": 1.05581533908844,
"learning_rate": 1.9473760300708403e-05,
"loss": 1.449,
"step": 288500
},
{
"epoch": 20.303498665167908,
"grad_norm": 1.0714962482452393,
"learning_rate": 1.932918895474917e-05,
"loss": 1.4521,
"step": 289000
},
{
"epoch": 20.33862582548827,
"grad_norm": 1.0655773878097534,
"learning_rate": 1.918461760878994e-05,
"loss": 1.4508,
"step": 289500
},
{
"epoch": 20.37375298580863,
"grad_norm": 1.0571376085281372,
"learning_rate": 1.9040046262830708e-05,
"loss": 1.4529,
"step": 290000
},
{
"epoch": 20.37375298580863,
"eval_accuracy": 0.7261482132790039,
"eval_loss": 1.223044753074646,
"eval_runtime": 141.5948,
"eval_samples_per_second": 852.475,
"eval_steps_per_second": 5.332,
"step": 290000
},
{
"epoch": 20.408880146128986,
"grad_norm": 1.0565483570098877,
"learning_rate": 1.889547491687148e-05,
"loss": 1.4504,
"step": 290500
},
{
"epoch": 20.444007306449347,
"grad_norm": 1.0659668445587158,
"learning_rate": 1.8750903570912246e-05,
"loss": 1.4525,
"step": 291000
},
{
"epoch": 20.479134466769708,
"grad_norm": 1.122076153755188,
"learning_rate": 1.8606332224953017e-05,
"loss": 1.4485,
"step": 291500
},
{
"epoch": 20.514261627090065,
"grad_norm": 1.0631004571914673,
"learning_rate": 1.8461760878993785e-05,
"loss": 1.4513,
"step": 292000
},
{
"epoch": 20.549388787410425,
"grad_norm": 1.0630178451538086,
"learning_rate": 1.8317189533034555e-05,
"loss": 1.4486,
"step": 292500
},
{
"epoch": 20.584515947730786,
"grad_norm": 1.0705844163894653,
"learning_rate": 1.8172618187075323e-05,
"loss": 1.4503,
"step": 293000
},
{
"epoch": 20.619643108051147,
"grad_norm": 1.0725988149642944,
"learning_rate": 1.8028046841116093e-05,
"loss": 1.4515,
"step": 293500
},
{
"epoch": 20.654770268371504,
"grad_norm": 1.0657289028167725,
"learning_rate": 1.788347549515686e-05,
"loss": 1.4475,
"step": 294000
},
{
"epoch": 20.689897428691864,
"grad_norm": 1.062153697013855,
"learning_rate": 1.7738904149197628e-05,
"loss": 1.4503,
"step": 294500
},
{
"epoch": 20.725024589012225,
"grad_norm": 1.0483520030975342,
"learning_rate": 1.75943328032384e-05,
"loss": 1.4492,
"step": 295000
},
{
"epoch": 20.760151749332586,
"grad_norm": 1.07576322555542,
"learning_rate": 1.7449761457279166e-05,
"loss": 1.4514,
"step": 295500
},
{
"epoch": 20.795278909652943,
"grad_norm": 1.0874416828155518,
"learning_rate": 1.7305190111319937e-05,
"loss": 1.4463,
"step": 296000
},
{
"epoch": 20.830406069973304,
"grad_norm": 1.0873123407363892,
"learning_rate": 1.7160618765360705e-05,
"loss": 1.4483,
"step": 296500
},
{
"epoch": 20.865533230293664,
"grad_norm": 1.0750373601913452,
"learning_rate": 1.7016047419401475e-05,
"loss": 1.4461,
"step": 297000
},
{
"epoch": 20.90066039061402,
"grad_norm": 1.0832915306091309,
"learning_rate": 1.6871476073442243e-05,
"loss": 1.45,
"step": 297500
},
{
"epoch": 20.935787550934382,
"grad_norm": 1.0408779382705688,
"learning_rate": 1.6726904727483014e-05,
"loss": 1.4483,
"step": 298000
},
{
"epoch": 20.970914711254743,
"grad_norm": 1.0541013479232788,
"learning_rate": 1.658233338152378e-05,
"loss": 1.4492,
"step": 298500
},
{
"epoch": 21.006041871575103,
"grad_norm": 1.0725327730178833,
"learning_rate": 1.6437762035564552e-05,
"loss": 1.4475,
"step": 299000
},
{
"epoch": 21.04116903189546,
"grad_norm": 1.0820538997650146,
"learning_rate": 1.629319068960532e-05,
"loss": 1.4481,
"step": 299500
},
{
"epoch": 21.07629619221582,
"grad_norm": 1.0279037952423096,
"learning_rate": 1.614861934364609e-05,
"loss": 1.4451,
"step": 300000
},
{
"epoch": 21.07629619221582,
"eval_accuracy": 0.7268696778575053,
"eval_loss": 1.2188650369644165,
"eval_runtime": 142.0191,
"eval_samples_per_second": 849.928,
"eval_steps_per_second": 5.316,
"step": 300000
},
{
"epoch": 21.111423352536182,
"grad_norm": 1.034382700920105,
"learning_rate": 1.6004047997686857e-05,
"loss": 1.4447,
"step": 300500
},
{
"epoch": 21.14655051285654,
"grad_norm": 1.0651057958602905,
"learning_rate": 1.5859476651727628e-05,
"loss": 1.4467,
"step": 301000
},
{
"epoch": 21.1816776731769,
"grad_norm": 1.086431622505188,
"learning_rate": 1.57149053057684e-05,
"loss": 1.4452,
"step": 301500
},
{
"epoch": 21.21680483349726,
"grad_norm": 1.0573840141296387,
"learning_rate": 1.5570333959809166e-05,
"loss": 1.4487,
"step": 302000
},
{
"epoch": 21.25193199381762,
"grad_norm": 1.0548052787780762,
"learning_rate": 1.5425762613849937e-05,
"loss": 1.4424,
"step": 302500
},
{
"epoch": 21.287059154137978,
"grad_norm": 1.0701725482940674,
"learning_rate": 1.5281191267890704e-05,
"loss": 1.4441,
"step": 303000
},
{
"epoch": 21.32218631445834,
"grad_norm": 1.0664572715759277,
"learning_rate": 1.5136619921931475e-05,
"loss": 1.4449,
"step": 303500
},
{
"epoch": 21.3573134747787,
"grad_norm": 1.0336108207702637,
"learning_rate": 1.4992048575972243e-05,
"loss": 1.4415,
"step": 304000
},
{
"epoch": 21.39244063509906,
"grad_norm": 1.098109245300293,
"learning_rate": 1.4847477230013013e-05,
"loss": 1.4453,
"step": 304500
},
{
"epoch": 21.427567795419417,
"grad_norm": 1.0706530809402466,
"learning_rate": 1.470290588405378e-05,
"loss": 1.4433,
"step": 305000
},
{
"epoch": 21.462694955739778,
"grad_norm": 1.086916208267212,
"learning_rate": 1.4558334538094551e-05,
"loss": 1.4441,
"step": 305500
},
{
"epoch": 21.49782211606014,
"grad_norm": 1.1005213260650635,
"learning_rate": 1.4413763192135319e-05,
"loss": 1.4429,
"step": 306000
},
{
"epoch": 21.532949276380496,
"grad_norm": 1.1273337602615356,
"learning_rate": 1.4269191846176088e-05,
"loss": 1.4429,
"step": 306500
},
{
"epoch": 21.568076436700856,
"grad_norm": 1.092872142791748,
"learning_rate": 1.4124620500216857e-05,
"loss": 1.4417,
"step": 307000
},
{
"epoch": 21.603203597021217,
"grad_norm": 1.0761197805404663,
"learning_rate": 1.3980049154257626e-05,
"loss": 1.4427,
"step": 307500
},
{
"epoch": 21.638330757341578,
"grad_norm": 1.1086857318878174,
"learning_rate": 1.3835477808298397e-05,
"loss": 1.447,
"step": 308000
},
{
"epoch": 21.673457917661935,
"grad_norm": 1.091773271560669,
"learning_rate": 1.3690906462339164e-05,
"loss": 1.4437,
"step": 308500
},
{
"epoch": 21.708585077982296,
"grad_norm": 1.0596261024475098,
"learning_rate": 1.3546335116379935e-05,
"loss": 1.4421,
"step": 309000
},
{
"epoch": 21.743712238302656,
"grad_norm": 1.0684653520584106,
"learning_rate": 1.3401763770420702e-05,
"loss": 1.4407,
"step": 309500
},
{
"epoch": 21.778839398623017,
"grad_norm": 1.0676227807998657,
"learning_rate": 1.3257192424461473e-05,
"loss": 1.443,
"step": 310000
},
{
"epoch": 21.778839398623017,
"eval_accuracy": 0.7277766831400472,
"eval_loss": 1.2135990858078003,
"eval_runtime": 142.0609,
"eval_samples_per_second": 849.678,
"eval_steps_per_second": 5.315,
"step": 310000
},
{
"epoch": 21.813966558943374,
"grad_norm": 1.0764987468719482,
"learning_rate": 1.311262107850224e-05,
"loss": 1.4392,
"step": 310500
},
{
"epoch": 21.849093719263735,
"grad_norm": 1.075104832649231,
"learning_rate": 1.2968049732543011e-05,
"loss": 1.439,
"step": 311000
},
{
"epoch": 21.884220879584095,
"grad_norm": 1.0967333316802979,
"learning_rate": 1.2823478386583779e-05,
"loss": 1.4428,
"step": 311500
},
{
"epoch": 21.919348039904452,
"grad_norm": 1.0619230270385742,
"learning_rate": 1.267890704062455e-05,
"loss": 1.4394,
"step": 312000
},
{
"epoch": 21.954475200224813,
"grad_norm": 1.074622631072998,
"learning_rate": 1.2534335694665317e-05,
"loss": 1.4416,
"step": 312500
},
{
"epoch": 21.989602360545174,
"grad_norm": 1.1134517192840576,
"learning_rate": 1.2389764348706088e-05,
"loss": 1.4424,
"step": 313000
},
{
"epoch": 22.024729520865534,
"grad_norm": 1.0811127424240112,
"learning_rate": 1.2245193002746857e-05,
"loss": 1.4396,
"step": 313500
},
{
"epoch": 22.05985668118589,
"grad_norm": 1.0637929439544678,
"learning_rate": 1.2100621656787626e-05,
"loss": 1.4368,
"step": 314000
},
{
"epoch": 22.094983841506252,
"grad_norm": 1.0761417150497437,
"learning_rate": 1.1956050310828395e-05,
"loss": 1.4403,
"step": 314500
},
{
"epoch": 22.130111001826613,
"grad_norm": 1.099725604057312,
"learning_rate": 1.1811478964869162e-05,
"loss": 1.4378,
"step": 315000
},
{
"epoch": 22.165238162146974,
"grad_norm": 1.0871621370315552,
"learning_rate": 1.1666907618909931e-05,
"loss": 1.4362,
"step": 315500
},
{
"epoch": 22.20036532246733,
"grad_norm": 1.057110071182251,
"learning_rate": 1.15223362729507e-05,
"loss": 1.4365,
"step": 316000
},
{
"epoch": 22.23549248278769,
"grad_norm": 1.1010171175003052,
"learning_rate": 1.137776492699147e-05,
"loss": 1.4387,
"step": 316500
},
{
"epoch": 22.270619643108052,
"grad_norm": 1.058934211730957,
"learning_rate": 1.123319358103224e-05,
"loss": 1.4394,
"step": 317000
},
{
"epoch": 22.30574680342841,
"grad_norm": 1.0834625959396362,
"learning_rate": 1.108862223507301e-05,
"loss": 1.4376,
"step": 317500
},
{
"epoch": 22.34087396374877,
"grad_norm": 1.0423885583877563,
"learning_rate": 1.0944050889113779e-05,
"loss": 1.4372,
"step": 318000
},
{
"epoch": 22.37600112406913,
"grad_norm": 1.0966408252716064,
"learning_rate": 1.0799479543154548e-05,
"loss": 1.4394,
"step": 318500
},
{
"epoch": 22.41112828438949,
"grad_norm": 1.0695067644119263,
"learning_rate": 1.0654908197195317e-05,
"loss": 1.4362,
"step": 319000
},
{
"epoch": 22.44625544470985,
"grad_norm": 1.134260892868042,
"learning_rate": 1.0510336851236086e-05,
"loss": 1.435,
"step": 319500
},
{
"epoch": 22.48138260503021,
"grad_norm": 1.1000449657440186,
"learning_rate": 1.0365765505276855e-05,
"loss": 1.4357,
"step": 320000
},
{
"epoch": 22.48138260503021,
"eval_accuracy": 0.7284053903950559,
"eval_loss": 1.210018277168274,
"eval_runtime": 142.872,
"eval_samples_per_second": 844.854,
"eval_steps_per_second": 5.284,
"step": 320000
},
{
"epoch": 22.51650976535057,
"grad_norm": 1.0945639610290527,
"learning_rate": 1.0221194159317624e-05,
"loss": 1.4366,
"step": 320500
},
{
"epoch": 22.55163692567093,
"grad_norm": 1.1207478046417236,
"learning_rate": 1.0076622813358393e-05,
"loss": 1.4399,
"step": 321000
},
{
"epoch": 22.586764085991287,
"grad_norm": 1.0734148025512695,
"learning_rate": 9.932051467399162e-06,
"loss": 1.4385,
"step": 321500
},
{
"epoch": 22.621891246311648,
"grad_norm": 1.0629650354385376,
"learning_rate": 9.787480121439931e-06,
"loss": 1.4352,
"step": 322000
},
{
"epoch": 22.65701840663201,
"grad_norm": 1.0973799228668213,
"learning_rate": 9.6429087754807e-06,
"loss": 1.4377,
"step": 322500
},
{
"epoch": 22.692145566952366,
"grad_norm": 1.0496876239776611,
"learning_rate": 9.49833742952147e-06,
"loss": 1.4372,
"step": 323000
},
{
"epoch": 22.727272727272727,
"grad_norm": 1.0563302040100098,
"learning_rate": 9.353766083562238e-06,
"loss": 1.4352,
"step": 323500
},
{
"epoch": 22.762399887593087,
"grad_norm": 1.1020653247833252,
"learning_rate": 9.209194737603008e-06,
"loss": 1.4331,
"step": 324000
},
{
"epoch": 22.797527047913448,
"grad_norm": 1.0854381322860718,
"learning_rate": 9.064623391643777e-06,
"loss": 1.4371,
"step": 324500
},
{
"epoch": 22.832654208233805,
"grad_norm": 1.0770697593688965,
"learning_rate": 8.920052045684546e-06,
"loss": 1.4361,
"step": 325000
},
{
"epoch": 22.867781368554166,
"grad_norm": 1.0986183881759644,
"learning_rate": 8.775480699725315e-06,
"loss": 1.4354,
"step": 325500
},
{
"epoch": 22.902908528874526,
"grad_norm": 1.1258665323257446,
"learning_rate": 8.630909353766084e-06,
"loss": 1.4376,
"step": 326000
},
{
"epoch": 22.938035689194887,
"grad_norm": 1.0938217639923096,
"learning_rate": 8.486338007806853e-06,
"loss": 1.4346,
"step": 326500
},
{
"epoch": 22.973162849515244,
"grad_norm": 1.1074005365371704,
"learning_rate": 8.341766661847622e-06,
"loss": 1.4306,
"step": 327000
},
{
"epoch": 23.008290009835605,
"grad_norm": 1.0893038511276245,
"learning_rate": 8.197195315888391e-06,
"loss": 1.4346,
"step": 327500
},
{
"epoch": 23.043417170155966,
"grad_norm": 1.1042568683624268,
"learning_rate": 8.05262396992916e-06,
"loss": 1.4335,
"step": 328000
},
{
"epoch": 23.078544330476323,
"grad_norm": 1.0980304479599,
"learning_rate": 7.90805262396993e-06,
"loss": 1.43,
"step": 328500
},
{
"epoch": 23.113671490796683,
"grad_norm": 1.0918561220169067,
"learning_rate": 7.763481278010698e-06,
"loss": 1.4333,
"step": 329000
},
{
"epoch": 23.148798651117044,
"grad_norm": 1.1151840686798096,
"learning_rate": 7.618909932051468e-06,
"loss": 1.4346,
"step": 329500
},
{
"epoch": 23.183925811437405,
"grad_norm": 1.0970309972763062,
"learning_rate": 7.474338586092237e-06,
"loss": 1.4327,
"step": 330000
},
{
"epoch": 23.183925811437405,
"eval_accuracy": 0.7289820707456471,
"eval_loss": 1.2067663669586182,
"eval_runtime": 141.4523,
"eval_samples_per_second": 853.333,
"eval_steps_per_second": 5.337,
"step": 330000
},
{
"epoch": 23.219052971757762,
"grad_norm": 1.1229480504989624,
"learning_rate": 7.3297672401330065e-06,
"loss": 1.429,
"step": 330500
},
{
"epoch": 23.254180132078123,
"grad_norm": 1.103352665901184,
"learning_rate": 7.1851958941737756e-06,
"loss": 1.4332,
"step": 331000
},
{
"epoch": 23.289307292398483,
"grad_norm": 1.0836817026138306,
"learning_rate": 7.040624548214545e-06,
"loss": 1.4331,
"step": 331500
},
{
"epoch": 23.324434452718844,
"grad_norm": 1.07485830783844,
"learning_rate": 6.896053202255314e-06,
"loss": 1.4351,
"step": 332000
},
{
"epoch": 23.3595616130392,
"grad_norm": 1.0870476961135864,
"learning_rate": 6.751481856296083e-06,
"loss": 1.4328,
"step": 332500
},
{
"epoch": 23.39468877335956,
"grad_norm": 1.0797423124313354,
"learning_rate": 6.606910510336851e-06,
"loss": 1.4297,
"step": 333000
},
{
"epoch": 23.429815933679922,
"grad_norm": 1.072199821472168,
"learning_rate": 6.46233916437762e-06,
"loss": 1.431,
"step": 333500
},
{
"epoch": 23.46494309400028,
"grad_norm": 1.0913589000701904,
"learning_rate": 6.317767818418389e-06,
"loss": 1.4317,
"step": 334000
},
{
"epoch": 23.50007025432064,
"grad_norm": 1.1012929677963257,
"learning_rate": 6.173196472459159e-06,
"loss": 1.4333,
"step": 334500
},
{
"epoch": 23.535197414641,
"grad_norm": 1.083678960800171,
"learning_rate": 6.028625126499928e-06,
"loss": 1.4331,
"step": 335000
},
{
"epoch": 23.57032457496136,
"grad_norm": 1.1238749027252197,
"learning_rate": 5.884053780540697e-06,
"loss": 1.4321,
"step": 335500
},
{
"epoch": 23.60545173528172,
"grad_norm": 1.0786747932434082,
"learning_rate": 5.739482434581466e-06,
"loss": 1.4292,
"step": 336000
},
{
"epoch": 23.64057889560208,
"grad_norm": 1.1055099964141846,
"learning_rate": 5.5949110886222355e-06,
"loss": 1.4271,
"step": 336500
},
{
"epoch": 23.67570605592244,
"grad_norm": 1.1266461610794067,
"learning_rate": 5.4503397426630046e-06,
"loss": 1.4317,
"step": 337000
},
{
"epoch": 23.7108332162428,
"grad_norm": 1.0842081308364868,
"learning_rate": 5.305768396703774e-06,
"loss": 1.4254,
"step": 337500
},
{
"epoch": 23.745960376563158,
"grad_norm": 1.0907618999481201,
"learning_rate": 5.161197050744543e-06,
"loss": 1.4276,
"step": 338000
},
{
"epoch": 23.78108753688352,
"grad_norm": 1.0893789529800415,
"learning_rate": 5.016625704785312e-06,
"loss": 1.4274,
"step": 338500
},
{
"epoch": 23.81621469720388,
"grad_norm": 1.108290195465088,
"learning_rate": 4.872054358826081e-06,
"loss": 1.4303,
"step": 339000
},
{
"epoch": 23.851341857524236,
"grad_norm": 1.1075944900512695,
"learning_rate": 4.72748301286685e-06,
"loss": 1.43,
"step": 339500
},
{
"epoch": 23.886469017844597,
"grad_norm": 1.0970890522003174,
"learning_rate": 4.582911666907619e-06,
"loss": 1.4309,
"step": 340000
},
{
"epoch": 23.886469017844597,
"eval_accuracy": 0.7294783881654981,
"eval_loss": 1.2039755582809448,
"eval_runtime": 142.3189,
"eval_samples_per_second": 848.137,
"eval_steps_per_second": 5.305,
"step": 340000
},
{
"epoch": 23.921596178164958,
"grad_norm": 1.102753758430481,
"learning_rate": 4.438340320948388e-06,
"loss": 1.4285,
"step": 340500
},
{
"epoch": 23.956723338485318,
"grad_norm": 1.0833710432052612,
"learning_rate": 4.293768974989157e-06,
"loss": 1.4296,
"step": 341000
},
{
"epoch": 23.991850498805675,
"grad_norm": 1.0727189779281616,
"learning_rate": 4.149197629029926e-06,
"loss": 1.4298,
"step": 341500
},
{
"epoch": 24.026977659126036,
"grad_norm": 1.0745258331298828,
"learning_rate": 4.004626283070695e-06,
"loss": 1.4306,
"step": 342000
},
{
"epoch": 24.062104819446397,
"grad_norm": 1.1169886589050293,
"learning_rate": 3.8600549371114645e-06,
"loss": 1.4281,
"step": 342500
},
{
"epoch": 24.097231979766757,
"grad_norm": 1.1066441535949707,
"learning_rate": 3.7154835911522336e-06,
"loss": 1.4271,
"step": 343000
},
{
"epoch": 24.132359140087114,
"grad_norm": 1.0809811353683472,
"learning_rate": 3.5709122451930026e-06,
"loss": 1.4288,
"step": 343500
},
{
"epoch": 24.167486300407475,
"grad_norm": 1.0997700691223145,
"learning_rate": 3.426340899233772e-06,
"loss": 1.4311,
"step": 344000
},
{
"epoch": 24.202613460727836,
"grad_norm": 1.0773580074310303,
"learning_rate": 3.2817695532745412e-06,
"loss": 1.4286,
"step": 344500
},
{
"epoch": 24.237740621048193,
"grad_norm": 1.1262527704238892,
"learning_rate": 3.1371982073153103e-06,
"loss": 1.4309,
"step": 345000
},
{
"epoch": 24.272867781368554,
"grad_norm": 1.123502254486084,
"learning_rate": 2.9926268613560794e-06,
"loss": 1.4282,
"step": 345500
},
{
"epoch": 24.307994941688914,
"grad_norm": 1.1002156734466553,
"learning_rate": 2.8480555153968485e-06,
"loss": 1.4275,
"step": 346000
},
{
"epoch": 24.343122102009275,
"grad_norm": 1.07537043094635,
"learning_rate": 2.7034841694376176e-06,
"loss": 1.4253,
"step": 346500
},
{
"epoch": 24.378249262329632,
"grad_norm": 1.1036957502365112,
"learning_rate": 2.5589128234783866e-06,
"loss": 1.4279,
"step": 347000
},
{
"epoch": 24.413376422649993,
"grad_norm": 1.1081724166870117,
"learning_rate": 2.414341477519156e-06,
"loss": 1.427,
"step": 347500
},
{
"epoch": 24.448503582970353,
"grad_norm": 1.0847169160842896,
"learning_rate": 2.269770131559925e-06,
"loss": 1.4282,
"step": 348000
},
{
"epoch": 24.483630743290714,
"grad_norm": 1.0844043493270874,
"learning_rate": 2.125198785600694e-06,
"loss": 1.4258,
"step": 348500
},
{
"epoch": 24.51875790361107,
"grad_norm": 1.0586968660354614,
"learning_rate": 1.980627439641463e-06,
"loss": 1.4246,
"step": 349000
},
{
"epoch": 24.553885063931432,
"grad_norm": 1.0943015813827515,
"learning_rate": 1.8360560936822323e-06,
"loss": 1.425,
"step": 349500
},
{
"epoch": 24.589012224251793,
"grad_norm": 1.080601453781128,
"learning_rate": 1.6914847477230013e-06,
"loss": 1.4281,
"step": 350000
},
{
"epoch": 24.589012224251793,
"eval_accuracy": 0.7300085771105469,
"eval_loss": 1.2004581689834595,
"eval_runtime": 142.3807,
"eval_samples_per_second": 847.769,
"eval_steps_per_second": 5.303,
"step": 350000
},
{
"epoch": 24.62413938457215,
"grad_norm": 1.1096524000167847,
"learning_rate": 1.5469134017637704e-06,
"loss": 1.4259,
"step": 350500
},
{
"epoch": 24.65926654489251,
"grad_norm": 1.1088886260986328,
"learning_rate": 1.4023420558045395e-06,
"loss": 1.4247,
"step": 351000
},
{
"epoch": 24.69439370521287,
"grad_norm": 1.0978398323059082,
"learning_rate": 1.2577707098453088e-06,
"loss": 1.4257,
"step": 351500
},
{
"epoch": 24.72952086553323,
"grad_norm": 1.131113886833191,
"learning_rate": 1.1131993638860779e-06,
"loss": 1.4275,
"step": 352000
},
{
"epoch": 24.76464802585359,
"grad_norm": 1.1421183347702026,
"learning_rate": 9.68628017926847e-07,
"loss": 1.4258,
"step": 352500
},
{
"epoch": 24.79977518617395,
"grad_norm": 1.0884921550750732,
"learning_rate": 8.240566719676161e-07,
"loss": 1.4264,
"step": 353000
},
{
"epoch": 24.83490234649431,
"grad_norm": 1.0992377996444702,
"learning_rate": 6.794853260083851e-07,
"loss": 1.4257,
"step": 353500
},
{
"epoch": 24.870029506814667,
"grad_norm": 1.0836498737335205,
"learning_rate": 5.349139800491543e-07,
"loss": 1.4275,
"step": 354000
},
{
"epoch": 24.905156667135028,
"grad_norm": 1.1038559675216675,
"learning_rate": 3.903426340899234e-07,
"loss": 1.4262,
"step": 354500
},
{
"epoch": 24.94028382745539,
"grad_norm": 1.1141635179519653,
"learning_rate": 2.457712881306925e-07,
"loss": 1.4226,
"step": 355000
},
{
"epoch": 24.97541098777575,
"grad_norm": 1.0651347637176514,
"learning_rate": 1.0119994217146162e-07,
"loss": 1.4269,
"step": 355500
},
{
"epoch": 25.0,
"step": 355850,
"total_flos": 7.547177208649421e+18,
"train_loss": 1.849292092869351,
"train_runtime": 122993.5623,
"train_samples_per_second": 462.899,
"train_steps_per_second": 2.893
}
],
"logging_steps": 500,
"max_steps": 355850,
"num_input_tokens_seen": 0,
"num_train_epochs": 25,
"save_steps": 10000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 7.547177208649421e+18,
"train_batch_size": 160,
"trial_name": null,
"trial_params": null
}