V10-40G / checkpoint-1100 /trainer_state.json
gotzmann's picture
..
5a9f14d
raw
history blame contribute delete
No virus
192 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.6779661016949152,
"eval_steps": 500,
"global_step": 1100,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0006163328197226503,
"grad_norm": 7.7007341384887695,
"learning_rate": 2.45398773006135e-07,
"loss": 1.9014,
"step": 1
},
{
"epoch": 0.0012326656394453005,
"grad_norm": 7.831578254699707,
"learning_rate": 4.9079754601227e-07,
"loss": 1.9231,
"step": 2
},
{
"epoch": 0.0018489984591679508,
"grad_norm": 7.626069068908691,
"learning_rate": 7.36196319018405e-07,
"loss": 1.9178,
"step": 3
},
{
"epoch": 0.002465331278890601,
"grad_norm": 7.49867057800293,
"learning_rate": 9.8159509202454e-07,
"loss": 1.9453,
"step": 4
},
{
"epoch": 0.0030816640986132513,
"grad_norm": 7.060399532318115,
"learning_rate": 1.226993865030675e-06,
"loss": 1.8867,
"step": 5
},
{
"epoch": 0.0036979969183359015,
"grad_norm": 6.3593010902404785,
"learning_rate": 1.47239263803681e-06,
"loss": 1.9149,
"step": 6
},
{
"epoch": 0.004314329738058551,
"grad_norm": 5.817101955413818,
"learning_rate": 1.717791411042945e-06,
"loss": 1.8836,
"step": 7
},
{
"epoch": 0.004930662557781202,
"grad_norm": 4.121643543243408,
"learning_rate": 1.96319018404908e-06,
"loss": 1.8643,
"step": 8
},
{
"epoch": 0.005546995377503852,
"grad_norm": 3.0739834308624268,
"learning_rate": 2.208588957055215e-06,
"loss": 1.7948,
"step": 9
},
{
"epoch": 0.0061633281972265025,
"grad_norm": 2.5905909538269043,
"learning_rate": 2.45398773006135e-06,
"loss": 1.7899,
"step": 10
},
{
"epoch": 0.006779661016949152,
"grad_norm": 4.185056209564209,
"learning_rate": 2.699386503067485e-06,
"loss": 1.8681,
"step": 11
},
{
"epoch": 0.007395993836671803,
"grad_norm": 4.769819259643555,
"learning_rate": 2.94478527607362e-06,
"loss": 1.8309,
"step": 12
},
{
"epoch": 0.008012326656394453,
"grad_norm": 4.520302772521973,
"learning_rate": 3.190184049079755e-06,
"loss": 1.7631,
"step": 13
},
{
"epoch": 0.008628659476117103,
"grad_norm": 4.179266452789307,
"learning_rate": 3.43558282208589e-06,
"loss": 1.878,
"step": 14
},
{
"epoch": 0.009244992295839754,
"grad_norm": 3.0450217723846436,
"learning_rate": 3.680981595092025e-06,
"loss": 1.7691,
"step": 15
},
{
"epoch": 0.009861325115562404,
"grad_norm": 2.297912359237671,
"learning_rate": 3.92638036809816e-06,
"loss": 1.778,
"step": 16
},
{
"epoch": 0.010477657935285054,
"grad_norm": 1.6684463024139404,
"learning_rate": 4.171779141104294e-06,
"loss": 1.7915,
"step": 17
},
{
"epoch": 0.011093990755007704,
"grad_norm": 1.660719871520996,
"learning_rate": 4.41717791411043e-06,
"loss": 1.7559,
"step": 18
},
{
"epoch": 0.011710323574730355,
"grad_norm": 1.5748722553253174,
"learning_rate": 4.662576687116564e-06,
"loss": 1.7705,
"step": 19
},
{
"epoch": 0.012326656394453005,
"grad_norm": 1.7511894702911377,
"learning_rate": 4.9079754601227e-06,
"loss": 1.7254,
"step": 20
},
{
"epoch": 0.012942989214175655,
"grad_norm": 1.5858523845672607,
"learning_rate": 5.153374233128835e-06,
"loss": 1.7507,
"step": 21
},
{
"epoch": 0.013559322033898305,
"grad_norm": 1.4151811599731445,
"learning_rate": 5.39877300613497e-06,
"loss": 1.6743,
"step": 22
},
{
"epoch": 0.014175654853620955,
"grad_norm": 1.2708033323287964,
"learning_rate": 5.644171779141104e-06,
"loss": 1.6913,
"step": 23
},
{
"epoch": 0.014791987673343606,
"grad_norm": 1.2384111881256104,
"learning_rate": 5.88957055214724e-06,
"loss": 1.7353,
"step": 24
},
{
"epoch": 0.015408320493066256,
"grad_norm": 1.2881749868392944,
"learning_rate": 6.134969325153375e-06,
"loss": 1.767,
"step": 25
},
{
"epoch": 0.016024653312788906,
"grad_norm": 1.2464679479599,
"learning_rate": 6.38036809815951e-06,
"loss": 1.7554,
"step": 26
},
{
"epoch": 0.016640986132511557,
"grad_norm": 1.24666166305542,
"learning_rate": 6.625766871165644e-06,
"loss": 1.6588,
"step": 27
},
{
"epoch": 0.017257318952234205,
"grad_norm": 1.20462167263031,
"learning_rate": 6.87116564417178e-06,
"loss": 1.7058,
"step": 28
},
{
"epoch": 0.017873651771956857,
"grad_norm": 1.073009967803955,
"learning_rate": 7.116564417177915e-06,
"loss": 1.7216,
"step": 29
},
{
"epoch": 0.01848998459167951,
"grad_norm": 1.090959906578064,
"learning_rate": 7.36196319018405e-06,
"loss": 1.6912,
"step": 30
},
{
"epoch": 0.019106317411402157,
"grad_norm": 1.118828535079956,
"learning_rate": 7.6073619631901856e-06,
"loss": 1.7231,
"step": 31
},
{
"epoch": 0.019722650231124808,
"grad_norm": 1.1499671936035156,
"learning_rate": 7.85276073619632e-06,
"loss": 1.658,
"step": 32
},
{
"epoch": 0.020338983050847456,
"grad_norm": 1.1840991973876953,
"learning_rate": 8.098159509202455e-06,
"loss": 1.679,
"step": 33
},
{
"epoch": 0.020955315870570108,
"grad_norm": 1.049609661102295,
"learning_rate": 8.343558282208589e-06,
"loss": 1.7687,
"step": 34
},
{
"epoch": 0.02157164869029276,
"grad_norm": 1.0060491561889648,
"learning_rate": 8.588957055214725e-06,
"loss": 1.685,
"step": 35
},
{
"epoch": 0.022187981510015407,
"grad_norm": 1.0304157733917236,
"learning_rate": 8.83435582822086e-06,
"loss": 1.6884,
"step": 36
},
{
"epoch": 0.02280431432973806,
"grad_norm": 0.9977303743362427,
"learning_rate": 9.079754601226994e-06,
"loss": 1.6903,
"step": 37
},
{
"epoch": 0.02342064714946071,
"grad_norm": 1.008644938468933,
"learning_rate": 9.325153374233129e-06,
"loss": 1.6521,
"step": 38
},
{
"epoch": 0.02403697996918336,
"grad_norm": 1.046408772468567,
"learning_rate": 9.570552147239264e-06,
"loss": 1.7299,
"step": 39
},
{
"epoch": 0.02465331278890601,
"grad_norm": 1.0336506366729736,
"learning_rate": 9.8159509202454e-06,
"loss": 1.6146,
"step": 40
},
{
"epoch": 0.025269645608628658,
"grad_norm": 1.0193053483963013,
"learning_rate": 1.0061349693251534e-05,
"loss": 1.6733,
"step": 41
},
{
"epoch": 0.02588597842835131,
"grad_norm": 1.0004470348358154,
"learning_rate": 1.030674846625767e-05,
"loss": 1.7239,
"step": 42
},
{
"epoch": 0.02650231124807396,
"grad_norm": 1.0116422176361084,
"learning_rate": 1.0552147239263804e-05,
"loss": 1.7232,
"step": 43
},
{
"epoch": 0.02711864406779661,
"grad_norm": 0.9728295207023621,
"learning_rate": 1.079754601226994e-05,
"loss": 1.6871,
"step": 44
},
{
"epoch": 0.02773497688751926,
"grad_norm": 0.9763592481613159,
"learning_rate": 1.1042944785276076e-05,
"loss": 1.6709,
"step": 45
},
{
"epoch": 0.02835130970724191,
"grad_norm": 0.9567804932594299,
"learning_rate": 1.1288343558282208e-05,
"loss": 1.6082,
"step": 46
},
{
"epoch": 0.02896764252696456,
"grad_norm": 0.9812591671943665,
"learning_rate": 1.1533742331288344e-05,
"loss": 1.6607,
"step": 47
},
{
"epoch": 0.029583975346687212,
"grad_norm": 1.1011967658996582,
"learning_rate": 1.177914110429448e-05,
"loss": 1.7133,
"step": 48
},
{
"epoch": 0.03020030816640986,
"grad_norm": 0.997948944568634,
"learning_rate": 1.2024539877300614e-05,
"loss": 1.6553,
"step": 49
},
{
"epoch": 0.030816640986132512,
"grad_norm": 0.9882333874702454,
"learning_rate": 1.226993865030675e-05,
"loss": 1.6809,
"step": 50
},
{
"epoch": 0.03143297380585516,
"grad_norm": 1.0237276554107666,
"learning_rate": 1.2515337423312886e-05,
"loss": 1.7517,
"step": 51
},
{
"epoch": 0.03204930662557781,
"grad_norm": 0.9705806374549866,
"learning_rate": 1.276073619631902e-05,
"loss": 1.7056,
"step": 52
},
{
"epoch": 0.03266563944530046,
"grad_norm": 0.9457777738571167,
"learning_rate": 1.3006134969325156e-05,
"loss": 1.6469,
"step": 53
},
{
"epoch": 0.033281972265023115,
"grad_norm": 1.01027250289917,
"learning_rate": 1.3251533742331288e-05,
"loss": 1.6744,
"step": 54
},
{
"epoch": 0.03389830508474576,
"grad_norm": 0.9895347356796265,
"learning_rate": 1.3496932515337424e-05,
"loss": 1.6506,
"step": 55
},
{
"epoch": 0.03451463790446841,
"grad_norm": 1.055320143699646,
"learning_rate": 1.374233128834356e-05,
"loss": 1.6655,
"step": 56
},
{
"epoch": 0.035130970724191066,
"grad_norm": 1.0253647565841675,
"learning_rate": 1.3987730061349694e-05,
"loss": 1.6899,
"step": 57
},
{
"epoch": 0.035747303543913714,
"grad_norm": 0.9785537123680115,
"learning_rate": 1.423312883435583e-05,
"loss": 1.6384,
"step": 58
},
{
"epoch": 0.03636363636363636,
"grad_norm": 1.0246118307113647,
"learning_rate": 1.4478527607361965e-05,
"loss": 1.7097,
"step": 59
},
{
"epoch": 0.03697996918335902,
"grad_norm": 1.0297993421554565,
"learning_rate": 1.47239263803681e-05,
"loss": 1.7113,
"step": 60
},
{
"epoch": 0.037596302003081665,
"grad_norm": 1.0864713191986084,
"learning_rate": 1.4969325153374235e-05,
"loss": 1.689,
"step": 61
},
{
"epoch": 0.03821263482280431,
"grad_norm": 0.8881139159202576,
"learning_rate": 1.5214723926380371e-05,
"loss": 1.6708,
"step": 62
},
{
"epoch": 0.03882896764252696,
"grad_norm": 1.0513116121292114,
"learning_rate": 1.5460122699386504e-05,
"loss": 1.6636,
"step": 63
},
{
"epoch": 0.039445300462249616,
"grad_norm": 0.9132838249206543,
"learning_rate": 1.570552147239264e-05,
"loss": 1.6633,
"step": 64
},
{
"epoch": 0.040061633281972264,
"grad_norm": 0.9439066648483276,
"learning_rate": 1.5950920245398772e-05,
"loss": 1.708,
"step": 65
},
{
"epoch": 0.04067796610169491,
"grad_norm": 0.8892993927001953,
"learning_rate": 1.619631901840491e-05,
"loss": 1.6697,
"step": 66
},
{
"epoch": 0.04129429892141757,
"grad_norm": 0.8928653597831726,
"learning_rate": 1.6441717791411043e-05,
"loss": 1.6138,
"step": 67
},
{
"epoch": 0.041910631741140215,
"grad_norm": 0.9380534887313843,
"learning_rate": 1.6687116564417178e-05,
"loss": 1.6451,
"step": 68
},
{
"epoch": 0.042526964560862864,
"grad_norm": 0.9998738169670105,
"learning_rate": 1.6932515337423315e-05,
"loss": 1.7299,
"step": 69
},
{
"epoch": 0.04314329738058552,
"grad_norm": 0.9376276731491089,
"learning_rate": 1.717791411042945e-05,
"loss": 1.6444,
"step": 70
},
{
"epoch": 0.04375963020030817,
"grad_norm": 0.9520479440689087,
"learning_rate": 1.7423312883435583e-05,
"loss": 1.667,
"step": 71
},
{
"epoch": 0.044375963020030815,
"grad_norm": 0.8968117237091064,
"learning_rate": 1.766871165644172e-05,
"loss": 1.6785,
"step": 72
},
{
"epoch": 0.04499229583975347,
"grad_norm": 0.9188838601112366,
"learning_rate": 1.7914110429447855e-05,
"loss": 1.6208,
"step": 73
},
{
"epoch": 0.04560862865947612,
"grad_norm": 0.8976795077323914,
"learning_rate": 1.815950920245399e-05,
"loss": 1.5895,
"step": 74
},
{
"epoch": 0.046224961479198766,
"grad_norm": 0.8823075294494629,
"learning_rate": 1.8404907975460123e-05,
"loss": 1.646,
"step": 75
},
{
"epoch": 0.04684129429892142,
"grad_norm": 0.8592455387115479,
"learning_rate": 1.8650306748466257e-05,
"loss": 1.5647,
"step": 76
},
{
"epoch": 0.04745762711864407,
"grad_norm": 0.8513899445533752,
"learning_rate": 1.8895705521472395e-05,
"loss": 1.6174,
"step": 77
},
{
"epoch": 0.04807395993836672,
"grad_norm": 0.9042425155639648,
"learning_rate": 1.914110429447853e-05,
"loss": 1.634,
"step": 78
},
{
"epoch": 0.048690292758089365,
"grad_norm": 0.8949453234672546,
"learning_rate": 1.9386503067484663e-05,
"loss": 1.5943,
"step": 79
},
{
"epoch": 0.04930662557781202,
"grad_norm": 0.8401027321815491,
"learning_rate": 1.96319018404908e-05,
"loss": 1.5745,
"step": 80
},
{
"epoch": 0.04992295839753467,
"grad_norm": 0.8911848664283752,
"learning_rate": 1.9877300613496935e-05,
"loss": 1.6336,
"step": 81
},
{
"epoch": 0.050539291217257316,
"grad_norm": 0.8529688119888306,
"learning_rate": 2.012269938650307e-05,
"loss": 1.6164,
"step": 82
},
{
"epoch": 0.05115562403697997,
"grad_norm": 0.9154197573661804,
"learning_rate": 2.0368098159509203e-05,
"loss": 1.6758,
"step": 83
},
{
"epoch": 0.05177195685670262,
"grad_norm": 0.866750955581665,
"learning_rate": 2.061349693251534e-05,
"loss": 1.5921,
"step": 84
},
{
"epoch": 0.05238828967642527,
"grad_norm": 0.8651569485664368,
"learning_rate": 2.0858895705521474e-05,
"loss": 1.6166,
"step": 85
},
{
"epoch": 0.05300462249614792,
"grad_norm": 0.9068999290466309,
"learning_rate": 2.110429447852761e-05,
"loss": 1.6648,
"step": 86
},
{
"epoch": 0.05362095531587057,
"grad_norm": 0.8784379959106445,
"learning_rate": 2.1349693251533746e-05,
"loss": 1.6612,
"step": 87
},
{
"epoch": 0.05423728813559322,
"grad_norm": 0.8599914908409119,
"learning_rate": 2.159509202453988e-05,
"loss": 1.6372,
"step": 88
},
{
"epoch": 0.054853620955315874,
"grad_norm": 0.8836574554443359,
"learning_rate": 2.184049079754601e-05,
"loss": 1.6742,
"step": 89
},
{
"epoch": 0.05546995377503852,
"grad_norm": 0.9129705429077148,
"learning_rate": 2.2085889570552152e-05,
"loss": 1.6252,
"step": 90
},
{
"epoch": 0.05608628659476117,
"grad_norm": 0.8584738373756409,
"learning_rate": 2.2331288343558283e-05,
"loss": 1.6668,
"step": 91
},
{
"epoch": 0.05670261941448382,
"grad_norm": 0.8560867309570312,
"learning_rate": 2.2576687116564417e-05,
"loss": 1.5992,
"step": 92
},
{
"epoch": 0.05731895223420647,
"grad_norm": 0.8884056806564331,
"learning_rate": 2.2822085889570554e-05,
"loss": 1.6916,
"step": 93
},
{
"epoch": 0.05793528505392912,
"grad_norm": 0.8534591197967529,
"learning_rate": 2.3067484662576688e-05,
"loss": 1.627,
"step": 94
},
{
"epoch": 0.05855161787365177,
"grad_norm": 0.8151088356971741,
"learning_rate": 2.3312883435582822e-05,
"loss": 1.5843,
"step": 95
},
{
"epoch": 0.059167950693374424,
"grad_norm": 0.8539857864379883,
"learning_rate": 2.355828220858896e-05,
"loss": 1.624,
"step": 96
},
{
"epoch": 0.05978428351309707,
"grad_norm": 0.8533106446266174,
"learning_rate": 2.3803680981595094e-05,
"loss": 1.7183,
"step": 97
},
{
"epoch": 0.06040061633281972,
"grad_norm": 0.9716623425483704,
"learning_rate": 2.4049079754601228e-05,
"loss": 1.7222,
"step": 98
},
{
"epoch": 0.061016949152542375,
"grad_norm": 0.8098641037940979,
"learning_rate": 2.4294478527607366e-05,
"loss": 1.6397,
"step": 99
},
{
"epoch": 0.061633281972265024,
"grad_norm": 0.8411961793899536,
"learning_rate": 2.45398773006135e-05,
"loss": 1.6258,
"step": 100
},
{
"epoch": 0.06224961479198767,
"grad_norm": 0.911243736743927,
"learning_rate": 2.4785276073619634e-05,
"loss": 1.6559,
"step": 101
},
{
"epoch": 0.06286594761171033,
"grad_norm": 0.9482967257499695,
"learning_rate": 2.503067484662577e-05,
"loss": 1.7146,
"step": 102
},
{
"epoch": 0.06348228043143297,
"grad_norm": 0.8271360397338867,
"learning_rate": 2.5276073619631905e-05,
"loss": 1.5776,
"step": 103
},
{
"epoch": 0.06409861325115562,
"grad_norm": 0.8761653900146484,
"learning_rate": 2.552147239263804e-05,
"loss": 1.6654,
"step": 104
},
{
"epoch": 0.06471494607087827,
"grad_norm": 0.8657371401786804,
"learning_rate": 2.576687116564417e-05,
"loss": 1.7236,
"step": 105
},
{
"epoch": 0.06533127889060092,
"grad_norm": 0.810516357421875,
"learning_rate": 2.601226993865031e-05,
"loss": 1.6357,
"step": 106
},
{
"epoch": 0.06594761171032358,
"grad_norm": 0.8823238015174866,
"learning_rate": 2.6257668711656442e-05,
"loss": 1.6358,
"step": 107
},
{
"epoch": 0.06656394453004623,
"grad_norm": 1.0349003076553345,
"learning_rate": 2.6503067484662576e-05,
"loss": 1.6785,
"step": 108
},
{
"epoch": 0.06718027734976888,
"grad_norm": 0.7891526222229004,
"learning_rate": 2.6748466257668714e-05,
"loss": 1.6366,
"step": 109
},
{
"epoch": 0.06779661016949153,
"grad_norm": 0.8495930433273315,
"learning_rate": 2.6993865030674848e-05,
"loss": 1.6392,
"step": 110
},
{
"epoch": 0.06841294298921417,
"grad_norm": 0.8164300322532654,
"learning_rate": 2.7239263803680982e-05,
"loss": 1.6739,
"step": 111
},
{
"epoch": 0.06902927580893682,
"grad_norm": 0.8112594485282898,
"learning_rate": 2.748466257668712e-05,
"loss": 1.6826,
"step": 112
},
{
"epoch": 0.06964560862865947,
"grad_norm": 0.8023330569267273,
"learning_rate": 2.7730061349693253e-05,
"loss": 1.6632,
"step": 113
},
{
"epoch": 0.07026194144838213,
"grad_norm": 0.788388729095459,
"learning_rate": 2.7975460122699388e-05,
"loss": 1.6048,
"step": 114
},
{
"epoch": 0.07087827426810478,
"grad_norm": 0.7653685808181763,
"learning_rate": 2.8220858895705525e-05,
"loss": 1.5936,
"step": 115
},
{
"epoch": 0.07149460708782743,
"grad_norm": 0.7973049283027649,
"learning_rate": 2.846625766871166e-05,
"loss": 1.6586,
"step": 116
},
{
"epoch": 0.07211093990755008,
"grad_norm": 0.7768449783325195,
"learning_rate": 2.8711656441717793e-05,
"loss": 1.5832,
"step": 117
},
{
"epoch": 0.07272727272727272,
"grad_norm": 0.7748964428901672,
"learning_rate": 2.895705521472393e-05,
"loss": 1.6792,
"step": 118
},
{
"epoch": 0.07334360554699537,
"grad_norm": 0.8378171920776367,
"learning_rate": 2.9202453987730065e-05,
"loss": 1.6247,
"step": 119
},
{
"epoch": 0.07395993836671803,
"grad_norm": 0.7740994691848755,
"learning_rate": 2.94478527607362e-05,
"loss": 1.6401,
"step": 120
},
{
"epoch": 0.07457627118644068,
"grad_norm": 0.7645397186279297,
"learning_rate": 2.9693251533742337e-05,
"loss": 1.6607,
"step": 121
},
{
"epoch": 0.07519260400616333,
"grad_norm": 0.7948278188705444,
"learning_rate": 2.993865030674847e-05,
"loss": 1.6642,
"step": 122
},
{
"epoch": 0.07580893682588598,
"grad_norm": 0.7775226831436157,
"learning_rate": 3.01840490797546e-05,
"loss": 1.6617,
"step": 123
},
{
"epoch": 0.07642526964560863,
"grad_norm": 0.8228120803833008,
"learning_rate": 3.0429447852760742e-05,
"loss": 1.5793,
"step": 124
},
{
"epoch": 0.07704160246533127,
"grad_norm": 0.8086126446723938,
"learning_rate": 3.067484662576687e-05,
"loss": 1.6518,
"step": 125
},
{
"epoch": 0.07765793528505392,
"grad_norm": 0.7439844608306885,
"learning_rate": 3.092024539877301e-05,
"loss": 1.6284,
"step": 126
},
{
"epoch": 0.07827426810477658,
"grad_norm": 0.8310662508010864,
"learning_rate": 3.116564417177915e-05,
"loss": 1.6417,
"step": 127
},
{
"epoch": 0.07889060092449923,
"grad_norm": 0.8119212985038757,
"learning_rate": 3.141104294478528e-05,
"loss": 1.6033,
"step": 128
},
{
"epoch": 0.07950693374422188,
"grad_norm": 0.7821022272109985,
"learning_rate": 3.1656441717791416e-05,
"loss": 1.6896,
"step": 129
},
{
"epoch": 0.08012326656394453,
"grad_norm": 0.7965499758720398,
"learning_rate": 3.1901840490797544e-05,
"loss": 1.6492,
"step": 130
},
{
"epoch": 0.08073959938366718,
"grad_norm": 0.7747324109077454,
"learning_rate": 3.2147239263803684e-05,
"loss": 1.6301,
"step": 131
},
{
"epoch": 0.08135593220338982,
"grad_norm": 0.7778410911560059,
"learning_rate": 3.239263803680982e-05,
"loss": 1.5949,
"step": 132
},
{
"epoch": 0.08197226502311249,
"grad_norm": 0.8136535286903381,
"learning_rate": 3.263803680981595e-05,
"loss": 1.5787,
"step": 133
},
{
"epoch": 0.08258859784283513,
"grad_norm": 0.7390981316566467,
"learning_rate": 3.288343558282209e-05,
"loss": 1.6162,
"step": 134
},
{
"epoch": 0.08320493066255778,
"grad_norm": 0.8133728504180908,
"learning_rate": 3.312883435582822e-05,
"loss": 1.6204,
"step": 135
},
{
"epoch": 0.08382126348228043,
"grad_norm": 0.7785617113113403,
"learning_rate": 3.3374233128834355e-05,
"loss": 1.664,
"step": 136
},
{
"epoch": 0.08443759630200308,
"grad_norm": 0.8097333908081055,
"learning_rate": 3.3619631901840496e-05,
"loss": 1.6337,
"step": 137
},
{
"epoch": 0.08505392912172573,
"grad_norm": 0.7560635209083557,
"learning_rate": 3.386503067484663e-05,
"loss": 1.6624,
"step": 138
},
{
"epoch": 0.08567026194144839,
"grad_norm": 0.735256016254425,
"learning_rate": 3.4110429447852764e-05,
"loss": 1.6295,
"step": 139
},
{
"epoch": 0.08628659476117104,
"grad_norm": 0.8018369674682617,
"learning_rate": 3.43558282208589e-05,
"loss": 1.6347,
"step": 140
},
{
"epoch": 0.08690292758089369,
"grad_norm": 0.7610778212547302,
"learning_rate": 3.460122699386503e-05,
"loss": 1.6675,
"step": 141
},
{
"epoch": 0.08751926040061633,
"grad_norm": 0.7676482200622559,
"learning_rate": 3.4846625766871166e-05,
"loss": 1.7137,
"step": 142
},
{
"epoch": 0.08813559322033898,
"grad_norm": 0.7479026913642883,
"learning_rate": 3.509202453987731e-05,
"loss": 1.7188,
"step": 143
},
{
"epoch": 0.08875192604006163,
"grad_norm": 0.7722792625427246,
"learning_rate": 3.533742331288344e-05,
"loss": 1.6932,
"step": 144
},
{
"epoch": 0.08936825885978428,
"grad_norm": 0.7511256337165833,
"learning_rate": 3.5582822085889576e-05,
"loss": 1.6144,
"step": 145
},
{
"epoch": 0.08998459167950694,
"grad_norm": 0.7605886459350586,
"learning_rate": 3.582822085889571e-05,
"loss": 1.635,
"step": 146
},
{
"epoch": 0.09060092449922959,
"grad_norm": 0.7359566688537598,
"learning_rate": 3.6073619631901844e-05,
"loss": 1.6192,
"step": 147
},
{
"epoch": 0.09121725731895224,
"grad_norm": 0.7442858815193176,
"learning_rate": 3.631901840490798e-05,
"loss": 1.5986,
"step": 148
},
{
"epoch": 0.09183359013867488,
"grad_norm": 0.7447007298469543,
"learning_rate": 3.656441717791412e-05,
"loss": 1.5948,
"step": 149
},
{
"epoch": 0.09244992295839753,
"grad_norm": 0.7689185738563538,
"learning_rate": 3.6809815950920246e-05,
"loss": 1.6534,
"step": 150
},
{
"epoch": 0.09306625577812018,
"grad_norm": 0.781383752822876,
"learning_rate": 3.705521472392638e-05,
"loss": 1.6269,
"step": 151
},
{
"epoch": 0.09368258859784284,
"grad_norm": 0.7619852423667908,
"learning_rate": 3.7300613496932514e-05,
"loss": 1.5797,
"step": 152
},
{
"epoch": 0.09429892141756549,
"grad_norm": 0.7769779562950134,
"learning_rate": 3.7546012269938655e-05,
"loss": 1.5976,
"step": 153
},
{
"epoch": 0.09491525423728814,
"grad_norm": 0.7349963188171387,
"learning_rate": 3.779141104294479e-05,
"loss": 1.6322,
"step": 154
},
{
"epoch": 0.09553158705701079,
"grad_norm": 0.7396400570869446,
"learning_rate": 3.8036809815950924e-05,
"loss": 1.6398,
"step": 155
},
{
"epoch": 0.09614791987673343,
"grad_norm": 0.7363472580909729,
"learning_rate": 3.828220858895706e-05,
"loss": 1.6313,
"step": 156
},
{
"epoch": 0.09676425269645608,
"grad_norm": 0.7341015338897705,
"learning_rate": 3.852760736196319e-05,
"loss": 1.7016,
"step": 157
},
{
"epoch": 0.09738058551617873,
"grad_norm": 0.6993916034698486,
"learning_rate": 3.8773006134969326e-05,
"loss": 1.5703,
"step": 158
},
{
"epoch": 0.09799691833590139,
"grad_norm": 0.7287641167640686,
"learning_rate": 3.901840490797547e-05,
"loss": 1.6438,
"step": 159
},
{
"epoch": 0.09861325115562404,
"grad_norm": 0.7177438139915466,
"learning_rate": 3.92638036809816e-05,
"loss": 1.6235,
"step": 160
},
{
"epoch": 0.09922958397534669,
"grad_norm": 0.6875810623168945,
"learning_rate": 3.9509202453987735e-05,
"loss": 1.593,
"step": 161
},
{
"epoch": 0.09984591679506934,
"grad_norm": 0.7038248777389526,
"learning_rate": 3.975460122699387e-05,
"loss": 1.6023,
"step": 162
},
{
"epoch": 0.10046224961479198,
"grad_norm": 0.687907338142395,
"learning_rate": 4e-05,
"loss": 1.608,
"step": 163
},
{
"epoch": 0.10107858243451463,
"grad_norm": 0.7126633524894714,
"learning_rate": 3.999995363513562e-05,
"loss": 1.6547,
"step": 164
},
{
"epoch": 0.1016949152542373,
"grad_norm": 0.7042025923728943,
"learning_rate": 3.999981454075744e-05,
"loss": 1.667,
"step": 165
},
{
"epoch": 0.10231124807395994,
"grad_norm": 0.6862210631370544,
"learning_rate": 3.999958271751038e-05,
"loss": 1.6081,
"step": 166
},
{
"epoch": 0.10292758089368259,
"grad_norm": 0.6989213824272156,
"learning_rate": 3.999925816646928e-05,
"loss": 1.6115,
"step": 167
},
{
"epoch": 0.10354391371340524,
"grad_norm": 0.690365195274353,
"learning_rate": 3.9998840889138925e-05,
"loss": 1.5594,
"step": 168
},
{
"epoch": 0.10416024653312789,
"grad_norm": 0.6726807951927185,
"learning_rate": 3.9998330887454e-05,
"loss": 1.6238,
"step": 169
},
{
"epoch": 0.10477657935285054,
"grad_norm": 0.6911548972129822,
"learning_rate": 3.9997728163779134e-05,
"loss": 1.6105,
"step": 170
},
{
"epoch": 0.10539291217257318,
"grad_norm": 0.6829053163528442,
"learning_rate": 3.999703272090884e-05,
"loss": 1.6448,
"step": 171
},
{
"epoch": 0.10600924499229585,
"grad_norm": 0.6991954445838928,
"learning_rate": 3.9996244562067525e-05,
"loss": 1.5788,
"step": 172
},
{
"epoch": 0.1066255778120185,
"grad_norm": 0.6998047828674316,
"learning_rate": 3.9995363690909487e-05,
"loss": 1.7358,
"step": 173
},
{
"epoch": 0.10724191063174114,
"grad_norm": 0.7109401822090149,
"learning_rate": 3.999439011151887e-05,
"loss": 1.5783,
"step": 174
},
{
"epoch": 0.10785824345146379,
"grad_norm": 0.7161952257156372,
"learning_rate": 3.9993323828409666e-05,
"loss": 1.6098,
"step": 175
},
{
"epoch": 0.10847457627118644,
"grad_norm": 0.7027195692062378,
"learning_rate": 3.999216484652567e-05,
"loss": 1.6,
"step": 176
},
{
"epoch": 0.10909090909090909,
"grad_norm": 0.6891278028488159,
"learning_rate": 3.999091317124049e-05,
"loss": 1.6253,
"step": 177
},
{
"epoch": 0.10970724191063175,
"grad_norm": 0.6901974081993103,
"learning_rate": 3.998956880835751e-05,
"loss": 1.6493,
"step": 178
},
{
"epoch": 0.1103235747303544,
"grad_norm": 0.7213894724845886,
"learning_rate": 3.998813176410985e-05,
"loss": 1.6286,
"step": 179
},
{
"epoch": 0.11093990755007704,
"grad_norm": 0.6912412643432617,
"learning_rate": 3.998660204516033e-05,
"loss": 1.5866,
"step": 180
},
{
"epoch": 0.11155624036979969,
"grad_norm": 0.6759384274482727,
"learning_rate": 3.998497965860149e-05,
"loss": 1.6171,
"step": 181
},
{
"epoch": 0.11217257318952234,
"grad_norm": 0.7188528180122375,
"learning_rate": 3.9983264611955496e-05,
"loss": 1.7641,
"step": 182
},
{
"epoch": 0.11278890600924499,
"grad_norm": 0.6806765198707581,
"learning_rate": 3.9981456913174135e-05,
"loss": 1.6582,
"step": 183
},
{
"epoch": 0.11340523882896764,
"grad_norm": 0.6784571409225464,
"learning_rate": 3.997955657063878e-05,
"loss": 1.5771,
"step": 184
},
{
"epoch": 0.1140215716486903,
"grad_norm": 0.7074373364448547,
"learning_rate": 3.997756359316035e-05,
"loss": 1.5754,
"step": 185
},
{
"epoch": 0.11463790446841295,
"grad_norm": 0.6583380699157715,
"learning_rate": 3.997547798997924e-05,
"loss": 1.6057,
"step": 186
},
{
"epoch": 0.1152542372881356,
"grad_norm": 0.6994054913520813,
"learning_rate": 3.9973299770765345e-05,
"loss": 1.5706,
"step": 187
},
{
"epoch": 0.11587057010785824,
"grad_norm": 0.6755475401878357,
"learning_rate": 3.997102894561793e-05,
"loss": 1.6144,
"step": 188
},
{
"epoch": 0.11648690292758089,
"grad_norm": 0.6866325736045837,
"learning_rate": 3.9968665525065654e-05,
"loss": 1.6255,
"step": 189
},
{
"epoch": 0.11710323574730354,
"grad_norm": 0.6652048230171204,
"learning_rate": 3.996620952006648e-05,
"loss": 1.605,
"step": 190
},
{
"epoch": 0.1177195685670262,
"grad_norm": 0.6833721399307251,
"learning_rate": 3.996366094200765e-05,
"loss": 1.6686,
"step": 191
},
{
"epoch": 0.11833590138674885,
"grad_norm": 0.6793074607849121,
"learning_rate": 3.9961019802705603e-05,
"loss": 1.637,
"step": 192
},
{
"epoch": 0.1189522342064715,
"grad_norm": 0.6983605623245239,
"learning_rate": 3.995828611440595e-05,
"loss": 1.6507,
"step": 193
},
{
"epoch": 0.11956856702619414,
"grad_norm": 0.6783537864685059,
"learning_rate": 3.995545988978339e-05,
"loss": 1.538,
"step": 194
},
{
"epoch": 0.12018489984591679,
"grad_norm": 0.6404557824134827,
"learning_rate": 3.9952541141941695e-05,
"loss": 1.5394,
"step": 195
},
{
"epoch": 0.12080123266563944,
"grad_norm": 0.6974167823791504,
"learning_rate": 3.994952988441358e-05,
"loss": 1.6152,
"step": 196
},
{
"epoch": 0.12141756548536209,
"grad_norm": 0.7140214443206787,
"learning_rate": 3.9946426131160706e-05,
"loss": 1.6507,
"step": 197
},
{
"epoch": 0.12203389830508475,
"grad_norm": 0.6546657681465149,
"learning_rate": 3.9943229896573594e-05,
"loss": 1.6298,
"step": 198
},
{
"epoch": 0.1226502311248074,
"grad_norm": 0.6407018303871155,
"learning_rate": 3.993994119547153e-05,
"loss": 1.5801,
"step": 199
},
{
"epoch": 0.12326656394453005,
"grad_norm": 0.6628043055534363,
"learning_rate": 3.993656004310253e-05,
"loss": 1.5879,
"step": 200
},
{
"epoch": 0.1238828967642527,
"grad_norm": 0.6629613637924194,
"learning_rate": 3.993308645514327e-05,
"loss": 1.5517,
"step": 201
},
{
"epoch": 0.12449922958397534,
"grad_norm": 0.6973757743835449,
"learning_rate": 3.992952044769898e-05,
"loss": 1.6574,
"step": 202
},
{
"epoch": 0.125115562403698,
"grad_norm": 0.677291750907898,
"learning_rate": 3.9925862037303425e-05,
"loss": 1.5528,
"step": 203
},
{
"epoch": 0.12573189522342065,
"grad_norm": 0.6733261942863464,
"learning_rate": 3.992211124091876e-05,
"loss": 1.6105,
"step": 204
},
{
"epoch": 0.1263482280431433,
"grad_norm": 0.6975075602531433,
"learning_rate": 3.9918268075935514e-05,
"loss": 1.6032,
"step": 205
},
{
"epoch": 0.12696456086286595,
"grad_norm": 0.6801272034645081,
"learning_rate": 3.991433256017246e-05,
"loss": 1.6238,
"step": 206
},
{
"epoch": 0.1275808936825886,
"grad_norm": 0.6633975505828857,
"learning_rate": 3.9910304711876556e-05,
"loss": 1.5586,
"step": 207
},
{
"epoch": 0.12819722650231125,
"grad_norm": 0.6919019818305969,
"learning_rate": 3.990618454972288e-05,
"loss": 1.5929,
"step": 208
},
{
"epoch": 0.1288135593220339,
"grad_norm": 0.6731187105178833,
"learning_rate": 3.9901972092814504e-05,
"loss": 1.6332,
"step": 209
},
{
"epoch": 0.12942989214175654,
"grad_norm": 0.6671748757362366,
"learning_rate": 3.989766736068242e-05,
"loss": 1.6109,
"step": 210
},
{
"epoch": 0.1300462249614792,
"grad_norm": 0.6681280136108398,
"learning_rate": 3.989327037328548e-05,
"loss": 1.5844,
"step": 211
},
{
"epoch": 0.13066255778120184,
"grad_norm": 0.7066627740859985,
"learning_rate": 3.988878115101024e-05,
"loss": 1.6193,
"step": 212
},
{
"epoch": 0.13127889060092449,
"grad_norm": 0.6514896750450134,
"learning_rate": 3.9884199714670914e-05,
"loss": 1.6017,
"step": 213
},
{
"epoch": 0.13189522342064716,
"grad_norm": 0.689854621887207,
"learning_rate": 3.987952608550928e-05,
"loss": 1.6162,
"step": 214
},
{
"epoch": 0.1325115562403698,
"grad_norm": 0.6585028171539307,
"learning_rate": 3.9874760285194554e-05,
"loss": 1.5866,
"step": 215
},
{
"epoch": 0.13312788906009246,
"grad_norm": 0.6556795239448547,
"learning_rate": 3.9869902335823296e-05,
"loss": 1.6076,
"step": 216
},
{
"epoch": 0.1337442218798151,
"grad_norm": 0.6920654773712158,
"learning_rate": 3.986495225991934e-05,
"loss": 1.593,
"step": 217
},
{
"epoch": 0.13436055469953775,
"grad_norm": 0.678615152835846,
"learning_rate": 3.985991008043363e-05,
"loss": 1.6143,
"step": 218
},
{
"epoch": 0.1349768875192604,
"grad_norm": 0.6753882169723511,
"learning_rate": 3.985477582074417e-05,
"loss": 1.6432,
"step": 219
},
{
"epoch": 0.13559322033898305,
"grad_norm": 0.6829013228416443,
"learning_rate": 3.984954950465589e-05,
"loss": 1.5916,
"step": 220
},
{
"epoch": 0.1362095531587057,
"grad_norm": 0.6789626479148865,
"learning_rate": 3.984423115640051e-05,
"loss": 1.5781,
"step": 221
},
{
"epoch": 0.13682588597842835,
"grad_norm": 0.7018295526504517,
"learning_rate": 3.983882080063652e-05,
"loss": 1.6833,
"step": 222
},
{
"epoch": 0.137442218798151,
"grad_norm": 0.6346052289009094,
"learning_rate": 3.983331846244892e-05,
"loss": 1.5468,
"step": 223
},
{
"epoch": 0.13805855161787364,
"grad_norm": 0.6694582104682922,
"learning_rate": 3.982772416734925e-05,
"loss": 1.5935,
"step": 224
},
{
"epoch": 0.1386748844375963,
"grad_norm": 0.6665736436843872,
"learning_rate": 3.982203794127538e-05,
"loss": 1.5285,
"step": 225
},
{
"epoch": 0.13929121725731894,
"grad_norm": 0.6625133156776428,
"learning_rate": 3.9816259810591416e-05,
"loss": 1.6179,
"step": 226
},
{
"epoch": 0.13990755007704161,
"grad_norm": 0.6552127599716187,
"learning_rate": 3.981038980208759e-05,
"loss": 1.6083,
"step": 227
},
{
"epoch": 0.14052388289676426,
"grad_norm": 0.6420478224754333,
"learning_rate": 3.98044279429801e-05,
"loss": 1.5898,
"step": 228
},
{
"epoch": 0.1411402157164869,
"grad_norm": 0.6572580337524414,
"learning_rate": 3.979837426091105e-05,
"loss": 1.6002,
"step": 229
},
{
"epoch": 0.14175654853620956,
"grad_norm": 0.6583701968193054,
"learning_rate": 3.979222878394823e-05,
"loss": 1.5608,
"step": 230
},
{
"epoch": 0.1423728813559322,
"grad_norm": 0.670501708984375,
"learning_rate": 3.978599154058508e-05,
"loss": 1.6096,
"step": 231
},
{
"epoch": 0.14298921417565486,
"grad_norm": 0.6815168857574463,
"learning_rate": 3.977966255974048e-05,
"loss": 1.6108,
"step": 232
},
{
"epoch": 0.1436055469953775,
"grad_norm": 0.6337635517120361,
"learning_rate": 3.9773241870758665e-05,
"loss": 1.6007,
"step": 233
},
{
"epoch": 0.14422187981510015,
"grad_norm": 0.692473292350769,
"learning_rate": 3.976672950340909e-05,
"loss": 1.633,
"step": 234
},
{
"epoch": 0.1448382126348228,
"grad_norm": 0.6315524578094482,
"learning_rate": 3.9760125487886235e-05,
"loss": 1.5258,
"step": 235
},
{
"epoch": 0.14545454545454545,
"grad_norm": 0.6339032053947449,
"learning_rate": 3.975342985480955e-05,
"loss": 1.561,
"step": 236
},
{
"epoch": 0.1460708782742681,
"grad_norm": 0.6732640862464905,
"learning_rate": 3.974664263522322e-05,
"loss": 1.6279,
"step": 237
},
{
"epoch": 0.14668721109399074,
"grad_norm": 0.6514915227890015,
"learning_rate": 3.973976386059612e-05,
"loss": 1.5986,
"step": 238
},
{
"epoch": 0.1473035439137134,
"grad_norm": 0.6529800295829773,
"learning_rate": 3.973279356282159e-05,
"loss": 1.5458,
"step": 239
},
{
"epoch": 0.14791987673343607,
"grad_norm": 0.6552094221115112,
"learning_rate": 3.9725731774217316e-05,
"loss": 1.5856,
"step": 240
},
{
"epoch": 0.14853620955315872,
"grad_norm": 0.6603219509124756,
"learning_rate": 3.971857852752518e-05,
"loss": 1.6069,
"step": 241
},
{
"epoch": 0.14915254237288136,
"grad_norm": 0.6710034608840942,
"learning_rate": 3.971133385591113e-05,
"loss": 1.6145,
"step": 242
},
{
"epoch": 0.149768875192604,
"grad_norm": 0.648938775062561,
"learning_rate": 3.970399779296498e-05,
"loss": 1.5991,
"step": 243
},
{
"epoch": 0.15038520801232666,
"grad_norm": 0.6794759035110474,
"learning_rate": 3.9696570372700284e-05,
"loss": 1.5898,
"step": 244
},
{
"epoch": 0.1510015408320493,
"grad_norm": 0.6564098596572876,
"learning_rate": 3.968905162955417e-05,
"loss": 1.6538,
"step": 245
},
{
"epoch": 0.15161787365177196,
"grad_norm": 0.6441242694854736,
"learning_rate": 3.968144159838719e-05,
"loss": 1.5993,
"step": 246
},
{
"epoch": 0.1522342064714946,
"grad_norm": 0.6316927671432495,
"learning_rate": 3.967374031448317e-05,
"loss": 1.5541,
"step": 247
},
{
"epoch": 0.15285053929121725,
"grad_norm": 0.6498556137084961,
"learning_rate": 3.966594781354899e-05,
"loss": 1.5799,
"step": 248
},
{
"epoch": 0.1534668721109399,
"grad_norm": 0.6671091914176941,
"learning_rate": 3.965806413171447e-05,
"loss": 1.5254,
"step": 249
},
{
"epoch": 0.15408320493066255,
"grad_norm": 0.6488400101661682,
"learning_rate": 3.965008930553221e-05,
"loss": 1.5769,
"step": 250
},
{
"epoch": 0.1546995377503852,
"grad_norm": 0.6346517205238342,
"learning_rate": 3.964202337197737e-05,
"loss": 1.5823,
"step": 251
},
{
"epoch": 0.15531587057010784,
"grad_norm": 0.6554057598114014,
"learning_rate": 3.963386636844755e-05,
"loss": 1.7108,
"step": 252
},
{
"epoch": 0.15593220338983052,
"grad_norm": 0.6496933102607727,
"learning_rate": 3.9625618332762583e-05,
"loss": 1.6616,
"step": 253
},
{
"epoch": 0.15654853620955317,
"grad_norm": 0.6630948185920715,
"learning_rate": 3.9617279303164374e-05,
"loss": 1.6548,
"step": 254
},
{
"epoch": 0.15716486902927582,
"grad_norm": 0.6255600452423096,
"learning_rate": 3.960884931831673e-05,
"loss": 1.5547,
"step": 255
},
{
"epoch": 0.15778120184899846,
"grad_norm": 0.6128103733062744,
"learning_rate": 3.960032841730514e-05,
"loss": 1.5376,
"step": 256
},
{
"epoch": 0.1583975346687211,
"grad_norm": 0.6571413278579712,
"learning_rate": 3.959171663963667e-05,
"loss": 1.6524,
"step": 257
},
{
"epoch": 0.15901386748844376,
"grad_norm": 0.6260244846343994,
"learning_rate": 3.958301402523969e-05,
"loss": 1.5668,
"step": 258
},
{
"epoch": 0.1596302003081664,
"grad_norm": 0.6612125039100647,
"learning_rate": 3.957422061446377e-05,
"loss": 1.6241,
"step": 259
},
{
"epoch": 0.16024653312788906,
"grad_norm": 0.6285637021064758,
"learning_rate": 3.956533644807943e-05,
"loss": 1.5283,
"step": 260
},
{
"epoch": 0.1608628659476117,
"grad_norm": 0.6467995047569275,
"learning_rate": 3.9556361567277995e-05,
"loss": 1.5803,
"step": 261
},
{
"epoch": 0.16147919876733435,
"grad_norm": 0.6415068507194519,
"learning_rate": 3.9547296013671365e-05,
"loss": 1.6484,
"step": 262
},
{
"epoch": 0.162095531587057,
"grad_norm": 0.6287012100219727,
"learning_rate": 3.953813982929187e-05,
"loss": 1.5626,
"step": 263
},
{
"epoch": 0.16271186440677965,
"grad_norm": 0.6504274606704712,
"learning_rate": 3.952889305659203e-05,
"loss": 1.5822,
"step": 264
},
{
"epoch": 0.1633281972265023,
"grad_norm": 0.6381985545158386,
"learning_rate": 3.951955573844437e-05,
"loss": 1.5835,
"step": 265
},
{
"epoch": 0.16394453004622497,
"grad_norm": 0.6651602387428284,
"learning_rate": 3.951012791814126e-05,
"loss": 1.6438,
"step": 266
},
{
"epoch": 0.16456086286594762,
"grad_norm": 0.6480642557144165,
"learning_rate": 3.950060963939465e-05,
"loss": 1.5801,
"step": 267
},
{
"epoch": 0.16517719568567027,
"grad_norm": 0.6354721188545227,
"learning_rate": 3.9491000946335904e-05,
"loss": 1.5894,
"step": 268
},
{
"epoch": 0.16579352850539292,
"grad_norm": 0.6333348751068115,
"learning_rate": 3.9481301883515605e-05,
"loss": 1.589,
"step": 269
},
{
"epoch": 0.16640986132511557,
"grad_norm": 0.6304234266281128,
"learning_rate": 3.947151249590333e-05,
"loss": 1.5163,
"step": 270
},
{
"epoch": 0.1670261941448382,
"grad_norm": 0.6365682482719421,
"learning_rate": 3.946163282888743e-05,
"loss": 1.5537,
"step": 271
},
{
"epoch": 0.16764252696456086,
"grad_norm": 0.6680180430412292,
"learning_rate": 3.945166292827486e-05,
"loss": 1.5785,
"step": 272
},
{
"epoch": 0.1682588597842835,
"grad_norm": 0.6543309092521667,
"learning_rate": 3.944160284029092e-05,
"loss": 1.6562,
"step": 273
},
{
"epoch": 0.16887519260400616,
"grad_norm": 0.6407747268676758,
"learning_rate": 3.943145261157908e-05,
"loss": 1.592,
"step": 274
},
{
"epoch": 0.1694915254237288,
"grad_norm": 0.6593438982963562,
"learning_rate": 3.942121228920072e-05,
"loss": 1.567,
"step": 275
},
{
"epoch": 0.17010785824345145,
"grad_norm": 0.6593472361564636,
"learning_rate": 3.9410881920634974e-05,
"loss": 1.6531,
"step": 276
},
{
"epoch": 0.1707241910631741,
"grad_norm": 0.6622447371482849,
"learning_rate": 3.9400461553778455e-05,
"loss": 1.5938,
"step": 277
},
{
"epoch": 0.17134052388289678,
"grad_norm": 0.6343011856079102,
"learning_rate": 3.938995123694504e-05,
"loss": 1.5481,
"step": 278
},
{
"epoch": 0.17195685670261943,
"grad_norm": 0.6643282771110535,
"learning_rate": 3.937935101886568e-05,
"loss": 1.5363,
"step": 279
},
{
"epoch": 0.17257318952234207,
"grad_norm": 0.6595264077186584,
"learning_rate": 3.9368660948688135e-05,
"loss": 1.5831,
"step": 280
},
{
"epoch": 0.17318952234206472,
"grad_norm": 0.6727797389030457,
"learning_rate": 3.935788107597678e-05,
"loss": 1.5718,
"step": 281
},
{
"epoch": 0.17380585516178737,
"grad_norm": 0.6271126866340637,
"learning_rate": 3.934701145071235e-05,
"loss": 1.6389,
"step": 282
},
{
"epoch": 0.17442218798151002,
"grad_norm": 0.6499379277229309,
"learning_rate": 3.93360521232917e-05,
"loss": 1.5553,
"step": 283
},
{
"epoch": 0.17503852080123267,
"grad_norm": 0.6241937875747681,
"learning_rate": 3.932500314452762e-05,
"loss": 1.5052,
"step": 284
},
{
"epoch": 0.17565485362095531,
"grad_norm": 0.6471588015556335,
"learning_rate": 3.931386456564854e-05,
"loss": 1.6057,
"step": 285
},
{
"epoch": 0.17627118644067796,
"grad_norm": 0.654523491859436,
"learning_rate": 3.9302636438298334e-05,
"loss": 1.6026,
"step": 286
},
{
"epoch": 0.1768875192604006,
"grad_norm": 0.6462265253067017,
"learning_rate": 3.929131881453606e-05,
"loss": 1.6083,
"step": 287
},
{
"epoch": 0.17750385208012326,
"grad_norm": 0.6588643789291382,
"learning_rate": 3.9279911746835726e-05,
"loss": 1.6134,
"step": 288
},
{
"epoch": 0.1781201848998459,
"grad_norm": 0.6225916147232056,
"learning_rate": 3.926841528808604e-05,
"loss": 1.5746,
"step": 289
},
{
"epoch": 0.17873651771956856,
"grad_norm": 0.647373616695404,
"learning_rate": 3.925682949159019e-05,
"loss": 1.6009,
"step": 290
},
{
"epoch": 0.17935285053929123,
"grad_norm": 0.635344922542572,
"learning_rate": 3.9245154411065566e-05,
"loss": 1.583,
"step": 291
},
{
"epoch": 0.17996918335901388,
"grad_norm": 0.6447535753250122,
"learning_rate": 3.9233390100643515e-05,
"loss": 1.6787,
"step": 292
},
{
"epoch": 0.18058551617873653,
"grad_norm": 0.6271030306816101,
"learning_rate": 3.92215366148691e-05,
"loss": 1.5895,
"step": 293
},
{
"epoch": 0.18120184899845918,
"grad_norm": 0.652759850025177,
"learning_rate": 3.9209594008700846e-05,
"loss": 1.6372,
"step": 294
},
{
"epoch": 0.18181818181818182,
"grad_norm": 0.6784529089927673,
"learning_rate": 3.9197562337510485e-05,
"loss": 1.6218,
"step": 295
},
{
"epoch": 0.18243451463790447,
"grad_norm": 0.637200117111206,
"learning_rate": 3.91854416570827e-05,
"loss": 1.5875,
"step": 296
},
{
"epoch": 0.18305084745762712,
"grad_norm": 0.63194340467453,
"learning_rate": 3.9173232023614866e-05,
"loss": 1.5507,
"step": 297
},
{
"epoch": 0.18366718027734977,
"grad_norm": 0.6260672211647034,
"learning_rate": 3.916093349371678e-05,
"loss": 1.5402,
"step": 298
},
{
"epoch": 0.18428351309707242,
"grad_norm": 0.6279014348983765,
"learning_rate": 3.91485461244104e-05,
"loss": 1.5537,
"step": 299
},
{
"epoch": 0.18489984591679506,
"grad_norm": 0.6495095491409302,
"learning_rate": 3.913606997312961e-05,
"loss": 1.6009,
"step": 300
},
{
"epoch": 0.1855161787365177,
"grad_norm": 0.6379627585411072,
"learning_rate": 3.9123505097719914e-05,
"loss": 1.6117,
"step": 301
},
{
"epoch": 0.18613251155624036,
"grad_norm": 0.6391900181770325,
"learning_rate": 3.9110851556438174e-05,
"loss": 1.6022,
"step": 302
},
{
"epoch": 0.186748844375963,
"grad_norm": 0.6260542273521423,
"learning_rate": 3.909810940795237e-05,
"loss": 1.5987,
"step": 303
},
{
"epoch": 0.18736517719568568,
"grad_norm": 0.6432803273200989,
"learning_rate": 3.90852787113413e-05,
"loss": 1.6286,
"step": 304
},
{
"epoch": 0.18798151001540833,
"grad_norm": 0.6564033031463623,
"learning_rate": 3.9072359526094315e-05,
"loss": 1.5856,
"step": 305
},
{
"epoch": 0.18859784283513098,
"grad_norm": 0.6385253071784973,
"learning_rate": 3.9059351912111044e-05,
"loss": 1.5412,
"step": 306
},
{
"epoch": 0.18921417565485363,
"grad_norm": 0.6348027586936951,
"learning_rate": 3.9046255929701116e-05,
"loss": 1.6031,
"step": 307
},
{
"epoch": 0.18983050847457628,
"grad_norm": 0.6338834762573242,
"learning_rate": 3.903307163958386e-05,
"loss": 1.5937,
"step": 308
},
{
"epoch": 0.19044684129429892,
"grad_norm": 0.6354376077651978,
"learning_rate": 3.901979910288809e-05,
"loss": 1.5642,
"step": 309
},
{
"epoch": 0.19106317411402157,
"grad_norm": 0.6317285299301147,
"learning_rate": 3.900643838115171e-05,
"loss": 1.5423,
"step": 310
},
{
"epoch": 0.19167950693374422,
"grad_norm": 0.6343933343887329,
"learning_rate": 3.899298953632154e-05,
"loss": 1.5861,
"step": 311
},
{
"epoch": 0.19229583975346687,
"grad_norm": 0.6410644054412842,
"learning_rate": 3.897945263075296e-05,
"loss": 1.5917,
"step": 312
},
{
"epoch": 0.19291217257318952,
"grad_norm": 0.6140625476837158,
"learning_rate": 3.896582772720966e-05,
"loss": 1.5459,
"step": 313
},
{
"epoch": 0.19352850539291216,
"grad_norm": 0.6385317444801331,
"learning_rate": 3.8952114888863304e-05,
"loss": 1.5529,
"step": 314
},
{
"epoch": 0.1941448382126348,
"grad_norm": 0.6224085688591003,
"learning_rate": 3.8938314179293305e-05,
"loss": 1.5449,
"step": 315
},
{
"epoch": 0.19476117103235746,
"grad_norm": 0.6194934248924255,
"learning_rate": 3.892442566248644e-05,
"loss": 1.5461,
"step": 316
},
{
"epoch": 0.19537750385208014,
"grad_norm": 0.624581515789032,
"learning_rate": 3.8910449402836645e-05,
"loss": 1.5707,
"step": 317
},
{
"epoch": 0.19599383667180278,
"grad_norm": 0.6206932663917542,
"learning_rate": 3.889638546514465e-05,
"loss": 1.5833,
"step": 318
},
{
"epoch": 0.19661016949152543,
"grad_norm": 0.636967122554779,
"learning_rate": 3.8882233914617715e-05,
"loss": 1.6226,
"step": 319
},
{
"epoch": 0.19722650231124808,
"grad_norm": 0.6057730913162231,
"learning_rate": 3.8867994816869316e-05,
"loss": 1.5537,
"step": 320
},
{
"epoch": 0.19784283513097073,
"grad_norm": 0.6264992356300354,
"learning_rate": 3.885366823791883e-05,
"loss": 1.5785,
"step": 321
},
{
"epoch": 0.19845916795069338,
"grad_norm": 0.6459900736808777,
"learning_rate": 3.883925424419126e-05,
"loss": 1.6063,
"step": 322
},
{
"epoch": 0.19907550077041603,
"grad_norm": 0.6335688829421997,
"learning_rate": 3.882475290251686e-05,
"loss": 1.5388,
"step": 323
},
{
"epoch": 0.19969183359013867,
"grad_norm": 0.6406273245811462,
"learning_rate": 3.8810164280130944e-05,
"loss": 1.5848,
"step": 324
},
{
"epoch": 0.20030816640986132,
"grad_norm": 0.6211891770362854,
"learning_rate": 3.879548844467343e-05,
"loss": 1.4908,
"step": 325
},
{
"epoch": 0.20092449922958397,
"grad_norm": 0.666120707988739,
"learning_rate": 3.878072546418865e-05,
"loss": 1.5351,
"step": 326
},
{
"epoch": 0.20154083204930662,
"grad_norm": 0.6317727565765381,
"learning_rate": 3.8765875407124956e-05,
"loss": 1.5363,
"step": 327
},
{
"epoch": 0.20215716486902927,
"grad_norm": 0.6246560215950012,
"learning_rate": 3.8750938342334436e-05,
"loss": 1.5637,
"step": 328
},
{
"epoch": 0.2027734976887519,
"grad_norm": 0.6172101497650146,
"learning_rate": 3.873591433907259e-05,
"loss": 1.5253,
"step": 329
},
{
"epoch": 0.2033898305084746,
"grad_norm": 0.6300083994865417,
"learning_rate": 3.8720803466998e-05,
"loss": 1.5946,
"step": 330
},
{
"epoch": 0.20400616332819724,
"grad_norm": 0.6064188480377197,
"learning_rate": 3.8705605796172025e-05,
"loss": 1.5502,
"step": 331
},
{
"epoch": 0.20462249614791989,
"grad_norm": 0.6302444338798523,
"learning_rate": 3.869032139705846e-05,
"loss": 1.5739,
"step": 332
},
{
"epoch": 0.20523882896764253,
"grad_norm": 0.6242470741271973,
"learning_rate": 3.8674950340523204e-05,
"loss": 1.556,
"step": 333
},
{
"epoch": 0.20585516178736518,
"grad_norm": 0.6166197061538696,
"learning_rate": 3.8659492697833965e-05,
"loss": 1.5887,
"step": 334
},
{
"epoch": 0.20647149460708783,
"grad_norm": 0.6221072673797607,
"learning_rate": 3.864394854065989e-05,
"loss": 1.5845,
"step": 335
},
{
"epoch": 0.20708782742681048,
"grad_norm": 0.6159368753433228,
"learning_rate": 3.8628317941071254e-05,
"loss": 1.5656,
"step": 336
},
{
"epoch": 0.20770416024653313,
"grad_norm": 0.6356878280639648,
"learning_rate": 3.861260097153912e-05,
"loss": 1.5358,
"step": 337
},
{
"epoch": 0.20832049306625577,
"grad_norm": 0.6479353904724121,
"learning_rate": 3.859679770493499e-05,
"loss": 1.5515,
"step": 338
},
{
"epoch": 0.20893682588597842,
"grad_norm": 0.6261966824531555,
"learning_rate": 3.858090821453052e-05,
"loss": 1.5361,
"step": 339
},
{
"epoch": 0.20955315870570107,
"grad_norm": 0.6239880323410034,
"learning_rate": 3.8564932573997096e-05,
"loss": 1.6195,
"step": 340
},
{
"epoch": 0.21016949152542372,
"grad_norm": 0.6246241927146912,
"learning_rate": 3.854887085740557e-05,
"loss": 1.5846,
"step": 341
},
{
"epoch": 0.21078582434514637,
"grad_norm": 0.6406985521316528,
"learning_rate": 3.853272313922587e-05,
"loss": 1.5648,
"step": 342
},
{
"epoch": 0.21140215716486904,
"grad_norm": 0.6283363699913025,
"learning_rate": 3.851648949432667e-05,
"loss": 1.5856,
"step": 343
},
{
"epoch": 0.2120184899845917,
"grad_norm": 0.6151086091995239,
"learning_rate": 3.8500169997975055e-05,
"loss": 1.5237,
"step": 344
},
{
"epoch": 0.21263482280431434,
"grad_norm": 0.6174421310424805,
"learning_rate": 3.848376472583613e-05,
"loss": 1.5409,
"step": 345
},
{
"epoch": 0.213251155624037,
"grad_norm": 0.6334297060966492,
"learning_rate": 3.8467273753972734e-05,
"loss": 1.5677,
"step": 346
},
{
"epoch": 0.21386748844375963,
"grad_norm": 0.6260072588920593,
"learning_rate": 3.845069715884502e-05,
"loss": 1.5605,
"step": 347
},
{
"epoch": 0.21448382126348228,
"grad_norm": 0.6221638321876526,
"learning_rate": 3.843403501731016e-05,
"loss": 1.5543,
"step": 348
},
{
"epoch": 0.21510015408320493,
"grad_norm": 0.6135300993919373,
"learning_rate": 3.841728740662194e-05,
"loss": 1.499,
"step": 349
},
{
"epoch": 0.21571648690292758,
"grad_norm": 0.6266738772392273,
"learning_rate": 3.840045440443043e-05,
"loss": 1.5937,
"step": 350
},
{
"epoch": 0.21633281972265023,
"grad_norm": 0.6476675271987915,
"learning_rate": 3.8383536088781624e-05,
"loss": 1.5651,
"step": 351
},
{
"epoch": 0.21694915254237288,
"grad_norm": 0.6242912411689758,
"learning_rate": 3.836653253811704e-05,
"loss": 1.6536,
"step": 352
},
{
"epoch": 0.21756548536209552,
"grad_norm": 0.6051536202430725,
"learning_rate": 3.834944383127344e-05,
"loss": 1.4944,
"step": 353
},
{
"epoch": 0.21818181818181817,
"grad_norm": 0.6046092510223389,
"learning_rate": 3.833227004748237e-05,
"loss": 1.5516,
"step": 354
},
{
"epoch": 0.21879815100154082,
"grad_norm": 0.6210584044456482,
"learning_rate": 3.831501126636983e-05,
"loss": 1.6006,
"step": 355
},
{
"epoch": 0.2194144838212635,
"grad_norm": 0.6627335548400879,
"learning_rate": 3.829766756795595e-05,
"loss": 1.6129,
"step": 356
},
{
"epoch": 0.22003081664098614,
"grad_norm": 0.6231202483177185,
"learning_rate": 3.8280239032654536e-05,
"loss": 1.5539,
"step": 357
},
{
"epoch": 0.2206471494607088,
"grad_norm": 0.6369742751121521,
"learning_rate": 3.826272574127276e-05,
"loss": 1.5917,
"step": 358
},
{
"epoch": 0.22126348228043144,
"grad_norm": 0.6062166094779968,
"learning_rate": 3.8245127775010764e-05,
"loss": 1.5284,
"step": 359
},
{
"epoch": 0.2218798151001541,
"grad_norm": 0.6495872139930725,
"learning_rate": 3.822744521546127e-05,
"loss": 1.5648,
"step": 360
},
{
"epoch": 0.22249614791987674,
"grad_norm": 0.6449238061904907,
"learning_rate": 3.820967814460923e-05,
"loss": 1.5595,
"step": 361
},
{
"epoch": 0.22311248073959938,
"grad_norm": 0.6381934881210327,
"learning_rate": 3.819182664483144e-05,
"loss": 1.5733,
"step": 362
},
{
"epoch": 0.22372881355932203,
"grad_norm": 0.6340575218200684,
"learning_rate": 3.817389079889612e-05,
"loss": 1.5488,
"step": 363
},
{
"epoch": 0.22434514637904468,
"grad_norm": 0.627213716506958,
"learning_rate": 3.815587068996258e-05,
"loss": 1.5744,
"step": 364
},
{
"epoch": 0.22496147919876733,
"grad_norm": 0.6066584587097168,
"learning_rate": 3.8137766401580814e-05,
"loss": 1.5482,
"step": 365
},
{
"epoch": 0.22557781201848998,
"grad_norm": 0.6272307634353638,
"learning_rate": 3.811957801769111e-05,
"loss": 1.5628,
"step": 366
},
{
"epoch": 0.22619414483821262,
"grad_norm": 0.6388189196586609,
"learning_rate": 3.810130562262366e-05,
"loss": 1.5816,
"step": 367
},
{
"epoch": 0.22681047765793527,
"grad_norm": 0.613997220993042,
"learning_rate": 3.808294930109818e-05,
"loss": 1.5346,
"step": 368
},
{
"epoch": 0.22742681047765795,
"grad_norm": 0.623281717300415,
"learning_rate": 3.80645091382235e-05,
"loss": 1.5748,
"step": 369
},
{
"epoch": 0.2280431432973806,
"grad_norm": 0.6061915755271912,
"learning_rate": 3.804598521949718e-05,
"loss": 1.539,
"step": 370
},
{
"epoch": 0.22865947611710324,
"grad_norm": 0.6243897080421448,
"learning_rate": 3.802737763080513e-05,
"loss": 1.5695,
"step": 371
},
{
"epoch": 0.2292758089368259,
"grad_norm": 0.6500725150108337,
"learning_rate": 3.800868645842118e-05,
"loss": 1.6556,
"step": 372
},
{
"epoch": 0.22989214175654854,
"grad_norm": 0.6233870983123779,
"learning_rate": 3.798991178900671e-05,
"loss": 1.5392,
"step": 373
},
{
"epoch": 0.2305084745762712,
"grad_norm": 0.6300720572471619,
"learning_rate": 3.797105370961019e-05,
"loss": 1.5576,
"step": 374
},
{
"epoch": 0.23112480739599384,
"grad_norm": 0.6239476203918457,
"learning_rate": 3.795211230766687e-05,
"loss": 1.5261,
"step": 375
},
{
"epoch": 0.23174114021571648,
"grad_norm": 0.6265757083892822,
"learning_rate": 3.793308767099829e-05,
"loss": 1.5849,
"step": 376
},
{
"epoch": 0.23235747303543913,
"grad_norm": 0.6283276677131653,
"learning_rate": 3.791397988781194e-05,
"loss": 1.5687,
"step": 377
},
{
"epoch": 0.23297380585516178,
"grad_norm": 0.624076247215271,
"learning_rate": 3.789478904670078e-05,
"loss": 1.5565,
"step": 378
},
{
"epoch": 0.23359013867488443,
"grad_norm": 0.6197231411933899,
"learning_rate": 3.7875515236642894e-05,
"loss": 1.5715,
"step": 379
},
{
"epoch": 0.23420647149460708,
"grad_norm": 0.6188759207725525,
"learning_rate": 3.785615854700104e-05,
"loss": 1.5728,
"step": 380
},
{
"epoch": 0.23482280431432973,
"grad_norm": 0.6138371229171753,
"learning_rate": 3.7836719067522236e-05,
"loss": 1.5571,
"step": 381
},
{
"epoch": 0.2354391371340524,
"grad_norm": 0.6327232718467712,
"learning_rate": 3.781719688833739e-05,
"loss": 1.5941,
"step": 382
},
{
"epoch": 0.23605546995377505,
"grad_norm": 0.6079659461975098,
"learning_rate": 3.779759209996079e-05,
"loss": 1.5373,
"step": 383
},
{
"epoch": 0.2366718027734977,
"grad_norm": 0.6112452745437622,
"learning_rate": 3.777790479328979e-05,
"loss": 1.5547,
"step": 384
},
{
"epoch": 0.23728813559322035,
"grad_norm": 0.6212001442909241,
"learning_rate": 3.7758135059604315e-05,
"loss": 1.537,
"step": 385
},
{
"epoch": 0.237904468412943,
"grad_norm": 0.6198443174362183,
"learning_rate": 3.773828299056647e-05,
"loss": 1.532,
"step": 386
},
{
"epoch": 0.23852080123266564,
"grad_norm": 0.6353772282600403,
"learning_rate": 3.77183486782201e-05,
"loss": 1.5347,
"step": 387
},
{
"epoch": 0.2391371340523883,
"grad_norm": 0.6267209053039551,
"learning_rate": 3.7698332214990385e-05,
"loss": 1.527,
"step": 388
},
{
"epoch": 0.23975346687211094,
"grad_norm": 0.634272575378418,
"learning_rate": 3.767823369368336e-05,
"loss": 1.5845,
"step": 389
},
{
"epoch": 0.24036979969183359,
"grad_norm": 0.6304289102554321,
"learning_rate": 3.765805320748558e-05,
"loss": 1.5901,
"step": 390
},
{
"epoch": 0.24098613251155623,
"grad_norm": 0.6337018013000488,
"learning_rate": 3.763779084996357e-05,
"loss": 1.5677,
"step": 391
},
{
"epoch": 0.24160246533127888,
"grad_norm": 0.6266562342643738,
"learning_rate": 3.761744671506349e-05,
"loss": 1.6029,
"step": 392
},
{
"epoch": 0.24221879815100153,
"grad_norm": 0.6246030926704407,
"learning_rate": 3.7597020897110634e-05,
"loss": 1.586,
"step": 393
},
{
"epoch": 0.24283513097072418,
"grad_norm": 0.6081348657608032,
"learning_rate": 3.7576513490809044e-05,
"loss": 1.5734,
"step": 394
},
{
"epoch": 0.24345146379044685,
"grad_norm": 0.6142352819442749,
"learning_rate": 3.7555924591241015e-05,
"loss": 1.5099,
"step": 395
},
{
"epoch": 0.2440677966101695,
"grad_norm": 0.6177673935890198,
"learning_rate": 3.753525429386671e-05,
"loss": 1.5074,
"step": 396
},
{
"epoch": 0.24468412942989215,
"grad_norm": 0.6194700002670288,
"learning_rate": 3.751450269452368e-05,
"loss": 1.5212,
"step": 397
},
{
"epoch": 0.2453004622496148,
"grad_norm": 0.6167085766792297,
"learning_rate": 3.749366988942645e-05,
"loss": 1.6025,
"step": 398
},
{
"epoch": 0.24591679506933745,
"grad_norm": 0.6241602897644043,
"learning_rate": 3.747275597516601e-05,
"loss": 1.5204,
"step": 399
},
{
"epoch": 0.2465331278890601,
"grad_norm": 0.6154288053512573,
"learning_rate": 3.745176104870945e-05,
"loss": 1.5505,
"step": 400
},
{
"epoch": 0.24714946070878274,
"grad_norm": 0.6391974091529846,
"learning_rate": 3.743068520739947e-05,
"loss": 1.5339,
"step": 401
},
{
"epoch": 0.2477657935285054,
"grad_norm": 0.6121848225593567,
"learning_rate": 3.740952854895392e-05,
"loss": 1.4906,
"step": 402
},
{
"epoch": 0.24838212634822804,
"grad_norm": 0.6097393035888672,
"learning_rate": 3.738829117146535e-05,
"loss": 1.4699,
"step": 403
},
{
"epoch": 0.2489984591679507,
"grad_norm": 0.6633890271186829,
"learning_rate": 3.736697317340059e-05,
"loss": 1.5798,
"step": 404
},
{
"epoch": 0.24961479198767333,
"grad_norm": 0.6471389532089233,
"learning_rate": 3.7345574653600234e-05,
"loss": 1.4527,
"step": 405
},
{
"epoch": 0.250231124807396,
"grad_norm": 0.7222802639007568,
"learning_rate": 3.7324095711278235e-05,
"loss": 1.5641,
"step": 406
},
{
"epoch": 0.25084745762711863,
"grad_norm": 0.6318141222000122,
"learning_rate": 3.730253644602142e-05,
"loss": 1.5671,
"step": 407
},
{
"epoch": 0.2514637904468413,
"grad_norm": 0.6795703172683716,
"learning_rate": 3.7280896957789024e-05,
"loss": 1.5776,
"step": 408
},
{
"epoch": 0.2520801232665639,
"grad_norm": 0.6108106374740601,
"learning_rate": 3.725917734691224e-05,
"loss": 1.5307,
"step": 409
},
{
"epoch": 0.2526964560862866,
"grad_norm": 0.6504859328269958,
"learning_rate": 3.7237377714093756e-05,
"loss": 1.5231,
"step": 410
},
{
"epoch": 0.2533127889060092,
"grad_norm": 0.6572467684745789,
"learning_rate": 3.7215498160407275e-05,
"loss": 1.5766,
"step": 411
},
{
"epoch": 0.2539291217257319,
"grad_norm": 0.6699705123901367,
"learning_rate": 3.719353878729704e-05,
"loss": 1.6119,
"step": 412
},
{
"epoch": 0.2545454545454545,
"grad_norm": 0.6572472453117371,
"learning_rate": 3.717149969657739e-05,
"loss": 1.5487,
"step": 413
},
{
"epoch": 0.2551617873651772,
"grad_norm": 0.6471747756004333,
"learning_rate": 3.7149380990432284e-05,
"loss": 1.5511,
"step": 414
},
{
"epoch": 0.25577812018489987,
"grad_norm": 0.6392568945884705,
"learning_rate": 3.712718277141479e-05,
"loss": 1.5491,
"step": 415
},
{
"epoch": 0.2563944530046225,
"grad_norm": 0.6324256658554077,
"learning_rate": 3.710490514244664e-05,
"loss": 1.5234,
"step": 416
},
{
"epoch": 0.25701078582434517,
"grad_norm": 0.6545376181602478,
"learning_rate": 3.708254820681778e-05,
"loss": 1.5703,
"step": 417
},
{
"epoch": 0.2576271186440678,
"grad_norm": 0.6385619044303894,
"learning_rate": 3.7060112068185834e-05,
"loss": 1.5831,
"step": 418
},
{
"epoch": 0.25824345146379046,
"grad_norm": 0.6294272541999817,
"learning_rate": 3.703759683057564e-05,
"loss": 1.5387,
"step": 419
},
{
"epoch": 0.2588597842835131,
"grad_norm": 0.6257953643798828,
"learning_rate": 3.7015002598378805e-05,
"loss": 1.4815,
"step": 420
},
{
"epoch": 0.25947611710323576,
"grad_norm": 0.6191906332969666,
"learning_rate": 3.6992329476353184e-05,
"loss": 1.5048,
"step": 421
},
{
"epoch": 0.2600924499229584,
"grad_norm": 0.6304193139076233,
"learning_rate": 3.696957756962238e-05,
"loss": 1.4664,
"step": 422
},
{
"epoch": 0.26070878274268106,
"grad_norm": 0.6742473244667053,
"learning_rate": 3.694674698367532e-05,
"loss": 1.581,
"step": 423
},
{
"epoch": 0.2613251155624037,
"grad_norm": 0.6408929824829102,
"learning_rate": 3.6923837824365704e-05,
"loss": 1.579,
"step": 424
},
{
"epoch": 0.26194144838212635,
"grad_norm": 0.6516648530960083,
"learning_rate": 3.6900850197911526e-05,
"loss": 1.5845,
"step": 425
},
{
"epoch": 0.26255778120184897,
"grad_norm": 0.6402381658554077,
"learning_rate": 3.687778421089461e-05,
"loss": 1.6036,
"step": 426
},
{
"epoch": 0.26317411402157165,
"grad_norm": 0.6402097344398499,
"learning_rate": 3.68546399702601e-05,
"loss": 1.5358,
"step": 427
},
{
"epoch": 0.2637904468412943,
"grad_norm": 0.6428090929985046,
"learning_rate": 3.683141758331594e-05,
"loss": 1.5649,
"step": 428
},
{
"epoch": 0.26440677966101694,
"grad_norm": 0.6168109774589539,
"learning_rate": 3.680811715773243e-05,
"loss": 1.4968,
"step": 429
},
{
"epoch": 0.2650231124807396,
"grad_norm": 0.6354753971099854,
"learning_rate": 3.678473880154166e-05,
"loss": 1.5276,
"step": 430
},
{
"epoch": 0.26563944530046224,
"grad_norm": 0.6141413450241089,
"learning_rate": 3.676128262313707e-05,
"loss": 1.5411,
"step": 431
},
{
"epoch": 0.2662557781201849,
"grad_norm": 0.6401364207267761,
"learning_rate": 3.673774873127291e-05,
"loss": 1.6012,
"step": 432
},
{
"epoch": 0.26687211093990754,
"grad_norm": 0.6499303579330444,
"learning_rate": 3.671413723506375e-05,
"loss": 1.5878,
"step": 433
},
{
"epoch": 0.2674884437596302,
"grad_norm": 0.626645565032959,
"learning_rate": 3.669044824398397e-05,
"loss": 1.53,
"step": 434
},
{
"epoch": 0.26810477657935283,
"grad_norm": 0.6259933114051819,
"learning_rate": 3.666668186786726e-05,
"loss": 1.5172,
"step": 435
},
{
"epoch": 0.2687211093990755,
"grad_norm": 0.6535432934761047,
"learning_rate": 3.664283821690611e-05,
"loss": 1.5639,
"step": 436
},
{
"epoch": 0.26933744221879813,
"grad_norm": 0.627822756767273,
"learning_rate": 3.661891740165127e-05,
"loss": 1.6058,
"step": 437
},
{
"epoch": 0.2699537750385208,
"grad_norm": 0.6237924695014954,
"learning_rate": 3.6594919533011274e-05,
"loss": 1.5589,
"step": 438
},
{
"epoch": 0.2705701078582434,
"grad_norm": 0.6268482804298401,
"learning_rate": 3.657084472225193e-05,
"loss": 1.5538,
"step": 439
},
{
"epoch": 0.2711864406779661,
"grad_norm": 0.6410224437713623,
"learning_rate": 3.654669308099575e-05,
"loss": 1.5947,
"step": 440
},
{
"epoch": 0.2718027734976888,
"grad_norm": 0.6285886168479919,
"learning_rate": 3.6522464721221513e-05,
"loss": 1.5114,
"step": 441
},
{
"epoch": 0.2724191063174114,
"grad_norm": 0.6456573605537415,
"learning_rate": 3.6498159755263667e-05,
"loss": 1.5966,
"step": 442
},
{
"epoch": 0.2730354391371341,
"grad_norm": 0.6269593834877014,
"learning_rate": 3.647377829581186e-05,
"loss": 1.6099,
"step": 443
},
{
"epoch": 0.2736517719568567,
"grad_norm": 0.63033127784729,
"learning_rate": 3.64493204559104e-05,
"loss": 1.5018,
"step": 444
},
{
"epoch": 0.27426810477657937,
"grad_norm": 0.6265948414802551,
"learning_rate": 3.6424786348957726e-05,
"loss": 1.5309,
"step": 445
},
{
"epoch": 0.274884437596302,
"grad_norm": 0.6247050166130066,
"learning_rate": 3.6400176088705903e-05,
"loss": 1.4771,
"step": 446
},
{
"epoch": 0.27550077041602467,
"grad_norm": 0.6381626725196838,
"learning_rate": 3.637548978926005e-05,
"loss": 1.5488,
"step": 447
},
{
"epoch": 0.2761171032357473,
"grad_norm": 0.6295813918113708,
"learning_rate": 3.6350727565077876e-05,
"loss": 1.556,
"step": 448
},
{
"epoch": 0.27673343605546996,
"grad_norm": 0.6318914294242859,
"learning_rate": 3.6325889530969095e-05,
"loss": 1.52,
"step": 449
},
{
"epoch": 0.2773497688751926,
"grad_norm": 0.622588574886322,
"learning_rate": 3.6300975802094904e-05,
"loss": 1.5562,
"step": 450
},
{
"epoch": 0.27796610169491526,
"grad_norm": 0.6113806962966919,
"learning_rate": 3.627598649396748e-05,
"loss": 1.5129,
"step": 451
},
{
"epoch": 0.2785824345146379,
"grad_norm": 0.628746747970581,
"learning_rate": 3.6250921722449417e-05,
"loss": 1.5225,
"step": 452
},
{
"epoch": 0.27919876733436055,
"grad_norm": 0.642052948474884,
"learning_rate": 3.6225781603753175e-05,
"loss": 1.5206,
"step": 453
},
{
"epoch": 0.27981510015408323,
"grad_norm": 0.6341302394866943,
"learning_rate": 3.620056625444057e-05,
"loss": 1.5517,
"step": 454
},
{
"epoch": 0.28043143297380585,
"grad_norm": 0.631202757358551,
"learning_rate": 3.617527579142224e-05,
"loss": 1.5652,
"step": 455
},
{
"epoch": 0.2810477657935285,
"grad_norm": 0.6409715414047241,
"learning_rate": 3.6149910331957076e-05,
"loss": 1.5595,
"step": 456
},
{
"epoch": 0.28166409861325115,
"grad_norm": 0.6304307579994202,
"learning_rate": 3.612446999365167e-05,
"loss": 1.5613,
"step": 457
},
{
"epoch": 0.2822804314329738,
"grad_norm": 0.6630215644836426,
"learning_rate": 3.609895489445983e-05,
"loss": 1.5297,
"step": 458
},
{
"epoch": 0.28289676425269644,
"grad_norm": 0.6226741075515747,
"learning_rate": 3.607336515268194e-05,
"loss": 1.4715,
"step": 459
},
{
"epoch": 0.2835130970724191,
"grad_norm": 0.64727783203125,
"learning_rate": 3.60477008869645e-05,
"loss": 1.6325,
"step": 460
},
{
"epoch": 0.28412942989214174,
"grad_norm": 0.6112896203994751,
"learning_rate": 3.6021962216299544e-05,
"loss": 1.5171,
"step": 461
},
{
"epoch": 0.2847457627118644,
"grad_norm": 0.6166220903396606,
"learning_rate": 3.599614926002406e-05,
"loss": 1.5326,
"step": 462
},
{
"epoch": 0.28536209553158703,
"grad_norm": 0.6457735300064087,
"learning_rate": 3.5970262137819456e-05,
"loss": 1.5413,
"step": 463
},
{
"epoch": 0.2859784283513097,
"grad_norm": 0.6203924417495728,
"learning_rate": 3.594430096971104e-05,
"loss": 1.534,
"step": 464
},
{
"epoch": 0.28659476117103233,
"grad_norm": 0.6635544300079346,
"learning_rate": 3.591826587606742e-05,
"loss": 1.5581,
"step": 465
},
{
"epoch": 0.287211093990755,
"grad_norm": 0.6290273666381836,
"learning_rate": 3.589215697759995e-05,
"loss": 1.507,
"step": 466
},
{
"epoch": 0.2878274268104777,
"grad_norm": 0.6142866015434265,
"learning_rate": 3.586597439536216e-05,
"loss": 1.4979,
"step": 467
},
{
"epoch": 0.2884437596302003,
"grad_norm": 0.6286876201629639,
"learning_rate": 3.583971825074927e-05,
"loss": 1.4555,
"step": 468
},
{
"epoch": 0.289060092449923,
"grad_norm": 0.6208572387695312,
"learning_rate": 3.581338866549752e-05,
"loss": 1.5307,
"step": 469
},
{
"epoch": 0.2896764252696456,
"grad_norm": 0.6354894638061523,
"learning_rate": 3.5786985761683685e-05,
"loss": 1.4588,
"step": 470
},
{
"epoch": 0.2902927580893683,
"grad_norm": 0.6662725210189819,
"learning_rate": 3.576050966172447e-05,
"loss": 1.4847,
"step": 471
},
{
"epoch": 0.2909090909090909,
"grad_norm": 0.6339364647865295,
"learning_rate": 3.5733960488375946e-05,
"loss": 1.5006,
"step": 472
},
{
"epoch": 0.29152542372881357,
"grad_norm": 0.6413895487785339,
"learning_rate": 3.5707338364733e-05,
"loss": 1.5562,
"step": 473
},
{
"epoch": 0.2921417565485362,
"grad_norm": 0.6387337446212769,
"learning_rate": 3.568064341422874e-05,
"loss": 1.549,
"step": 474
},
{
"epoch": 0.29275808936825887,
"grad_norm": 0.6315402388572693,
"learning_rate": 3.5653875760633946e-05,
"loss": 1.5277,
"step": 475
},
{
"epoch": 0.2933744221879815,
"grad_norm": 0.649841845035553,
"learning_rate": 3.562703552805649e-05,
"loss": 1.5936,
"step": 476
},
{
"epoch": 0.29399075500770416,
"grad_norm": 0.681604266166687,
"learning_rate": 3.560012284094074e-05,
"loss": 1.5763,
"step": 477
},
{
"epoch": 0.2946070878274268,
"grad_norm": 0.6153651475906372,
"learning_rate": 3.5573137824067e-05,
"loss": 1.5296,
"step": 478
},
{
"epoch": 0.29522342064714946,
"grad_norm": 0.6196587681770325,
"learning_rate": 3.5546080602550935e-05,
"loss": 1.5017,
"step": 479
},
{
"epoch": 0.29583975346687214,
"grad_norm": 0.6492789387702942,
"learning_rate": 3.5518951301843e-05,
"loss": 1.4908,
"step": 480
},
{
"epoch": 0.29645608628659476,
"grad_norm": 0.6555952429771423,
"learning_rate": 3.5491750047727814e-05,
"loss": 1.5554,
"step": 481
},
{
"epoch": 0.29707241910631743,
"grad_norm": 0.6530175805091858,
"learning_rate": 3.5464476966323626e-05,
"loss": 1.5217,
"step": 482
},
{
"epoch": 0.29768875192604005,
"grad_norm": 0.6493626236915588,
"learning_rate": 3.543713218408171e-05,
"loss": 1.5261,
"step": 483
},
{
"epoch": 0.2983050847457627,
"grad_norm": 0.6442253589630127,
"learning_rate": 3.540971582778578e-05,
"loss": 1.4622,
"step": 484
},
{
"epoch": 0.29892141756548535,
"grad_norm": 0.6336924433708191,
"learning_rate": 3.538222802455139e-05,
"loss": 1.5069,
"step": 485
},
{
"epoch": 0.299537750385208,
"grad_norm": 0.6494876146316528,
"learning_rate": 3.535466890182539e-05,
"loss": 1.5268,
"step": 486
},
{
"epoch": 0.30015408320493064,
"grad_norm": 0.6452393531799316,
"learning_rate": 3.532703858738525e-05,
"loss": 1.586,
"step": 487
},
{
"epoch": 0.3007704160246533,
"grad_norm": 0.639706015586853,
"learning_rate": 3.5299337209338566e-05,
"loss": 1.5198,
"step": 488
},
{
"epoch": 0.30138674884437594,
"grad_norm": 0.6566674113273621,
"learning_rate": 3.52715648961224e-05,
"loss": 1.6219,
"step": 489
},
{
"epoch": 0.3020030816640986,
"grad_norm": 0.6216531991958618,
"learning_rate": 3.52437217765027e-05,
"loss": 1.4874,
"step": 490
},
{
"epoch": 0.30261941448382124,
"grad_norm": 0.6326376795768738,
"learning_rate": 3.5215807979573724e-05,
"loss": 1.5384,
"step": 491
},
{
"epoch": 0.3032357473035439,
"grad_norm": 0.6411398649215698,
"learning_rate": 3.5187823634757395e-05,
"loss": 1.5534,
"step": 492
},
{
"epoch": 0.3038520801232666,
"grad_norm": 0.6091376543045044,
"learning_rate": 3.5159768871802766e-05,
"loss": 1.5432,
"step": 493
},
{
"epoch": 0.3044684129429892,
"grad_norm": 0.6145322322845459,
"learning_rate": 3.513164382078535e-05,
"loss": 1.5224,
"step": 494
},
{
"epoch": 0.3050847457627119,
"grad_norm": 0.6333751678466797,
"learning_rate": 3.5103448612106575e-05,
"loss": 1.5535,
"step": 495
},
{
"epoch": 0.3057010785824345,
"grad_norm": 0.6383638978004456,
"learning_rate": 3.507518337649314e-05,
"loss": 1.5962,
"step": 496
},
{
"epoch": 0.3063174114021572,
"grad_norm": 0.6318841576576233,
"learning_rate": 3.504684824499642e-05,
"loss": 1.5213,
"step": 497
},
{
"epoch": 0.3069337442218798,
"grad_norm": 0.6321629881858826,
"learning_rate": 3.501844334899188e-05,
"loss": 1.5732,
"step": 498
},
{
"epoch": 0.3075500770416025,
"grad_norm": 0.6273744702339172,
"learning_rate": 3.498996882017843e-05,
"loss": 1.4679,
"step": 499
},
{
"epoch": 0.3081664098613251,
"grad_norm": 0.6182377934455872,
"learning_rate": 3.4961424790577835e-05,
"loss": 1.4724,
"step": 500
},
{
"epoch": 0.3087827426810478,
"grad_norm": 0.6165096759796143,
"learning_rate": 3.4932811392534104e-05,
"loss": 1.5005,
"step": 501
},
{
"epoch": 0.3093990755007704,
"grad_norm": 0.5994207859039307,
"learning_rate": 3.490412875871287e-05,
"loss": 1.4517,
"step": 502
},
{
"epoch": 0.31001540832049307,
"grad_norm": 0.6823679804801941,
"learning_rate": 3.487537702210077e-05,
"loss": 1.5238,
"step": 503
},
{
"epoch": 0.3106317411402157,
"grad_norm": 0.6435548067092896,
"learning_rate": 3.4846556316004844e-05,
"loss": 1.5634,
"step": 504
},
{
"epoch": 0.31124807395993837,
"grad_norm": 0.6374288201332092,
"learning_rate": 3.481766677405191e-05,
"loss": 1.5566,
"step": 505
},
{
"epoch": 0.31186440677966104,
"grad_norm": 0.6297028064727783,
"learning_rate": 3.478870853018792e-05,
"loss": 1.5282,
"step": 506
},
{
"epoch": 0.31248073959938366,
"grad_norm": 0.6188157200813293,
"learning_rate": 3.475968171867741e-05,
"loss": 1.5,
"step": 507
},
{
"epoch": 0.31309707241910634,
"grad_norm": 0.6542200446128845,
"learning_rate": 3.473058647410277e-05,
"loss": 1.5111,
"step": 508
},
{
"epoch": 0.31371340523882896,
"grad_norm": 0.6281660199165344,
"learning_rate": 3.47014229313637e-05,
"loss": 1.5577,
"step": 509
},
{
"epoch": 0.31432973805855163,
"grad_norm": 0.6386356353759766,
"learning_rate": 3.467219122567661e-05,
"loss": 1.5276,
"step": 510
},
{
"epoch": 0.31494607087827425,
"grad_norm": 0.6302286982536316,
"learning_rate": 3.464289149257387e-05,
"loss": 1.5313,
"step": 511
},
{
"epoch": 0.31556240369799693,
"grad_norm": 0.6208285093307495,
"learning_rate": 3.4613523867903324e-05,
"loss": 1.5475,
"step": 512
},
{
"epoch": 0.31617873651771955,
"grad_norm": 0.6123420000076294,
"learning_rate": 3.458408848782754e-05,
"loss": 1.4761,
"step": 513
},
{
"epoch": 0.3167950693374422,
"grad_norm": 0.648521900177002,
"learning_rate": 3.4554585488823265e-05,
"loss": 1.4817,
"step": 514
},
{
"epoch": 0.31741140215716485,
"grad_norm": 0.67547607421875,
"learning_rate": 3.452501500768076e-05,
"loss": 1.5742,
"step": 515
},
{
"epoch": 0.3180277349768875,
"grad_norm": 0.6527137160301208,
"learning_rate": 3.449537718150316e-05,
"loss": 1.4839,
"step": 516
},
{
"epoch": 0.31864406779661014,
"grad_norm": 0.6243513822555542,
"learning_rate": 3.4465672147705844e-05,
"loss": 1.4877,
"step": 517
},
{
"epoch": 0.3192604006163328,
"grad_norm": 0.6275103092193604,
"learning_rate": 3.4435900044015796e-05,
"loss": 1.5742,
"step": 518
},
{
"epoch": 0.3198767334360555,
"grad_norm": 0.6227298378944397,
"learning_rate": 3.440606100847096e-05,
"loss": 1.5862,
"step": 519
},
{
"epoch": 0.3204930662557781,
"grad_norm": 0.6253839731216431,
"learning_rate": 3.437615517941964e-05,
"loss": 1.5021,
"step": 520
},
{
"epoch": 0.3211093990755008,
"grad_norm": 0.6459096074104309,
"learning_rate": 3.43461826955198e-05,
"loss": 1.5576,
"step": 521
},
{
"epoch": 0.3217257318952234,
"grad_norm": 0.642366349697113,
"learning_rate": 3.431614369573844e-05,
"loss": 1.5324,
"step": 522
},
{
"epoch": 0.3223420647149461,
"grad_norm": 0.6393497586250305,
"learning_rate": 3.4286038319351e-05,
"loss": 1.4959,
"step": 523
},
{
"epoch": 0.3229583975346687,
"grad_norm": 0.6141563057899475,
"learning_rate": 3.4255866705940635e-05,
"loss": 1.5329,
"step": 524
},
{
"epoch": 0.3235747303543914,
"grad_norm": 0.6277351379394531,
"learning_rate": 3.4225628995397615e-05,
"loss": 1.4421,
"step": 525
},
{
"epoch": 0.324191063174114,
"grad_norm": 0.6193028092384338,
"learning_rate": 3.419532532791869e-05,
"loss": 1.4832,
"step": 526
},
{
"epoch": 0.3248073959938367,
"grad_norm": 0.6569095849990845,
"learning_rate": 3.4164955844006396e-05,
"loss": 1.5458,
"step": 527
},
{
"epoch": 0.3254237288135593,
"grad_norm": 0.6416183114051819,
"learning_rate": 3.413452068446843e-05,
"loss": 1.5214,
"step": 528
},
{
"epoch": 0.326040061633282,
"grad_norm": 0.6409138441085815,
"learning_rate": 3.410401999041701e-05,
"loss": 1.5498,
"step": 529
},
{
"epoch": 0.3266563944530046,
"grad_norm": 0.6253319382667542,
"learning_rate": 3.4073453903268166e-05,
"loss": 1.5245,
"step": 530
},
{
"epoch": 0.32727272727272727,
"grad_norm": 0.6222389340400696,
"learning_rate": 3.404282256474117e-05,
"loss": 1.4386,
"step": 531
},
{
"epoch": 0.32788906009244995,
"grad_norm": 0.6330371499061584,
"learning_rate": 3.40121261168578e-05,
"loss": 1.4607,
"step": 532
},
{
"epoch": 0.32850539291217257,
"grad_norm": 0.6958795189857483,
"learning_rate": 3.3981364701941714e-05,
"loss": 1.5438,
"step": 533
},
{
"epoch": 0.32912172573189524,
"grad_norm": 0.6625787019729614,
"learning_rate": 3.39505384626178e-05,
"loss": 1.5514,
"step": 534
},
{
"epoch": 0.32973805855161786,
"grad_norm": 0.6192840337753296,
"learning_rate": 3.39196475418115e-05,
"loss": 1.4313,
"step": 535
},
{
"epoch": 0.33035439137134054,
"grad_norm": 0.6188201308250427,
"learning_rate": 3.388869208274815e-05,
"loss": 1.4991,
"step": 536
},
{
"epoch": 0.33097072419106316,
"grad_norm": 0.6207196712493896,
"learning_rate": 3.385767222895232e-05,
"loss": 1.5068,
"step": 537
},
{
"epoch": 0.33158705701078584,
"grad_norm": 0.6164208650588989,
"learning_rate": 3.382658812424712e-05,
"loss": 1.4934,
"step": 538
},
{
"epoch": 0.33220338983050846,
"grad_norm": 0.6260819435119629,
"learning_rate": 3.379543991275361e-05,
"loss": 1.4864,
"step": 539
},
{
"epoch": 0.33281972265023113,
"grad_norm": 0.663669228553772,
"learning_rate": 3.3764227738890026e-05,
"loss": 1.4923,
"step": 540
},
{
"epoch": 0.33343605546995375,
"grad_norm": 0.6321974992752075,
"learning_rate": 3.37329517473712e-05,
"loss": 1.4754,
"step": 541
},
{
"epoch": 0.3340523882896764,
"grad_norm": 0.6340722441673279,
"learning_rate": 3.370161208320784e-05,
"loss": 1.5135,
"step": 542
},
{
"epoch": 0.3346687211093991,
"grad_norm": 0.6164632439613342,
"learning_rate": 3.367020889170588e-05,
"loss": 1.5456,
"step": 543
},
{
"epoch": 0.3352850539291217,
"grad_norm": 0.6347816586494446,
"learning_rate": 3.363874231846579e-05,
"loss": 1.5307,
"step": 544
},
{
"epoch": 0.3359013867488444,
"grad_norm": 0.6264616847038269,
"learning_rate": 3.36072125093819e-05,
"loss": 1.4914,
"step": 545
},
{
"epoch": 0.336517719568567,
"grad_norm": 0.6221004724502563,
"learning_rate": 3.357561961064175e-05,
"loss": 1.5095,
"step": 546
},
{
"epoch": 0.3371340523882897,
"grad_norm": 0.6659384965896606,
"learning_rate": 3.354396376872538e-05,
"loss": 1.5756,
"step": 547
},
{
"epoch": 0.3377503852080123,
"grad_norm": 0.6458001732826233,
"learning_rate": 3.3512245130404683e-05,
"loss": 1.5201,
"step": 548
},
{
"epoch": 0.338366718027735,
"grad_norm": 0.6084640622138977,
"learning_rate": 3.348046384274269e-05,
"loss": 1.5696,
"step": 549
},
{
"epoch": 0.3389830508474576,
"grad_norm": 0.6264820098876953,
"learning_rate": 3.344862005309291e-05,
"loss": 1.4964,
"step": 550
},
{
"epoch": 0.3395993836671803,
"grad_norm": 0.6191871166229248,
"learning_rate": 3.341671390909864e-05,
"loss": 1.5339,
"step": 551
},
{
"epoch": 0.3402157164869029,
"grad_norm": 0.6216069459915161,
"learning_rate": 3.338474555869228e-05,
"loss": 1.4627,
"step": 552
},
{
"epoch": 0.3408320493066256,
"grad_norm": 0.6235500574111938,
"learning_rate": 3.335271515009467e-05,
"loss": 1.473,
"step": 553
},
{
"epoch": 0.3414483821263482,
"grad_norm": 0.6313657164573669,
"learning_rate": 3.332062283181434e-05,
"loss": 1.4555,
"step": 554
},
{
"epoch": 0.3420647149460709,
"grad_norm": 0.6590183973312378,
"learning_rate": 3.328846875264691e-05,
"loss": 1.5455,
"step": 555
},
{
"epoch": 0.34268104776579356,
"grad_norm": 0.6327791810035706,
"learning_rate": 3.3256253061674326e-05,
"loss": 1.5265,
"step": 556
},
{
"epoch": 0.3432973805855162,
"grad_norm": 0.6291818618774414,
"learning_rate": 3.32239759082642e-05,
"loss": 1.5322,
"step": 557
},
{
"epoch": 0.34391371340523885,
"grad_norm": 0.6328380703926086,
"learning_rate": 3.3191637442069115e-05,
"loss": 1.4937,
"step": 558
},
{
"epoch": 0.3445300462249615,
"grad_norm": 0.6432162523269653,
"learning_rate": 3.315923781302592e-05,
"loss": 1.4924,
"step": 559
},
{
"epoch": 0.34514637904468415,
"grad_norm": 0.6319786310195923,
"learning_rate": 3.3126777171355083e-05,
"loss": 1.4655,
"step": 560
},
{
"epoch": 0.34576271186440677,
"grad_norm": 0.6463144421577454,
"learning_rate": 3.309425566755991e-05,
"loss": 1.5037,
"step": 561
},
{
"epoch": 0.34637904468412944,
"grad_norm": 0.6314020752906799,
"learning_rate": 3.306167345242591e-05,
"loss": 1.5287,
"step": 562
},
{
"epoch": 0.34699537750385206,
"grad_norm": 0.6312958002090454,
"learning_rate": 3.30290306770201e-05,
"loss": 1.5239,
"step": 563
},
{
"epoch": 0.34761171032357474,
"grad_norm": 0.631670355796814,
"learning_rate": 3.299632749269025e-05,
"loss": 1.5446,
"step": 564
},
{
"epoch": 0.34822804314329736,
"grad_norm": 0.6398617029190063,
"learning_rate": 3.296356405106423e-05,
"loss": 1.5112,
"step": 565
},
{
"epoch": 0.34884437596302004,
"grad_norm": 0.62071293592453,
"learning_rate": 3.293074050404929e-05,
"loss": 1.4851,
"step": 566
},
{
"epoch": 0.34946070878274266,
"grad_norm": 0.6425772309303284,
"learning_rate": 3.289785700383138e-05,
"loss": 1.486,
"step": 567
},
{
"epoch": 0.35007704160246533,
"grad_norm": 0.638628363609314,
"learning_rate": 3.2864913702874384e-05,
"loss": 1.5042,
"step": 568
},
{
"epoch": 0.350693374422188,
"grad_norm": 0.6274445056915283,
"learning_rate": 3.2831910753919477e-05,
"loss": 1.5057,
"step": 569
},
{
"epoch": 0.35130970724191063,
"grad_norm": 0.6332955956459045,
"learning_rate": 3.2798848309984385e-05,
"loss": 1.5038,
"step": 570
},
{
"epoch": 0.3519260400616333,
"grad_norm": 0.624282717704773,
"learning_rate": 3.276572652436268e-05,
"loss": 1.5155,
"step": 571
},
{
"epoch": 0.3525423728813559,
"grad_norm": 0.6485286951065063,
"learning_rate": 3.2732545550623066e-05,
"loss": 1.5301,
"step": 572
},
{
"epoch": 0.3531587057010786,
"grad_norm": 0.6295774579048157,
"learning_rate": 3.269930554260869e-05,
"loss": 1.5521,
"step": 573
},
{
"epoch": 0.3537750385208012,
"grad_norm": 0.6386708617210388,
"learning_rate": 3.266600665443638e-05,
"loss": 1.499,
"step": 574
},
{
"epoch": 0.3543913713405239,
"grad_norm": 0.6355401873588562,
"learning_rate": 3.2632649040496005e-05,
"loss": 1.5603,
"step": 575
},
{
"epoch": 0.3550077041602465,
"grad_norm": 0.6240369081497192,
"learning_rate": 3.259923285544966e-05,
"loss": 1.496,
"step": 576
},
{
"epoch": 0.3556240369799692,
"grad_norm": 0.6588268876075745,
"learning_rate": 3.256575825423106e-05,
"loss": 1.5202,
"step": 577
},
{
"epoch": 0.3562403697996918,
"grad_norm": 0.6418221592903137,
"learning_rate": 3.253222539204472e-05,
"loss": 1.4892,
"step": 578
},
{
"epoch": 0.3568567026194145,
"grad_norm": 0.6440759301185608,
"learning_rate": 3.249863442436532e-05,
"loss": 1.5012,
"step": 579
},
{
"epoch": 0.3574730354391371,
"grad_norm": 0.6364291906356812,
"learning_rate": 3.2464985506936914e-05,
"loss": 1.4539,
"step": 580
},
{
"epoch": 0.3580893682588598,
"grad_norm": 0.6213045120239258,
"learning_rate": 3.243127879577225e-05,
"loss": 1.4471,
"step": 581
},
{
"epoch": 0.35870570107858246,
"grad_norm": 0.6480481028556824,
"learning_rate": 3.2397514447152035e-05,
"loss": 1.4998,
"step": 582
},
{
"epoch": 0.3593220338983051,
"grad_norm": 0.6508931517601013,
"learning_rate": 3.2363692617624226e-05,
"loss": 1.5233,
"step": 583
},
{
"epoch": 0.35993836671802776,
"grad_norm": 0.6378174424171448,
"learning_rate": 3.232981346400327e-05,
"loss": 1.5147,
"step": 584
},
{
"epoch": 0.3605546995377504,
"grad_norm": 0.6446996927261353,
"learning_rate": 3.2295877143369394e-05,
"loss": 1.5091,
"step": 585
},
{
"epoch": 0.36117103235747305,
"grad_norm": 0.6296038031578064,
"learning_rate": 3.2261883813067906e-05,
"loss": 1.573,
"step": 586
},
{
"epoch": 0.3617873651771957,
"grad_norm": 0.6350198984146118,
"learning_rate": 3.222783363070841e-05,
"loss": 1.4667,
"step": 587
},
{
"epoch": 0.36240369799691835,
"grad_norm": 0.6288630962371826,
"learning_rate": 3.2193726754164114e-05,
"loss": 1.4666,
"step": 588
},
{
"epoch": 0.36302003081664097,
"grad_norm": 0.6324213743209839,
"learning_rate": 3.21595633415711e-05,
"loss": 1.5423,
"step": 589
},
{
"epoch": 0.36363636363636365,
"grad_norm": 0.6186695694923401,
"learning_rate": 3.2125343551327545e-05,
"loss": 1.4596,
"step": 590
},
{
"epoch": 0.36425269645608627,
"grad_norm": 0.6489487290382385,
"learning_rate": 3.209106754209306e-05,
"loss": 1.5015,
"step": 591
},
{
"epoch": 0.36486902927580894,
"grad_norm": 0.6401476263999939,
"learning_rate": 3.2056735472787896e-05,
"loss": 1.5509,
"step": 592
},
{
"epoch": 0.36548536209553156,
"grad_norm": 0.6308756470680237,
"learning_rate": 3.202234750259222e-05,
"loss": 1.4969,
"step": 593
},
{
"epoch": 0.36610169491525424,
"grad_norm": 0.62753826379776,
"learning_rate": 3.19879037909454e-05,
"loss": 1.4621,
"step": 594
},
{
"epoch": 0.3667180277349769,
"grad_norm": 0.6245593428611755,
"learning_rate": 3.1953404497545224e-05,
"loss": 1.4296,
"step": 595
},
{
"epoch": 0.36733436055469953,
"grad_norm": 0.6454522609710693,
"learning_rate": 3.191884978234721e-05,
"loss": 1.4412,
"step": 596
},
{
"epoch": 0.3679506933744222,
"grad_norm": 0.6471550464630127,
"learning_rate": 3.188423980556382e-05,
"loss": 1.5135,
"step": 597
},
{
"epoch": 0.36856702619414483,
"grad_norm": 0.6443375945091248,
"learning_rate": 3.1849574727663734e-05,
"loss": 1.5463,
"step": 598
},
{
"epoch": 0.3691833590138675,
"grad_norm": 0.6420419216156006,
"learning_rate": 3.181485470937114e-05,
"loss": 1.5464,
"step": 599
},
{
"epoch": 0.3697996918335901,
"grad_norm": 0.6404433250427246,
"learning_rate": 3.17800799116649e-05,
"loss": 1.5228,
"step": 600
},
{
"epoch": 0.3704160246533128,
"grad_norm": 0.6368747353553772,
"learning_rate": 3.174525049577793e-05,
"loss": 1.5595,
"step": 601
},
{
"epoch": 0.3710323574730354,
"grad_norm": 0.6601626873016357,
"learning_rate": 3.171036662319631e-05,
"loss": 1.4279,
"step": 602
},
{
"epoch": 0.3716486902927581,
"grad_norm": 0.6518347859382629,
"learning_rate": 3.167542845565866e-05,
"loss": 1.5115,
"step": 603
},
{
"epoch": 0.3722650231124807,
"grad_norm": 0.6589585542678833,
"learning_rate": 3.164043615515531e-05,
"loss": 1.4223,
"step": 604
},
{
"epoch": 0.3728813559322034,
"grad_norm": 0.6309995055198669,
"learning_rate": 3.160538988392759e-05,
"loss": 1.4763,
"step": 605
},
{
"epoch": 0.373497688751926,
"grad_norm": 0.6362992525100708,
"learning_rate": 3.157028980446707e-05,
"loss": 1.4877,
"step": 606
},
{
"epoch": 0.3741140215716487,
"grad_norm": 0.6353477239608765,
"learning_rate": 3.1535136079514785e-05,
"loss": 1.4983,
"step": 607
},
{
"epoch": 0.37473035439137137,
"grad_norm": 0.6715290546417236,
"learning_rate": 3.1499928872060504e-05,
"loss": 1.5051,
"step": 608
},
{
"epoch": 0.375346687211094,
"grad_norm": 0.6573690176010132,
"learning_rate": 3.146466834534197e-05,
"loss": 1.5581,
"step": 609
},
{
"epoch": 0.37596302003081666,
"grad_norm": 0.6558327674865723,
"learning_rate": 3.142935466284413e-05,
"loss": 1.4989,
"step": 610
},
{
"epoch": 0.3765793528505393,
"grad_norm": 0.6386260390281677,
"learning_rate": 3.13939879882984e-05,
"loss": 1.491,
"step": 611
},
{
"epoch": 0.37719568567026196,
"grad_norm": 0.6379230618476868,
"learning_rate": 3.135856848568188e-05,
"loss": 1.5419,
"step": 612
},
{
"epoch": 0.3778120184899846,
"grad_norm": 0.6291123032569885,
"learning_rate": 3.132309631921662e-05,
"loss": 1.4688,
"step": 613
},
{
"epoch": 0.37842835130970726,
"grad_norm": 0.6273422241210938,
"learning_rate": 3.128757165336885e-05,
"loss": 1.4147,
"step": 614
},
{
"epoch": 0.3790446841294299,
"grad_norm": 0.6356596946716309,
"learning_rate": 3.125199465284817e-05,
"loss": 1.5015,
"step": 615
},
{
"epoch": 0.37966101694915255,
"grad_norm": 0.6427242159843445,
"learning_rate": 3.1216365482606894e-05,
"loss": 1.5036,
"step": 616
},
{
"epoch": 0.3802773497688752,
"grad_norm": 0.6553143858909607,
"learning_rate": 3.118068430783917e-05,
"loss": 1.5595,
"step": 617
},
{
"epoch": 0.38089368258859785,
"grad_norm": 0.6268466114997864,
"learning_rate": 3.114495129398028e-05,
"loss": 1.4878,
"step": 618
},
{
"epoch": 0.38151001540832047,
"grad_norm": 0.6335549354553223,
"learning_rate": 3.1109166606705865e-05,
"loss": 1.537,
"step": 619
},
{
"epoch": 0.38212634822804314,
"grad_norm": 0.6393272876739502,
"learning_rate": 3.107333041193114e-05,
"loss": 1.5054,
"step": 620
},
{
"epoch": 0.3827426810477658,
"grad_norm": 0.6280071139335632,
"learning_rate": 3.103744287581012e-05,
"loss": 1.4897,
"step": 621
},
{
"epoch": 0.38335901386748844,
"grad_norm": 0.6517126560211182,
"learning_rate": 3.100150416473491e-05,
"loss": 1.4843,
"step": 622
},
{
"epoch": 0.3839753466872111,
"grad_norm": 0.6666373610496521,
"learning_rate": 3.096551444533483e-05,
"loss": 1.5471,
"step": 623
},
{
"epoch": 0.38459167950693374,
"grad_norm": 0.6469865441322327,
"learning_rate": 3.092947388447574e-05,
"loss": 1.549,
"step": 624
},
{
"epoch": 0.3852080123266564,
"grad_norm": 0.6489653587341309,
"learning_rate": 3.089338264925921e-05,
"loss": 1.5347,
"step": 625
},
{
"epoch": 0.38582434514637903,
"grad_norm": 0.6462615728378296,
"learning_rate": 3.085724090702177e-05,
"loss": 1.5098,
"step": 626
},
{
"epoch": 0.3864406779661017,
"grad_norm": 0.648750901222229,
"learning_rate": 3.08210488253341e-05,
"loss": 1.545,
"step": 627
},
{
"epoch": 0.38705701078582433,
"grad_norm": 0.6393259167671204,
"learning_rate": 3.07848065720003e-05,
"loss": 1.4958,
"step": 628
},
{
"epoch": 0.387673343605547,
"grad_norm": 0.6603264808654785,
"learning_rate": 3.0748514315057087e-05,
"loss": 1.4882,
"step": 629
},
{
"epoch": 0.3882896764252696,
"grad_norm": 0.6588591933250427,
"learning_rate": 3.071217222277303e-05,
"loss": 1.4983,
"step": 630
},
{
"epoch": 0.3889060092449923,
"grad_norm": 0.6373116970062256,
"learning_rate": 3.0675780463647737e-05,
"loss": 1.513,
"step": 631
},
{
"epoch": 0.3895223420647149,
"grad_norm": 0.6421071290969849,
"learning_rate": 3.0639339206411104e-05,
"loss": 1.539,
"step": 632
},
{
"epoch": 0.3901386748844376,
"grad_norm": 0.6400432586669922,
"learning_rate": 3.060284862002253e-05,
"loss": 1.4893,
"step": 633
},
{
"epoch": 0.3907550077041603,
"grad_norm": 0.6378886103630066,
"learning_rate": 3.056630887367013e-05,
"loss": 1.4926,
"step": 634
},
{
"epoch": 0.3913713405238829,
"grad_norm": 0.6466785669326782,
"learning_rate": 3.0529720136769924e-05,
"loss": 1.5619,
"step": 635
},
{
"epoch": 0.39198767334360557,
"grad_norm": 0.6507505774497986,
"learning_rate": 3.0493082578965114e-05,
"loss": 1.4699,
"step": 636
},
{
"epoch": 0.3926040061633282,
"grad_norm": 0.6481563448905945,
"learning_rate": 3.0456396370125233e-05,
"loss": 1.5298,
"step": 637
},
{
"epoch": 0.39322033898305087,
"grad_norm": 0.6414892077445984,
"learning_rate": 3.041966168034539e-05,
"loss": 1.4908,
"step": 638
},
{
"epoch": 0.3938366718027735,
"grad_norm": 0.6348177790641785,
"learning_rate": 3.038287867994547e-05,
"loss": 1.5126,
"step": 639
},
{
"epoch": 0.39445300462249616,
"grad_norm": 0.6342907547950745,
"learning_rate": 3.034604753946936e-05,
"loss": 1.4447,
"step": 640
},
{
"epoch": 0.3950693374422188,
"grad_norm": 0.6346410512924194,
"learning_rate": 3.0309168429684148e-05,
"loss": 1.4592,
"step": 641
},
{
"epoch": 0.39568567026194146,
"grad_norm": 0.642825186252594,
"learning_rate": 3.027224152157932e-05,
"loss": 1.5082,
"step": 642
},
{
"epoch": 0.3963020030816641,
"grad_norm": 0.6557008028030396,
"learning_rate": 3.0235266986366e-05,
"loss": 1.4764,
"step": 643
},
{
"epoch": 0.39691833590138675,
"grad_norm": 0.6577332019805908,
"learning_rate": 3.0198244995476087e-05,
"loss": 1.5069,
"step": 644
},
{
"epoch": 0.3975346687211094,
"grad_norm": 0.6441546678543091,
"learning_rate": 3.0161175720561574e-05,
"loss": 1.5226,
"step": 645
},
{
"epoch": 0.39815100154083205,
"grad_norm": 0.6265890598297119,
"learning_rate": 3.012405933349363e-05,
"loss": 1.454,
"step": 646
},
{
"epoch": 0.3987673343605547,
"grad_norm": 0.617996871471405,
"learning_rate": 3.0086896006361886e-05,
"loss": 1.3763,
"step": 647
},
{
"epoch": 0.39938366718027735,
"grad_norm": 0.7179849743843079,
"learning_rate": 3.0049685911473608e-05,
"loss": 1.5737,
"step": 648
},
{
"epoch": 0.4,
"grad_norm": 0.6592563390731812,
"learning_rate": 3.0012429221352898e-05,
"loss": 1.489,
"step": 649
},
{
"epoch": 0.40061633281972264,
"grad_norm": 0.6569672226905823,
"learning_rate": 2.9975126108739875e-05,
"loss": 1.5116,
"step": 650
},
{
"epoch": 0.4012326656394453,
"grad_norm": 0.6473748683929443,
"learning_rate": 2.9937776746589945e-05,
"loss": 1.5001,
"step": 651
},
{
"epoch": 0.40184899845916794,
"grad_norm": 0.6282142400741577,
"learning_rate": 2.9900381308072894e-05,
"loss": 1.4954,
"step": 652
},
{
"epoch": 0.4024653312788906,
"grad_norm": 0.6429985761642456,
"learning_rate": 2.986293996657218e-05,
"loss": 1.4735,
"step": 653
},
{
"epoch": 0.40308166409861323,
"grad_norm": 0.6365038156509399,
"learning_rate": 2.9825452895684073e-05,
"loss": 1.4961,
"step": 654
},
{
"epoch": 0.4036979969183359,
"grad_norm": 0.6219319105148315,
"learning_rate": 2.978792026921686e-05,
"loss": 1.4487,
"step": 655
},
{
"epoch": 0.40431432973805853,
"grad_norm": 0.6654730439186096,
"learning_rate": 2.9750342261190063e-05,
"loss": 1.5337,
"step": 656
},
{
"epoch": 0.4049306625577812,
"grad_norm": 0.6424438953399658,
"learning_rate": 2.9712719045833603e-05,
"loss": 1.4577,
"step": 657
},
{
"epoch": 0.4055469953775038,
"grad_norm": 0.645980715751648,
"learning_rate": 2.967505079758701e-05,
"loss": 1.5231,
"step": 658
},
{
"epoch": 0.4061633281972265,
"grad_norm": 0.6435828804969788,
"learning_rate": 2.963733769109861e-05,
"loss": 1.4566,
"step": 659
},
{
"epoch": 0.4067796610169492,
"grad_norm": 0.6797378659248352,
"learning_rate": 2.95995799012247e-05,
"loss": 1.4502,
"step": 660
},
{
"epoch": 0.4073959938366718,
"grad_norm": 0.6313846707344055,
"learning_rate": 2.956177760302876e-05,
"loss": 1.4217,
"step": 661
},
{
"epoch": 0.4080123266563945,
"grad_norm": 0.6844395995140076,
"learning_rate": 2.952393097178065e-05,
"loss": 1.4947,
"step": 662
},
{
"epoch": 0.4086286594761171,
"grad_norm": 0.6478317379951477,
"learning_rate": 2.9486040182955737e-05,
"loss": 1.5423,
"step": 663
},
{
"epoch": 0.40924499229583977,
"grad_norm": 0.6268302202224731,
"learning_rate": 2.9448105412234178e-05,
"loss": 1.4477,
"step": 664
},
{
"epoch": 0.4098613251155624,
"grad_norm": 0.6396114826202393,
"learning_rate": 2.94101268355e-05,
"loss": 1.435,
"step": 665
},
{
"epoch": 0.41047765793528507,
"grad_norm": 0.6264893412590027,
"learning_rate": 2.937210462884037e-05,
"loss": 1.4967,
"step": 666
},
{
"epoch": 0.4110939907550077,
"grad_norm": 0.651495635509491,
"learning_rate": 2.9334038968544738e-05,
"loss": 1.5682,
"step": 667
},
{
"epoch": 0.41171032357473036,
"grad_norm": 0.6641359925270081,
"learning_rate": 2.9295930031104006e-05,
"loss": 1.551,
"step": 668
},
{
"epoch": 0.412326656394453,
"grad_norm": 0.6501958966255188,
"learning_rate": 2.9257777993209756e-05,
"loss": 1.5307,
"step": 669
},
{
"epoch": 0.41294298921417566,
"grad_norm": 0.651185154914856,
"learning_rate": 2.9219583031753392e-05,
"loss": 1.4893,
"step": 670
},
{
"epoch": 0.4135593220338983,
"grad_norm": 0.6622418761253357,
"learning_rate": 2.9181345323825342e-05,
"loss": 1.55,
"step": 671
},
{
"epoch": 0.41417565485362096,
"grad_norm": 0.665107786655426,
"learning_rate": 2.9143065046714202e-05,
"loss": 1.4776,
"step": 672
},
{
"epoch": 0.41479198767334363,
"grad_norm": 0.6447892785072327,
"learning_rate": 2.910474237790598e-05,
"loss": 1.5724,
"step": 673
},
{
"epoch": 0.41540832049306625,
"grad_norm": 0.6534367203712463,
"learning_rate": 2.9066377495083192e-05,
"loss": 1.4929,
"step": 674
},
{
"epoch": 0.41602465331278893,
"grad_norm": 0.6608549356460571,
"learning_rate": 2.9027970576124108e-05,
"loss": 1.5393,
"step": 675
},
{
"epoch": 0.41664098613251155,
"grad_norm": 0.6643514037132263,
"learning_rate": 2.898952179910189e-05,
"loss": 1.464,
"step": 676
},
{
"epoch": 0.4172573189522342,
"grad_norm": 0.6583855748176575,
"learning_rate": 2.8951031342283757e-05,
"loss": 1.4958,
"step": 677
},
{
"epoch": 0.41787365177195684,
"grad_norm": 0.6277612447738647,
"learning_rate": 2.8912499384130196e-05,
"loss": 1.4656,
"step": 678
},
{
"epoch": 0.4184899845916795,
"grad_norm": 0.6526576280593872,
"learning_rate": 2.8873926103294115e-05,
"loss": 1.4862,
"step": 679
},
{
"epoch": 0.41910631741140214,
"grad_norm": 0.6871560215950012,
"learning_rate": 2.8835311678619997e-05,
"loss": 1.5168,
"step": 680
},
{
"epoch": 0.4197226502311248,
"grad_norm": 0.6310210824012756,
"learning_rate": 2.8796656289143114e-05,
"loss": 1.4582,
"step": 681
},
{
"epoch": 0.42033898305084744,
"grad_norm": 0.6626084446907043,
"learning_rate": 2.875796011408863e-05,
"loss": 1.5576,
"step": 682
},
{
"epoch": 0.4209553158705701,
"grad_norm": 0.6169494390487671,
"learning_rate": 2.8719223332870864e-05,
"loss": 1.4072,
"step": 683
},
{
"epoch": 0.42157164869029273,
"grad_norm": 0.6335439085960388,
"learning_rate": 2.868044612509236e-05,
"loss": 1.4381,
"step": 684
},
{
"epoch": 0.4221879815100154,
"grad_norm": 0.6794674396514893,
"learning_rate": 2.8641628670543116e-05,
"loss": 1.5033,
"step": 685
},
{
"epoch": 0.4228043143297381,
"grad_norm": 0.6509531736373901,
"learning_rate": 2.860277114919974e-05,
"loss": 1.5414,
"step": 686
},
{
"epoch": 0.4234206471494607,
"grad_norm": 0.6503686904907227,
"learning_rate": 2.8563873741224602e-05,
"loss": 1.5238,
"step": 687
},
{
"epoch": 0.4240369799691834,
"grad_norm": 0.6381911039352417,
"learning_rate": 2.8524936626964996e-05,
"loss": 1.5677,
"step": 688
},
{
"epoch": 0.424653312788906,
"grad_norm": 0.652459979057312,
"learning_rate": 2.8485959986952344e-05,
"loss": 1.5336,
"step": 689
},
{
"epoch": 0.4252696456086287,
"grad_norm": 0.6382361054420471,
"learning_rate": 2.8446944001901298e-05,
"loss": 1.4167,
"step": 690
},
{
"epoch": 0.4258859784283513,
"grad_norm": 0.6898105144500732,
"learning_rate": 2.8407888852708932e-05,
"loss": 1.4322,
"step": 691
},
{
"epoch": 0.426502311248074,
"grad_norm": 0.687101423740387,
"learning_rate": 2.8368794720453935e-05,
"loss": 1.4322,
"step": 692
},
{
"epoch": 0.4271186440677966,
"grad_norm": 0.6702199578285217,
"learning_rate": 2.832966178639571e-05,
"loss": 1.4865,
"step": 693
},
{
"epoch": 0.42773497688751927,
"grad_norm": 0.6446210741996765,
"learning_rate": 2.8290490231973575e-05,
"loss": 1.4972,
"step": 694
},
{
"epoch": 0.4283513097072419,
"grad_norm": 0.6693007946014404,
"learning_rate": 2.8251280238805913e-05,
"loss": 1.4929,
"step": 695
},
{
"epoch": 0.42896764252696457,
"grad_norm": 0.6394990682601929,
"learning_rate": 2.8212031988689333e-05,
"loss": 1.4326,
"step": 696
},
{
"epoch": 0.4295839753466872,
"grad_norm": 0.6386460065841675,
"learning_rate": 2.81727456635978e-05,
"loss": 1.4112,
"step": 697
},
{
"epoch": 0.43020030816640986,
"grad_norm": 0.6588038802146912,
"learning_rate": 2.8133421445681843e-05,
"loss": 1.4654,
"step": 698
},
{
"epoch": 0.43081664098613254,
"grad_norm": 0.6661253571510315,
"learning_rate": 2.8094059517267646e-05,
"loss": 1.5444,
"step": 699
},
{
"epoch": 0.43143297380585516,
"grad_norm": 0.6428791284561157,
"learning_rate": 2.805466006085628e-05,
"loss": 1.4637,
"step": 700
},
{
"epoch": 0.43204930662557783,
"grad_norm": 0.6493973135948181,
"learning_rate": 2.801522325912277e-05,
"loss": 1.4362,
"step": 701
},
{
"epoch": 0.43266563944530045,
"grad_norm": 0.6501812934875488,
"learning_rate": 2.7975749294915315e-05,
"loss": 1.4655,
"step": 702
},
{
"epoch": 0.43328197226502313,
"grad_norm": 0.6604790687561035,
"learning_rate": 2.793623835125443e-05,
"loss": 1.4996,
"step": 703
},
{
"epoch": 0.43389830508474575,
"grad_norm": 0.6836243867874146,
"learning_rate": 2.7896690611332053e-05,
"loss": 1.4963,
"step": 704
},
{
"epoch": 0.4345146379044684,
"grad_norm": 0.6414328813552856,
"learning_rate": 2.785710625851075e-05,
"loss": 1.4454,
"step": 705
},
{
"epoch": 0.43513097072419105,
"grad_norm": 0.651253342628479,
"learning_rate": 2.781748547632284e-05,
"loss": 1.5247,
"step": 706
},
{
"epoch": 0.4357473035439137,
"grad_norm": 0.6446231007575989,
"learning_rate": 2.7777828448469536e-05,
"loss": 1.4538,
"step": 707
},
{
"epoch": 0.43636363636363634,
"grad_norm": 0.6915920972824097,
"learning_rate": 2.7738135358820114e-05,
"loss": 1.5617,
"step": 708
},
{
"epoch": 0.436979969183359,
"grad_norm": 0.6558811068534851,
"learning_rate": 2.769840639141104e-05,
"loss": 1.4662,
"step": 709
},
{
"epoch": 0.43759630200308164,
"grad_norm": 0.6515608429908752,
"learning_rate": 2.7658641730445142e-05,
"loss": 1.4382,
"step": 710
},
{
"epoch": 0.4382126348228043,
"grad_norm": 0.6421152353286743,
"learning_rate": 2.7618841560290722e-05,
"loss": 1.4465,
"step": 711
},
{
"epoch": 0.438828967642527,
"grad_norm": 0.6631745100021362,
"learning_rate": 2.7579006065480737e-05,
"loss": 1.4599,
"step": 712
},
{
"epoch": 0.4394453004622496,
"grad_norm": 0.6604713797569275,
"learning_rate": 2.7539135430711908e-05,
"loss": 1.4806,
"step": 713
},
{
"epoch": 0.4400616332819723,
"grad_norm": 0.6700551509857178,
"learning_rate": 2.74992298408439e-05,
"loss": 1.4934,
"step": 714
},
{
"epoch": 0.4406779661016949,
"grad_norm": 0.6602810621261597,
"learning_rate": 2.7459289480898443e-05,
"loss": 1.5138,
"step": 715
},
{
"epoch": 0.4412942989214176,
"grad_norm": 0.6400670409202576,
"learning_rate": 2.7419314536058467e-05,
"loss": 1.4378,
"step": 716
},
{
"epoch": 0.4419106317411402,
"grad_norm": 0.6445081830024719,
"learning_rate": 2.7379305191667263e-05,
"loss": 1.4935,
"step": 717
},
{
"epoch": 0.4425269645608629,
"grad_norm": 0.6491249799728394,
"learning_rate": 2.7339261633227612e-05,
"loss": 1.4695,
"step": 718
},
{
"epoch": 0.4431432973805855,
"grad_norm": 0.6502711176872253,
"learning_rate": 2.7299184046400936e-05,
"loss": 1.4849,
"step": 719
},
{
"epoch": 0.4437596302003082,
"grad_norm": 0.6470304131507874,
"learning_rate": 2.725907261700642e-05,
"loss": 1.4573,
"step": 720
},
{
"epoch": 0.4443759630200308,
"grad_norm": 0.6609827280044556,
"learning_rate": 2.721892753102016e-05,
"loss": 1.481,
"step": 721
},
{
"epoch": 0.44499229583975347,
"grad_norm": 0.6376433372497559,
"learning_rate": 2.71787489745743e-05,
"loss": 1.4438,
"step": 722
},
{
"epoch": 0.4456086286594761,
"grad_norm": 0.6593858003616333,
"learning_rate": 2.7138537133956186e-05,
"loss": 1.5037,
"step": 723
},
{
"epoch": 0.44622496147919877,
"grad_norm": 0.6915332078933716,
"learning_rate": 2.7098292195607456e-05,
"loss": 1.4216,
"step": 724
},
{
"epoch": 0.44684129429892144,
"grad_norm": 0.6552739143371582,
"learning_rate": 2.7058014346123225e-05,
"loss": 1.4874,
"step": 725
},
{
"epoch": 0.44745762711864406,
"grad_norm": 0.6488803625106812,
"learning_rate": 2.70177037722512e-05,
"loss": 1.5355,
"step": 726
},
{
"epoch": 0.44807395993836674,
"grad_norm": 0.6355862021446228,
"learning_rate": 2.69773606608908e-05,
"loss": 1.4827,
"step": 727
},
{
"epoch": 0.44869029275808936,
"grad_norm": 0.652966320514679,
"learning_rate": 2.6936985199092335e-05,
"loss": 1.4875,
"step": 728
},
{
"epoch": 0.44930662557781204,
"grad_norm": 0.6532738208770752,
"learning_rate": 2.6896577574056065e-05,
"loss": 1.5267,
"step": 729
},
{
"epoch": 0.44992295839753466,
"grad_norm": 0.6622675657272339,
"learning_rate": 2.6856137973131403e-05,
"loss": 1.449,
"step": 730
},
{
"epoch": 0.45053929121725733,
"grad_norm": 0.641776442527771,
"learning_rate": 2.681566658381601e-05,
"loss": 1.472,
"step": 731
},
{
"epoch": 0.45115562403697995,
"grad_norm": 0.6786929965019226,
"learning_rate": 2.6775163593754937e-05,
"loss": 1.5539,
"step": 732
},
{
"epoch": 0.45177195685670263,
"grad_norm": 0.6311996579170227,
"learning_rate": 2.6734629190739744e-05,
"loss": 1.4317,
"step": 733
},
{
"epoch": 0.45238828967642525,
"grad_norm": 0.6562221050262451,
"learning_rate": 2.669406356270764e-05,
"loss": 1.5049,
"step": 734
},
{
"epoch": 0.4530046224961479,
"grad_norm": 0.6616199612617493,
"learning_rate": 2.6653466897740607e-05,
"loss": 1.4478,
"step": 735
},
{
"epoch": 0.45362095531587054,
"grad_norm": 0.6472796201705933,
"learning_rate": 2.6612839384064546e-05,
"loss": 1.4283,
"step": 736
},
{
"epoch": 0.4542372881355932,
"grad_norm": 0.6530714631080627,
"learning_rate": 2.657218121004836e-05,
"loss": 1.4406,
"step": 737
},
{
"epoch": 0.4548536209553159,
"grad_norm": 0.709610104560852,
"learning_rate": 2.653149256420312e-05,
"loss": 1.4922,
"step": 738
},
{
"epoch": 0.4554699537750385,
"grad_norm": 0.6500923037528992,
"learning_rate": 2.6490773635181188e-05,
"loss": 1.4193,
"step": 739
},
{
"epoch": 0.4560862865947612,
"grad_norm": 0.6618855595588684,
"learning_rate": 2.6450024611775318e-05,
"loss": 1.4259,
"step": 740
},
{
"epoch": 0.4567026194144838,
"grad_norm": 0.6506283283233643,
"learning_rate": 2.6409245682917814e-05,
"loss": 1.5005,
"step": 741
},
{
"epoch": 0.4573189522342065,
"grad_norm": 0.6780362725257874,
"learning_rate": 2.6368437037679625e-05,
"loss": 1.4403,
"step": 742
},
{
"epoch": 0.4579352850539291,
"grad_norm": 0.6576229333877563,
"learning_rate": 2.632759886526947e-05,
"loss": 1.4727,
"step": 743
},
{
"epoch": 0.4585516178736518,
"grad_norm": 0.671320915222168,
"learning_rate": 2.6286731355033e-05,
"loss": 1.4791,
"step": 744
},
{
"epoch": 0.4591679506933744,
"grad_norm": 0.6556963920593262,
"learning_rate": 2.6245834696451857e-05,
"loss": 1.4887,
"step": 745
},
{
"epoch": 0.4597842835130971,
"grad_norm": 0.651106059551239,
"learning_rate": 2.6204909079142847e-05,
"loss": 1.4462,
"step": 746
},
{
"epoch": 0.4604006163328197,
"grad_norm": 0.6544069051742554,
"learning_rate": 2.6163954692857047e-05,
"loss": 1.5272,
"step": 747
},
{
"epoch": 0.4610169491525424,
"grad_norm": 0.6927053332328796,
"learning_rate": 2.61229717274789e-05,
"loss": 1.4843,
"step": 748
},
{
"epoch": 0.461633281972265,
"grad_norm": 0.6328755617141724,
"learning_rate": 2.6081960373025378e-05,
"loss": 1.3956,
"step": 749
},
{
"epoch": 0.4622496147919877,
"grad_norm": 0.6715585589408875,
"learning_rate": 2.604092081964507e-05,
"loss": 1.4925,
"step": 750
},
{
"epoch": 0.46286594761171035,
"grad_norm": 0.6592444181442261,
"learning_rate": 2.5999853257617315e-05,
"loss": 1.5232,
"step": 751
},
{
"epoch": 0.46348228043143297,
"grad_norm": 0.6313751339912415,
"learning_rate": 2.5958757877351294e-05,
"loss": 1.4398,
"step": 752
},
{
"epoch": 0.46409861325115565,
"grad_norm": 0.6490797996520996,
"learning_rate": 2.5917634869385185e-05,
"loss": 1.4558,
"step": 753
},
{
"epoch": 0.46471494607087827,
"grad_norm": 0.6704680919647217,
"learning_rate": 2.5876484424385263e-05,
"loss": 1.5176,
"step": 754
},
{
"epoch": 0.46533127889060094,
"grad_norm": 0.6712368726730347,
"learning_rate": 2.5835306733144996e-05,
"loss": 1.4243,
"step": 755
},
{
"epoch": 0.46594761171032356,
"grad_norm": 0.6699070334434509,
"learning_rate": 2.57941019865842e-05,
"loss": 1.5105,
"step": 756
},
{
"epoch": 0.46656394453004624,
"grad_norm": 0.6497899293899536,
"learning_rate": 2.575287037574812e-05,
"loss": 1.4815,
"step": 757
},
{
"epoch": 0.46718027734976886,
"grad_norm": 0.6557024717330933,
"learning_rate": 2.5711612091806566e-05,
"loss": 1.4426,
"step": 758
},
{
"epoch": 0.46779661016949153,
"grad_norm": 0.6698507070541382,
"learning_rate": 2.567032732605301e-05,
"loss": 1.4753,
"step": 759
},
{
"epoch": 0.46841294298921415,
"grad_norm": 0.6543661952018738,
"learning_rate": 2.5629016269903706e-05,
"loss": 1.498,
"step": 760
},
{
"epoch": 0.46902927580893683,
"grad_norm": 0.6658725738525391,
"learning_rate": 2.5587679114896806e-05,
"loss": 1.4045,
"step": 761
},
{
"epoch": 0.46964560862865945,
"grad_norm": 0.6653224229812622,
"learning_rate": 2.5546316052691467e-05,
"loss": 1.469,
"step": 762
},
{
"epoch": 0.4702619414483821,
"grad_norm": 0.6615314483642578,
"learning_rate": 2.550492727506696e-05,
"loss": 1.5018,
"step": 763
},
{
"epoch": 0.4708782742681048,
"grad_norm": 0.6513172388076782,
"learning_rate": 2.5463512973921814e-05,
"loss": 1.4181,
"step": 764
},
{
"epoch": 0.4714946070878274,
"grad_norm": 0.6554753184318542,
"learning_rate": 2.5422073341272852e-05,
"loss": 1.3937,
"step": 765
},
{
"epoch": 0.4721109399075501,
"grad_norm": 0.6689674258232117,
"learning_rate": 2.538060856925437e-05,
"loss": 1.4254,
"step": 766
},
{
"epoch": 0.4727272727272727,
"grad_norm": 0.6607903242111206,
"learning_rate": 2.5339118850117244e-05,
"loss": 1.4113,
"step": 767
},
{
"epoch": 0.4733436055469954,
"grad_norm": 0.6706930994987488,
"learning_rate": 2.5297604376227963e-05,
"loss": 1.4264,
"step": 768
},
{
"epoch": 0.473959938366718,
"grad_norm": 0.6535710692405701,
"learning_rate": 2.5256065340067843e-05,
"loss": 1.4139,
"step": 769
},
{
"epoch": 0.4745762711864407,
"grad_norm": 0.6666339039802551,
"learning_rate": 2.521450193423206e-05,
"loss": 1.4982,
"step": 770
},
{
"epoch": 0.4751926040061633,
"grad_norm": 0.6591233611106873,
"learning_rate": 2.5172914351428765e-05,
"loss": 1.472,
"step": 771
},
{
"epoch": 0.475808936825886,
"grad_norm": 0.6556514501571655,
"learning_rate": 2.513130278447824e-05,
"loss": 1.4035,
"step": 772
},
{
"epoch": 0.4764252696456086,
"grad_norm": 0.681161642074585,
"learning_rate": 2.508966742631195e-05,
"loss": 1.5286,
"step": 773
},
{
"epoch": 0.4770416024653313,
"grad_norm": 0.651854395866394,
"learning_rate": 2.504800846997165e-05,
"loss": 1.4774,
"step": 774
},
{
"epoch": 0.4776579352850539,
"grad_norm": 0.6415034532546997,
"learning_rate": 2.500632610860855e-05,
"loss": 1.4045,
"step": 775
},
{
"epoch": 0.4782742681047766,
"grad_norm": 0.6623095273971558,
"learning_rate": 2.4964620535482336e-05,
"loss": 1.4454,
"step": 776
},
{
"epoch": 0.47889060092449925,
"grad_norm": 0.6609742045402527,
"learning_rate": 2.492289194396034e-05,
"loss": 1.4954,
"step": 777
},
{
"epoch": 0.4795069337442219,
"grad_norm": 0.6443544030189514,
"learning_rate": 2.4881140527516613e-05,
"loss": 1.4089,
"step": 778
},
{
"epoch": 0.48012326656394455,
"grad_norm": 0.6567139029502869,
"learning_rate": 2.483936647973103e-05,
"loss": 1.4591,
"step": 779
},
{
"epoch": 0.48073959938366717,
"grad_norm": 0.6636015772819519,
"learning_rate": 2.4797569994288385e-05,
"loss": 1.4427,
"step": 780
},
{
"epoch": 0.48135593220338985,
"grad_norm": 0.6707240343093872,
"learning_rate": 2.475575126497753e-05,
"loss": 1.5628,
"step": 781
},
{
"epoch": 0.48197226502311247,
"grad_norm": 0.667204737663269,
"learning_rate": 2.4713910485690428e-05,
"loss": 1.4458,
"step": 782
},
{
"epoch": 0.48258859784283514,
"grad_norm": 0.6726914644241333,
"learning_rate": 2.4672047850421295e-05,
"loss": 1.4424,
"step": 783
},
{
"epoch": 0.48320493066255776,
"grad_norm": 0.6603702902793884,
"learning_rate": 2.4630163553265657e-05,
"loss": 1.4882,
"step": 784
},
{
"epoch": 0.48382126348228044,
"grad_norm": 0.6715510487556458,
"learning_rate": 2.4588257788419503e-05,
"loss": 1.4697,
"step": 785
},
{
"epoch": 0.48443759630200306,
"grad_norm": 0.6559399366378784,
"learning_rate": 2.454633075017834e-05,
"loss": 1.4835,
"step": 786
},
{
"epoch": 0.48505392912172574,
"grad_norm": 0.6464762687683105,
"learning_rate": 2.4504382632936303e-05,
"loss": 1.4401,
"step": 787
},
{
"epoch": 0.48567026194144836,
"grad_norm": 0.6455738544464111,
"learning_rate": 2.446241363118528e-05,
"loss": 1.39,
"step": 788
},
{
"epoch": 0.48628659476117103,
"grad_norm": 0.6899600625038147,
"learning_rate": 2.4420423939513973e-05,
"loss": 1.4898,
"step": 789
},
{
"epoch": 0.4869029275808937,
"grad_norm": 0.6539801955223083,
"learning_rate": 2.4378413752607025e-05,
"loss": 1.4481,
"step": 790
},
{
"epoch": 0.4875192604006163,
"grad_norm": 0.6555554866790771,
"learning_rate": 2.4336383265244088e-05,
"loss": 1.4525,
"step": 791
},
{
"epoch": 0.488135593220339,
"grad_norm": 0.6425526142120361,
"learning_rate": 2.4294332672298947e-05,
"loss": 1.3769,
"step": 792
},
{
"epoch": 0.4887519260400616,
"grad_norm": 0.6495502591133118,
"learning_rate": 2.4252262168738617e-05,
"loss": 1.4287,
"step": 793
},
{
"epoch": 0.4893682588597843,
"grad_norm": 0.6833025217056274,
"learning_rate": 2.4210171949622408e-05,
"loss": 1.4681,
"step": 794
},
{
"epoch": 0.4899845916795069,
"grad_norm": 0.6515230536460876,
"learning_rate": 2.416806221010105e-05,
"loss": 1.4676,
"step": 795
},
{
"epoch": 0.4906009244992296,
"grad_norm": 0.6599568128585815,
"learning_rate": 2.412593314541578e-05,
"loss": 1.4813,
"step": 796
},
{
"epoch": 0.4912172573189522,
"grad_norm": 0.6508007645606995,
"learning_rate": 2.408378495089744e-05,
"loss": 1.4416,
"step": 797
},
{
"epoch": 0.4918335901386749,
"grad_norm": 0.6619030237197876,
"learning_rate": 2.404161782196556e-05,
"loss": 1.4391,
"step": 798
},
{
"epoch": 0.4924499229583975,
"grad_norm": 0.6672455668449402,
"learning_rate": 2.3999431954127457e-05,
"loss": 1.5233,
"step": 799
},
{
"epoch": 0.4930662557781202,
"grad_norm": 0.6848391890525818,
"learning_rate": 2.3957227542977334e-05,
"loss": 1.5218,
"step": 800
},
{
"epoch": 0.4936825885978428,
"grad_norm": 0.6891319751739502,
"learning_rate": 2.3915004784195386e-05,
"loss": 1.5312,
"step": 801
},
{
"epoch": 0.4942989214175655,
"grad_norm": 0.664033055305481,
"learning_rate": 2.3872763873546835e-05,
"loss": 1.4585,
"step": 802
},
{
"epoch": 0.49491525423728816,
"grad_norm": 0.6545723676681519,
"learning_rate": 2.3830505006881125e-05,
"loss": 1.4515,
"step": 803
},
{
"epoch": 0.4955315870570108,
"grad_norm": 0.6647583246231079,
"learning_rate": 2.3788228380130884e-05,
"loss": 1.4261,
"step": 804
},
{
"epoch": 0.49614791987673346,
"grad_norm": 0.6671835780143738,
"learning_rate": 2.374593418931114e-05,
"loss": 1.5139,
"step": 805
},
{
"epoch": 0.4967642526964561,
"grad_norm": 0.6669169664382935,
"learning_rate": 2.370362263051833e-05,
"loss": 1.4518,
"step": 806
},
{
"epoch": 0.49738058551617875,
"grad_norm": 0.6806299686431885,
"learning_rate": 2.3661293899929417e-05,
"loss": 1.4638,
"step": 807
},
{
"epoch": 0.4979969183359014,
"grad_norm": 0.6680865287780762,
"learning_rate": 2.3618948193800996e-05,
"loss": 1.4511,
"step": 808
},
{
"epoch": 0.49861325115562405,
"grad_norm": 0.6399280428886414,
"learning_rate": 2.3576585708468355e-05,
"loss": 1.3566,
"step": 809
},
{
"epoch": 0.49922958397534667,
"grad_norm": 0.6574198603630066,
"learning_rate": 2.353420664034457e-05,
"loss": 1.5564,
"step": 810
},
{
"epoch": 0.49984591679506934,
"grad_norm": 0.6828180551528931,
"learning_rate": 2.3491811185919644e-05,
"loss": 1.4455,
"step": 811
},
{
"epoch": 0.500462249614792,
"grad_norm": 0.6625897288322449,
"learning_rate": 2.3449399541759498e-05,
"loss": 1.3979,
"step": 812
},
{
"epoch": 0.5010785824345146,
"grad_norm": 0.6758010983467102,
"learning_rate": 2.3406971904505155e-05,
"loss": 1.4646,
"step": 813
},
{
"epoch": 0.5016949152542373,
"grad_norm": 0.6834417581558228,
"learning_rate": 2.336452847087179e-05,
"loss": 1.4211,
"step": 814
},
{
"epoch": 0.50231124807396,
"grad_norm": 0.6549422144889832,
"learning_rate": 2.3322069437647795e-05,
"loss": 1.3809,
"step": 815
},
{
"epoch": 0.5029275808936826,
"grad_norm": 0.6464531421661377,
"learning_rate": 2.3279595001693903e-05,
"loss": 1.4303,
"step": 816
},
{
"epoch": 0.5035439137134052,
"grad_norm": 0.6724410653114319,
"learning_rate": 2.3237105359942272e-05,
"loss": 1.5311,
"step": 817
},
{
"epoch": 0.5041602465331279,
"grad_norm": 0.6607643365859985,
"learning_rate": 2.319460070939553e-05,
"loss": 1.4646,
"step": 818
},
{
"epoch": 0.5047765793528506,
"grad_norm": 0.6936811208724976,
"learning_rate": 2.315208124712593e-05,
"loss": 1.491,
"step": 819
},
{
"epoch": 0.5053929121725732,
"grad_norm": 0.6466771364212036,
"learning_rate": 2.3109547170274373e-05,
"loss": 1.415,
"step": 820
},
{
"epoch": 0.5060092449922958,
"grad_norm": 0.6754940152168274,
"learning_rate": 2.306699867604953e-05,
"loss": 1.4786,
"step": 821
},
{
"epoch": 0.5066255778120184,
"grad_norm": 0.6594055891036987,
"learning_rate": 2.3024435961726926e-05,
"loss": 1.4387,
"step": 822
},
{
"epoch": 0.5072419106317412,
"grad_norm": 0.6514335870742798,
"learning_rate": 2.2981859224647997e-05,
"loss": 1.4451,
"step": 823
},
{
"epoch": 0.5078582434514638,
"grad_norm": 0.659854531288147,
"learning_rate": 2.2939268662219213e-05,
"loss": 1.4717,
"step": 824
},
{
"epoch": 0.5084745762711864,
"grad_norm": 0.6732680201530457,
"learning_rate": 2.289666447191114e-05,
"loss": 1.5037,
"step": 825
},
{
"epoch": 0.509090909090909,
"grad_norm": 0.657238781452179,
"learning_rate": 2.2854046851257525e-05,
"loss": 1.4832,
"step": 826
},
{
"epoch": 0.5097072419106318,
"grad_norm": 0.6703278422355652,
"learning_rate": 2.281141599785439e-05,
"loss": 1.4503,
"step": 827
},
{
"epoch": 0.5103235747303544,
"grad_norm": 0.6670629382133484,
"learning_rate": 2.2768772109359114e-05,
"loss": 1.4201,
"step": 828
},
{
"epoch": 0.510939907550077,
"grad_norm": 0.6621464490890503,
"learning_rate": 2.2726115383489495e-05,
"loss": 1.4666,
"step": 829
},
{
"epoch": 0.5115562403697997,
"grad_norm": 0.6738835573196411,
"learning_rate": 2.2683446018022882e-05,
"loss": 1.5386,
"step": 830
},
{
"epoch": 0.5121725731895224,
"grad_norm": 0.6561502814292908,
"learning_rate": 2.2640764210795187e-05,
"loss": 1.4534,
"step": 831
},
{
"epoch": 0.512788906009245,
"grad_norm": 0.6672171354293823,
"learning_rate": 2.2598070159700043e-05,
"loss": 1.4285,
"step": 832
},
{
"epoch": 0.5134052388289676,
"grad_norm": 0.6675829291343689,
"learning_rate": 2.2555364062687844e-05,
"loss": 1.4009,
"step": 833
},
{
"epoch": 0.5140215716486903,
"grad_norm": 0.6768284440040588,
"learning_rate": 2.2512646117764828e-05,
"loss": 1.3978,
"step": 834
},
{
"epoch": 0.514637904468413,
"grad_norm": 0.6675807237625122,
"learning_rate": 2.2469916522992155e-05,
"loss": 1.4627,
"step": 835
},
{
"epoch": 0.5152542372881356,
"grad_norm": 0.6885505318641663,
"learning_rate": 2.242717547648502e-05,
"loss": 1.4702,
"step": 836
},
{
"epoch": 0.5158705701078582,
"grad_norm": 0.6633325219154358,
"learning_rate": 2.238442317641171e-05,
"loss": 1.4615,
"step": 837
},
{
"epoch": 0.5164869029275809,
"grad_norm": 0.6807160973548889,
"learning_rate": 2.234165982099268e-05,
"loss": 1.5114,
"step": 838
},
{
"epoch": 0.5171032357473035,
"grad_norm": 0.6771484017372131,
"learning_rate": 2.229888560849964e-05,
"loss": 1.4715,
"step": 839
},
{
"epoch": 0.5177195685670262,
"grad_norm": 0.6473548412322998,
"learning_rate": 2.2256100737254657e-05,
"loss": 1.3995,
"step": 840
},
{
"epoch": 0.5183359013867489,
"grad_norm": 0.7034615874290466,
"learning_rate": 2.22133054056292e-05,
"loss": 1.5146,
"step": 841
},
{
"epoch": 0.5189522342064715,
"grad_norm": 0.7064915895462036,
"learning_rate": 2.217049981204325e-05,
"loss": 1.4643,
"step": 842
},
{
"epoch": 0.5195685670261941,
"grad_norm": 0.859814465045929,
"learning_rate": 2.2127684154964357e-05,
"loss": 1.4419,
"step": 843
},
{
"epoch": 0.5201848998459168,
"grad_norm": 0.6511772871017456,
"learning_rate": 2.2084858632906733e-05,
"loss": 1.4144,
"step": 844
},
{
"epoch": 0.5208012326656395,
"grad_norm": 0.6685275435447693,
"learning_rate": 2.2042023444430333e-05,
"loss": 1.4596,
"step": 845
},
{
"epoch": 0.5214175654853621,
"grad_norm": 0.6598917841911316,
"learning_rate": 2.199917878813992e-05,
"loss": 1.4823,
"step": 846
},
{
"epoch": 0.5220338983050847,
"grad_norm": 0.6725688576698303,
"learning_rate": 2.195632486268418e-05,
"loss": 1.5023,
"step": 847
},
{
"epoch": 0.5226502311248074,
"grad_norm": 0.7149176597595215,
"learning_rate": 2.1913461866754747e-05,
"loss": 1.4431,
"step": 848
},
{
"epoch": 0.5232665639445301,
"grad_norm": 0.6676431894302368,
"learning_rate": 2.1870589999085302e-05,
"loss": 1.3993,
"step": 849
},
{
"epoch": 0.5238828967642527,
"grad_norm": 0.6679561734199524,
"learning_rate": 2.182770945845071e-05,
"loss": 1.4462,
"step": 850
},
{
"epoch": 0.5244992295839753,
"grad_norm": 0.6490969061851501,
"learning_rate": 2.178482044366599e-05,
"loss": 1.435,
"step": 851
},
{
"epoch": 0.5251155624036979,
"grad_norm": 0.6597004532814026,
"learning_rate": 2.1741923153585497e-05,
"loss": 1.4516,
"step": 852
},
{
"epoch": 0.5257318952234207,
"grad_norm": 0.6608878374099731,
"learning_rate": 2.1699017787101924e-05,
"loss": 1.4543,
"step": 853
},
{
"epoch": 0.5263482280431433,
"grad_norm": 0.6896352767944336,
"learning_rate": 2.165610454314542e-05,
"loss": 1.4563,
"step": 854
},
{
"epoch": 0.5269645608628659,
"grad_norm": 0.6917398571968079,
"learning_rate": 2.1613183620682662e-05,
"loss": 1.3637,
"step": 855
},
{
"epoch": 0.5275808936825886,
"grad_norm": 0.6737967133522034,
"learning_rate": 2.157025521871593e-05,
"loss": 1.3772,
"step": 856
},
{
"epoch": 0.5281972265023113,
"grad_norm": 0.6648988723754883,
"learning_rate": 2.152731953628216e-05,
"loss": 1.4338,
"step": 857
},
{
"epoch": 0.5288135593220339,
"grad_norm": 0.6846292018890381,
"learning_rate": 2.1484376772452087e-05,
"loss": 1.499,
"step": 858
},
{
"epoch": 0.5294298921417565,
"grad_norm": 0.661718487739563,
"learning_rate": 2.144142712632923e-05,
"loss": 1.4039,
"step": 859
},
{
"epoch": 0.5300462249614792,
"grad_norm": 0.679577648639679,
"learning_rate": 2.1398470797049055e-05,
"loss": 1.4868,
"step": 860
},
{
"epoch": 0.5306625577812019,
"grad_norm": 0.6697185635566711,
"learning_rate": 2.1355507983778e-05,
"loss": 1.4118,
"step": 861
},
{
"epoch": 0.5312788906009245,
"grad_norm": 0.6589436531066895,
"learning_rate": 2.131253888571255e-05,
"loss": 1.4483,
"step": 862
},
{
"epoch": 0.5318952234206471,
"grad_norm": 0.6604962348937988,
"learning_rate": 2.126956370207836e-05,
"loss": 1.4177,
"step": 863
},
{
"epoch": 0.5325115562403698,
"grad_norm": 0.6640836596488953,
"learning_rate": 2.1226582632129286e-05,
"loss": 1.4351,
"step": 864
},
{
"epoch": 0.5331278890600925,
"grad_norm": 0.659049928188324,
"learning_rate": 2.1183595875146467e-05,
"loss": 1.4109,
"step": 865
},
{
"epoch": 0.5337442218798151,
"grad_norm": 0.6786544322967529,
"learning_rate": 2.1140603630437434e-05,
"loss": 1.4525,
"step": 866
},
{
"epoch": 0.5343605546995378,
"grad_norm": 0.6710315346717834,
"learning_rate": 2.1097606097335133e-05,
"loss": 1.4424,
"step": 867
},
{
"epoch": 0.5349768875192604,
"grad_norm": 0.658591091632843,
"learning_rate": 2.1054603475197045e-05,
"loss": 1.4215,
"step": 868
},
{
"epoch": 0.535593220338983,
"grad_norm": 0.6709784865379333,
"learning_rate": 2.1011595963404253e-05,
"loss": 1.4753,
"step": 869
},
{
"epoch": 0.5362095531587057,
"grad_norm": 0.6819916367530823,
"learning_rate": 2.096858376136049e-05,
"loss": 1.4439,
"step": 870
},
{
"epoch": 0.5368258859784284,
"grad_norm": 0.6897315382957458,
"learning_rate": 2.092556706849125e-05,
"loss": 1.4687,
"step": 871
},
{
"epoch": 0.537442218798151,
"grad_norm": 0.6903764009475708,
"learning_rate": 2.0882546084242856e-05,
"loss": 1.4768,
"step": 872
},
{
"epoch": 0.5380585516178736,
"grad_norm": 0.677721381187439,
"learning_rate": 2.0839521008081508e-05,
"loss": 1.4324,
"step": 873
},
{
"epoch": 0.5386748844375963,
"grad_norm": 0.6687971353530884,
"learning_rate": 2.0796492039492394e-05,
"loss": 1.4655,
"step": 874
},
{
"epoch": 0.539291217257319,
"grad_norm": 0.6661518216133118,
"learning_rate": 2.0753459377978737e-05,
"loss": 1.4712,
"step": 875
},
{
"epoch": 0.5399075500770416,
"grad_norm": 0.6647510528564453,
"learning_rate": 2.071042322306089e-05,
"loss": 1.423,
"step": 876
},
{
"epoch": 0.5405238828967642,
"grad_norm": 0.7283526062965393,
"learning_rate": 2.066738377427541e-05,
"loss": 1.4547,
"step": 877
},
{
"epoch": 0.5411402157164868,
"grad_norm": 0.679385781288147,
"learning_rate": 2.0624341231174113e-05,
"loss": 1.4217,
"step": 878
},
{
"epoch": 0.5417565485362096,
"grad_norm": 0.6685990691184998,
"learning_rate": 2.0581295793323156e-05,
"loss": 1.4135,
"step": 879
},
{
"epoch": 0.5423728813559322,
"grad_norm": 0.6681914329528809,
"learning_rate": 2.0538247660302143e-05,
"loss": 1.4704,
"step": 880
},
{
"epoch": 0.5429892141756548,
"grad_norm": 0.6571676135063171,
"learning_rate": 2.0495197031703154e-05,
"loss": 1.3697,
"step": 881
},
{
"epoch": 0.5436055469953776,
"grad_norm": 0.6493473052978516,
"learning_rate": 2.0452144107129844e-05,
"loss": 1.4209,
"step": 882
},
{
"epoch": 0.5442218798151002,
"grad_norm": 0.6795789003372192,
"learning_rate": 2.0409089086196506e-05,
"loss": 1.451,
"step": 883
},
{
"epoch": 0.5448382126348228,
"grad_norm": 0.6869490742683411,
"learning_rate": 2.0366032168527178e-05,
"loss": 1.5008,
"step": 884
},
{
"epoch": 0.5454545454545454,
"grad_norm": 0.6746764183044434,
"learning_rate": 2.032297355375465e-05,
"loss": 1.3978,
"step": 885
},
{
"epoch": 0.5460708782742681,
"grad_norm": 0.6667096018791199,
"learning_rate": 2.027991344151963e-05,
"loss": 1.4117,
"step": 886
},
{
"epoch": 0.5466872110939908,
"grad_norm": 0.6780974268913269,
"learning_rate": 2.0236852031469735e-05,
"loss": 1.4576,
"step": 887
},
{
"epoch": 0.5473035439137134,
"grad_norm": 0.6567050814628601,
"learning_rate": 2.0193789523258608e-05,
"loss": 1.4527,
"step": 888
},
{
"epoch": 0.547919876733436,
"grad_norm": 0.6834569573402405,
"learning_rate": 2.0150726116544982e-05,
"loss": 1.4345,
"step": 889
},
{
"epoch": 0.5485362095531587,
"grad_norm": 0.6610395908355713,
"learning_rate": 2.0107662010991764e-05,
"loss": 1.4347,
"step": 890
},
{
"epoch": 0.5491525423728814,
"grad_norm": 0.6741618514060974,
"learning_rate": 2.006459740626509e-05,
"loss": 1.452,
"step": 891
},
{
"epoch": 0.549768875192604,
"grad_norm": 0.6818397045135498,
"learning_rate": 2.002153250203342e-05,
"loss": 1.408,
"step": 892
},
{
"epoch": 0.5503852080123267,
"grad_norm": 0.6544389128684998,
"learning_rate": 1.997846749796659e-05,
"loss": 1.4037,
"step": 893
},
{
"epoch": 0.5510015408320493,
"grad_norm": 0.6719116568565369,
"learning_rate": 1.9935402593734915e-05,
"loss": 1.4347,
"step": 894
},
{
"epoch": 0.551617873651772,
"grad_norm": 0.6798861026763916,
"learning_rate": 1.9892337989008243e-05,
"loss": 1.4861,
"step": 895
},
{
"epoch": 0.5522342064714946,
"grad_norm": 0.6727917790412903,
"learning_rate": 1.984927388345502e-05,
"loss": 1.4481,
"step": 896
},
{
"epoch": 0.5528505392912173,
"grad_norm": 0.6841053366661072,
"learning_rate": 1.9806210476741396e-05,
"loss": 1.4536,
"step": 897
},
{
"epoch": 0.5534668721109399,
"grad_norm": 0.6897479891777039,
"learning_rate": 1.976314796853027e-05,
"loss": 1.4889,
"step": 898
},
{
"epoch": 0.5540832049306625,
"grad_norm": 0.657565176486969,
"learning_rate": 1.9720086558480372e-05,
"loss": 1.4495,
"step": 899
},
{
"epoch": 0.5546995377503852,
"grad_norm": 0.6826606392860413,
"learning_rate": 1.9677026446245352e-05,
"loss": 1.4127,
"step": 900
},
{
"epoch": 0.5553158705701079,
"grad_norm": 0.6761550307273865,
"learning_rate": 1.9633967831472835e-05,
"loss": 1.4465,
"step": 901
},
{
"epoch": 0.5559322033898305,
"grad_norm": 0.6841764450073242,
"learning_rate": 1.95909109138035e-05,
"loss": 1.4282,
"step": 902
},
{
"epoch": 0.5565485362095531,
"grad_norm": 0.6865301728248596,
"learning_rate": 1.9547855892870166e-05,
"loss": 1.4903,
"step": 903
},
{
"epoch": 0.5571648690292758,
"grad_norm": 0.6651366353034973,
"learning_rate": 1.9504802968296853e-05,
"loss": 1.4387,
"step": 904
},
{
"epoch": 0.5577812018489985,
"grad_norm": 0.6853525042533875,
"learning_rate": 1.9461752339697863e-05,
"loss": 1.4799,
"step": 905
},
{
"epoch": 0.5583975346687211,
"grad_norm": 0.6708539724349976,
"learning_rate": 1.9418704206676847e-05,
"loss": 1.4576,
"step": 906
},
{
"epoch": 0.5590138674884437,
"grad_norm": 0.6524003744125366,
"learning_rate": 1.937565876882589e-05,
"loss": 1.4474,
"step": 907
},
{
"epoch": 0.5596302003081665,
"grad_norm": 0.660274863243103,
"learning_rate": 1.93326162257246e-05,
"loss": 1.4079,
"step": 908
},
{
"epoch": 0.5602465331278891,
"grad_norm": 0.665828287601471,
"learning_rate": 1.9289576776939114e-05,
"loss": 1.4248,
"step": 909
},
{
"epoch": 0.5608628659476117,
"grad_norm": 0.6802533268928528,
"learning_rate": 1.9246540622021266e-05,
"loss": 1.4001,
"step": 910
},
{
"epoch": 0.5614791987673343,
"grad_norm": 0.674877941608429,
"learning_rate": 1.920350796050762e-05,
"loss": 1.4604,
"step": 911
},
{
"epoch": 0.562095531587057,
"grad_norm": 0.6802154183387756,
"learning_rate": 1.9160478991918496e-05,
"loss": 1.441,
"step": 912
},
{
"epoch": 0.5627118644067797,
"grad_norm": 0.6677930951118469,
"learning_rate": 1.9117453915757148e-05,
"loss": 1.3754,
"step": 913
},
{
"epoch": 0.5633281972265023,
"grad_norm": 0.683760941028595,
"learning_rate": 1.907443293150876e-05,
"loss": 1.4508,
"step": 914
},
{
"epoch": 0.5639445300462249,
"grad_norm": 0.6907650232315063,
"learning_rate": 1.9031416238639518e-05,
"loss": 1.4885,
"step": 915
},
{
"epoch": 0.5645608628659476,
"grad_norm": 0.672909140586853,
"learning_rate": 1.8988404036595757e-05,
"loss": 1.3758,
"step": 916
},
{
"epoch": 0.5651771956856703,
"grad_norm": 0.6761062741279602,
"learning_rate": 1.894539652480296e-05,
"loss": 1.4423,
"step": 917
},
{
"epoch": 0.5657935285053929,
"grad_norm": 0.7031134366989136,
"learning_rate": 1.8902393902664873e-05,
"loss": 1.3982,
"step": 918
},
{
"epoch": 0.5664098613251156,
"grad_norm": 0.6689425706863403,
"learning_rate": 1.885939636956257e-05,
"loss": 1.4437,
"step": 919
},
{
"epoch": 0.5670261941448382,
"grad_norm": 0.6898466944694519,
"learning_rate": 1.881640412485353e-05,
"loss": 1.4017,
"step": 920
},
{
"epoch": 0.5676425269645609,
"grad_norm": 0.6723157167434692,
"learning_rate": 1.877341736787072e-05,
"loss": 1.4168,
"step": 921
},
{
"epoch": 0.5682588597842835,
"grad_norm": 0.6810134649276733,
"learning_rate": 1.8730436297921645e-05,
"loss": 1.4349,
"step": 922
},
{
"epoch": 0.5688751926040062,
"grad_norm": 0.6878021955490112,
"learning_rate": 1.8687461114287455e-05,
"loss": 1.333,
"step": 923
},
{
"epoch": 0.5694915254237288,
"grad_norm": 0.7006982564926147,
"learning_rate": 1.8644492016222015e-05,
"loss": 1.4425,
"step": 924
},
{
"epoch": 0.5701078582434514,
"grad_norm": 0.6941596269607544,
"learning_rate": 1.8601529202950948e-05,
"loss": 1.4304,
"step": 925
},
{
"epoch": 0.5707241910631741,
"grad_norm": 0.6846546530723572,
"learning_rate": 1.8558572873670774e-05,
"loss": 1.4505,
"step": 926
},
{
"epoch": 0.5713405238828968,
"grad_norm": 0.6761274337768555,
"learning_rate": 1.8515623227547923e-05,
"loss": 1.4106,
"step": 927
},
{
"epoch": 0.5719568567026194,
"grad_norm": 0.6644502878189087,
"learning_rate": 1.8472680463717842e-05,
"loss": 1.4591,
"step": 928
},
{
"epoch": 0.572573189522342,
"grad_norm": 0.6804108023643494,
"learning_rate": 1.8429744781284072e-05,
"loss": 1.4747,
"step": 929
},
{
"epoch": 0.5731895223420647,
"grad_norm": 0.6691780686378479,
"learning_rate": 1.8386816379317344e-05,
"loss": 1.4035,
"step": 930
},
{
"epoch": 0.5738058551617874,
"grad_norm": 0.7028989195823669,
"learning_rate": 1.8343895456854588e-05,
"loss": 1.448,
"step": 931
},
{
"epoch": 0.57442218798151,
"grad_norm": 0.6800969243049622,
"learning_rate": 1.830098221289808e-05,
"loss": 1.3695,
"step": 932
},
{
"epoch": 0.5750385208012326,
"grad_norm": 0.6787469983100891,
"learning_rate": 1.8258076846414506e-05,
"loss": 1.4078,
"step": 933
},
{
"epoch": 0.5756548536209554,
"grad_norm": 0.6691258549690247,
"learning_rate": 1.8215179556334012e-05,
"loss": 1.3818,
"step": 934
},
{
"epoch": 0.576271186440678,
"grad_norm": 0.677056074142456,
"learning_rate": 1.8172290541549297e-05,
"loss": 1.42,
"step": 935
},
{
"epoch": 0.5768875192604006,
"grad_norm": 0.6807578206062317,
"learning_rate": 1.8129410000914698e-05,
"loss": 1.4405,
"step": 936
},
{
"epoch": 0.5775038520801232,
"grad_norm": 0.6643014550209045,
"learning_rate": 1.8086538133245267e-05,
"loss": 1.3877,
"step": 937
},
{
"epoch": 0.578120184899846,
"grad_norm": 0.6755670309066772,
"learning_rate": 1.8043675137315824e-05,
"loss": 1.4315,
"step": 938
},
{
"epoch": 0.5787365177195686,
"grad_norm": 0.6702075004577637,
"learning_rate": 1.8000821211860077e-05,
"loss": 1.4442,
"step": 939
},
{
"epoch": 0.5793528505392912,
"grad_norm": 0.6891489028930664,
"learning_rate": 1.7957976555569677e-05,
"loss": 1.4554,
"step": 940
},
{
"epoch": 0.5799691833590138,
"grad_norm": 0.6698282361030579,
"learning_rate": 1.7915141367093274e-05,
"loss": 1.4076,
"step": 941
},
{
"epoch": 0.5805855161787365,
"grad_norm": 0.6736328601837158,
"learning_rate": 1.787231584503565e-05,
"loss": 1.4102,
"step": 942
},
{
"epoch": 0.5812018489984592,
"grad_norm": 0.6851081848144531,
"learning_rate": 1.782950018795676e-05,
"loss": 1.4017,
"step": 943
},
{
"epoch": 0.5818181818181818,
"grad_norm": 0.6863130331039429,
"learning_rate": 1.7786694594370806e-05,
"loss": 1.4394,
"step": 944
},
{
"epoch": 0.5824345146379045,
"grad_norm": 0.6876046061515808,
"learning_rate": 1.7743899262745353e-05,
"loss": 1.3686,
"step": 945
},
{
"epoch": 0.5830508474576271,
"grad_norm": 0.6842202544212341,
"learning_rate": 1.7701114391500364e-05,
"loss": 1.3856,
"step": 946
},
{
"epoch": 0.5836671802773498,
"grad_norm": 0.6818761229515076,
"learning_rate": 1.7658340179007335e-05,
"loss": 1.3733,
"step": 947
},
{
"epoch": 0.5842835130970724,
"grad_norm": 0.6609686613082886,
"learning_rate": 1.7615576823588294e-05,
"loss": 1.3586,
"step": 948
},
{
"epoch": 0.5848998459167951,
"grad_norm": 0.6683045625686646,
"learning_rate": 1.7572824523514982e-05,
"loss": 1.454,
"step": 949
},
{
"epoch": 0.5855161787365177,
"grad_norm": 0.6885985732078552,
"learning_rate": 1.7530083477007855e-05,
"loss": 1.443,
"step": 950
},
{
"epoch": 0.5861325115562404,
"grad_norm": 0.67160564661026,
"learning_rate": 1.748735388223518e-05,
"loss": 1.4051,
"step": 951
},
{
"epoch": 0.586748844375963,
"grad_norm": 0.6979023218154907,
"learning_rate": 1.7444635937312156e-05,
"loss": 1.4371,
"step": 952
},
{
"epoch": 0.5873651771956857,
"grad_norm": 0.6684184074401855,
"learning_rate": 1.740192984029996e-05,
"loss": 1.4285,
"step": 953
},
{
"epoch": 0.5879815100154083,
"grad_norm": 0.6790677309036255,
"learning_rate": 1.735923578920482e-05,
"loss": 1.4607,
"step": 954
},
{
"epoch": 0.588597842835131,
"grad_norm": 0.692985475063324,
"learning_rate": 1.7316553981977128e-05,
"loss": 1.3717,
"step": 955
},
{
"epoch": 0.5892141756548536,
"grad_norm": 0.6927673816680908,
"learning_rate": 1.7273884616510505e-05,
"loss": 1.4044,
"step": 956
},
{
"epoch": 0.5898305084745763,
"grad_norm": 0.686356246471405,
"learning_rate": 1.7231227890640893e-05,
"loss": 1.3903,
"step": 957
},
{
"epoch": 0.5904468412942989,
"grad_norm": 0.6897464394569397,
"learning_rate": 1.7188584002145612e-05,
"loss": 1.4586,
"step": 958
},
{
"epoch": 0.5910631741140215,
"grad_norm": 0.6666277050971985,
"learning_rate": 1.7145953148742478e-05,
"loss": 1.4593,
"step": 959
},
{
"epoch": 0.5916795069337443,
"grad_norm": 0.6714256405830383,
"learning_rate": 1.7103335528088866e-05,
"loss": 1.3904,
"step": 960
},
{
"epoch": 0.5922958397534669,
"grad_norm": 0.6808894872665405,
"learning_rate": 1.7060731337780794e-05,
"loss": 1.4326,
"step": 961
},
{
"epoch": 0.5929121725731895,
"grad_norm": 0.6975859999656677,
"learning_rate": 1.7018140775352007e-05,
"loss": 1.4409,
"step": 962
},
{
"epoch": 0.5935285053929121,
"grad_norm": 0.6830939054489136,
"learning_rate": 1.6975564038273084e-05,
"loss": 1.3683,
"step": 963
},
{
"epoch": 0.5941448382126349,
"grad_norm": 0.6693740487098694,
"learning_rate": 1.6933001323950476e-05,
"loss": 1.3316,
"step": 964
},
{
"epoch": 0.5947611710323575,
"grad_norm": 0.6849698424339294,
"learning_rate": 1.689045282972563e-05,
"loss": 1.3855,
"step": 965
},
{
"epoch": 0.5953775038520801,
"grad_norm": 0.7089459300041199,
"learning_rate": 1.6847918752874082e-05,
"loss": 1.4248,
"step": 966
},
{
"epoch": 0.5959938366718027,
"grad_norm": 0.6675121784210205,
"learning_rate": 1.680539929060448e-05,
"loss": 1.4441,
"step": 967
},
{
"epoch": 0.5966101694915255,
"grad_norm": 0.7075740098953247,
"learning_rate": 1.6762894640057735e-05,
"loss": 1.4399,
"step": 968
},
{
"epoch": 0.5972265023112481,
"grad_norm": 0.6729096174240112,
"learning_rate": 1.6720404998306094e-05,
"loss": 1.4364,
"step": 969
},
{
"epoch": 0.5978428351309707,
"grad_norm": 0.6988831162452698,
"learning_rate": 1.6677930562352208e-05,
"loss": 1.4147,
"step": 970
},
{
"epoch": 0.5984591679506934,
"grad_norm": 0.6925584673881531,
"learning_rate": 1.6635471529128215e-05,
"loss": 1.4554,
"step": 971
},
{
"epoch": 0.599075500770416,
"grad_norm": 0.6823211312294006,
"learning_rate": 1.6593028095494842e-05,
"loss": 1.4105,
"step": 972
},
{
"epoch": 0.5996918335901387,
"grad_norm": 0.6767581105232239,
"learning_rate": 1.655060045824051e-05,
"loss": 1.4272,
"step": 973
},
{
"epoch": 0.6003081664098613,
"grad_norm": 0.6897989511489868,
"learning_rate": 1.6508188814080367e-05,
"loss": 1.4132,
"step": 974
},
{
"epoch": 0.600924499229584,
"grad_norm": 0.6938064694404602,
"learning_rate": 1.6465793359655433e-05,
"loss": 1.4464,
"step": 975
},
{
"epoch": 0.6015408320493066,
"grad_norm": 0.6732364892959595,
"learning_rate": 1.6423414291531655e-05,
"loss": 1.4384,
"step": 976
},
{
"epoch": 0.6021571648690293,
"grad_norm": 0.6800574660301208,
"learning_rate": 1.638105180619901e-05,
"loss": 1.4428,
"step": 977
},
{
"epoch": 0.6027734976887519,
"grad_norm": 0.6989326477050781,
"learning_rate": 1.6338706100070586e-05,
"loss": 1.4086,
"step": 978
},
{
"epoch": 0.6033898305084746,
"grad_norm": 0.6868564486503601,
"learning_rate": 1.629637736948168e-05,
"loss": 1.4247,
"step": 979
},
{
"epoch": 0.6040061633281972,
"grad_norm": 0.6902734041213989,
"learning_rate": 1.6254065810688866e-05,
"loss": 1.3884,
"step": 980
},
{
"epoch": 0.6046224961479199,
"grad_norm": 0.6859354972839355,
"learning_rate": 1.6211771619869123e-05,
"loss": 1.3885,
"step": 981
},
{
"epoch": 0.6052388289676425,
"grad_norm": 0.6829075813293457,
"learning_rate": 1.616949499311888e-05,
"loss": 1.4694,
"step": 982
},
{
"epoch": 0.6058551617873652,
"grad_norm": 0.6718934774398804,
"learning_rate": 1.6127236126453168e-05,
"loss": 1.4215,
"step": 983
},
{
"epoch": 0.6064714946070878,
"grad_norm": 0.6713712215423584,
"learning_rate": 1.608499521580462e-05,
"loss": 1.4091,
"step": 984
},
{
"epoch": 0.6070878274268104,
"grad_norm": 0.6932412385940552,
"learning_rate": 1.6042772457022666e-05,
"loss": 1.4868,
"step": 985
},
{
"epoch": 0.6077041602465332,
"grad_norm": 0.7100671529769897,
"learning_rate": 1.6000568045872556e-05,
"loss": 1.4582,
"step": 986
},
{
"epoch": 0.6083204930662558,
"grad_norm": 0.6786303520202637,
"learning_rate": 1.5958382178034447e-05,
"loss": 1.4771,
"step": 987
},
{
"epoch": 0.6089368258859784,
"grad_norm": 0.6749539375305176,
"learning_rate": 1.5916215049102564e-05,
"loss": 1.3796,
"step": 988
},
{
"epoch": 0.609553158705701,
"grad_norm": 0.6852856874465942,
"learning_rate": 1.5874066854584224e-05,
"loss": 1.3981,
"step": 989
},
{
"epoch": 0.6101694915254238,
"grad_norm": 0.6997564435005188,
"learning_rate": 1.5831937789898957e-05,
"loss": 1.4493,
"step": 990
},
{
"epoch": 0.6107858243451464,
"grad_norm": 0.6845008134841919,
"learning_rate": 1.57898280503776e-05,
"loss": 1.3884,
"step": 991
},
{
"epoch": 0.611402157164869,
"grad_norm": 0.7117551565170288,
"learning_rate": 1.5747737831261393e-05,
"loss": 1.453,
"step": 992
},
{
"epoch": 0.6120184899845916,
"grad_norm": 0.6957306861877441,
"learning_rate": 1.5705667327701056e-05,
"loss": 1.4064,
"step": 993
},
{
"epoch": 0.6126348228043144,
"grad_norm": 0.6860647797584534,
"learning_rate": 1.566361673475592e-05,
"loss": 1.3506,
"step": 994
},
{
"epoch": 0.613251155624037,
"grad_norm": 0.6879005432128906,
"learning_rate": 1.562158624739298e-05,
"loss": 1.4843,
"step": 995
},
{
"epoch": 0.6138674884437596,
"grad_norm": 0.7000822424888611,
"learning_rate": 1.557957606048603e-05,
"loss": 1.407,
"step": 996
},
{
"epoch": 0.6144838212634823,
"grad_norm": 0.698491632938385,
"learning_rate": 1.5537586368814727e-05,
"loss": 1.3917,
"step": 997
},
{
"epoch": 0.615100154083205,
"grad_norm": 0.7287001609802246,
"learning_rate": 1.54956173670637e-05,
"loss": 1.4017,
"step": 998
},
{
"epoch": 0.6157164869029276,
"grad_norm": 0.7118046879768372,
"learning_rate": 1.545366924982167e-05,
"loss": 1.402,
"step": 999
},
{
"epoch": 0.6163328197226502,
"grad_norm": 0.6940359473228455,
"learning_rate": 1.5411742211580504e-05,
"loss": 1.3956,
"step": 1000
},
{
"epoch": 0.6169491525423729,
"grad_norm": 0.6867558360099792,
"learning_rate": 1.5369836446734343e-05,
"loss": 1.4436,
"step": 1001
},
{
"epoch": 0.6175654853620955,
"grad_norm": 0.6873385310173035,
"learning_rate": 1.532795214957872e-05,
"loss": 1.4537,
"step": 1002
},
{
"epoch": 0.6181818181818182,
"grad_norm": 0.6997383832931519,
"learning_rate": 1.528608951430958e-05,
"loss": 1.463,
"step": 1003
},
{
"epoch": 0.6187981510015408,
"grad_norm": 0.687658965587616,
"learning_rate": 1.5244248735022473e-05,
"loss": 1.4255,
"step": 1004
},
{
"epoch": 0.6194144838212635,
"grad_norm": 0.6884608268737793,
"learning_rate": 1.5202430005711625e-05,
"loss": 1.4706,
"step": 1005
},
{
"epoch": 0.6200308166409861,
"grad_norm": 0.7008056640625,
"learning_rate": 1.5160633520268978e-05,
"loss": 1.5052,
"step": 1006
},
{
"epoch": 0.6206471494607088,
"grad_norm": 0.7026891708374023,
"learning_rate": 1.5118859472483389e-05,
"loss": 1.441,
"step": 1007
},
{
"epoch": 0.6212634822804314,
"grad_norm": 0.6971073150634766,
"learning_rate": 1.507710805603966e-05,
"loss": 1.3532,
"step": 1008
},
{
"epoch": 0.6218798151001541,
"grad_norm": 0.7128373384475708,
"learning_rate": 1.5035379464517667e-05,
"loss": 1.4875,
"step": 1009
},
{
"epoch": 0.6224961479198767,
"grad_norm": 0.6831132173538208,
"learning_rate": 1.4993673891391456e-05,
"loss": 1.4559,
"step": 1010
},
{
"epoch": 0.6231124807395994,
"grad_norm": 0.6850417852401733,
"learning_rate": 1.4951991530028352e-05,
"loss": 1.3441,
"step": 1011
},
{
"epoch": 0.6237288135593221,
"grad_norm": 0.6863908767700195,
"learning_rate": 1.4910332573688062e-05,
"loss": 1.4899,
"step": 1012
},
{
"epoch": 0.6243451463790447,
"grad_norm": 0.6850789785385132,
"learning_rate": 1.4868697215521764e-05,
"loss": 1.4014,
"step": 1013
},
{
"epoch": 0.6249614791987673,
"grad_norm": 0.7038957476615906,
"learning_rate": 1.482708564857124e-05,
"loss": 1.486,
"step": 1014
},
{
"epoch": 0.6255778120184899,
"grad_norm": 0.6841074824333191,
"learning_rate": 1.4785498065767953e-05,
"loss": 1.3738,
"step": 1015
},
{
"epoch": 0.6261941448382127,
"grad_norm": 0.7064815759658813,
"learning_rate": 1.4743934659932162e-05,
"loss": 1.5064,
"step": 1016
},
{
"epoch": 0.6268104776579353,
"grad_norm": 0.6969325542449951,
"learning_rate": 1.4702395623772042e-05,
"loss": 1.4584,
"step": 1017
},
{
"epoch": 0.6274268104776579,
"grad_norm": 0.681100606918335,
"learning_rate": 1.4660881149882768e-05,
"loss": 1.4532,
"step": 1018
},
{
"epoch": 0.6280431432973805,
"grad_norm": 0.6736499071121216,
"learning_rate": 1.461939143074563e-05,
"loss": 1.4733,
"step": 1019
},
{
"epoch": 0.6286594761171033,
"grad_norm": 0.7187907695770264,
"learning_rate": 1.4577926658727155e-05,
"loss": 1.4732,
"step": 1020
},
{
"epoch": 0.6292758089368259,
"grad_norm": 0.6810031533241272,
"learning_rate": 1.4536487026078188e-05,
"loss": 1.4505,
"step": 1021
},
{
"epoch": 0.6298921417565485,
"grad_norm": 0.7185724973678589,
"learning_rate": 1.449507272493304e-05,
"loss": 1.41,
"step": 1022
},
{
"epoch": 0.6305084745762712,
"grad_norm": 0.6889476180076599,
"learning_rate": 1.4453683947308538e-05,
"loss": 1.3529,
"step": 1023
},
{
"epoch": 0.6311248073959939,
"grad_norm": 0.7004572749137878,
"learning_rate": 1.44123208851032e-05,
"loss": 1.4009,
"step": 1024
},
{
"epoch": 0.6317411402157165,
"grad_norm": 0.7015326619148254,
"learning_rate": 1.4370983730096307e-05,
"loss": 1.4323,
"step": 1025
},
{
"epoch": 0.6323574730354391,
"grad_norm": 0.689512312412262,
"learning_rate": 1.4329672673946995e-05,
"loss": 1.3846,
"step": 1026
},
{
"epoch": 0.6329738058551618,
"grad_norm": 0.7007417678833008,
"learning_rate": 1.4288387908193434e-05,
"loss": 1.4277,
"step": 1027
},
{
"epoch": 0.6335901386748845,
"grad_norm": 0.701409101486206,
"learning_rate": 1.4247129624251883e-05,
"loss": 1.4366,
"step": 1028
},
{
"epoch": 0.6342064714946071,
"grad_norm": 0.6901214122772217,
"learning_rate": 1.4205898013415804e-05,
"loss": 1.4107,
"step": 1029
},
{
"epoch": 0.6348228043143297,
"grad_norm": 0.6846891641616821,
"learning_rate": 1.416469326685501e-05,
"loss": 1.4469,
"step": 1030
},
{
"epoch": 0.6354391371340524,
"grad_norm": 0.6826861500740051,
"learning_rate": 1.412351557561475e-05,
"loss": 1.3618,
"step": 1031
},
{
"epoch": 0.636055469953775,
"grad_norm": 0.712218701839447,
"learning_rate": 1.4082365130614824e-05,
"loss": 1.4264,
"step": 1032
},
{
"epoch": 0.6366718027734977,
"grad_norm": 0.6926703453063965,
"learning_rate": 1.4041242122648714e-05,
"loss": 1.3978,
"step": 1033
},
{
"epoch": 0.6372881355932203,
"grad_norm": 0.6992529034614563,
"learning_rate": 1.4000146742382692e-05,
"loss": 1.4868,
"step": 1034
},
{
"epoch": 0.637904468412943,
"grad_norm": 0.6841732263565063,
"learning_rate": 1.3959079180354933e-05,
"loss": 1.3619,
"step": 1035
},
{
"epoch": 0.6385208012326656,
"grad_norm": 0.6914638876914978,
"learning_rate": 1.3918039626974626e-05,
"loss": 1.4077,
"step": 1036
},
{
"epoch": 0.6391371340523883,
"grad_norm": 0.6879279613494873,
"learning_rate": 1.3877028272521102e-05,
"loss": 1.3889,
"step": 1037
},
{
"epoch": 0.639753466872111,
"grad_norm": 0.7122974395751953,
"learning_rate": 1.3836045307142967e-05,
"loss": 1.4497,
"step": 1038
},
{
"epoch": 0.6403697996918336,
"grad_norm": 0.6926254630088806,
"learning_rate": 1.3795090920857161e-05,
"loss": 1.3819,
"step": 1039
},
{
"epoch": 0.6409861325115562,
"grad_norm": 0.6959458589553833,
"learning_rate": 1.3754165303548148e-05,
"loss": 1.4077,
"step": 1040
},
{
"epoch": 0.6416024653312788,
"grad_norm": 0.6710784435272217,
"learning_rate": 1.3713268644967012e-05,
"loss": 1.3921,
"step": 1041
},
{
"epoch": 0.6422187981510016,
"grad_norm": 0.6844033002853394,
"learning_rate": 1.367240113473053e-05,
"loss": 1.4421,
"step": 1042
},
{
"epoch": 0.6428351309707242,
"grad_norm": 0.6891530752182007,
"learning_rate": 1.3631562962320382e-05,
"loss": 1.3631,
"step": 1043
},
{
"epoch": 0.6434514637904468,
"grad_norm": 0.6909316182136536,
"learning_rate": 1.3590754317082196e-05,
"loss": 1.3655,
"step": 1044
},
{
"epoch": 0.6440677966101694,
"grad_norm": 0.7180591225624084,
"learning_rate": 1.3549975388224686e-05,
"loss": 1.4067,
"step": 1045
},
{
"epoch": 0.6446841294298922,
"grad_norm": 0.7052772641181946,
"learning_rate": 1.3509226364818817e-05,
"loss": 1.3907,
"step": 1046
},
{
"epoch": 0.6453004622496148,
"grad_norm": 0.7179615497589111,
"learning_rate": 1.3468507435796885e-05,
"loss": 1.3709,
"step": 1047
},
{
"epoch": 0.6459167950693374,
"grad_norm": 0.6939482092857361,
"learning_rate": 1.3427818789951647e-05,
"loss": 1.3833,
"step": 1048
},
{
"epoch": 0.6465331278890601,
"grad_norm": 0.6862821578979492,
"learning_rate": 1.338716061593546e-05,
"loss": 1.3881,
"step": 1049
},
{
"epoch": 0.6471494607087828,
"grad_norm": 0.6936004161834717,
"learning_rate": 1.3346533102259391e-05,
"loss": 1.3849,
"step": 1050
},
{
"epoch": 0.6477657935285054,
"grad_norm": 0.698454737663269,
"learning_rate": 1.3305936437292366e-05,
"loss": 1.3328,
"step": 1051
},
{
"epoch": 0.648382126348228,
"grad_norm": 0.6913847327232361,
"learning_rate": 1.3265370809260265e-05,
"loss": 1.4123,
"step": 1052
},
{
"epoch": 0.6489984591679507,
"grad_norm": 0.7037560343742371,
"learning_rate": 1.322483640624507e-05,
"loss": 1.41,
"step": 1053
},
{
"epoch": 0.6496147919876734,
"grad_norm": 0.7199169397354126,
"learning_rate": 1.3184333416183996e-05,
"loss": 1.3898,
"step": 1054
},
{
"epoch": 0.650231124807396,
"grad_norm": 0.7095668315887451,
"learning_rate": 1.3143862026868604e-05,
"loss": 1.431,
"step": 1055
},
{
"epoch": 0.6508474576271186,
"grad_norm": 0.691674530506134,
"learning_rate": 1.3103422425943942e-05,
"loss": 1.4326,
"step": 1056
},
{
"epoch": 0.6514637904468413,
"grad_norm": 0.6917114853858948,
"learning_rate": 1.3063014800907677e-05,
"loss": 1.4941,
"step": 1057
},
{
"epoch": 0.652080123266564,
"grad_norm": 0.7097762227058411,
"learning_rate": 1.3022639339109201e-05,
"loss": 1.4593,
"step": 1058
},
{
"epoch": 0.6526964560862866,
"grad_norm": 0.6939966082572937,
"learning_rate": 1.2982296227748804e-05,
"loss": 1.4865,
"step": 1059
},
{
"epoch": 0.6533127889060092,
"grad_norm": 0.6821755170822144,
"learning_rate": 1.2941985653876778e-05,
"loss": 1.4389,
"step": 1060
},
{
"epoch": 0.6539291217257319,
"grad_norm": 0.7058581709861755,
"learning_rate": 1.2901707804392556e-05,
"loss": 1.4054,
"step": 1061
},
{
"epoch": 0.6545454545454545,
"grad_norm": 0.6868889927864075,
"learning_rate": 1.286146286604382e-05,
"loss": 1.4012,
"step": 1062
},
{
"epoch": 0.6551617873651772,
"grad_norm": 0.6890780925750732,
"learning_rate": 1.2821251025425698e-05,
"loss": 1.4019,
"step": 1063
},
{
"epoch": 0.6557781201848999,
"grad_norm": 0.7047430872917175,
"learning_rate": 1.2781072468979848e-05,
"loss": 1.4217,
"step": 1064
},
{
"epoch": 0.6563944530046225,
"grad_norm": 0.6973077654838562,
"learning_rate": 1.2740927382993586e-05,
"loss": 1.4174,
"step": 1065
},
{
"epoch": 0.6570107858243451,
"grad_norm": 0.696974515914917,
"learning_rate": 1.2700815953599067e-05,
"loss": 1.4529,
"step": 1066
},
{
"epoch": 0.6576271186440678,
"grad_norm": 0.6864145398139954,
"learning_rate": 1.2660738366772391e-05,
"loss": 1.4023,
"step": 1067
},
{
"epoch": 0.6582434514637905,
"grad_norm": 0.6852766871452332,
"learning_rate": 1.2620694808332743e-05,
"loss": 1.3387,
"step": 1068
},
{
"epoch": 0.6588597842835131,
"grad_norm": 0.7034934163093567,
"learning_rate": 1.2580685463941538e-05,
"loss": 1.3917,
"step": 1069
},
{
"epoch": 0.6594761171032357,
"grad_norm": 0.7008872628211975,
"learning_rate": 1.2540710519101564e-05,
"loss": 1.4584,
"step": 1070
},
{
"epoch": 0.6600924499229583,
"grad_norm": 0.6999143362045288,
"learning_rate": 1.2500770159156104e-05,
"loss": 1.4401,
"step": 1071
},
{
"epoch": 0.6607087827426811,
"grad_norm": 0.6936800479888916,
"learning_rate": 1.2460864569288095e-05,
"loss": 1.3586,
"step": 1072
},
{
"epoch": 0.6613251155624037,
"grad_norm": 0.6965266466140747,
"learning_rate": 1.2420993934519268e-05,
"loss": 1.3917,
"step": 1073
},
{
"epoch": 0.6619414483821263,
"grad_norm": 0.7005475163459778,
"learning_rate": 1.2381158439709285e-05,
"loss": 1.4211,
"step": 1074
},
{
"epoch": 0.662557781201849,
"grad_norm": 0.6943978071212769,
"learning_rate": 1.2341358269554866e-05,
"loss": 1.3793,
"step": 1075
},
{
"epoch": 0.6631741140215717,
"grad_norm": 0.6871793866157532,
"learning_rate": 1.2301593608588958e-05,
"loss": 1.3604,
"step": 1076
},
{
"epoch": 0.6637904468412943,
"grad_norm": 0.7159802913665771,
"learning_rate": 1.2261864641179894e-05,
"loss": 1.4797,
"step": 1077
},
{
"epoch": 0.6644067796610169,
"grad_norm": 0.7132952809333801,
"learning_rate": 1.2222171551530466e-05,
"loss": 1.423,
"step": 1078
},
{
"epoch": 0.6650231124807396,
"grad_norm": 0.7126781344413757,
"learning_rate": 1.2182514523677163e-05,
"loss": 1.3755,
"step": 1079
},
{
"epoch": 0.6656394453004623,
"grad_norm": 0.7203714847564697,
"learning_rate": 1.2142893741489257e-05,
"loss": 1.4024,
"step": 1080
},
{
"epoch": 0.6662557781201849,
"grad_norm": 0.7033785581588745,
"learning_rate": 1.210330938866795e-05,
"loss": 1.408,
"step": 1081
},
{
"epoch": 0.6668721109399075,
"grad_norm": 0.6952004432678223,
"learning_rate": 1.2063761648745573e-05,
"loss": 1.4091,
"step": 1082
},
{
"epoch": 0.6674884437596302,
"grad_norm": 0.6984594464302063,
"learning_rate": 1.2024250705084683e-05,
"loss": 1.4586,
"step": 1083
},
{
"epoch": 0.6681047765793529,
"grad_norm": 0.6936200857162476,
"learning_rate": 1.1984776740877239e-05,
"loss": 1.3687,
"step": 1084
},
{
"epoch": 0.6687211093990755,
"grad_norm": 0.6909807324409485,
"learning_rate": 1.194533993914373e-05,
"loss": 1.3636,
"step": 1085
},
{
"epoch": 0.6693374422187982,
"grad_norm": 0.6899669766426086,
"learning_rate": 1.1905940482732357e-05,
"loss": 1.4395,
"step": 1086
},
{
"epoch": 0.6699537750385208,
"grad_norm": 0.6890370845794678,
"learning_rate": 1.1866578554318168e-05,
"loss": 1.396,
"step": 1087
},
{
"epoch": 0.6705701078582434,
"grad_norm": 0.7054473161697388,
"learning_rate": 1.1827254336402205e-05,
"loss": 1.4086,
"step": 1088
},
{
"epoch": 0.6711864406779661,
"grad_norm": 0.679929256439209,
"learning_rate": 1.1787968011310675e-05,
"loss": 1.4122,
"step": 1089
},
{
"epoch": 0.6718027734976888,
"grad_norm": 0.7349550127983093,
"learning_rate": 1.1748719761194093e-05,
"loss": 1.5266,
"step": 1090
},
{
"epoch": 0.6724191063174114,
"grad_norm": 0.7072716951370239,
"learning_rate": 1.1709509768026432e-05,
"loss": 1.3337,
"step": 1091
},
{
"epoch": 0.673035439137134,
"grad_norm": 0.6961799263954163,
"learning_rate": 1.1670338213604299e-05,
"loss": 1.4086,
"step": 1092
},
{
"epoch": 0.6736517719568567,
"grad_norm": 0.705967903137207,
"learning_rate": 1.1631205279546079e-05,
"loss": 1.4231,
"step": 1093
},
{
"epoch": 0.6742681047765794,
"grad_norm": 0.6936699748039246,
"learning_rate": 1.1592111147291071e-05,
"loss": 1.3823,
"step": 1094
},
{
"epoch": 0.674884437596302,
"grad_norm": 0.6913039684295654,
"learning_rate": 1.1553055998098715e-05,
"loss": 1.381,
"step": 1095
},
{
"epoch": 0.6755007704160246,
"grad_norm": 0.6969724893569946,
"learning_rate": 1.1514040013047656e-05,
"loss": 1.3794,
"step": 1096
},
{
"epoch": 0.6761171032357473,
"grad_norm": 0.7696253657341003,
"learning_rate": 1.1475063373035002e-05,
"loss": 1.44,
"step": 1097
},
{
"epoch": 0.67673343605547,
"grad_norm": 0.7233097553253174,
"learning_rate": 1.1436126258775406e-05,
"loss": 1.3911,
"step": 1098
},
{
"epoch": 0.6773497688751926,
"grad_norm": 0.7103934288024902,
"learning_rate": 1.1397228850800262e-05,
"loss": 1.4205,
"step": 1099
},
{
"epoch": 0.6779661016949152,
"grad_norm": 0.6893105506896973,
"learning_rate": 1.1358371329456887e-05,
"loss": 1.37,
"step": 1100
}
],
"logging_steps": 1.0,
"max_steps": 1622,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.4145732702162125e+19,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}