DataPRM-ScienceAgentBench / trainer_state.json
consultantQ's picture
Upload folder using huggingface_hub
bea6243 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 279,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.010752688172043012,
"grad_norm": 19.659445278083677,
"learning_rate": 7.142857142857143e-07,
"loss": 0.8801538944244385,
"step": 1
},
{
"epoch": 0.021505376344086023,
"grad_norm": 21.183622990197563,
"learning_rate": 1.4285714285714286e-06,
"loss": 0.9106118679046631,
"step": 2
},
{
"epoch": 0.03225806451612903,
"grad_norm": 9.680562864297194,
"learning_rate": 2.1428571428571427e-06,
"loss": 0.89518141746521,
"step": 3
},
{
"epoch": 0.043010752688172046,
"grad_norm": 8.463557758001873,
"learning_rate": 2.8571428571428573e-06,
"loss": 0.8813375234603882,
"step": 4
},
{
"epoch": 0.053763440860215055,
"grad_norm": 8.000585049846563,
"learning_rate": 3.5714285714285718e-06,
"loss": 0.8478757739067078,
"step": 5
},
{
"epoch": 0.06451612903225806,
"grad_norm": 6.392760393586216,
"learning_rate": 4.2857142857142855e-06,
"loss": 0.7342901825904846,
"step": 6
},
{
"epoch": 0.07526881720430108,
"grad_norm": 4.291231505071156,
"learning_rate": 5e-06,
"loss": 0.6894474029541016,
"step": 7
},
{
"epoch": 0.08602150537634409,
"grad_norm": 3.3545639671049603,
"learning_rate": 5.7142857142857145e-06,
"loss": 0.6144750118255615,
"step": 8
},
{
"epoch": 0.0967741935483871,
"grad_norm": 1.3577798099653526,
"learning_rate": 6.4285714285714295e-06,
"loss": 0.5370485186576843,
"step": 9
},
{
"epoch": 0.10752688172043011,
"grad_norm": 1.2774532761940443,
"learning_rate": 7.1428571428571436e-06,
"loss": 0.5807953476905823,
"step": 10
},
{
"epoch": 0.11827956989247312,
"grad_norm": 1.0447472651159857,
"learning_rate": 7.857142857142858e-06,
"loss": 0.5280188918113708,
"step": 11
},
{
"epoch": 0.12903225806451613,
"grad_norm": 1.3301070676426008,
"learning_rate": 8.571428571428571e-06,
"loss": 0.5441749691963196,
"step": 12
},
{
"epoch": 0.13978494623655913,
"grad_norm": 1.0565188353675214,
"learning_rate": 9.285714285714288e-06,
"loss": 0.5112073421478271,
"step": 13
},
{
"epoch": 0.15053763440860216,
"grad_norm": 1.0753192556611026,
"learning_rate": 1e-05,
"loss": 0.5184136629104614,
"step": 14
},
{
"epoch": 0.16129032258064516,
"grad_norm": 0.8209457367101312,
"learning_rate": 9.999648647603774e-06,
"loss": 0.5023149251937866,
"step": 15
},
{
"epoch": 0.17204301075268819,
"grad_norm": 0.7282985426170734,
"learning_rate": 9.998594639794502e-06,
"loss": 0.48974722623825073,
"step": 16
},
{
"epoch": 0.1827956989247312,
"grad_norm": 0.7177574512039909,
"learning_rate": 9.996838124703448e-06,
"loss": 0.47443556785583496,
"step": 17
},
{
"epoch": 0.1935483870967742,
"grad_norm": 0.7867041042254088,
"learning_rate": 9.994379349192927e-06,
"loss": 0.48248669505119324,
"step": 18
},
{
"epoch": 0.20430107526881722,
"grad_norm": 0.7192865433337082,
"learning_rate": 9.991218658821609e-06,
"loss": 0.46316730976104736,
"step": 19
},
{
"epoch": 0.21505376344086022,
"grad_norm": 0.8279557032400371,
"learning_rate": 9.987356497795944e-06,
"loss": 0.5312884449958801,
"step": 20
},
{
"epoch": 0.22580645161290322,
"grad_norm": 0.7907337063911983,
"learning_rate": 9.982793408907747e-06,
"loss": 0.48629915714263916,
"step": 21
},
{
"epoch": 0.23655913978494625,
"grad_norm": 0.634234358721607,
"learning_rate": 9.977530033457906e-06,
"loss": 0.47711604833602905,
"step": 22
},
{
"epoch": 0.24731182795698925,
"grad_norm": 0.7748372497671804,
"learning_rate": 9.971567111166246e-06,
"loss": 0.4741020202636719,
"step": 23
},
{
"epoch": 0.25806451612903225,
"grad_norm": 0.6968307669517513,
"learning_rate": 9.964905480067585e-06,
"loss": 0.48086652159690857,
"step": 24
},
{
"epoch": 0.26881720430107525,
"grad_norm": 0.6205197336090342,
"learning_rate": 9.957546076393944e-06,
"loss": 0.45759013295173645,
"step": 25
},
{
"epoch": 0.27956989247311825,
"grad_norm": 0.5942106287222008,
"learning_rate": 9.949489934442966e-06,
"loss": 0.454744428396225,
"step": 26
},
{
"epoch": 0.2903225806451613,
"grad_norm": 0.6374339549026193,
"learning_rate": 9.940738186432565e-06,
"loss": 0.4751560091972351,
"step": 27
},
{
"epoch": 0.3010752688172043,
"grad_norm": 0.5187971760627564,
"learning_rate": 9.931292062341793e-06,
"loss": 0.39658498764038086,
"step": 28
},
{
"epoch": 0.3118279569892473,
"grad_norm": 0.5360240389517962,
"learning_rate": 9.921152889737985e-06,
"loss": 0.46163371205329895,
"step": 29
},
{
"epoch": 0.3225806451612903,
"grad_norm": 0.5585203893563241,
"learning_rate": 9.910322093590177e-06,
"loss": 0.46062612533569336,
"step": 30
},
{
"epoch": 0.3333333333333333,
"grad_norm": 0.5610561072744311,
"learning_rate": 9.898801196068839e-06,
"loss": 0.4484734833240509,
"step": 31
},
{
"epoch": 0.34408602150537637,
"grad_norm": 0.5533641378752504,
"learning_rate": 9.886591816331953e-06,
"loss": 0.41486942768096924,
"step": 32
},
{
"epoch": 0.3548387096774194,
"grad_norm": 0.5365723882895206,
"learning_rate": 9.87369567029745e-06,
"loss": 0.43502286076545715,
"step": 33
},
{
"epoch": 0.3655913978494624,
"grad_norm": 0.5536007189636761,
"learning_rate": 9.860114570402055e-06,
"loss": 0.4295029640197754,
"step": 34
},
{
"epoch": 0.3763440860215054,
"grad_norm": 0.537839879636324,
"learning_rate": 9.845850425346563e-06,
"loss": 0.45029592514038086,
"step": 35
},
{
"epoch": 0.3870967741935484,
"grad_norm": 0.5299693550104672,
"learning_rate": 9.830905239827592e-06,
"loss": 0.38297146558761597,
"step": 36
},
{
"epoch": 0.3978494623655914,
"grad_norm": 0.5451092272235792,
"learning_rate": 9.815281114255841e-06,
"loss": 0.4300140142440796,
"step": 37
},
{
"epoch": 0.40860215053763443,
"grad_norm": 0.6192689983212899,
"learning_rate": 9.798980244460892e-06,
"loss": 0.47214269638061523,
"step": 38
},
{
"epoch": 0.41935483870967744,
"grad_norm": 0.5870329353069306,
"learning_rate": 9.782004921382612e-06,
"loss": 0.43783068656921387,
"step": 39
},
{
"epoch": 0.43010752688172044,
"grad_norm": 0.5903637860392952,
"learning_rate": 9.764357530749178e-06,
"loss": 0.4084509611129761,
"step": 40
},
{
"epoch": 0.44086021505376344,
"grad_norm": 0.61705449873411,
"learning_rate": 9.74604055274178e-06,
"loss": 0.42812418937683105,
"step": 41
},
{
"epoch": 0.45161290322580644,
"grad_norm": 1.156325694795339,
"learning_rate": 9.727056561646067e-06,
"loss": 0.4196915626525879,
"step": 42
},
{
"epoch": 0.46236559139784944,
"grad_norm": 0.580641483078281,
"learning_rate": 9.707408225490343e-06,
"loss": 0.4251587986946106,
"step": 43
},
{
"epoch": 0.4731182795698925,
"grad_norm": 0.5510691280763641,
"learning_rate": 9.687098305670606e-06,
"loss": 0.4205142855644226,
"step": 44
},
{
"epoch": 0.4838709677419355,
"grad_norm": 0.5225683335691166,
"learning_rate": 9.66612965656245e-06,
"loss": 0.37856176495552063,
"step": 45
},
{
"epoch": 0.4946236559139785,
"grad_norm": 0.5740926054795668,
"learning_rate": 9.644505225119922e-06,
"loss": 0.4447917342185974,
"step": 46
},
{
"epoch": 0.5053763440860215,
"grad_norm": 0.5873250280785997,
"learning_rate": 9.622228050461345e-06,
"loss": 0.3922080397605896,
"step": 47
},
{
"epoch": 0.5161290322580645,
"grad_norm": 0.6219913849412524,
"learning_rate": 9.599301263442194e-06,
"loss": 0.43804931640625,
"step": 48
},
{
"epoch": 0.5268817204301075,
"grad_norm": 0.5613998719760911,
"learning_rate": 9.575728086215093e-06,
"loss": 0.39639532566070557,
"step": 49
},
{
"epoch": 0.5376344086021505,
"grad_norm": 0.5892845271021349,
"learning_rate": 9.551511831776966e-06,
"loss": 0.4323637783527374,
"step": 50
},
{
"epoch": 0.5483870967741935,
"grad_norm": 0.6452327069861976,
"learning_rate": 9.526655903503423e-06,
"loss": 0.4111916422843933,
"step": 51
},
{
"epoch": 0.5591397849462365,
"grad_norm": 0.6139840452181067,
"learning_rate": 9.501163794670445e-06,
"loss": 0.41024255752563477,
"step": 52
},
{
"epoch": 0.5698924731182796,
"grad_norm": 0.5311229042653062,
"learning_rate": 9.475039087963443e-06,
"loss": 0.39047467708587646,
"step": 53
},
{
"epoch": 0.5806451612903226,
"grad_norm": 0.5686779350086011,
"learning_rate": 9.448285454973739e-06,
"loss": 0.4083505868911743,
"step": 54
},
{
"epoch": 0.5913978494623656,
"grad_norm": 0.5795028580373828,
"learning_rate": 9.420906655682553e-06,
"loss": 0.41430288553237915,
"step": 55
},
{
"epoch": 0.6021505376344086,
"grad_norm": 0.5426200748349074,
"learning_rate": 9.392906537932582e-06,
"loss": 0.39692986011505127,
"step": 56
},
{
"epoch": 0.6129032258064516,
"grad_norm": 0.5096666206067523,
"learning_rate": 9.364289036887214e-06,
"loss": 0.39456355571746826,
"step": 57
},
{
"epoch": 0.6236559139784946,
"grad_norm": 0.6214693543974321,
"learning_rate": 9.335058174477472e-06,
"loss": 0.43065640330314636,
"step": 58
},
{
"epoch": 0.6344086021505376,
"grad_norm": 0.6080775052566749,
"learning_rate": 9.305218058836778e-06,
"loss": 0.4233320653438568,
"step": 59
},
{
"epoch": 0.6451612903225806,
"grad_norm": 0.5891380183146641,
"learning_rate": 9.274772883723587e-06,
"loss": 0.44017481803894043,
"step": 60
},
{
"epoch": 0.6559139784946236,
"grad_norm": 0.5317932432917034,
"learning_rate": 9.24372692793199e-06,
"loss": 0.39426931738853455,
"step": 61
},
{
"epoch": 0.6666666666666666,
"grad_norm": 0.6464375491620328,
"learning_rate": 9.21208455469037e-06,
"loss": 0.4302670657634735,
"step": 62
},
{
"epoch": 0.6774193548387096,
"grad_norm": 0.6191245668003084,
"learning_rate": 9.179850211048193e-06,
"loss": 0.3996105492115021,
"step": 63
},
{
"epoch": 0.6881720430107527,
"grad_norm": 0.5209850545783036,
"learning_rate": 9.14702842725101e-06,
"loss": 0.3926236629486084,
"step": 64
},
{
"epoch": 0.6989247311827957,
"grad_norm": 0.5469602060829455,
"learning_rate": 9.113623816103775e-06,
"loss": 0.4129854142665863,
"step": 65
},
{
"epoch": 0.7096774193548387,
"grad_norm": 0.6436176685575749,
"learning_rate": 9.079641072322555e-06,
"loss": 0.4262722432613373,
"step": 66
},
{
"epoch": 0.7204301075268817,
"grad_norm": 0.6241121753870925,
"learning_rate": 9.045084971874738e-06,
"loss": 0.4259442687034607,
"step": 67
},
{
"epoch": 0.7311827956989247,
"grad_norm": 0.520706057326722,
"learning_rate": 9.009960371307798e-06,
"loss": 0.40881264209747314,
"step": 68
},
{
"epoch": 0.7419354838709677,
"grad_norm": 0.5406305590251429,
"learning_rate": 8.974272207066767e-06,
"loss": 0.4136165976524353,
"step": 69
},
{
"epoch": 0.7526881720430108,
"grad_norm": 0.5589669382622552,
"learning_rate": 8.938025494800454e-06,
"loss": 0.3692484498023987,
"step": 70
},
{
"epoch": 0.7634408602150538,
"grad_norm": 0.5484586018961858,
"learning_rate": 8.901225328656543e-06,
"loss": 0.392721563577652,
"step": 71
},
{
"epoch": 0.7741935483870968,
"grad_norm": 0.5720189126246105,
"learning_rate": 8.863876880565656e-06,
"loss": 0.4131432771682739,
"step": 72
},
{
"epoch": 0.7849462365591398,
"grad_norm": 0.5475650690180592,
"learning_rate": 8.825985399514488e-06,
"loss": 0.39399850368499756,
"step": 73
},
{
"epoch": 0.7956989247311828,
"grad_norm": 0.571446157132029,
"learning_rate": 8.787556210808101e-06,
"loss": 0.3765556216239929,
"step": 74
},
{
"epoch": 0.8064516129032258,
"grad_norm": 0.5096610649520538,
"learning_rate": 8.748594715321512e-06,
"loss": 0.38086819648742676,
"step": 75
},
{
"epoch": 0.8172043010752689,
"grad_norm": 0.5696584948279285,
"learning_rate": 8.70910638874064e-06,
"loss": 0.40993741154670715,
"step": 76
},
{
"epoch": 0.8279569892473119,
"grad_norm": 0.5448260281274895,
"learning_rate": 8.669096780792754e-06,
"loss": 0.3912901282310486,
"step": 77
},
{
"epoch": 0.8387096774193549,
"grad_norm": 0.584810442023009,
"learning_rate": 8.628571514466502e-06,
"loss": 0.39722931385040283,
"step": 78
},
{
"epoch": 0.8494623655913979,
"grad_norm": 0.5590387314291252,
"learning_rate": 8.587536285221656e-06,
"loss": 0.38008394837379456,
"step": 79
},
{
"epoch": 0.8602150537634409,
"grad_norm": 0.5199976478034525,
"learning_rate": 8.545996860188668e-06,
"loss": 0.3779188394546509,
"step": 80
},
{
"epoch": 0.8709677419354839,
"grad_norm": 0.6237910452530793,
"learning_rate": 8.503959077358143e-06,
"loss": 0.40831005573272705,
"step": 81
},
{
"epoch": 0.8817204301075269,
"grad_norm": 0.5103230716146216,
"learning_rate": 8.46142884476038e-06,
"loss": 0.3830574154853821,
"step": 82
},
{
"epoch": 0.8924731182795699,
"grad_norm": 0.6022470754794768,
"learning_rate": 8.418412139635026e-06,
"loss": 0.3760561943054199,
"step": 83
},
{
"epoch": 0.9032258064516129,
"grad_norm": 0.5676501864622416,
"learning_rate": 8.374915007591053e-06,
"loss": 0.3981160521507263,
"step": 84
},
{
"epoch": 0.9139784946236559,
"grad_norm": 0.62733814659435,
"learning_rate": 8.330943561757092e-06,
"loss": 0.38736072182655334,
"step": 85
},
{
"epoch": 0.9247311827956989,
"grad_norm": 0.581917413328386,
"learning_rate": 8.286503981922284e-06,
"loss": 0.40054696798324585,
"step": 86
},
{
"epoch": 0.9354838709677419,
"grad_norm": 0.5069721523225358,
"learning_rate": 8.241602513667775e-06,
"loss": 0.36416178941726685,
"step": 87
},
{
"epoch": 0.946236559139785,
"grad_norm": 0.6484709424945745,
"learning_rate": 8.19624546748895e-06,
"loss": 0.3703634738922119,
"step": 88
},
{
"epoch": 0.956989247311828,
"grad_norm": 0.5102251649420729,
"learning_rate": 8.150439217908557e-06,
"loss": 0.39143842458724976,
"step": 89
},
{
"epoch": 0.967741935483871,
"grad_norm": 0.5485517368687335,
"learning_rate": 8.104190202580811e-06,
"loss": 0.3922935128211975,
"step": 90
},
{
"epoch": 0.978494623655914,
"grad_norm": 0.5680193851368315,
"learning_rate": 8.057504921386661e-06,
"loss": 0.3872210383415222,
"step": 91
},
{
"epoch": 0.989247311827957,
"grad_norm": 0.5405961708214808,
"learning_rate": 8.010389935520269e-06,
"loss": 0.4088854193687439,
"step": 92
},
{
"epoch": 1.0,
"grad_norm": 0.5234958056904553,
"learning_rate": 7.962851866566912e-06,
"loss": 0.3855600953102112,
"step": 93
},
{
"epoch": 1.010752688172043,
"grad_norm": 0.5751268630788401,
"learning_rate": 7.914897395572362e-06,
"loss": 0.32572415471076965,
"step": 94
},
{
"epoch": 1.021505376344086,
"grad_norm": 0.5181295042077261,
"learning_rate": 7.866533262103937e-06,
"loss": 0.3046826720237732,
"step": 95
},
{
"epoch": 1.032258064516129,
"grad_norm": 0.5508917328926465,
"learning_rate": 7.817766263303312e-06,
"loss": 0.29789605736732483,
"step": 96
},
{
"epoch": 1.043010752688172,
"grad_norm": 0.5870437887252276,
"learning_rate": 7.768603252931243e-06,
"loss": 0.3265905976295471,
"step": 97
},
{
"epoch": 1.053763440860215,
"grad_norm": 0.5133318653975599,
"learning_rate": 7.719051140404327e-06,
"loss": 0.31667008996009827,
"step": 98
},
{
"epoch": 1.064516129032258,
"grad_norm": 0.5679147172610466,
"learning_rate": 7.669116889823955e-06,
"loss": 0.2714020311832428,
"step": 99
},
{
"epoch": 1.075268817204301,
"grad_norm": 0.6225396396665326,
"learning_rate": 7.6188075189975644e-06,
"loss": 0.3139406740665436,
"step": 100
},
{
"epoch": 1.086021505376344,
"grad_norm": 0.714553736071244,
"learning_rate": 7.568130098452352e-06,
"loss": 0.2961091995239258,
"step": 101
},
{
"epoch": 1.096774193548387,
"grad_norm": 0.523889168005371,
"learning_rate": 7.517091750441576e-06,
"loss": 0.27245283126831055,
"step": 102
},
{
"epoch": 1.10752688172043,
"grad_norm": 0.639295771779283,
"learning_rate": 7.465699647943586e-06,
"loss": 0.29422637820243835,
"step": 103
},
{
"epoch": 1.118279569892473,
"grad_norm": 0.6150692376811179,
"learning_rate": 7.413961013653725e-06,
"loss": 0.29594993591308594,
"step": 104
},
{
"epoch": 1.129032258064516,
"grad_norm": 0.6901790920161586,
"learning_rate": 7.361883118969248e-06,
"loss": 0.3163909912109375,
"step": 105
},
{
"epoch": 1.139784946236559,
"grad_norm": 0.5947317629259291,
"learning_rate": 7.309473282967387e-06,
"loss": 0.2941242456436157,
"step": 106
},
{
"epoch": 1.1505376344086022,
"grad_norm": 0.5873657820382887,
"learning_rate": 7.256738871376733e-06,
"loss": 0.29748812317848206,
"step": 107
},
{
"epoch": 1.1612903225806452,
"grad_norm": 0.5914901548215277,
"learning_rate": 7.203687295542032e-06,
"loss": 0.3174619972705841,
"step": 108
},
{
"epoch": 1.1720430107526882,
"grad_norm": 0.5757880846821776,
"learning_rate": 7.1503260113826035e-06,
"loss": 0.31319767236709595,
"step": 109
},
{
"epoch": 1.1827956989247312,
"grad_norm": 0.7019483863871168,
"learning_rate": 7.09666251834447e-06,
"loss": 0.3302695155143738,
"step": 110
},
{
"epoch": 1.1935483870967742,
"grad_norm": 0.5613831627968954,
"learning_rate": 7.042704358346375e-06,
"loss": 0.30153942108154297,
"step": 111
},
{
"epoch": 1.2043010752688172,
"grad_norm": 0.5906296533769152,
"learning_rate": 6.988459114719849e-06,
"loss": 0.3321286141872406,
"step": 112
},
{
"epoch": 1.2150537634408602,
"grad_norm": 0.6185930436417313,
"learning_rate": 6.933934411143419e-06,
"loss": 0.3183019161224365,
"step": 113
},
{
"epoch": 1.2258064516129032,
"grad_norm": 0.5223193968800454,
"learning_rate": 6.879137910571191e-06,
"loss": 0.28522804379463196,
"step": 114
},
{
"epoch": 1.2365591397849462,
"grad_norm": 0.5306097120719284,
"learning_rate": 6.824077314155877e-06,
"loss": 0.27248305082321167,
"step": 115
},
{
"epoch": 1.2473118279569892,
"grad_norm": 0.5410094711890291,
"learning_rate": 6.768760360166471e-06,
"loss": 0.29798364639282227,
"step": 116
},
{
"epoch": 1.2580645161290323,
"grad_norm": 0.594316663926518,
"learning_rate": 6.713194822900707e-06,
"loss": 0.27839338779449463,
"step": 117
},
{
"epoch": 1.2688172043010753,
"grad_norm": 0.6110653082771621,
"learning_rate": 6.657388511592453e-06,
"loss": 0.30802613496780396,
"step": 118
},
{
"epoch": 1.2795698924731183,
"grad_norm": 0.5260300979877027,
"learning_rate": 6.601349269314188e-06,
"loss": 0.29552769660949707,
"step": 119
},
{
"epoch": 1.2903225806451613,
"grad_norm": 0.541255561766811,
"learning_rate": 6.545084971874738e-06,
"loss": 0.28333544731140137,
"step": 120
},
{
"epoch": 1.3010752688172043,
"grad_norm": 0.5631658544435217,
"learning_rate": 6.488603526712391e-06,
"loss": 0.3072783946990967,
"step": 121
},
{
"epoch": 1.3118279569892473,
"grad_norm": 0.5999697642319023,
"learning_rate": 6.431912871783587e-06,
"loss": 0.2956047058105469,
"step": 122
},
{
"epoch": 1.3225806451612903,
"grad_norm": 0.5550402984265157,
"learning_rate": 6.3750209744473105e-06,
"loss": 0.3064526319503784,
"step": 123
},
{
"epoch": 1.3333333333333333,
"grad_norm": 0.6418624402271341,
"learning_rate": 6.3179358303453386e-06,
"loss": 0.30682218074798584,
"step": 124
},
{
"epoch": 1.3440860215053765,
"grad_norm": 0.5305043074037712,
"learning_rate": 6.260665462278544e-06,
"loss": 0.3036462068557739,
"step": 125
},
{
"epoch": 1.3548387096774195,
"grad_norm": 0.5732839993822031,
"learning_rate": 6.203217919079343e-06,
"loss": 0.32605141401290894,
"step": 126
},
{
"epoch": 1.3655913978494625,
"grad_norm": 0.5272514261375651,
"learning_rate": 6.145601274480521e-06,
"loss": 0.2800780236721039,
"step": 127
},
{
"epoch": 1.3763440860215055,
"grad_norm": 0.5383888711112677,
"learning_rate": 6.08782362598054e-06,
"loss": 0.29685452580451965,
"step": 128
},
{
"epoch": 1.3870967741935485,
"grad_norm": 0.5260069329849035,
"learning_rate": 6.029893093705492e-06,
"loss": 0.30100512504577637,
"step": 129
},
{
"epoch": 1.3978494623655915,
"grad_norm": 0.5194176965790988,
"learning_rate": 5.971817819267914e-06,
"loss": 0.28969407081604004,
"step": 130
},
{
"epoch": 1.4086021505376345,
"grad_norm": 0.5795379145956892,
"learning_rate": 5.9136059646225375e-06,
"loss": 0.3088012635707855,
"step": 131
},
{
"epoch": 1.4193548387096775,
"grad_norm": 0.5605938731509191,
"learning_rate": 5.855265710919211e-06,
"loss": 0.2954477369785309,
"step": 132
},
{
"epoch": 1.4301075268817205,
"grad_norm": 0.5513949309182947,
"learning_rate": 5.796805257353109e-06,
"loss": 0.31166109442710876,
"step": 133
},
{
"epoch": 1.4408602150537635,
"grad_norm": 0.6271394444282732,
"learning_rate": 5.738232820012407e-06,
"loss": 0.33265459537506104,
"step": 134
},
{
"epoch": 1.4516129032258065,
"grad_norm": 0.5513926189063966,
"learning_rate": 5.679556630723592e-06,
"loss": 0.271618515253067,
"step": 135
},
{
"epoch": 1.4623655913978495,
"grad_norm": 0.5799088021416959,
"learning_rate": 5.620784935894548e-06,
"loss": 0.2902123034000397,
"step": 136
},
{
"epoch": 1.4731182795698925,
"grad_norm": 0.5662089313253346,
"learning_rate": 5.561925995355595e-06,
"loss": 0.2844197154045105,
"step": 137
},
{
"epoch": 1.4838709677419355,
"grad_norm": 0.9937555515678825,
"learning_rate": 5.5029880811986546e-06,
"loss": 0.2922922372817993,
"step": 138
},
{
"epoch": 1.4946236559139785,
"grad_norm": 0.591368004900944,
"learning_rate": 5.443979476614674e-06,
"loss": 0.28497251868247986,
"step": 139
},
{
"epoch": 1.5053763440860215,
"grad_norm": 0.6076112891374724,
"learning_rate": 5.384908474729501e-06,
"loss": 0.31142377853393555,
"step": 140
},
{
"epoch": 1.5161290322580645,
"grad_norm": 0.5765965412966652,
"learning_rate": 5.325783377438357e-06,
"loss": 0.3015650510787964,
"step": 141
},
{
"epoch": 1.5268817204301075,
"grad_norm": 0.5232918774810716,
"learning_rate": 5.266612494239088e-06,
"loss": 0.275716096162796,
"step": 142
},
{
"epoch": 1.5376344086021505,
"grad_norm": 0.5680197113944581,
"learning_rate": 5.207404141064334e-06,
"loss": 0.2873622179031372,
"step": 143
},
{
"epoch": 1.5483870967741935,
"grad_norm": 0.49399262225350316,
"learning_rate": 5.148166639112799e-06,
"loss": 0.26688140630722046,
"step": 144
},
{
"epoch": 1.5591397849462365,
"grad_norm": 0.5949244178581083,
"learning_rate": 5.088908313679788e-06,
"loss": 0.29102981090545654,
"step": 145
},
{
"epoch": 1.5698924731182795,
"grad_norm": 0.6289302836737687,
"learning_rate": 5.029637492987153e-06,
"loss": 0.2931191027164459,
"step": 146
},
{
"epoch": 1.5806451612903225,
"grad_norm": 0.6088833365757095,
"learning_rate": 4.970362507012848e-06,
"loss": 0.2844638228416443,
"step": 147
},
{
"epoch": 1.5913978494623655,
"grad_norm": 0.6971118150263104,
"learning_rate": 4.911091686320213e-06,
"loss": 0.3226025104522705,
"step": 148
},
{
"epoch": 1.6021505376344085,
"grad_norm": 0.5612811561354285,
"learning_rate": 4.8518333608872015e-06,
"loss": 0.28764593601226807,
"step": 149
},
{
"epoch": 1.6129032258064515,
"grad_norm": 0.6065928167636621,
"learning_rate": 4.792595858935668e-06,
"loss": 0.3277990221977234,
"step": 150
},
{
"epoch": 1.6236559139784945,
"grad_norm": 0.5135483316527438,
"learning_rate": 4.733387505760913e-06,
"loss": 0.2962847948074341,
"step": 151
},
{
"epoch": 1.6344086021505375,
"grad_norm": 0.6160906809336857,
"learning_rate": 4.674216622561645e-06,
"loss": 0.294375479221344,
"step": 152
},
{
"epoch": 1.6451612903225805,
"grad_norm": 0.6058840168616271,
"learning_rate": 4.6150915252705005e-06,
"loss": 0.3054540157318115,
"step": 153
},
{
"epoch": 1.6559139784946235,
"grad_norm": 0.5298336251451402,
"learning_rate": 4.556020523385326e-06,
"loss": 0.30477648973464966,
"step": 154
},
{
"epoch": 1.6666666666666665,
"grad_norm": 0.5241054596736249,
"learning_rate": 4.497011918801347e-06,
"loss": 0.2855217456817627,
"step": 155
},
{
"epoch": 1.6774193548387095,
"grad_norm": 0.48994778885097257,
"learning_rate": 4.438074004644407e-06,
"loss": 0.30031388998031616,
"step": 156
},
{
"epoch": 1.6881720430107527,
"grad_norm": 0.6034882888193746,
"learning_rate": 4.379215064105454e-06,
"loss": 0.3072971999645233,
"step": 157
},
{
"epoch": 1.6989247311827957,
"grad_norm": 0.5057395078651564,
"learning_rate": 4.32044336927641e-06,
"loss": 0.27074217796325684,
"step": 158
},
{
"epoch": 1.7096774193548387,
"grad_norm": 0.548027010828449,
"learning_rate": 4.261767179987595e-06,
"loss": 0.28892627358436584,
"step": 159
},
{
"epoch": 1.7204301075268817,
"grad_norm": 0.5949799690870136,
"learning_rate": 4.203194742646893e-06,
"loss": 0.31755733489990234,
"step": 160
},
{
"epoch": 1.7311827956989247,
"grad_norm": 0.5988557766268561,
"learning_rate": 4.1447342890807905e-06,
"loss": 0.28983038663864136,
"step": 161
},
{
"epoch": 1.7419354838709677,
"grad_norm": 0.5153375977084437,
"learning_rate": 4.086394035377463e-06,
"loss": 0.2821354269981384,
"step": 162
},
{
"epoch": 1.7526881720430108,
"grad_norm": 0.5758453410329892,
"learning_rate": 4.028182180732088e-06,
"loss": 0.26659733057022095,
"step": 163
},
{
"epoch": 1.7634408602150538,
"grad_norm": 0.5515188533069812,
"learning_rate": 3.970106906294509e-06,
"loss": 0.29241591691970825,
"step": 164
},
{
"epoch": 1.7741935483870968,
"grad_norm": 0.5465960078319545,
"learning_rate": 3.912176374019462e-06,
"loss": 0.28333449363708496,
"step": 165
},
{
"epoch": 1.7849462365591398,
"grad_norm": 0.5696267902501757,
"learning_rate": 3.85439872551948e-06,
"loss": 0.28454482555389404,
"step": 166
},
{
"epoch": 1.7956989247311828,
"grad_norm": 0.5537225691065368,
"learning_rate": 3.796782080920659e-06,
"loss": 0.29573291540145874,
"step": 167
},
{
"epoch": 1.8064516129032258,
"grad_norm": 0.5317821119388748,
"learning_rate": 3.7393345377214584e-06,
"loss": 0.3006817400455475,
"step": 168
},
{
"epoch": 1.817204301075269,
"grad_norm": 0.5193143293546626,
"learning_rate": 3.682064169654663e-06,
"loss": 0.28759250044822693,
"step": 169
},
{
"epoch": 1.827956989247312,
"grad_norm": 0.5024936169709733,
"learning_rate": 3.6249790255526916e-06,
"loss": 0.279812753200531,
"step": 170
},
{
"epoch": 1.838709677419355,
"grad_norm": 0.522744432939167,
"learning_rate": 3.568087128216414e-06,
"loss": 0.2699696123600006,
"step": 171
},
{
"epoch": 1.849462365591398,
"grad_norm": 0.6740080747223203,
"learning_rate": 3.511396473287611e-06,
"loss": 0.3212928771972656,
"step": 172
},
{
"epoch": 1.860215053763441,
"grad_norm": 0.5393424030521861,
"learning_rate": 3.4549150281252635e-06,
"loss": 0.28264331817626953,
"step": 173
},
{
"epoch": 1.870967741935484,
"grad_norm": 0.5389635395816308,
"learning_rate": 3.398650730685813e-06,
"loss": 0.2982739508152008,
"step": 174
},
{
"epoch": 1.881720430107527,
"grad_norm": 0.5527956538218779,
"learning_rate": 3.3426114884075488e-06,
"loss": 0.2892792820930481,
"step": 175
},
{
"epoch": 1.89247311827957,
"grad_norm": 0.5783362380337232,
"learning_rate": 3.2868051770992935e-06,
"loss": 0.2867039442062378,
"step": 176
},
{
"epoch": 1.903225806451613,
"grad_norm": 0.4505107682619091,
"learning_rate": 3.2312396398335312e-06,
"loss": 0.2998119592666626,
"step": 177
},
{
"epoch": 1.913978494623656,
"grad_norm": 0.544159970587014,
"learning_rate": 3.175922685844125e-06,
"loss": 0.2946498394012451,
"step": 178
},
{
"epoch": 1.924731182795699,
"grad_norm": 0.5601933382330992,
"learning_rate": 3.1208620894288105e-06,
"loss": 0.30848872661590576,
"step": 179
},
{
"epoch": 1.935483870967742,
"grad_norm": 0.5692086999688832,
"learning_rate": 3.0660655888565827e-06,
"loss": 0.30292433500289917,
"step": 180
},
{
"epoch": 1.946236559139785,
"grad_norm": 0.5204759182695003,
"learning_rate": 3.0115408852801535e-06,
"loss": 0.2992181181907654,
"step": 181
},
{
"epoch": 1.956989247311828,
"grad_norm": 0.547974934732699,
"learning_rate": 2.9572956416536267e-06,
"loss": 0.3052282929420471,
"step": 182
},
{
"epoch": 1.967741935483871,
"grad_norm": 0.5036888640423901,
"learning_rate": 2.9033374816555338e-06,
"loss": 0.27619490027427673,
"step": 183
},
{
"epoch": 1.978494623655914,
"grad_norm": 0.5857097319940581,
"learning_rate": 2.8496739886173994e-06,
"loss": 0.2969588041305542,
"step": 184
},
{
"epoch": 1.989247311827957,
"grad_norm": 0.5125923920870736,
"learning_rate": 2.7963127044579697e-06,
"loss": 0.28236857056617737,
"step": 185
},
{
"epoch": 2.0,
"grad_norm": 0.5460296926482203,
"learning_rate": 2.743261128623269e-06,
"loss": 0.2802872657775879,
"step": 186
},
{
"epoch": 2.010752688172043,
"grad_norm": 0.5627765344162555,
"learning_rate": 2.6905267170326143e-06,
"loss": 0.2465861439704895,
"step": 187
},
{
"epoch": 2.021505376344086,
"grad_norm": 0.5241819547135816,
"learning_rate": 2.6381168810307536e-06,
"loss": 0.2593002915382385,
"step": 188
},
{
"epoch": 2.032258064516129,
"grad_norm": 0.5146155536713976,
"learning_rate": 2.5860389863462765e-06,
"loss": 0.22058740258216858,
"step": 189
},
{
"epoch": 2.043010752688172,
"grad_norm": 0.527929699946712,
"learning_rate": 2.534300352056416e-06,
"loss": 0.24380052089691162,
"step": 190
},
{
"epoch": 2.053763440860215,
"grad_norm": 0.5680981651809573,
"learning_rate": 2.4829082495584244e-06,
"loss": 0.2181762158870697,
"step": 191
},
{
"epoch": 2.064516129032258,
"grad_norm": 0.5210266779766828,
"learning_rate": 2.4318699015476495e-06,
"loss": 0.21483556926250458,
"step": 192
},
{
"epoch": 2.075268817204301,
"grad_norm": 0.5228849523676178,
"learning_rate": 2.3811924810024385e-06,
"loss": 0.2233523428440094,
"step": 193
},
{
"epoch": 2.086021505376344,
"grad_norm": 1.2592287942902403,
"learning_rate": 2.330883110176049e-06,
"loss": 0.21375712752342224,
"step": 194
},
{
"epoch": 2.096774193548387,
"grad_norm": 0.5541159570889329,
"learning_rate": 2.2809488595956746e-06,
"loss": 0.22336843609809875,
"step": 195
},
{
"epoch": 2.10752688172043,
"grad_norm": 0.6275490404272853,
"learning_rate": 2.2313967470687593e-06,
"loss": 0.23877255618572235,
"step": 196
},
{
"epoch": 2.118279569892473,
"grad_norm": 0.5287134027247969,
"learning_rate": 2.18223373669669e-06,
"loss": 0.23572032153606415,
"step": 197
},
{
"epoch": 2.129032258064516,
"grad_norm": 0.6005092115812523,
"learning_rate": 2.1334667378960642e-06,
"loss": 0.25271525979042053,
"step": 198
},
{
"epoch": 2.139784946236559,
"grad_norm": 0.5729023596023511,
"learning_rate": 2.0851026044276405e-06,
"loss": 0.2403365820646286,
"step": 199
},
{
"epoch": 2.150537634408602,
"grad_norm": 0.502960648659189,
"learning_rate": 2.0371481334330913e-06,
"loss": 0.22194445133209229,
"step": 200
},
{
"epoch": 2.161290322580645,
"grad_norm": 1.0792975178234447,
"learning_rate": 1.9896100644797316e-06,
"loss": 0.20418012142181396,
"step": 201
},
{
"epoch": 2.172043010752688,
"grad_norm": 0.563900355756192,
"learning_rate": 1.9424950786133414e-06,
"loss": 0.2197030484676361,
"step": 202
},
{
"epoch": 2.182795698924731,
"grad_norm": 0.6049813205320395,
"learning_rate": 1.8958097974191909e-06,
"loss": 0.23065771162509918,
"step": 203
},
{
"epoch": 2.193548387096774,
"grad_norm": 0.5693258359321052,
"learning_rate": 1.8495607820914451e-06,
"loss": 0.22469733655452728,
"step": 204
},
{
"epoch": 2.204301075268817,
"grad_norm": 0.49687778546847566,
"learning_rate": 1.8037545325110506e-06,
"loss": 0.22447475790977478,
"step": 205
},
{
"epoch": 2.21505376344086,
"grad_norm": 0.6870685608162285,
"learning_rate": 1.7583974863322272e-06,
"loss": 0.2232748568058014,
"step": 206
},
{
"epoch": 2.225806451612903,
"grad_norm": 0.5587791759635568,
"learning_rate": 1.7134960180777171e-06,
"loss": 0.21617698669433594,
"step": 207
},
{
"epoch": 2.236559139784946,
"grad_norm": 0.5406726897347297,
"learning_rate": 1.6690564382429104e-06,
"loss": 0.22749440371990204,
"step": 208
},
{
"epoch": 2.247311827956989,
"grad_norm": 0.6059659457413509,
"learning_rate": 1.6250849924089485e-06,
"loss": 0.24092203378677368,
"step": 209
},
{
"epoch": 2.258064516129032,
"grad_norm": 0.5440152722934113,
"learning_rate": 1.581587860364977e-06,
"loss": 0.20948372781276703,
"step": 210
},
{
"epoch": 2.268817204301075,
"grad_norm": 0.48547134095880384,
"learning_rate": 1.5385711552396227e-06,
"loss": 0.21325388550758362,
"step": 211
},
{
"epoch": 2.279569892473118,
"grad_norm": 0.6324122951694761,
"learning_rate": 1.4960409226418576e-06,
"loss": 0.22965453565120697,
"step": 212
},
{
"epoch": 2.2903225806451615,
"grad_norm": 0.5469586332098659,
"learning_rate": 1.4540031398113335e-06,
"loss": 0.22315660119056702,
"step": 213
},
{
"epoch": 2.3010752688172045,
"grad_norm": 0.7520121411441703,
"learning_rate": 1.4124637147783431e-06,
"loss": 0.23476293683052063,
"step": 214
},
{
"epoch": 2.3118279569892475,
"grad_norm": 0.535533468892341,
"learning_rate": 1.371428485533498e-06,
"loss": 0.21404291689395905,
"step": 215
},
{
"epoch": 2.3225806451612905,
"grad_norm": 0.5561238179269572,
"learning_rate": 1.3309032192072463e-06,
"loss": 0.23378810286521912,
"step": 216
},
{
"epoch": 2.3333333333333335,
"grad_norm": 0.5663331713554731,
"learning_rate": 1.29089361125936e-06,
"loss": 0.2105513960123062,
"step": 217
},
{
"epoch": 2.3440860215053765,
"grad_norm": 0.5851084790781899,
"learning_rate": 1.251405284678488e-06,
"loss": 0.22501936554908752,
"step": 218
},
{
"epoch": 2.3548387096774195,
"grad_norm": 0.5546173616985522,
"learning_rate": 1.2124437891918995e-06,
"loss": 0.21492747962474823,
"step": 219
},
{
"epoch": 2.3655913978494625,
"grad_norm": 0.5542629930888266,
"learning_rate": 1.1740146004855141e-06,
"loss": 0.22757509350776672,
"step": 220
},
{
"epoch": 2.3763440860215055,
"grad_norm": 0.5368948762143404,
"learning_rate": 1.1361231194343436e-06,
"loss": 0.20776845514774323,
"step": 221
},
{
"epoch": 2.3870967741935485,
"grad_norm": 0.5362067849303692,
"learning_rate": 1.0987746713434578e-06,
"loss": 0.21643775701522827,
"step": 222
},
{
"epoch": 2.3978494623655915,
"grad_norm": 0.7478981160056104,
"learning_rate": 1.0619745051995473e-06,
"loss": 0.2321006953716278,
"step": 223
},
{
"epoch": 2.4086021505376345,
"grad_norm": 0.5589077284378241,
"learning_rate": 1.0257277929332332e-06,
"loss": 0.2335091233253479,
"step": 224
},
{
"epoch": 2.4193548387096775,
"grad_norm": 0.5449990060855128,
"learning_rate": 9.900396286922025e-07,
"loss": 0.22067061066627502,
"step": 225
},
{
"epoch": 2.4301075268817205,
"grad_norm": 0.545790815861183,
"learning_rate": 9.549150281252633e-07,
"loss": 0.24003136157989502,
"step": 226
},
{
"epoch": 2.4408602150537635,
"grad_norm": 0.5197037503280709,
"learning_rate": 9.203589276774438e-07,
"loss": 0.2232522964477539,
"step": 227
},
{
"epoch": 2.4516129032258065,
"grad_norm": 0.5783224759821376,
"learning_rate": 8.86376183896226e-07,
"loss": 0.2168661653995514,
"step": 228
},
{
"epoch": 2.4623655913978495,
"grad_norm": 0.5772298663365216,
"learning_rate": 8.529715727489912e-07,
"loss": 0.22855976223945618,
"step": 229
},
{
"epoch": 2.4731182795698925,
"grad_norm": 0.5901747678986362,
"learning_rate": 8.201497889518073e-07,
"loss": 0.23543155193328857,
"step": 230
},
{
"epoch": 2.4838709677419355,
"grad_norm": 0.4859350321815666,
"learning_rate": 7.879154453096305e-07,
"loss": 0.20444487035274506,
"step": 231
},
{
"epoch": 2.4946236559139785,
"grad_norm": 0.5733816209081586,
"learning_rate": 7.562730720680111e-07,
"loss": 0.22317257523536682,
"step": 232
},
{
"epoch": 2.5053763440860215,
"grad_norm": 0.6158850373560834,
"learning_rate": 7.25227116276413e-07,
"loss": 0.21304547786712646,
"step": 233
},
{
"epoch": 2.5161290322580645,
"grad_norm": 0.5303929149495401,
"learning_rate": 6.947819411632223e-07,
"loss": 0.22398605942726135,
"step": 234
},
{
"epoch": 2.5268817204301075,
"grad_norm": 0.5807287890882401,
"learning_rate": 6.649418255225298e-07,
"loss": 0.22487413883209229,
"step": 235
},
{
"epoch": 2.5376344086021505,
"grad_norm": 0.5151341689499322,
"learning_rate": 6.357109631127889e-07,
"loss": 0.20791815221309662,
"step": 236
},
{
"epoch": 2.5483870967741935,
"grad_norm": 0.5723989318815612,
"learning_rate": 6.07093462067419e-07,
"loss": 0.22853252291679382,
"step": 237
},
{
"epoch": 2.5591397849462365,
"grad_norm": 0.5151529413040788,
"learning_rate": 5.79093344317449e-07,
"loss": 0.23513737320899963,
"step": 238
},
{
"epoch": 2.5698924731182795,
"grad_norm": 0.6125760680997326,
"learning_rate": 5.517145450262639e-07,
"loss": 0.23416566848754883,
"step": 239
},
{
"epoch": 2.5806451612903225,
"grad_norm": 0.5169337829356161,
"learning_rate": 5.249609120365579e-07,
"loss": 0.2236918807029724,
"step": 240
},
{
"epoch": 2.5913978494623655,
"grad_norm": 0.5150874552544314,
"learning_rate": 4.988362053295564e-07,
"loss": 0.22128576040267944,
"step": 241
},
{
"epoch": 2.6021505376344085,
"grad_norm": 0.6254184464054292,
"learning_rate": 4.733440964965791e-07,
"loss": 0.23455382883548737,
"step": 242
},
{
"epoch": 2.6129032258064515,
"grad_norm": 0.5612944471466359,
"learning_rate": 4.484881682230341e-07,
"loss": 0.22176575660705566,
"step": 243
},
{
"epoch": 2.6236559139784945,
"grad_norm": 0.5178079276744024,
"learning_rate": 4.242719137849077e-07,
"loss": 0.21548646688461304,
"step": 244
},
{
"epoch": 2.6344086021505375,
"grad_norm": 0.5659047109155118,
"learning_rate": 4.00698736557808e-07,
"loss": 0.23057113587856293,
"step": 245
},
{
"epoch": 2.6451612903225805,
"grad_norm": 0.5098023427109231,
"learning_rate": 3.777719495386567e-07,
"loss": 0.2156354784965515,
"step": 246
},
{
"epoch": 2.6559139784946235,
"grad_norm": 0.49515330466811,
"learning_rate": 3.5549477488007853e-07,
"loss": 0.19704020023345947,
"step": 247
},
{
"epoch": 2.6666666666666665,
"grad_norm": 0.5297791040512543,
"learning_rate": 3.3387034343755063e-07,
"loss": 0.21541355550289154,
"step": 248
},
{
"epoch": 2.6774193548387095,
"grad_norm": 0.5134148705186912,
"learning_rate": 3.1290169432939556e-07,
"loss": 0.20781481266021729,
"step": 249
},
{
"epoch": 2.688172043010753,
"grad_norm": 0.6591675869543023,
"learning_rate": 2.925917745096568e-07,
"loss": 0.23452350497245789,
"step": 250
},
{
"epoch": 2.698924731182796,
"grad_norm": 0.5940779586998381,
"learning_rate": 2.7294343835393366e-07,
"loss": 0.23721283674240112,
"step": 251
},
{
"epoch": 2.709677419354839,
"grad_norm": 0.6904575634426214,
"learning_rate": 2.539594472582213e-07,
"loss": 0.23339498043060303,
"step": 252
},
{
"epoch": 2.720430107526882,
"grad_norm": 0.549700857643845,
"learning_rate": 2.3564246925082358e-07,
"loss": 0.2422715723514557,
"step": 253
},
{
"epoch": 2.731182795698925,
"grad_norm": 0.551951066945312,
"learning_rate": 2.179950786173879e-07,
"loss": 0.22343403100967407,
"step": 254
},
{
"epoch": 2.741935483870968,
"grad_norm": 0.5112230065122543,
"learning_rate": 2.01019755539108e-07,
"loss": 0.2022184431552887,
"step": 255
},
{
"epoch": 2.752688172043011,
"grad_norm": 0.6011326795775324,
"learning_rate": 1.8471888574415953e-07,
"loss": 0.21843938529491425,
"step": 256
},
{
"epoch": 2.763440860215054,
"grad_norm": 0.5598013580182175,
"learning_rate": 1.690947601724091e-07,
"loss": 0.2313247174024582,
"step": 257
},
{
"epoch": 2.774193548387097,
"grad_norm": 0.579260801737671,
"learning_rate": 1.5414957465343883e-07,
"loss": 0.23432737588882446,
"step": 258
},
{
"epoch": 2.78494623655914,
"grad_norm": 0.6080486926763811,
"learning_rate": 1.3988542959794627e-07,
"loss": 0.2501685619354248,
"step": 259
},
{
"epoch": 2.795698924731183,
"grad_norm": 0.6231912247704571,
"learning_rate": 1.2630432970255014e-07,
"loss": 0.22736617922782898,
"step": 260
},
{
"epoch": 2.806451612903226,
"grad_norm": 0.6526633675286432,
"learning_rate": 1.1340818366804728e-07,
"loss": 0.2217504233121872,
"step": 261
},
{
"epoch": 2.817204301075269,
"grad_norm": 0.5774442194442434,
"learning_rate": 1.0119880393116177e-07,
"loss": 0.21880705654621124,
"step": 262
},
{
"epoch": 2.827956989247312,
"grad_norm": 0.5510924764674725,
"learning_rate": 8.967790640982466e-08,
"loss": 0.23220205307006836,
"step": 263
},
{
"epoch": 2.838709677419355,
"grad_norm": 0.5940155308333281,
"learning_rate": 7.884711026201586e-08,
"loss": 0.22368629276752472,
"step": 264
},
{
"epoch": 2.849462365591398,
"grad_norm": 0.48018958702032566,
"learning_rate": 6.870793765820783e-08,
"loss": 0.2191024124622345,
"step": 265
},
{
"epoch": 2.860215053763441,
"grad_norm": 0.5348968055346366,
"learning_rate": 5.92618135674361e-08,
"loss": 0.2141711413860321,
"step": 266
},
{
"epoch": 2.870967741935484,
"grad_norm": 0.5058952998875674,
"learning_rate": 5.0510065557034526e-08,
"loss": 0.21273520588874817,
"step": 267
},
{
"epoch": 2.881720430107527,
"grad_norm": 0.5748923929140294,
"learning_rate": 4.245392360605727e-08,
"loss": 0.2339775264263153,
"step": 268
},
{
"epoch": 2.89247311827957,
"grad_norm": 0.543587597117276,
"learning_rate": 3.5094519932415417e-08,
"loss": 0.22123411297798157,
"step": 269
},
{
"epoch": 2.903225806451613,
"grad_norm": 0.5451715623802244,
"learning_rate": 2.843288883375539e-08,
"loss": 0.22225311398506165,
"step": 270
},
{
"epoch": 2.913978494623656,
"grad_norm": 0.5639808086322913,
"learning_rate": 2.2469966542096323e-08,
"loss": 0.220473051071167,
"step": 271
},
{
"epoch": 2.924731182795699,
"grad_norm": 0.5603998497829358,
"learning_rate": 1.7206591092253642e-08,
"loss": 0.22855965793132782,
"step": 272
},
{
"epoch": 2.935483870967742,
"grad_norm": 0.5700123425140283,
"learning_rate": 1.264350220405719e-08,
"loss": 0.226323202252388,
"step": 273
},
{
"epoch": 2.946236559139785,
"grad_norm": 0.5193108167577539,
"learning_rate": 8.781341178393244e-09,
"loss": 0.21938025951385498,
"step": 274
},
{
"epoch": 2.956989247311828,
"grad_norm": 0.6419513279403505,
"learning_rate": 5.620650807073857e-09,
"loss": 0.24610912799835205,
"step": 275
},
{
"epoch": 2.967741935483871,
"grad_norm": 0.6308682159730912,
"learning_rate": 3.1618752965534295e-09,
"loss": 0.22690746188163757,
"step": 276
},
{
"epoch": 2.978494623655914,
"grad_norm": 0.5263395659056124,
"learning_rate": 1.4053602054991954e-09,
"loss": 0.20472858846187592,
"step": 277
},
{
"epoch": 2.989247311827957,
"grad_norm": 0.5329091838420249,
"learning_rate": 3.513523962256349e-10,
"loss": 0.2296607792377472,
"step": 278
},
{
"epoch": 3.0,
"grad_norm": 0.6481688008102051,
"learning_rate": 0.0,
"loss": 0.22974838316440582,
"step": 279
}
],
"logging_steps": 1,
"max_steps": 279,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 142925655572480.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}