Succ_21_87000 / trainer_state.json
sgarrett's picture
Upload 14 files
279a9c8 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 151.8324607329843,
"eval_steps": 500,
"global_step": 87000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.8726003490401396,
"grad_norm": 0.931673526763916,
"learning_rate": 0.0009956369982547994,
"loss": 2.6507,
"step": 500
},
{
"epoch": 1.7452006980802792,
"grad_norm": 1.8975260257720947,
"learning_rate": 0.0009912739965095986,
"loss": 1.6969,
"step": 1000
},
{
"epoch": 2.6178010471204187,
"grad_norm": 0.5923702120780945,
"learning_rate": 0.000986910994764398,
"loss": 1.287,
"step": 1500
},
{
"epoch": 3.4904013961605584,
"grad_norm": 0.79058438539505,
"learning_rate": 0.000982547993019197,
"loss": 0.9398,
"step": 2000
},
{
"epoch": 4.363001745200698,
"grad_norm": 0.7330807447433472,
"learning_rate": 0.0009781849912739965,
"loss": 0.6968,
"step": 2500
},
{
"epoch": 5.2356020942408374,
"grad_norm": 0.6996338963508606,
"learning_rate": 0.0009738219895287959,
"loss": 0.4816,
"step": 3000
},
{
"epoch": 6.108202443280978,
"grad_norm": 0.5383062362670898,
"learning_rate": 0.0009694589877835951,
"loss": 0.3189,
"step": 3500
},
{
"epoch": 6.980802792321117,
"grad_norm": 0.6486771106719971,
"learning_rate": 0.0009650959860383944,
"loss": 0.2218,
"step": 4000
},
{
"epoch": 7.853403141361256,
"grad_norm": 0.507366955280304,
"learning_rate": 0.0009607329842931938,
"loss": 0.1542,
"step": 4500
},
{
"epoch": 8.726003490401396,
"grad_norm": 0.8452386260032654,
"learning_rate": 0.000956369982547993,
"loss": 0.1285,
"step": 5000
},
{
"epoch": 9.598603839441536,
"grad_norm": 0.43142977356910706,
"learning_rate": 0.0009520069808027923,
"loss": 0.1175,
"step": 5500
},
{
"epoch": 10.471204188481675,
"grad_norm": 0.2331060916185379,
"learning_rate": 0.0009476439790575916,
"loss": 0.1053,
"step": 6000
},
{
"epoch": 11.343804537521814,
"grad_norm": 0.4272315800189972,
"learning_rate": 0.000943280977312391,
"loss": 0.092,
"step": 6500
},
{
"epoch": 12.216404886561955,
"grad_norm": 0.2999955117702484,
"learning_rate": 0.0009389179755671902,
"loss": 0.0878,
"step": 7000
},
{
"epoch": 13.089005235602095,
"grad_norm": 2.55757474899292,
"learning_rate": 0.0009345549738219895,
"loss": 0.0884,
"step": 7500
},
{
"epoch": 13.961605584642234,
"grad_norm": 0.33913654088974,
"learning_rate": 0.0009301919720767889,
"loss": 0.0805,
"step": 8000
},
{
"epoch": 14.834205933682373,
"grad_norm": 0.3642922341823578,
"learning_rate": 0.0009258289703315882,
"loss": 0.0721,
"step": 8500
},
{
"epoch": 15.706806282722512,
"grad_norm": 0.7718423008918762,
"learning_rate": 0.0009214659685863874,
"loss": 0.0687,
"step": 9000
},
{
"epoch": 16.57940663176265,
"grad_norm": 0.5820666551589966,
"learning_rate": 0.0009171029668411868,
"loss": 0.0644,
"step": 9500
},
{
"epoch": 17.452006980802793,
"grad_norm": 0.2773011028766632,
"learning_rate": 0.000912739965095986,
"loss": 0.0598,
"step": 10000
},
{
"epoch": 18.324607329842934,
"grad_norm": 0.25250518321990967,
"learning_rate": 0.0009083769633507853,
"loss": 0.0653,
"step": 10500
},
{
"epoch": 19.19720767888307,
"grad_norm": 0.24120768904685974,
"learning_rate": 0.0009040139616055847,
"loss": 0.0632,
"step": 11000
},
{
"epoch": 20.069808027923212,
"grad_norm": 0.2897244989871979,
"learning_rate": 0.0008996509598603839,
"loss": 0.0594,
"step": 11500
},
{
"epoch": 20.94240837696335,
"grad_norm": 0.25145065784454346,
"learning_rate": 0.0008952879581151833,
"loss": 0.0556,
"step": 12000
},
{
"epoch": 21.81500872600349,
"grad_norm": 0.27175939083099365,
"learning_rate": 0.0008909249563699826,
"loss": 0.0481,
"step": 12500
},
{
"epoch": 22.68760907504363,
"grad_norm": 0.8626015782356262,
"learning_rate": 0.0008865619546247818,
"loss": 0.0466,
"step": 13000
},
{
"epoch": 23.56020942408377,
"grad_norm": 0.18672889471054077,
"learning_rate": 0.0008821989528795812,
"loss": 0.0512,
"step": 13500
},
{
"epoch": 24.43280977312391,
"grad_norm": 0.2387542873620987,
"learning_rate": 0.0008778359511343804,
"loss": 0.0512,
"step": 14000
},
{
"epoch": 25.305410122164048,
"grad_norm": 0.2665075957775116,
"learning_rate": 0.0008734729493891797,
"loss": 0.0483,
"step": 14500
},
{
"epoch": 26.17801047120419,
"grad_norm": 0.16715960204601288,
"learning_rate": 0.0008691099476439791,
"loss": 0.0441,
"step": 15000
},
{
"epoch": 27.050610820244327,
"grad_norm": 0.1875993311405182,
"learning_rate": 0.0008647469458987784,
"loss": 0.0397,
"step": 15500
},
{
"epoch": 27.923211169284468,
"grad_norm": 0.25451162457466125,
"learning_rate": 0.0008603839441535776,
"loss": 0.039,
"step": 16000
},
{
"epoch": 28.79581151832461,
"grad_norm": 0.307054728269577,
"learning_rate": 0.000856020942408377,
"loss": 0.0443,
"step": 16500
},
{
"epoch": 29.668411867364746,
"grad_norm": 0.2467741221189499,
"learning_rate": 0.0008516579406631763,
"loss": 0.038,
"step": 17000
},
{
"epoch": 30.541012216404887,
"grad_norm": 0.23178012669086456,
"learning_rate": 0.0008472949389179755,
"loss": 0.0351,
"step": 17500
},
{
"epoch": 31.413612565445025,
"grad_norm": 0.35061487555503845,
"learning_rate": 0.0008429319371727748,
"loss": 0.0323,
"step": 18000
},
{
"epoch": 32.28621291448517,
"grad_norm": 0.3363337218761444,
"learning_rate": 0.0008385689354275742,
"loss": 0.038,
"step": 18500
},
{
"epoch": 33.1588132635253,
"grad_norm": 0.24015972018241882,
"learning_rate": 0.0008342059336823735,
"loss": 0.0361,
"step": 19000
},
{
"epoch": 34.031413612565444,
"grad_norm": 0.5147364735603333,
"learning_rate": 0.0008298429319371727,
"loss": 0.0318,
"step": 19500
},
{
"epoch": 34.904013961605585,
"grad_norm": 0.17477251589298248,
"learning_rate": 0.0008254799301919721,
"loss": 0.0307,
"step": 20000
},
{
"epoch": 35.776614310645726,
"grad_norm": 0.2511900067329407,
"learning_rate": 0.0008211169284467714,
"loss": 0.0348,
"step": 20500
},
{
"epoch": 36.64921465968587,
"grad_norm": 0.2036200314760208,
"learning_rate": 0.0008167539267015707,
"loss": 0.0298,
"step": 21000
},
{
"epoch": 37.521815008726,
"grad_norm": 0.16981545090675354,
"learning_rate": 0.00081239092495637,
"loss": 0.0319,
"step": 21500
},
{
"epoch": 38.39441535776614,
"grad_norm": 0.22329097986221313,
"learning_rate": 0.0008080279232111692,
"loss": 0.0359,
"step": 22000
},
{
"epoch": 39.26701570680628,
"grad_norm": 0.13258185982704163,
"learning_rate": 0.0008036649214659686,
"loss": 0.0266,
"step": 22500
},
{
"epoch": 40.139616055846425,
"grad_norm": 0.12790647149085999,
"learning_rate": 0.000799301919720768,
"loss": 0.026,
"step": 23000
},
{
"epoch": 41.01221640488656,
"grad_norm": 0.23635344207286835,
"learning_rate": 0.0007949389179755671,
"loss": 0.0279,
"step": 23500
},
{
"epoch": 41.8848167539267,
"grad_norm": 0.11364254355430603,
"learning_rate": 0.0007905759162303665,
"loss": 0.0257,
"step": 24000
},
{
"epoch": 42.75741710296684,
"grad_norm": 0.2781168520450592,
"learning_rate": 0.0007862129144851659,
"loss": 0.0295,
"step": 24500
},
{
"epoch": 43.63001745200698,
"grad_norm": 0.106789730489254,
"learning_rate": 0.0007818499127399651,
"loss": 0.0308,
"step": 25000
},
{
"epoch": 44.50261780104712,
"grad_norm": 0.16404911875724792,
"learning_rate": 0.0007774869109947644,
"loss": 0.0222,
"step": 25500
},
{
"epoch": 45.37521815008726,
"grad_norm": 0.14249293506145477,
"learning_rate": 0.0007731239092495637,
"loss": 0.0225,
"step": 26000
},
{
"epoch": 46.2478184991274,
"grad_norm": 0.1853444129228592,
"learning_rate": 0.0007687609075043631,
"loss": 0.0261,
"step": 26500
},
{
"epoch": 47.12041884816754,
"grad_norm": 0.1456003040075302,
"learning_rate": 0.0007643979057591623,
"loss": 0.0252,
"step": 27000
},
{
"epoch": 47.99301919720768,
"grad_norm": 0.16386698186397552,
"learning_rate": 0.0007600349040139616,
"loss": 0.0259,
"step": 27500
},
{
"epoch": 48.86561954624782,
"grad_norm": 0.12221992015838623,
"learning_rate": 0.000755671902268761,
"loss": 0.0253,
"step": 28000
},
{
"epoch": 49.738219895287955,
"grad_norm": 0.14093224704265594,
"learning_rate": 0.0007513089005235602,
"loss": 0.0203,
"step": 28500
},
{
"epoch": 50.610820244328096,
"grad_norm": 0.1189383938908577,
"learning_rate": 0.0007469458987783595,
"loss": 0.0207,
"step": 29000
},
{
"epoch": 51.48342059336824,
"grad_norm": 0.1471104919910431,
"learning_rate": 0.0007425828970331589,
"loss": 0.0209,
"step": 29500
},
{
"epoch": 52.35602094240838,
"grad_norm": 0.08947575837373734,
"learning_rate": 0.0007382198952879581,
"loss": 0.0234,
"step": 30000
},
{
"epoch": 53.22862129144852,
"grad_norm": 0.18746259808540344,
"learning_rate": 0.0007338568935427574,
"loss": 0.0245,
"step": 30500
},
{
"epoch": 54.10122164048865,
"grad_norm": 0.1539311408996582,
"learning_rate": 0.0007294938917975568,
"loss": 0.0214,
"step": 31000
},
{
"epoch": 54.973821989528794,
"grad_norm": 0.11201947182416916,
"learning_rate": 0.000725130890052356,
"loss": 0.0194,
"step": 31500
},
{
"epoch": 55.846422338568935,
"grad_norm": 0.16618479788303375,
"learning_rate": 0.0007207678883071554,
"loss": 0.0185,
"step": 32000
},
{
"epoch": 56.719022687609076,
"grad_norm": 0.1569599211215973,
"learning_rate": 0.0007164048865619547,
"loss": 0.0234,
"step": 32500
},
{
"epoch": 57.59162303664922,
"grad_norm": 0.11062045395374298,
"learning_rate": 0.0007120418848167539,
"loss": 0.0187,
"step": 33000
},
{
"epoch": 58.46422338568935,
"grad_norm": 0.1617700457572937,
"learning_rate": 0.0007076788830715533,
"loss": 0.0176,
"step": 33500
},
{
"epoch": 59.33682373472949,
"grad_norm": 0.11750755459070206,
"learning_rate": 0.0007033158813263525,
"loss": 0.0188,
"step": 34000
},
{
"epoch": 60.20942408376963,
"grad_norm": 0.24937282502651215,
"learning_rate": 0.0006989528795811518,
"loss": 0.0225,
"step": 34500
},
{
"epoch": 61.082024432809774,
"grad_norm": 0.22410957515239716,
"learning_rate": 0.0006945898778359512,
"loss": 0.0192,
"step": 35000
},
{
"epoch": 61.954624781849915,
"grad_norm": 0.18029357492923737,
"learning_rate": 0.0006902268760907505,
"loss": 0.0169,
"step": 35500
},
{
"epoch": 62.82722513089005,
"grad_norm": 0.14077898859977722,
"learning_rate": 0.0006858638743455497,
"loss": 0.0167,
"step": 36000
},
{
"epoch": 63.69982547993019,
"grad_norm": 0.12804169952869415,
"learning_rate": 0.0006815008726003491,
"loss": 0.0167,
"step": 36500
},
{
"epoch": 64.57242582897034,
"grad_norm": 0.11356078833341599,
"learning_rate": 0.0006771378708551484,
"loss": 0.0183,
"step": 37000
},
{
"epoch": 65.44502617801047,
"grad_norm": 0.07214757055044174,
"learning_rate": 0.0006727748691099476,
"loss": 0.0167,
"step": 37500
},
{
"epoch": 66.3176265270506,
"grad_norm": 0.20497408509254456,
"learning_rate": 0.0006684118673647469,
"loss": 0.0186,
"step": 38000
},
{
"epoch": 67.19022687609075,
"grad_norm": 0.09408937394618988,
"learning_rate": 0.0006640488656195463,
"loss": 0.0169,
"step": 38500
},
{
"epoch": 68.06282722513089,
"grad_norm": 0.12423662841320038,
"learning_rate": 0.0006596858638743456,
"loss": 0.0179,
"step": 39000
},
{
"epoch": 68.93542757417103,
"grad_norm": 0.4406953752040863,
"learning_rate": 0.0006553228621291448,
"loss": 0.019,
"step": 39500
},
{
"epoch": 69.80802792321117,
"grad_norm": 0.11233725398778915,
"learning_rate": 0.0006509598603839442,
"loss": 0.0151,
"step": 40000
},
{
"epoch": 70.68062827225131,
"grad_norm": 0.08892803639173508,
"learning_rate": 0.0006465968586387435,
"loss": 0.0137,
"step": 40500
},
{
"epoch": 71.55322862129145,
"grad_norm": 0.11712398380041122,
"learning_rate": 0.0006422338568935428,
"loss": 0.0145,
"step": 41000
},
{
"epoch": 72.4258289703316,
"grad_norm": 0.11560557782649994,
"learning_rate": 0.0006378708551483421,
"loss": 0.0181,
"step": 41500
},
{
"epoch": 73.29842931937172,
"grad_norm": 0.10780952870845795,
"learning_rate": 0.0006335078534031413,
"loss": 0.0162,
"step": 42000
},
{
"epoch": 74.17102966841186,
"grad_norm": 0.06540343165397644,
"learning_rate": 0.0006291448516579407,
"loss": 0.0157,
"step": 42500
},
{
"epoch": 75.043630017452,
"grad_norm": 0.08087161183357239,
"learning_rate": 0.00062478184991274,
"loss": 0.0181,
"step": 43000
},
{
"epoch": 75.91623036649214,
"grad_norm": 0.08909798413515091,
"learning_rate": 0.0006204188481675392,
"loss": 0.0136,
"step": 43500
},
{
"epoch": 76.78883071553228,
"grad_norm": 0.09045905619859695,
"learning_rate": 0.0006160558464223386,
"loss": 0.0124,
"step": 44000
},
{
"epoch": 77.66143106457243,
"grad_norm": 0.1375885307788849,
"learning_rate": 0.000611692844677138,
"loss": 0.013,
"step": 44500
},
{
"epoch": 78.53403141361257,
"grad_norm": 0.06584478169679642,
"learning_rate": 0.0006073298429319371,
"loss": 0.0149,
"step": 45000
},
{
"epoch": 79.40663176265271,
"grad_norm": 0.10484465211629868,
"learning_rate": 0.0006029668411867365,
"loss": 0.0141,
"step": 45500
},
{
"epoch": 80.27923211169285,
"grad_norm": 0.11171326786279678,
"learning_rate": 0.0005986038394415358,
"loss": 0.0142,
"step": 46000
},
{
"epoch": 81.15183246073299,
"grad_norm": 0.12636296451091766,
"learning_rate": 0.000594240837696335,
"loss": 0.0154,
"step": 46500
},
{
"epoch": 82.02443280977312,
"grad_norm": 0.06967689841985703,
"learning_rate": 0.0005898778359511344,
"loss": 0.0144,
"step": 47000
},
{
"epoch": 82.89703315881326,
"grad_norm": 0.09147424250841141,
"learning_rate": 0.0005855148342059337,
"loss": 0.0127,
"step": 47500
},
{
"epoch": 83.7696335078534,
"grad_norm": 0.6871227025985718,
"learning_rate": 0.000581151832460733,
"loss": 0.0113,
"step": 48000
},
{
"epoch": 84.64223385689354,
"grad_norm": 0.07160250097513199,
"learning_rate": 0.0005767888307155323,
"loss": 0.014,
"step": 48500
},
{
"epoch": 85.51483420593368,
"grad_norm": 0.06839723885059357,
"learning_rate": 0.0005724258289703316,
"loss": 0.0121,
"step": 49000
},
{
"epoch": 86.38743455497382,
"grad_norm": 0.05615156516432762,
"learning_rate": 0.0005680628272251309,
"loss": 0.0136,
"step": 49500
},
{
"epoch": 87.26003490401396,
"grad_norm": 0.10079669952392578,
"learning_rate": 0.0005636998254799302,
"loss": 0.0116,
"step": 50000
},
{
"epoch": 88.1326352530541,
"grad_norm": 0.09019599854946136,
"learning_rate": 0.0005593368237347295,
"loss": 0.0122,
"step": 50500
},
{
"epoch": 89.00523560209425,
"grad_norm": 0.05459260195493698,
"learning_rate": 0.0005549738219895288,
"loss": 0.0113,
"step": 51000
},
{
"epoch": 89.87783595113439,
"grad_norm": 0.07002388685941696,
"learning_rate": 0.0005506108202443281,
"loss": 0.0112,
"step": 51500
},
{
"epoch": 90.75043630017451,
"grad_norm": 0.12079566717147827,
"learning_rate": 0.0005462478184991274,
"loss": 0.0115,
"step": 52000
},
{
"epoch": 91.62303664921465,
"grad_norm": 0.08160880208015442,
"learning_rate": 0.0005418848167539267,
"loss": 0.0135,
"step": 52500
},
{
"epoch": 92.4956369982548,
"grad_norm": 0.053758975118398666,
"learning_rate": 0.000537521815008726,
"loss": 0.0111,
"step": 53000
},
{
"epoch": 93.36823734729494,
"grad_norm": 0.08977492153644562,
"learning_rate": 0.0005331588132635254,
"loss": 0.0118,
"step": 53500
},
{
"epoch": 94.24083769633508,
"grad_norm": 0.14162831008434296,
"learning_rate": 0.0005287958115183245,
"loss": 0.0118,
"step": 54000
},
{
"epoch": 95.11343804537522,
"grad_norm": 0.09927529096603394,
"learning_rate": 0.0005244328097731239,
"loss": 0.011,
"step": 54500
},
{
"epoch": 95.98603839441536,
"grad_norm": 0.08714314550161362,
"learning_rate": 0.0005200698080279233,
"loss": 0.0101,
"step": 55000
},
{
"epoch": 96.8586387434555,
"grad_norm": 0.09934234619140625,
"learning_rate": 0.0005157068062827224,
"loss": 0.0109,
"step": 55500
},
{
"epoch": 97.73123909249564,
"grad_norm": 0.11947502195835114,
"learning_rate": 0.0005113438045375218,
"loss": 0.0114,
"step": 56000
},
{
"epoch": 98.60383944153578,
"grad_norm": 0.0667385384440422,
"learning_rate": 0.0005069808027923212,
"loss": 0.0099,
"step": 56500
},
{
"epoch": 99.47643979057591,
"grad_norm": 0.11594051122665405,
"learning_rate": 0.0005026178010471204,
"loss": 0.0117,
"step": 57000
},
{
"epoch": 100.34904013961605,
"grad_norm": 0.06558683514595032,
"learning_rate": 0.0004982547993019197,
"loss": 0.0104,
"step": 57500
},
{
"epoch": 101.22164048865619,
"grad_norm": 0.07644475996494293,
"learning_rate": 0.000493891797556719,
"loss": 0.0098,
"step": 58000
},
{
"epoch": 102.09424083769633,
"grad_norm": 0.06626095622777939,
"learning_rate": 0.0004895287958115183,
"loss": 0.0095,
"step": 58500
},
{
"epoch": 102.96684118673647,
"grad_norm": 0.10808754712343216,
"learning_rate": 0.0004851657940663176,
"loss": 0.0108,
"step": 59000
},
{
"epoch": 103.83944153577661,
"grad_norm": 0.05127561092376709,
"learning_rate": 0.000480802792321117,
"loss": 0.01,
"step": 59500
},
{
"epoch": 104.71204188481676,
"grad_norm": 0.12128196656703949,
"learning_rate": 0.00047643979057591625,
"loss": 0.0092,
"step": 60000
},
{
"epoch": 105.5846422338569,
"grad_norm": 0.06882330775260925,
"learning_rate": 0.0004720767888307155,
"loss": 0.0111,
"step": 60500
},
{
"epoch": 106.45724258289704,
"grad_norm": 0.05268734693527222,
"learning_rate": 0.00046771378708551485,
"loss": 0.0094,
"step": 61000
},
{
"epoch": 107.32984293193718,
"grad_norm": 0.046268824487924576,
"learning_rate": 0.00046335078534031417,
"loss": 0.0082,
"step": 61500
},
{
"epoch": 108.2024432809773,
"grad_norm": 0.04883289709687233,
"learning_rate": 0.00045898778359511344,
"loss": 0.0085,
"step": 62000
},
{
"epoch": 109.07504363001745,
"grad_norm": 0.0723048597574234,
"learning_rate": 0.00045462478184991276,
"loss": 0.0093,
"step": 62500
},
{
"epoch": 109.94764397905759,
"grad_norm": 0.06026541814208031,
"learning_rate": 0.00045026178010471203,
"loss": 0.0095,
"step": 63000
},
{
"epoch": 110.82024432809773,
"grad_norm": 0.058908674865961075,
"learning_rate": 0.00044589877835951136,
"loss": 0.0085,
"step": 63500
},
{
"epoch": 111.69284467713787,
"grad_norm": 0.05758107081055641,
"learning_rate": 0.0004415357766143107,
"loss": 0.01,
"step": 64000
},
{
"epoch": 112.56544502617801,
"grad_norm": 0.06559444963932037,
"learning_rate": 0.00043717277486910995,
"loss": 0.0088,
"step": 64500
},
{
"epoch": 113.43804537521815,
"grad_norm": 0.05080035701394081,
"learning_rate": 0.0004328097731239092,
"loss": 0.0083,
"step": 65000
},
{
"epoch": 114.3106457242583,
"grad_norm": 0.0523524135351181,
"learning_rate": 0.0004284467713787086,
"loss": 0.0078,
"step": 65500
},
{
"epoch": 115.18324607329843,
"grad_norm": 0.14169646799564362,
"learning_rate": 0.00042408376963350787,
"loss": 0.0085,
"step": 66000
},
{
"epoch": 116.05584642233858,
"grad_norm": 0.05305915325880051,
"learning_rate": 0.00041972076788830714,
"loss": 0.0091,
"step": 66500
},
{
"epoch": 116.9284467713787,
"grad_norm": 0.05915080010890961,
"learning_rate": 0.00041535776614310646,
"loss": 0.0085,
"step": 67000
},
{
"epoch": 117.80104712041884,
"grad_norm": 0.07950141280889511,
"learning_rate": 0.0004109947643979058,
"loss": 0.0075,
"step": 67500
},
{
"epoch": 118.67364746945898,
"grad_norm": 0.083484947681427,
"learning_rate": 0.00040663176265270506,
"loss": 0.0086,
"step": 68000
},
{
"epoch": 119.54624781849913,
"grad_norm": 0.1149265244603157,
"learning_rate": 0.0004022687609075044,
"loss": 0.0087,
"step": 68500
},
{
"epoch": 120.41884816753927,
"grad_norm": 0.10079418867826462,
"learning_rate": 0.00039790575916230365,
"loss": 0.0087,
"step": 69000
},
{
"epoch": 121.29144851657941,
"grad_norm": 0.04444717988371849,
"learning_rate": 0.000393542757417103,
"loss": 0.0083,
"step": 69500
},
{
"epoch": 122.16404886561955,
"grad_norm": 0.015783872455358505,
"learning_rate": 0.0003891797556719023,
"loss": 0.0074,
"step": 70000
},
{
"epoch": 123.03664921465969,
"grad_norm": 0.07311473041772842,
"learning_rate": 0.00038481675392670157,
"loss": 0.007,
"step": 70500
},
{
"epoch": 123.90924956369983,
"grad_norm": 0.03907659277319908,
"learning_rate": 0.00038045375218150084,
"loss": 0.0071,
"step": 71000
},
{
"epoch": 124.78184991273997,
"grad_norm": 0.031013870611786842,
"learning_rate": 0.0003760907504363002,
"loss": 0.0071,
"step": 71500
},
{
"epoch": 125.6544502617801,
"grad_norm": 0.05435263365507126,
"learning_rate": 0.0003717277486910995,
"loss": 0.0077,
"step": 72000
},
{
"epoch": 126.52705061082024,
"grad_norm": 0.015180529095232487,
"learning_rate": 0.00036736474694589876,
"loss": 0.0072,
"step": 72500
},
{
"epoch": 127.39965095986038,
"grad_norm": 0.02177223190665245,
"learning_rate": 0.0003630017452006981,
"loss": 0.0074,
"step": 73000
},
{
"epoch": 128.27225130890054,
"grad_norm": 0.04897777736186981,
"learning_rate": 0.0003586387434554974,
"loss": 0.0076,
"step": 73500
},
{
"epoch": 129.14485165794068,
"grad_norm": 0.02429993264377117,
"learning_rate": 0.0003542757417102967,
"loss": 0.0067,
"step": 74000
},
{
"epoch": 130.0174520069808,
"grad_norm": 0.019401997327804565,
"learning_rate": 0.000349912739965096,
"loss": 0.0071,
"step": 74500
},
{
"epoch": 130.89005235602093,
"grad_norm": 0.05895571410655975,
"learning_rate": 0.00034554973821989527,
"loss": 0.0067,
"step": 75000
},
{
"epoch": 131.76265270506107,
"grad_norm": 0.08564560860395432,
"learning_rate": 0.0003411867364746946,
"loss": 0.0066,
"step": 75500
},
{
"epoch": 132.6352530541012,
"grad_norm": 0.0574815534055233,
"learning_rate": 0.0003368237347294939,
"loss": 0.0063,
"step": 76000
},
{
"epoch": 133.50785340314135,
"grad_norm": 0.08703868091106415,
"learning_rate": 0.0003324607329842932,
"loss": 0.0064,
"step": 76500
},
{
"epoch": 134.3804537521815,
"grad_norm": 0.1667858362197876,
"learning_rate": 0.00032809773123909246,
"loss": 0.0074,
"step": 77000
},
{
"epoch": 135.25305410122164,
"grad_norm": 0.07352814823389053,
"learning_rate": 0.00032373472949389184,
"loss": 0.0067,
"step": 77500
},
{
"epoch": 136.12565445026178,
"grad_norm": 0.10321661829948425,
"learning_rate": 0.0003193717277486911,
"loss": 0.0065,
"step": 78000
},
{
"epoch": 136.99825479930192,
"grad_norm": 0.046309106051921844,
"learning_rate": 0.0003150087260034904,
"loss": 0.0065,
"step": 78500
},
{
"epoch": 137.87085514834206,
"grad_norm": 0.014806479215621948,
"learning_rate": 0.0003106457242582897,
"loss": 0.0061,
"step": 79000
},
{
"epoch": 138.7434554973822,
"grad_norm": 0.041382819414138794,
"learning_rate": 0.000306282722513089,
"loss": 0.0059,
"step": 79500
},
{
"epoch": 139.61605584642234,
"grad_norm": 0.07493896782398224,
"learning_rate": 0.0003019197207678883,
"loss": 0.0061,
"step": 80000
},
{
"epoch": 140.48865619546248,
"grad_norm": 0.052245959639549255,
"learning_rate": 0.0002975567190226876,
"loss": 0.006,
"step": 80500
},
{
"epoch": 141.36125654450262,
"grad_norm": 0.10472971946001053,
"learning_rate": 0.0002931937172774869,
"loss": 0.0061,
"step": 81000
},
{
"epoch": 142.23385689354276,
"grad_norm": 0.033921029418706894,
"learning_rate": 0.0002888307155322862,
"loss": 0.0059,
"step": 81500
},
{
"epoch": 143.1064572425829,
"grad_norm": 0.06276967376470566,
"learning_rate": 0.00028446771378708553,
"loss": 0.0059,
"step": 82000
},
{
"epoch": 143.97905759162305,
"grad_norm": 0.022356705740094185,
"learning_rate": 0.0002801047120418848,
"loss": 0.0056,
"step": 82500
},
{
"epoch": 144.8516579406632,
"grad_norm": 0.008057367987930775,
"learning_rate": 0.0002757417102966841,
"loss": 0.0054,
"step": 83000
},
{
"epoch": 145.72425828970333,
"grad_norm": 0.00805575679987669,
"learning_rate": 0.00027137870855148345,
"loss": 0.0054,
"step": 83500
},
{
"epoch": 146.59685863874344,
"grad_norm": 0.07681386172771454,
"learning_rate": 0.0002670157068062827,
"loss": 0.0053,
"step": 84000
},
{
"epoch": 147.46945898778358,
"grad_norm": 0.041651804000139236,
"learning_rate": 0.00026265270506108205,
"loss": 0.0054,
"step": 84500
},
{
"epoch": 148.34205933682372,
"grad_norm": 0.09435189515352249,
"learning_rate": 0.0002582897033158813,
"loss": 0.0057,
"step": 85000
},
{
"epoch": 149.21465968586386,
"grad_norm": 0.02968364767730236,
"learning_rate": 0.00025392670157068064,
"loss": 0.0053,
"step": 85500
},
{
"epoch": 150.087260034904,
"grad_norm": 0.04307426139712334,
"learning_rate": 0.00024956369982547996,
"loss": 0.0053,
"step": 86000
},
{
"epoch": 150.95986038394415,
"grad_norm": 0.08466316014528275,
"learning_rate": 0.00024520069808027923,
"loss": 0.0051,
"step": 86500
},
{
"epoch": 151.8324607329843,
"grad_norm": 0.14644889533519745,
"learning_rate": 0.00024083769633507853,
"loss": 0.005,
"step": 87000
}
],
"logging_steps": 500,
"max_steps": 114600,
"num_input_tokens_seen": 0,
"num_train_epochs": 200,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.5133036050382848e+18,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}