|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 1109, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.009017132551848512, |
|
"grad_norm": 1.8547641038894653, |
|
"learning_rate": 9.00900900900901e-06, |
|
"loss": 0.9349, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.018034265103697024, |
|
"grad_norm": 1.347398281097412, |
|
"learning_rate": 1.801801801801802e-05, |
|
"loss": 0.7242, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.027051397655545536, |
|
"grad_norm": 1.2487881183624268, |
|
"learning_rate": 2.702702702702703e-05, |
|
"loss": 0.4507, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.03606853020739405, |
|
"grad_norm": 1.1477875709533691, |
|
"learning_rate": 3.603603603603604e-05, |
|
"loss": 0.33, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.04508566275924256, |
|
"grad_norm": 0.923848569393158, |
|
"learning_rate": 4.5045045045045046e-05, |
|
"loss": 0.234, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.05410279531109107, |
|
"grad_norm": 0.9260658025741577, |
|
"learning_rate": 5.405405405405406e-05, |
|
"loss": 0.2077, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.06311992786293959, |
|
"grad_norm": 1.255967378616333, |
|
"learning_rate": 6.306306306306306e-05, |
|
"loss": 0.1886, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.0721370604147881, |
|
"grad_norm": 0.7384198904037476, |
|
"learning_rate": 7.207207207207208e-05, |
|
"loss": 0.1645, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.0811541929666366, |
|
"grad_norm": 0.672892153263092, |
|
"learning_rate": 8.108108108108109e-05, |
|
"loss": 0.1591, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.09017132551848513, |
|
"grad_norm": 0.8307783007621765, |
|
"learning_rate": 9.009009009009009e-05, |
|
"loss": 0.1595, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.09918845807033363, |
|
"grad_norm": 0.7313665151596069, |
|
"learning_rate": 9.90990990990991e-05, |
|
"loss": 0.1406, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.10820559062218214, |
|
"grad_norm": 0.985645592212677, |
|
"learning_rate": 9.997993520894937e-05, |
|
"loss": 0.1385, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.11722272317403065, |
|
"grad_norm": 0.9960182905197144, |
|
"learning_rate": 9.99105961120544e-05, |
|
"loss": 0.137, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.12623985572587917, |
|
"grad_norm": 0.5437267422676086, |
|
"learning_rate": 9.979180368331558e-05, |
|
"loss": 0.1269, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.13525698827772767, |
|
"grad_norm": 0.5780126452445984, |
|
"learning_rate": 9.962367562682496e-05, |
|
"loss": 0.1282, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.1442741208295762, |
|
"grad_norm": 0.8550663590431213, |
|
"learning_rate": 9.940637853030572e-05, |
|
"loss": 0.1267, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.1532912533814247, |
|
"grad_norm": 0.5869548916816711, |
|
"learning_rate": 9.914012770005072e-05, |
|
"loss": 0.1224, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.1623083859332732, |
|
"grad_norm": 0.6128000617027283, |
|
"learning_rate": 9.882518694758875e-05, |
|
"loss": 0.1059, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.17132551848512173, |
|
"grad_norm": 0.852254331111908, |
|
"learning_rate": 9.846186832828989e-05, |
|
"loss": 0.0997, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.18034265103697025, |
|
"grad_norm": 0.6299287676811218, |
|
"learning_rate": 9.805053183216923e-05, |
|
"loss": 0.1038, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.18935978358881875, |
|
"grad_norm": 0.8325192928314209, |
|
"learning_rate": 9.759158502719481e-05, |
|
"loss": 0.1207, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.19837691614066727, |
|
"grad_norm": 0.46743109822273254, |
|
"learning_rate": 9.708548265545375e-05, |
|
"loss": 0.105, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.2073940486925158, |
|
"grad_norm": 0.6069552302360535, |
|
"learning_rate": 9.653272618257631e-05, |
|
"loss": 0.0968, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.2164111812443643, |
|
"grad_norm": 0.6297630071640015, |
|
"learning_rate": 9.593386330086458e-05, |
|
"loss": 0.0982, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.2254283137962128, |
|
"grad_norm": 0.574450671672821, |
|
"learning_rate": 9.528948738661784e-05, |
|
"loss": 0.0996, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.2344454463480613, |
|
"grad_norm": 0.8703469038009644, |
|
"learning_rate": 9.460023691219277e-05, |
|
"loss": 0.0919, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.24346257889990983, |
|
"grad_norm": 0.6118870377540588, |
|
"learning_rate": 9.386679481338033e-05, |
|
"loss": 0.094, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.25247971145175835, |
|
"grad_norm": 0.47032231092453003, |
|
"learning_rate": 9.308988781272694e-05, |
|
"loss": 0.0918, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.26149684400360684, |
|
"grad_norm": 0.5307496786117554, |
|
"learning_rate": 9.227028569946996e-05, |
|
"loss": 0.095, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.27051397655545534, |
|
"grad_norm": 0.37769177556037903, |
|
"learning_rate": 9.140880056680088e-05, |
|
"loss": 0.0892, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.2795311091073039, |
|
"grad_norm": 0.5687931776046753, |
|
"learning_rate": 9.050628600721234e-05, |
|
"loss": 0.0853, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.2885482416591524, |
|
"grad_norm": 0.6189451217651367, |
|
"learning_rate": 8.956363626672595e-05, |
|
"loss": 0.0838, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.2975653742110009, |
|
"grad_norm": 0.6863543391227722, |
|
"learning_rate": 8.858178535883905e-05, |
|
"loss": 0.0825, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.3065825067628494, |
|
"grad_norm": 0.8236618041992188, |
|
"learning_rate": 8.756170613906833e-05, |
|
"loss": 0.0832, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.3155996393146979, |
|
"grad_norm": 0.5764484405517578, |
|
"learning_rate": 8.650440934100728e-05, |
|
"loss": 0.0775, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.3246167718665464, |
|
"grad_norm": 0.40208104252815247, |
|
"learning_rate": 8.541094257485265e-05, |
|
"loss": 0.0751, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.33363390441839497, |
|
"grad_norm": 0.6935873031616211, |
|
"learning_rate": 8.428238928939207e-05, |
|
"loss": 0.0793, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.34265103697024346, |
|
"grad_norm": 0.5251246094703674, |
|
"learning_rate": 8.311986769848141e-05, |
|
"loss": 0.0713, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.35166816952209196, |
|
"grad_norm": 0.4371775686740875, |
|
"learning_rate": 8.192452967307576e-05, |
|
"loss": 0.0712, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.3606853020739405, |
|
"grad_norm": 0.6153225302696228, |
|
"learning_rate": 8.069755959991142e-05, |
|
"loss": 0.0694, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.369702434625789, |
|
"grad_norm": 0.4341914653778076, |
|
"learning_rate": 7.944017320797013e-05, |
|
"loss": 0.075, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.3787195671776375, |
|
"grad_norm": 0.4805607199668884, |
|
"learning_rate": 7.815361636388827e-05, |
|
"loss": 0.0665, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.38773669972948605, |
|
"grad_norm": 0.508466362953186, |
|
"learning_rate": 7.683916383750436e-05, |
|
"loss": 0.0692, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.39675383228133454, |
|
"grad_norm": 0.7154425978660583, |
|
"learning_rate": 7.549811803876825e-05, |
|
"loss": 0.0693, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.40577096483318303, |
|
"grad_norm": 0.3025909960269928, |
|
"learning_rate": 7.413180772726348e-05, |
|
"loss": 0.0721, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.4147880973850316, |
|
"grad_norm": 0.5098099708557129, |
|
"learning_rate": 7.274158669562126e-05, |
|
"loss": 0.0691, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.4238052299368801, |
|
"grad_norm": 0.31458744406700134, |
|
"learning_rate": 7.13288324281309e-05, |
|
"loss": 0.0668, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.4328223624887286, |
|
"grad_norm": 0.5931514501571655, |
|
"learning_rate": 6.989494473587554e-05, |
|
"loss": 0.0656, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.4418394950405771, |
|
"grad_norm": 0.3595449924468994, |
|
"learning_rate": 6.844134436974567e-05, |
|
"loss": 0.0646, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.4508566275924256, |
|
"grad_norm": 0.5599433779716492, |
|
"learning_rate": 6.696947161270476e-05, |
|
"loss": 0.0687, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.4598737601442741, |
|
"grad_norm": 0.4869447946548462, |
|
"learning_rate": 6.548078485270152e-05, |
|
"loss": 0.0683, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.4688908926961226, |
|
"grad_norm": 0.36777040362358093, |
|
"learning_rate": 6.397675913764347e-05, |
|
"loss": 0.0658, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.47790802524797116, |
|
"grad_norm": 0.5203920602798462, |
|
"learning_rate": 6.245888471386263e-05, |
|
"loss": 0.0627, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.48692515779981965, |
|
"grad_norm": 0.46625274419784546, |
|
"learning_rate": 6.0928665549522554e-05, |
|
"loss": 0.0622, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.49594229035166815, |
|
"grad_norm": 0.7451359629631042, |
|
"learning_rate": 5.9387617844429e-05, |
|
"loss": 0.0618, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.5049594229035167, |
|
"grad_norm": 0.392704576253891, |
|
"learning_rate": 5.78372685277209e-05, |
|
"loss": 0.0589, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.5139765554553652, |
|
"grad_norm": 0.44599130749702454, |
|
"learning_rate": 5.627915374493061e-05, |
|
"loss": 0.0615, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.5229936880072137, |
|
"grad_norm": 0.3163304328918457, |
|
"learning_rate": 5.4714817335911894e-05, |
|
"loss": 0.0503, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.5320108205590622, |
|
"grad_norm": 0.6041306257247925, |
|
"learning_rate": 5.314580930514431e-05, |
|
"loss": 0.0598, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.5410279531109107, |
|
"grad_norm": 0.38383838534355164, |
|
"learning_rate": 5.157368428592933e-05, |
|
"loss": 0.0589, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.5500450856627592, |
|
"grad_norm": 0.2501381039619446, |
|
"learning_rate": 5e-05, |
|
"loss": 0.049, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.5590622182146078, |
|
"grad_norm": 0.3422386348247528, |
|
"learning_rate": 4.8426315714070684e-05, |
|
"loss": 0.0538, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.5680793507664562, |
|
"grad_norm": 0.5101218223571777, |
|
"learning_rate": 4.6854190694855694e-05, |
|
"loss": 0.0497, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.5770964833183048, |
|
"grad_norm": 0.37502148747444153, |
|
"learning_rate": 4.528518266408811e-05, |
|
"loss": 0.052, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.5861136158701533, |
|
"grad_norm": 0.3856571912765503, |
|
"learning_rate": 4.3720846255069406e-05, |
|
"loss": 0.0568, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.5951307484220018, |
|
"grad_norm": 0.3386954069137573, |
|
"learning_rate": 4.21627314722791e-05, |
|
"loss": 0.0493, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.6041478809738503, |
|
"grad_norm": 0.28927159309387207, |
|
"learning_rate": 4.0612382155571026e-05, |
|
"loss": 0.0574, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.6131650135256989, |
|
"grad_norm": 0.3935062289237976, |
|
"learning_rate": 3.907133445047747e-05, |
|
"loss": 0.0434, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.6221821460775473, |
|
"grad_norm": 0.39534249901771545, |
|
"learning_rate": 3.75411152861374e-05, |
|
"loss": 0.0487, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.6311992786293958, |
|
"grad_norm": 0.35437795519828796, |
|
"learning_rate": 3.602324086235655e-05, |
|
"loss": 0.0512, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.6402164111812444, |
|
"grad_norm": 0.2841149866580963, |
|
"learning_rate": 3.451921514729848e-05, |
|
"loss": 0.0391, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.6492335437330928, |
|
"grad_norm": 0.3514391779899597, |
|
"learning_rate": 3.303052838729525e-05, |
|
"loss": 0.0484, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.6582506762849414, |
|
"grad_norm": 0.29207444190979004, |
|
"learning_rate": 3.155865563025433e-05, |
|
"loss": 0.0443, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.6672678088367899, |
|
"grad_norm": 0.27113544940948486, |
|
"learning_rate": 3.010505526412447e-05, |
|
"loss": 0.0446, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.6762849413886384, |
|
"grad_norm": 0.4424648880958557, |
|
"learning_rate": 2.867116757186911e-05, |
|
"loss": 0.0431, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.6853020739404869, |
|
"grad_norm": 0.4714156687259674, |
|
"learning_rate": 2.7258413304378734e-05, |
|
"loss": 0.0468, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.6943192064923355, |
|
"grad_norm": 0.3349123001098633, |
|
"learning_rate": 2.5868192272736514e-05, |
|
"loss": 0.0433, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.7033363390441839, |
|
"grad_norm": 0.5246461629867554, |
|
"learning_rate": 2.450188196123177e-05, |
|
"loss": 0.041, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.7123534715960325, |
|
"grad_norm": 0.37693607807159424, |
|
"learning_rate": 2.3160836162495653e-05, |
|
"loss": 0.0512, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.721370604147881, |
|
"grad_norm": 0.33574390411376953, |
|
"learning_rate": 2.1846383636111743e-05, |
|
"loss": 0.0452, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.7303877366997295, |
|
"grad_norm": 0.22386135160923004, |
|
"learning_rate": 2.0559826792029884e-05, |
|
"loss": 0.0424, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.739404869251578, |
|
"grad_norm": 0.40371736884117126, |
|
"learning_rate": 1.9302440400088606e-05, |
|
"loss": 0.0357, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.7484220018034266, |
|
"grad_norm": 0.4526534676551819, |
|
"learning_rate": 1.8075470326924243e-05, |
|
"loss": 0.0432, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.757439134355275, |
|
"grad_norm": 0.2903415858745575, |
|
"learning_rate": 1.6880132301518598e-05, |
|
"loss": 0.0375, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.7664562669071235, |
|
"grad_norm": 0.30989566445350647, |
|
"learning_rate": 1.5717610710607948e-05, |
|
"loss": 0.0377, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.7754733994589721, |
|
"grad_norm": 0.3357095420360565, |
|
"learning_rate": 1.458905742514734e-05, |
|
"loss": 0.0376, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.7844905320108205, |
|
"grad_norm": 0.47947272658348083, |
|
"learning_rate": 1.3495590658992718e-05, |
|
"loss": 0.037, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.7935076645626691, |
|
"grad_norm": 0.3474498391151428, |
|
"learning_rate": 1.2438293860931677e-05, |
|
"loss": 0.0397, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.8025247971145176, |
|
"grad_norm": 0.4065302610397339, |
|
"learning_rate": 1.1418214641160958e-05, |
|
"loss": 0.0378, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.8115419296663661, |
|
"grad_norm": 0.20412218570709229, |
|
"learning_rate": 1.0436363733274057e-05, |
|
"loss": 0.0333, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.8205590622182146, |
|
"grad_norm": 0.2562933564186096, |
|
"learning_rate": 9.493713992787672e-06, |
|
"loss": 0.0396, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.8295761947700632, |
|
"grad_norm": 0.3250836133956909, |
|
"learning_rate": 8.591199433199126e-06, |
|
"loss": 0.0351, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.8385933273219116, |
|
"grad_norm": 0.31816214323043823, |
|
"learning_rate": 7.72971430053005e-06, |
|
"loss": 0.043, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.8476104598737602, |
|
"grad_norm": 0.28158149123191833, |
|
"learning_rate": 6.910112187273066e-06, |
|
"loss": 0.0324, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.8566275924256087, |
|
"grad_norm": 0.34491991996765137, |
|
"learning_rate": 6.133205186619695e-06, |
|
"loss": 0.046, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.8656447249774571, |
|
"grad_norm": 0.3850158154964447, |
|
"learning_rate": 5.399763087807236e-06, |
|
"loss": 0.0339, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.8746618575293057, |
|
"grad_norm": 0.3283858597278595, |
|
"learning_rate": 4.710512613382151e-06, |
|
"loss": 0.0388, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.8836789900811542, |
|
"grad_norm": 0.4049375653266907, |
|
"learning_rate": 4.0661366991354365e-06, |
|
"loss": 0.0346, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.8926961226330027, |
|
"grad_norm": 0.43446460366249084, |
|
"learning_rate": 3.4672738174236884e-06, |
|
"loss": 0.0346, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.9017132551848512, |
|
"grad_norm": 0.2909717857837677, |
|
"learning_rate": 2.914517344546258e-06, |
|
"loss": 0.0372, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.9107303877366997, |
|
"grad_norm": 0.42221134901046753, |
|
"learning_rate": 2.4084149728051952e-06, |
|
"loss": 0.0317, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.9197475202885482, |
|
"grad_norm": 0.3318589925765991, |
|
"learning_rate": 1.9494681678307703e-06, |
|
"loss": 0.0343, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.9287646528403968, |
|
"grad_norm": 0.3418586552143097, |
|
"learning_rate": 1.538131671710108e-06, |
|
"loss": 0.0364, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.9377817853922452, |
|
"grad_norm": 0.2581607699394226, |
|
"learning_rate": 1.1748130524112666e-06, |
|
"loss": 0.0347, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.9467989179440938, |
|
"grad_norm": 0.37032172083854675, |
|
"learning_rate": 8.59872299949288e-07, |
|
"loss": 0.0402, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.9558160504959423, |
|
"grad_norm": 0.2215413898229599, |
|
"learning_rate": 5.936214696942887e-07, |
|
"loss": 0.0357, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.9648331830477908, |
|
"grad_norm": 0.30502429604530334, |
|
"learning_rate": 3.7632437317505207e-07, |
|
"loss": 0.0332, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.9738503155996393, |
|
"grad_norm": 0.24881534278392792, |
|
"learning_rate": 2.0819631668442253e-07, |
|
"loss": 0.033, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.9828674481514879, |
|
"grad_norm": 0.7615265846252441, |
|
"learning_rate": 8.940388794559939e-08, |
|
"loss": 0.0378, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.9918845807033363, |
|
"grad_norm": 0.4428829848766327, |
|
"learning_rate": 2.0064791050633526e-08, |
|
"loss": 0.0334, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 1109, |
|
"total_flos": 2.052344509146071e+17, |
|
"train_loss": 0.08958469805575578, |
|
"train_runtime": 1690.0157, |
|
"train_samples_per_second": 5.249, |
|
"train_steps_per_second": 0.656 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 1109, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.052344509146071e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|