|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.0, |
|
"eval_steps": 500, |
|
"global_step": 368, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 1359.2177158077461, |
|
"learning_rate": 2.702702702702703e-10, |
|
"logits/chosen": -1.3332719802856445, |
|
"logits/rejected": -1.246394395828247, |
|
"logps/chosen": -286.9539794921875, |
|
"logps/rejected": -263.3782958984375, |
|
"loss": 0.7007, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.0, |
|
"rewards/margins": 0.0, |
|
"rewards/rejected": 0.0, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 1659.9667867776882, |
|
"learning_rate": 2.702702702702703e-09, |
|
"logits/chosen": -1.6176425218582153, |
|
"logits/rejected": -1.3966516256332397, |
|
"logps/chosen": -342.49456787109375, |
|
"logps/rejected": -294.5242614746094, |
|
"loss": 0.7618, |
|
"rewards/accuracies": 0.4340277910232544, |
|
"rewards/chosen": 0.040204986929893494, |
|
"rewards/margins": 0.027825627475976944, |
|
"rewards/rejected": 0.012379361316561699, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 1365.2835589869765, |
|
"learning_rate": 5.405405405405406e-09, |
|
"logits/chosen": -1.4892847537994385, |
|
"logits/rejected": -1.311715006828308, |
|
"logps/chosen": -314.7313537597656, |
|
"logps/rejected": -279.33746337890625, |
|
"loss": 0.7414, |
|
"rewards/accuracies": 0.5406249761581421, |
|
"rewards/chosen": 0.03467293456196785, |
|
"rewards/margins": 0.06634848564863205, |
|
"rewards/rejected": -0.0316755548119545, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 1495.215254347967, |
|
"learning_rate": 8.108108108108109e-09, |
|
"logits/chosen": -1.5478875637054443, |
|
"logits/rejected": -1.3803514242172241, |
|
"logps/chosen": -324.8534240722656, |
|
"logps/rejected": -286.27276611328125, |
|
"loss": 0.7533, |
|
"rewards/accuracies": 0.5218750238418579, |
|
"rewards/chosen": 0.05265723541378975, |
|
"rewards/margins": 0.04645369574427605, |
|
"rewards/rejected": 0.006203538738191128, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 1385.2297985942057, |
|
"learning_rate": 9.997973265157192e-09, |
|
"logits/chosen": -1.5357484817504883, |
|
"logits/rejected": -1.3583762645721436, |
|
"logps/chosen": -325.4037780761719, |
|
"logps/rejected": -285.64508056640625, |
|
"loss": 0.7565, |
|
"rewards/accuracies": 0.4937500059604645, |
|
"rewards/chosen": -0.010429712943732738, |
|
"rewards/margins": -0.014903778210282326, |
|
"rewards/rejected": 0.0044740648008883, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 1517.4779086138956, |
|
"learning_rate": 9.961988113473708e-09, |
|
"logits/chosen": -1.5413590669631958, |
|
"logits/rejected": -1.3948113918304443, |
|
"logps/chosen": -337.04327392578125, |
|
"logps/rejected": -297.2876892089844, |
|
"loss": 0.7454, |
|
"rewards/accuracies": 0.47187501192092896, |
|
"rewards/chosen": -0.022386690601706505, |
|
"rewards/margins": -0.045912474393844604, |
|
"rewards/rejected": 0.0235257837921381, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 1317.1513962511742, |
|
"learning_rate": 9.881337335184878e-09, |
|
"logits/chosen": -1.5795971155166626, |
|
"logits/rejected": -1.43094801902771, |
|
"logps/chosen": -319.79681396484375, |
|
"logps/rejected": -285.0555725097656, |
|
"loss": 0.6982, |
|
"rewards/accuracies": 0.606249988079071, |
|
"rewards/chosen": 0.046097733080387115, |
|
"rewards/margins": 0.21919748187065125, |
|
"rewards/rejected": -0.17309975624084473, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 1446.8726182465537, |
|
"learning_rate": 9.756746912994832e-09, |
|
"logits/chosen": -1.5057274103164673, |
|
"logits/rejected": -1.343697190284729, |
|
"logps/chosen": -312.12432861328125, |
|
"logps/rejected": -275.1330871582031, |
|
"loss": 0.6973, |
|
"rewards/accuracies": 0.5874999761581421, |
|
"rewards/chosen": -0.016833370551466942, |
|
"rewards/margins": 0.13973814249038696, |
|
"rewards/rejected": -0.15657152235507965, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 1216.5073312303066, |
|
"learning_rate": 9.589338354885628e-09, |
|
"logits/chosen": -1.5987694263458252, |
|
"logits/rejected": -1.4453301429748535, |
|
"logps/chosen": -323.3362731933594, |
|
"logps/rejected": -288.1485595703125, |
|
"loss": 0.6695, |
|
"rewards/accuracies": 0.565625011920929, |
|
"rewards/chosen": 0.04707593470811844, |
|
"rewards/margins": 0.22710590064525604, |
|
"rewards/rejected": -0.180029958486557, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 1161.721503342428, |
|
"learning_rate": 9.380618598797472e-09, |
|
"logits/chosen": -1.6032607555389404, |
|
"logits/rejected": -1.4058634042739868, |
|
"logps/chosen": -319.96173095703125, |
|
"logps/rejected": -281.8211364746094, |
|
"loss": 0.6501, |
|
"rewards/accuracies": 0.628125011920929, |
|
"rewards/chosen": 0.11853907257318497, |
|
"rewards/margins": 0.32674410939216614, |
|
"rewards/rejected": -0.20820502936840057, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 1255.206435160775, |
|
"learning_rate": 9.132466447838596e-09, |
|
"logits/chosen": -1.5419257879257202, |
|
"logits/rejected": -1.3666443824768066, |
|
"logps/chosen": -321.9004821777344, |
|
"logps/rejected": -282.74346923828125, |
|
"loss": 0.6241, |
|
"rewards/accuracies": 0.7437499761581421, |
|
"rewards/chosen": 0.21292057633399963, |
|
"rewards/margins": 0.4663107991218567, |
|
"rewards/rejected": -0.25339022278785706, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 1196.6092743226689, |
|
"learning_rate": 8.847115658129039e-09, |
|
"logits/chosen": -1.50592839717865, |
|
"logits/rejected": -1.3777477741241455, |
|
"logps/chosen": -318.1675720214844, |
|
"logps/rejected": -287.3236083984375, |
|
"loss": 0.6068, |
|
"rewards/accuracies": 0.6499999761581421, |
|
"rewards/chosen": 0.17643409967422485, |
|
"rewards/margins": 0.4380553662776947, |
|
"rewards/rejected": -0.26162129640579224, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 1069.9581355208816, |
|
"learning_rate": 8.527134831514116e-09, |
|
"logits/chosen": -1.5797988176345825, |
|
"logits/rejected": -1.4248679876327515, |
|
"logps/chosen": -331.4310302734375, |
|
"logps/rejected": -297.9786682128906, |
|
"loss": 0.613, |
|
"rewards/accuracies": 0.6187499761581421, |
|
"rewards/chosen": 0.19524307548999786, |
|
"rewards/margins": 0.3385895788669586, |
|
"rewards/rejected": -0.14334650337696075, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"grad_norm": 1001.0236709718766, |
|
"learning_rate": 8.175404294144481e-09, |
|
"logits/chosen": -1.6208512783050537, |
|
"logits/rejected": -1.4344335794448853, |
|
"logps/chosen": -317.17388916015625, |
|
"logps/rejected": -271.62255859375, |
|
"loss": 0.5838, |
|
"rewards/accuracies": 0.6499999761581421, |
|
"rewards/chosen": 0.3348919153213501, |
|
"rewards/margins": 0.4819467067718506, |
|
"rewards/rejected": -0.14705480635166168, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 1056.0149896521477, |
|
"learning_rate": 7.79509016905158e-09, |
|
"logits/chosen": -1.566674828529358, |
|
"logits/rejected": -1.417965292930603, |
|
"logps/chosen": -331.0799255371094, |
|
"logps/rejected": -294.3246154785156, |
|
"loss": 0.5738, |
|
"rewards/accuracies": 0.7281249761581421, |
|
"rewards/chosen": 0.5121167898178101, |
|
"rewards/margins": 0.6085957288742065, |
|
"rewards/rejected": -0.09647894650697708, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 1085.2712824812697, |
|
"learning_rate": 7.389615876105773e-09, |
|
"logits/chosen": -1.5494105815887451, |
|
"logits/rejected": -1.4207683801651, |
|
"logps/chosen": -314.55694580078125, |
|
"logps/rejected": -291.8951110839844, |
|
"loss": 0.5814, |
|
"rewards/accuracies": 0.6656249761581421, |
|
"rewards/chosen": 0.5086601972579956, |
|
"rewards/margins": 0.5659324526786804, |
|
"rewards/rejected": -0.057272158563137054, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 1147.2094257934837, |
|
"learning_rate": 6.962631315901861e-09, |
|
"logits/chosen": -1.5222028493881226, |
|
"logits/rejected": -1.406696081161499, |
|
"logps/chosen": -317.96502685546875, |
|
"logps/rejected": -291.0884704589844, |
|
"loss": 0.5667, |
|
"rewards/accuracies": 0.675000011920929, |
|
"rewards/chosen": 0.5517348051071167, |
|
"rewards/margins": 0.5213624835014343, |
|
"rewards/rejected": 0.030372310429811478, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 1097.0692572479243, |
|
"learning_rate": 6.517980014965139e-09, |
|
"logits/chosen": -1.5993129014968872, |
|
"logits/rejected": -1.41109299659729, |
|
"logps/chosen": -331.37066650390625, |
|
"logps/rejected": -289.573486328125, |
|
"loss": 0.5518, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": 0.6186414957046509, |
|
"rewards/margins": 0.6974190473556519, |
|
"rewards/rejected": -0.0787774994969368, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 1005.7280013223308, |
|
"learning_rate": 6.059664528022266e-09, |
|
"logits/chosen": -1.6032111644744873, |
|
"logits/rejected": -1.4532817602157593, |
|
"logps/chosen": -315.0304260253906, |
|
"logps/rejected": -276.7928771972656, |
|
"loss": 0.5485, |
|
"rewards/accuracies": 0.746874988079071, |
|
"rewards/chosen": 0.6357426047325134, |
|
"rewards/margins": 0.7282391786575317, |
|
"rewards/rejected": -0.09249657392501831, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 1081.11663192182, |
|
"learning_rate": 5.591810408770492e-09, |
|
"logits/chosen": -1.5532522201538086, |
|
"logits/rejected": -1.3794432878494263, |
|
"logps/chosen": -315.5338439941406, |
|
"logps/rejected": -278.7769470214844, |
|
"loss": 0.5324, |
|
"rewards/accuracies": 0.7124999761581421, |
|
"rewards/chosen": 0.6404975056648254, |
|
"rewards/margins": 0.7475987076759338, |
|
"rewards/rejected": -0.10710116475820541, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 994.8740373609369, |
|
"learning_rate": 5.118629073464423e-09, |
|
"logits/chosen": -1.5689371824264526, |
|
"logits/rejected": -1.3581821918487549, |
|
"logps/chosen": -325.8009033203125, |
|
"logps/rejected": -282.7605895996094, |
|
"loss": 0.5279, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": 0.8169169425964355, |
|
"rewards/margins": 0.8118550181388855, |
|
"rewards/rejected": 0.005061971955001354, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 1046.062764450064, |
|
"learning_rate": 4.644379891605983e-09, |
|
"logits/chosen": -1.6105190515518188, |
|
"logits/rejected": -1.4331891536712646, |
|
"logps/chosen": -324.65234375, |
|
"logps/rejected": -291.387451171875, |
|
"loss": 0.5257, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.7297837138175964, |
|
"rewards/margins": 0.761337399482727, |
|
"rewards/rejected": -0.03155365586280823, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 1062.5909355592705, |
|
"learning_rate": 4.173331844980362e-09, |
|
"logits/chosen": -1.542718529701233, |
|
"logits/rejected": -1.4185158014297485, |
|
"logps/chosen": -323.8753356933594, |
|
"logps/rejected": -293.48626708984375, |
|
"loss": 0.5141, |
|
"rewards/accuracies": 0.731249988079071, |
|
"rewards/chosen": 0.7222188711166382, |
|
"rewards/margins": 0.7484906911849976, |
|
"rewards/rejected": -0.026271820068359375, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 1054.001314243816, |
|
"learning_rate": 3.7097251001664824e-09, |
|
"logits/chosen": -1.5353752374649048, |
|
"logits/rejected": -1.3762162923812866, |
|
"logps/chosen": -323.80047607421875, |
|
"logps/rejected": -287.00726318359375, |
|
"loss": 0.5117, |
|
"rewards/accuracies": 0.715624988079071, |
|
"rewards/chosen": 0.8148177266120911, |
|
"rewards/margins": 0.8335307240486145, |
|
"rewards/rejected": -0.018712949007749557, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"grad_norm": 1057.7644150535455, |
|
"learning_rate": 3.2577328404292057e-09, |
|
"logits/chosen": -1.5480402708053589, |
|
"logits/rejected": -1.4182308912277222, |
|
"logps/chosen": -312.3599548339844, |
|
"logps/rejected": -285.9615478515625, |
|
"loss": 0.5004, |
|
"rewards/accuracies": 0.7437499761581421, |
|
"rewards/chosen": 0.8805425763130188, |
|
"rewards/margins": 0.8217647671699524, |
|
"rewards/rejected": 0.05877774953842163, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"grad_norm": 1134.7078580204877, |
|
"learning_rate": 2.821423700565763e-09, |
|
"logits/chosen": -1.597772479057312, |
|
"logits/rejected": -1.4192153215408325, |
|
"logps/chosen": -350.49432373046875, |
|
"logps/rejected": -306.66290283203125, |
|
"loss": 0.497, |
|
"rewards/accuracies": 0.8031250238418579, |
|
"rewards/chosen": 1.035740613937378, |
|
"rewards/margins": 1.0768238306045532, |
|
"rewards/rejected": -0.04108327627182007, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"grad_norm": 1065.7113049843235, |
|
"learning_rate": 2.4047251428513483e-09, |
|
"logits/chosen": -1.615321397781372, |
|
"logits/rejected": -1.4609696865081787, |
|
"logps/chosen": -325.0920104980469, |
|
"logps/rejected": -291.08197021484375, |
|
"loss": 0.513, |
|
"rewards/accuracies": 0.7406250238418579, |
|
"rewards/chosen": 0.9730992317199707, |
|
"rewards/margins": 0.9078540802001953, |
|
"rewards/rejected": 0.06524516642093658, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 864.6386915761193, |
|
"learning_rate": 2.011388103757442e-09, |
|
"logits/chosen": -1.5269956588745117, |
|
"logits/rejected": -1.3828452825546265, |
|
"logps/chosen": -316.21124267578125, |
|
"logps/rejected": -285.77667236328125, |
|
"loss": 0.4948, |
|
"rewards/accuracies": 0.737500011920929, |
|
"rewards/chosen": 0.9972602725028992, |
|
"rewards/margins": 0.9109752774238586, |
|
"rewards/rejected": 0.08628497272729874, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 990.2860964249656, |
|
"learning_rate": 1.644953229677474e-09, |
|
"logits/chosen": -1.601299524307251, |
|
"logits/rejected": -1.4185166358947754, |
|
"logps/chosen": -325.8226623535156, |
|
"logps/rejected": -284.75567626953125, |
|
"loss": 0.5046, |
|
"rewards/accuracies": 0.7906249761581421, |
|
"rewards/chosen": 1.0880569219589233, |
|
"rewards/margins": 1.0006572008132935, |
|
"rewards/rejected": 0.08739979565143585, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"grad_norm": 1035.2441434858022, |
|
"learning_rate": 1.308719005590957e-09, |
|
"logits/chosen": -1.5150429010391235, |
|
"logits/rejected": -1.400657057762146, |
|
"logps/chosen": -318.3056640625, |
|
"logps/rejected": -282.54803466796875, |
|
"loss": 0.5012, |
|
"rewards/accuracies": 0.7562500238418579, |
|
"rewards/chosen": 0.9477123022079468, |
|
"rewards/margins": 0.9480409622192383, |
|
"rewards/rejected": -0.00032869577989913523, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 952.917065524399, |
|
"learning_rate": 1.005712063557776e-09, |
|
"logits/chosen": -1.6322643756866455, |
|
"logits/rejected": -1.4532310962677002, |
|
"logps/chosen": -323.9804382324219, |
|
"logps/rejected": -290.63629150390625, |
|
"loss": 0.505, |
|
"rewards/accuracies": 0.731249988079071, |
|
"rewards/chosen": 0.903466522693634, |
|
"rewards/margins": 0.8820658922195435, |
|
"rewards/rejected": 0.02140064910054207, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 946.6373160602382, |
|
"learning_rate": 7.386599383124321e-10, |
|
"logits/chosen": -1.5624831914901733, |
|
"logits/rejected": -1.3783166408538818, |
|
"logps/chosen": -321.67547607421875, |
|
"logps/rejected": -285.8205261230469, |
|
"loss": 0.5001, |
|
"rewards/accuracies": 0.7562500238418579, |
|
"rewards/chosen": 0.9382842779159546, |
|
"rewards/margins": 0.9216762781143188, |
|
"rewards/rejected": 0.01660792902112007, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 995.8621536271032, |
|
"learning_rate": 5.099665152003929e-10, |
|
"logits/chosen": -1.5874156951904297, |
|
"logits/rejected": -1.3745605945587158, |
|
"logps/chosen": -333.58453369140625, |
|
"logps/rejected": -289.9964294433594, |
|
"loss": 0.4994, |
|
"rewards/accuracies": 0.7749999761581421, |
|
"rewards/chosen": 1.030099868774414, |
|
"rewards/margins": 1.0678800344467163, |
|
"rewards/rejected": -0.03778017312288284, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 920.3253000772652, |
|
"learning_rate": 3.216903914633745e-10, |
|
"logits/chosen": -1.5468734502792358, |
|
"logits/rejected": -1.4234826564788818, |
|
"logps/chosen": -325.117919921875, |
|
"logps/rejected": -296.07086181640625, |
|
"loss": 0.5015, |
|
"rewards/accuracies": 0.737500011920929, |
|
"rewards/chosen": 0.9095224142074585, |
|
"rewards/margins": 0.8073747754096985, |
|
"rewards/rejected": 0.1021476536989212, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 988.4564193799931, |
|
"learning_rate": 1.7552634565570324e-10, |
|
"logits/chosen": -1.5551540851593018, |
|
"logits/rejected": -1.3869084119796753, |
|
"logps/chosen": -329.73553466796875, |
|
"logps/rejected": -292.8804931640625, |
|
"loss": 0.4969, |
|
"rewards/accuracies": 0.7749999761581421, |
|
"rewards/chosen": 1.0978221893310547, |
|
"rewards/margins": 1.0354284048080444, |
|
"rewards/rejected": 0.0623936764895916, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 978.2289689829179, |
|
"learning_rate": 7.279008199590543e-11, |
|
"logits/chosen": -1.5398132801055908, |
|
"logits/rejected": -1.3765273094177246, |
|
"logps/chosen": -326.29425048828125, |
|
"logps/rejected": -292.0558166503906, |
|
"loss": 0.4951, |
|
"rewards/accuracies": 0.762499988079071, |
|
"rewards/chosen": 1.0201548337936401, |
|
"rewards/margins": 1.052316665649414, |
|
"rewards/rejected": -0.03216180205345154, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 921.2000894790062, |
|
"learning_rate": 1.4406386978128017e-11, |
|
"logits/chosen": -1.629494309425354, |
|
"logits/rejected": -1.434570074081421, |
|
"logps/chosen": -330.90911865234375, |
|
"logps/rejected": -291.73638916015625, |
|
"loss": 0.4873, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": 1.1673272848129272, |
|
"rewards/margins": 1.1191251277923584, |
|
"rewards/rejected": 0.048202164471149445, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"step": 368, |
|
"total_flos": 0.0, |
|
"train_loss": 0.577453197668428, |
|
"train_runtime": 9953.5163, |
|
"train_samples_per_second": 9.463, |
|
"train_steps_per_second": 0.037 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 368, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 100, |
|
"total_flos": 0.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|