|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.935251798561151, |
|
"eval_steps": 500, |
|
"global_step": 102, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.05755395683453238, |
|
"grad_norm": 12.61336898803711, |
|
"learning_rate": 9.090909090909091e-07, |
|
"logits/chosen": -1.8628841638565063, |
|
"logits/rejected": -2.055020332336426, |
|
"logps/chosen": -171.465087890625, |
|
"logps/rejected": -237.2239532470703, |
|
"loss": 0.6931, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.0, |
|
"rewards/margins": 0.0, |
|
"rewards/rejected": 0.0, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.11510791366906475, |
|
"grad_norm": 13.303223609924316, |
|
"learning_rate": 1.8181818181818183e-06, |
|
"logits/chosen": -1.77495539188385, |
|
"logits/rejected": -2.0203144550323486, |
|
"logps/chosen": -174.85792541503906, |
|
"logps/rejected": -223.60960388183594, |
|
"loss": 0.7202, |
|
"rewards/accuracies": 0.421875, |
|
"rewards/chosen": 0.027576472610235214, |
|
"rewards/margins": -0.017962537705898285, |
|
"rewards/rejected": 0.0455390103161335, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.17266187050359713, |
|
"grad_norm": 13.142106056213379, |
|
"learning_rate": 2.7272727272727272e-06, |
|
"logits/chosen": -1.6846373081207275, |
|
"logits/rejected": -1.8930776119232178, |
|
"logps/chosen": -148.6173553466797, |
|
"logps/rejected": -215.2635498046875, |
|
"loss": 0.6931, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": 0.01209343783557415, |
|
"rewards/margins": 0.026968184858560562, |
|
"rewards/rejected": -0.014874744229018688, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.2302158273381295, |
|
"grad_norm": 14.098532676696777, |
|
"learning_rate": 3.6363636363636366e-06, |
|
"logits/chosen": -1.664399266242981, |
|
"logits/rejected": -1.8190324306488037, |
|
"logps/chosen": -162.35986328125, |
|
"logps/rejected": -209.52432250976562, |
|
"loss": 0.7227, |
|
"rewards/accuracies": 0.453125, |
|
"rewards/chosen": -0.0490722618997097, |
|
"rewards/margins": -0.019629262387752533, |
|
"rewards/rejected": -0.029443001374602318, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.28776978417266186, |
|
"grad_norm": 17.35445213317871, |
|
"learning_rate": 4.5454545454545455e-06, |
|
"logits/chosen": -1.794135332107544, |
|
"logits/rejected": -2.1967906951904297, |
|
"logps/chosen": -175.1381378173828, |
|
"logps/rejected": -199.35000610351562, |
|
"loss": 0.7281, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": -0.0006946558132767677, |
|
"rewards/margins": -0.02462945692241192, |
|
"rewards/rejected": 0.023934796452522278, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.34532374100719426, |
|
"grad_norm": 15.025739669799805, |
|
"learning_rate": 4.998510351377676e-06, |
|
"logits/chosen": -1.6492798328399658, |
|
"logits/rejected": -1.9510689973831177, |
|
"logps/chosen": -168.0927276611328, |
|
"logps/rejected": -215.88006591796875, |
|
"loss": 0.6629, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.04117076098918915, |
|
"rewards/margins": 0.11135232448577881, |
|
"rewards/rejected": -0.07018155604600906, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.4028776978417266, |
|
"grad_norm": 13.217601776123047, |
|
"learning_rate": 4.986603811737982e-06, |
|
"logits/chosen": -1.6234389543533325, |
|
"logits/rejected": -1.7974214553833008, |
|
"logps/chosen": -181.07872009277344, |
|
"logps/rejected": -224.6402130126953, |
|
"loss": 0.6747, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": 0.05587499216198921, |
|
"rewards/margins": 0.07424326241016388, |
|
"rewards/rejected": -0.018368273973464966, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.460431654676259, |
|
"grad_norm": 13.030729293823242, |
|
"learning_rate": 4.9628474725421845e-06, |
|
"logits/chosen": -1.7801084518432617, |
|
"logits/rejected": -2.158487319946289, |
|
"logps/chosen": -183.85427856445312, |
|
"logps/rejected": -220.0618133544922, |
|
"loss": 0.6627, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.03484422340989113, |
|
"rewards/margins": 0.11216042935848236, |
|
"rewards/rejected": -0.14700466394424438, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.5179856115107914, |
|
"grad_norm": 10.820516586303711, |
|
"learning_rate": 4.927354543565131e-06, |
|
"logits/chosen": -1.6137356758117676, |
|
"logits/rejected": -1.8953951597213745, |
|
"logps/chosen": -162.70101928710938, |
|
"logps/rejected": -197.60382080078125, |
|
"loss": 0.6371, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": 0.11542239040136337, |
|
"rewards/margins": 0.1481863409280777, |
|
"rewards/rejected": -0.03276393562555313, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.5755395683453237, |
|
"grad_norm": 13.496746063232422, |
|
"learning_rate": 4.880294164776785e-06, |
|
"logits/chosen": -1.5672615766525269, |
|
"logits/rejected": -1.799597978591919, |
|
"logps/chosen": -166.3477020263672, |
|
"logps/rejected": -206.35385131835938, |
|
"loss": 0.6767, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.03490065410733223, |
|
"rewards/margins": 0.07657002657651901, |
|
"rewards/rejected": -0.041669368743896484, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.6330935251798561, |
|
"grad_norm": 13.735678672790527, |
|
"learning_rate": 4.821890600313256e-06, |
|
"logits/chosen": -1.8001123666763306, |
|
"logits/rejected": -2.105792284011841, |
|
"logps/chosen": -183.22373962402344, |
|
"logps/rejected": -225.92710876464844, |
|
"loss": 0.6326, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.05107605457305908, |
|
"rewards/margins": 0.1594647318124771, |
|
"rewards/rejected": -0.10838868468999863, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.6906474820143885, |
|
"grad_norm": 11.953166961669922, |
|
"learning_rate": 4.752422169756048e-06, |
|
"logits/chosen": -1.9147757291793823, |
|
"logits/rejected": -2.2474513053894043, |
|
"logps/chosen": -181.51597595214844, |
|
"logps/rejected": -195.81033325195312, |
|
"loss": 0.6138, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": 0.13331472873687744, |
|
"rewards/margins": 0.21828582882881165, |
|
"rewards/rejected": -0.0849711000919342, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.7482014388489209, |
|
"grad_norm": 10.812925338745117, |
|
"learning_rate": 4.672219921812517e-06, |
|
"logits/chosen": -2.0098390579223633, |
|
"logits/rejected": -2.057783365249634, |
|
"logps/chosen": -166.85260009765625, |
|
"logps/rejected": -219.13427734375, |
|
"loss": 0.6185, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": 0.12382392585277557, |
|
"rewards/margins": 0.23543882369995117, |
|
"rewards/rejected": -0.11161486804485321, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.8057553956834532, |
|
"grad_norm": 11.20254135131836, |
|
"learning_rate": 4.581666056718016e-06, |
|
"logits/chosen": -1.809260606765747, |
|
"logits/rejected": -1.9223175048828125, |
|
"logps/chosen": -194.20700073242188, |
|
"logps/rejected": -236.03829956054688, |
|
"loss": 0.6222, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": 0.13630633056163788, |
|
"rewards/margins": 0.2403251975774765, |
|
"rewards/rejected": -0.10401885211467743, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.8633093525179856, |
|
"grad_norm": 11.394420623779297, |
|
"learning_rate": 4.481192104877727e-06, |
|
"logits/chosen": -1.6475874185562134, |
|
"logits/rejected": -2.0424444675445557, |
|
"logps/chosen": -177.6649169921875, |
|
"logps/rejected": -217.6787872314453, |
|
"loss": 0.4954, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.23437190055847168, |
|
"rewards/margins": 0.5494118928909302, |
|
"rewards/rejected": -0.3150399923324585, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.920863309352518, |
|
"grad_norm": 13.055252075195312, |
|
"learning_rate": 4.3712768704277535e-06, |
|
"logits/chosen": -1.9410960674285889, |
|
"logits/rejected": -2.216390371322632, |
|
"logps/chosen": -178.47877502441406, |
|
"logps/rejected": -221.4510498046875, |
|
"loss": 0.4999, |
|
"rewards/accuracies": 0.84375, |
|
"rewards/chosen": 0.19329071044921875, |
|
"rewards/margins": 0.5714992880821228, |
|
"rewards/rejected": -0.37820857763290405, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.9784172661870504, |
|
"grad_norm": 8.89469051361084, |
|
"learning_rate": 4.252444149515374e-06, |
|
"logits/chosen": -1.7217860221862793, |
|
"logits/rejected": -1.8292845487594604, |
|
"logps/chosen": -166.56455993652344, |
|
"logps/rejected": -224.22413635253906, |
|
"loss": 0.4695, |
|
"rewards/accuracies": 0.828125, |
|
"rewards/chosen": 0.21696212887763977, |
|
"rewards/margins": 0.6354631781578064, |
|
"rewards/rejected": -0.41850101947784424, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 1.0359712230215827, |
|
"grad_norm": 9.03423023223877, |
|
"learning_rate": 4.125260234171861e-06, |
|
"logits/chosen": -1.7452164888381958, |
|
"logits/rejected": -1.9169570207595825, |
|
"logps/chosen": -176.8074951171875, |
|
"logps/rejected": -240.5430145263672, |
|
"loss": 0.4663, |
|
"rewards/accuracies": 0.84375, |
|
"rewards/chosen": 0.1004013791680336, |
|
"rewards/margins": 0.6946967840194702, |
|
"rewards/rejected": -0.5942954421043396, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 1.0935251798561152, |
|
"grad_norm": 8.077057838439941, |
|
"learning_rate": 3.990331213673064e-06, |
|
"logits/chosen": -1.7885024547576904, |
|
"logits/rejected": -2.0291481018066406, |
|
"logps/chosen": -182.01747131347656, |
|
"logps/rejected": -220.93408203125, |
|
"loss": 0.4458, |
|
"rewards/accuracies": 0.78125, |
|
"rewards/chosen": 0.14817766845226288, |
|
"rewards/margins": 0.8059278726577759, |
|
"rewards/rejected": -0.6577501893043518, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 1.1510791366906474, |
|
"grad_norm": 7.7763776779174805, |
|
"learning_rate": 3.848300086247998e-06, |
|
"logits/chosen": -1.771430492401123, |
|
"logits/rejected": -1.8969953060150146, |
|
"logps/chosen": -169.00811767578125, |
|
"logps/rejected": -216.32608032226562, |
|
"loss": 0.3395, |
|
"rewards/accuracies": 0.921875, |
|
"rewards/chosen": 0.42292630672454834, |
|
"rewards/margins": 1.200685739517212, |
|
"rewards/rejected": -0.7777595520019531, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 1.20863309352518, |
|
"grad_norm": 9.472132682800293, |
|
"learning_rate": 3.6998436948994664e-06, |
|
"logits/chosen": -1.7478959560394287, |
|
"logits/rejected": -1.946545124053955, |
|
"logps/chosen": -154.75123596191406, |
|
"logps/rejected": -207.8878631591797, |
|
"loss": 0.3934, |
|
"rewards/accuracies": 0.890625, |
|
"rewards/chosen": 0.17304691672325134, |
|
"rewards/margins": 0.9744219779968262, |
|
"rewards/rejected": -0.8013750314712524, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 1.2661870503597124, |
|
"grad_norm": 8.541696548461914, |
|
"learning_rate": 3.545669501938913e-06, |
|
"logits/chosen": -1.6284449100494385, |
|
"logits/rejected": -1.819665551185608, |
|
"logps/chosen": -177.72425842285156, |
|
"logps/rejected": -221.01028442382812, |
|
"loss": 0.4418, |
|
"rewards/accuracies": 0.796875, |
|
"rewards/chosen": 0.3692091405391693, |
|
"rewards/margins": 0.9766747951507568, |
|
"rewards/rejected": -0.6074656248092651, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 1.3237410071942446, |
|
"grad_norm": 6.5243048667907715, |
|
"learning_rate": 3.386512217606339e-06, |
|
"logits/chosen": -1.5771324634552002, |
|
"logits/rejected": -1.7775179147720337, |
|
"logps/chosen": -170.24432373046875, |
|
"logps/rejected": -210.37942504882812, |
|
"loss": 0.35, |
|
"rewards/accuracies": 0.859375, |
|
"rewards/chosen": 0.14126329123973846, |
|
"rewards/margins": 1.228642225265503, |
|
"rewards/rejected": -1.087378978729248, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 1.381294964028777, |
|
"grad_norm": 7.318809509277344, |
|
"learning_rate": 3.2231302988414198e-06, |
|
"logits/chosen": -1.797095775604248, |
|
"logits/rejected": -2.164658784866333, |
|
"logps/chosen": -172.25515747070312, |
|
"logps/rejected": -195.97921752929688, |
|
"loss": 0.3598, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.2658257484436035, |
|
"rewards/margins": 1.1805520057678223, |
|
"rewards/rejected": -0.914726197719574, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 1.4388489208633093, |
|
"grad_norm": 7.181938171386719, |
|
"learning_rate": 3.056302334890786e-06, |
|
"logits/chosen": -1.7573457956314087, |
|
"logits/rejected": -2.1520862579345703, |
|
"logps/chosen": -170.33551025390625, |
|
"logps/rejected": -202.55755615234375, |
|
"loss": 0.3145, |
|
"rewards/accuracies": 0.90625, |
|
"rewards/chosen": 0.3353237211704254, |
|
"rewards/margins": 1.4618067741394043, |
|
"rewards/rejected": -1.1264830827713013, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.4964028776978417, |
|
"grad_norm": 10.611798286437988, |
|
"learning_rate": 2.886823336975703e-06, |
|
"logits/chosen": -1.9649319648742676, |
|
"logits/rejected": -2.0540201663970947, |
|
"logps/chosen": -172.51666259765625, |
|
"logps/rejected": -220.8997039794922, |
|
"loss": 0.3824, |
|
"rewards/accuracies": 0.828125, |
|
"rewards/chosen": 0.0204707570374012, |
|
"rewards/margins": 1.466935396194458, |
|
"rewards/rejected": -1.4464645385742188, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 1.5539568345323742, |
|
"grad_norm": 7.461566925048828, |
|
"learning_rate": 2.7155009497015487e-06, |
|
"logits/chosen": -1.7212716341018677, |
|
"logits/rejected": -1.9288042783737183, |
|
"logps/chosen": -161.3657989501953, |
|
"logps/rejected": -203.9424285888672, |
|
"loss": 0.3495, |
|
"rewards/accuracies": 0.890625, |
|
"rewards/chosen": 0.2896910011768341, |
|
"rewards/margins": 1.4831678867340088, |
|
"rewards/rejected": -1.193476915359497, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 1.6115107913669064, |
|
"grad_norm": 6.363231658935547, |
|
"learning_rate": 2.5431516022634718e-06, |
|
"logits/chosen": -1.7244608402252197, |
|
"logits/rejected": -1.8467546701431274, |
|
"logps/chosen": -169.42726135253906, |
|
"logps/rejected": -238.03623962402344, |
|
"loss": 0.3193, |
|
"rewards/accuracies": 0.84375, |
|
"rewards/chosen": 0.5018333792686462, |
|
"rewards/margins": 1.5871676206588745, |
|
"rewards/rejected": -1.085334300994873, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 1.6690647482014387, |
|
"grad_norm": 8.272695541381836, |
|
"learning_rate": 2.3705966177894763e-06, |
|
"logits/chosen": -1.8467073440551758, |
|
"logits/rejected": -2.1156005859375, |
|
"logps/chosen": -169.63739013671875, |
|
"logps/rejected": -229.26177978515625, |
|
"loss": 0.3311, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": 0.4218375086784363, |
|
"rewards/margins": 1.7388478517532349, |
|
"rewards/rejected": -1.3170104026794434, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 1.7266187050359711, |
|
"grad_norm": 6.028374195098877, |
|
"learning_rate": 2.1986582993616926e-06, |
|
"logits/chosen": -1.7396399974822998, |
|
"logits/rejected": -2.01945424079895, |
|
"logps/chosen": -169.17230224609375, |
|
"logps/rejected": -220.03109741210938, |
|
"loss": 0.2872, |
|
"rewards/accuracies": 0.90625, |
|
"rewards/chosen": 0.15267056226730347, |
|
"rewards/margins": 1.99850594997406, |
|
"rewards/rejected": -1.8458354473114014, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.7841726618705036, |
|
"grad_norm": 8.13830280303955, |
|
"learning_rate": 2.0281560113677085e-06, |
|
"logits/chosen": -1.8339189291000366, |
|
"logits/rejected": -2.0238192081451416, |
|
"logps/chosen": -176.9852752685547, |
|
"logps/rejected": -219.31089782714844, |
|
"loss": 0.3258, |
|
"rewards/accuracies": 0.90625, |
|
"rewards/chosen": 0.418798565864563, |
|
"rewards/margins": 1.8581160306930542, |
|
"rewards/rejected": -1.4393174648284912, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 1.841726618705036, |
|
"grad_norm": 6.037229537963867, |
|
"learning_rate": 1.8599022748561324e-06, |
|
"logits/chosen": -1.6754798889160156, |
|
"logits/rejected": -1.8604053258895874, |
|
"logps/chosen": -179.31600952148438, |
|
"logps/rejected": -222.77664184570312, |
|
"loss": 0.2994, |
|
"rewards/accuracies": 0.921875, |
|
"rewards/chosen": 0.05394846200942993, |
|
"rewards/margins": 1.9718104600906372, |
|
"rewards/rejected": -1.9178617000579834, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 1.8992805755395683, |
|
"grad_norm": 5.1347336769104, |
|
"learning_rate": 1.694698895503774e-06, |
|
"logits/chosen": -1.6910380125045776, |
|
"logits/rejected": -1.9908016920089722, |
|
"logps/chosen": -168.50946044921875, |
|
"logps/rejected": -230.62281799316406, |
|
"loss": 0.2577, |
|
"rewards/accuracies": 0.921875, |
|
"rewards/chosen": 0.30626192688941956, |
|
"rewards/margins": 2.129185199737549, |
|
"rewards/rejected": -1.822923183441162, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 1.9568345323741008, |
|
"grad_norm": 4.9467926025390625, |
|
"learning_rate": 1.5333331426464532e-06, |
|
"logits/chosen": -1.836257815361023, |
|
"logits/rejected": -2.202897310256958, |
|
"logps/chosen": -191.49977111816406, |
|
"logps/rejected": -229.22738647460938, |
|
"loss": 0.2053, |
|
"rewards/accuracies": 0.96875, |
|
"rewards/chosen": 0.3933337330818176, |
|
"rewards/margins": 2.6139490604400635, |
|
"rewards/rejected": -2.2206156253814697, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 2.014388489208633, |
|
"grad_norm": 8.162074089050293, |
|
"learning_rate": 1.3765739975820964e-06, |
|
"logits/chosen": -1.7075754404067993, |
|
"logits/rejected": -1.7874010801315308, |
|
"logps/chosen": -185.13323974609375, |
|
"logps/rejected": -241.02822875976562, |
|
"loss": 0.3402, |
|
"rewards/accuracies": 0.84375, |
|
"rewards/chosen": 0.3749930262565613, |
|
"rewards/margins": 1.841801404953003, |
|
"rewards/rejected": -1.4668081998825073, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 2.0719424460431655, |
|
"grad_norm": 9.112337112426758, |
|
"learning_rate": 1.225168489024661e-06, |
|
"logits/chosen": -1.6381630897521973, |
|
"logits/rejected": -1.8705121278762817, |
|
"logps/chosen": -164.97901916503906, |
|
"logps/rejected": -204.80026245117188, |
|
"loss": 0.3119, |
|
"rewards/accuracies": 0.890625, |
|
"rewards/chosen": 0.2434196174144745, |
|
"rewards/margins": 1.9632219076156616, |
|
"rewards/rejected": -1.7198021411895752, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 2.129496402877698, |
|
"grad_norm": 6.385829448699951, |
|
"learning_rate": 1.079838133172111e-06, |
|
"logits/chosen": -1.7317883968353271, |
|
"logits/rejected": -1.9150619506835938, |
|
"logps/chosen": -176.2250518798828, |
|
"logps/rejected": -224.98941040039062, |
|
"loss": 0.2232, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 0.1401495635509491, |
|
"rewards/margins": 2.5023159980773926, |
|
"rewards/rejected": -2.3621668815612793, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 2.1870503597122304, |
|
"grad_norm": 5.108960151672363, |
|
"learning_rate": 9.412754953531664e-07, |
|
"logits/chosen": -1.7440515756607056, |
|
"logits/rejected": -1.87562096118927, |
|
"logps/chosen": -171.93978881835938, |
|
"logps/rejected": -214.25946044921875, |
|
"loss": 0.2777, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -0.15876302123069763, |
|
"rewards/margins": 2.08390212059021, |
|
"rewards/rejected": -2.2426650524139404, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 2.2446043165467624, |
|
"grad_norm": 4.441920280456543, |
|
"learning_rate": 8.101408896381141e-07, |
|
"logits/chosen": -1.8361996412277222, |
|
"logits/rejected": -2.2859373092651367, |
|
"logps/chosen": -178.8001251220703, |
|
"logps/rejected": -210.408447265625, |
|
"loss": 0.2124, |
|
"rewards/accuracies": 0.953125, |
|
"rewards/chosen": 0.34240958094596863, |
|
"rewards/margins": 2.488708734512329, |
|
"rewards/rejected": -2.146299362182617, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 2.302158273381295, |
|
"grad_norm": 5.0956807136535645, |
|
"learning_rate": 6.870592321415595e-07, |
|
"logits/chosen": -1.7229351997375488, |
|
"logits/rejected": -1.9915900230407715, |
|
"logps/chosen": -176.20111083984375, |
|
"logps/rejected": -228.6021728515625, |
|
"loss": 0.1949, |
|
"rewards/accuracies": 0.96875, |
|
"rewards/chosen": 0.3722834885120392, |
|
"rewards/margins": 2.7083969116210938, |
|
"rewards/rejected": -2.336113214492798, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 2.3597122302158273, |
|
"grad_norm": 4.845056056976318, |
|
"learning_rate": 5.72617063012551e-07, |
|
"logits/chosen": -1.7617988586425781, |
|
"logits/rejected": -1.9842320680618286, |
|
"logps/chosen": -169.4710235595703, |
|
"logps/rejected": -221.92437744140625, |
|
"loss": 0.2402, |
|
"rewards/accuracies": 0.90625, |
|
"rewards/chosen": 0.4934796690940857, |
|
"rewards/margins": 2.6191768646240234, |
|
"rewards/rejected": -2.125697374343872, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 2.41726618705036, |
|
"grad_norm": 6.618442535400391, |
|
"learning_rate": 4.673597513036684e-07, |
|
"logits/chosen": -1.8196489810943604, |
|
"logits/rejected": -2.119847536087036, |
|
"logps/chosen": -194.7741241455078, |
|
"logps/rejected": -226.7880401611328, |
|
"loss": 0.2859, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.15423324704170227, |
|
"rewards/margins": 1.936818242073059, |
|
"rewards/rejected": -1.7825851440429688, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 2.4748201438848922, |
|
"grad_norm": 5.660532474517822, |
|
"learning_rate": 3.717888960391222e-07, |
|
"logits/chosen": -1.6743967533111572, |
|
"logits/rejected": -2.142153263092041, |
|
"logps/chosen": -182.84117126464844, |
|
"logps/rejected": -236.10464477539062, |
|
"loss": 0.2294, |
|
"rewards/accuracies": 0.953125, |
|
"rewards/chosen": -0.0025496408343315125, |
|
"rewards/margins": 2.134979486465454, |
|
"rewards/rejected": -2.137529134750366, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 2.5323741007194247, |
|
"grad_norm": 4.540892124176025, |
|
"learning_rate": 2.8635993586697555e-07, |
|
"logits/chosen": -1.559401273727417, |
|
"logits/rejected": -1.7534209489822388, |
|
"logps/chosen": -168.13064575195312, |
|
"logps/rejected": -207.1958770751953, |
|
"loss": 0.2516, |
|
"rewards/accuracies": 0.84375, |
|
"rewards/chosen": -0.20388221740722656, |
|
"rewards/margins": 2.3258824348449707, |
|
"rewards/rejected": -2.5297646522521973, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 2.5899280575539567, |
|
"grad_norm": 5.678586483001709, |
|
"learning_rate": 2.1147997868658427e-07, |
|
"logits/chosen": -1.5231993198394775, |
|
"logits/rejected": -1.73148512840271, |
|
"logps/chosen": -158.38270568847656, |
|
"logps/rejected": -204.1846466064453, |
|
"loss": 0.2254, |
|
"rewards/accuracies": 0.953125, |
|
"rewards/chosen": 0.09025146812200546, |
|
"rewards/margins": 2.35001277923584, |
|
"rewards/rejected": -2.2597615718841553, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 2.647482014388489, |
|
"grad_norm": 7.349920749664307, |
|
"learning_rate": 1.4750586159405917e-07, |
|
"logits/chosen": -1.716179370880127, |
|
"logits/rejected": -1.8096524477005005, |
|
"logps/chosen": -162.72821044921875, |
|
"logps/rejected": -225.792236328125, |
|
"loss": 0.3215, |
|
"rewards/accuracies": 0.84375, |
|
"rewards/chosen": 0.1088458001613617, |
|
"rewards/margins": 1.9956300258636475, |
|
"rewards/rejected": -1.886784315109253, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 2.7050359712230216, |
|
"grad_norm": 5.460008144378662, |
|
"learning_rate": 9.474245039099883e-08, |
|
"logits/chosen": -1.9032938480377197, |
|
"logits/rejected": -2.2183752059936523, |
|
"logps/chosen": -185.80279541015625, |
|
"logps/rejected": -227.5401611328125, |
|
"loss": 0.2675, |
|
"rewards/accuracies": 0.890625, |
|
"rewards/chosen": 0.1709720939397812, |
|
"rewards/margins": 2.227972984313965, |
|
"rewards/rejected": -2.0570011138916016, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 2.762589928057554, |
|
"grad_norm": 7.240005016326904, |
|
"learning_rate": 5.344118676011173e-08, |
|
"logits/chosen": -1.8280268907546997, |
|
"logits/rejected": -2.053773880004883, |
|
"logps/chosen": -182.4041748046875, |
|
"logps/rejected": -230.9427032470703, |
|
"loss": 0.2482, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.03328900411725044, |
|
"rewards/margins": 2.2742063999176025, |
|
"rewards/rejected": -2.240917205810547, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 2.8201438848920866, |
|
"grad_norm": 8.40621566772461, |
|
"learning_rate": 2.3798890031092037e-08, |
|
"logits/chosen": -1.7997097969055176, |
|
"logits/rejected": -1.9944825172424316, |
|
"logps/chosen": -169.30648803710938, |
|
"logps/rejected": -205.72390747070312, |
|
"loss": 0.2811, |
|
"rewards/accuracies": 0.890625, |
|
"rewards/chosen": 0.2229904681444168, |
|
"rewards/margins": 2.305001735687256, |
|
"rewards/rejected": -2.0820112228393555, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 2.8776978417266186, |
|
"grad_norm": 4.172658443450928, |
|
"learning_rate": 5.9568192468811844e-09, |
|
"logits/chosen": -1.6826629638671875, |
|
"logits/rejected": -1.8101749420166016, |
|
"logps/chosen": -172.32896423339844, |
|
"logps/rejected": -232.7934112548828, |
|
"loss": 0.169, |
|
"rewards/accuracies": 0.96875, |
|
"rewards/chosen": 0.24307750165462494, |
|
"rewards/margins": 2.7273943424224854, |
|
"rewards/rejected": -2.484316825866699, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 2.935251798561151, |
|
"grad_norm": 4.340843200683594, |
|
"learning_rate": 0.0, |
|
"logits/chosen": -1.5438573360443115, |
|
"logits/rejected": -1.7152578830718994, |
|
"logps/chosen": -160.0493927001953, |
|
"logps/rejected": -215.81707763671875, |
|
"loss": 0.19, |
|
"rewards/accuracies": 0.953125, |
|
"rewards/chosen": 0.4251682162284851, |
|
"rewards/margins": 2.6738100051879883, |
|
"rewards/rejected": -2.2486419677734375, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 2.935251798561151, |
|
"step": 102, |
|
"total_flos": 8.94448793688146e+16, |
|
"train_loss": 0.4110284693100873, |
|
"train_runtime": 896.119, |
|
"train_samples_per_second": 3.713, |
|
"train_steps_per_second": 0.114 |
|
} |
|
], |
|
"logging_steps": 2, |
|
"max_steps": 102, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": false, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 8.94448793688146e+16, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|