|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.012587817034378349, |
|
"eval_steps": 500, |
|
"global_step": 370, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 3.4021127119941484e-05, |
|
"grad_norm": 33.091712951660156, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"logits/chosen": -3.133824586868286, |
|
"logits/rejected": -3.1356313228607178, |
|
"logps/chosen": -317.8847961425781, |
|
"logps/rejected": -306.8866271972656, |
|
"loss": 0.8544, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.10576057434082031, |
|
"rewards/margins": 0.2207910716533661, |
|
"rewards/rejected": -0.3265516459941864, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 6.804225423988297e-05, |
|
"grad_norm": 38.49541473388672, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"logits/chosen": -3.1433358192443848, |
|
"logits/rejected": -3.1124918460845947, |
|
"logps/chosen": -212.93820190429688, |
|
"logps/rejected": -198.22511291503906, |
|
"loss": 1.02, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.28638574481010437, |
|
"rewards/margins": -0.37813109159469604, |
|
"rewards/rejected": 0.0917452871799469, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.00010206338135982445, |
|
"grad_norm": 27.203516006469727, |
|
"learning_rate": 3e-06, |
|
"logits/chosen": -3.2104451656341553, |
|
"logits/rejected": -3.2103281021118164, |
|
"logps/chosen": -235.89254760742188, |
|
"logps/rejected": -206.0758514404297, |
|
"loss": 0.605, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.40017586946487427, |
|
"rewards/margins": 0.24887371063232422, |
|
"rewards/rejected": 0.15130215883255005, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.00013608450847976594, |
|
"grad_norm": 26.42736053466797, |
|
"learning_rate": 4.000000000000001e-06, |
|
"logits/chosen": -3.0741634368896484, |
|
"logits/rejected": -2.953117847442627, |
|
"logps/chosen": -428.7364196777344, |
|
"logps/rejected": -192.357666015625, |
|
"loss": 0.4914, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.4332427978515625, |
|
"rewards/margins": 0.5804362893104553, |
|
"rewards/rejected": -0.1471935510635376, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.00017010563559970743, |
|
"grad_norm": 35.623634338378906, |
|
"learning_rate": 5e-06, |
|
"logits/chosen": -2.985285758972168, |
|
"logits/rejected": -2.8854246139526367, |
|
"logps/chosen": -342.38604736328125, |
|
"logps/rejected": -211.78237915039062, |
|
"loss": 0.8273, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -0.3475755751132965, |
|
"rewards/margins": -0.19371166825294495, |
|
"rewards/rejected": -0.15386392176151276, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0002041267627196489, |
|
"grad_norm": 31.70409393310547, |
|
"learning_rate": 6e-06, |
|
"logits/chosen": -3.074589967727661, |
|
"logits/rejected": -2.85581636428833, |
|
"logps/chosen": -593.6435546875, |
|
"logps/rejected": -259.5284729003906, |
|
"loss": 0.5473, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.2495414763689041, |
|
"rewards/margins": 0.35888367891311646, |
|
"rewards/rejected": -0.10934218764305115, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.00023814788983959038, |
|
"grad_norm": 32.41214370727539, |
|
"learning_rate": 7.000000000000001e-06, |
|
"logits/chosen": -2.976675033569336, |
|
"logits/rejected": -2.978133201599121, |
|
"logps/chosen": -218.11688232421875, |
|
"logps/rejected": -231.7698974609375, |
|
"loss": 0.7124, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -0.08671779930591583, |
|
"rewards/margins": 0.21950224041938782, |
|
"rewards/rejected": -0.30622005462646484, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.00027216901695953187, |
|
"grad_norm": 27.637351989746094, |
|
"learning_rate": 8.000000000000001e-06, |
|
"logits/chosen": -2.997600793838501, |
|
"logits/rejected": -3.1078853607177734, |
|
"logps/chosen": -165.71444702148438, |
|
"logps/rejected": -335.5938720703125, |
|
"loss": 0.6085, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.057393088936805725, |
|
"rewards/margins": 0.34109365940093994, |
|
"rewards/rejected": -0.3984867036342621, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.00030619014407947336, |
|
"grad_norm": 32.246559143066406, |
|
"learning_rate": 9e-06, |
|
"logits/chosen": -3.0760390758514404, |
|
"logits/rejected": -2.938197135925293, |
|
"logps/chosen": -463.1333312988281, |
|
"logps/rejected": -173.99951171875, |
|
"loss": 0.6398, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.5581258535385132, |
|
"rewards/margins": 0.7653559446334839, |
|
"rewards/rejected": -0.20723000168800354, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.00034021127119941485, |
|
"grad_norm": 39.84626388549805, |
|
"learning_rate": 1e-05, |
|
"logits/chosen": -3.192596673965454, |
|
"logits/rejected": -2.9751501083374023, |
|
"logps/chosen": -473.14532470703125, |
|
"logps/rejected": -110.90155029296875, |
|
"loss": 0.9794, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -0.3494076132774353, |
|
"rewards/margins": -0.31546899676322937, |
|
"rewards/rejected": -0.03393859416246414, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.00037423239831935634, |
|
"grad_norm": 39.444889068603516, |
|
"learning_rate": 1.1000000000000001e-05, |
|
"logits/chosen": -3.096883535385132, |
|
"logits/rejected": -3.043616771697998, |
|
"logps/chosen": -391.7703552246094, |
|
"logps/rejected": -271.8331604003906, |
|
"loss": 0.8029, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.6240158081054688, |
|
"rewards/margins": 0.4651091992855072, |
|
"rewards/rejected": 0.15890654921531677, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.0004082535254392978, |
|
"grad_norm": 40.78356170654297, |
|
"learning_rate": 1.2e-05, |
|
"logits/chosen": -3.000021457672119, |
|
"logits/rejected": -3.0893115997314453, |
|
"logps/chosen": -255.94607543945312, |
|
"logps/rejected": -405.3093566894531, |
|
"loss": 0.6297, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.33697396516799927, |
|
"rewards/margins": 0.3315456509590149, |
|
"rewards/rejected": 0.005428321659564972, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.00044227465255923927, |
|
"grad_norm": 35.69674301147461, |
|
"learning_rate": 1.3000000000000001e-05, |
|
"logits/chosen": -3.011014699935913, |
|
"logits/rejected": -3.0214414596557617, |
|
"logps/chosen": -223.4480438232422, |
|
"logps/rejected": -261.62469482421875, |
|
"loss": 0.815, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.7416718006134033, |
|
"rewards/margins": 0.19626998901367188, |
|
"rewards/rejected": 0.5454018115997314, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.00047629577967918076, |
|
"grad_norm": 32.88405990600586, |
|
"learning_rate": 1.4000000000000001e-05, |
|
"logits/chosen": -3.065549373626709, |
|
"logits/rejected": -3.1094508171081543, |
|
"logps/chosen": -169.08047485351562, |
|
"logps/rejected": -227.85162353515625, |
|
"loss": 0.8115, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 0.37659722566604614, |
|
"rewards/margins": -0.11390496790409088, |
|
"rewards/rejected": 0.4905022084712982, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.0005103169067991223, |
|
"grad_norm": 20.918012619018555, |
|
"learning_rate": 1.5e-05, |
|
"logits/chosen": -3.045846462249756, |
|
"logits/rejected": -2.9114990234375, |
|
"logps/chosen": -573.1253662109375, |
|
"logps/rejected": -273.9326477050781, |
|
"loss": 0.3105, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.7755428552627563, |
|
"rewards/margins": 1.082038164138794, |
|
"rewards/rejected": -0.3064952790737152, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0005443380339190637, |
|
"grad_norm": 38.434749603271484, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"logits/chosen": -3.1250810623168945, |
|
"logits/rejected": -2.96527099609375, |
|
"logps/chosen": -396.9795837402344, |
|
"logps/rejected": -174.92918395996094, |
|
"loss": 0.9971, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.2244594693183899, |
|
"rewards/margins": -0.37520748376846313, |
|
"rewards/rejected": 0.5996670126914978, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.0005783591610390052, |
|
"grad_norm": 37.71885681152344, |
|
"learning_rate": 1.7000000000000003e-05, |
|
"logits/chosen": -2.970050096511841, |
|
"logits/rejected": -3.0346829891204834, |
|
"logps/chosen": -236.67071533203125, |
|
"logps/rejected": -403.0212097167969, |
|
"loss": 1.0307, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.9259212613105774, |
|
"rewards/margins": -0.23367927968502045, |
|
"rewards/rejected": 1.1596004962921143, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.0006123802881589467, |
|
"grad_norm": 25.964290618896484, |
|
"learning_rate": 1.8e-05, |
|
"logits/chosen": -3.0690722465515137, |
|
"logits/rejected": -3.0348947048187256, |
|
"logps/chosen": -312.81317138671875, |
|
"logps/rejected": -259.230712890625, |
|
"loss": 0.6989, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.6348915100097656, |
|
"rewards/margins": 0.6418477892875671, |
|
"rewards/rejected": -0.006956290453672409, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.0006464014152788882, |
|
"grad_norm": 27.076108932495117, |
|
"learning_rate": 1.9e-05, |
|
"logits/chosen": -3.135136842727661, |
|
"logits/rejected": -2.911625862121582, |
|
"logps/chosen": -473.7462158203125, |
|
"logps/rejected": -141.9491729736328, |
|
"loss": 0.6027, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.9424793720245361, |
|
"rewards/margins": 0.9380611181259155, |
|
"rewards/rejected": 0.00441819429397583, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.0006804225423988297, |
|
"grad_norm": 51.820159912109375, |
|
"learning_rate": 2e-05, |
|
"logits/chosen": -2.8403027057647705, |
|
"logits/rejected": -3.0664803981781006, |
|
"logps/chosen": -219.79153442382812, |
|
"logps/rejected": -231.70004272460938, |
|
"loss": 1.1022, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.052263081073760986, |
|
"rewards/margins": -0.6957410573959351, |
|
"rewards/rejected": 0.7480041980743408, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0007144436695187712, |
|
"grad_norm": 45.9146614074707, |
|
"learning_rate": 2.1e-05, |
|
"logits/chosen": -2.9764764308929443, |
|
"logits/rejected": -2.9391565322875977, |
|
"logps/chosen": -458.0147705078125, |
|
"logps/rejected": -218.37648010253906, |
|
"loss": 1.0772, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.8049196004867554, |
|
"rewards/margins": -0.448437362909317, |
|
"rewards/rejected": 1.2533570528030396, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.0007484647966387127, |
|
"grad_norm": 22.854082107543945, |
|
"learning_rate": 2.2000000000000003e-05, |
|
"logits/chosen": -3.1029181480407715, |
|
"logits/rejected": -3.044419527053833, |
|
"logps/chosen": -154.8139190673828, |
|
"logps/rejected": -155.980712890625, |
|
"loss": 0.4065, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.45841503143310547, |
|
"rewards/margins": 0.8753025531768799, |
|
"rewards/rejected": -0.41688746213912964, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.0007824859237586541, |
|
"grad_norm": 43.94650650024414, |
|
"learning_rate": 2.3000000000000003e-05, |
|
"logits/chosen": -2.9984076023101807, |
|
"logits/rejected": -3.114562511444092, |
|
"logps/chosen": -263.8316650390625, |
|
"logps/rejected": -340.1427307128906, |
|
"loss": 1.0427, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 0.6933631896972656, |
|
"rewards/margins": -0.14637237787246704, |
|
"rewards/rejected": 0.8397356271743774, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.0008165070508785956, |
|
"grad_norm": 49.11759948730469, |
|
"learning_rate": 2.4e-05, |
|
"logits/chosen": -3.025609016418457, |
|
"logits/rejected": -3.0236239433288574, |
|
"logps/chosen": -412.2383117675781, |
|
"logps/rejected": -643.138427734375, |
|
"loss": 1.0374, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.1175220012664795, |
|
"rewards/margins": -0.31503716111183167, |
|
"rewards/rejected": 1.4325592517852783, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.000850528177998537, |
|
"grad_norm": 37.57383728027344, |
|
"learning_rate": 2.5e-05, |
|
"logits/chosen": -3.024618625640869, |
|
"logits/rejected": -3.123701572418213, |
|
"logps/chosen": -201.70530700683594, |
|
"logps/rejected": -224.99038696289062, |
|
"loss": 0.7477, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 0.9206753373146057, |
|
"rewards/margins": 0.5947309732437134, |
|
"rewards/rejected": 0.32594436407089233, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0008845493051184785, |
|
"grad_norm": 40.297794342041016, |
|
"learning_rate": 2.6000000000000002e-05, |
|
"logits/chosen": -3.1100215911865234, |
|
"logits/rejected": -3.0195024013519287, |
|
"logps/chosen": -226.2320098876953, |
|
"logps/rejected": -193.0676727294922, |
|
"loss": 0.8151, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.9122087359428406, |
|
"rewards/margins": 0.0203973650932312, |
|
"rewards/rejected": 0.8918113708496094, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.00091857043223842, |
|
"grad_norm": 43.394229888916016, |
|
"learning_rate": 2.7000000000000002e-05, |
|
"logits/chosen": -3.0187630653381348, |
|
"logits/rejected": -3.053678035736084, |
|
"logps/chosen": -153.02883911132812, |
|
"logps/rejected": -260.0540771484375, |
|
"loss": 1.3457, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.7145153284072876, |
|
"rewards/margins": -0.806611180305481, |
|
"rewards/rejected": 1.5211265087127686, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.0009525915593583615, |
|
"grad_norm": 28.59680938720703, |
|
"learning_rate": 2.8000000000000003e-05, |
|
"logits/chosen": -3.032409191131592, |
|
"logits/rejected": -3.0502161979675293, |
|
"logps/chosen": -240.1644287109375, |
|
"logps/rejected": -245.81436157226562, |
|
"loss": 0.5501, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.0826761722564697, |
|
"rewards/margins": 0.3433869779109955, |
|
"rewards/rejected": 1.7392891645431519, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.000986612686478303, |
|
"grad_norm": 37.069583892822266, |
|
"learning_rate": 2.9e-05, |
|
"logits/chosen": -3.0516104698181152, |
|
"logits/rejected": -3.144774913787842, |
|
"logps/chosen": -296.03125, |
|
"logps/rejected": -292.9052429199219, |
|
"loss": 1.1087, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.8324771523475647, |
|
"rewards/margins": -0.34529268741607666, |
|
"rewards/rejected": 1.1777698993682861, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.0010206338135982445, |
|
"grad_norm": 30.238658905029297, |
|
"learning_rate": 3e-05, |
|
"logits/chosen": -3.064671039581299, |
|
"logits/rejected": -2.8963613510131836, |
|
"logps/chosen": -415.6085510253906, |
|
"logps/rejected": -149.9669952392578, |
|
"loss": 0.4892, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.0249770879745483, |
|
"rewards/margins": 0.7549355030059814, |
|
"rewards/rejected": 0.27004164457321167, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.001054654940718186, |
|
"grad_norm": 40.35402297973633, |
|
"learning_rate": 3.1e-05, |
|
"logits/chosen": -3.1762304306030273, |
|
"logits/rejected": -3.061676025390625, |
|
"logps/chosen": -585.9200439453125, |
|
"logps/rejected": -238.16307067871094, |
|
"loss": 0.587, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.726571559906006, |
|
"rewards/margins": 0.9808880090713501, |
|
"rewards/rejected": 1.7456837892532349, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.0010886760678381275, |
|
"grad_norm": 33.78396987915039, |
|
"learning_rate": 3.2000000000000005e-05, |
|
"logits/chosen": -3.05906081199646, |
|
"logits/rejected": -2.9525132179260254, |
|
"logps/chosen": -580.87646484375, |
|
"logps/rejected": -223.65457153320312, |
|
"loss": 0.8173, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.3172895908355713, |
|
"rewards/margins": 0.07103176414966583, |
|
"rewards/rejected": 1.2462577819824219, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.001122697194958069, |
|
"grad_norm": 40.97666549682617, |
|
"learning_rate": 3.3e-05, |
|
"logits/chosen": -3.084476947784424, |
|
"logits/rejected": -3.11270809173584, |
|
"logps/chosen": -285.040283203125, |
|
"logps/rejected": -251.92686462402344, |
|
"loss": 1.0645, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.4826446771621704, |
|
"rewards/margins": -0.13718131184577942, |
|
"rewards/rejected": 1.6198259592056274, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.0011567183220780105, |
|
"grad_norm": 33.32075881958008, |
|
"learning_rate": 3.4000000000000007e-05, |
|
"logits/chosen": -3.037482500076294, |
|
"logits/rejected": -3.0169787406921387, |
|
"logps/chosen": -467.7403869628906, |
|
"logps/rejected": -308.62274169921875, |
|
"loss": 0.5226, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.4038407802581787, |
|
"rewards/margins": 0.8624236583709717, |
|
"rewards/rejected": 1.541416883468628, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.001190739449197952, |
|
"grad_norm": 26.35866928100586, |
|
"learning_rate": 3.5e-05, |
|
"logits/chosen": -2.969816207885742, |
|
"logits/rejected": -2.9069106578826904, |
|
"logps/chosen": -312.856689453125, |
|
"logps/rejected": -183.8594512939453, |
|
"loss": 0.4495, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.1757181882858276, |
|
"rewards/margins": 0.803225576877594, |
|
"rewards/rejected": 0.37249261140823364, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.0012247605763178934, |
|
"grad_norm": 44.118099212646484, |
|
"learning_rate": 3.6e-05, |
|
"logits/chosen": -3.091865301132202, |
|
"logits/rejected": -3.0135951042175293, |
|
"logps/chosen": -724.1255493164062, |
|
"logps/rejected": -335.2945861816406, |
|
"loss": 0.6341, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.7523369789123535, |
|
"rewards/margins": 0.3912923336029053, |
|
"rewards/rejected": 1.3610445261001587, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.001258781703437835, |
|
"grad_norm": 26.80854606628418, |
|
"learning_rate": 3.7e-05, |
|
"logits/chosen": -3.0334794521331787, |
|
"logits/rejected": -2.9878833293914795, |
|
"logps/chosen": -406.20745849609375, |
|
"logps/rejected": -377.2811584472656, |
|
"loss": 0.5577, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.6232261657714844, |
|
"rewards/margins": 0.9015346765518188, |
|
"rewards/rejected": 1.7216914892196655, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.0012928028305577764, |
|
"grad_norm": 31.345367431640625, |
|
"learning_rate": 3.8e-05, |
|
"logits/chosen": -3.0986504554748535, |
|
"logits/rejected": -3.062534809112549, |
|
"logps/chosen": -412.54498291015625, |
|
"logps/rejected": -386.44158935546875, |
|
"loss": 0.5479, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.985120177268982, |
|
"rewards/margins": 0.8609979748725891, |
|
"rewards/rejected": 1.1241222620010376, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.001326823957677718, |
|
"grad_norm": 15.26794719696045, |
|
"learning_rate": 3.9000000000000006e-05, |
|
"logits/chosen": -3.0730984210968018, |
|
"logits/rejected": -3.0951075553894043, |
|
"logps/chosen": -305.12982177734375, |
|
"logps/rejected": -285.45001220703125, |
|
"loss": 0.2547, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.55714750289917, |
|
"rewards/margins": 1.3561341762542725, |
|
"rewards/rejected": 1.2010130882263184, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.0013608450847976594, |
|
"grad_norm": 22.881174087524414, |
|
"learning_rate": 4e-05, |
|
"logits/chosen": -3.162843942642212, |
|
"logits/rejected": -3.15628981590271, |
|
"logps/chosen": -277.48626708984375, |
|
"logps/rejected": -219.9507293701172, |
|
"loss": 0.5427, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.6990911960601807, |
|
"rewards/margins": 0.5041837692260742, |
|
"rewards/rejected": 2.1949074268341064, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.001394866211917601, |
|
"grad_norm": 15.179062843322754, |
|
"learning_rate": 4.1e-05, |
|
"logits/chosen": -3.138144016265869, |
|
"logits/rejected": -3.1033153533935547, |
|
"logps/chosen": -234.6917724609375, |
|
"logps/rejected": -141.10888671875, |
|
"loss": 0.2626, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.799344539642334, |
|
"rewards/margins": 1.4650464057922363, |
|
"rewards/rejected": 1.3342978954315186, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.0014288873390375424, |
|
"grad_norm": 21.09362030029297, |
|
"learning_rate": 4.2e-05, |
|
"logits/chosen": -3.0385375022888184, |
|
"logits/rejected": -3.083676338195801, |
|
"logps/chosen": -168.33697509765625, |
|
"logps/rejected": -179.30813598632812, |
|
"loss": 0.4002, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.3925881385803223, |
|
"rewards/margins": 0.8477088809013367, |
|
"rewards/rejected": 1.5448791980743408, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.0014629084661574839, |
|
"grad_norm": 38.255069732666016, |
|
"learning_rate": 4.3e-05, |
|
"logits/chosen": -3.080360174179077, |
|
"logits/rejected": -3.029470443725586, |
|
"logps/chosen": -267.5482177734375, |
|
"logps/rejected": -161.17111206054688, |
|
"loss": 0.8857, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.8575999736785889, |
|
"rewards/margins": -0.24116593599319458, |
|
"rewards/rejected": 2.0987658500671387, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.0014969295932774254, |
|
"grad_norm": 27.97095489501953, |
|
"learning_rate": 4.4000000000000006e-05, |
|
"logits/chosen": -3.1322476863861084, |
|
"logits/rejected": -3.1334149837493896, |
|
"logps/chosen": -94.13690185546875, |
|
"logps/rejected": -78.98738098144531, |
|
"loss": 0.9987, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.6137428283691406, |
|
"rewards/margins": -0.4127828776836395, |
|
"rewards/rejected": 2.0265257358551025, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.0015309507203973669, |
|
"grad_norm": 45.25028991699219, |
|
"learning_rate": 4.5e-05, |
|
"logits/chosen": -3.121274709701538, |
|
"logits/rejected": -3.1236798763275146, |
|
"logps/chosen": -516.3858642578125, |
|
"logps/rejected": -212.694580078125, |
|
"loss": 0.9457, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 2.419821262359619, |
|
"rewards/margins": -0.38284340500831604, |
|
"rewards/rejected": 2.8026647567749023, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.0015649718475173081, |
|
"grad_norm": 23.58992576599121, |
|
"learning_rate": 4.600000000000001e-05, |
|
"logits/chosen": -3.113922357559204, |
|
"logits/rejected": -3.0506935119628906, |
|
"logps/chosen": -318.288330078125, |
|
"logps/rejected": -124.70735168457031, |
|
"loss": 0.5671, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.407118320465088, |
|
"rewards/margins": 0.30923187732696533, |
|
"rewards/rejected": 2.097886323928833, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.0015989929746372496, |
|
"grad_norm": 36.93040084838867, |
|
"learning_rate": 4.7e-05, |
|
"logits/chosen": -3.079291582107544, |
|
"logits/rejected": -3.1158454418182373, |
|
"logps/chosen": -246.26065063476562, |
|
"logps/rejected": -336.46649169921875, |
|
"loss": 0.9352, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 2.214073657989502, |
|
"rewards/margins": -0.2623659074306488, |
|
"rewards/rejected": 2.4764394760131836, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.0016330141017571911, |
|
"grad_norm": 32.5983772277832, |
|
"learning_rate": 4.8e-05, |
|
"logits/chosen": -3.155709981918335, |
|
"logits/rejected": -3.1829142570495605, |
|
"logps/chosen": -204.55328369140625, |
|
"logps/rejected": -219.52647399902344, |
|
"loss": 0.8805, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.155839681625366, |
|
"rewards/margins": -0.002133697271347046, |
|
"rewards/rejected": 2.157973289489746, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.0016670352288771326, |
|
"grad_norm": 35.626869201660156, |
|
"learning_rate": 4.9e-05, |
|
"logits/chosen": -3.10532546043396, |
|
"logits/rejected": -3.113755464553833, |
|
"logps/chosen": -390.5479736328125, |
|
"logps/rejected": -351.7720031738281, |
|
"loss": 0.7657, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.5670056343078613, |
|
"rewards/margins": -0.04349440336227417, |
|
"rewards/rejected": 2.6105000972747803, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.001701056355997074, |
|
"grad_norm": 37.51042938232422, |
|
"learning_rate": 5e-05, |
|
"logits/chosen": -3.122190237045288, |
|
"logits/rejected": -3.0837321281433105, |
|
"logps/chosen": -251.68600463867188, |
|
"logps/rejected": -242.4346923828125, |
|
"loss": 0.9539, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.0532169342041016, |
|
"rewards/margins": -0.32518482208251953, |
|
"rewards/rejected": 2.378401756286621, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0017350774831170156, |
|
"grad_norm": 43.16423416137695, |
|
"learning_rate": 4.999939076763487e-05, |
|
"logits/chosen": -3.0424001216888428, |
|
"logits/rejected": -3.0246453285217285, |
|
"logps/chosen": -228.8073272705078, |
|
"logps/rejected": -399.57781982421875, |
|
"loss": 1.204, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.1423535346984863, |
|
"rewards/margins": -0.628890335559845, |
|
"rewards/rejected": 2.7712438106536865, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.001769098610236957, |
|
"grad_norm": 32.16966247558594, |
|
"learning_rate": 4.999756310023261e-05, |
|
"logits/chosen": -3.049051284790039, |
|
"logits/rejected": -3.061192750930786, |
|
"logps/chosen": -219.27503967285156, |
|
"logps/rejected": -238.98226928710938, |
|
"loss": 0.9616, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.1868481636047363, |
|
"rewards/margins": -0.2618050277233124, |
|
"rewards/rejected": 2.448652744293213, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.0018031197373568986, |
|
"grad_norm": 38.5218505859375, |
|
"learning_rate": 4.999451708687114e-05, |
|
"logits/chosen": -3.088409185409546, |
|
"logits/rejected": -3.0845394134521484, |
|
"logps/chosen": -131.70074462890625, |
|
"logps/rejected": -205.93624877929688, |
|
"loss": 1.0419, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.822286605834961, |
|
"rewards/margins": -0.18211710453033447, |
|
"rewards/rejected": 2.004403829574585, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.00183714086447684, |
|
"grad_norm": 31.46576690673828, |
|
"learning_rate": 4.999025287600886e-05, |
|
"logits/chosen": -3.143968105316162, |
|
"logits/rejected": -3.032649278640747, |
|
"logps/chosen": -403.4815673828125, |
|
"logps/rejected": -195.69281005859375, |
|
"loss": 0.567, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.2466557025909424, |
|
"rewards/margins": 0.6339659094810486, |
|
"rewards/rejected": 1.612689733505249, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.0018711619915967816, |
|
"grad_norm": 22.339370727539062, |
|
"learning_rate": 4.99847706754774e-05, |
|
"logits/chosen": -3.1364409923553467, |
|
"logits/rejected": -3.0678646564483643, |
|
"logps/chosen": -272.07330322265625, |
|
"logps/rejected": -94.81082153320312, |
|
"loss": 0.4856, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 1.9992039203643799, |
|
"rewards/margins": 0.5075141191482544, |
|
"rewards/rejected": 1.491689682006836, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.001905183118716723, |
|
"grad_norm": 38.88309860229492, |
|
"learning_rate": 4.997807075247146e-05, |
|
"logits/chosen": -3.0754241943359375, |
|
"logits/rejected": -3.1445722579956055, |
|
"logps/chosen": -161.1916961669922, |
|
"logps/rejected": -230.43589782714844, |
|
"loss": 0.8925, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.070554256439209, |
|
"rewards/margins": -0.027551889419555664, |
|
"rewards/rejected": 2.0981061458587646, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.0019392042458366645, |
|
"grad_norm": 29.006982803344727, |
|
"learning_rate": 4.997015343353585e-05, |
|
"logits/chosen": -3.079509973526001, |
|
"logits/rejected": -3.119426727294922, |
|
"logps/chosen": -236.40664672851562, |
|
"logps/rejected": -259.257568359375, |
|
"loss": 0.5812, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.4261646270751953, |
|
"rewards/margins": 0.7371253371238708, |
|
"rewards/rejected": 1.6890392303466797, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.001973225372956606, |
|
"grad_norm": 18.193532943725586, |
|
"learning_rate": 4.996101910454953e-05, |
|
"logits/chosen": -3.070995569229126, |
|
"logits/rejected": -2.8317549228668213, |
|
"logps/chosen": -730.03759765625, |
|
"logps/rejected": -133.17929077148438, |
|
"loss": 0.3401, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 3.711276054382324, |
|
"rewards/margins": 1.7915010452270508, |
|
"rewards/rejected": 1.9197750091552734, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.0020072465000765477, |
|
"grad_norm": 35.50777053833008, |
|
"learning_rate": 4.995066821070679e-05, |
|
"logits/chosen": -3.093824863433838, |
|
"logits/rejected": -3.083796739578247, |
|
"logps/chosen": -238.11883544921875, |
|
"logps/rejected": -151.69790649414062, |
|
"loss": 0.6251, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.0124969482421875, |
|
"rewards/margins": 0.640554666519165, |
|
"rewards/rejected": 1.371942400932312, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.002041267627196489, |
|
"grad_norm": 40.34638977050781, |
|
"learning_rate": 4.993910125649561e-05, |
|
"logits/chosen": -3.1637063026428223, |
|
"logits/rejected": -3.055285930633545, |
|
"logps/chosen": -365.89434814453125, |
|
"logps/rejected": -187.11392211914062, |
|
"loss": 1.1514, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.089510679244995, |
|
"rewards/margins": -0.3508464992046356, |
|
"rewards/rejected": 2.440357208251953, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.0020752887543164307, |
|
"grad_norm": 31.27332878112793, |
|
"learning_rate": 4.992631880567301e-05, |
|
"logits/chosen": -2.9931535720825195, |
|
"logits/rejected": -2.994088649749756, |
|
"logps/chosen": -127.87042236328125, |
|
"logps/rejected": -236.82327270507812, |
|
"loss": 0.8745, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.9324662685394287, |
|
"rewards/margins": -0.040864601731300354, |
|
"rewards/rejected": 1.9733309745788574, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.002109309881436372, |
|
"grad_norm": 24.1131591796875, |
|
"learning_rate": 4.991232148123761e-05, |
|
"logits/chosen": -3.094635009765625, |
|
"logits/rejected": -3.060533046722412, |
|
"logps/chosen": -369.5483093261719, |
|
"logps/rejected": -167.15155029296875, |
|
"loss": 0.46, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.38918399810791, |
|
"rewards/margins": 1.1014902591705322, |
|
"rewards/rejected": 1.2876936197280884, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.0021433310085563133, |
|
"grad_norm": 49.346961975097656, |
|
"learning_rate": 4.989710996539926e-05, |
|
"logits/chosen": -3.1134653091430664, |
|
"logits/rejected": -3.0323257446289062, |
|
"logps/chosen": -273.39361572265625, |
|
"logps/rejected": -157.84652709960938, |
|
"loss": 1.023, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 1.1458964347839355, |
|
"rewards/margins": -0.5081260204315186, |
|
"rewards/rejected": 1.654022455215454, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.002177352135676255, |
|
"grad_norm": 41.145633697509766, |
|
"learning_rate": 4.988068499954578e-05, |
|
"logits/chosen": -2.9834160804748535, |
|
"logits/rejected": -2.9958994388580322, |
|
"logps/chosen": -334.5246887207031, |
|
"logps/rejected": -386.3387145996094, |
|
"loss": 1.5388, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.305959939956665, |
|
"rewards/margins": 0.07403802871704102, |
|
"rewards/rejected": 2.231921911239624, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.0022113732627961962, |
|
"grad_norm": 33.07326126098633, |
|
"learning_rate": 4.9863047384206835e-05, |
|
"logits/chosen": -3.0006370544433594, |
|
"logits/rejected": -3.020918846130371, |
|
"logps/chosen": -344.7704162597656, |
|
"logps/rejected": -200.45114135742188, |
|
"loss": 0.5845, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.643819808959961, |
|
"rewards/margins": 0.6079027652740479, |
|
"rewards/rejected": 1.035917043685913, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.002245394389916138, |
|
"grad_norm": 51.295345306396484, |
|
"learning_rate": 4.984419797901491e-05, |
|
"logits/chosen": -3.039350986480713, |
|
"logits/rejected": -3.030729293823242, |
|
"logps/chosen": -307.3952331542969, |
|
"logps/rejected": -238.56224060058594, |
|
"loss": 1.2331, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 0.6686372756958008, |
|
"rewards/margins": -0.6910279989242554, |
|
"rewards/rejected": 1.3596653938293457, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.0022794155170360792, |
|
"grad_norm": 27.732166290283203, |
|
"learning_rate": 4.982413770266342e-05, |
|
"logits/chosen": -3.092078685760498, |
|
"logits/rejected": -3.1104204654693604, |
|
"logps/chosen": -206.7743682861328, |
|
"logps/rejected": -112.66790771484375, |
|
"loss": 0.5552, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.273049235343933, |
|
"rewards/margins": 0.48671725392341614, |
|
"rewards/rejected": 0.7863319516181946, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.002313436644156021, |
|
"grad_norm": 46.03068161010742, |
|
"learning_rate": 4.980286753286195e-05, |
|
"logits/chosen": -2.9219119548797607, |
|
"logits/rejected": -2.973864793777466, |
|
"logps/chosen": -455.3492431640625, |
|
"logps/rejected": -422.7337646484375, |
|
"loss": 0.8967, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.788267731666565, |
|
"rewards/margins": -0.24189908802509308, |
|
"rewards/rejected": 2.0301668643951416, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.002347457771275962, |
|
"grad_norm": 24.625539779663086, |
|
"learning_rate": 4.978038850628854e-05, |
|
"logits/chosen": -2.9824914932250977, |
|
"logits/rejected": -2.923306703567505, |
|
"logps/chosen": -320.2785339355469, |
|
"logps/rejected": -199.95407104492188, |
|
"loss": 0.5542, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.4030070304870605, |
|
"rewards/margins": 1.2227776050567627, |
|
"rewards/rejected": 1.180229663848877, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.002381478898395904, |
|
"grad_norm": 52.87644958496094, |
|
"learning_rate": 4.975670171853926e-05, |
|
"logits/chosen": -2.9576680660247803, |
|
"logits/rejected": -3.113696336746216, |
|
"logps/chosen": -226.1188201904297, |
|
"logps/rejected": -472.0656433105469, |
|
"loss": 1.5976, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 1.8890628814697266, |
|
"rewards/margins": -1.2377982139587402, |
|
"rewards/rejected": 3.126861095428467, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.002415500025515845, |
|
"grad_norm": 41.89684295654297, |
|
"learning_rate": 4.9731808324074717e-05, |
|
"logits/chosen": -2.963118553161621, |
|
"logits/rejected": -2.95782732963562, |
|
"logps/chosen": -376.9123229980469, |
|
"logps/rejected": -408.6007080078125, |
|
"loss": 0.8305, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.2748968601226807, |
|
"rewards/margins": 1.4728538990020752, |
|
"rewards/rejected": 0.8020429611206055, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.002449521152635787, |
|
"grad_norm": 33.66620635986328, |
|
"learning_rate": 4.9705709536163824e-05, |
|
"logits/chosen": -2.951019287109375, |
|
"logits/rejected": -2.8525712490081787, |
|
"logps/chosen": -334.75689697265625, |
|
"logps/rejected": -256.43963623046875, |
|
"loss": 0.5852, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.2168543338775635, |
|
"rewards/margins": 1.3764129877090454, |
|
"rewards/rejected": 0.8404413461685181, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.002483542279755728, |
|
"grad_norm": 48.537376403808594, |
|
"learning_rate": 4.96784066268247e-05, |
|
"logits/chosen": -2.96649169921875, |
|
"logits/rejected": -2.877800703048706, |
|
"logps/chosen": -342.9018859863281, |
|
"logps/rejected": -166.73826599121094, |
|
"loss": 1.3147, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 1.4513309001922607, |
|
"rewards/margins": -0.9881281852722168, |
|
"rewards/rejected": 2.4394590854644775, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.00251756340687567, |
|
"grad_norm": 35.8758544921875, |
|
"learning_rate": 4.964990092676263e-05, |
|
"logits/chosen": -3.007376194000244, |
|
"logits/rejected": -3.0710248947143555, |
|
"logps/chosen": -322.75604248046875, |
|
"logps/rejected": -288.9136047363281, |
|
"loss": 1.0483, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.328976035118103, |
|
"rewards/margins": -0.057621777057647705, |
|
"rewards/rejected": 1.3865978717803955, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.002551584533995611, |
|
"grad_norm": 52.42377471923828, |
|
"learning_rate": 4.962019382530521e-05, |
|
"logits/chosen": -2.952801465988159, |
|
"logits/rejected": -2.9580321311950684, |
|
"logps/chosen": -272.15576171875, |
|
"logps/rejected": -230.2147674560547, |
|
"loss": 1.4341, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 1.2654911279678345, |
|
"rewards/margins": -0.9303737282752991, |
|
"rewards/rejected": 2.1958649158477783, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.002585605661115553, |
|
"grad_norm": 54.157386779785156, |
|
"learning_rate": 4.9589286770334654e-05, |
|
"logits/chosen": -2.997375965118408, |
|
"logits/rejected": -3.091611385345459, |
|
"logps/chosen": -219.73159790039062, |
|
"logps/rejected": -503.797119140625, |
|
"loss": 1.3568, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 1.520772933959961, |
|
"rewards/margins": -0.4154503643512726, |
|
"rewards/rejected": 1.9362232685089111, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.002619626788235494, |
|
"grad_norm": 40.647274017333984, |
|
"learning_rate": 4.9557181268217227e-05, |
|
"logits/chosen": -3.0128493309020996, |
|
"logits/rejected": -2.9981260299682617, |
|
"logps/chosen": -319.1977233886719, |
|
"logps/rejected": -250.11123657226562, |
|
"loss": 0.7922, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.9633705615997314, |
|
"rewards/margins": -0.027412444353103638, |
|
"rewards/rejected": 1.9907830953598022, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.002653647915355436, |
|
"grad_norm": 46.99341583251953, |
|
"learning_rate": 4.952387888372979e-05, |
|
"logits/chosen": -3.0307843685150146, |
|
"logits/rejected": -3.070122480392456, |
|
"logps/chosen": -134.52757263183594, |
|
"logps/rejected": -225.51242065429688, |
|
"loss": 1.182, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 1.2856706380844116, |
|
"rewards/margins": -0.7930065393447876, |
|
"rewards/rejected": 2.078677177429199, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.002687669042475377, |
|
"grad_norm": 26.92460823059082, |
|
"learning_rate": 4.94893812399836e-05, |
|
"logits/chosen": -2.995112895965576, |
|
"logits/rejected": -2.950528144836426, |
|
"logps/chosen": -560.2005615234375, |
|
"logps/rejected": -383.92523193359375, |
|
"loss": 0.5477, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.8620712757110596, |
|
"rewards/margins": 0.3632301390171051, |
|
"rewards/rejected": 1.4988411664962769, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.002721690169595319, |
|
"grad_norm": 18.80535125732422, |
|
"learning_rate": 4.9453690018345144e-05, |
|
"logits/chosen": -3.0887956619262695, |
|
"logits/rejected": -3.013925790786743, |
|
"logps/chosen": -201.7890167236328, |
|
"logps/rejected": -120.17092895507812, |
|
"loss": 0.6049, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.207493305206299, |
|
"rewards/margins": 0.5624717473983765, |
|
"rewards/rejected": 1.645021677017212, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.00275571129671526, |
|
"grad_norm": 35.741817474365234, |
|
"learning_rate": 4.94168069583542e-05, |
|
"logits/chosen": -3.087347984313965, |
|
"logits/rejected": -2.9956464767456055, |
|
"logps/chosen": -485.24749755859375, |
|
"logps/rejected": -269.6015930175781, |
|
"loss": 1.0357, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.650765299797058, |
|
"rewards/margins": -0.448674738407135, |
|
"rewards/rejected": 2.099440097808838, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.002789732423835202, |
|
"grad_norm": 25.36280059814453, |
|
"learning_rate": 4.937873385763908e-05, |
|
"logits/chosen": -2.9839279651641846, |
|
"logits/rejected": -3.0026051998138428, |
|
"logps/chosen": -221.44424438476562, |
|
"logps/rejected": -193.8582305908203, |
|
"loss": 0.9423, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.4655239582061768, |
|
"rewards/margins": -0.0549183189868927, |
|
"rewards/rejected": 1.520442247390747, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.002823753550955143, |
|
"grad_norm": 12.532795906066895, |
|
"learning_rate": 4.933947257182901e-05, |
|
"logits/chosen": -3.0249555110931396, |
|
"logits/rejected": -3.0148353576660156, |
|
"logps/chosen": -386.8382263183594, |
|
"logps/rejected": -106.46695709228516, |
|
"loss": 0.3462, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 3.745790481567383, |
|
"rewards/margins": 1.7120976448059082, |
|
"rewards/rejected": 2.0336928367614746, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.002857774678075085, |
|
"grad_norm": 39.60961151123047, |
|
"learning_rate": 4.929902501446366e-05, |
|
"logits/chosen": -3.058694362640381, |
|
"logits/rejected": -3.0021092891693115, |
|
"logps/chosen": -269.97113037109375, |
|
"logps/rejected": -247.43728637695312, |
|
"loss": 1.5676, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 1.230128526687622, |
|
"rewards/margins": -1.297512412071228, |
|
"rewards/rejected": 2.5276408195495605, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.002891795805195026, |
|
"grad_norm": 19.17630958557129, |
|
"learning_rate": 4.925739315689991e-05, |
|
"logits/chosen": -3.1123204231262207, |
|
"logits/rejected": -3.076711416244507, |
|
"logps/chosen": -323.8536376953125, |
|
"logps/rejected": -180.48394775390625, |
|
"loss": 0.5561, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.0740671157836914, |
|
"rewards/margins": 0.7356327772140503, |
|
"rewards/rejected": 1.3384342193603516, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.0029258169323149678, |
|
"grad_norm": 15.29592227935791, |
|
"learning_rate": 4.9214579028215776e-05, |
|
"logits/chosen": -2.995969772338867, |
|
"logits/rejected": -2.975437879562378, |
|
"logps/chosen": -198.14756774902344, |
|
"logps/rejected": -97.39586639404297, |
|
"loss": 0.4645, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.481008529663086, |
|
"rewards/margins": 0.6242896914482117, |
|
"rewards/rejected": 1.8567187786102295, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.002959838059434909, |
|
"grad_norm": 37.80276870727539, |
|
"learning_rate": 4.917058471511149e-05, |
|
"logits/chosen": -3.0480551719665527, |
|
"logits/rejected": -3.0806996822357178, |
|
"logps/chosen": -224.3466796875, |
|
"logps/rejected": -420.4508361816406, |
|
"loss": 1.1967, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 1.6259909868240356, |
|
"rewards/margins": -0.6426498889923096, |
|
"rewards/rejected": 2.2686409950256348, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.0029938591865548508, |
|
"grad_norm": 20.813907623291016, |
|
"learning_rate": 4.912541236180779e-05, |
|
"logits/chosen": -3.045354127883911, |
|
"logits/rejected": -3.0551252365112305, |
|
"logps/chosen": -190.39498901367188, |
|
"logps/rejected": -417.9426574707031, |
|
"loss": 0.4819, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.733407497406006, |
|
"rewards/margins": 0.5519695281982422, |
|
"rewards/rejected": 2.1814379692077637, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.003027880313674792, |
|
"grad_norm": 18.250917434692383, |
|
"learning_rate": 4.907906416994146e-05, |
|
"logits/chosen": -3.1118156909942627, |
|
"logits/rejected": -3.124022960662842, |
|
"logps/chosen": -423.84698486328125, |
|
"logps/rejected": -217.9359130859375, |
|
"loss": 0.3913, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.9902727603912354, |
|
"rewards/margins": 1.3608616590499878, |
|
"rewards/rejected": 1.629410982131958, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.0030619014407947337, |
|
"grad_norm": 18.630115509033203, |
|
"learning_rate": 4.9031542398457974e-05, |
|
"logits/chosen": -3.0090408325195312, |
|
"logits/rejected": -3.0369369983673096, |
|
"logps/chosen": -435.40478515625, |
|
"logps/rejected": -191.86627197265625, |
|
"loss": 0.7079, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.947678327560425, |
|
"rewards/margins": 0.8492887020111084, |
|
"rewards/rejected": 2.0983893871307373, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.003095922567914675, |
|
"grad_norm": 19.552978515625, |
|
"learning_rate": 4.898284936350144e-05, |
|
"logits/chosen": -3.0316624641418457, |
|
"logits/rejected": -3.011552572250366, |
|
"logps/chosen": -432.43402099609375, |
|
"logps/rejected": -214.85935974121094, |
|
"loss": 0.7816, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.7960076332092285, |
|
"rewards/margins": 0.5724576115608215, |
|
"rewards/rejected": 2.2235498428344727, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.0031299436950346163, |
|
"grad_norm": 7.452019691467285, |
|
"learning_rate": 4.893298743830168e-05, |
|
"logits/chosen": -3.038140296936035, |
|
"logits/rejected": -3.00756573677063, |
|
"logps/chosen": -168.3916473388672, |
|
"logps/rejected": -119.79571533203125, |
|
"loss": 0.2097, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.8402457237243652, |
|
"rewards/margins": 1.5049405097961426, |
|
"rewards/rejected": 1.3353054523468018, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.003163964822154558, |
|
"grad_norm": 16.27883529663086, |
|
"learning_rate": 4.888195905305859e-05, |
|
"logits/chosen": -3.030277729034424, |
|
"logits/rejected": -3.0196878910064697, |
|
"logps/chosen": -295.2430114746094, |
|
"logps/rejected": -229.96961975097656, |
|
"loss": 0.5378, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.0215210914611816, |
|
"rewards/margins": 0.6857559084892273, |
|
"rewards/rejected": 1.3357652425765991, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.0031979859492744993, |
|
"grad_norm": 27.870994567871094, |
|
"learning_rate": 4.882976669482367e-05, |
|
"logits/chosen": -3.0157508850097656, |
|
"logits/rejected": -3.062469005584717, |
|
"logps/chosen": -594.1408081054688, |
|
"logps/rejected": -284.8695068359375, |
|
"loss": 0.829, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 2.395214796066284, |
|
"rewards/margins": -0.06876510381698608, |
|
"rewards/rejected": 2.463979959487915, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.003232007076394441, |
|
"grad_norm": 34.57992172241211, |
|
"learning_rate": 4.877641290737884e-05, |
|
"logits/chosen": -3.0241267681121826, |
|
"logits/rejected": -3.072436809539795, |
|
"logps/chosen": -206.46072387695312, |
|
"logps/rejected": -490.99163818359375, |
|
"loss": 0.9507, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 3.0148510932922363, |
|
"rewards/margins": -0.41296035051345825, |
|
"rewards/rejected": 3.427811622619629, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.0032660282035143822, |
|
"grad_norm": 31.449804306030273, |
|
"learning_rate": 4.8721900291112415e-05, |
|
"logits/chosen": -2.978365659713745, |
|
"logits/rejected": -2.9975037574768066, |
|
"logps/chosen": -353.8739929199219, |
|
"logps/rejected": -265.1501159667969, |
|
"loss": 0.7224, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 3.355400562286377, |
|
"rewards/margins": 0.41021838784217834, |
|
"rewards/rejected": 2.9451825618743896, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.003300049330634324, |
|
"grad_norm": 20.54676628112793, |
|
"learning_rate": 4.8666231502892415e-05, |
|
"logits/chosen": -3.0715813636779785, |
|
"logits/rejected": -3.0223703384399414, |
|
"logps/chosen": -417.56280517578125, |
|
"logps/rejected": -126.94705963134766, |
|
"loss": 0.5149, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.667041540145874, |
|
"rewards/margins": 0.9312877655029297, |
|
"rewards/rejected": 1.7357536554336548, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.0033340704577542652, |
|
"grad_norm": 34.95238494873047, |
|
"learning_rate": 4.860940925593703e-05, |
|
"logits/chosen": -3.038242816925049, |
|
"logits/rejected": -3.026169776916504, |
|
"logps/chosen": -320.06005859375, |
|
"logps/rejected": -534.7603759765625, |
|
"loss": 0.6213, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 3.1372852325439453, |
|
"rewards/margins": 0.3597479462623596, |
|
"rewards/rejected": 2.7775373458862305, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.003368091584874207, |
|
"grad_norm": 22.04984474182129, |
|
"learning_rate": 4.855143631968242e-05, |
|
"logits/chosen": -3.0394222736358643, |
|
"logits/rejected": -3.040712356567383, |
|
"logps/chosen": -405.9694519042969, |
|
"logps/rejected": -224.75143432617188, |
|
"loss": 0.472, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 3.6535189151763916, |
|
"rewards/margins": 1.3992540836334229, |
|
"rewards/rejected": 2.2542648315429688, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.003402112711994148, |
|
"grad_norm": 10.456045150756836, |
|
"learning_rate": 4.849231551964771e-05, |
|
"logits/chosen": -3.0558083057403564, |
|
"logits/rejected": -3.0164895057678223, |
|
"logps/chosen": -215.77474975585938, |
|
"logps/rejected": -114.40177917480469, |
|
"loss": 0.3431, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 3.2260184288024902, |
|
"rewards/margins": 1.2816791534423828, |
|
"rewards/rejected": 1.9443392753601074, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.00343613383911409, |
|
"grad_norm": 26.428396224975586, |
|
"learning_rate": 4.843204973729729e-05, |
|
"logits/chosen": -2.9279229640960693, |
|
"logits/rejected": -2.941478729248047, |
|
"logps/chosen": -344.0908203125, |
|
"logps/rejected": -369.3677978515625, |
|
"loss": 0.851, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 4.361503601074219, |
|
"rewards/margins": 2.4121603965759277, |
|
"rewards/rejected": 1.949343204498291, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.003470154966234031, |
|
"grad_norm": 20.163902282714844, |
|
"learning_rate": 4.837064190990036e-05, |
|
"logits/chosen": -2.994917869567871, |
|
"logits/rejected": -2.9446635246276855, |
|
"logps/chosen": -348.66937255859375, |
|
"logps/rejected": -345.49578857421875, |
|
"loss": 0.4779, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 3.1643588542938232, |
|
"rewards/margins": 1.304919958114624, |
|
"rewards/rejected": 1.8594388961791992, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.003504176093353973, |
|
"grad_norm": 10.98054027557373, |
|
"learning_rate": 4.830809503038781e-05, |
|
"logits/chosen": -2.9521918296813965, |
|
"logits/rejected": -2.9645028114318848, |
|
"logps/chosen": -226.67506408691406, |
|
"logps/rejected": -326.85455322265625, |
|
"loss": 0.1768, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 3.5958943367004395, |
|
"rewards/margins": 1.691298007965088, |
|
"rewards/rejected": 1.904596209526062, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.003538197220473914, |
|
"grad_norm": 10.843141555786133, |
|
"learning_rate": 4.8244412147206284e-05, |
|
"logits/chosen": -2.9416680335998535, |
|
"logits/rejected": -2.935615062713623, |
|
"logps/chosen": -242.29098510742188, |
|
"logps/rejected": -196.54017639160156, |
|
"loss": 0.4134, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.624265670776367, |
|
"rewards/margins": 1.845750331878662, |
|
"rewards/rejected": 0.778515100479126, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.003572218347593856, |
|
"grad_norm": 51.45275115966797, |
|
"learning_rate": 4.817959636416969e-05, |
|
"logits/chosen": -2.987889289855957, |
|
"logits/rejected": -2.9992475509643555, |
|
"logps/chosen": -265.65509033203125, |
|
"logps/rejected": -504.4512023925781, |
|
"loss": 2.0434, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 1.9749891757965088, |
|
"rewards/margins": -1.6412996053695679, |
|
"rewards/rejected": 3.6162891387939453, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.003606239474713797, |
|
"grad_norm": 34.740169525146484, |
|
"learning_rate": 4.8113650840307834e-05, |
|
"logits/chosen": -2.9495849609375, |
|
"logits/rejected": -2.9650793075561523, |
|
"logps/chosen": -267.8492126464844, |
|
"logps/rejected": -229.05166625976562, |
|
"loss": 1.4427, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 1.6815515756607056, |
|
"rewards/margins": -1.025644063949585, |
|
"rewards/rejected": 2.70719575881958, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.003640260601833739, |
|
"grad_norm": 33.90592956542969, |
|
"learning_rate": 4.8046578789712515e-05, |
|
"logits/chosen": -2.9517102241516113, |
|
"logits/rejected": -2.892892360687256, |
|
"logps/chosen": -271.4212646484375, |
|
"logps/rejected": -432.76318359375, |
|
"loss": 0.9206, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.004805564880371, |
|
"rewards/margins": -0.28335341811180115, |
|
"rewards/rejected": 2.288158893585205, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.00367428172895368, |
|
"grad_norm": 26.77821922302246, |
|
"learning_rate": 4.797838348138086e-05, |
|
"logits/chosen": -2.961052656173706, |
|
"logits/rejected": -2.963571548461914, |
|
"logps/chosen": -235.24127197265625, |
|
"logps/rejected": -195.713623046875, |
|
"loss": 0.9046, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 0.960392951965332, |
|
"rewards/margins": -0.3249235153198242, |
|
"rewards/rejected": 1.2853164672851562, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.003708302856073622, |
|
"grad_norm": 20.291223526000977, |
|
"learning_rate": 4.790906823905599e-05, |
|
"logits/chosen": -2.922367811203003, |
|
"logits/rejected": -2.9089303016662598, |
|
"logps/chosen": -432.6478576660156, |
|
"logps/rejected": -147.66995239257812, |
|
"loss": 0.5442, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.312082052230835, |
|
"rewards/margins": 0.7899919748306274, |
|
"rewards/rejected": 1.522089958190918, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.003742323983193563, |
|
"grad_norm": 21.555723190307617, |
|
"learning_rate": 4.783863644106502e-05, |
|
"logits/chosen": -2.893828868865967, |
|
"logits/rejected": -2.9130947589874268, |
|
"logps/chosen": -188.0413360595703, |
|
"logps/rejected": -163.48797607421875, |
|
"loss": 0.7379, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 1.8718959093093872, |
|
"rewards/margins": 0.09508019685745239, |
|
"rewards/rejected": 1.77681565284729, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.003776345110313505, |
|
"grad_norm": 32.17390060424805, |
|
"learning_rate": 4.776709152015443e-05, |
|
"logits/chosen": -2.9426565170288086, |
|
"logits/rejected": -2.961993455886841, |
|
"logps/chosen": -292.1125183105469, |
|
"logps/rejected": -345.0081481933594, |
|
"loss": 1.1772, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.0569343566894531, |
|
"rewards/margins": -0.44999200105667114, |
|
"rewards/rejected": 1.5069262981414795, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.003810366237433446, |
|
"grad_norm": 33.43811798095703, |
|
"learning_rate": 4.769443696332272e-05, |
|
"logits/chosen": -2.935262680053711, |
|
"logits/rejected": -2.9003987312316895, |
|
"logps/chosen": -299.31158447265625, |
|
"logps/rejected": -369.65911865234375, |
|
"loss": 0.9117, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.7615939378738403, |
|
"rewards/margins": 0.04655948281288147, |
|
"rewards/rejected": 1.7150346040725708, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.003844387364553388, |
|
"grad_norm": 14.72221851348877, |
|
"learning_rate": 4.762067631165049e-05, |
|
"logits/chosen": -2.9283885955810547, |
|
"logits/rejected": -2.9265716075897217, |
|
"logps/chosen": -172.1368865966797, |
|
"logps/rejected": -198.41604614257812, |
|
"loss": 0.4199, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.7335453033447266, |
|
"rewards/margins": 0.9735270738601685, |
|
"rewards/rejected": 1.7600181102752686, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.003878408491673329, |
|
"grad_norm": 26.454801559448242, |
|
"learning_rate": 4.754581316012785e-05, |
|
"logits/chosen": -3.0016942024230957, |
|
"logits/rejected": -2.9508540630340576, |
|
"logps/chosen": -301.06683349609375, |
|
"logps/rejected": -207.55111694335938, |
|
"loss": 0.6857, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.0953211784362793, |
|
"rewards/margins": 0.12809863686561584, |
|
"rewards/rejected": 1.9672226905822754, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.00391242961879327, |
|
"grad_norm": 25.513513565063477, |
|
"learning_rate": 4.7469851157479177e-05, |
|
"logits/chosen": -2.9934637546539307, |
|
"logits/rejected": -3.0119104385375977, |
|
"logps/chosen": -660.9867553710938, |
|
"logps/rejected": -195.93348693847656, |
|
"loss": 0.5308, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.8832085132598877, |
|
"rewards/margins": 0.5043414831161499, |
|
"rewards/rejected": 2.3788669109344482, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.003946450745913212, |
|
"grad_norm": 16.804649353027344, |
|
"learning_rate": 4.7392794005985326e-05, |
|
"logits/chosen": -2.970472574234009, |
|
"logits/rejected": -2.901583194732666, |
|
"logps/chosen": -214.0200958251953, |
|
"logps/rejected": -311.00836181640625, |
|
"loss": 0.5421, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.2354931831359863, |
|
"rewards/margins": 0.49688705801963806, |
|
"rewards/rejected": 1.7386059761047363, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.003980471873033154, |
|
"grad_norm": 23.340364456176758, |
|
"learning_rate": 4.731464546130314e-05, |
|
"logits/chosen": -3.0023207664489746, |
|
"logits/rejected": -2.985395908355713, |
|
"logps/chosen": -91.83181762695312, |
|
"logps/rejected": -107.75381469726562, |
|
"loss": 1.0601, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 1.4270564317703247, |
|
"rewards/margins": -0.45852407813072205, |
|
"rewards/rejected": 1.88558030128479, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.0040144930001530955, |
|
"grad_norm": 18.47270965576172, |
|
"learning_rate": 4.723540933228244e-05, |
|
"logits/chosen": -2.899160861968994, |
|
"logits/rejected": -2.9231584072113037, |
|
"logps/chosen": -432.610595703125, |
|
"logps/rejected": -297.2770690917969, |
|
"loss": 0.5938, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 3.0332915782928467, |
|
"rewards/margins": 1.6466760635375977, |
|
"rewards/rejected": 1.3866153955459595, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.004048514127273036, |
|
"grad_norm": 27.180835723876953, |
|
"learning_rate": 4.715508948078037e-05, |
|
"logits/chosen": -3.047947645187378, |
|
"logits/rejected": -3.05257248878479, |
|
"logps/chosen": -174.9864959716797, |
|
"logps/rejected": -159.964111328125, |
|
"loss": 1.3, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.1789213418960571, |
|
"rewards/margins": -0.4154563546180725, |
|
"rewards/rejected": 1.5943776369094849, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.004082535254392978, |
|
"grad_norm": 36.64414978027344, |
|
"learning_rate": 4.707368982147318e-05, |
|
"logits/chosen": -3.0685150623321533, |
|
"logits/rejected": -3.0044500827789307, |
|
"logps/chosen": -522.5695190429688, |
|
"logps/rejected": -213.46484375, |
|
"loss": 0.7759, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.9243935346603394, |
|
"rewards/margins": 0.016531378030776978, |
|
"rewards/rejected": 1.9078620672225952, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.00411655638151292, |
|
"grad_norm": 7.666904449462891, |
|
"learning_rate": 4.6991214321665414e-05, |
|
"logits/chosen": -2.9543278217315674, |
|
"logits/rejected": -2.9607834815979004, |
|
"logps/chosen": -258.1210021972656, |
|
"logps/rejected": -198.5812225341797, |
|
"loss": 0.1974, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.4223241806030273, |
|
"rewards/margins": 2.2252554893493652, |
|
"rewards/rejected": 0.1970686912536621, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.0041505775086328614, |
|
"grad_norm": 19.53645896911621, |
|
"learning_rate": 4.690766700109659e-05, |
|
"logits/chosen": -2.9542903900146484, |
|
"logits/rejected": -2.9877171516418457, |
|
"logps/chosen": -532.9715576171875, |
|
"logps/rejected": -224.84103393554688, |
|
"loss": 0.4557, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.20086407661438, |
|
"rewards/margins": 0.7117323875427246, |
|
"rewards/rejected": 1.4891316890716553, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.004184598635752802, |
|
"grad_norm": 16.438650131225586, |
|
"learning_rate": 4.682305193174524e-05, |
|
"logits/chosen": -3.004117727279663, |
|
"logits/rejected": -2.9864871501922607, |
|
"logps/chosen": -207.78421020507812, |
|
"logps/rejected": -124.6191635131836, |
|
"loss": 0.5509, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.6520434617996216, |
|
"rewards/margins": 0.7864970564842224, |
|
"rewards/rejected": 0.865546464920044, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.004218619762872744, |
|
"grad_norm": 14.840346336364746, |
|
"learning_rate": 4.6737373237630476e-05, |
|
"logits/chosen": -3.018615245819092, |
|
"logits/rejected": -3.0123519897460938, |
|
"logps/chosen": -115.13555908203125, |
|
"logps/rejected": -88.72488403320312, |
|
"loss": 0.5478, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.4716414213180542, |
|
"rewards/margins": 0.9200279712677002, |
|
"rewards/rejected": 0.5516133308410645, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.004252640889992686, |
|
"grad_norm": 25.749723434448242, |
|
"learning_rate": 4.665063509461097e-05, |
|
"logits/chosen": -3.0335187911987305, |
|
"logits/rejected": -2.9879348278045654, |
|
"logps/chosen": -428.715576171875, |
|
"logps/rejected": -315.72943115234375, |
|
"loss": 0.9619, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.7211081981658936, |
|
"rewards/margins": 0.08669501543045044, |
|
"rewards/rejected": 2.634413242340088, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.0042866620171126265, |
|
"grad_norm": 11.342658042907715, |
|
"learning_rate": 4.656284173018144e-05, |
|
"logits/chosen": -3.0191352367401123, |
|
"logits/rejected": -2.9867730140686035, |
|
"logps/chosen": -282.3454895019531, |
|
"logps/rejected": -163.94921875, |
|
"loss": 0.3235, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.988835573196411, |
|
"rewards/margins": 1.4542242288589478, |
|
"rewards/rejected": 1.5346113443374634, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.004320683144232568, |
|
"grad_norm": 22.939863204956055, |
|
"learning_rate": 4.6473997423266614e-05, |
|
"logits/chosen": -2.9171319007873535, |
|
"logits/rejected": -2.9400453567504883, |
|
"logps/chosen": -294.5677490234375, |
|
"logps/rejected": -219.3333282470703, |
|
"loss": 0.5542, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.0927131175994873, |
|
"rewards/margins": 0.75679612159729, |
|
"rewards/rejected": 1.3359171152114868, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.00435470427135251, |
|
"grad_norm": 10.7678804397583, |
|
"learning_rate": 4.638410650401267e-05, |
|
"logits/chosen": -3.050967216491699, |
|
"logits/rejected": -2.973550796508789, |
|
"logps/chosen": -427.3642883300781, |
|
"logps/rejected": -214.50796508789062, |
|
"loss": 0.2155, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 3.5660037994384766, |
|
"rewards/margins": 2.3330581188201904, |
|
"rewards/rejected": 1.2329455614089966, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.004388725398472452, |
|
"grad_norm": 25.86681365966797, |
|
"learning_rate": 4.629317335357619e-05, |
|
"logits/chosen": -3.0208535194396973, |
|
"logits/rejected": -3.0417640209198, |
|
"logps/chosen": -103.72763061523438, |
|
"logps/rejected": -151.43617248535156, |
|
"loss": 1.3143, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.1442210078239441, |
|
"rewards/margins": -0.46983587741851807, |
|
"rewards/rejected": 0.32561492919921875, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.0044227465255923925, |
|
"grad_norm": 5.466204643249512, |
|
"learning_rate": 4.620120240391065e-05, |
|
"logits/chosen": -3.016930341720581, |
|
"logits/rejected": -2.971389055252075, |
|
"logps/chosen": -496.30938720703125, |
|
"logps/rejected": -235.5283660888672, |
|
"loss": 0.133, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 4.568811416625977, |
|
"rewards/margins": 2.570054054260254, |
|
"rewards/rejected": 1.9987574815750122, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.004456767652712334, |
|
"grad_norm": 28.28066062927246, |
|
"learning_rate": 4.610819813755038e-05, |
|
"logits/chosen": -3.012406349182129, |
|
"logits/rejected": -3.038177728652954, |
|
"logps/chosen": -166.3546142578125, |
|
"logps/rejected": -136.15911865234375, |
|
"loss": 1.3695, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.7164108753204346, |
|
"rewards/margins": -0.628231942653656, |
|
"rewards/rejected": 2.3446426391601562, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.004490788779832276, |
|
"grad_norm": 23.825101852416992, |
|
"learning_rate": 4.601416508739211e-05, |
|
"logits/chosen": -2.9706547260284424, |
|
"logits/rejected": -2.977961540222168, |
|
"logps/chosen": -177.7816162109375, |
|
"logps/rejected": -162.42628479003906, |
|
"loss": 0.9562, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.5946948528289795, |
|
"rewards/margins": -0.007066547870635986, |
|
"rewards/rejected": 1.6017615795135498, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.004524809906952218, |
|
"grad_norm": 15.178232192993164, |
|
"learning_rate": 4.591910783647404e-05, |
|
"logits/chosen": -3.079338550567627, |
|
"logits/rejected": -3.0130653381347656, |
|
"logps/chosen": -190.42550659179688, |
|
"logps/rejected": -197.51080322265625, |
|
"loss": 0.4386, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.8792479038238525, |
|
"rewards/margins": 0.7111226916313171, |
|
"rewards/rejected": 1.1681251525878906, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.0045588310340721585, |
|
"grad_norm": 30.27988624572754, |
|
"learning_rate": 4.5823031017752485e-05, |
|
"logits/chosen": -3.0107874870300293, |
|
"logits/rejected": -3.0137808322906494, |
|
"logps/chosen": -489.22198486328125, |
|
"logps/rejected": -383.8040466308594, |
|
"loss": 0.7358, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.4854037761688232, |
|
"rewards/margins": 0.40515825152397156, |
|
"rewards/rejected": 2.0802454948425293, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.0045928521611921, |
|
"grad_norm": 36.758949279785156, |
|
"learning_rate": 4.572593931387604e-05, |
|
"logits/chosen": -3.046233654022217, |
|
"logits/rejected": -2.991231679916382, |
|
"logps/chosen": -156.7913818359375, |
|
"logps/rejected": -328.3721923828125, |
|
"loss": 2.1706, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.42976588010787964, |
|
"rewards/margins": -2.022183418273926, |
|
"rewards/rejected": 2.45194935798645, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.004626873288312042, |
|
"grad_norm": 21.850040435791016, |
|
"learning_rate": 4.562783745695738e-05, |
|
"logits/chosen": -2.989779472351074, |
|
"logits/rejected": -2.9347622394561768, |
|
"logps/chosen": -335.12872314453125, |
|
"logps/rejected": -146.79830932617188, |
|
"loss": 0.7109, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.7486190795898438, |
|
"rewards/margins": 0.7050973773002625, |
|
"rewards/rejected": 1.043521761894226, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.004660894415431984, |
|
"grad_norm": 44.39814758300781, |
|
"learning_rate": 4.5528730228342605e-05, |
|
"logits/chosen": -2.944559097290039, |
|
"logits/rejected": -2.9720685482025146, |
|
"logps/chosen": -269.68988037109375, |
|
"logps/rejected": -548.735595703125, |
|
"loss": 2.2232, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 1.4295103549957275, |
|
"rewards/margins": -1.3700931072235107, |
|
"rewards/rejected": 2.7996034622192383, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.004694915542551924, |
|
"grad_norm": 9.128776550292969, |
|
"learning_rate": 4.542862245837821e-05, |
|
"logits/chosen": -3.013176441192627, |
|
"logits/rejected": -3.0005812644958496, |
|
"logps/chosen": -226.3272247314453, |
|
"logps/rejected": -123.2714614868164, |
|
"loss": 0.3682, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.1151604652404785, |
|
"rewards/margins": 1.2523483037948608, |
|
"rewards/rejected": -0.13718795776367188, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.004728936669671866, |
|
"grad_norm": 14.181815147399902, |
|
"learning_rate": 4.532751902617569e-05, |
|
"logits/chosen": -2.985856533050537, |
|
"logits/rejected": -2.995152235031128, |
|
"logps/chosen": -346.6661376953125, |
|
"logps/rejected": -293.36962890625, |
|
"loss": 0.2832, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 3.5159454345703125, |
|
"rewards/margins": 1.9295426607131958, |
|
"rewards/rejected": 1.5864028930664062, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.004762957796791808, |
|
"grad_norm": 32.54734802246094, |
|
"learning_rate": 4.522542485937369e-05, |
|
"logits/chosen": -3.041367530822754, |
|
"logits/rejected": -3.034242630004883, |
|
"logps/chosen": -180.97171020507812, |
|
"logps/rejected": -217.47726440429688, |
|
"loss": 1.6368, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 0.3136112093925476, |
|
"rewards/margins": -1.286853313446045, |
|
"rewards/rejected": 1.6004644632339478, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.0047969789239117495, |
|
"grad_norm": 22.16848373413086, |
|
"learning_rate": 4.512234493389785e-05, |
|
"logits/chosen": -3.0714263916015625, |
|
"logits/rejected": -3.0722179412841797, |
|
"logps/chosen": -366.7979431152344, |
|
"logps/rejected": -219.58306884765625, |
|
"loss": 0.5429, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.0975884199142456, |
|
"rewards/margins": 0.6359337568283081, |
|
"rewards/rejected": 0.4616546630859375, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.00483100005103169, |
|
"grad_norm": 27.500123977661133, |
|
"learning_rate": 4.5018284273718336e-05, |
|
"logits/chosen": -2.957406997680664, |
|
"logits/rejected": -2.958174705505371, |
|
"logps/chosen": -609.5079345703125, |
|
"logps/rejected": -272.30181884765625, |
|
"loss": 0.647, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.9717493057250977, |
|
"rewards/margins": 1.2977182865142822, |
|
"rewards/rejected": 1.6740310192108154, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.004865021178151632, |
|
"grad_norm": 16.75972557067871, |
|
"learning_rate": 4.491324795060491e-05, |
|
"logits/chosen": -3.080034017562866, |
|
"logits/rejected": -3.0754172801971436, |
|
"logps/chosen": -376.17236328125, |
|
"logps/rejected": -203.8787841796875, |
|
"loss": 0.38, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.3156166076660156, |
|
"rewards/margins": 1.1856859922409058, |
|
"rewards/rejected": 1.1299307346343994, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.004899042305271574, |
|
"grad_norm": 13.316751480102539, |
|
"learning_rate": 4.480724108387977e-05, |
|
"logits/chosen": -3.0658347606658936, |
|
"logits/rejected": -3.0963737964630127, |
|
"logps/chosen": -282.4508056640625, |
|
"logps/rejected": -191.8719482421875, |
|
"loss": 0.3433, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.7471672296524048, |
|
"rewards/margins": 1.0950264930725098, |
|
"rewards/rejected": -0.3478591740131378, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.0049330634323915155, |
|
"grad_norm": 29.934419631958008, |
|
"learning_rate": 4.4700268840168045e-05, |
|
"logits/chosen": -2.9548258781433105, |
|
"logits/rejected": -3.0186848640441895, |
|
"logps/chosen": -405.0225830078125, |
|
"logps/rejected": -353.18408203125, |
|
"loss": 1.5146, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.42316895723342896, |
|
"rewards/margins": -0.660403847694397, |
|
"rewards/rejected": 1.0835727453231812, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.004967084559511456, |
|
"grad_norm": 30.49190902709961, |
|
"learning_rate": 4.4592336433146e-05, |
|
"logits/chosen": -3.0669848918914795, |
|
"logits/rejected": -3.057908773422241, |
|
"logps/chosen": -325.12432861328125, |
|
"logps/rejected": -398.8703918457031, |
|
"loss": 0.7205, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.377676248550415, |
|
"rewards/margins": 0.8879678845405579, |
|
"rewards/rejected": 1.4897081851959229, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.005001105686631398, |
|
"grad_norm": 15.67647647857666, |
|
"learning_rate": 4.448344912328686e-05, |
|
"logits/chosen": -3.0414624214172363, |
|
"logits/rejected": -3.0715274810791016, |
|
"logps/chosen": -338.4504089355469, |
|
"logps/rejected": -118.74220275878906, |
|
"loss": 0.5679, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.4492450952529907, |
|
"rewards/margins": 0.8044933080673218, |
|
"rewards/rejected": -0.35524818301200867, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.00503512681375134, |
|
"grad_norm": 21.485815048217773, |
|
"learning_rate": 4.4373612217604496e-05, |
|
"logits/chosen": -3.09837007522583, |
|
"logits/rejected": -3.157946825027466, |
|
"logps/chosen": -376.14630126953125, |
|
"logps/rejected": -201.80422973632812, |
|
"loss": 0.4138, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.0123474597930908, |
|
"rewards/margins": 1.332427740097046, |
|
"rewards/rejected": -0.3200801908969879, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.0050691479408712815, |
|
"grad_norm": 33.10397720336914, |
|
"learning_rate": 4.426283106939474e-05, |
|
"logits/chosen": -3.089867115020752, |
|
"logits/rejected": -3.1238722801208496, |
|
"logps/chosen": -452.83819580078125, |
|
"logps/rejected": -128.71438598632812, |
|
"loss": 0.9454, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.6251472234725952, |
|
"rewards/margins": -0.13446199893951416, |
|
"rewards/rejected": -0.49068525433540344, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.005103169067991222, |
|
"grad_norm": 17.779081344604492, |
|
"learning_rate": 4.415111107797445e-05, |
|
"logits/chosen": -3.123142957687378, |
|
"logits/rejected": -3.123056173324585, |
|
"logps/chosen": -251.64035034179688, |
|
"logps/rejected": -185.87037658691406, |
|
"loss": 0.6144, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.13248348236083984, |
|
"rewards/margins": 0.6361640691757202, |
|
"rewards/rejected": -0.5036805868148804, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.005137190195111164, |
|
"grad_norm": 29.36853790283203, |
|
"learning_rate": 4.403845768841842e-05, |
|
"logits/chosen": -3.114872455596924, |
|
"logits/rejected": -3.1005492210388184, |
|
"logps/chosen": -352.29071044921875, |
|
"logps/rejected": -248.60684204101562, |
|
"loss": 0.9394, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.9047115445137024, |
|
"rewards/margins": 0.29889756441116333, |
|
"rewards/rejected": 0.6058140397071838, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.005171211322231106, |
|
"grad_norm": 37.393760681152344, |
|
"learning_rate": 4.3924876391293915e-05, |
|
"logits/chosen": -3.100764513015747, |
|
"logits/rejected": -3.0552031993865967, |
|
"logps/chosen": -176.60968017578125, |
|
"logps/rejected": -177.10955810546875, |
|
"loss": 1.7618, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": -1.2755136489868164, |
|
"rewards/margins": -1.5124303102493286, |
|
"rewards/rejected": 0.23691676557064056, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.005205232449351047, |
|
"grad_norm": 20.11163330078125, |
|
"learning_rate": 4.381037272239311e-05, |
|
"logits/chosen": -3.0348846912384033, |
|
"logits/rejected": -3.0846118927001953, |
|
"logps/chosen": -458.5091857910156, |
|
"logps/rejected": -223.36878967285156, |
|
"loss": 0.5437, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.8197979927062988, |
|
"rewards/margins": 1.692866563796997, |
|
"rewards/rejected": -0.873068630695343, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.005239253576470988, |
|
"grad_norm": 32.02238464355469, |
|
"learning_rate": 4.36949522624633e-05, |
|
"logits/chosen": -3.030576705932617, |
|
"logits/rejected": -3.0167465209960938, |
|
"logps/chosen": -238.9589385986328, |
|
"logps/rejected": -289.1723327636719, |
|
"loss": 1.1852, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.3048284649848938, |
|
"rewards/margins": 0.32317912578582764, |
|
"rewards/rejected": -0.018350541591644287, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.00527327470359093, |
|
"grad_norm": 10.38000202178955, |
|
"learning_rate": 4.357862063693486e-05, |
|
"logits/chosen": -3.062821865081787, |
|
"logits/rejected": -3.0488338470458984, |
|
"logps/chosen": -569.1879272460938, |
|
"logps/rejected": -318.0606994628906, |
|
"loss": 0.207, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 1.9250755310058594, |
|
"rewards/margins": 2.6815903186798096, |
|
"rewards/rejected": -0.7565147280693054, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.005307295830710872, |
|
"grad_norm": 21.376619338989258, |
|
"learning_rate": 4.3461383515647106e-05, |
|
"logits/chosen": -3.0080904960632324, |
|
"logits/rejected": -3.0347626209259033, |
|
"logps/chosen": -714.813232421875, |
|
"logps/rejected": -279.81219482421875, |
|
"loss": 0.4824, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.5823137164115906, |
|
"rewards/margins": 0.8629709482192993, |
|
"rewards/rejected": -0.28065717220306396, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.0053413169578308125, |
|
"grad_norm": 16.18117332458496, |
|
"learning_rate": 4.334324661257191e-05, |
|
"logits/chosen": -3.165111780166626, |
|
"logits/rejected": -3.1727468967437744, |
|
"logps/chosen": -135.48158264160156, |
|
"logps/rejected": -165.6737823486328, |
|
"loss": 0.4282, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.3546949326992035, |
|
"rewards/margins": 0.8620704412460327, |
|
"rewards/rejected": -0.5073755383491516, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.005375338084950754, |
|
"grad_norm": 19.98881721496582, |
|
"learning_rate": 4.3224215685535294e-05, |
|
"logits/chosen": -3.125204563140869, |
|
"logits/rejected": -3.1065356731414795, |
|
"logps/chosen": -155.87408447265625, |
|
"logps/rejected": -92.91828155517578, |
|
"loss": 0.8494, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.582575798034668, |
|
"rewards/margins": 0.04557628929615021, |
|
"rewards/rejected": -0.6281521320343018, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.005409359212070696, |
|
"grad_norm": 24.46631622314453, |
|
"learning_rate": 4.3104296535936695e-05, |
|
"logits/chosen": -3.017587184906006, |
|
"logits/rejected": -3.0635039806365967, |
|
"logps/chosen": -424.2115478515625, |
|
"logps/rejected": -233.75881958007812, |
|
"loss": 1.0511, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.9063742160797119, |
|
"rewards/margins": 2.066019296646118, |
|
"rewards/rejected": -1.1596449613571167, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.005443380339190638, |
|
"grad_norm": 19.825090408325195, |
|
"learning_rate": 4.2983495008466276e-05, |
|
"logits/chosen": -3.1046340465545654, |
|
"logits/rejected": -3.111210346221924, |
|
"logps/chosen": -762.91015625, |
|
"logps/rejected": -304.2664489746094, |
|
"loss": 0.3557, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.6634994745254517, |
|
"rewards/margins": 1.7218773365020752, |
|
"rewards/rejected": -1.0583778619766235, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.0054774014663105785, |
|
"grad_norm": 18.325929641723633, |
|
"learning_rate": 4.2861816990820084e-05, |
|
"logits/chosen": -3.143399238586426, |
|
"logits/rejected": -3.0676584243774414, |
|
"logps/chosen": -557.0424194335938, |
|
"logps/rejected": -183.75619506835938, |
|
"loss": 0.4645, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.5502029657363892, |
|
"rewards/margins": 1.1281495094299316, |
|
"rewards/rejected": -0.577946662902832, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.00551142259343052, |
|
"grad_norm": 26.475664138793945, |
|
"learning_rate": 4.273926841341302e-05, |
|
"logits/chosen": -3.0684218406677246, |
|
"logits/rejected": -3.061774492263794, |
|
"logps/chosen": -616.07763671875, |
|
"logps/rejected": -462.61627197265625, |
|
"loss": 0.6446, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.3071922063827515, |
|
"rewards/margins": 1.3884514570236206, |
|
"rewards/rejected": -0.08125917613506317, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.005545443720550462, |
|
"grad_norm": 33.35295486450195, |
|
"learning_rate": 4.261585524908987e-05, |
|
"logits/chosen": -3.061275005340576, |
|
"logits/rejected": -3.0361580848693848, |
|
"logps/chosen": -632.3155517578125, |
|
"logps/rejected": -289.16925048828125, |
|
"loss": 0.7509, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.2999813556671143, |
|
"rewards/margins": 1.7425051927566528, |
|
"rewards/rejected": 0.5574760437011719, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.005579464847670404, |
|
"grad_norm": 15.048567771911621, |
|
"learning_rate": 4.249158351283414e-05, |
|
"logits/chosen": -3.075890064239502, |
|
"logits/rejected": -3.026381015777588, |
|
"logps/chosen": -254.39315795898438, |
|
"logps/rejected": -182.19528198242188, |
|
"loss": 0.3302, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.07112771272659302, |
|
"rewards/margins": 1.416032075881958, |
|
"rewards/rejected": -1.3449044227600098, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.0056134859747903445, |
|
"grad_norm": 41.53990173339844, |
|
"learning_rate": 4.2366459261474933e-05, |
|
"logits/chosen": -3.0547707080841064, |
|
"logits/rejected": -3.0147359371185303, |
|
"logps/chosen": -468.70928955078125, |
|
"logps/rejected": -353.0340881347656, |
|
"loss": 1.4417, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 0.5168285369873047, |
|
"rewards/margins": -0.23437988758087158, |
|
"rewards/rejected": 0.7512083649635315, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.005647507101910286, |
|
"grad_norm": 5.120975494384766, |
|
"learning_rate": 4.224048859339175e-05, |
|
"logits/chosen": -2.974299192428589, |
|
"logits/rejected": -3.049945592880249, |
|
"logps/chosen": -579.917236328125, |
|
"logps/rejected": -196.25302124023438, |
|
"loss": 0.0918, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 3.5269789695739746, |
|
"rewards/margins": 3.765716552734375, |
|
"rewards/rejected": -0.23873747885227203, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.005681528229030228, |
|
"grad_norm": 31.05389404296875, |
|
"learning_rate": 4.211367764821722e-05, |
|
"logits/chosen": -3.0451102256774902, |
|
"logits/rejected": -3.0483558177948, |
|
"logps/chosen": -369.182373046875, |
|
"logps/rejected": -175.77587890625, |
|
"loss": 0.8954, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -1.2048249244689941, |
|
"rewards/margins": 0.31681936979293823, |
|
"rewards/rejected": -1.5216444730758667, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.00571554935615017, |
|
"grad_norm": 17.161266326904297, |
|
"learning_rate": 4.198603260653792e-05, |
|
"logits/chosen": -2.9860920906066895, |
|
"logits/rejected": -3.027653455734253, |
|
"logps/chosen": -256.9324035644531, |
|
"logps/rejected": -334.290771484375, |
|
"loss": 0.4836, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.07286377251148224, |
|
"rewards/margins": 0.6210994720458984, |
|
"rewards/rejected": -0.6939632892608643, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.00574957048327011, |
|
"grad_norm": 14.865567207336426, |
|
"learning_rate": 4.185755968959308e-05, |
|
"logits/chosen": -3.0316944122314453, |
|
"logits/rejected": -3.0366439819335938, |
|
"logps/chosen": -259.6930236816406, |
|
"logps/rejected": -163.08534240722656, |
|
"loss": 0.7172, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.03256016969680786, |
|
"rewards/margins": 0.8183696866035461, |
|
"rewards/rejected": -0.7858095169067383, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.005783591610390052, |
|
"grad_norm": 25.655973434448242, |
|
"learning_rate": 4.172826515897146e-05, |
|
"logits/chosen": -3.0672049522399902, |
|
"logits/rejected": -3.027251720428467, |
|
"logps/chosen": -435.9547119140625, |
|
"logps/rejected": -167.56182861328125, |
|
"loss": 0.8137, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.17178192734718323, |
|
"rewards/margins": 0.6809127926826477, |
|
"rewards/rejected": -0.5091308951377869, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.005817612737509994, |
|
"grad_norm": 38.359947204589844, |
|
"learning_rate": 4.1598155316306044e-05, |
|
"logits/chosen": -3.0813450813293457, |
|
"logits/rejected": -2.9670357704162598, |
|
"logps/chosen": -474.51318359375, |
|
"logps/rejected": -419.6246032714844, |
|
"loss": 0.5563, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 3.3214197158813477, |
|
"rewards/margins": 3.06449818611145, |
|
"rewards/rejected": 0.2569214105606079, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.0058516338646299355, |
|
"grad_norm": 38.196041107177734, |
|
"learning_rate": 4.146723650296701e-05, |
|
"logits/chosen": -3.083744764328003, |
|
"logits/rejected": -3.0692083835601807, |
|
"logps/chosen": -319.418212890625, |
|
"logps/rejected": -399.4487609863281, |
|
"loss": 1.3367, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 0.23198705911636353, |
|
"rewards/margins": -0.49163129925727844, |
|
"rewards/rejected": 0.7236183285713196, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.005885654991749876, |
|
"grad_norm": 30.843332290649414, |
|
"learning_rate": 4.133551509975264e-05, |
|
"logits/chosen": -3.1385626792907715, |
|
"logits/rejected": -3.1076760292053223, |
|
"logps/chosen": -159.91226196289062, |
|
"logps/rejected": -186.83062744140625, |
|
"loss": 1.4466, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -1.123528003692627, |
|
"rewards/margins": -0.7851051092147827, |
|
"rewards/rejected": -0.3384227454662323, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.005919676118869818, |
|
"grad_norm": 21.5479679107666, |
|
"learning_rate": 4.1202997526578276e-05, |
|
"logits/chosen": -3.0526177883148193, |
|
"logits/rejected": -3.051259994506836, |
|
"logps/chosen": -252.21401977539062, |
|
"logps/rejected": -222.28805541992188, |
|
"loss": 0.8296, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.05638125538825989, |
|
"rewards/margins": 0.732616662979126, |
|
"rewards/rejected": -0.6762354373931885, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.00595369724598976, |
|
"grad_norm": 14.012537956237793, |
|
"learning_rate": 4.1069690242163484e-05, |
|
"logits/chosen": -3.069938898086548, |
|
"logits/rejected": -3.0810136795043945, |
|
"logps/chosen": -468.80712890625, |
|
"logps/rejected": -306.1033020019531, |
|
"loss": 0.2554, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.6414734125137329, |
|
"rewards/margins": 1.8135206699371338, |
|
"rewards/rejected": -1.1720472574234009, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.0059877183731097015, |
|
"grad_norm": 34.83932876586914, |
|
"learning_rate": 4.093559974371725e-05, |
|
"logits/chosen": -2.9957642555236816, |
|
"logits/rejected": -3.0063579082489014, |
|
"logps/chosen": -460.42413330078125, |
|
"logps/rejected": -250.7933349609375, |
|
"loss": 0.5806, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.6806167364120483, |
|
"rewards/margins": 0.625430703163147, |
|
"rewards/rejected": 0.055186063051223755, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.006021739500229642, |
|
"grad_norm": 11.975921630859375, |
|
"learning_rate": 4.080073256662127e-05, |
|
"logits/chosen": -3.0853240489959717, |
|
"logits/rejected": -3.062232494354248, |
|
"logps/chosen": -191.1412811279297, |
|
"logps/rejected": -127.30520629882812, |
|
"loss": 0.3486, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.016726315021514893, |
|
"rewards/margins": 1.4842876195907593, |
|
"rewards/rejected": -1.5010138750076294, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.006055760627349584, |
|
"grad_norm": 18.293811798095703, |
|
"learning_rate": 4.066509528411152e-05, |
|
"logits/chosen": -3.059425115585327, |
|
"logits/rejected": -3.0602657794952393, |
|
"logps/chosen": -292.20257568359375, |
|
"logps/rejected": -128.2594451904297, |
|
"loss": 0.5182, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.13988780975341797, |
|
"rewards/margins": 0.9404779672622681, |
|
"rewards/rejected": -0.8005901575088501, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.006089781754469526, |
|
"grad_norm": 14.906744003295898, |
|
"learning_rate": 4.052869450695776e-05, |
|
"logits/chosen": -3.025784492492676, |
|
"logits/rejected": -3.0592339038848877, |
|
"logps/chosen": -377.5328674316406, |
|
"logps/rejected": -332.77716064453125, |
|
"loss": 0.2417, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.292724132537842, |
|
"rewards/margins": 1.75880765914917, |
|
"rewards/rejected": 0.5339164733886719, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.0061238028815894675, |
|
"grad_norm": 14.212754249572754, |
|
"learning_rate": 4.039153688314145e-05, |
|
"logits/chosen": -2.969928026199341, |
|
"logits/rejected": -2.9694008827209473, |
|
"logps/chosen": -280.77630615234375, |
|
"logps/rejected": -165.06353759765625, |
|
"loss": 0.5345, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.9551172256469727, |
|
"rewards/margins": 1.6668689250946045, |
|
"rewards/rejected": -0.7117516994476318, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.006157824008709408, |
|
"grad_norm": 35.84933090209961, |
|
"learning_rate": 4.02536290975317e-05, |
|
"logits/chosen": -3.0387415885925293, |
|
"logits/rejected": -3.0296339988708496, |
|
"logps/chosen": -308.9189147949219, |
|
"logps/rejected": -265.5447692871094, |
|
"loss": 1.3423, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.3892940282821655, |
|
"rewards/margins": -0.6154876947402954, |
|
"rewards/rejected": 1.004781723022461, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.00619184513582935, |
|
"grad_norm": 28.74908447265625, |
|
"learning_rate": 4.011497787155938e-05, |
|
"logits/chosen": -3.0556583404541016, |
|
"logits/rejected": -3.050330877304077, |
|
"logps/chosen": -277.91668701171875, |
|
"logps/rejected": -350.3586120605469, |
|
"loss": 1.1391, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.2823715209960938, |
|
"rewards/margins": -0.13678458333015442, |
|
"rewards/rejected": 1.4191560745239258, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.006225866262949292, |
|
"grad_norm": 11.06321907043457, |
|
"learning_rate": 3.997558996288965e-05, |
|
"logits/chosen": -3.007373332977295, |
|
"logits/rejected": -3.0230917930603027, |
|
"logps/chosen": -240.30911254882812, |
|
"logps/rejected": -276.2512512207031, |
|
"loss": 0.2578, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.5546003580093384, |
|
"rewards/margins": 1.3124656677246094, |
|
"rewards/rejected": -0.757865309715271, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.0062598873900692326, |
|
"grad_norm": 25.019960403442383, |
|
"learning_rate": 3.983547216509254e-05, |
|
"logits/chosen": -3.000227212905884, |
|
"logits/rejected": -3.008676528930664, |
|
"logps/chosen": -375.60894775390625, |
|
"logps/rejected": -163.33262634277344, |
|
"loss": 0.6255, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.9179801940917969, |
|
"rewards/margins": 1.235597848892212, |
|
"rewards/rejected": -0.31761759519577026, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.006293908517189174, |
|
"grad_norm": 47.152008056640625, |
|
"learning_rate": 3.969463130731183e-05, |
|
"logits/chosen": -3.0159125328063965, |
|
"logits/rejected": -3.0475568771362305, |
|
"logps/chosen": -247.3843231201172, |
|
"logps/rejected": -397.0201416015625, |
|
"loss": 2.6598, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -0.1503952443599701, |
|
"rewards/margins": -2.3472542762756348, |
|
"rewards/rejected": 2.1968588829040527, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.006327929644309116, |
|
"grad_norm": 24.639726638793945, |
|
"learning_rate": 3.955307425393224e-05, |
|
"logits/chosen": -3.0540177822113037, |
|
"logits/rejected": -3.0610384941101074, |
|
"logps/chosen": -155.3057861328125, |
|
"logps/rejected": -140.0192413330078, |
|
"loss": 1.0936, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 0.19507810473442078, |
|
"rewards/margins": -0.6317047476768494, |
|
"rewards/rejected": 0.8267828822135925, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.006361950771429058, |
|
"grad_norm": 24.448938369750977, |
|
"learning_rate": 3.941080790424484e-05, |
|
"logits/chosen": -3.0413625240325928, |
|
"logits/rejected": -3.0307819843292236, |
|
"logps/chosen": -119.50736999511719, |
|
"logps/rejected": -150.93804931640625, |
|
"loss": 1.5665, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 0.8793508410453796, |
|
"rewards/margins": -1.1721866130828857, |
|
"rewards/rejected": 2.05153751373291, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.0063959718985489985, |
|
"grad_norm": 23.76529884338379, |
|
"learning_rate": 3.92678391921108e-05, |
|
"logits/chosen": -3.020270347595215, |
|
"logits/rejected": -3.0138823986053467, |
|
"logps/chosen": -300.4285888671875, |
|
"logps/rejected": -316.291015625, |
|
"loss": 0.7366, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.6291970014572144, |
|
"rewards/margins": 1.3483175039291382, |
|
"rewards/rejected": 0.28087958693504333, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.00642999302566894, |
|
"grad_norm": 13.355870246887207, |
|
"learning_rate": 3.912417508562345e-05, |
|
"logits/chosen": -3.037825107574463, |
|
"logits/rejected": -3.0023696422576904, |
|
"logps/chosen": -383.6678771972656, |
|
"logps/rejected": -146.22149658203125, |
|
"loss": 0.3185, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 1.9375550746917725, |
|
"rewards/margins": 1.46681547164917, |
|
"rewards/rejected": 0.47073960304260254, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.006464014152788882, |
|
"grad_norm": 11.14699649810791, |
|
"learning_rate": 3.897982258676867e-05, |
|
"logits/chosen": -3.0212783813476562, |
|
"logits/rejected": -3.0273423194885254, |
|
"logps/chosen": -347.14788818359375, |
|
"logps/rejected": -208.30084228515625, |
|
"loss": 0.2532, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 1.1937689781188965, |
|
"rewards/margins": 1.3640395402908325, |
|
"rewards/rejected": -0.17027053236961365, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.006498035279908824, |
|
"grad_norm": 6.4625630378723145, |
|
"learning_rate": 3.883478873108361e-05, |
|
"logits/chosen": -3.0402722358703613, |
|
"logits/rejected": -3.0695693492889404, |
|
"logps/chosen": -654.4649658203125, |
|
"logps/rejected": -184.7763214111328, |
|
"loss": 0.1096, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 3.3329529762268066, |
|
"rewards/margins": 2.8273959159851074, |
|
"rewards/rejected": 0.5055570602416992, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.0065320564070287645, |
|
"grad_norm": 4.210840225219727, |
|
"learning_rate": 3.868908058731376e-05, |
|
"logits/chosen": -2.9779717922210693, |
|
"logits/rejected": -3.0510642528533936, |
|
"logps/chosen": -692.6348876953125, |
|
"logps/rejected": -187.54519653320312, |
|
"loss": 0.085, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 4.205834865570068, |
|
"rewards/margins": 3.758317708969116, |
|
"rewards/rejected": 0.4475172162055969, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.006566077534148706, |
|
"grad_norm": 20.636354446411133, |
|
"learning_rate": 3.85427052570685e-05, |
|
"logits/chosen": -3.0168797969818115, |
|
"logits/rejected": -3.081496000289917, |
|
"logps/chosen": -947.9866333007812, |
|
"logps/rejected": -168.07162475585938, |
|
"loss": 0.6248, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 3.4627633094787598, |
|
"rewards/margins": 2.3802175521850586, |
|
"rewards/rejected": 1.0825456380844116, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.006600098661268648, |
|
"grad_norm": 14.299696922302246, |
|
"learning_rate": 3.8395669874474915e-05, |
|
"logits/chosen": -3.0828664302825928, |
|
"logits/rejected": -3.0772957801818848, |
|
"logps/chosen": -207.02049255371094, |
|
"logps/rejected": -338.0000915527344, |
|
"loss": 0.3508, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.4636499881744385, |
|
"rewards/margins": 1.1103200912475586, |
|
"rewards/rejected": 1.3533298969268799, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.00663411978838859, |
|
"grad_norm": 46.808406829833984, |
|
"learning_rate": 3.824798160583012e-05, |
|
"logits/chosen": -2.941770076751709, |
|
"logits/rejected": -2.964123487472534, |
|
"logps/chosen": -507.2760314941406, |
|
"logps/rejected": -661.114013671875, |
|
"loss": 2.4408, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.7217434644699097, |
|
"rewards/margins": -1.7116930484771729, |
|
"rewards/rejected": 3.433436632156372, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.0066681409155085304, |
|
"grad_norm": 35.63595962524414, |
|
"learning_rate": 3.8099647649251986e-05, |
|
"logits/chosen": -3.0318660736083984, |
|
"logits/rejected": -3.0759072303771973, |
|
"logps/chosen": -129.00115966796875, |
|
"logps/rejected": -273.1650390625, |
|
"loss": 1.6014, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 1.0606502294540405, |
|
"rewards/margins": -1.1689484119415283, |
|
"rewards/rejected": 2.2295985221862793, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.006702162042628472, |
|
"grad_norm": 37.316131591796875, |
|
"learning_rate": 3.795067523432826e-05, |
|
"logits/chosen": -3.072580575942993, |
|
"logits/rejected": -3.0176706314086914, |
|
"logps/chosen": -524.81640625, |
|
"logps/rejected": -310.4704284667969, |
|
"loss": 1.3092, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.23217511177062988, |
|
"rewards/margins": 0.07222485542297363, |
|
"rewards/rejected": 0.15995030105113983, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.006736183169748414, |
|
"grad_norm": 31.383419036865234, |
|
"learning_rate": 3.780107162176429e-05, |
|
"logits/chosen": -3.037428140640259, |
|
"logits/rejected": -3.008864641189575, |
|
"logps/chosen": -257.443603515625, |
|
"logps/rejected": -191.2501983642578, |
|
"loss": 0.928, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.2667136788368225, |
|
"rewards/margins": 0.2050628662109375, |
|
"rewards/rejected": 0.061650827527046204, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.006770204296868356, |
|
"grad_norm": 22.12115478515625, |
|
"learning_rate": 3.765084410302909e-05, |
|
"logits/chosen": -3.0683722496032715, |
|
"logits/rejected": -3.1092453002929688, |
|
"logps/chosen": -204.34571838378906, |
|
"logps/rejected": -125.01068115234375, |
|
"loss": 0.7936, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.25386929512023926, |
|
"rewards/margins": 0.3415624797344208, |
|
"rewards/rejected": -0.0876932144165039, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.006804225423988296, |
|
"grad_norm": 24.464948654174805, |
|
"learning_rate": 3.7500000000000003e-05, |
|
"logits/chosen": -3.024453639984131, |
|
"logits/rejected": -3.0440425872802734, |
|
"logps/chosen": -357.72430419921875, |
|
"logps/rejected": -370.7812805175781, |
|
"loss": 0.6607, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.4781608581542969, |
|
"rewards/margins": 1.404123306274414, |
|
"rewards/rejected": 0.07403749227523804, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.006838246551108238, |
|
"grad_norm": 10.937922477722168, |
|
"learning_rate": 3.7348546664605777e-05, |
|
"logits/chosen": -3.023815393447876, |
|
"logits/rejected": -3.0414845943450928, |
|
"logps/chosen": -425.942138671875, |
|
"logps/rejected": -245.46630859375, |
|
"loss": 0.2367, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.1465694904327393, |
|
"rewards/margins": 1.9644802808761597, |
|
"rewards/rejected": 0.18208922445774078, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.00687226767822818, |
|
"grad_norm": 21.211931228637695, |
|
"learning_rate": 3.719649147846832e-05, |
|
"logits/chosen": -3.0427141189575195, |
|
"logits/rejected": -2.9640493392944336, |
|
"logps/chosen": -388.6680908203125, |
|
"logps/rejected": -215.27015686035156, |
|
"loss": 0.3829, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.5961097478866577, |
|
"rewards/margins": 1.3369483947753906, |
|
"rewards/rejected": 0.25916144251823425, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.0069062888053481215, |
|
"grad_norm": 31.90744400024414, |
|
"learning_rate": 3.704384185254288e-05, |
|
"logits/chosen": -3.09157657623291, |
|
"logits/rejected": -3.071305990219116, |
|
"logps/chosen": -423.3403625488281, |
|
"logps/rejected": -325.9812927246094, |
|
"loss": 0.8221, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.09947669506073, |
|
"rewards/margins": -0.017909586429595947, |
|
"rewards/rejected": 1.1173863410949707, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.006940309932468062, |
|
"grad_norm": 34.60957717895508, |
|
"learning_rate": 3.689060522675689e-05, |
|
"logits/chosen": -3.041463851928711, |
|
"logits/rejected": -3.016758441925049, |
|
"logps/chosen": -390.4910583496094, |
|
"logps/rejected": -255.284912109375, |
|
"loss": 0.8552, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.645746648311615, |
|
"rewards/margins": 0.39533135294914246, |
|
"rewards/rejected": 0.25041520595550537, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.006974331059588004, |
|
"grad_norm": 29.063993453979492, |
|
"learning_rate": 3.673678906964727e-05, |
|
"logits/chosen": -3.116567850112915, |
|
"logits/rejected": -3.10011625289917, |
|
"logps/chosen": -330.48529052734375, |
|
"logps/rejected": -386.4382019042969, |
|
"loss": 0.391, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.1223270893096924, |
|
"rewards/margins": 1.8259775638580322, |
|
"rewards/rejected": 0.2963497042655945, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.007008352186707946, |
|
"grad_norm": 17.059940338134766, |
|
"learning_rate": 3.6582400877996546e-05, |
|
"logits/chosen": -3.108125686645508, |
|
"logits/rejected": -3.078683376312256, |
|
"logps/chosen": -332.83935546875, |
|
"logps/rejected": -100.14358520507812, |
|
"loss": 0.3988, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.511772871017456, |
|
"rewards/margins": 0.9818691611289978, |
|
"rewards/rejected": 0.5299036502838135, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.0070423733138278875, |
|
"grad_norm": 48.1053581237793, |
|
"learning_rate": 3.642744817646736e-05, |
|
"logits/chosen": -3.0949149131774902, |
|
"logits/rejected": -3.1091861724853516, |
|
"logps/chosen": -321.2143859863281, |
|
"logps/rejected": -155.83506774902344, |
|
"loss": 1.1682, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 0.3836750388145447, |
|
"rewards/margins": -0.15334665775299072, |
|
"rewards/rejected": 0.5370216369628906, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.007076394440947828, |
|
"grad_norm": 7.495742321014404, |
|
"learning_rate": 3.627193851723577e-05, |
|
"logits/chosen": -3.0586297512054443, |
|
"logits/rejected": -3.048053503036499, |
|
"logps/chosen": -173.07139587402344, |
|
"logps/rejected": -191.9176025390625, |
|
"loss": 0.2395, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.4388919770717621, |
|
"rewards/margins": 1.3739150762557983, |
|
"rewards/rejected": -0.9350231289863586, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.00711041556806777, |
|
"grad_norm": 25.139368057250977, |
|
"learning_rate": 3.611587947962319e-05, |
|
"logits/chosen": -3.0632824897766113, |
|
"logits/rejected": -3.0348927974700928, |
|
"logps/chosen": -305.40679931640625, |
|
"logps/rejected": -198.31448364257812, |
|
"loss": 0.8952, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.3874067068099976, |
|
"rewards/margins": 0.3460991084575653, |
|
"rewards/rejected": 1.0413076877593994, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.007144436695187712, |
|
"grad_norm": 29.285905838012695, |
|
"learning_rate": 3.5959278669726935e-05, |
|
"logits/chosen": -3.069756507873535, |
|
"logits/rejected": -3.009371757507324, |
|
"logps/chosen": -436.599853515625, |
|
"logps/rejected": -228.88070678710938, |
|
"loss": 0.8311, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.9359901547431946, |
|
"rewards/margins": 0.4009828567504883, |
|
"rewards/rejected": 0.5350073575973511, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.0071784578223076535, |
|
"grad_norm": 28.730527877807617, |
|
"learning_rate": 3.580214372004956e-05, |
|
"logits/chosen": -3.039961338043213, |
|
"logits/rejected": -2.9948177337646484, |
|
"logps/chosen": -321.87506103515625, |
|
"logps/rejected": -272.5018310546875, |
|
"loss": 0.6225, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.5199428796768188, |
|
"rewards/margins": 0.7615558505058289, |
|
"rewards/rejected": 0.7583869695663452, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.007212478949427594, |
|
"grad_norm": 23.884435653686523, |
|
"learning_rate": 3.564448228912682e-05, |
|
"logits/chosen": -3.1587538719177246, |
|
"logits/rejected": -3.1054487228393555, |
|
"logps/chosen": -347.6966552734375, |
|
"logps/rejected": -121.53209686279297, |
|
"loss": 0.9833, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.9999253749847412, |
|
"rewards/margins": -0.003412783145904541, |
|
"rewards/rejected": 2.003338098526001, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.007246500076547536, |
|
"grad_norm": 34.988765716552734, |
|
"learning_rate": 3.548630206115443e-05, |
|
"logits/chosen": -3.123871326446533, |
|
"logits/rejected": -3.1092422008514404, |
|
"logps/chosen": -288.57171630859375, |
|
"logps/rejected": -172.01113891601562, |
|
"loss": 0.9646, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 0.4253236949443817, |
|
"rewards/margins": -0.3028186857700348, |
|
"rewards/rejected": 0.7281423807144165, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.007280521203667478, |
|
"grad_norm": 10.853194236755371, |
|
"learning_rate": 3.532761074561355e-05, |
|
"logits/chosen": -3.0363683700561523, |
|
"logits/rejected": -3.0223159790039062, |
|
"logps/chosen": -289.89605712890625, |
|
"logps/rejected": -208.051025390625, |
|
"loss": 0.276, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 1.5353323221206665, |
|
"rewards/margins": 1.4551451206207275, |
|
"rewards/rejected": 0.08018721640110016, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.0073145423307874186, |
|
"grad_norm": 22.112659454345703, |
|
"learning_rate": 3.516841607689501e-05, |
|
"logits/chosen": -3.0768256187438965, |
|
"logits/rejected": -3.0702104568481445, |
|
"logps/chosen": -288.7914123535156, |
|
"logps/rejected": -123.36109161376953, |
|
"loss": 0.5534, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.0879135131835938, |
|
"rewards/margins": 0.5657981634140015, |
|
"rewards/rejected": 0.5221153497695923, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.00734856345790736, |
|
"grad_norm": 21.93831443786621, |
|
"learning_rate": 3.5008725813922386e-05, |
|
"logits/chosen": -3.053438901901245, |
|
"logits/rejected": -2.93066668510437, |
|
"logps/chosen": -365.1923522949219, |
|
"logps/rejected": -230.60353088378906, |
|
"loss": 0.6113, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.4077308177948, |
|
"rewards/margins": 0.4580451548099518, |
|
"rewards/rejected": 1.9496855735778809, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.007382584585027302, |
|
"grad_norm": 20.58770751953125, |
|
"learning_rate": 3.484854773977378e-05, |
|
"logits/chosen": -3.0227174758911133, |
|
"logits/rejected": -3.012328624725342, |
|
"logps/chosen": -255.4609375, |
|
"logps/rejected": -173.58782958984375, |
|
"loss": 0.5454, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.5801427364349365, |
|
"rewards/margins": 1.1155328750610352, |
|
"rewards/rejected": 0.46460971236228943, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.007416605712147244, |
|
"grad_norm": 28.43218994140625, |
|
"learning_rate": 3.4687889661302576e-05, |
|
"logits/chosen": -3.0358567237854004, |
|
"logits/rejected": -3.1040730476379395, |
|
"logps/chosen": -195.6716766357422, |
|
"logps/rejected": -222.753173828125, |
|
"loss": 0.8159, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.174198627471924, |
|
"rewards/margins": 0.9547780752182007, |
|
"rewards/rejected": 1.2194204330444336, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.0074506268392671845, |
|
"grad_norm": 11.295846939086914, |
|
"learning_rate": 3.452675940875686e-05, |
|
"logits/chosen": -3.0799617767333984, |
|
"logits/rejected": -3.064857244491577, |
|
"logps/chosen": -228.47509765625, |
|
"logps/rejected": -122.39473724365234, |
|
"loss": 0.2802, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.210326910018921, |
|
"rewards/margins": 1.343271255493164, |
|
"rewards/rejected": 0.8670557141304016, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.007484647966387126, |
|
"grad_norm": 29.458600997924805, |
|
"learning_rate": 3.436516483539781e-05, |
|
"logits/chosen": -3.106647491455078, |
|
"logits/rejected": -3.057008743286133, |
|
"logps/chosen": -370.58282470703125, |
|
"logps/rejected": -309.05596923828125, |
|
"loss": 0.8483, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.4642128944396973, |
|
"rewards/margins": 0.5211576819419861, |
|
"rewards/rejected": 1.9430551528930664, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.007518669093507068, |
|
"grad_norm": 9.961498260498047, |
|
"learning_rate": 3.4203113817116957e-05, |
|
"logits/chosen": -2.9950881004333496, |
|
"logits/rejected": -3.043566942214966, |
|
"logps/chosen": -145.6614990234375, |
|
"logps/rejected": -359.437255859375, |
|
"loss": 0.1927, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.1120011806488037, |
|
"rewards/margins": 2.6634767055511475, |
|
"rewards/rejected": -0.5514757633209229, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.00755269022062701, |
|
"grad_norm": 18.790021896362305, |
|
"learning_rate": 3.4040614252052305e-05, |
|
"logits/chosen": -3.0158464908599854, |
|
"logits/rejected": -3.033033847808838, |
|
"logps/chosen": -357.15777587890625, |
|
"logps/rejected": -226.76284790039062, |
|
"loss": 0.42, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 3.0548276901245117, |
|
"rewards/margins": 1.0073288679122925, |
|
"rewards/rejected": 2.0474987030029297, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.0075867113477469505, |
|
"grad_norm": 17.877826690673828, |
|
"learning_rate": 3.387767406020343e-05, |
|
"logits/chosen": -3.0780677795410156, |
|
"logits/rejected": -3.0696065425872803, |
|
"logps/chosen": -458.90887451171875, |
|
"logps/rejected": -151.19442749023438, |
|
"loss": 0.5504, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.527190685272217, |
|
"rewards/margins": 0.6874034404754639, |
|
"rewards/rejected": 1.839787244796753, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.007620732474866892, |
|
"grad_norm": 28.49088478088379, |
|
"learning_rate": 3.3714301183045385e-05, |
|
"logits/chosen": -3.0982367992401123, |
|
"logits/rejected": -3.123861312866211, |
|
"logps/chosen": -401.4808349609375, |
|
"logps/rejected": -395.56134033203125, |
|
"loss": 0.8387, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.667259693145752, |
|
"rewards/margins": 0.14681124687194824, |
|
"rewards/rejected": 2.5204484462738037, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.007654753601986834, |
|
"grad_norm": 25.027799606323242, |
|
"learning_rate": 3.355050358314172e-05, |
|
"logits/chosen": -3.101318359375, |
|
"logits/rejected": -3.1230921745300293, |
|
"logps/chosen": -150.42347717285156, |
|
"logps/rejected": -217.7826690673828, |
|
"loss": 0.8813, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.7596235275268555, |
|
"rewards/margins": 0.40374666452407837, |
|
"rewards/rejected": 1.3558769226074219, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.007688774729106776, |
|
"grad_norm": 30.265897750854492, |
|
"learning_rate": 3.338628924375638e-05, |
|
"logits/chosen": -3.042361259460449, |
|
"logits/rejected": -2.9984130859375, |
|
"logps/chosen": -274.414306640625, |
|
"logps/rejected": -151.527099609375, |
|
"loss": 1.0779, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.9679371118545532, |
|
"rewards/margins": 0.3534848690032959, |
|
"rewards/rejected": 1.6144523620605469, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.0077227958562267164, |
|
"grad_norm": 28.559492111206055, |
|
"learning_rate": 3.322166616846458e-05, |
|
"logits/chosen": -3.0308961868286133, |
|
"logits/rejected": -3.048680543899536, |
|
"logps/chosen": -469.332275390625, |
|
"logps/rejected": -246.54237365722656, |
|
"loss": 0.8713, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 3.8970139026641846, |
|
"rewards/margins": 0.9500312209129333, |
|
"rewards/rejected": 2.9469828605651855, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.007756816983346658, |
|
"grad_norm": 30.437313079833984, |
|
"learning_rate": 3.305664238076278e-05, |
|
"logits/chosen": -3.0552141666412354, |
|
"logits/rejected": -3.0249457359313965, |
|
"logps/chosen": -190.10910034179688, |
|
"logps/rejected": -176.27610778808594, |
|
"loss": 1.7815, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.2772684693336487, |
|
"rewards/margins": -0.6817531585693359, |
|
"rewards/rejected": 0.9590215682983398, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.0077908381104666, |
|
"grad_norm": 32.68939971923828, |
|
"learning_rate": 3.289122592367757e-05, |
|
"logits/chosen": -3.1223931312561035, |
|
"logits/rejected": -3.024728536605835, |
|
"logps/chosen": -200.835693359375, |
|
"logps/rejected": -254.39468383789062, |
|
"loss": 0.9178, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 1.862541675567627, |
|
"rewards/margins": -0.2349603772163391, |
|
"rewards/rejected": 2.0975019931793213, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.00782485923758654, |
|
"grad_norm": 12.67287826538086, |
|
"learning_rate": 3.272542485937369e-05, |
|
"logits/chosen": -3.0310213565826416, |
|
"logits/rejected": -3.0267879962921143, |
|
"logps/chosen": -682.5564575195312, |
|
"logps/rejected": -286.54583740234375, |
|
"loss": 0.2568, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 4.804429531097412, |
|
"rewards/margins": 2.462975025177002, |
|
"rewards/rejected": 2.34145450592041, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.007858880364706482, |
|
"grad_norm": 43.85115432739258, |
|
"learning_rate": 3.2559247268761115e-05, |
|
"logits/chosen": -3.037172317504883, |
|
"logits/rejected": -3.0447936058044434, |
|
"logps/chosen": -138.02731323242188, |
|
"logps/rejected": -285.2200012207031, |
|
"loss": 1.4127, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 1.5211700201034546, |
|
"rewards/margins": -0.8457721471786499, |
|
"rewards/rejected": 2.3669421672821045, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.007892901491826424, |
|
"grad_norm": 13.436664581298828, |
|
"learning_rate": 3.239270125110117e-05, |
|
"logits/chosen": -3.179615020751953, |
|
"logits/rejected": -3.1418628692626953, |
|
"logps/chosen": -267.392578125, |
|
"logps/rejected": -207.12918090820312, |
|
"loss": 0.2831, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 1.59194815158844, |
|
"rewards/margins": 1.297247290611267, |
|
"rewards/rejected": 0.29470083117485046, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.007926922618946366, |
|
"grad_norm": 30.355941772460938, |
|
"learning_rate": 3.222579492361179e-05, |
|
"logits/chosen": -3.03082275390625, |
|
"logits/rejected": -3.0827555656433105, |
|
"logps/chosen": -355.8211364746094, |
|
"logps/rejected": -223.76837158203125, |
|
"loss": 0.8163, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 2.9960384368896484, |
|
"rewards/margins": 0.4547451436519623, |
|
"rewards/rejected": 2.5412933826446533, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.007960943746066308, |
|
"grad_norm": 39.661808013916016, |
|
"learning_rate": 3.205853642107192e-05, |
|
"logits/chosen": -2.9772939682006836, |
|
"logits/rejected": -3.041161060333252, |
|
"logps/chosen": -108.25926208496094, |
|
"logps/rejected": -340.5980224609375, |
|
"loss": 1.8546, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 2.25307559967041, |
|
"rewards/margins": -1.0127062797546387, |
|
"rewards/rejected": 3.265781879425049, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.00799496487318625, |
|
"grad_norm": 18.539447784423828, |
|
"learning_rate": 3.1890933895424976e-05, |
|
"logits/chosen": -2.995600700378418, |
|
"logits/rejected": -2.936952829360962, |
|
"logps/chosen": -416.49627685546875, |
|
"logps/rejected": -259.64202880859375, |
|
"loss": 0.3332, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 3.67849063873291, |
|
"rewards/margins": 0.9409371018409729, |
|
"rewards/rejected": 2.737553358078003, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.008028986000306191, |
|
"grad_norm": 8.98495101928711, |
|
"learning_rate": 3.172299551538164e-05, |
|
"logits/chosen": -3.117161512374878, |
|
"logits/rejected": -3.05356502532959, |
|
"logps/chosen": -482.1029357910156, |
|
"logps/rejected": -127.43675994873047, |
|
"loss": 0.1397, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 4.75900936126709, |
|
"rewards/margins": 3.078279972076416, |
|
"rewards/rejected": 1.680729627609253, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.008063007127426131, |
|
"grad_norm": 38.39946365356445, |
|
"learning_rate": 3.155472946602162e-05, |
|
"logits/chosen": -3.073312759399414, |
|
"logits/rejected": -3.0811350345611572, |
|
"logps/chosen": -349.9805908203125, |
|
"logps/rejected": -227.2250213623047, |
|
"loss": 0.6411, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 3.138963222503662, |
|
"rewards/margins": 1.1131224632263184, |
|
"rewards/rejected": 2.0258407592773438, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.008097028254546073, |
|
"grad_norm": 34.01177978515625, |
|
"learning_rate": 3.138614394839476e-05, |
|
"logits/chosen": -3.036881446838379, |
|
"logits/rejected": -3.066068172454834, |
|
"logps/chosen": -360.13787841796875, |
|
"logps/rejected": -342.33453369140625, |
|
"loss": 0.6854, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 3.1536881923675537, |
|
"rewards/margins": 0.4615260064601898, |
|
"rewards/rejected": 2.692162036895752, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.008131049381666014, |
|
"grad_norm": 15.110796928405762, |
|
"learning_rate": 3.121724717912138e-05, |
|
"logits/chosen": -3.133331775665283, |
|
"logits/rejected": -3.1206603050231934, |
|
"logps/chosen": -253.7147216796875, |
|
"logps/rejected": -174.19720458984375, |
|
"loss": 0.3162, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.769083261489868, |
|
"rewards/margins": 1.2836055755615234, |
|
"rewards/rejected": 1.4854779243469238, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.008165070508785956, |
|
"grad_norm": 18.21829605102539, |
|
"learning_rate": 3.104804738999169e-05, |
|
"logits/chosen": -3.0911126136779785, |
|
"logits/rejected": -3.1077239513397217, |
|
"logps/chosen": -307.99774169921875, |
|
"logps/rejected": -313.49267578125, |
|
"loss": 0.3279, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.8375076055526733, |
|
"rewards/margins": 1.11543607711792, |
|
"rewards/rejected": -0.27792832255363464, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.008199091635905898, |
|
"grad_norm": 28.513282775878906, |
|
"learning_rate": 3.087855282756475e-05, |
|
"logits/chosen": -3.083726406097412, |
|
"logits/rejected": -3.0311951637268066, |
|
"logps/chosen": -597.4566040039062, |
|
"logps/rejected": -413.25457763671875, |
|
"loss": 0.7872, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 4.096184253692627, |
|
"rewards/margins": 1.410437822341919, |
|
"rewards/rejected": 2.685746192932129, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.00823311276302584, |
|
"grad_norm": 20.769445419311523, |
|
"learning_rate": 3.0708771752766394e-05, |
|
"logits/chosen": -3.0382697582244873, |
|
"logits/rejected": -3.0615615844726562, |
|
"logps/chosen": -306.8995056152344, |
|
"logps/rejected": -207.62193298339844, |
|
"loss": 0.6248, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.5817922353744507, |
|
"rewards/margins": 0.49707847833633423, |
|
"rewards/rejected": -1.0788707733154297, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.008267133890145781, |
|
"grad_norm": 17.814634323120117, |
|
"learning_rate": 3.053871244048669e-05, |
|
"logits/chosen": -3.0551891326904297, |
|
"logits/rejected": -3.072455883026123, |
|
"logps/chosen": -145.9087371826172, |
|
"logps/rejected": -194.36148071289062, |
|
"loss": 0.2275, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.6005315780639648, |
|
"rewards/margins": 1.8227686882019043, |
|
"rewards/rejected": -1.2222371101379395, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.008301155017265723, |
|
"grad_norm": 29.448062896728516, |
|
"learning_rate": 3.0368383179176585e-05, |
|
"logits/chosen": -3.074300765991211, |
|
"logits/rejected": -3.0786709785461426, |
|
"logps/chosen": -283.5516357421875, |
|
"logps/rejected": -280.1275634765625, |
|
"loss": 0.7537, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.5777359008789062, |
|
"rewards/margins": 0.6331707239151001, |
|
"rewards/rejected": 0.9445652961730957, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.008335176144385663, |
|
"grad_norm": 35.73444366455078, |
|
"learning_rate": 3.0197792270443982e-05, |
|
"logits/chosen": -2.983248710632324, |
|
"logits/rejected": -3.028512954711914, |
|
"logps/chosen": -197.35414123535156, |
|
"logps/rejected": -337.7020263671875, |
|
"loss": 1.5799, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.4744613766670227, |
|
"rewards/margins": -0.6541084051132202, |
|
"rewards/rejected": 0.1796470284461975, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.008369197271505605, |
|
"grad_norm": 44.81789779663086, |
|
"learning_rate": 3.002694802864912e-05, |
|
"logits/chosen": -3.100214958190918, |
|
"logits/rejected": -3.0922040939331055, |
|
"logps/chosen": -221.59645080566406, |
|
"logps/rejected": -564.4658203125, |
|
"loss": 1.3614, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -1.5563018321990967, |
|
"rewards/margins": -0.38281095027923584, |
|
"rewards/rejected": -1.1734910011291504, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.008403218398625546, |
|
"grad_norm": 14.416508674621582, |
|
"learning_rate": 2.98558587804993e-05, |
|
"logits/chosen": -3.058180093765259, |
|
"logits/rejected": -3.0165953636169434, |
|
"logps/chosen": -292.0586242675781, |
|
"logps/rejected": -181.1424102783203, |
|
"loss": 0.4068, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.9377683401107788, |
|
"rewards/margins": 1.167628526687622, |
|
"rewards/rejected": -2.1053967475891113, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.008437239525745488, |
|
"grad_norm": 39.20392990112305, |
|
"learning_rate": 2.9684532864643122e-05, |
|
"logits/chosen": -3.090451717376709, |
|
"logits/rejected": -3.0921058654785156, |
|
"logps/chosen": -473.1267395019531, |
|
"logps/rejected": -281.45892333984375, |
|
"loss": 2.3962, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -2.7255377769470215, |
|
"rewards/margins": -1.2629203796386719, |
|
"rewards/rejected": -1.4626175165176392, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.00847126065286543, |
|
"grad_norm": 23.516746520996094, |
|
"learning_rate": 2.9512978631264006e-05, |
|
"logits/chosen": -3.067230463027954, |
|
"logits/rejected": -3.0887598991394043, |
|
"logps/chosen": -267.3660583496094, |
|
"logps/rejected": -251.80201721191406, |
|
"loss": 0.5022, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.7226058840751648, |
|
"rewards/margins": 0.4901663064956665, |
|
"rewards/rejected": -1.2127721309661865, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.008505281779985371, |
|
"grad_norm": 44.44258117675781, |
|
"learning_rate": 2.9341204441673266e-05, |
|
"logits/chosen": -3.1237659454345703, |
|
"logits/rejected": -3.1184511184692383, |
|
"logps/chosen": -313.210205078125, |
|
"logps/rejected": -211.3685302734375, |
|
"loss": 1.8753, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -2.308335304260254, |
|
"rewards/margins": -0.37920308113098145, |
|
"rewards/rejected": -1.929132342338562, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.008539302907105313, |
|
"grad_norm": 26.670148849487305, |
|
"learning_rate": 2.916921866790256e-05, |
|
"logits/chosen": -3.162524700164795, |
|
"logits/rejected": -3.1264796257019043, |
|
"logps/chosen": -387.97235107421875, |
|
"logps/rejected": -299.582763671875, |
|
"loss": 0.7968, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -2.017115592956543, |
|
"rewards/margins": 0.06582087278366089, |
|
"rewards/rejected": -2.0829365253448486, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.008573324034225253, |
|
"grad_norm": 28.174758911132812, |
|
"learning_rate": 2.8997029692295874e-05, |
|
"logits/chosen": -3.002204179763794, |
|
"logits/rejected": -3.038792133331299, |
|
"logps/chosen": -344.5345458984375, |
|
"logps/rejected": -280.2668762207031, |
|
"loss": 0.6465, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.32356908917427063, |
|
"rewards/margins": 0.603424072265625, |
|
"rewards/rejected": -0.9269931316375732, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.008607345161345195, |
|
"grad_norm": 32.33949661254883, |
|
"learning_rate": 2.8824645907100954e-05, |
|
"logits/chosen": -3.0804738998413086, |
|
"logits/rejected": -3.124377489089966, |
|
"logps/chosen": -257.32330322265625, |
|
"logps/rejected": -402.580810546875, |
|
"loss": 0.8504, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -1.3171865940093994, |
|
"rewards/margins": -0.11607035994529724, |
|
"rewards/rejected": -1.2011162042617798, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.008641366288465136, |
|
"grad_norm": 21.82145118713379, |
|
"learning_rate": 2.8652075714060295e-05, |
|
"logits/chosen": -2.903639316558838, |
|
"logits/rejected": -3.0638492107391357, |
|
"logps/chosen": -356.0044250488281, |
|
"logps/rejected": -176.8843536376953, |
|
"loss": 1.055, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.4542312622070312, |
|
"rewards/margins": 0.3908478319644928, |
|
"rewards/rejected": -1.8450790643692017, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.008675387415585078, |
|
"grad_norm": 17.95180320739746, |
|
"learning_rate": 2.8479327524001636e-05, |
|
"logits/chosen": -3.127072811126709, |
|
"logits/rejected": -3.074962854385376, |
|
"logps/chosen": -332.04638671875, |
|
"logps/rejected": -160.70286560058594, |
|
"loss": 0.476, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.376025915145874, |
|
"rewards/margins": 1.413560390472412, |
|
"rewards/rejected": -2.7895865440368652, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.00870940854270502, |
|
"grad_norm": 31.030893325805664, |
|
"learning_rate": 2.8306409756428064e-05, |
|
"logits/chosen": -3.0781168937683105, |
|
"logits/rejected": -3.0802314281463623, |
|
"logps/chosen": -229.36236572265625, |
|
"logps/rejected": -180.08914184570312, |
|
"loss": 0.9595, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -2.6754817962646484, |
|
"rewards/margins": -0.2480696439743042, |
|
"rewards/rejected": -2.427412271499634, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.008743429669824962, |
|
"grad_norm": 24.756603240966797, |
|
"learning_rate": 2.8133330839107608e-05, |
|
"logits/chosen": -3.094456911087036, |
|
"logits/rejected": -3.0564780235290527, |
|
"logps/chosen": -326.5870056152344, |
|
"logps/rejected": -205.408935546875, |
|
"loss": 0.6335, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.9964733719825745, |
|
"rewards/margins": 0.971878170967102, |
|
"rewards/rejected": -1.9683516025543213, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.008777450796944903, |
|
"grad_norm": 7.824972629547119, |
|
"learning_rate": 2.7960099207662532e-05, |
|
"logits/chosen": -3.0179147720336914, |
|
"logits/rejected": -3.0747182369232178, |
|
"logps/chosen": -522.4074096679688, |
|
"logps/rejected": -203.90457153320312, |
|
"loss": 0.1417, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.22739183902740479, |
|
"rewards/margins": 3.4399912357330322, |
|
"rewards/rejected": -3.212599277496338, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.008811471924064845, |
|
"grad_norm": 40.6921501159668, |
|
"learning_rate": 2.7786723305158136e-05, |
|
"logits/chosen": -3.0709447860717773, |
|
"logits/rejected": -3.1174564361572266, |
|
"logps/chosen": -369.0797119140625, |
|
"logps/rejected": -284.5461730957031, |
|
"loss": 1.9964, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -1.55587899684906, |
|
"rewards/margins": -1.3374121189117432, |
|
"rewards/rejected": -0.21846693754196167, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.008845493051184785, |
|
"grad_norm": 8.104869842529297, |
|
"learning_rate": 2.761321158169134e-05, |
|
"logits/chosen": -3.0564663410186768, |
|
"logits/rejected": -3.0263187885284424, |
|
"logps/chosen": -358.8256530761719, |
|
"logps/rejected": -238.61795043945312, |
|
"loss": 0.1023, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.533338725566864, |
|
"rewards/margins": 3.087118148803711, |
|
"rewards/rejected": -2.5537796020507812, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.008879514178304727, |
|
"grad_norm": 34.5290641784668, |
|
"learning_rate": 2.7439572493978736e-05, |
|
"logits/chosen": -3.097294569015503, |
|
"logits/rejected": -3.068530559539795, |
|
"logps/chosen": -467.70587158203125, |
|
"logps/rejected": -319.7679443359375, |
|
"loss": 1.2843, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -1.0861297845840454, |
|
"rewards/margins": 0.6818221211433411, |
|
"rewards/rejected": -1.7679519653320312, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.008913535305424668, |
|
"grad_norm": 34.942047119140625, |
|
"learning_rate": 2.726581450494451e-05, |
|
"logits/chosen": -3.118527412414551, |
|
"logits/rejected": -3.1248326301574707, |
|
"logps/chosen": -367.74273681640625, |
|
"logps/rejected": -299.168701171875, |
|
"loss": 1.1626, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.9521721601486206, |
|
"rewards/margins": 0.18353843688964844, |
|
"rewards/rejected": -1.1357104778289795, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.00894755643254461, |
|
"grad_norm": 9.154927253723145, |
|
"learning_rate": 2.7091946083307896e-05, |
|
"logits/chosen": -3.0849175453186035, |
|
"logits/rejected": -3.0990471839904785, |
|
"logps/chosen": -346.5392761230469, |
|
"logps/rejected": -165.90834045410156, |
|
"loss": 0.1701, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.819671630859375, |
|
"rewards/margins": 1.9401745796203613, |
|
"rewards/rejected": -2.7598462104797363, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.008981577559664552, |
|
"grad_norm": 39.93187713623047, |
|
"learning_rate": 2.6917975703170466e-05, |
|
"logits/chosen": -3.0132288932800293, |
|
"logits/rejected": -3.050952911376953, |
|
"logps/chosen": -224.81666564941406, |
|
"logps/rejected": -336.09326171875, |
|
"loss": 1.1261, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.545545220375061, |
|
"rewards/margins": -0.04098162055015564, |
|
"rewards/rejected": -0.5045635104179382, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.009015598686784494, |
|
"grad_norm": 15.97708797454834, |
|
"learning_rate": 2.674391184360313e-05, |
|
"logits/chosen": -3.0694761276245117, |
|
"logits/rejected": -3.052687168121338, |
|
"logps/chosen": -505.47064208984375, |
|
"logps/rejected": -280.6697998046875, |
|
"loss": 0.3027, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.38131940364837646, |
|
"rewards/margins": 1.2190617322921753, |
|
"rewards/rejected": -0.837742269039154, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.009049619813904435, |
|
"grad_norm": 27.057903289794922, |
|
"learning_rate": 2.656976298823284e-05, |
|
"logits/chosen": -3.1242947578430176, |
|
"logits/rejected": -3.1150031089782715, |
|
"logps/chosen": -299.47406005859375, |
|
"logps/rejected": -222.78358459472656, |
|
"loss": 0.7096, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.32237398624420166, |
|
"rewards/margins": 0.2573413550853729, |
|
"rewards/rejected": -0.579715371131897, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.009083640941024377, |
|
"grad_norm": 36.52804946899414, |
|
"learning_rate": 2.6395537624829096e-05, |
|
"logits/chosen": -3.101689338684082, |
|
"logits/rejected": -3.09779691696167, |
|
"logps/chosen": -363.6987609863281, |
|
"logps/rejected": -393.2881164550781, |
|
"loss": 0.8801, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.650049090385437, |
|
"rewards/margins": 0.190473735332489, |
|
"rewards/rejected": -0.8405227661132812, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.009117662068144317, |
|
"grad_norm": 9.045161247253418, |
|
"learning_rate": 2.6221244244890336e-05, |
|
"logits/chosen": -3.114185094833374, |
|
"logits/rejected": -3.1218295097351074, |
|
"logps/chosen": -345.9595947265625, |
|
"logps/rejected": -191.24752807617188, |
|
"loss": 0.2649, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.04784354567527771, |
|
"rewards/margins": 2.0645532608032227, |
|
"rewards/rejected": -2.112396717071533, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.009151683195264259, |
|
"grad_norm": 35.42766571044922, |
|
"learning_rate": 2.604689134322999e-05, |
|
"logits/chosen": -3.088737964630127, |
|
"logits/rejected": -3.043372392654419, |
|
"logps/chosen": -255.34486389160156, |
|
"logps/rejected": -266.3675842285156, |
|
"loss": 1.133, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -1.9069218635559082, |
|
"rewards/margins": -0.6148130893707275, |
|
"rewards/rejected": -1.2921090126037598, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.0091857043223842, |
|
"grad_norm": 28.35894012451172, |
|
"learning_rate": 2.587248741756253e-05, |
|
"logits/chosen": -3.097867488861084, |
|
"logits/rejected": -3.120126724243164, |
|
"logps/chosen": -164.0009002685547, |
|
"logps/rejected": -266.9195556640625, |
|
"loss": 1.0151, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -2.1401047706604004, |
|
"rewards/margins": -0.5045443177223206, |
|
"rewards/rejected": -1.6355602741241455, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.009219725449504142, |
|
"grad_norm": 8.758069038391113, |
|
"learning_rate": 2.5698040968089225e-05, |
|
"logits/chosen": -3.1079535484313965, |
|
"logits/rejected": -3.0330348014831543, |
|
"logps/chosen": -427.437255859375, |
|
"logps/rejected": -140.69039916992188, |
|
"loss": 0.143, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.42043229937553406, |
|
"rewards/margins": 2.4307219982147217, |
|
"rewards/rejected": -2.851154327392578, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.009253746576624084, |
|
"grad_norm": 21.08401107788086, |
|
"learning_rate": 2.5523560497083926e-05, |
|
"logits/chosen": -3.079854726791382, |
|
"logits/rejected": -3.0848076343536377, |
|
"logps/chosen": -249.33953857421875, |
|
"logps/rejected": -199.57180786132812, |
|
"loss": 0.6952, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.5158636569976807, |
|
"rewards/margins": 0.10717989504337311, |
|
"rewards/rejected": -1.6230435371398926, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.009287767703744025, |
|
"grad_norm": 22.102176666259766, |
|
"learning_rate": 2.5349054508478637e-05, |
|
"logits/chosen": -3.10518741607666, |
|
"logits/rejected": -3.0030434131622314, |
|
"logps/chosen": -467.40283203125, |
|
"logps/rejected": -433.1641845703125, |
|
"loss": 0.4981, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.7289947271347046, |
|
"rewards/margins": 1.3333637714385986, |
|
"rewards/rejected": -0.6043689846992493, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.009321788830863967, |
|
"grad_norm": 34.28799057006836, |
|
"learning_rate": 2.517453150744904e-05, |
|
"logits/chosen": -3.0793797969818115, |
|
"logits/rejected": -3.1031153202056885, |
|
"logps/chosen": -188.7920379638672, |
|
"logps/rejected": -470.43646240234375, |
|
"loss": 1.0595, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -1.7734549045562744, |
|
"rewards/margins": -0.33487796783447266, |
|
"rewards/rejected": -1.4385769367218018, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.009355809957983909, |
|
"grad_norm": 28.724674224853516, |
|
"learning_rate": 2.5e-05, |
|
"logits/chosen": -3.1051037311553955, |
|
"logits/rejected": -3.086857318878174, |
|
"logps/chosen": -421.162353515625, |
|
"logps/rejected": -292.6861572265625, |
|
"loss": 0.6997, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.3912460207939148, |
|
"rewards/margins": 1.05302095413208, |
|
"rewards/rejected": -0.6617748737335205, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.009389831085103849, |
|
"grad_norm": 26.460264205932617, |
|
"learning_rate": 2.4825468492550964e-05, |
|
"logits/chosen": -3.1109108924865723, |
|
"logits/rejected": -3.0332915782928467, |
|
"logps/chosen": -329.5637512207031, |
|
"logps/rejected": -315.29931640625, |
|
"loss": 0.7131, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.8493518829345703, |
|
"rewards/margins": 0.2182472199201584, |
|
"rewards/rejected": -1.0675990581512451, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.00942385221222379, |
|
"grad_norm": 29.5801944732666, |
|
"learning_rate": 2.4650945491521372e-05, |
|
"logits/chosen": -3.078686237335205, |
|
"logits/rejected": -3.0991415977478027, |
|
"logps/chosen": -310.9006042480469, |
|
"logps/rejected": -257.3263244628906, |
|
"loss": 0.9834, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.0626027584075928, |
|
"rewards/margins": 0.5319604873657227, |
|
"rewards/rejected": 0.5306423306465149, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.009457873339343732, |
|
"grad_norm": 15.408950805664062, |
|
"learning_rate": 2.447643950291608e-05, |
|
"logits/chosen": -3.024104595184326, |
|
"logits/rejected": -3.0790274143218994, |
|
"logps/chosen": -836.8731689453125, |
|
"logps/rejected": -507.77618408203125, |
|
"loss": 0.2132, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 3.0483808517456055, |
|
"rewards/margins": 2.7617428302764893, |
|
"rewards/rejected": 0.2866382598876953, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.009491894466463674, |
|
"grad_norm": 17.838191986083984, |
|
"learning_rate": 2.4301959031910784e-05, |
|
"logits/chosen": -3.0706381797790527, |
|
"logits/rejected": -3.064626693725586, |
|
"logps/chosen": -304.050537109375, |
|
"logps/rejected": -231.68185424804688, |
|
"loss": 0.53, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.18637660145759583, |
|
"rewards/margins": 1.330854058265686, |
|
"rewards/rejected": -1.5172306299209595, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.009525915593583616, |
|
"grad_norm": 31.0172119140625, |
|
"learning_rate": 2.4127512582437485e-05, |
|
"logits/chosen": -3.084573268890381, |
|
"logits/rejected": -3.0737102031707764, |
|
"logps/chosen": -321.79327392578125, |
|
"logps/rejected": -226.30621337890625, |
|
"loss": 1.3898, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.6131480932235718, |
|
"rewards/margins": 0.363433837890625, |
|
"rewards/rejected": -0.9765819311141968, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.009559936720703557, |
|
"grad_norm": 66.75733184814453, |
|
"learning_rate": 2.3953108656770016e-05, |
|
"logits/chosen": -3.0968070030212402, |
|
"logits/rejected": -3.024610757827759, |
|
"logps/chosen": -643.58984375, |
|
"logps/rejected": -335.47503662109375, |
|
"loss": 0.6745, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.9356346130371094, |
|
"rewards/margins": 1.7457501888275146, |
|
"rewards/rejected": 0.18988442420959473, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.009593957847823499, |
|
"grad_norm": 31.20706558227539, |
|
"learning_rate": 2.377875575510967e-05, |
|
"logits/chosen": -3.0294013023376465, |
|
"logits/rejected": -3.0743305683135986, |
|
"logps/chosen": -257.7310485839844, |
|
"logps/rejected": -310.9118347167969, |
|
"loss": 0.9646, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -0.9907014966011047, |
|
"rewards/margins": -0.2872907519340515, |
|
"rewards/rejected": -0.7034107446670532, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.009627978974943439, |
|
"grad_norm": 33.48945236206055, |
|
"learning_rate": 2.3604462375170906e-05, |
|
"logits/chosen": -3.0838496685028076, |
|
"logits/rejected": -3.0826075077056885, |
|
"logps/chosen": -322.6458435058594, |
|
"logps/rejected": -217.98764038085938, |
|
"loss": 1.1084, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.5904526114463806, |
|
"rewards/margins": 0.3243420124053955, |
|
"rewards/rejected": -0.9147946238517761, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.00966200010206338, |
|
"grad_norm": 20.60328483581543, |
|
"learning_rate": 2.3430237011767167e-05, |
|
"logits/chosen": -3.115434408187866, |
|
"logits/rejected": -2.9417314529418945, |
|
"logps/chosen": -313.62603759765625, |
|
"logps/rejected": -236.8177032470703, |
|
"loss": 0.5128, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.5609833002090454, |
|
"rewards/margins": 1.0160855054855347, |
|
"rewards/rejected": -0.4551021456718445, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.009696021229183322, |
|
"grad_norm": 29.198223114013672, |
|
"learning_rate": 2.3256088156396868e-05, |
|
"logits/chosen": -3.066559314727783, |
|
"logits/rejected": -3.044053077697754, |
|
"logps/chosen": -309.3338317871094, |
|
"logps/rejected": -246.8656005859375, |
|
"loss": 1.3914, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -1.0285968780517578, |
|
"rewards/margins": -0.560611367225647, |
|
"rewards/rejected": -0.4679855406284332, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.009730042356303264, |
|
"grad_norm": 35.85755157470703, |
|
"learning_rate": 2.3082024296829536e-05, |
|
"logits/chosen": -3.118199348449707, |
|
"logits/rejected": -3.0979912281036377, |
|
"logps/chosen": -464.64447021484375, |
|
"logps/rejected": -460.1026916503906, |
|
"loss": 1.1007, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -0.3913114666938782, |
|
"rewards/margins": -0.08727627992630005, |
|
"rewards/rejected": -0.30403515696525574, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.009764063483423206, |
|
"grad_norm": 38.31180191040039, |
|
"learning_rate": 2.2908053916692117e-05, |
|
"logits/chosen": -3.112149715423584, |
|
"logits/rejected": -3.1065833568573, |
|
"logps/chosen": -455.6182556152344, |
|
"logps/rejected": -500.2666015625, |
|
"loss": 1.1182, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.275381088256836, |
|
"rewards/margins": 1.6004549264907837, |
|
"rewards/rejected": -0.32507359981536865, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.009798084610543148, |
|
"grad_norm": 28.276689529418945, |
|
"learning_rate": 2.2734185495055503e-05, |
|
"logits/chosen": -3.0288684368133545, |
|
"logits/rejected": -3.0792016983032227, |
|
"logps/chosen": -253.03741455078125, |
|
"logps/rejected": -485.63287353515625, |
|
"loss": 0.5788, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.6279298067092896, |
|
"rewards/margins": 0.8981667757034302, |
|
"rewards/rejected": 0.7297630906105042, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.00983210573766309, |
|
"grad_norm": 14.531770706176758, |
|
"learning_rate": 2.2560427506021266e-05, |
|
"logits/chosen": -3.099992036819458, |
|
"logits/rejected": -2.9924557209014893, |
|
"logps/chosen": -666.0892944335938, |
|
"logps/rejected": -261.25830078125, |
|
"loss": 0.2154, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.346898078918457, |
|
"rewards/margins": 2.1159393787384033, |
|
"rewards/rejected": 0.23095856606960297, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.009866126864783031, |
|
"grad_norm": 13.025914192199707, |
|
"learning_rate": 2.238678841830867e-05, |
|
"logits/chosen": -3.0328335762023926, |
|
"logits/rejected": -3.024500846862793, |
|
"logps/chosen": -384.6082458496094, |
|
"logps/rejected": -202.70977783203125, |
|
"loss": 0.2644, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.5193145871162415, |
|
"rewards/margins": 1.8208515644073486, |
|
"rewards/rejected": -1.3015371561050415, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.009900147991902971, |
|
"grad_norm": 34.91489028930664, |
|
"learning_rate": 2.2213276694841866e-05, |
|
"logits/chosen": -3.0935463905334473, |
|
"logits/rejected": -3.1344003677368164, |
|
"logps/chosen": -187.15899658203125, |
|
"logps/rejected": -246.30255126953125, |
|
"loss": 2.1358, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -1.405801773071289, |
|
"rewards/margins": -1.8487468957901, |
|
"rewards/rejected": 0.44294512271881104, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.009934169119022913, |
|
"grad_norm": 29.857711791992188, |
|
"learning_rate": 2.2039900792337474e-05, |
|
"logits/chosen": -3.033463954925537, |
|
"logits/rejected": -3.0471982955932617, |
|
"logps/chosen": -601.5758056640625, |
|
"logps/rejected": -313.49847412109375, |
|
"loss": 1.342, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.47221651673316956, |
|
"rewards/margins": 0.46173369884490967, |
|
"rewards/rejected": 0.0104827880859375, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.009968190246142854, |
|
"grad_norm": 21.967784881591797, |
|
"learning_rate": 2.186666916089239e-05, |
|
"logits/chosen": -3.1273081302642822, |
|
"logits/rejected": -3.1298046112060547, |
|
"logps/chosen": -284.71270751953125, |
|
"logps/rejected": -316.1896057128906, |
|
"loss": 0.9021, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.5063320398330688, |
|
"rewards/margins": 1.5186200141906738, |
|
"rewards/rejected": -1.0122878551483154, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.010002211373262796, |
|
"grad_norm": 36.973548889160156, |
|
"learning_rate": 2.1693590243571938e-05, |
|
"logits/chosen": -3.036428928375244, |
|
"logits/rejected": -3.0828795433044434, |
|
"logps/chosen": -224.53245544433594, |
|
"logps/rejected": -232.72964477539062, |
|
"loss": 0.9055, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -0.6889886856079102, |
|
"rewards/margins": -0.2959771752357483, |
|
"rewards/rejected": -0.39301151037216187, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.010036232500382738, |
|
"grad_norm": 39.14424514770508, |
|
"learning_rate": 2.1520672475998373e-05, |
|
"logits/chosen": -3.053661823272705, |
|
"logits/rejected": -3.104318857192993, |
|
"logps/chosen": -317.2615966796875, |
|
"logps/rejected": -443.0313415527344, |
|
"loss": 1.1603, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.10411298274993896, |
|
"rewards/margins": -0.2241775393486023, |
|
"rewards/rejected": 0.32829055190086365, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.01007025362750268, |
|
"grad_norm": 6.142483234405518, |
|
"learning_rate": 2.1347924285939714e-05, |
|
"logits/chosen": -3.0842039585113525, |
|
"logits/rejected": -3.0498149394989014, |
|
"logps/chosen": -656.255859375, |
|
"logps/rejected": -208.490478515625, |
|
"loss": 0.1191, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.4482297897338867, |
|
"rewards/margins": 3.8186123371124268, |
|
"rewards/rejected": -1.3703827857971191, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.010104274754622621, |
|
"grad_norm": 17.855514526367188, |
|
"learning_rate": 2.117535409289905e-05, |
|
"logits/chosen": -3.129765033721924, |
|
"logits/rejected": -3.0323197841644287, |
|
"logps/chosen": -742.1157836914062, |
|
"logps/rejected": -210.71441650390625, |
|
"loss": 0.291, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.1236053705215454, |
|
"rewards/margins": 1.71109938621521, |
|
"rewards/rejected": -0.587493896484375, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.010138295881742563, |
|
"grad_norm": 12.640122413635254, |
|
"learning_rate": 2.1002970307704132e-05, |
|
"logits/chosen": -2.989623546600342, |
|
"logits/rejected": -2.9980311393737793, |
|
"logps/chosen": -655.4100341796875, |
|
"logps/rejected": -275.30499267578125, |
|
"loss": 0.1916, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.2308914214372635, |
|
"rewards/margins": 1.7737789154052734, |
|
"rewards/rejected": -1.5428874492645264, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.010172317008862503, |
|
"grad_norm": 261.4872131347656, |
|
"learning_rate": 2.0830781332097446e-05, |
|
"logits/chosen": -3.142162799835205, |
|
"logits/rejected": -3.109694242477417, |
|
"logps/chosen": -452.865966796875, |
|
"logps/rejected": -418.84246826171875, |
|
"loss": 1.1575, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.17142334580421448, |
|
"rewards/margins": -0.32879793643951416, |
|
"rewards/rejected": 0.1573745757341385, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.010206338135982445, |
|
"grad_norm": 8.690618515014648, |
|
"learning_rate": 2.0658795558326743e-05, |
|
"logits/chosen": -3.038311004638672, |
|
"logits/rejected": -3.076416015625, |
|
"logps/chosen": -384.8786315917969, |
|
"logps/rejected": -319.6917419433594, |
|
"loss": 0.16, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.17558330297470093, |
|
"rewards/margins": 1.7798576354980469, |
|
"rewards/rejected": -1.955440878868103, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.010240359263102386, |
|
"grad_norm": 26.635164260864258, |
|
"learning_rate": 2.0487021368736003e-05, |
|
"logits/chosen": -3.059584140777588, |
|
"logits/rejected": -3.065708875656128, |
|
"logps/chosen": -453.93829345703125, |
|
"logps/rejected": -404.24237060546875, |
|
"loss": 0.4096, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.6757233142852783, |
|
"rewards/margins": 1.9321625232696533, |
|
"rewards/rejected": -1.256439208984375, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.010274380390222328, |
|
"grad_norm": 18.957597732543945, |
|
"learning_rate": 2.031546713535688e-05, |
|
"logits/chosen": -3.0744192600250244, |
|
"logits/rejected": -3.0237321853637695, |
|
"logps/chosen": -271.17169189453125, |
|
"logps/rejected": -143.23562622070312, |
|
"loss": 0.5565, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.8414280414581299, |
|
"rewards/margins": 0.7053337097167969, |
|
"rewards/rejected": -1.5467617511749268, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.01030840151734227, |
|
"grad_norm": 9.486793518066406, |
|
"learning_rate": 2.0144141219500705e-05, |
|
"logits/chosen": -3.1045191287994385, |
|
"logits/rejected": -3.109177589416504, |
|
"logps/chosen": -536.4464721679688, |
|
"logps/rejected": -439.09405517578125, |
|
"loss": 0.1462, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 1.618796944618225, |
|
"rewards/margins": 2.926623821258545, |
|
"rewards/rejected": -1.3078268766403198, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.010342422644462211, |
|
"grad_norm": 26.20615577697754, |
|
"learning_rate": 1.9973051971350888e-05, |
|
"logits/chosen": -3.0797457695007324, |
|
"logits/rejected": -3.088682174682617, |
|
"logps/chosen": -199.53915405273438, |
|
"logps/rejected": -279.2613525390625, |
|
"loss": 1.1081, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -1.1805038452148438, |
|
"rewards/margins": -0.5185167193412781, |
|
"rewards/rejected": -0.6619871258735657, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.010376443771582153, |
|
"grad_norm": 18.54012680053711, |
|
"learning_rate": 1.980220772955602e-05, |
|
"logits/chosen": -3.0850045680999756, |
|
"logits/rejected": -3.09476637840271, |
|
"logps/chosen": -333.60491943359375, |
|
"logps/rejected": -295.305908203125, |
|
"loss": 0.3441, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.4468861222267151, |
|
"rewards/margins": 1.9292519092559814, |
|
"rewards/rejected": -1.4823658466339111, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.010410464898702095, |
|
"grad_norm": 18.82195281982422, |
|
"learning_rate": 1.963161682082342e-05, |
|
"logits/chosen": -3.1529183387756348, |
|
"logits/rejected": -3.1466455459594727, |
|
"logps/chosen": -266.6943664550781, |
|
"logps/rejected": -221.9112548828125, |
|
"loss": 0.4877, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.0366407632827759, |
|
"rewards/margins": 1.299647331237793, |
|
"rewards/rejected": -0.26300662755966187, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.010444486025822035, |
|
"grad_norm": 20.001497268676758, |
|
"learning_rate": 1.946128755951332e-05, |
|
"logits/chosen": -3.081289529800415, |
|
"logits/rejected": -3.064981460571289, |
|
"logps/chosen": -213.6260986328125, |
|
"logps/rejected": -203.31961059570312, |
|
"loss": 0.5308, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.6731459498405457, |
|
"rewards/margins": 0.42021849751472473, |
|
"rewards/rejected": -1.0933644771575928, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.010478507152941977, |
|
"grad_norm": 20.88237762451172, |
|
"learning_rate": 1.9291228247233605e-05, |
|
"logits/chosen": -3.081801652908325, |
|
"logits/rejected": -3.058558225631714, |
|
"logps/chosen": -216.79476928710938, |
|
"logps/rejected": -202.99365234375, |
|
"loss": 0.6972, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.14889180660247803, |
|
"rewards/margins": 0.8070772886276245, |
|
"rewards/rejected": -0.658185601234436, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.010512528280061918, |
|
"grad_norm": 6.233214378356934, |
|
"learning_rate": 1.912144717243525e-05, |
|
"logits/chosen": -3.0151007175445557, |
|
"logits/rejected": -2.987820863723755, |
|
"logps/chosen": -460.4840087890625, |
|
"logps/rejected": -179.51492309570312, |
|
"loss": 0.1057, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.363726407289505, |
|
"rewards/margins": 2.340445041656494, |
|
"rewards/rejected": -1.9767186641693115, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.01054654940718186, |
|
"grad_norm": 27.89762306213379, |
|
"learning_rate": 1.895195261000831e-05, |
|
"logits/chosen": -3.1034395694732666, |
|
"logits/rejected": -3.0906271934509277, |
|
"logps/chosen": -224.1537628173828, |
|
"logps/rejected": -144.4778594970703, |
|
"loss": 1.521, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.8860660791397095, |
|
"rewards/margins": -0.28250885009765625, |
|
"rewards/rejected": -0.6035573482513428, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.010580570534301802, |
|
"grad_norm": 40.458892822265625, |
|
"learning_rate": 1.8782752820878634e-05, |
|
"logits/chosen": -3.121913194656372, |
|
"logits/rejected": -3.0847465991973877, |
|
"logps/chosen": -325.77825927734375, |
|
"logps/rejected": -400.49200439453125, |
|
"loss": 1.6415, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.15024378895759583, |
|
"rewards/margins": -0.814440131187439, |
|
"rewards/rejected": 0.9646838903427124, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.010614591661421743, |
|
"grad_norm": 18.067075729370117, |
|
"learning_rate": 1.8613856051605243e-05, |
|
"logits/chosen": -3.118083953857422, |
|
"logits/rejected": -3.129699468612671, |
|
"logps/chosen": -292.23211669921875, |
|
"logps/rejected": -207.62652587890625, |
|
"loss": 0.5406, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.8503540754318237, |
|
"rewards/margins": 1.4205927848815918, |
|
"rewards/rejected": -2.270946979522705, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.010648612788541685, |
|
"grad_norm": 21.936250686645508, |
|
"learning_rate": 1.8445270533978388e-05, |
|
"logits/chosen": -3.071958541870117, |
|
"logits/rejected": -3.045109510421753, |
|
"logps/chosen": -211.42446899414062, |
|
"logps/rejected": -221.9840545654297, |
|
"loss": 0.8057, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.086125135421753, |
|
"rewards/margins": 0.7886361479759216, |
|
"rewards/rejected": 0.2974889874458313, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.010682633915661625, |
|
"grad_norm": 40.1848030090332, |
|
"learning_rate": 1.827700448461836e-05, |
|
"logits/chosen": -3.0537211894989014, |
|
"logits/rejected": -3.0256903171539307, |
|
"logps/chosen": -342.0927734375, |
|
"logps/rejected": -434.037353515625, |
|
"loss": 1.7644, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": 1.124205231666565, |
|
"rewards/margins": -0.8373633027076721, |
|
"rewards/rejected": 1.9615685939788818, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.010716655042781567, |
|
"grad_norm": 21.452964782714844, |
|
"learning_rate": 1.8109066104575023e-05, |
|
"logits/chosen": -3.0191309452056885, |
|
"logits/rejected": -3.028496503829956, |
|
"logps/chosen": -170.28369140625, |
|
"logps/rejected": -191.6688232421875, |
|
"loss": 0.8156, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -1.139381766319275, |
|
"rewards/margins": 0.021857857704162598, |
|
"rewards/rejected": -1.1612396240234375, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.010750676169901508, |
|
"grad_norm": 28.763195037841797, |
|
"learning_rate": 1.7941463578928086e-05, |
|
"logits/chosen": -3.0125069618225098, |
|
"logits/rejected": -2.9951272010803223, |
|
"logps/chosen": -162.54168701171875, |
|
"logps/rejected": -188.5244903564453, |
|
"loss": 1.2491, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -1.842651128768921, |
|
"rewards/margins": -0.6431676149368286, |
|
"rewards/rejected": -1.1994833946228027, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.01078469729702145, |
|
"grad_norm": 29.91036605834961, |
|
"learning_rate": 1.7774205076388206e-05, |
|
"logits/chosen": -3.0470335483551025, |
|
"logits/rejected": -3.040771007537842, |
|
"logps/chosen": -444.4345703125, |
|
"logps/rejected": -469.51080322265625, |
|
"loss": 0.6804, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.18205317854881287, |
|
"rewards/margins": 0.12004223465919495, |
|
"rewards/rejected": -0.3020954132080078, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.010818718424141392, |
|
"grad_norm": 28.914196014404297, |
|
"learning_rate": 1.7607298748898842e-05, |
|
"logits/chosen": -2.9763503074645996, |
|
"logits/rejected": -3.0277490615844727, |
|
"logps/chosen": -353.9087219238281, |
|
"logps/rejected": -278.2948913574219, |
|
"loss": 1.1554, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -0.05427783727645874, |
|
"rewards/margins": 0.21850907802581787, |
|
"rewards/rejected": -0.2727869153022766, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.010852739551261334, |
|
"grad_norm": 9.441596031188965, |
|
"learning_rate": 1.744075273123889e-05, |
|
"logits/chosen": -3.0829198360443115, |
|
"logits/rejected": -3.0089874267578125, |
|
"logps/chosen": -542.9083862304688, |
|
"logps/rejected": -268.1593933105469, |
|
"loss": 0.1301, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 1.6913819313049316, |
|
"rewards/margins": 2.48256516456604, |
|
"rewards/rejected": -0.7911832928657532, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.010886760678381275, |
|
"grad_norm": 10.971582412719727, |
|
"learning_rate": 1.7274575140626318e-05, |
|
"logits/chosen": -3.0915372371673584, |
|
"logits/rejected": -3.0577850341796875, |
|
"logps/chosen": -388.7813415527344, |
|
"logps/rejected": -392.7474365234375, |
|
"loss": 0.1695, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 1.556161642074585, |
|
"rewards/margins": 2.4083662033081055, |
|
"rewards/rejected": -0.8522045612335205, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.010920781805501217, |
|
"grad_norm": 12.355932235717773, |
|
"learning_rate": 1.7108774076322443e-05, |
|
"logits/chosen": -3.1104588508605957, |
|
"logits/rejected": -3.080660820007324, |
|
"logps/chosen": -206.50411987304688, |
|
"logps/rejected": -217.62088012695312, |
|
"loss": 0.3153, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.02197819948196411, |
|
"rewards/margins": 1.7108386754989624, |
|
"rewards/rejected": -1.688860535621643, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.010954802932621157, |
|
"grad_norm": 21.490591049194336, |
|
"learning_rate": 1.6943357619237226e-05, |
|
"logits/chosen": -3.0338594913482666, |
|
"logits/rejected": -2.994076728820801, |
|
"logps/chosen": -409.81402587890625, |
|
"logps/rejected": -183.5924072265625, |
|
"loss": 0.7803, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.31724247336387634, |
|
"rewards/margins": 0.8221940994262695, |
|
"rewards/rejected": -1.1394366025924683, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.010988824059741099, |
|
"grad_norm": 18.730703353881836, |
|
"learning_rate": 1.677833383153542e-05, |
|
"logits/chosen": -3.0578455924987793, |
|
"logits/rejected": -3.0455541610717773, |
|
"logps/chosen": -418.79669189453125, |
|
"logps/rejected": -261.9721984863281, |
|
"loss": 0.6056, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.0711638927459717, |
|
"rewards/margins": 1.3760437965393066, |
|
"rewards/rejected": -0.30487996339797974, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.01102284518686104, |
|
"grad_norm": 23.7857608795166, |
|
"learning_rate": 1.6613710756243626e-05, |
|
"logits/chosen": -3.0486016273498535, |
|
"logits/rejected": -2.983973979949951, |
|
"logps/chosen": -284.9371032714844, |
|
"logps/rejected": -253.89512634277344, |
|
"loss": 0.6952, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.7594280242919922, |
|
"rewards/margins": 0.8340873122215271, |
|
"rewards/rejected": -0.07465922832489014, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.011056866313980982, |
|
"grad_norm": 30.229217529296875, |
|
"learning_rate": 1.6449496416858284e-05, |
|
"logits/chosen": -3.059001922607422, |
|
"logits/rejected": -3.0561468601226807, |
|
"logps/chosen": -344.63458251953125, |
|
"logps/rejected": -310.771728515625, |
|
"loss": 0.8006, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.3245196342468262, |
|
"rewards/margins": -0.025284990668296814, |
|
"rewards/rejected": 1.3498046398162842, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.011090887441100924, |
|
"grad_norm": 23.202922821044922, |
|
"learning_rate": 1.6285698816954624e-05, |
|
"logits/chosen": -3.0508437156677246, |
|
"logits/rejected": -3.077416181564331, |
|
"logps/chosen": -178.1719970703125, |
|
"logps/rejected": -273.89520263671875, |
|
"loss": 0.9761, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -0.5066389441490173, |
|
"rewards/margins": -0.041522592306137085, |
|
"rewards/rejected": -0.46511632204055786, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.011124908568220866, |
|
"grad_norm": 19.027143478393555, |
|
"learning_rate": 1.612232593979658e-05, |
|
"logits/chosen": -3.050765037536621, |
|
"logits/rejected": -3.0737788677215576, |
|
"logps/chosen": -273.1347351074219, |
|
"logps/rejected": -219.3900909423828, |
|
"loss": 0.4248, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.2776317298412323, |
|
"rewards/margins": 1.1471221446990967, |
|
"rewards/rejected": -1.4247537851333618, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.011158929695340807, |
|
"grad_norm": 10.706300735473633, |
|
"learning_rate": 1.5959385747947698e-05, |
|
"logits/chosen": -3.0759172439575195, |
|
"logits/rejected": -3.0675525665283203, |
|
"logps/chosen": -364.5429382324219, |
|
"logps/rejected": -149.35906982421875, |
|
"loss": 0.2888, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.7931306958198547, |
|
"rewards/margins": 1.4744360446929932, |
|
"rewards/rejected": -0.6813052892684937, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.011192950822460749, |
|
"grad_norm": 18.652679443359375, |
|
"learning_rate": 1.5796886182883053e-05, |
|
"logits/chosen": -2.9733424186706543, |
|
"logits/rejected": -3.0072081089019775, |
|
"logps/chosen": -551.43994140625, |
|
"logps/rejected": -393.4814758300781, |
|
"loss": 0.4622, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.2505404949188232, |
|
"rewards/margins": 0.7959327697753906, |
|
"rewards/rejected": 0.454607754945755, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.011226971949580689, |
|
"grad_norm": 20.83335304260254, |
|
"learning_rate": 1.56348351646022e-05, |
|
"logits/chosen": -2.9565417766571045, |
|
"logits/rejected": -2.996302366256714, |
|
"logps/chosen": -211.92990112304688, |
|
"logps/rejected": -285.9666748046875, |
|
"loss": 0.7127, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.0533920526504517, |
|
"rewards/margins": 0.010834500193595886, |
|
"rewards/rejected": -1.064226508140564, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.01126099307670063, |
|
"grad_norm": 27.274747848510742, |
|
"learning_rate": 1.547324059124315e-05, |
|
"logits/chosen": -3.022899866104126, |
|
"logits/rejected": -3.034066677093506, |
|
"logps/chosen": -279.33935546875, |
|
"logps/rejected": -259.1253662109375, |
|
"loss": 0.8729, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.5358911156654358, |
|
"rewards/margins": 0.3108682632446289, |
|
"rewards/rejected": 0.22502289712429047, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.011295014203820572, |
|
"grad_norm": 5.232192516326904, |
|
"learning_rate": 1.5312110338697426e-05, |
|
"logits/chosen": -2.9983677864074707, |
|
"logits/rejected": -2.987232208251953, |
|
"logps/chosen": -589.9959106445312, |
|
"logps/rejected": -182.52914428710938, |
|
"loss": 0.085, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 2.1566078662872314, |
|
"rewards/margins": 2.945901870727539, |
|
"rewards/rejected": -0.7892938852310181, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.011329035330940514, |
|
"grad_norm": 6.4158406257629395, |
|
"learning_rate": 1.5151452260226224e-05, |
|
"logits/chosen": -3.026732921600342, |
|
"logits/rejected": -3.0291504859924316, |
|
"logps/chosen": -345.53887939453125, |
|
"logps/rejected": -212.79861450195312, |
|
"loss": 0.1613, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 1.472572684288025, |
|
"rewards/margins": 2.6776487827301025, |
|
"rewards/rejected": -1.2050762176513672, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.011363056458060456, |
|
"grad_norm": 35.42536926269531, |
|
"learning_rate": 1.4991274186077632e-05, |
|
"logits/chosen": -3.0808024406433105, |
|
"logits/rejected": -3.098182439804077, |
|
"logps/chosen": -345.3873596191406, |
|
"logps/rejected": -328.4718017578125, |
|
"loss": 1.1931, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -1.2231285572052002, |
|
"rewards/margins": -0.6565406918525696, |
|
"rewards/rejected": -0.5665878057479858, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.011397077585180397, |
|
"grad_norm": 19.117053985595703, |
|
"learning_rate": 1.4831583923104999e-05, |
|
"logits/chosen": -3.036012649536133, |
|
"logits/rejected": -2.988028049468994, |
|
"logps/chosen": -415.3401794433594, |
|
"logps/rejected": -192.91635131835938, |
|
"loss": 0.4422, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.9942634105682373, |
|
"rewards/margins": 1.7577967643737793, |
|
"rewards/rejected": 0.23646679520606995, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.01143109871230034, |
|
"grad_norm": 20.918046951293945, |
|
"learning_rate": 1.467238925438646e-05, |
|
"logits/chosen": -3.0097744464874268, |
|
"logits/rejected": -3.056058406829834, |
|
"logps/chosen": -338.6103515625, |
|
"logps/rejected": -315.48883056640625, |
|
"loss": 0.5847, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.8200477361679077, |
|
"rewards/margins": 0.47310030460357666, |
|
"rewards/rejected": 0.34694749116897583, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.01146511983942028, |
|
"grad_norm": 27.27528953552246, |
|
"learning_rate": 1.4513697938845572e-05, |
|
"logits/chosen": -2.995814323425293, |
|
"logits/rejected": -2.9693636894226074, |
|
"logps/chosen": -303.2396240234375, |
|
"logps/rejected": -243.76248168945312, |
|
"loss": 0.8536, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.6613872647285461, |
|
"rewards/margins": 0.23751217126846313, |
|
"rewards/rejected": 0.42387503385543823, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.01149914096654022, |
|
"grad_norm": 35.35068130493164, |
|
"learning_rate": 1.4355517710873184e-05, |
|
"logits/chosen": -3.057015895843506, |
|
"logits/rejected": -3.009333848953247, |
|
"logps/chosen": -276.9962158203125, |
|
"logps/rejected": -268.44415283203125, |
|
"loss": 2.1207, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.11296120285987854, |
|
"rewards/margins": -1.4428050518035889, |
|
"rewards/rejected": 1.3298439979553223, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.011533162093660163, |
|
"grad_norm": 25.704814910888672, |
|
"learning_rate": 1.4197856279950438e-05, |
|
"logits/chosen": -2.9994757175445557, |
|
"logits/rejected": -3.013826847076416, |
|
"logps/chosen": -315.23138427734375, |
|
"logps/rejected": -241.8716278076172, |
|
"loss": 0.6619, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.558853268623352, |
|
"rewards/margins": 1.597301959991455, |
|
"rewards/rejected": -1.038448691368103, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.011567183220780104, |
|
"grad_norm": 20.282594680786133, |
|
"learning_rate": 1.4040721330273062e-05, |
|
"logits/chosen": -3.0336432456970215, |
|
"logits/rejected": -3.0258731842041016, |
|
"logps/chosen": -255.81939697265625, |
|
"logps/rejected": -169.42164611816406, |
|
"loss": 0.6552, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.40217888355255127, |
|
"rewards/margins": 0.9305892586708069, |
|
"rewards/rejected": -0.5284103155136108, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.011601204347900046, |
|
"grad_norm": 3.7436513900756836, |
|
"learning_rate": 1.388412052037682e-05, |
|
"logits/chosen": -3.025284767150879, |
|
"logits/rejected": -3.004758358001709, |
|
"logps/chosen": -497.70721435546875, |
|
"logps/rejected": -230.55877685546875, |
|
"loss": 0.0685, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 1.3800829648971558, |
|
"rewards/margins": 3.0754141807556152, |
|
"rewards/rejected": -1.695331335067749, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.011635225475019988, |
|
"grad_norm": 15.924072265625, |
|
"learning_rate": 1.3728061482764238e-05, |
|
"logits/chosen": -3.087947368621826, |
|
"logits/rejected": -3.0811054706573486, |
|
"logps/chosen": -380.2074890136719, |
|
"logps/rejected": -352.7001647949219, |
|
"loss": 0.315, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.3308167457580566, |
|
"rewards/margins": 1.5659781694412231, |
|
"rewards/rejected": -0.23516160249710083, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.01166924660213993, |
|
"grad_norm": 35.99821472167969, |
|
"learning_rate": 1.3572551823532654e-05, |
|
"logits/chosen": -3.097698450088501, |
|
"logits/rejected": -3.0401198863983154, |
|
"logps/chosen": -256.2811279296875, |
|
"logps/rejected": -387.823486328125, |
|
"loss": 1.068, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.0692390501499176, |
|
"rewards/margins": -0.10297507047653198, |
|
"rewards/rejected": 0.033736035227775574, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.011703267729259871, |
|
"grad_norm": 13.269058227539062, |
|
"learning_rate": 1.3417599122003464e-05, |
|
"logits/chosen": -3.063401222229004, |
|
"logits/rejected": -3.0312092304229736, |
|
"logps/chosen": -382.0020446777344, |
|
"logps/rejected": -195.6297607421875, |
|
"loss": 0.4706, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.2467381954193115, |
|
"rewards/margins": 1.155901551246643, |
|
"rewards/rejected": 0.09083673357963562, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.011737288856379811, |
|
"grad_norm": 10.342382431030273, |
|
"learning_rate": 1.3263210930352737e-05, |
|
"logits/chosen": -3.0519509315490723, |
|
"logits/rejected": -3.012207508087158, |
|
"logps/chosen": -191.2939910888672, |
|
"logps/rejected": -126.82672119140625, |
|
"loss": 0.2744, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.5850515961647034, |
|
"rewards/margins": 1.6264656782150269, |
|
"rewards/rejected": -1.0414141416549683, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.011771309983499753, |
|
"grad_norm": 20.50326919555664, |
|
"learning_rate": 1.3109394773243117e-05, |
|
"logits/chosen": -3.0429341793060303, |
|
"logits/rejected": -2.9837708473205566, |
|
"logps/chosen": -461.16131591796875, |
|
"logps/rejected": -208.45947265625, |
|
"loss": 0.4989, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.6980278491973877, |
|
"rewards/margins": 1.450954556465149, |
|
"rewards/rejected": -0.7529267072677612, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.011805331110619694, |
|
"grad_norm": 25.575138092041016, |
|
"learning_rate": 1.2956158147457115e-05, |
|
"logits/chosen": -2.9965171813964844, |
|
"logits/rejected": -3.0495352745056152, |
|
"logps/chosen": -403.685791015625, |
|
"logps/rejected": -208.33901977539062, |
|
"loss": 1.1954, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.11700379848480225, |
|
"rewards/margins": 0.3518708050251007, |
|
"rewards/rejected": -0.4688745439052582, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.011839352237739636, |
|
"grad_norm": 17.425708770751953, |
|
"learning_rate": 1.280350852153168e-05, |
|
"logits/chosen": -3.0574798583984375, |
|
"logits/rejected": -3.11017107963562, |
|
"logps/chosen": -289.36285400390625, |
|
"logps/rejected": -351.8661804199219, |
|
"loss": 0.3493, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.2467730045318604, |
|
"rewards/margins": 2.5243654251098633, |
|
"rewards/rejected": -1.2775925397872925, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.011873373364859578, |
|
"grad_norm": 32.092559814453125, |
|
"learning_rate": 1.2651453335394231e-05, |
|
"logits/chosen": -3.0130808353424072, |
|
"logits/rejected": -3.0014123916625977, |
|
"logps/chosen": -204.8560028076172, |
|
"logps/rejected": -186.54983520507812, |
|
"loss": 1.6882, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -0.09507843852043152, |
|
"rewards/margins": -1.015676736831665, |
|
"rewards/rejected": 0.9205982685089111, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.01190739449197952, |
|
"grad_norm": 27.803937911987305, |
|
"learning_rate": 1.2500000000000006e-05, |
|
"logits/chosen": -2.9791035652160645, |
|
"logits/rejected": -2.9889700412750244, |
|
"logps/chosen": -480.4871520996094, |
|
"logps/rejected": -195.9858856201172, |
|
"loss": 1.0231, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.4147427082061768, |
|
"rewards/margins": 0.5014397501945496, |
|
"rewards/rejected": 1.9133028984069824, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.011941415619099461, |
|
"grad_norm": 19.826087951660156, |
|
"learning_rate": 1.234915589697091e-05, |
|
"logits/chosen": -3.141842842102051, |
|
"logits/rejected": -3.1288604736328125, |
|
"logps/chosen": -247.04510498046875, |
|
"logps/rejected": -192.33494567871094, |
|
"loss": 0.7628, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.4595264792442322, |
|
"rewards/margins": 0.13754770159721375, |
|
"rewards/rejected": 0.32197877764701843, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.011975436746219403, |
|
"grad_norm": 22.731517791748047, |
|
"learning_rate": 1.2198928378235716e-05, |
|
"logits/chosen": -3.001542091369629, |
|
"logits/rejected": -3.0244500637054443, |
|
"logps/chosen": -213.20513916015625, |
|
"logps/rejected": -399.69024658203125, |
|
"loss": 0.6189, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.21581439673900604, |
|
"rewards/margins": 0.29288235306739807, |
|
"rewards/rejected": -0.07706794142723083, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.012009457873339343, |
|
"grad_norm": 25.124303817749023, |
|
"learning_rate": 1.2049324765671749e-05, |
|
"logits/chosen": -3.0697011947631836, |
|
"logits/rejected": -3.0779526233673096, |
|
"logps/chosen": -354.6813049316406, |
|
"logps/rejected": -199.56146240234375, |
|
"loss": 0.8109, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -0.12926557660102844, |
|
"rewards/margins": -0.1552991420030594, |
|
"rewards/rejected": 0.02603357657790184, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.012043479000459285, |
|
"grad_norm": 17.857685089111328, |
|
"learning_rate": 1.1900352350748026e-05, |
|
"logits/chosen": -3.080061674118042, |
|
"logits/rejected": -3.030620574951172, |
|
"logps/chosen": -397.3528747558594, |
|
"logps/rejected": -219.3166961669922, |
|
"loss": 0.3711, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.5567859411239624, |
|
"rewards/margins": 1.1941144466400146, |
|
"rewards/rejected": -0.6373286247253418, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.012077500127579226, |
|
"grad_norm": 13.247682571411133, |
|
"learning_rate": 1.175201839416988e-05, |
|
"logits/chosen": -3.049644708633423, |
|
"logits/rejected": -3.0940423011779785, |
|
"logps/chosen": -257.4561462402344, |
|
"logps/rejected": -282.6706848144531, |
|
"loss": 0.273, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.035192608833313, |
|
"rewards/margins": 2.3947765827178955, |
|
"rewards/rejected": -1.359583854675293, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.012111521254699168, |
|
"grad_norm": 19.514381408691406, |
|
"learning_rate": 1.1604330125525079e-05, |
|
"logits/chosen": -3.0747852325439453, |
|
"logits/rejected": -3.0953102111816406, |
|
"logps/chosen": -287.0201416015625, |
|
"logps/rejected": -241.919677734375, |
|
"loss": 0.8225, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.1913723647594452, |
|
"rewards/margins": 0.6621765494346619, |
|
"rewards/rejected": -0.47080421447753906, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.01214554238181911, |
|
"grad_norm": 24.5207462310791, |
|
"learning_rate": 1.1457294742931507e-05, |
|
"logits/chosen": -3.119217872619629, |
|
"logits/rejected": -3.0973029136657715, |
|
"logps/chosen": -314.3831787109375, |
|
"logps/rejected": -241.65138244628906, |
|
"loss": 0.6746, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.13125188648700714, |
|
"rewards/margins": 0.20069465041160583, |
|
"rewards/rejected": -0.0694427490234375, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.012179563508939052, |
|
"grad_norm": 14.060046195983887, |
|
"learning_rate": 1.1310919412686247e-05, |
|
"logits/chosen": -3.079954147338867, |
|
"logits/rejected": -3.0620617866516113, |
|
"logps/chosen": -614.0543823242188, |
|
"logps/rejected": -203.45040893554688, |
|
"loss": 0.3499, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 0.7918987274169922, |
|
"rewards/margins": 1.1504579782485962, |
|
"rewards/rejected": -0.358559250831604, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.012213584636058993, |
|
"grad_norm": 21.438091278076172, |
|
"learning_rate": 1.11652112689164e-05, |
|
"logits/chosen": -3.038268566131592, |
|
"logits/rejected": -3.041386604309082, |
|
"logps/chosen": -116.21704864501953, |
|
"logps/rejected": -188.45834350585938, |
|
"loss": 0.6917, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.015452951192855835, |
|
"rewards/margins": 0.3833298087120056, |
|
"rewards/rejected": -0.39878273010253906, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.012247605763178935, |
|
"grad_norm": 19.021968841552734, |
|
"learning_rate": 1.1020177413231334e-05, |
|
"logits/chosen": -3.105311632156372, |
|
"logits/rejected": -3.0912036895751953, |
|
"logps/chosen": -310.29547119140625, |
|
"logps/rejected": -235.15786743164062, |
|
"loss": 0.7319, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.29602566361427307, |
|
"rewards/margins": 0.21368007361888885, |
|
"rewards/rejected": 0.08234557509422302, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.012281626890298875, |
|
"grad_norm": 13.478549003601074, |
|
"learning_rate": 1.0875824914376553e-05, |
|
"logits/chosen": -3.084926128387451, |
|
"logits/rejected": -3.035360813140869, |
|
"logps/chosen": -363.0761413574219, |
|
"logps/rejected": -187.72837829589844, |
|
"loss": 0.3199, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": 1.0591999292373657, |
|
"rewards/margins": 1.126670479774475, |
|
"rewards/rejected": -0.06747055053710938, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.012315648017418817, |
|
"grad_norm": 40.014034271240234, |
|
"learning_rate": 1.0732160807889211e-05, |
|
"logits/chosen": -3.0882506370544434, |
|
"logits/rejected": -3.0681185722351074, |
|
"logps/chosen": -585.7984008789062, |
|
"logps/rejected": -391.25836181640625, |
|
"loss": 0.8768, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.8234740495681763, |
|
"rewards/margins": -0.049838095903396606, |
|
"rewards/rejected": 0.8733119964599609, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.012349669144538758, |
|
"grad_norm": 22.66176414489746, |
|
"learning_rate": 1.058919209575517e-05, |
|
"logits/chosen": -3.0657193660736084, |
|
"logits/rejected": -3.062636375427246, |
|
"logps/chosen": -248.1912841796875, |
|
"logps/rejected": -292.0156555175781, |
|
"loss": 0.7295, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.12826082110404968, |
|
"rewards/margins": 0.16077953577041626, |
|
"rewards/rejected": -0.03251868486404419, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.0123836902716587, |
|
"grad_norm": 28.293270111083984, |
|
"learning_rate": 1.0446925746067768e-05, |
|
"logits/chosen": -3.0678696632385254, |
|
"logits/rejected": -3.0742006301879883, |
|
"logps/chosen": -287.19683837890625, |
|
"logps/rejected": -394.8347473144531, |
|
"loss": 0.7652, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.5613548755645752, |
|
"rewards/margins": 0.3589908480644226, |
|
"rewards/rejected": 1.2023639678955078, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.012417711398778642, |
|
"grad_norm": 19.34810447692871, |
|
"learning_rate": 1.0305368692688174e-05, |
|
"logits/chosen": -3.0600197315216064, |
|
"logits/rejected": -3.087749481201172, |
|
"logps/chosen": -234.78482055664062, |
|
"logps/rejected": -205.5214385986328, |
|
"loss": 0.6881, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.4390994310379028, |
|
"rewards/margins": 0.13296374678611755, |
|
"rewards/rejected": 1.306135654449463, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.012451732525898583, |
|
"grad_norm": 33.45370101928711, |
|
"learning_rate": 1.0164527834907467e-05, |
|
"logits/chosen": -3.0540003776550293, |
|
"logits/rejected": -3.0027832984924316, |
|
"logps/chosen": -379.1693115234375, |
|
"logps/rejected": -422.2978210449219, |
|
"loss": 1.964, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.17507779598236084, |
|
"rewards/margins": 0.6117099523544312, |
|
"rewards/rejected": -0.4366319179534912, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.012485753653018525, |
|
"grad_norm": 24.815614700317383, |
|
"learning_rate": 1.0024410037110357e-05, |
|
"logits/chosen": -3.097560167312622, |
|
"logits/rejected": -3.097411632537842, |
|
"logps/chosen": -388.7486572265625, |
|
"logps/rejected": -369.4508056640625, |
|
"loss": 0.9352, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 1.432456612586975, |
|
"rewards/margins": 0.7668805122375488, |
|
"rewards/rejected": 0.6655762195587158, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.012519774780138465, |
|
"grad_norm": 23.309213638305664, |
|
"learning_rate": 9.88502212844063e-06, |
|
"logits/chosen": -3.033217430114746, |
|
"logits/rejected": -3.0999505519866943, |
|
"logps/chosen": -287.3164978027344, |
|
"logps/rejected": -341.2158203125, |
|
"loss": 0.7234, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 1.7660259008407593, |
|
"rewards/margins": 0.8636859059333801, |
|
"rewards/rejected": 0.9023399353027344, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.012553795907258407, |
|
"grad_norm": 12.917970657348633, |
|
"learning_rate": 9.746370902468311e-06, |
|
"logits/chosen": -3.058816432952881, |
|
"logits/rejected": -3.067315101623535, |
|
"logps/chosen": -158.72190856933594, |
|
"logps/rejected": -202.031005859375, |
|
"loss": 0.3616, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -0.1266048550605774, |
|
"rewards/margins": 1.6018997430801392, |
|
"rewards/rejected": -1.7285046577453613, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.012587817034378349, |
|
"grad_norm": 20.528718948364258, |
|
"learning_rate": 9.608463116858542e-06, |
|
"logits/chosen": -3.0635955333709717, |
|
"logits/rejected": -3.0514440536499023, |
|
"logps/chosen": -516.4857788085938, |
|
"logps/rejected": -249.27166748046875, |
|
"loss": 0.469, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 0.9522740244865417, |
|
"rewards/margins": 1.3290956020355225, |
|
"rewards/rejected": -0.37682172656059265, |
|
"step": 370 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 500, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 10, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|