{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 80000000000, "global_step": 1718, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 47.75, "learning_rate": 2.9069767441860468e-08, "logits/chosen": -0.30066978931427, "logits/rejected": 0.05270648002624512, "logps/chosen": -292.3870544433594, "logps/rejected": -519.2880859375, "loss": 0.6931, "rewards/accuracies": 0.0, "rewards/chosen": 0.0, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": 0.0, "rewards/student_margin": 0.0, "rewards/teacher_margin": 0.0, "step": 1 }, { "epoch": 0.01, "grad_norm": 50.75, "learning_rate": 2.906976744186047e-07, "logits/chosen": -0.37918758392333984, "logits/rejected": -0.042655572295188904, "logps/chosen": -231.2487030029297, "logps/rejected": -516.9002685546875, "loss": 0.726, "rewards/accuracies": 0.3333333432674408, "rewards/chosen": 0.05387793853878975, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": 0.07821387052536011, "rewards/student_margin": -0.024335933849215508, "rewards/teacher_margin": 0.0, "step": 10 }, { "epoch": 0.01, "grad_norm": 42.75, "learning_rate": 5.813953488372094e-07, "logits/chosen": -0.7622300386428833, "logits/rejected": 0.022924324497580528, "logps/chosen": -245.8932647705078, "logps/rejected": -559.360595703125, "loss": 0.7117, "rewards/accuracies": 0.4333333373069763, "rewards/chosen": 0.04971719905734062, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -0.02675577439367771, "rewards/student_margin": 0.07647298276424408, "rewards/teacher_margin": 0.0, "step": 20 }, { "epoch": 0.02, "grad_norm": 34.25, "learning_rate": 8.720930232558141e-07, "logits/chosen": -0.6480653285980225, "logits/rejected": -0.05629728361964226, "logps/chosen": -207.52798461914062, "logps/rejected": -569.8907470703125, "loss": 0.6704, "rewards/accuracies": 0.6666666865348816, "rewards/chosen": -0.04712574928998947, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -0.31786206364631653, "rewards/student_margin": 0.27073630690574646, "rewards/teacher_margin": 0.0, "step": 30 }, { "epoch": 0.02, "grad_norm": 33.25, "learning_rate": 1.1627906976744188e-06, "logits/chosen": -0.46851086616516113, "logits/rejected": 0.08061075955629349, "logps/chosen": -213.6239013671875, "logps/rejected": -531.3410034179688, "loss": 0.5929, "rewards/accuracies": 0.7666666507720947, "rewards/chosen": -0.09198583662509918, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -0.5451609492301941, "rewards/student_margin": 0.4531751275062561, "rewards/teacher_margin": 0.0, "step": 40 }, { "epoch": 0.03, "grad_norm": 28.125, "learning_rate": 1.4534883720930234e-06, "logits/chosen": -0.6713495850563049, "logits/rejected": -0.03319794684648514, "logps/chosen": -178.42832946777344, "logps/rejected": -427.0347595214844, "loss": 0.5249, "rewards/accuracies": 0.5666666626930237, "rewards/chosen": -0.16660943627357483, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -0.6883589029312134, "rewards/student_margin": 0.5217493772506714, "rewards/teacher_margin": 0.0, "step": 50 }, { "epoch": 0.03, "grad_norm": 23.625, "learning_rate": 1.7441860465116282e-06, "logits/chosen": -0.8985533714294434, "logits/rejected": 0.039773982018232346, "logps/chosen": -175.9773406982422, "logps/rejected": -535.4810791015625, "loss": 0.4963, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.20179271697998047, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.35043203830719, "rewards/student_margin": 1.148639440536499, "rewards/teacher_margin": 0.0, "step": 60 }, { "epoch": 0.04, "grad_norm": 27.875, "learning_rate": 2.0348837209302328e-06, "logits/chosen": -0.5139203071594238, "logits/rejected": 0.02224038355052471, "logps/chosen": -244.5893096923828, "logps/rejected": -573.5228271484375, "loss": 0.4526, "rewards/accuracies": 0.7999999523162842, "rewards/chosen": -0.348739355802536, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.8875904083251953, "rewards/student_margin": 1.5388509035110474, "rewards/teacher_margin": 0.0, "step": 70 }, { "epoch": 0.05, "grad_norm": 18.875, "learning_rate": 2.3255813953488376e-06, "logits/chosen": -0.7934290170669556, "logits/rejected": -0.0012261777883395553, "logps/chosen": -182.20286560058594, "logps/rejected": -427.95037841796875, "loss": 0.4462, "rewards/accuracies": 0.7333333492279053, "rewards/chosen": -0.0741361603140831, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.3484480381011963, "rewards/student_margin": 1.2743120193481445, "rewards/teacher_margin": 0.0, "step": 80 }, { "epoch": 0.05, "grad_norm": 26.25, "learning_rate": 2.6162790697674424e-06, "logits/chosen": -0.7236814498901367, "logits/rejected": 0.027469992637634277, "logps/chosen": -186.38136291503906, "logps/rejected": -399.2889099121094, "loss": 0.4499, "rewards/accuracies": 0.7333332896232605, "rewards/chosen": -0.031203720718622208, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.2050061225891113, "rewards/student_margin": 1.1738024950027466, "rewards/teacher_margin": 0.0, "step": 90 }, { "epoch": 0.06, "grad_norm": 30.0, "learning_rate": 2.9069767441860468e-06, "logits/chosen": -0.4575803279876709, "logits/rejected": -0.1135629191994667, "logps/chosen": -259.8243713378906, "logps/rejected": -481.0874938964844, "loss": 0.4093, "rewards/accuracies": 0.8666666746139526, "rewards/chosen": -0.21589164435863495, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.5411407947540283, "rewards/student_margin": 1.3252493143081665, "rewards/teacher_margin": 0.0, "step": 100 }, { "epoch": 0.06, "grad_norm": 20.5, "learning_rate": 3.1976744186046516e-06, "logits/chosen": -0.3742029070854187, "logits/rejected": -0.12473519891500473, "logps/chosen": -186.20205688476562, "logps/rejected": -392.6462097167969, "loss": 0.3743, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 0.20217347145080566, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.5840258598327637, "rewards/student_margin": 1.7861993312835693, "rewards/teacher_margin": 0.0, "step": 110 }, { "epoch": 0.07, "grad_norm": 19.125, "learning_rate": 3.4883720930232564e-06, "logits/chosen": -0.6747792959213257, "logits/rejected": 0.16625359654426575, "logps/chosen": -229.23672485351562, "logps/rejected": -525.2789306640625, "loss": 0.3796, "rewards/accuracies": 0.8333333730697632, "rewards/chosen": -0.06482670456171036, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.0777878761291504, "rewards/student_margin": 2.012960910797119, "rewards/teacher_margin": 0.0, "step": 120 }, { "epoch": 0.08, "grad_norm": 36.5, "learning_rate": 3.7790697674418607e-06, "logits/chosen": -0.6423686742782593, "logits/rejected": -0.049126945436000824, "logps/chosen": -207.14761352539062, "logps/rejected": -422.570556640625, "loss": 0.3735, "rewards/accuracies": 0.9000000953674316, "rewards/chosen": -0.0891571044921875, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.900564193725586, "rewards/student_margin": 1.8114070892333984, "rewards/teacher_margin": 0.0, "step": 130 }, { "epoch": 0.08, "grad_norm": 22.625, "learning_rate": 4.0697674418604655e-06, "logits/chosen": -0.6022151112556458, "logits/rejected": -0.05691419914364815, "logps/chosen": -294.1002502441406, "logps/rejected": -565.58642578125, "loss": 0.3309, "rewards/accuracies": 0.7999999523162842, "rewards/chosen": -0.2062998265028, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.2702088356018066, "rewards/student_margin": 2.0639092922210693, "rewards/teacher_margin": 0.0, "step": 140 }, { "epoch": 0.09, "grad_norm": 21.5, "learning_rate": 4.36046511627907e-06, "logits/chosen": -0.5255838632583618, "logits/rejected": -0.025744330137968063, "logps/chosen": -206.9980010986328, "logps/rejected": -500.0045471191406, "loss": 0.4189, "rewards/accuracies": 0.800000011920929, "rewards/chosen": 0.19613902270793915, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.1442222595214844, "rewards/student_margin": 2.3403611183166504, "rewards/teacher_margin": 0.0, "step": 150 }, { "epoch": 0.09, "grad_norm": 28.625, "learning_rate": 4.651162790697675e-06, "logits/chosen": -0.595037043094635, "logits/rejected": -0.05052659660577774, "logps/chosen": -228.1767578125, "logps/rejected": -517.8463134765625, "loss": 0.3434, "rewards/accuracies": 0.8333333134651184, "rewards/chosen": 0.011402442120015621, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.753488302230835, "rewards/student_margin": 2.764890670776367, "rewards/teacher_margin": 0.0, "step": 160 }, { "epoch": 0.1, "grad_norm": 22.625, "learning_rate": 4.941860465116279e-06, "logits/chosen": -0.8135232925415039, "logits/rejected": 0.09407637268304825, "logps/chosen": -175.84527587890625, "logps/rejected": -517.9186401367188, "loss": 0.3348, "rewards/accuracies": 1.0, "rewards/chosen": 0.3143172264099121, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -3.1565933227539062, "rewards/student_margin": 3.4709103107452393, "rewards/teacher_margin": 0.0, "step": 170 }, { "epoch": 0.1, "grad_norm": 31.25, "learning_rate": 4.999669659981585e-06, "logits/chosen": -0.6553431153297424, "logits/rejected": 0.02322854846715927, "logps/chosen": -190.44459533691406, "logps/rejected": -498.41845703125, "loss": 0.3838, "rewards/accuracies": 0.7666666507720947, "rewards/chosen": -0.03836626559495926, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.3416810035705566, "rewards/student_margin": 2.303314685821533, "rewards/teacher_margin": 0.0, "step": 180 }, { "epoch": 0.11, "grad_norm": 23.625, "learning_rate": 4.998327803275328e-06, "logits/chosen": -0.5196131467819214, "logits/rejected": 0.03156811743974686, "logps/chosen": -244.62362670898438, "logps/rejected": -550.338134765625, "loss": 0.3397, "rewards/accuracies": 0.8666666150093079, "rewards/chosen": 0.1986195147037506, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.157165050506592, "rewards/student_margin": 2.3557844161987305, "rewards/teacher_margin": 0.0, "step": 190 }, { "epoch": 0.12, "grad_norm": 23.0, "learning_rate": 4.995954337283145e-06, "logits/chosen": -0.46547913551330566, "logits/rejected": 0.002417749259620905, "logps/chosen": -220.58139038085938, "logps/rejected": -457.37957763671875, "loss": 0.3513, "rewards/accuracies": 0.7666667103767395, "rewards/chosen": 0.23564238846302032, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.046499729156494, "rewards/student_margin": 2.282142162322998, "rewards/teacher_margin": 0.0, "step": 200 }, { "epoch": 0.12, "grad_norm": 29.625, "learning_rate": 4.992550242056393e-06, "logits/chosen": -0.7077560424804688, "logits/rejected": 0.23227286338806152, "logps/chosen": -250.8577117919922, "logps/rejected": -470.998046875, "loss": 0.3619, "rewards/accuracies": 0.76666659116745, "rewards/chosen": 0.30192822217941284, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.5922541618347168, "rewards/student_margin": 1.8941824436187744, "rewards/teacher_margin": 0.0, "step": 210 }, { "epoch": 0.13, "grad_norm": 13.4375, "learning_rate": 4.988116923213761e-06, "logits/chosen": -0.5510163903236389, "logits/rejected": 0.061091721057891846, "logps/chosen": -231.3572235107422, "logps/rejected": -456.94219970703125, "loss": 0.3619, "rewards/accuracies": 0.8333333134651184, "rewards/chosen": 0.7294235825538635, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.8068240880966187, "rewards/student_margin": 2.536247491836548, "rewards/teacher_margin": 0.0, "step": 220 }, { "epoch": 0.13, "grad_norm": 23.125, "learning_rate": 4.982656211360852e-06, "logits/chosen": -0.6307552456855774, "logits/rejected": 0.11015477031469345, "logps/chosen": -210.99728393554688, "logps/rejected": -536.0760498046875, "loss": 0.3218, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 0.6962422132492065, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.4401557445526123, "rewards/student_margin": 3.1363983154296875, "rewards/teacher_margin": 0.0, "step": 230 }, { "epoch": 0.14, "grad_norm": 24.75, "learning_rate": 4.976170361334305e-06, "logits/chosen": -0.47195425629615784, "logits/rejected": -0.022416137158870697, "logps/chosen": -255.0712127685547, "logps/rejected": -450.69989013671875, "loss": 0.3947, "rewards/accuracies": 0.8333333134651184, "rewards/chosen": 0.3207937180995941, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.8813111782073975, "rewards/student_margin": 2.2021050453186035, "rewards/teacher_margin": 0.0, "step": 240 }, { "epoch": 0.15, "grad_norm": 39.25, "learning_rate": 4.968662051270712e-06, "logits/chosen": -0.7277335524559021, "logits/rejected": 0.09811758249998093, "logps/chosen": -212.38381958007812, "logps/rejected": -535.5774536132812, "loss": 0.3208, "rewards/accuracies": 0.8333333134651184, "rewards/chosen": 0.7789496183395386, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.0195579528808594, "rewards/student_margin": 2.7985076904296875, "rewards/teacher_margin": 0.0, "step": 250 }, { "epoch": 0.15, "grad_norm": 17.125, "learning_rate": 4.960134381500771e-06, "logits/chosen": -0.6490304470062256, "logits/rejected": 0.1224866732954979, "logps/chosen": -208.1616973876953, "logps/rejected": -549.3804931640625, "loss": 0.3667, "rewards/accuracies": 0.9333332777023315, "rewards/chosen": 1.0950639247894287, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.4546568393707275, "rewards/student_margin": 3.5497207641601562, "rewards/teacher_margin": 0.0, "step": 260 }, { "epoch": 0.16, "grad_norm": 27.5, "learning_rate": 4.950590873269098e-06, "logits/chosen": -0.5456488132476807, "logits/rejected": 0.020637910813093185, "logps/chosen": -259.57196044921875, "logps/rejected": -409.33123779296875, "loss": 0.3453, "rewards/accuracies": 0.7666666507720947, "rewards/chosen": 0.4798993468284607, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.331217885017395, "rewards/student_margin": 1.811117172241211, "rewards/teacher_margin": 0.0, "step": 270 }, { "epoch": 0.16, "grad_norm": 35.0, "learning_rate": 4.940035467280229e-06, "logits/chosen": -0.4896344542503357, "logits/rejected": 0.024026235565543175, "logps/chosen": -237.4656524658203, "logps/rejected": -509.08404541015625, "loss": 0.3292, "rewards/accuracies": 0.9333332777023315, "rewards/chosen": 0.728676974773407, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.34112811088562, "rewards/student_margin": 3.069805145263672, "rewards/teacher_margin": 0.0, "step": 280 }, { "epoch": 0.17, "grad_norm": 26.5, "learning_rate": 4.928472522071432e-06, "logits/chosen": -0.6238890290260315, "logits/rejected": 0.1964757889509201, "logps/chosen": -213.3905029296875, "logps/rejected": -478.8919372558594, "loss": 0.3695, "rewards/accuracies": 0.76666659116745, "rewards/chosen": 0.6496525406837463, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.2729125022888184, "rewards/student_margin": 1.9225651025772095, "rewards/teacher_margin": 0.0, "step": 290 }, { "epoch": 0.17, "grad_norm": 29.375, "learning_rate": 4.915906812212976e-06, "logits/chosen": -0.6193239092826843, "logits/rejected": 0.05876749008893967, "logps/chosen": -202.52391052246094, "logps/rejected": -458.86492919921875, "loss": 0.2946, "rewards/accuracies": 0.8666666150093079, "rewards/chosen": 0.8241189122200012, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.658251166343689, "rewards/student_margin": 2.482369899749756, "rewards/teacher_margin": 0.0, "step": 300 }, { "epoch": 0.18, "grad_norm": 16.125, "learning_rate": 4.9023435263366145e-06, "logits/chosen": -0.9334913492202759, "logits/rejected": -0.058604609221220016, "logps/chosen": -168.749267578125, "logps/rejected": -640.4456176757812, "loss": 0.2114, "rewards/accuracies": 0.966666579246521, "rewards/chosen": 0.5594726800918579, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -3.131923198699951, "rewards/student_margin": 3.6913959980010986, "rewards/teacher_margin": 0.0, "step": 310 }, { "epoch": 0.19, "grad_norm": 15.125, "learning_rate": 4.887788264993099e-06, "logits/chosen": -0.6044551730155945, "logits/rejected": -0.04338273033499718, "logps/chosen": -224.3017578125, "logps/rejected": -549.9353637695312, "loss": 0.3094, "rewards/accuracies": 0.8333333730697632, "rewards/chosen": 0.7394543290138245, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.3911728858947754, "rewards/student_margin": 3.130627393722534, "rewards/teacher_margin": 0.0, "step": 320 }, { "epoch": 0.19, "grad_norm": 23.375, "learning_rate": 4.872247038339603e-06, "logits/chosen": -0.7818952798843384, "logits/rejected": 0.06497599929571152, "logps/chosen": -192.23880004882812, "logps/rejected": -626.657470703125, "loss": 0.2646, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 1.0649678707122803, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -3.6517505645751953, "rewards/student_margin": 4.716717720031738, "rewards/teacher_margin": 0.0, "step": 330 }, { "epoch": 0.2, "grad_norm": 30.625, "learning_rate": 4.855726263657999e-06, "logits/chosen": -0.29226094484329224, "logits/rejected": -0.13234995305538177, "logps/chosen": -264.1004333496094, "logps/rejected": -362.9111633300781, "loss": 0.4197, "rewards/accuracies": 0.800000011920929, "rewards/chosen": 0.6886833906173706, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.4170998334884644, "rewards/student_margin": 2.105782985687256, "rewards/teacher_margin": 0.0, "step": 340 }, { "epoch": 0.2, "grad_norm": 24.25, "learning_rate": 4.838232762705044e-06, "logits/chosen": -0.4636480212211609, "logits/rejected": 0.027485806494951248, "logps/chosen": -194.08128356933594, "logps/rejected": -531.1504516601562, "loss": 0.2651, "rewards/accuracies": 0.9333332777023315, "rewards/chosen": 1.0223414897918701, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.3282663822174072, "rewards/student_margin": 3.3506076335906982, "rewards/teacher_margin": 0.0, "step": 350 }, { "epoch": 0.21, "grad_norm": 27.5, "learning_rate": 4.819773758895538e-06, "logits/chosen": -0.7649397253990173, "logits/rejected": 0.06376661360263824, "logps/chosen": -163.4183807373047, "logps/rejected": -592.5127563476562, "loss": 0.3665, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 0.8098726272583008, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -3.550152540206909, "rewards/student_margin": 4.360025405883789, "rewards/teacher_margin": 0.0, "step": 360 }, { "epoch": 0.22, "grad_norm": 22.125, "learning_rate": 4.8003568743196294e-06, "logits/chosen": -0.5304759740829468, "logits/rejected": 0.034225065261125565, "logps/chosen": -187.56813049316406, "logps/rejected": -569.3456420898438, "loss": 0.2797, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 0.6876587867736816, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.1767399311065674, "rewards/student_margin": 2.864398956298828, "rewards/teacher_margin": 0.0, "step": 370 }, { "epoch": 0.22, "grad_norm": 29.625, "learning_rate": 4.779990126595505e-06, "logits/chosen": -0.5957476496696472, "logits/rejected": -0.04448353499174118, "logps/chosen": -200.004150390625, "logps/rejected": -613.2229614257812, "loss": 0.3936, "rewards/accuracies": 0.9333332777023315, "rewards/chosen": 0.7026926279067993, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.0374011993408203, "rewards/student_margin": 2.740093946456909, "rewards/teacher_margin": 0.0, "step": 380 }, { "epoch": 0.23, "grad_norm": 26.375, "learning_rate": 4.758681925558756e-06, "logits/chosen": -0.5602328777313232, "logits/rejected": -0.012245875783264637, "logps/chosen": -185.31021118164062, "logps/rejected": -407.41766357421875, "loss": 0.3227, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 0.620998740196228, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.894086480140686, "rewards/student_margin": 2.515085458755493, "rewards/teacher_margin": 0.0, "step": 390 }, { "epoch": 0.23, "grad_norm": 20.125, "learning_rate": 4.736441069789786e-06, "logits/chosen": -0.46345287561416626, "logits/rejected": 0.001672474667429924, "logps/chosen": -228.4940948486328, "logps/rejected": -361.48175048828125, "loss": 0.2846, "rewards/accuracies": 0.6999999284744263, "rewards/chosen": 0.4205475449562073, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.4956649541854858, "rewards/student_margin": 1.9162124395370483, "rewards/teacher_margin": 0.0, "step": 400 }, { "epoch": 0.24, "grad_norm": 16.5, "learning_rate": 4.713276742980708e-06, "logits/chosen": -0.4360102713108063, "logits/rejected": -0.10899927467107773, "logps/chosen": -249.0923309326172, "logps/rejected": -560.4241333007812, "loss": 0.2865, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 0.8636870384216309, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.2510733604431152, "rewards/student_margin": 3.114760637283325, "rewards/teacher_margin": 0.0, "step": 410 }, { "epoch": 0.24, "grad_norm": 30.0, "learning_rate": 4.6891985101432085e-06, "logits/chosen": -0.3729998469352722, "logits/rejected": -0.028511980548501015, "logps/chosen": -207.8327178955078, "logps/rejected": -356.3282775878906, "loss": 0.3602, "rewards/accuracies": 0.7333332896232605, "rewards/chosen": 0.9164339303970337, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -0.7431859374046326, "rewards/student_margin": 1.659619927406311, "rewards/teacher_margin": 0.0, "step": 420 }, { "epoch": 0.25, "grad_norm": 29.0, "learning_rate": 4.664216313658969e-06, "logits/chosen": -0.324567973613739, "logits/rejected": 0.016650067642331123, "logps/chosen": -206.4794921875, "logps/rejected": -445.509033203125, "loss": 0.3204, "rewards/accuracies": 0.8666666746139526, "rewards/chosen": 1.2572410106658936, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -0.936699390411377, "rewards/student_margin": 2.1939404010772705, "rewards/teacher_margin": 0.0, "step": 430 }, { "epoch": 0.26, "grad_norm": 28.0, "learning_rate": 4.638340469174256e-06, "logits/chosen": -0.5071316957473755, "logits/rejected": 0.044037409126758575, "logps/chosen": -180.57435607910156, "logps/rejected": -601.9741821289062, "loss": 0.2881, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 1.1313551664352417, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.505422830581665, "rewards/student_margin": 3.6367783546447754, "rewards/teacher_margin": 0.0, "step": 440 }, { "epoch": 0.26, "grad_norm": 18.625, "learning_rate": 4.611581661340386e-06, "logits/chosen": -0.5572972297668457, "logits/rejected": 0.05740237236022949, "logps/chosen": -202.59849548339844, "logps/rejected": -579.298095703125, "loss": 0.2705, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 1.076784372329712, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.998802661895752, "rewards/student_margin": 4.075586795806885, "rewards/teacher_margin": 0.0, "step": 450 }, { "epoch": 0.27, "grad_norm": 23.625, "learning_rate": 4.5839509394018185e-06, "logits/chosen": -0.37410736083984375, "logits/rejected": 0.017481276765465736, "logps/chosen": -243.3333282470703, "logps/rejected": -485.5693359375, "loss": 0.2463, "rewards/accuracies": 0.8666666150093079, "rewards/chosen": 0.691962480545044, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.156360149383545, "rewards/student_margin": 2.8483223915100098, "rewards/teacher_margin": 0.0, "step": 460 }, { "epoch": 0.27, "grad_norm": 11.9375, "learning_rate": 4.555459712633707e-06, "logits/chosen": -0.6188653707504272, "logits/rejected": 0.16356995701789856, "logps/chosen": -238.00772094726562, "logps/rejected": -447.47442626953125, "loss": 0.273, "rewards/accuracies": 0.8333333730697632, "rewards/chosen": 1.3531479835510254, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.5606005191802979, "rewards/student_margin": 2.913748264312744, "rewards/teacher_margin": 0.0, "step": 470 }, { "epoch": 0.28, "grad_norm": 15.5625, "learning_rate": 4.526119745630779e-06, "logits/chosen": -0.5436848402023315, "logits/rejected": 0.04653601720929146, "logps/chosen": -228.1831817626953, "logps/rejected": -620.6707153320312, "loss": 0.2793, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 1.1916218996047974, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.1671457290649414, "rewards/student_margin": 3.3587677478790283, "rewards/teacher_margin": 0.0, "step": 480 }, { "epoch": 0.29, "grad_norm": 20.125, "learning_rate": 4.4959431534495015e-06, "logits/chosen": -0.1781628429889679, "logits/rejected": 0.06491532176733017, "logps/chosen": -266.6177673339844, "logps/rejected": -403.18841552734375, "loss": 0.2836, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 2.089085340499878, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -0.6247254610061646, "rewards/student_margin": 2.713810682296753, "rewards/teacher_margin": 0.0, "step": 490 }, { "epoch": 0.29, "grad_norm": 28.625, "learning_rate": 4.464942396605531e-06, "logits/chosen": -0.5647803544998169, "logits/rejected": 0.09310182183980942, "logps/chosen": -185.3010711669922, "logps/rejected": -453.74188232421875, "loss": 0.2228, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.3027641773223877, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.6533056497573853, "rewards/student_margin": 2.9560697078704834, "rewards/teacher_margin": 0.0, "step": 500 }, { "epoch": 0.3, "grad_norm": 27.625, "learning_rate": 4.433130275928522e-06, "logits/chosen": -0.508124828338623, "logits/rejected": 0.05562957376241684, "logps/chosen": -202.5834197998047, "logps/rejected": -409.6363220214844, "loss": 0.2788, "rewards/accuracies": 0.8333333730697632, "rewards/chosen": 1.8609619140625, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.7967350482940674, "rewards/student_margin": 3.6576972007751465, "rewards/teacher_margin": 0.0, "step": 510 }, { "epoch": 0.3, "grad_norm": 17.0, "learning_rate": 4.400519927276411e-06, "logits/chosen": -0.48706334829330444, "logits/rejected": -0.07246413826942444, "logps/chosen": -216.9883270263672, "logps/rejected": -562.3739013671875, "loss": 0.2144, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.0860559940338135, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -3.157426357269287, "rewards/student_margin": 4.2434821128845215, "rewards/teacher_margin": 0.0, "step": 520 }, { "epoch": 0.31, "grad_norm": 21.5, "learning_rate": 4.367124816111351e-06, "logits/chosen": -0.3559921979904175, "logits/rejected": -0.11402563750743866, "logps/chosen": -218.14559936523438, "logps/rejected": -421.6756286621094, "loss": 0.2122, "rewards/accuracies": 0.8666666150093079, "rewards/chosen": 1.3421014547348022, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.9218908548355103, "rewards/student_margin": 3.2639918327331543, "rewards/teacher_margin": 0.0, "step": 530 }, { "epoch": 0.31, "grad_norm": 11.375, "learning_rate": 4.332958731939562e-06, "logits/chosen": -0.7146890163421631, "logits/rejected": 0.2430504560470581, "logps/chosen": -173.04917907714844, "logps/rejected": -493.38641357421875, "loss": 0.2895, "rewards/accuracies": 0.8333333134651184, "rewards/chosen": 1.2438677549362183, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.6418049335479736, "rewards/student_margin": 2.8856725692749023, "rewards/teacher_margin": 0.0, "step": 540 }, { "epoch": 0.32, "grad_norm": 29.75, "learning_rate": 4.2980357826173665e-06, "logits/chosen": -0.37195292115211487, "logits/rejected": 0.02155585028231144, "logps/chosen": -223.03817749023438, "logps/rejected": -389.7994689941406, "loss": 0.3108, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.6178417205810547, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.4387261867523193, "rewards/student_margin": 3.056567907333374, "rewards/teacher_margin": 0.0, "step": 550 }, { "epoch": 0.33, "grad_norm": 21.875, "learning_rate": 4.262370388525777e-06, "logits/chosen": -0.3869599401950836, "logits/rejected": 0.07817984372377396, "logps/chosen": -175.74595642089844, "logps/rejected": -394.0006408691406, "loss": 0.2445, "rewards/accuracies": 0.966666579246521, "rewards/chosen": 2.0001840591430664, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.1199393272399902, "rewards/student_margin": 4.120123863220215, "rewards/teacher_margin": 0.0, "step": 560 }, { "epoch": 0.33, "grad_norm": 32.25, "learning_rate": 4.225977276616028e-06, "logits/chosen": -0.5689536333084106, "logits/rejected": 0.11373758316040039, "logps/chosen": -219.41482543945312, "logps/rejected": -426.211669921875, "loss": 0.2718, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.7283222675323486, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.4601569175720215, "rewards/student_margin": 4.188479423522949, "rewards/teacher_margin": 0.0, "step": 570 }, { "epoch": 0.34, "grad_norm": 12.5, "learning_rate": 4.1888714743285305e-06, "logits/chosen": -0.3077758550643921, "logits/rejected": -0.01714705489575863, "logps/chosen": -228.143798828125, "logps/rejected": -587.8424682617188, "loss": 0.2311, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 1.6572096347808838, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -3.5316357612609863, "rewards/student_margin": 5.188845157623291, "rewards/teacher_margin": 0.0, "step": 580 }, { "epoch": 0.34, "grad_norm": 15.1875, "learning_rate": 4.151068303387739e-06, "logits/chosen": -0.442771852016449, "logits/rejected": 0.08343404531478882, "logps/chosen": -172.50587463378906, "logps/rejected": -430.36859130859375, "loss": 0.2294, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.2904380559921265, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.5538195371627808, "rewards/student_margin": 2.8442578315734863, "rewards/teacher_margin": 0.0, "step": 590 }, { "epoch": 0.35, "grad_norm": 17.125, "learning_rate": 4.1125833734755115e-06, "logits/chosen": -0.36492887139320374, "logits/rejected": 0.17705991864204407, "logps/chosen": -279.50311279296875, "logps/rejected": -664.5555419921875, "loss": 0.2703, "rewards/accuracies": 0.9000000953674316, "rewards/chosen": 1.2323994636535645, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.9321751594543457, "rewards/student_margin": 4.16457462310791, "rewards/teacher_margin": 0.0, "step": 600 }, { "epoch": 0.36, "grad_norm": 13.75, "learning_rate": 4.073432575785554e-06, "logits/chosen": -0.1946389228105545, "logits/rejected": 0.013921762816607952, "logps/chosen": -248.22073364257812, "logps/rejected": -516.2274169921875, "loss": 0.2433, "rewards/accuracies": 0.8333333134651184, "rewards/chosen": 1.737766981124878, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.9633979797363281, "rewards/student_margin": 3.7011654376983643, "rewards/teacher_margin": 0.0, "step": 610 }, { "epoch": 0.36, "grad_norm": 27.125, "learning_rate": 4.033632076461634e-06, "logits/chosen": -0.35111454129219055, "logits/rejected": 0.09419815242290497, "logps/chosen": -237.6559295654297, "logps/rejected": -470.4676818847656, "loss": 0.3378, "rewards/accuracies": 0.9333332777023315, "rewards/chosen": 1.6999142169952393, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.3078672885894775, "rewards/student_margin": 3.007781505584717, "rewards/teacher_margin": 0.0, "step": 620 }, { "epoch": 0.37, "grad_norm": 15.1875, "learning_rate": 3.9931983099222595e-06, "logits/chosen": -0.47325119376182556, "logits/rejected": -0.02582048997282982, "logps/chosen": -222.41415405273438, "logps/rejected": -465.8368225097656, "loss": 0.2413, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 1.1449840068817139, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.2412731647491455, "rewards/student_margin": 2.3862571716308594, "rewards/teacher_margin": 0.0, "step": 630 }, { "epoch": 0.37, "grad_norm": 18.875, "learning_rate": 3.9521479720745785e-06, "logits/chosen": -0.45588645339012146, "logits/rejected": 0.014990913681685925, "logps/chosen": -208.7266845703125, "logps/rejected": -565.1739501953125, "loss": 0.2011, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.211085319519043, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.549320936203003, "rewards/student_margin": 3.760406017303467, "rewards/teacher_margin": 0.0, "step": 640 }, { "epoch": 0.38, "grad_norm": 20.25, "learning_rate": 3.910498013420303e-06, "logits/chosen": -0.5036629438400269, "logits/rejected": -0.04427925869822502, "logps/chosen": -158.07492065429688, "logps/rejected": -435.2080078125, "loss": 0.2739, "rewards/accuracies": 0.800000011920929, "rewards/chosen": 1.043824315071106, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.8605940341949463, "rewards/student_margin": 2.9044182300567627, "rewards/teacher_margin": 0.0, "step": 650 }, { "epoch": 0.38, "grad_norm": 29.0, "learning_rate": 3.8682656320565205e-06, "logits/chosen": -0.41261768341064453, "logits/rejected": 0.030822629109025, "logps/chosen": -225.4918670654297, "logps/rejected": -392.59613037109375, "loss": 0.2616, "rewards/accuracies": 0.8666666746139526, "rewards/chosen": 1.8061786890029907, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.9655482769012451, "rewards/student_margin": 3.771726608276367, "rewards/teacher_margin": 0.0, "step": 660 }, { "epoch": 0.39, "grad_norm": 22.0, "learning_rate": 3.8254682665742425e-06, "logits/chosen": -0.31573912501335144, "logits/rejected": 0.005792762152850628, "logps/chosen": -240.8995361328125, "logps/rejected": -505.09735107421875, "loss": 0.2795, "rewards/accuracies": 0.8333333134651184, "rewards/chosen": 1.1914173364639282, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.8682523965835571, "rewards/student_margin": 3.0596697330474854, "rewards/teacher_margin": 0.0, "step": 670 }, { "epoch": 0.4, "grad_norm": 18.5, "learning_rate": 3.7821235888576744e-06, "logits/chosen": -0.2197073996067047, "logits/rejected": 0.02696964144706726, "logps/chosen": -244.7580108642578, "logps/rejected": -460.04400634765625, "loss": 0.1711, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.5450682640075684, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.898803472518921, "rewards/student_margin": 3.4438719749450684, "rewards/teacher_margin": 0.0, "step": 680 }, { "epoch": 0.4, "grad_norm": 9.5625, "learning_rate": 3.7382494967871384e-06, "logits/chosen": -0.4361536502838135, "logits/rejected": 0.13731250166893005, "logps/chosen": -246.27395629882812, "logps/rejected": -633.4682006835938, "loss": 0.278, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.635087251663208, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.891167640686035, "rewards/student_margin": 4.526254653930664, "rewards/teacher_margin": 0.0, "step": 690 }, { "epoch": 0.41, "grad_norm": 16.125, "learning_rate": 3.693864106848683e-06, "logits/chosen": -0.4011663794517517, "logits/rejected": 0.015058115124702454, "logps/chosen": -248.08349609375, "logps/rejected": -427.1165466308594, "loss": 0.1977, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.8699365854263306, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.281233787536621, "rewards/student_margin": 3.151170253753662, "rewards/teacher_margin": 0.0, "step": 700 }, { "epoch": 0.41, "grad_norm": 24.375, "learning_rate": 3.648985746653424e-06, "logits/chosen": -0.3061038553714752, "logits/rejected": -0.029941242188215256, "logps/chosen": -201.59161376953125, "logps/rejected": -463.8720703125, "loss": 0.1973, "rewards/accuracies": 0.8666666746139526, "rewards/chosen": 1.5523725748062134, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.9691741466522217, "rewards/student_margin": 3.521547317504883, "rewards/teacher_margin": 0.0, "step": 710 }, { "epoch": 0.42, "grad_norm": 12.875, "learning_rate": 3.6036329473697103e-06, "logits/chosen": -0.613228976726532, "logits/rejected": 0.009662175551056862, "logps/chosen": -196.20706176757812, "logps/rejected": -525.05908203125, "loss": 0.1909, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 1.3759281635284424, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.8588900566101074, "rewards/student_margin": 4.234818458557129, "rewards/teacher_margin": 0.0, "step": 720 }, { "epoch": 0.42, "grad_norm": 12.9375, "learning_rate": 3.5578244360712373e-06, "logits/chosen": -0.5177701115608215, "logits/rejected": 0.176934152841568, "logps/chosen": -206.4003143310547, "logps/rejected": -487.99249267578125, "loss": 0.2496, "rewards/accuracies": 0.9000000953674316, "rewards/chosen": 1.4539085626602173, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.346587657928467, "rewards/student_margin": 3.8004963397979736, "rewards/teacher_margin": 0.0, "step": 730 }, { "epoch": 0.43, "grad_norm": 8.375, "learning_rate": 3.511579128004261e-06, "logits/chosen": -0.44068461656570435, "logits/rejected": 0.005051338579505682, "logps/chosen": -272.7444763183594, "logps/rejected": -478.147705078125, "loss": 0.2245, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.060945987701416, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.0382800102233887, "rewards/student_margin": 3.0992257595062256, "rewards/teacher_margin": 0.0, "step": 740 }, { "epoch": 0.44, "grad_norm": 18.0, "learning_rate": 3.464916118777121e-06, "logits/chosen": -0.5848614573478699, "logits/rejected": 0.30563992261886597, "logps/chosen": -184.22329711914062, "logps/rejected": -468.75970458984375, "loss": 0.2374, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 1.795701026916504, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.5491726398468018, "rewards/student_margin": 3.3448734283447266, "rewards/teacher_margin": 0.0, "step": 750 }, { "epoch": 0.44, "grad_norm": 8.1875, "learning_rate": 3.417854676475279e-06, "logits/chosen": -0.38186392188072205, "logits/rejected": 0.059945814311504364, "logps/chosen": -260.58978271484375, "logps/rejected": -594.833251953125, "loss": 0.2078, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 1.7202898263931274, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.4396731853485107, "rewards/student_margin": 4.159963130950928, "rewards/teacher_margin": 0.0, "step": 760 }, { "epoch": 0.45, "grad_norm": 16.0, "learning_rate": 3.370414233705147e-06, "logits/chosen": -0.2833537757396698, "logits/rejected": 0.09210254997015, "logps/chosen": -223.1533203125, "logps/rejected": -582.6676025390625, "loss": 0.2096, "rewards/accuracies": 0.966666579246521, "rewards/chosen": 1.4522992372512817, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -3.2473137378692627, "rewards/student_margin": 4.699612617492676, "rewards/teacher_margin": 0.0, "step": 770 }, { "epoch": 0.45, "grad_norm": 8.375, "learning_rate": 3.322614379569975e-06, "logits/chosen": -0.302879273891449, "logits/rejected": -0.14690950512886047, "logps/chosen": -225.3035888671875, "logps/rejected": -357.20330810546875, "loss": 0.2293, "rewards/accuracies": 0.7666666507720947, "rewards/chosen": 1.1200844049453735, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.0596327781677246, "rewards/student_margin": 3.1797170639038086, "rewards/teacher_margin": 0.0, "step": 780 }, { "epoch": 0.46, "grad_norm": 19.5, "learning_rate": 3.27447485158112e-06, "logits/chosen": -0.4661865830421448, "logits/rejected": 0.12939366698265076, "logps/chosen": -226.401611328125, "logps/rejected": -421.63592529296875, "loss": 0.1703, "rewards/accuracies": 0.8666666150093079, "rewards/chosen": 1.3359860181808472, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.47092604637146, "rewards/student_margin": 3.806912660598755, "rewards/teacher_margin": 0.0, "step": 790 }, { "epoch": 0.47, "grad_norm": 24.125, "learning_rate": 3.226015527508036e-06, "logits/chosen": -0.5169921517372131, "logits/rejected": -0.004798925016075373, "logps/chosen": -241.3819580078125, "logps/rejected": -335.36492919921875, "loss": 0.2056, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.6830694675445557, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.760565996170044, "rewards/student_margin": 3.4436354637145996, "rewards/teacher_margin": 0.0, "step": 800 }, { "epoch": 0.47, "grad_norm": 19.875, "learning_rate": 3.1772564171703475e-06, "logits/chosen": -0.28334909677505493, "logits/rejected": -0.021695634350180626, "logps/chosen": -255.8955078125, "logps/rejected": -581.0257568359375, "loss": 0.3028, "rewards/accuracies": 0.8666666150093079, "rewards/chosen": 1.906981110572815, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -3.3529152870178223, "rewards/student_margin": 5.259896278381348, "rewards/teacher_margin": 0.0, "step": 810 }, { "epoch": 0.48, "grad_norm": 17.875, "learning_rate": 3.128217654175393e-06, "logits/chosen": -0.3028213381767273, "logits/rejected": 0.1735200583934784, "logps/chosen": -269.35162353515625, "logps/rejected": -464.0060119628906, "loss": 0.1942, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.3699723482131958, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.393336057662964, "rewards/student_margin": 3.7633087635040283, "rewards/teacher_margin": 0.0, "step": 820 }, { "epoch": 0.48, "grad_norm": 18.5, "learning_rate": 3.078919487604661e-06, "logits/chosen": -0.43081316351890564, "logits/rejected": -0.09694792330265045, "logps/chosen": -211.90707397460938, "logps/rejected": -447.4751892089844, "loss": 0.1913, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 1.855574369430542, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.7887474298477173, "rewards/student_margin": 3.644321918487549, "rewards/teacher_margin": 0.0, "step": 830 }, { "epoch": 0.49, "grad_norm": 12.75, "learning_rate": 3.029382273652544e-06, "logits/chosen": -0.4358023703098297, "logits/rejected": 0.1673169583082199, "logps/chosen": -241.4596710205078, "logps/rejected": -499.11956787109375, "loss": 0.1966, "rewards/accuracies": 0.8666666746139526, "rewards/chosen": 1.5806983709335327, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.8262617588043213, "rewards/student_margin": 3.4069600105285645, "rewards/teacher_margin": 0.0, "step": 840 }, { "epoch": 0.49, "grad_norm": 16.875, "learning_rate": 2.9796264672208535e-06, "logits/chosen": -0.44691354036331177, "logits/rejected": 0.14132077991962433, "logps/chosen": -217.2314453125, "logps/rejected": -476.49859619140625, "loss": 0.1794, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.7376991510391235, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.2412829399108887, "rewards/student_margin": 3.9789822101593018, "rewards/teacher_margin": 0.0, "step": 850 }, { "epoch": 0.5, "grad_norm": 12.375, "learning_rate": 2.929672613472589e-06, "logits/chosen": -0.6897366642951965, "logits/rejected": 0.006776002235710621, "logps/chosen": -216.3780517578125, "logps/rejected": -620.7020263671875, "loss": 0.1916, "rewards/accuracies": 0.8666666150093079, "rewards/chosen": 1.432846188545227, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -3.6632869243621826, "rewards/student_margin": 5.096132755279541, "rewards/teacher_margin": 0.0, "step": 860 }, { "epoch": 0.51, "grad_norm": 13.375, "learning_rate": 2.879541339348429e-06, "logits/chosen": -0.5285438895225525, "logits/rejected": 0.00928882509469986, "logps/chosen": -242.85098266601562, "logps/rejected": -481.8531188964844, "loss": 0.1813, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.5595648288726807, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.5762245655059814, "rewards/student_margin": 4.135788440704346, "rewards/teacher_margin": 0.0, "step": 870 }, { "epoch": 0.51, "grad_norm": 24.25, "learning_rate": 2.829253345049458e-06, "logits/chosen": -0.31855055689811707, "logits/rejected": 0.0941370353102684, "logps/chosen": -240.16650390625, "logps/rejected": -531.5496826171875, "loss": 0.2001, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 2.016098976135254, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.6434473991394043, "rewards/student_margin": 4.659546375274658, "rewards/teacher_margin": 0.0, "step": 880 }, { "epoch": 0.52, "grad_norm": 5.5, "learning_rate": 2.77882939548964e-06, "logits/chosen": -0.4587094187736511, "logits/rejected": 0.0055079348385334015, "logps/chosen": -223.0956573486328, "logps/rejected": -570.2337036132812, "loss": 0.2067, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 1.6644607782363892, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.6545493602752686, "rewards/student_margin": 4.3190107345581055, "rewards/teacher_margin": 0.0, "step": 890 }, { "epoch": 0.52, "grad_norm": 20.625, "learning_rate": 2.7282903117215724e-06, "logits/chosen": -0.26384082436561584, "logits/rejected": 0.112058125436306, "logps/chosen": -246.3653106689453, "logps/rejected": -487.1790466308594, "loss": 0.1951, "rewards/accuracies": 0.9333332777023315, "rewards/chosen": 1.8497905731201172, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.262843608856201, "rewards/student_margin": 4.112634658813477, "rewards/teacher_margin": 0.0, "step": 900 }, { "epoch": 0.53, "grad_norm": 17.75, "learning_rate": 2.6776569623390615e-06, "logits/chosen": -0.3648647367954254, "logits/rejected": 0.17572906613349915, "logps/chosen": -262.3744201660156, "logps/rejected": -555.7116088867188, "loss": 0.171, "rewards/accuracies": 0.8666666746139526, "rewards/chosen": 1.8194630146026611, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.9717178344726562, "rewards/student_margin": 4.791180610656738, "rewards/teacher_margin": 0.0, "step": 910 }, { "epoch": 0.54, "grad_norm": 12.9375, "learning_rate": 2.626950254860059e-06, "logits/chosen": -0.5600097179412842, "logits/rejected": 0.1498565971851349, "logps/chosen": -180.4383544921875, "logps/rejected": -550.03857421875, "loss": 0.2459, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 2.4329237937927246, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.4012227058410645, "rewards/student_margin": 4.834146499633789, "rewards/teacher_margin": 0.0, "step": 920 }, { "epoch": 0.54, "grad_norm": 15.875, "learning_rate": 2.576191127093538e-06, "logits/chosen": -0.3777901232242584, "logits/rejected": 0.046824149787425995, "logps/chosen": -214.46310424804688, "logps/rejected": -476.24627685546875, "loss": 0.1634, "rewards/accuracies": 0.8333333134651184, "rewards/chosen": 1.7608407735824585, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.415520191192627, "rewards/student_margin": 4.176361083984375, "rewards/teacher_margin": 0.0, "step": 930 }, { "epoch": 0.55, "grad_norm": 26.0, "learning_rate": 2.5254005384938495e-06, "logits/chosen": -0.49377965927124023, "logits/rejected": 0.03530919924378395, "logps/chosen": -214.46923828125, "logps/rejected": -443.4176330566406, "loss": 0.231, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 1.699275255203247, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.4718739986419678, "rewards/student_margin": 4.171149253845215, "rewards/teacher_margin": 0.0, "step": 940 }, { "epoch": 0.55, "grad_norm": 9.5, "learning_rate": 2.474599461506151e-06, "logits/chosen": -0.6275911331176758, "logits/rejected": 0.08756984770298004, "logps/chosen": -198.088623046875, "logps/rejected": -556.3299560546875, "loss": 0.1402, "rewards/accuracies": 1.0, "rewards/chosen": 1.5414060354232788, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -3.8458423614501953, "rewards/student_margin": 5.387248516082764, "rewards/teacher_margin": 0.0, "step": 950 }, { "epoch": 0.56, "grad_norm": 5.71875, "learning_rate": 2.4238088729064625e-06, "logits/chosen": -0.031927138566970825, "logits/rejected": -0.28637224435806274, "logps/chosen": -263.4320373535156, "logps/rejected": -391.6306457519531, "loss": 0.1843, "rewards/accuracies": 0.966666579246521, "rewards/chosen": 1.5371907949447632, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.255502939224243, "rewards/student_margin": 3.7926928997039795, "rewards/teacher_margin": 0.0, "step": 960 }, { "epoch": 0.56, "grad_norm": 62.25, "learning_rate": 2.3730497451399416e-06, "logits/chosen": -0.39691585302352905, "logits/rejected": 0.12687797844409943, "logps/chosen": -187.50283813476562, "logps/rejected": -645.21533203125, "loss": 0.2059, "rewards/accuracies": 1.0, "rewards/chosen": 1.4940803050994873, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -4.447799205780029, "rewards/student_margin": 5.941879749298096, "rewards/teacher_margin": 0.0, "step": 970 }, { "epoch": 0.57, "grad_norm": 17.0, "learning_rate": 2.3223430376609397e-06, "logits/chosen": -0.31980419158935547, "logits/rejected": 0.05356071516871452, "logps/chosen": -222.4664764404297, "logps/rejected": -434.98681640625, "loss": 0.2345, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 1.9460868835449219, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.792879104614258, "rewards/student_margin": 4.7389655113220215, "rewards/teacher_margin": 0.0, "step": 980 }, { "epoch": 0.58, "grad_norm": 8.9375, "learning_rate": 2.2717096882784275e-06, "logits/chosen": -0.5998212099075317, "logits/rejected": 0.12909957766532898, "logps/chosen": -176.48876953125, "logps/rejected": -488.4345703125, "loss": 0.2036, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 2.0280308723449707, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.9598499536514282, "rewards/student_margin": 3.9878811836242676, "rewards/teacher_margin": 0.0, "step": 990 }, { "epoch": 0.58, "grad_norm": 14.6875, "learning_rate": 2.2211706045103607e-06, "logits/chosen": -0.49985045194625854, "logits/rejected": 0.06370660662651062, "logps/chosen": -219.44418334960938, "logps/rejected": -586.0706787109375, "loss": 0.1749, "rewards/accuracies": 1.0, "rewards/chosen": 2.355344295501709, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.49812650680542, "rewards/student_margin": 4.853470802307129, "rewards/teacher_margin": 0.0, "step": 1000 }, { "epoch": 0.59, "grad_norm": 12.5625, "learning_rate": 2.1707466549505428e-06, "logits/chosen": -0.5008007287979126, "logits/rejected": -0.06781232357025146, "logps/chosen": -212.249267578125, "logps/rejected": -466.44964599609375, "loss": 0.1597, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 1.9009497165679932, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.340704917907715, "rewards/student_margin": 4.241654872894287, "rewards/teacher_margin": 0.0, "step": 1010 }, { "epoch": 0.59, "grad_norm": 7.78125, "learning_rate": 2.120458660651572e-06, "logits/chosen": -0.5323157906532288, "logits/rejected": -0.03192181885242462, "logps/chosen": -211.4275360107422, "logps/rejected": -644.5025024414062, "loss": 0.1576, "rewards/accuracies": 1.0, "rewards/chosen": 2.192241668701172, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -3.6768863201141357, "rewards/student_margin": 5.869128227233887, "rewards/teacher_margin": 0.0, "step": 1020 }, { "epoch": 0.6, "grad_norm": 27.375, "learning_rate": 2.070327386527412e-06, "logits/chosen": -0.3082345128059387, "logits/rejected": -0.09356869012117386, "logps/chosen": -276.2046203613281, "logps/rejected": -543.6260375976562, "loss": 0.2272, "rewards/accuracies": 0.8666666746139526, "rewards/chosen": 1.5081729888916016, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.5921823978424072, "rewards/student_margin": 4.100355625152588, "rewards/teacher_margin": 0.0, "step": 1030 }, { "epoch": 0.61, "grad_norm": 13.25, "learning_rate": 2.020373532779147e-06, "logits/chosen": -0.44038814306259155, "logits/rejected": 0.11050190776586533, "logps/chosen": -222.35556030273438, "logps/rejected": -402.10443115234375, "loss": 0.2101, "rewards/accuracies": 0.966666579246521, "rewards/chosen": 2.202882766723633, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.7409263849258423, "rewards/student_margin": 3.9438090324401855, "rewards/teacher_margin": 0.0, "step": 1040 }, { "epoch": 0.61, "grad_norm": 11.1875, "learning_rate": 1.9706177263474567e-06, "logits/chosen": -0.4345908761024475, "logits/rejected": 0.14452026784420013, "logps/chosen": -266.3000793457031, "logps/rejected": -518.9310913085938, "loss": 0.1659, "rewards/accuracies": 0.966666579246521, "rewards/chosen": 2.2457222938537598, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.8994420766830444, "rewards/student_margin": 4.145164489746094, "rewards/teacher_margin": 0.0, "step": 1050 }, { "epoch": 0.62, "grad_norm": 26.0, "learning_rate": 1.9210805123953397e-06, "logits/chosen": -0.47684893012046814, "logits/rejected": 0.010342525318264961, "logps/chosen": -221.49465942382812, "logps/rejected": -640.6463623046875, "loss": 0.1841, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.7234092950820923, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.9107723236083984, "rewards/student_margin": 4.634181499481201, "rewards/teacher_margin": 0.0, "step": 1060 }, { "epoch": 0.62, "grad_norm": 19.25, "learning_rate": 1.8717823458246082e-06, "logits/chosen": -0.4791850447654724, "logits/rejected": 0.2243192493915558, "logps/chosen": -210.71337890625, "logps/rejected": -536.1077880859375, "loss": 0.1652, "rewards/accuracies": 0.9333332777023315, "rewards/chosen": 1.8451074361801147, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.979658842086792, "rewards/student_margin": 4.824766159057617, "rewards/teacher_margin": 0.0, "step": 1070 }, { "epoch": 0.63, "grad_norm": 22.625, "learning_rate": 1.8227435828296525e-06, "logits/chosen": -0.7112604379653931, "logits/rejected": 0.23266398906707764, "logps/chosen": -200.66549682617188, "logps/rejected": -564.5877685546875, "loss": 0.1927, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 2.2678866386413574, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -3.094907283782959, "rewards/student_margin": 5.362793922424316, "rewards/teacher_margin": 0.0, "step": 1080 }, { "epoch": 0.63, "grad_norm": 17.625, "learning_rate": 1.7739844724919636e-06, "logits/chosen": -0.18364885449409485, "logits/rejected": -0.07437516748905182, "logps/chosen": -236.36068725585938, "logps/rejected": -423.6537170410156, "loss": 0.1789, "rewards/accuracies": 0.8666666150093079, "rewards/chosen": 2.1759696006774902, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.076977014541626, "rewards/student_margin": 4.252946376800537, "rewards/teacher_margin": 0.0, "step": 1090 }, { "epoch": 0.64, "grad_norm": 18.375, "learning_rate": 1.72552514841888e-06, "logits/chosen": -0.5487465858459473, "logits/rejected": -0.06100478023290634, "logps/chosen": -208.61264038085938, "logps/rejected": -503.5950622558594, "loss": 0.1366, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 1.9314453601837158, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.7049808502197266, "rewards/student_margin": 4.636425971984863, "rewards/teacher_margin": 0.0, "step": 1100 }, { "epoch": 0.65, "grad_norm": 14.125, "learning_rate": 1.6773856204300259e-06, "logits/chosen": -0.48537248373031616, "logits/rejected": 0.20526167750358582, "logps/chosen": -204.2689208984375, "logps/rejected": -438.7137145996094, "loss": 0.1781, "rewards/accuracies": 1.0, "rewards/chosen": 1.7250961065292358, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.3652234077453613, "rewards/student_margin": 4.090319633483887, "rewards/teacher_margin": 0.0, "step": 1110 }, { "epoch": 0.65, "grad_norm": 20.75, "learning_rate": 1.6295857662948535e-06, "logits/chosen": -0.41923952102661133, "logits/rejected": 0.08341416716575623, "logps/chosen": -190.0411376953125, "logps/rejected": -548.9283447265625, "loss": 0.19, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.4958313703536987, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.9623820781707764, "rewards/student_margin": 4.458213806152344, "rewards/teacher_margin": 0.0, "step": 1120 }, { "epoch": 0.66, "grad_norm": 20.75, "learning_rate": 1.5821453235247217e-06, "logits/chosen": -0.053016532212495804, "logits/rejected": 0.08401618152856827, "logps/chosen": -312.3536682128906, "logps/rejected": -552.4027099609375, "loss": 0.198, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 1.9978468418121338, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.899606227874756, "rewards/student_margin": 4.897453784942627, "rewards/teacher_margin": 0.0, "step": 1130 }, { "epoch": 0.66, "grad_norm": 11.375, "learning_rate": 1.5350838812228796e-06, "logits/chosen": -0.5477980375289917, "logits/rejected": 0.18308475613594055, "logps/chosen": -194.05624389648438, "logps/rejected": -481.322265625, "loss": 0.1794, "rewards/accuracies": 0.8666666150093079, "rewards/chosen": 2.268648624420166, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.6922143697738647, "rewards/student_margin": 3.9608635902404785, "rewards/teacher_margin": 0.0, "step": 1140 }, { "epoch": 0.67, "grad_norm": 11.5625, "learning_rate": 1.4884208719957396e-06, "logits/chosen": -0.4524906277656555, "logits/rejected": 0.048188693821430206, "logps/chosen": -240.8212890625, "logps/rejected": -457.342041015625, "loss": 0.1424, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 2.082411527633667, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.1976187229156494, "rewards/student_margin": 4.280030727386475, "rewards/teacher_margin": 0.0, "step": 1150 }, { "epoch": 0.68, "grad_norm": 23.125, "learning_rate": 1.4421755639287633e-06, "logits/chosen": -0.735596239566803, "logits/rejected": 0.0022054375149309635, "logps/chosen": -182.4052734375, "logps/rejected": -511.3103942871094, "loss": 0.1992, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 2.068901538848877, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.3740670680999756, "rewards/student_margin": 4.442968845367432, "rewards/teacher_margin": 0.0, "step": 1160 }, { "epoch": 0.68, "grad_norm": 7.28125, "learning_rate": 1.39636705263029e-06, "logits/chosen": -0.09597094357013702, "logits/rejected": -0.051341939717531204, "logps/chosen": -257.4917297363281, "logps/rejected": -411.2127990722656, "loss": 0.1143, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 2.1275782585144043, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.4862334728240967, "rewards/student_margin": 3.613811492919922, "rewards/teacher_margin": 0.0, "step": 1170 }, { "epoch": 0.69, "grad_norm": 4.875, "learning_rate": 1.3510142533465764e-06, "logits/chosen": -0.35563239455223083, "logits/rejected": 0.13452188670635223, "logps/chosen": -267.82806396484375, "logps/rejected": -522.6319580078125, "loss": 0.1118, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 2.2255101203918457, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.5478758811950684, "rewards/student_margin": 4.773386478424072, "rewards/teacher_margin": 0.0, "step": 1180 }, { "epoch": 0.69, "grad_norm": 28.5, "learning_rate": 1.306135893151318e-06, "logits/chosen": -0.37067538499832153, "logits/rejected": 0.1407340168952942, "logps/chosen": -292.832763671875, "logps/rejected": -464.6031799316406, "loss": 0.2175, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 2.117187976837158, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.6492514610290527, "rewards/student_margin": 4.766438961029053, "rewards/teacher_margin": 0.0, "step": 1190 }, { "epoch": 0.7, "grad_norm": 11.9375, "learning_rate": 1.2617505032128618e-06, "logits/chosen": -0.2240295708179474, "logits/rejected": 0.03423571586608887, "logps/chosen": -241.01223754882812, "logps/rejected": -511.5538635253906, "loss": 0.1503, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 1.904809594154358, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.1720213890075684, "rewards/student_margin": 4.076830863952637, "rewards/teacher_margin": 0.0, "step": 1200 }, { "epoch": 0.7, "grad_norm": 22.5, "learning_rate": 1.2178764111423258e-06, "logits/chosen": -0.453156054019928, "logits/rejected": -0.015243321657180786, "logps/chosen": -261.2116394042969, "logps/rejected": -541.9698486328125, "loss": 0.1609, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 2.2009761333465576, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.577160120010376, "rewards/student_margin": 4.778136253356934, "rewards/teacher_margin": 0.0, "step": 1210 }, { "epoch": 0.71, "grad_norm": 11.625, "learning_rate": 1.1745317334257577e-06, "logits/chosen": -0.3408849537372589, "logits/rejected": 0.10163626819849014, "logps/chosen": -180.9221649169922, "logps/rejected": -456.72412109375, "loss": 0.1747, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.4770768880844116, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.7443153858184814, "rewards/student_margin": 4.221392631530762, "rewards/teacher_margin": 0.0, "step": 1220 }, { "epoch": 0.72, "grad_norm": 18.0, "learning_rate": 1.1317343679434797e-06, "logits/chosen": -0.39155808091163635, "logits/rejected": 0.03318040445446968, "logps/chosen": -243.98617553710938, "logps/rejected": -480.6688537597656, "loss": 0.1238, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 1.888183355331421, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.6680731773376465, "rewards/student_margin": 4.556256294250488, "rewards/teacher_margin": 0.0, "step": 1230 }, { "epoch": 0.72, "grad_norm": 10.1875, "learning_rate": 1.0895019865796974e-06, "logits/chosen": -0.35370463132858276, "logits/rejected": -0.06857095658779144, "logps/chosen": -214.5222930908203, "logps/rejected": -411.405517578125, "loss": 0.2088, "rewards/accuracies": 0.8666666150093079, "rewards/chosen": 1.605942726135254, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.65793776512146, "rewards/student_margin": 3.263880968093872, "rewards/teacher_margin": 0.0, "step": 1240 }, { "epoch": 0.73, "grad_norm": 7.96875, "learning_rate": 1.0478520279254232e-06, "logits/chosen": -0.4441911578178406, "logits/rejected": 0.0003550305846147239, "logps/chosen": -173.74652099609375, "logps/rejected": -404.5919494628906, "loss": 0.1417, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.5375311374664307, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.0653061866760254, "rewards/student_margin": 3.602837324142456, "rewards/teacher_margin": 0.0, "step": 1250 }, { "epoch": 0.73, "grad_norm": 11.0, "learning_rate": 1.0068016900777411e-06, "logits/chosen": -0.14587707817554474, "logits/rejected": 0.1403418779373169, "logps/chosen": -265.76385498046875, "logps/rejected": -488.88360595703125, "loss": 0.1327, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 1.8325278759002686, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.28810453414917, "rewards/student_margin": 4.120633125305176, "rewards/teacher_margin": 0.0, "step": 1260 }, { "epoch": 0.74, "grad_norm": 15.875, "learning_rate": 9.663679235383663e-07, "logits/chosen": -0.4032668173313141, "logits/rejected": 0.15200474858283997, "logps/chosen": -253.056396484375, "logps/rejected": -567.1189575195312, "loss": 0.1393, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 1.7767219543457031, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.897172451019287, "rewards/student_margin": 4.67389440536499, "rewards/teacher_margin": 0.0, "step": 1270 }, { "epoch": 0.75, "grad_norm": 13.25, "learning_rate": 9.265674242144462e-07, "logits/chosen": -0.5168894529342651, "logits/rejected": 0.1268787384033203, "logps/chosen": -241.34536743164062, "logps/rejected": -535.0253295898438, "loss": 0.167, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 2.2057933807373047, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.5597376823425293, "rewards/student_margin": 4.765531539916992, "rewards/teacher_margin": 0.0, "step": 1280 }, { "epoch": 0.75, "grad_norm": 11.375, "learning_rate": 8.874166265244891e-07, "logits/chosen": -0.2563822865486145, "logits/rejected": -0.007131014950573444, "logps/chosen": -241.32510375976562, "logps/rejected": -438.27783203125, "loss": 0.1496, "rewards/accuracies": 0.966666579246521, "rewards/chosen": 2.1465556621551514, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.53509783744812, "rewards/student_margin": 4.681653022766113, "rewards/teacher_margin": 0.0, "step": 1290 }, { "epoch": 0.76, "grad_norm": 5.71875, "learning_rate": 8.48931696612261e-07, "logits/chosen": -0.03185834363102913, "logits/rejected": 0.0335637666285038, "logps/chosen": -292.47320556640625, "logps/rejected": -396.5122375488281, "loss": 0.1623, "rewards/accuracies": 1.0, "rewards/chosen": 2.144620656967163, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.0947091579437256, "rewards/student_margin": 4.239329814910889, "rewards/teacher_margin": 0.0, "step": 1300 }, { "epoch": 0.76, "grad_norm": 11.1875, "learning_rate": 8.111285256714702e-07, "logits/chosen": -0.5409881472587585, "logits/rejected": -0.020885199308395386, "logps/chosen": -259.0535583496094, "logps/rejected": -619.1828002929688, "loss": 0.1752, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 2.047285795211792, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -3.6237800121307373, "rewards/student_margin": 5.6710662841796875, "rewards/teacher_margin": 0.0, "step": 1310 }, { "epoch": 0.77, "grad_norm": 11.3125, "learning_rate": 7.740227233839726e-07, "logits/chosen": -0.23395772278308868, "logits/rejected": 0.0602710135281086, "logps/chosen": -236.9947967529297, "logps/rejected": -399.87115478515625, "loss": 0.1768, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 1.6983884572982788, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.7023701667785645, "rewards/student_margin": 4.400758743286133, "rewards/teacher_margin": 0.0, "step": 1320 }, { "epoch": 0.77, "grad_norm": 8.5, "learning_rate": 7.376296114742231e-07, "logits/chosen": -0.5401469469070435, "logits/rejected": 0.14081577956676483, "logps/chosen": -179.28860473632812, "logps/rejected": -485.43218994140625, "loss": 0.1542, "rewards/accuracies": 1.0, "rewards/chosen": 2.387885808944702, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.904003620147705, "rewards/student_margin": 5.291889190673828, "rewards/teacher_margin": 0.0, "step": 1330 }, { "epoch": 0.78, "grad_norm": 8.3125, "learning_rate": 7.019642173826335e-07, "logits/chosen": -0.42929285764694214, "logits/rejected": 0.14102640748023987, "logps/chosen": -236.9450225830078, "logps/rejected": -510.2264099121094, "loss": 0.1831, "rewards/accuracies": 0.966666579246521, "rewards/chosen": 2.448880672454834, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.9888415336608887, "rewards/student_margin": 5.437722206115723, "rewards/teacher_margin": 0.0, "step": 1340 }, { "epoch": 0.79, "grad_norm": 12.125, "learning_rate": 6.670412680604379e-07, "logits/chosen": -0.25430089235305786, "logits/rejected": 0.12993714213371277, "logps/chosen": -257.48284912109375, "logps/rejected": -497.19110107421875, "loss": 0.1467, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 2.0826573371887207, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.528546094894409, "rewards/student_margin": 4.611203193664551, "rewards/teacher_margin": 0.0, "step": 1350 }, { "epoch": 0.79, "grad_norm": 22.875, "learning_rate": 6.328751838886496e-07, "logits/chosen": -0.2653670907020569, "logits/rejected": 0.04784447327256203, "logps/chosen": -231.7043914794922, "logps/rejected": -405.35430908203125, "loss": 0.1555, "rewards/accuracies": 1.0, "rewards/chosen": 2.0900282859802246, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.961517572402954, "rewards/student_margin": 4.0515456199646, "rewards/teacher_margin": 0.0, "step": 1360 }, { "epoch": 0.8, "grad_norm": 12.0625, "learning_rate": 5.994800727235894e-07, "logits/chosen": -0.6459251642227173, "logits/rejected": -0.028898591175675392, "logps/chosen": -167.8853759765625, "logps/rejected": -490.18621826171875, "loss": 0.1421, "rewards/accuracies": 1.0, "rewards/chosen": 1.746824026107788, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.833235502243042, "rewards/student_margin": 4.580060005187988, "rewards/teacher_margin": 0.0, "step": 1370 }, { "epoch": 0.8, "grad_norm": 11.4375, "learning_rate": 5.668697240714782e-07, "logits/chosen": -0.736299991607666, "logits/rejected": -0.027988219633698463, "logps/chosen": -166.20730590820312, "logps/rejected": -544.7757568359375, "loss": 0.1785, "rewards/accuracies": 1.0, "rewards/chosen": 2.0296387672424316, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.3135805130004883, "rewards/student_margin": 4.34321928024292, "rewards/teacher_margin": 0.0, "step": 1380 }, { "epoch": 0.81, "grad_norm": 26.0, "learning_rate": 5.350576033944705e-07, "logits/chosen": -0.4334324896335602, "logits/rejected": 0.09274639934301376, "logps/chosen": -197.7583465576172, "logps/rejected": -443.3536071777344, "loss": 0.1698, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 2.787088632583618, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.1186861991882324, "rewards/student_margin": 4.9057745933532715, "rewards/teacher_margin": 0.0, "step": 1390 }, { "epoch": 0.81, "grad_norm": 12.5, "learning_rate": 5.040568465504991e-07, "logits/chosen": -0.23930701613426208, "logits/rejected": 0.13585281372070312, "logps/chosen": -247.615478515625, "logps/rejected": -547.8809814453125, "loss": 0.2305, "rewards/accuracies": 0.966666579246521, "rewards/chosen": 2.4914164543151855, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.0706675052642822, "rewards/student_margin": 4.562083721160889, "rewards/teacher_margin": 0.0, "step": 1400 }, { "epoch": 0.82, "grad_norm": 6.71875, "learning_rate": 4.738802543692214e-07, "logits/chosen": -0.33991652727127075, "logits/rejected": 0.23731207847595215, "logps/chosen": -219.88528442382812, "logps/rejected": -498.65618896484375, "loss": 0.1918, "rewards/accuracies": 0.966666579246521, "rewards/chosen": 2.3377609252929688, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.0855703353881836, "rewards/student_margin": 4.423331260681152, "rewards/teacher_margin": 0.0, "step": 1410 }, { "epoch": 0.83, "grad_norm": 17.625, "learning_rate": 4.445402873662932e-07, "logits/chosen": -0.33060532808303833, "logits/rejected": 0.07881384342908859, "logps/chosen": -235.1278839111328, "logps/rejected": -671.2401123046875, "loss": 0.1594, "rewards/accuracies": 0.8666666746139526, "rewards/chosen": 1.9622802734375, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.477834701538086, "rewards/student_margin": 4.440115451812744, "rewards/teacher_margin": 0.0, "step": 1420 }, { "epoch": 0.83, "grad_norm": 18.125, "learning_rate": 4.1604906059818234e-07, "logits/chosen": -0.5972268581390381, "logits/rejected": -0.00915351789444685, "logps/chosen": -217.2305450439453, "logps/rejected": -514.9454956054688, "loss": 0.1546, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 1.6480636596679688, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.5352678298950195, "rewards/student_margin": 4.183331489562988, "rewards/teacher_margin": 0.0, "step": 1430 }, { "epoch": 0.84, "grad_norm": 9.0625, "learning_rate": 3.8841833865961435e-07, "logits/chosen": -0.19409717619419098, "logits/rejected": 0.05707187205553055, "logps/chosen": -290.650390625, "logps/rejected": -460.62548828125, "loss": 0.1939, "rewards/accuracies": 0.9333332777023315, "rewards/chosen": 1.422797441482544, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.440934896469116, "rewards/student_margin": 3.8637325763702393, "rewards/teacher_margin": 0.0, "step": 1440 }, { "epoch": 0.84, "grad_norm": 11.5, "learning_rate": 3.6165953082574453e-07, "logits/chosen": -0.34082725644111633, "logits/rejected": 0.08756278455257416, "logps/chosen": -258.0528259277344, "logps/rejected": -627.5899047851562, "loss": 0.1415, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 1.9923893213272095, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.4780945777893066, "rewards/student_margin": 4.470483303070068, "rewards/teacher_margin": 0.0, "step": 1450 }, { "epoch": 0.85, "grad_norm": 14.25, "learning_rate": 3.3578368634103133e-07, "logits/chosen": -0.5575006008148193, "logits/rejected": 0.1742563545703888, "logps/chosen": -224.60537719726562, "logps/rejected": -587.3863525390625, "loss": 0.1389, "rewards/accuracies": 0.8666666746139526, "rewards/chosen": 1.9285333156585693, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -3.4504916667938232, "rewards/student_margin": 5.379024982452393, "rewards/teacher_margin": 0.0, "step": 1460 }, { "epoch": 0.86, "grad_norm": 15.1875, "learning_rate": 3.108014898567921e-07, "logits/chosen": -0.32235223054885864, "logits/rejected": 0.23691675066947937, "logps/chosen": -289.0404052734375, "logps/rejected": -518.9627685546875, "loss": 0.1884, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 1.5728108882904053, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.671419143676758, "rewards/student_margin": 4.244229793548584, "rewards/teacher_margin": 0.0, "step": 1470 }, { "epoch": 0.86, "grad_norm": 12.0, "learning_rate": 2.867232570192932e-07, "logits/chosen": -0.4675559103488922, "logits/rejected": 0.2386462688446045, "logps/chosen": -238.5657501220703, "logps/rejected": -556.7369384765625, "loss": 0.172, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 2.0571231842041016, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.1405348777770996, "rewards/student_margin": 4.197658061981201, "rewards/teacher_margin": 0.0, "step": 1480 }, { "epoch": 0.87, "grad_norm": 9.6875, "learning_rate": 2.635589302102143e-07, "logits/chosen": -0.5631746053695679, "logits/rejected": 0.1842821091413498, "logps/chosen": -202.35910034179688, "logps/rejected": -496.7588806152344, "loss": 0.1308, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 1.9014427661895752, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.115988254547119, "rewards/student_margin": 4.017430782318115, "rewards/teacher_margin": 0.0, "step": 1490 }, { "epoch": 0.87, "grad_norm": 12.875, "learning_rate": 2.413180744412449e-07, "logits/chosen": -0.45950451493263245, "logits/rejected": -0.005202120635658503, "logps/chosen": -199.29319763183594, "logps/rejected": -393.0441589355469, "loss": 0.1418, "rewards/accuracies": 1.0, "rewards/chosen": 2.295112371444702, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.899126648902893, "rewards/student_margin": 4.194239139556885, "rewards/teacher_margin": 0.0, "step": 1500 }, { "epoch": 0.88, "grad_norm": 7.25, "learning_rate": 2.200098734044953e-07, "logits/chosen": -0.3245978355407715, "logits/rejected": 0.15231186151504517, "logps/chosen": -282.319091796875, "logps/rejected": -564.8612060546875, "loss": 0.1391, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 1.9999994039535522, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.712496757507324, "rewards/student_margin": 4.712496757507324, "rewards/teacher_margin": 0.0, "step": 1510 }, { "epoch": 0.88, "grad_norm": 20.5, "learning_rate": 1.9964312568037102e-07, "logits/chosen": -0.43997129797935486, "logits/rejected": 0.11200802028179169, "logps/chosen": -215.91390991210938, "logps/rejected": -454.8565979003906, "loss": 0.2001, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.7922544479370117, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.544246196746826, "rewards/student_margin": 4.336500644683838, "rewards/teacher_margin": 0.0, "step": 1520 }, { "epoch": 0.89, "grad_norm": 7.9375, "learning_rate": 1.8022624110446223e-07, "logits/chosen": -0.23303954303264618, "logits/rejected": -0.25525903701782227, "logps/chosen": -233.7492218017578, "logps/rejected": -434.97943115234375, "loss": 0.1737, "rewards/accuracies": 0.9000000953674316, "rewards/chosen": 2.0419070720672607, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.659355878829956, "rewards/student_margin": 3.701263427734375, "rewards/teacher_margin": 0.0, "step": 1530 }, { "epoch": 0.9, "grad_norm": 11.3125, "learning_rate": 1.617672372949558e-07, "logits/chosen": -0.4891526699066162, "logits/rejected": 0.0062074740417301655, "logps/chosen": -184.78028869628906, "logps/rejected": -495.7847595214844, "loss": 0.1925, "rewards/accuracies": 0.8666666150093079, "rewards/chosen": 1.5070412158966064, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.09743070602417, "rewards/student_margin": 3.6044719219207764, "rewards/teacher_margin": 0.0, "step": 1540 }, { "epoch": 0.9, "grad_norm": 8.8125, "learning_rate": 1.442737363420016e-07, "logits/chosen": -0.47575002908706665, "logits/rejected": 0.09344211965799332, "logps/chosen": -215.34390258789062, "logps/rejected": -447.3370056152344, "loss": 0.1637, "rewards/accuracies": 0.9333332777023315, "rewards/chosen": 1.5486562252044678, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.476374864578247, "rewards/student_margin": 4.025030612945557, "rewards/teacher_margin": 0.0, "step": 1550 }, { "epoch": 0.91, "grad_norm": 5.09375, "learning_rate": 1.2775296166039759e-07, "logits/chosen": -0.34430769085884094, "logits/rejected": 0.11993454396724701, "logps/chosen": -198.77247619628906, "logps/rejected": -414.20648193359375, "loss": 0.1491, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 2.136995553970337, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.7208058834075928, "rewards/student_margin": 3.8578014373779297, "rewards/teacher_margin": 0.0, "step": 1560 }, { "epoch": 0.91, "grad_norm": 9.125, "learning_rate": 1.1221173500690074e-07, "logits/chosen": -0.43167194724082947, "logits/rejected": 0.05056245997548103, "logps/chosen": -202.82408142089844, "logps/rejected": -457.1270446777344, "loss": 0.1328, "rewards/accuracies": 0.966666579246521, "rewards/chosen": 1.7842384576797485, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -1.8982183933258057, "rewards/student_margin": 3.6824564933776855, "rewards/teacher_margin": 0.0, "step": 1570 }, { "epoch": 0.92, "grad_norm": 13.4375, "learning_rate": 9.765647366338615e-08, "logits/chosen": -0.617282509803772, "logits/rejected": 0.20323173701763153, "logps/chosen": -215.5322723388672, "logps/rejected": -516.0162353515625, "loss": 0.1082, "rewards/accuracies": 1.0, "rewards/chosen": 2.109098196029663, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.6933035850524902, "rewards/student_margin": 4.802401542663574, "rewards/teacher_margin": 0.0, "step": 1580 }, { "epoch": 0.93, "grad_norm": 13.1875, "learning_rate": 8.409318778702436e-08, "logits/chosen": -0.24509771168231964, "logits/rejected": 0.05450732633471489, "logps/chosen": -244.3855438232422, "logps/rejected": -399.8455505371094, "loss": 0.1269, "rewards/accuracies": 0.966666579246521, "rewards/chosen": 2.09488582611084, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.1163201332092285, "rewards/student_margin": 4.21120548248291, "rewards/teacher_margin": 0.0, "step": 1590 }, { "epoch": 0.93, "grad_norm": 12.625, "learning_rate": 7.152747792856795e-08, "logits/chosen": -0.5267202854156494, "logits/rejected": 0.10546509921550751, "logps/chosen": -246.46163940429688, "logps/rejected": -488.86358642578125, "loss": 0.1929, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 2.5057289600372314, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.4191794395446777, "rewards/student_margin": 4.924908638000488, "rewards/teacher_margin": 0.0, "step": 1600 }, { "epoch": 0.94, "grad_norm": 8.8125, "learning_rate": 5.99645327197712e-08, "logits/chosen": -0.5035473108291626, "logits/rejected": 0.06472502648830414, "logps/chosen": -175.3248748779297, "logps/rejected": -475.020263671875, "loss": 0.1096, "rewards/accuracies": 0.9333332777023315, "rewards/chosen": 2.2377569675445557, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.6387767791748047, "rewards/student_margin": 4.8765339851379395, "rewards/teacher_margin": 0.0, "step": 1610 }, { "epoch": 0.94, "grad_norm": 14.625, "learning_rate": 4.94091267309027e-08, "logits/chosen": -0.5968758463859558, "logits/rejected": 0.08792127668857574, "logps/chosen": -183.85728454589844, "logps/rejected": -632.9024658203125, "loss": 0.1953, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 2.25777530670166, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.826479434967041, "rewards/student_margin": 5.084254741668701, "rewards/teacher_margin": 0.0, "step": 1620 }, { "epoch": 0.95, "grad_norm": 14.125, "learning_rate": 3.9865618499229094e-08, "logits/chosen": -0.41523703932762146, "logits/rejected": -0.0010458737378939986, "logps/chosen": -246.58242797851562, "logps/rejected": -524.6209716796875, "loss": 0.1439, "rewards/accuracies": 0.9333332777023315, "rewards/chosen": 1.9719102382659912, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.955763339996338, "rewards/student_margin": 4.927673816680908, "rewards/teacher_margin": 0.0, "step": 1630 }, { "epoch": 0.95, "grad_norm": 7.59375, "learning_rate": 3.1337948729288633e-08, "logits/chosen": -0.45982232689857483, "logits/rejected": 0.08918474614620209, "logps/chosen": -246.1114501953125, "logps/rejected": -527.5103149414062, "loss": 0.1411, "rewards/accuracies": 1.0, "rewards/chosen": 2.207768440246582, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.371577501296997, "rewards/student_margin": 4.579345703125, "rewards/teacher_margin": 0.0, "step": 1640 }, { "epoch": 0.96, "grad_norm": 12.5625, "learning_rate": 2.3829638665695544e-08, "logits/chosen": -0.5832090377807617, "logits/rejected": 0.012857144698500633, "logps/chosen": -177.62669372558594, "logps/rejected": -534.47314453125, "loss": 0.1282, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 1.428046703338623, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.5331838130950928, "rewards/student_margin": 3.961230516433716, "rewards/teacher_margin": 0.0, "step": 1650 }, { "epoch": 0.97, "grad_norm": 9.6875, "learning_rate": 1.734378863914782e-08, "logits/chosen": -0.4249907433986664, "logits/rejected": 0.0664374977350235, "logps/chosen": -245.0904083251953, "logps/rejected": -566.2081909179688, "loss": 0.1449, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 2.0259430408477783, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.927861452102661, "rewards/student_margin": 4.953804969787598, "rewards/teacher_margin": 0.0, "step": 1660 }, { "epoch": 0.97, "grad_norm": 15.75, "learning_rate": 1.1883076786240177e-08, "logits/chosen": -0.5717610716819763, "logits/rejected": 0.050547920167446136, "logps/chosen": -175.99942016601562, "logps/rejected": -478.18524169921875, "loss": 0.1616, "rewards/accuracies": 0.9333333969116211, "rewards/chosen": 1.8787494897842407, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.2613539695739746, "rewards/student_margin": 4.140103340148926, "rewards/teacher_margin": 0.0, "step": 1670 }, { "epoch": 0.98, "grad_norm": 25.25, "learning_rate": 7.449757943606972e-09, "logits/chosen": -0.5395152568817139, "logits/rejected": 0.1687006652355194, "logps/chosen": -226.3690643310547, "logps/rejected": -453.1436462402344, "loss": 0.1588, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.7440685033798218, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.6785857677459717, "rewards/student_margin": 4.422654151916504, "rewards/teacher_margin": 0.0, "step": 1680 }, { "epoch": 0.98, "grad_norm": 16.0, "learning_rate": 4.045662716855303e-09, "logits/chosen": -0.2058514803647995, "logits/rejected": 0.07757274806499481, "logps/chosen": -237.3645477294922, "logps/rejected": -580.2776489257812, "loss": 0.1531, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 1.9641157388687134, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.1610915660858154, "rewards/student_margin": 4.12520694732666, "rewards/teacher_margin": 0.0, "step": 1690 }, { "epoch": 0.99, "grad_norm": 12.625, "learning_rate": 1.6721967246721704e-09, "logits/chosen": -0.36075156927108765, "logits/rejected": 0.044158972799777985, "logps/chosen": -210.068359375, "logps/rejected": -480.38726806640625, "loss": 0.1368, "rewards/accuracies": 0.9333332777023315, "rewards/chosen": 1.983689308166504, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.8466620445251465, "rewards/student_margin": 4.830351829528809, "rewards/teacher_margin": 0.0, "step": 1700 }, { "epoch": 1.0, "grad_norm": 10.6875, "learning_rate": 3.30340018415698e-10, "logits/chosen": -0.4800366461277008, "logits/rejected": 0.04889531061053276, "logps/chosen": -193.82672119140625, "logps/rejected": -558.3578491210938, "loss": 0.1185, "rewards/accuracies": 0.9666666984558105, "rewards/chosen": 2.046447515487671, "rewards/diff": 0.0, "rewards/diff_abs": 0.0, "rewards/rejected": -2.4438791275024414, "rewards/student_margin": 4.490326881408691, "rewards/teacher_margin": 0.0, "step": 1710 }, { "epoch": 1.0, "step": 1718, "total_flos": 0.0, "train_loss": 0.24084537111144683, "train_runtime": 1647.7299, "train_samples_per_second": 25.016, "train_steps_per_second": 1.043 } ], "logging_steps": 10, "max_steps": 1718, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 400, "total_flos": 0.0, "train_batch_size": 3, "trial_name": null, "trial_params": null }