diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,3126 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0, + "eval_steps": 80000000000, + "global_step": 1718, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0, + "grad_norm": 47.75, + "learning_rate": 2.9069767441860468e-08, + "logits/chosen": -0.30066978931427, + "logits/rejected": 0.05270648002624512, + "logps/chosen": -292.3870544433594, + "logps/rejected": -519.2880859375, + "loss": 0.6931, + "rewards/accuracies": 0.0, + "rewards/chosen": 0.0, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": 0.0, + "rewards/student_margin": 0.0, + "rewards/teacher_margin": 0.0, + "step": 1 + }, + { + "epoch": 0.01, + "grad_norm": 50.75, + "learning_rate": 2.906976744186047e-07, + "logits/chosen": -0.37918758392333984, + "logits/rejected": -0.042655572295188904, + "logps/chosen": -231.2487030029297, + "logps/rejected": -516.9002685546875, + "loss": 0.726, + "rewards/accuracies": 0.3333333432674408, + "rewards/chosen": 0.05387793853878975, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": 0.07821387052536011, + "rewards/student_margin": -0.024335933849215508, + "rewards/teacher_margin": 0.0, + "step": 10 + }, + { + "epoch": 0.01, + "grad_norm": 42.75, + "learning_rate": 5.813953488372094e-07, + "logits/chosen": -0.7622300386428833, + "logits/rejected": 0.022924324497580528, + "logps/chosen": -245.8932647705078, + "logps/rejected": -559.360595703125, + "loss": 0.7117, + "rewards/accuracies": 0.4333333373069763, + "rewards/chosen": 0.04971719905734062, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -0.02675577439367771, + "rewards/student_margin": 0.07647298276424408, + "rewards/teacher_margin": 0.0, + "step": 20 + }, + { + "epoch": 0.02, + "grad_norm": 34.25, + "learning_rate": 8.720930232558141e-07, + "logits/chosen": -0.6480653285980225, + "logits/rejected": -0.05629728361964226, + "logps/chosen": -207.52798461914062, + "logps/rejected": -569.8907470703125, + "loss": 0.6704, + "rewards/accuracies": 0.6666666865348816, + "rewards/chosen": -0.04712574928998947, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -0.31786206364631653, + "rewards/student_margin": 0.27073630690574646, + "rewards/teacher_margin": 0.0, + "step": 30 + }, + { + "epoch": 0.02, + "grad_norm": 33.25, + "learning_rate": 1.1627906976744188e-06, + "logits/chosen": -0.46851086616516113, + "logits/rejected": 0.08061075955629349, + "logps/chosen": -213.6239013671875, + "logps/rejected": -531.3410034179688, + "loss": 0.5929, + "rewards/accuracies": 0.7666666507720947, + "rewards/chosen": -0.09198583662509918, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -0.5451609492301941, + "rewards/student_margin": 0.4531751275062561, + "rewards/teacher_margin": 0.0, + "step": 40 + }, + { + "epoch": 0.03, + "grad_norm": 28.125, + "learning_rate": 1.4534883720930234e-06, + "logits/chosen": -0.6713495850563049, + "logits/rejected": -0.03319794684648514, + "logps/chosen": -178.42832946777344, + "logps/rejected": -427.0347595214844, + "loss": 0.5249, + "rewards/accuracies": 0.5666666626930237, + "rewards/chosen": -0.16660943627357483, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -0.6883589029312134, + "rewards/student_margin": 0.5217493772506714, + "rewards/teacher_margin": 0.0, + "step": 50 + }, + { + "epoch": 0.03, + "grad_norm": 23.625, + "learning_rate": 1.7441860465116282e-06, + "logits/chosen": -0.8985533714294434, + "logits/rejected": 0.039773982018232346, + "logps/chosen": -175.9773406982422, + "logps/rejected": -535.4810791015625, + "loss": 0.4963, + "rewards/accuracies": 0.800000011920929, + "rewards/chosen": -0.20179271697998047, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.35043203830719, + "rewards/student_margin": 1.148639440536499, + "rewards/teacher_margin": 0.0, + "step": 60 + }, + { + "epoch": 0.04, + "grad_norm": 27.875, + "learning_rate": 2.0348837209302328e-06, + "logits/chosen": -0.5139203071594238, + "logits/rejected": 0.02224038355052471, + "logps/chosen": -244.5893096923828, + "logps/rejected": -573.5228271484375, + "loss": 0.4526, + "rewards/accuracies": 0.7999999523162842, + "rewards/chosen": -0.348739355802536, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.8875904083251953, + "rewards/student_margin": 1.5388509035110474, + "rewards/teacher_margin": 0.0, + "step": 70 + }, + { + "epoch": 0.05, + "grad_norm": 18.875, + "learning_rate": 2.3255813953488376e-06, + "logits/chosen": -0.7934290170669556, + "logits/rejected": -0.0012261777883395553, + "logps/chosen": -182.20286560058594, + "logps/rejected": -427.95037841796875, + "loss": 0.4462, + "rewards/accuracies": 0.7333333492279053, + "rewards/chosen": -0.0741361603140831, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.3484480381011963, + "rewards/student_margin": 1.2743120193481445, + "rewards/teacher_margin": 0.0, + "step": 80 + }, + { + "epoch": 0.05, + "grad_norm": 26.25, + "learning_rate": 2.6162790697674424e-06, + "logits/chosen": -0.7236814498901367, + "logits/rejected": 0.027469992637634277, + "logps/chosen": -186.38136291503906, + "logps/rejected": -399.2889099121094, + "loss": 0.4499, + "rewards/accuracies": 0.7333332896232605, + "rewards/chosen": -0.031203720718622208, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.2050061225891113, + "rewards/student_margin": 1.1738024950027466, + "rewards/teacher_margin": 0.0, + "step": 90 + }, + { + "epoch": 0.06, + "grad_norm": 30.0, + "learning_rate": 2.9069767441860468e-06, + "logits/chosen": -0.4575803279876709, + "logits/rejected": -0.1135629191994667, + "logps/chosen": -259.8243713378906, + "logps/rejected": -481.0874938964844, + "loss": 0.4093, + "rewards/accuracies": 0.8666666746139526, + "rewards/chosen": -0.21589164435863495, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.5411407947540283, + "rewards/student_margin": 1.3252493143081665, + "rewards/teacher_margin": 0.0, + "step": 100 + }, + { + "epoch": 0.06, + "grad_norm": 20.5, + "learning_rate": 3.1976744186046516e-06, + "logits/chosen": -0.3742029070854187, + "logits/rejected": -0.12473519891500473, + "logps/chosen": -186.20205688476562, + "logps/rejected": -392.6462097167969, + "loss": 0.3743, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 0.20217347145080566, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.5840258598327637, + "rewards/student_margin": 1.7861993312835693, + "rewards/teacher_margin": 0.0, + "step": 110 + }, + { + "epoch": 0.07, + "grad_norm": 19.125, + "learning_rate": 3.4883720930232564e-06, + "logits/chosen": -0.6747792959213257, + "logits/rejected": 0.16625359654426575, + "logps/chosen": -229.23672485351562, + "logps/rejected": -525.2789306640625, + "loss": 0.3796, + "rewards/accuracies": 0.8333333730697632, + "rewards/chosen": -0.06482670456171036, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.0777878761291504, + "rewards/student_margin": 2.012960910797119, + "rewards/teacher_margin": 0.0, + "step": 120 + }, + { + "epoch": 0.08, + "grad_norm": 36.5, + "learning_rate": 3.7790697674418607e-06, + "logits/chosen": -0.6423686742782593, + "logits/rejected": -0.049126945436000824, + "logps/chosen": -207.14761352539062, + "logps/rejected": -422.570556640625, + "loss": 0.3735, + "rewards/accuracies": 0.9000000953674316, + "rewards/chosen": -0.0891571044921875, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.900564193725586, + "rewards/student_margin": 1.8114070892333984, + "rewards/teacher_margin": 0.0, + "step": 130 + }, + { + "epoch": 0.08, + "grad_norm": 22.625, + "learning_rate": 4.0697674418604655e-06, + "logits/chosen": -0.6022151112556458, + "logits/rejected": -0.05691419914364815, + "logps/chosen": -294.1002502441406, + "logps/rejected": -565.58642578125, + "loss": 0.3309, + "rewards/accuracies": 0.7999999523162842, + "rewards/chosen": -0.2062998265028, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.2702088356018066, + "rewards/student_margin": 2.0639092922210693, + "rewards/teacher_margin": 0.0, + "step": 140 + }, + { + "epoch": 0.09, + "grad_norm": 21.5, + "learning_rate": 4.36046511627907e-06, + "logits/chosen": -0.5255838632583618, + "logits/rejected": -0.025744330137968063, + "logps/chosen": -206.9980010986328, + "logps/rejected": -500.0045471191406, + "loss": 0.4189, + "rewards/accuracies": 0.800000011920929, + "rewards/chosen": 0.19613902270793915, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.1442222595214844, + "rewards/student_margin": 2.3403611183166504, + "rewards/teacher_margin": 0.0, + "step": 150 + }, + { + "epoch": 0.09, + "grad_norm": 28.625, + "learning_rate": 4.651162790697675e-06, + "logits/chosen": -0.595037043094635, + "logits/rejected": -0.05052659660577774, + "logps/chosen": -228.1767578125, + "logps/rejected": -517.8463134765625, + "loss": 0.3434, + "rewards/accuracies": 0.8333333134651184, + "rewards/chosen": 0.011402442120015621, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.753488302230835, + "rewards/student_margin": 2.764890670776367, + "rewards/teacher_margin": 0.0, + "step": 160 + }, + { + "epoch": 0.1, + "grad_norm": 22.625, + "learning_rate": 4.941860465116279e-06, + "logits/chosen": -0.8135232925415039, + "logits/rejected": 0.09407637268304825, + "logps/chosen": -175.84527587890625, + "logps/rejected": -517.9186401367188, + "loss": 0.3348, + "rewards/accuracies": 1.0, + "rewards/chosen": 0.3143172264099121, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -3.1565933227539062, + "rewards/student_margin": 3.4709103107452393, + "rewards/teacher_margin": 0.0, + "step": 170 + }, + { + "epoch": 0.1, + "grad_norm": 31.25, + "learning_rate": 4.999669659981585e-06, + "logits/chosen": -0.6553431153297424, + "logits/rejected": 0.02322854846715927, + "logps/chosen": -190.44459533691406, + "logps/rejected": -498.41845703125, + "loss": 0.3838, + "rewards/accuracies": 0.7666666507720947, + "rewards/chosen": -0.03836626559495926, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.3416810035705566, + "rewards/student_margin": 2.303314685821533, + "rewards/teacher_margin": 0.0, + "step": 180 + }, + { + "epoch": 0.11, + "grad_norm": 23.625, + "learning_rate": 4.998327803275328e-06, + "logits/chosen": -0.5196131467819214, + "logits/rejected": 0.03156811743974686, + "logps/chosen": -244.62362670898438, + "logps/rejected": -550.338134765625, + "loss": 0.3397, + "rewards/accuracies": 0.8666666150093079, + "rewards/chosen": 0.1986195147037506, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.157165050506592, + "rewards/student_margin": 2.3557844161987305, + "rewards/teacher_margin": 0.0, + "step": 190 + }, + { + "epoch": 0.12, + "grad_norm": 23.0, + "learning_rate": 4.995954337283145e-06, + "logits/chosen": -0.46547913551330566, + "logits/rejected": 0.002417749259620905, + "logps/chosen": -220.58139038085938, + "logps/rejected": -457.37957763671875, + "loss": 0.3513, + "rewards/accuracies": 0.7666667103767395, + "rewards/chosen": 0.23564238846302032, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.046499729156494, + "rewards/student_margin": 2.282142162322998, + "rewards/teacher_margin": 0.0, + "step": 200 + }, + { + "epoch": 0.12, + "grad_norm": 29.625, + "learning_rate": 4.992550242056393e-06, + "logits/chosen": -0.7077560424804688, + "logits/rejected": 0.23227286338806152, + "logps/chosen": -250.8577117919922, + "logps/rejected": -470.998046875, + "loss": 0.3619, + "rewards/accuracies": 0.76666659116745, + "rewards/chosen": 0.30192822217941284, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.5922541618347168, + "rewards/student_margin": 1.8941824436187744, + "rewards/teacher_margin": 0.0, + "step": 210 + }, + { + "epoch": 0.13, + "grad_norm": 13.4375, + "learning_rate": 4.988116923213761e-06, + "logits/chosen": -0.5510163903236389, + "logits/rejected": 0.061091721057891846, + "logps/chosen": -231.3572235107422, + "logps/rejected": -456.94219970703125, + "loss": 0.3619, + "rewards/accuracies": 0.8333333134651184, + "rewards/chosen": 0.7294235825538635, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.8068240880966187, + "rewards/student_margin": 2.536247491836548, + "rewards/teacher_margin": 0.0, + "step": 220 + }, + { + "epoch": 0.13, + "grad_norm": 23.125, + "learning_rate": 4.982656211360852e-06, + "logits/chosen": -0.6307552456855774, + "logits/rejected": 0.11015477031469345, + "logps/chosen": -210.99728393554688, + "logps/rejected": -536.0760498046875, + "loss": 0.3218, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 0.6962422132492065, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.4401557445526123, + "rewards/student_margin": 3.1363983154296875, + "rewards/teacher_margin": 0.0, + "step": 230 + }, + { + "epoch": 0.14, + "grad_norm": 24.75, + "learning_rate": 4.976170361334305e-06, + "logits/chosen": -0.47195425629615784, + "logits/rejected": -0.022416137158870697, + "logps/chosen": -255.0712127685547, + "logps/rejected": -450.69989013671875, + "loss": 0.3947, + "rewards/accuracies": 0.8333333134651184, + "rewards/chosen": 0.3207937180995941, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.8813111782073975, + "rewards/student_margin": 2.2021050453186035, + "rewards/teacher_margin": 0.0, + "step": 240 + }, + { + "epoch": 0.15, + "grad_norm": 39.25, + "learning_rate": 4.968662051270712e-06, + "logits/chosen": -0.7277335524559021, + "logits/rejected": 0.09811758249998093, + "logps/chosen": -212.38381958007812, + "logps/rejected": -535.5774536132812, + "loss": 0.3208, + "rewards/accuracies": 0.8333333134651184, + "rewards/chosen": 0.7789496183395386, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.0195579528808594, + "rewards/student_margin": 2.7985076904296875, + "rewards/teacher_margin": 0.0, + "step": 250 + }, + { + "epoch": 0.15, + "grad_norm": 17.125, + "learning_rate": 4.960134381500771e-06, + "logits/chosen": -0.6490304470062256, + "logits/rejected": 0.1224866732954979, + "logps/chosen": -208.1616973876953, + "logps/rejected": -549.3804931640625, + "loss": 0.3667, + "rewards/accuracies": 0.9333332777023315, + "rewards/chosen": 1.0950639247894287, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.4546568393707275, + "rewards/student_margin": 3.5497207641601562, + "rewards/teacher_margin": 0.0, + "step": 260 + }, + { + "epoch": 0.16, + "grad_norm": 27.5, + "learning_rate": 4.950590873269098e-06, + "logits/chosen": -0.5456488132476807, + "logits/rejected": 0.020637910813093185, + "logps/chosen": -259.57196044921875, + "logps/rejected": -409.33123779296875, + "loss": 0.3453, + "rewards/accuracies": 0.7666666507720947, + "rewards/chosen": 0.4798993468284607, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.331217885017395, + "rewards/student_margin": 1.811117172241211, + "rewards/teacher_margin": 0.0, + "step": 270 + }, + { + "epoch": 0.16, + "grad_norm": 35.0, + "learning_rate": 4.940035467280229e-06, + "logits/chosen": -0.4896344542503357, + "logits/rejected": 0.024026235565543175, + "logps/chosen": -237.4656524658203, + "logps/rejected": -509.08404541015625, + "loss": 0.3292, + "rewards/accuracies": 0.9333332777023315, + "rewards/chosen": 0.728676974773407, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.34112811088562, + "rewards/student_margin": 3.069805145263672, + "rewards/teacher_margin": 0.0, + "step": 280 + }, + { + "epoch": 0.17, + "grad_norm": 26.5, + "learning_rate": 4.928472522071432e-06, + "logits/chosen": -0.6238890290260315, + "logits/rejected": 0.1964757889509201, + "logps/chosen": -213.3905029296875, + "logps/rejected": -478.8919372558594, + "loss": 0.3695, + "rewards/accuracies": 0.76666659116745, + "rewards/chosen": 0.6496525406837463, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.2729125022888184, + "rewards/student_margin": 1.9225651025772095, + "rewards/teacher_margin": 0.0, + "step": 290 + }, + { + "epoch": 0.17, + "grad_norm": 29.375, + "learning_rate": 4.915906812212976e-06, + "logits/chosen": -0.6193239092826843, + "logits/rejected": 0.05876749008893967, + "logps/chosen": -202.52391052246094, + "logps/rejected": -458.86492919921875, + "loss": 0.2946, + "rewards/accuracies": 0.8666666150093079, + "rewards/chosen": 0.8241189122200012, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.658251166343689, + "rewards/student_margin": 2.482369899749756, + "rewards/teacher_margin": 0.0, + "step": 300 + }, + { + "epoch": 0.18, + "grad_norm": 16.125, + "learning_rate": 4.9023435263366145e-06, + "logits/chosen": -0.9334913492202759, + "logits/rejected": -0.058604609221220016, + "logps/chosen": -168.749267578125, + "logps/rejected": -640.4456176757812, + "loss": 0.2114, + "rewards/accuracies": 0.966666579246521, + "rewards/chosen": 0.5594726800918579, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -3.131923198699951, + "rewards/student_margin": 3.6913959980010986, + "rewards/teacher_margin": 0.0, + "step": 310 + }, + { + "epoch": 0.19, + "grad_norm": 15.125, + "learning_rate": 4.887788264993099e-06, + "logits/chosen": -0.6044551730155945, + "logits/rejected": -0.04338273033499718, + "logps/chosen": -224.3017578125, + "logps/rejected": -549.9353637695312, + "loss": 0.3094, + "rewards/accuracies": 0.8333333730697632, + "rewards/chosen": 0.7394543290138245, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.3911728858947754, + "rewards/student_margin": 3.130627393722534, + "rewards/teacher_margin": 0.0, + "step": 320 + }, + { + "epoch": 0.19, + "grad_norm": 23.375, + "learning_rate": 4.872247038339603e-06, + "logits/chosen": -0.7818952798843384, + "logits/rejected": 0.06497599929571152, + "logps/chosen": -192.23880004882812, + "logps/rejected": -626.657470703125, + "loss": 0.2646, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 1.0649678707122803, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -3.6517505645751953, + "rewards/student_margin": 4.716717720031738, + "rewards/teacher_margin": 0.0, + "step": 330 + }, + { + "epoch": 0.2, + "grad_norm": 30.625, + "learning_rate": 4.855726263657999e-06, + "logits/chosen": -0.29226094484329224, + "logits/rejected": -0.13234995305538177, + "logps/chosen": -264.1004333496094, + "logps/rejected": -362.9111633300781, + "loss": 0.4197, + "rewards/accuracies": 0.800000011920929, + "rewards/chosen": 0.6886833906173706, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.4170998334884644, + "rewards/student_margin": 2.105782985687256, + "rewards/teacher_margin": 0.0, + "step": 340 + }, + { + "epoch": 0.2, + "grad_norm": 24.25, + "learning_rate": 4.838232762705044e-06, + "logits/chosen": -0.4636480212211609, + "logits/rejected": 0.027485806494951248, + "logps/chosen": -194.08128356933594, + "logps/rejected": -531.1504516601562, + "loss": 0.2651, + "rewards/accuracies": 0.9333332777023315, + "rewards/chosen": 1.0223414897918701, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.3282663822174072, + "rewards/student_margin": 3.3506076335906982, + "rewards/teacher_margin": 0.0, + "step": 350 + }, + { + "epoch": 0.21, + "grad_norm": 27.5, + "learning_rate": 4.819773758895538e-06, + "logits/chosen": -0.7649397253990173, + "logits/rejected": 0.06376661360263824, + "logps/chosen": -163.4183807373047, + "logps/rejected": -592.5127563476562, + "loss": 0.3665, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 0.8098726272583008, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -3.550152540206909, + "rewards/student_margin": 4.360025405883789, + "rewards/teacher_margin": 0.0, + "step": 360 + }, + { + "epoch": 0.22, + "grad_norm": 22.125, + "learning_rate": 4.8003568743196294e-06, + "logits/chosen": -0.5304759740829468, + "logits/rejected": 0.034225065261125565, + "logps/chosen": -187.56813049316406, + "logps/rejected": -569.3456420898438, + "loss": 0.2797, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 0.6876587867736816, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.1767399311065674, + "rewards/student_margin": 2.864398956298828, + "rewards/teacher_margin": 0.0, + "step": 370 + }, + { + "epoch": 0.22, + "grad_norm": 29.625, + "learning_rate": 4.779990126595505e-06, + "logits/chosen": -0.5957476496696472, + "logits/rejected": -0.04448353499174118, + "logps/chosen": -200.004150390625, + "logps/rejected": -613.2229614257812, + "loss": 0.3936, + "rewards/accuracies": 0.9333332777023315, + "rewards/chosen": 0.7026926279067993, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.0374011993408203, + "rewards/student_margin": 2.740093946456909, + "rewards/teacher_margin": 0.0, + "step": 380 + }, + { + "epoch": 0.23, + "grad_norm": 26.375, + "learning_rate": 4.758681925558756e-06, + "logits/chosen": -0.5602328777313232, + "logits/rejected": -0.012245875783264637, + "logps/chosen": -185.31021118164062, + "logps/rejected": -407.41766357421875, + "loss": 0.3227, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 0.620998740196228, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.894086480140686, + "rewards/student_margin": 2.515085458755493, + "rewards/teacher_margin": 0.0, + "step": 390 + }, + { + "epoch": 0.23, + "grad_norm": 20.125, + "learning_rate": 4.736441069789786e-06, + "logits/chosen": -0.46345287561416626, + "logits/rejected": 0.001672474667429924, + "logps/chosen": -228.4940948486328, + "logps/rejected": -361.48175048828125, + "loss": 0.2846, + "rewards/accuracies": 0.6999999284744263, + "rewards/chosen": 0.4205475449562073, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.4956649541854858, + "rewards/student_margin": 1.9162124395370483, + "rewards/teacher_margin": 0.0, + "step": 400 + }, + { + "epoch": 0.24, + "grad_norm": 16.5, + "learning_rate": 4.713276742980708e-06, + "logits/chosen": -0.4360102713108063, + "logits/rejected": -0.10899927467107773, + "logps/chosen": -249.0923309326172, + "logps/rejected": -560.4241333007812, + "loss": 0.2865, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 0.8636870384216309, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.2510733604431152, + "rewards/student_margin": 3.114760637283325, + "rewards/teacher_margin": 0.0, + "step": 410 + }, + { + "epoch": 0.24, + "grad_norm": 30.0, + "learning_rate": 4.6891985101432085e-06, + "logits/chosen": -0.3729998469352722, + "logits/rejected": -0.028511980548501015, + "logps/chosen": -207.8327178955078, + "logps/rejected": -356.3282775878906, + "loss": 0.3602, + "rewards/accuracies": 0.7333332896232605, + "rewards/chosen": 0.9164339303970337, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -0.7431859374046326, + "rewards/student_margin": 1.659619927406311, + "rewards/teacher_margin": 0.0, + "step": 420 + }, + { + "epoch": 0.25, + "grad_norm": 29.0, + "learning_rate": 4.664216313658969e-06, + "logits/chosen": -0.324567973613739, + "logits/rejected": 0.016650067642331123, + "logps/chosen": -206.4794921875, + "logps/rejected": -445.509033203125, + "loss": 0.3204, + "rewards/accuracies": 0.8666666746139526, + "rewards/chosen": 1.2572410106658936, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -0.936699390411377, + "rewards/student_margin": 2.1939404010772705, + "rewards/teacher_margin": 0.0, + "step": 430 + }, + { + "epoch": 0.26, + "grad_norm": 28.0, + "learning_rate": 4.638340469174256e-06, + "logits/chosen": -0.5071316957473755, + "logits/rejected": 0.044037409126758575, + "logps/chosen": -180.57435607910156, + "logps/rejected": -601.9741821289062, + "loss": 0.2881, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 1.1313551664352417, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.505422830581665, + "rewards/student_margin": 3.6367783546447754, + "rewards/teacher_margin": 0.0, + "step": 440 + }, + { + "epoch": 0.26, + "grad_norm": 18.625, + "learning_rate": 4.611581661340386e-06, + "logits/chosen": -0.5572972297668457, + "logits/rejected": 0.05740237236022949, + "logps/chosen": -202.59849548339844, + "logps/rejected": -579.298095703125, + "loss": 0.2705, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 1.076784372329712, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.998802661895752, + "rewards/student_margin": 4.075586795806885, + "rewards/teacher_margin": 0.0, + "step": 450 + }, + { + "epoch": 0.27, + "grad_norm": 23.625, + "learning_rate": 4.5839509394018185e-06, + "logits/chosen": -0.37410736083984375, + "logits/rejected": 0.017481276765465736, + "logps/chosen": -243.3333282470703, + "logps/rejected": -485.5693359375, + "loss": 0.2463, + "rewards/accuracies": 0.8666666150093079, + "rewards/chosen": 0.691962480545044, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.156360149383545, + "rewards/student_margin": 2.8483223915100098, + "rewards/teacher_margin": 0.0, + "step": 460 + }, + { + "epoch": 0.27, + "grad_norm": 11.9375, + "learning_rate": 4.555459712633707e-06, + "logits/chosen": -0.6188653707504272, + "logits/rejected": 0.16356995701789856, + "logps/chosen": -238.00772094726562, + "logps/rejected": -447.47442626953125, + "loss": 0.273, + "rewards/accuracies": 0.8333333730697632, + "rewards/chosen": 1.3531479835510254, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.5606005191802979, + "rewards/student_margin": 2.913748264312744, + "rewards/teacher_margin": 0.0, + "step": 470 + }, + { + "epoch": 0.28, + "grad_norm": 15.5625, + "learning_rate": 4.526119745630779e-06, + "logits/chosen": -0.5436848402023315, + "logits/rejected": 0.04653601720929146, + "logps/chosen": -228.1831817626953, + "logps/rejected": -620.6707153320312, + "loss": 0.2793, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 1.1916218996047974, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.1671457290649414, + "rewards/student_margin": 3.3587677478790283, + "rewards/teacher_margin": 0.0, + "step": 480 + }, + { + "epoch": 0.29, + "grad_norm": 20.125, + "learning_rate": 4.4959431534495015e-06, + "logits/chosen": -0.1781628429889679, + "logits/rejected": 0.06491532176733017, + "logps/chosen": -266.6177673339844, + "logps/rejected": -403.18841552734375, + "loss": 0.2836, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 2.089085340499878, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -0.6247254610061646, + "rewards/student_margin": 2.713810682296753, + "rewards/teacher_margin": 0.0, + "step": 490 + }, + { + "epoch": 0.29, + "grad_norm": 28.625, + "learning_rate": 4.464942396605531e-06, + "logits/chosen": -0.5647803544998169, + "logits/rejected": 0.09310182183980942, + "logps/chosen": -185.3010711669922, + "logps/rejected": -453.74188232421875, + "loss": 0.2228, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.3027641773223877, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.6533056497573853, + "rewards/student_margin": 2.9560697078704834, + "rewards/teacher_margin": 0.0, + "step": 500 + }, + { + "epoch": 0.3, + "grad_norm": 27.625, + "learning_rate": 4.433130275928522e-06, + "logits/chosen": -0.508124828338623, + "logits/rejected": 0.05562957376241684, + "logps/chosen": -202.5834197998047, + "logps/rejected": -409.6363220214844, + "loss": 0.2788, + "rewards/accuracies": 0.8333333730697632, + "rewards/chosen": 1.8609619140625, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.7967350482940674, + "rewards/student_margin": 3.6576972007751465, + "rewards/teacher_margin": 0.0, + "step": 510 + }, + { + "epoch": 0.3, + "grad_norm": 17.0, + "learning_rate": 4.400519927276411e-06, + "logits/chosen": -0.48706334829330444, + "logits/rejected": -0.07246413826942444, + "logps/chosen": -216.9883270263672, + "logps/rejected": -562.3739013671875, + "loss": 0.2144, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.0860559940338135, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -3.157426357269287, + "rewards/student_margin": 4.2434821128845215, + "rewards/teacher_margin": 0.0, + "step": 520 + }, + { + "epoch": 0.31, + "grad_norm": 21.5, + "learning_rate": 4.367124816111351e-06, + "logits/chosen": -0.3559921979904175, + "logits/rejected": -0.11402563750743866, + "logps/chosen": -218.14559936523438, + "logps/rejected": -421.6756286621094, + "loss": 0.2122, + "rewards/accuracies": 0.8666666150093079, + "rewards/chosen": 1.3421014547348022, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.9218908548355103, + "rewards/student_margin": 3.2639918327331543, + "rewards/teacher_margin": 0.0, + "step": 530 + }, + { + "epoch": 0.31, + "grad_norm": 11.375, + "learning_rate": 4.332958731939562e-06, + "logits/chosen": -0.7146890163421631, + "logits/rejected": 0.2430504560470581, + "logps/chosen": -173.04917907714844, + "logps/rejected": -493.38641357421875, + "loss": 0.2895, + "rewards/accuracies": 0.8333333134651184, + "rewards/chosen": 1.2438677549362183, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.6418049335479736, + "rewards/student_margin": 2.8856725692749023, + "rewards/teacher_margin": 0.0, + "step": 540 + }, + { + "epoch": 0.32, + "grad_norm": 29.75, + "learning_rate": 4.2980357826173665e-06, + "logits/chosen": -0.37195292115211487, + "logits/rejected": 0.02155585028231144, + "logps/chosen": -223.03817749023438, + "logps/rejected": -389.7994689941406, + "loss": 0.3108, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.6178417205810547, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.4387261867523193, + "rewards/student_margin": 3.056567907333374, + "rewards/teacher_margin": 0.0, + "step": 550 + }, + { + "epoch": 0.33, + "grad_norm": 21.875, + "learning_rate": 4.262370388525777e-06, + "logits/chosen": -0.3869599401950836, + "logits/rejected": 0.07817984372377396, + "logps/chosen": -175.74595642089844, + "logps/rejected": -394.0006408691406, + "loss": 0.2445, + "rewards/accuracies": 0.966666579246521, + "rewards/chosen": 2.0001840591430664, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.1199393272399902, + "rewards/student_margin": 4.120123863220215, + "rewards/teacher_margin": 0.0, + "step": 560 + }, + { + "epoch": 0.33, + "grad_norm": 32.25, + "learning_rate": 4.225977276616028e-06, + "logits/chosen": -0.5689536333084106, + "logits/rejected": 0.11373758316040039, + "logps/chosen": -219.41482543945312, + "logps/rejected": -426.211669921875, + "loss": 0.2718, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.7283222675323486, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.4601569175720215, + "rewards/student_margin": 4.188479423522949, + "rewards/teacher_margin": 0.0, + "step": 570 + }, + { + "epoch": 0.34, + "grad_norm": 12.5, + "learning_rate": 4.1888714743285305e-06, + "logits/chosen": -0.3077758550643921, + "logits/rejected": -0.01714705489575863, + "logps/chosen": -228.143798828125, + "logps/rejected": -587.8424682617188, + "loss": 0.2311, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 1.6572096347808838, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -3.5316357612609863, + "rewards/student_margin": 5.188845157623291, + "rewards/teacher_margin": 0.0, + "step": 580 + }, + { + "epoch": 0.34, + "grad_norm": 15.1875, + "learning_rate": 4.151068303387739e-06, + "logits/chosen": -0.442771852016449, + "logits/rejected": 0.08343404531478882, + "logps/chosen": -172.50587463378906, + "logps/rejected": -430.36859130859375, + "loss": 0.2294, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.2904380559921265, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.5538195371627808, + "rewards/student_margin": 2.8442578315734863, + "rewards/teacher_margin": 0.0, + "step": 590 + }, + { + "epoch": 0.35, + "grad_norm": 17.125, + "learning_rate": 4.1125833734755115e-06, + "logits/chosen": -0.36492887139320374, + "logits/rejected": 0.17705991864204407, + "logps/chosen": -279.50311279296875, + "logps/rejected": -664.5555419921875, + "loss": 0.2703, + "rewards/accuracies": 0.9000000953674316, + "rewards/chosen": 1.2323994636535645, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.9321751594543457, + "rewards/student_margin": 4.16457462310791, + "rewards/teacher_margin": 0.0, + "step": 600 + }, + { + "epoch": 0.36, + "grad_norm": 13.75, + "learning_rate": 4.073432575785554e-06, + "logits/chosen": -0.1946389228105545, + "logits/rejected": 0.013921762816607952, + "logps/chosen": -248.22073364257812, + "logps/rejected": -516.2274169921875, + "loss": 0.2433, + "rewards/accuracies": 0.8333333134651184, + "rewards/chosen": 1.737766981124878, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.9633979797363281, + "rewards/student_margin": 3.7011654376983643, + "rewards/teacher_margin": 0.0, + "step": 610 + }, + { + "epoch": 0.36, + "grad_norm": 27.125, + "learning_rate": 4.033632076461634e-06, + "logits/chosen": -0.35111454129219055, + "logits/rejected": 0.09419815242290497, + "logps/chosen": -237.6559295654297, + "logps/rejected": -470.4676818847656, + "loss": 0.3378, + "rewards/accuracies": 0.9333332777023315, + "rewards/chosen": 1.6999142169952393, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.3078672885894775, + "rewards/student_margin": 3.007781505584717, + "rewards/teacher_margin": 0.0, + "step": 620 + }, + { + "epoch": 0.37, + "grad_norm": 15.1875, + "learning_rate": 3.9931983099222595e-06, + "logits/chosen": -0.47325119376182556, + "logits/rejected": -0.02582048997282982, + "logps/chosen": -222.41415405273438, + "logps/rejected": -465.8368225097656, + "loss": 0.2413, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 1.1449840068817139, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.2412731647491455, + "rewards/student_margin": 2.3862571716308594, + "rewards/teacher_margin": 0.0, + "step": 630 + }, + { + "epoch": 0.37, + "grad_norm": 18.875, + "learning_rate": 3.9521479720745785e-06, + "logits/chosen": -0.45588645339012146, + "logits/rejected": 0.014990913681685925, + "logps/chosen": -208.7266845703125, + "logps/rejected": -565.1739501953125, + "loss": 0.2011, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.211085319519043, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.549320936203003, + "rewards/student_margin": 3.760406017303467, + "rewards/teacher_margin": 0.0, + "step": 640 + }, + { + "epoch": 0.38, + "grad_norm": 20.25, + "learning_rate": 3.910498013420303e-06, + "logits/chosen": -0.5036629438400269, + "logits/rejected": -0.04427925869822502, + "logps/chosen": -158.07492065429688, + "logps/rejected": -435.2080078125, + "loss": 0.2739, + "rewards/accuracies": 0.800000011920929, + "rewards/chosen": 1.043824315071106, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.8605940341949463, + "rewards/student_margin": 2.9044182300567627, + "rewards/teacher_margin": 0.0, + "step": 650 + }, + { + "epoch": 0.38, + "grad_norm": 29.0, + "learning_rate": 3.8682656320565205e-06, + "logits/chosen": -0.41261768341064453, + "logits/rejected": 0.030822629109025, + "logps/chosen": -225.4918670654297, + "logps/rejected": -392.59613037109375, + "loss": 0.2616, + "rewards/accuracies": 0.8666666746139526, + "rewards/chosen": 1.8061786890029907, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.9655482769012451, + "rewards/student_margin": 3.771726608276367, + "rewards/teacher_margin": 0.0, + "step": 660 + }, + { + "epoch": 0.39, + "grad_norm": 22.0, + "learning_rate": 3.8254682665742425e-06, + "logits/chosen": -0.31573912501335144, + "logits/rejected": 0.005792762152850628, + "logps/chosen": -240.8995361328125, + "logps/rejected": -505.09735107421875, + "loss": 0.2795, + "rewards/accuracies": 0.8333333134651184, + "rewards/chosen": 1.1914173364639282, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.8682523965835571, + "rewards/student_margin": 3.0596697330474854, + "rewards/teacher_margin": 0.0, + "step": 670 + }, + { + "epoch": 0.4, + "grad_norm": 18.5, + "learning_rate": 3.7821235888576744e-06, + "logits/chosen": -0.2197073996067047, + "logits/rejected": 0.02696964144706726, + "logps/chosen": -244.7580108642578, + "logps/rejected": -460.04400634765625, + "loss": 0.1711, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.5450682640075684, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.898803472518921, + "rewards/student_margin": 3.4438719749450684, + "rewards/teacher_margin": 0.0, + "step": 680 + }, + { + "epoch": 0.4, + "grad_norm": 9.5625, + "learning_rate": 3.7382494967871384e-06, + "logits/chosen": -0.4361536502838135, + "logits/rejected": 0.13731250166893005, + "logps/chosen": -246.27395629882812, + "logps/rejected": -633.4682006835938, + "loss": 0.278, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.635087251663208, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.891167640686035, + "rewards/student_margin": 4.526254653930664, + "rewards/teacher_margin": 0.0, + "step": 690 + }, + { + "epoch": 0.41, + "grad_norm": 16.125, + "learning_rate": 3.693864106848683e-06, + "logits/chosen": -0.4011663794517517, + "logits/rejected": 0.015058115124702454, + "logps/chosen": -248.08349609375, + "logps/rejected": -427.1165466308594, + "loss": 0.1977, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.8699365854263306, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.281233787536621, + "rewards/student_margin": 3.151170253753662, + "rewards/teacher_margin": 0.0, + "step": 700 + }, + { + "epoch": 0.41, + "grad_norm": 24.375, + "learning_rate": 3.648985746653424e-06, + "logits/chosen": -0.3061038553714752, + "logits/rejected": -0.029941242188215256, + "logps/chosen": -201.59161376953125, + "logps/rejected": -463.8720703125, + "loss": 0.1973, + "rewards/accuracies": 0.8666666746139526, + "rewards/chosen": 1.5523725748062134, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.9691741466522217, + "rewards/student_margin": 3.521547317504883, + "rewards/teacher_margin": 0.0, + "step": 710 + }, + { + "epoch": 0.42, + "grad_norm": 12.875, + "learning_rate": 3.6036329473697103e-06, + "logits/chosen": -0.613228976726532, + "logits/rejected": 0.009662175551056862, + "logps/chosen": -196.20706176757812, + "logps/rejected": -525.05908203125, + "loss": 0.1909, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 1.3759281635284424, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.8588900566101074, + "rewards/student_margin": 4.234818458557129, + "rewards/teacher_margin": 0.0, + "step": 720 + }, + { + "epoch": 0.42, + "grad_norm": 12.9375, + "learning_rate": 3.5578244360712373e-06, + "logits/chosen": -0.5177701115608215, + "logits/rejected": 0.176934152841568, + "logps/chosen": -206.4003143310547, + "logps/rejected": -487.99249267578125, + "loss": 0.2496, + "rewards/accuracies": 0.9000000953674316, + "rewards/chosen": 1.4539085626602173, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.346587657928467, + "rewards/student_margin": 3.8004963397979736, + "rewards/teacher_margin": 0.0, + "step": 730 + }, + { + "epoch": 0.43, + "grad_norm": 8.375, + "learning_rate": 3.511579128004261e-06, + "logits/chosen": -0.44068461656570435, + "logits/rejected": 0.005051338579505682, + "logps/chosen": -272.7444763183594, + "logps/rejected": -478.147705078125, + "loss": 0.2245, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.060945987701416, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.0382800102233887, + "rewards/student_margin": 3.0992257595062256, + "rewards/teacher_margin": 0.0, + "step": 740 + }, + { + "epoch": 0.44, + "grad_norm": 18.0, + "learning_rate": 3.464916118777121e-06, + "logits/chosen": -0.5848614573478699, + "logits/rejected": 0.30563992261886597, + "logps/chosen": -184.22329711914062, + "logps/rejected": -468.75970458984375, + "loss": 0.2374, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 1.795701026916504, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.5491726398468018, + "rewards/student_margin": 3.3448734283447266, + "rewards/teacher_margin": 0.0, + "step": 750 + }, + { + "epoch": 0.44, + "grad_norm": 8.1875, + "learning_rate": 3.417854676475279e-06, + "logits/chosen": -0.38186392188072205, + "logits/rejected": 0.059945814311504364, + "logps/chosen": -260.58978271484375, + "logps/rejected": -594.833251953125, + "loss": 0.2078, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 1.7202898263931274, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.4396731853485107, + "rewards/student_margin": 4.159963130950928, + "rewards/teacher_margin": 0.0, + "step": 760 + }, + { + "epoch": 0.45, + "grad_norm": 16.0, + "learning_rate": 3.370414233705147e-06, + "logits/chosen": -0.2833537757396698, + "logits/rejected": 0.09210254997015, + "logps/chosen": -223.1533203125, + "logps/rejected": -582.6676025390625, + "loss": 0.2096, + "rewards/accuracies": 0.966666579246521, + "rewards/chosen": 1.4522992372512817, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -3.2473137378692627, + "rewards/student_margin": 4.699612617492676, + "rewards/teacher_margin": 0.0, + "step": 770 + }, + { + "epoch": 0.45, + "grad_norm": 8.375, + "learning_rate": 3.322614379569975e-06, + "logits/chosen": -0.302879273891449, + "logits/rejected": -0.14690950512886047, + "logps/chosen": -225.3035888671875, + "logps/rejected": -357.20330810546875, + "loss": 0.2293, + "rewards/accuracies": 0.7666666507720947, + "rewards/chosen": 1.1200844049453735, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.0596327781677246, + "rewards/student_margin": 3.1797170639038086, + "rewards/teacher_margin": 0.0, + "step": 780 + }, + { + "epoch": 0.46, + "grad_norm": 19.5, + "learning_rate": 3.27447485158112e-06, + "logits/chosen": -0.4661865830421448, + "logits/rejected": 0.12939366698265076, + "logps/chosen": -226.401611328125, + "logps/rejected": -421.63592529296875, + "loss": 0.1703, + "rewards/accuracies": 0.8666666150093079, + "rewards/chosen": 1.3359860181808472, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.47092604637146, + "rewards/student_margin": 3.806912660598755, + "rewards/teacher_margin": 0.0, + "step": 790 + }, + { + "epoch": 0.47, + "grad_norm": 24.125, + "learning_rate": 3.226015527508036e-06, + "logits/chosen": -0.5169921517372131, + "logits/rejected": -0.004798925016075373, + "logps/chosen": -241.3819580078125, + "logps/rejected": -335.36492919921875, + "loss": 0.2056, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.6830694675445557, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.760565996170044, + "rewards/student_margin": 3.4436354637145996, + "rewards/teacher_margin": 0.0, + "step": 800 + }, + { + "epoch": 0.47, + "grad_norm": 19.875, + "learning_rate": 3.1772564171703475e-06, + "logits/chosen": -0.28334909677505493, + "logits/rejected": -0.021695634350180626, + "logps/chosen": -255.8955078125, + "logps/rejected": -581.0257568359375, + "loss": 0.3028, + "rewards/accuracies": 0.8666666150093079, + "rewards/chosen": 1.906981110572815, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -3.3529152870178223, + "rewards/student_margin": 5.259896278381348, + "rewards/teacher_margin": 0.0, + "step": 810 + }, + { + "epoch": 0.48, + "grad_norm": 17.875, + "learning_rate": 3.128217654175393e-06, + "logits/chosen": -0.3028213381767273, + "logits/rejected": 0.1735200583934784, + "logps/chosen": -269.35162353515625, + "logps/rejected": -464.0060119628906, + "loss": 0.1942, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.3699723482131958, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.393336057662964, + "rewards/student_margin": 3.7633087635040283, + "rewards/teacher_margin": 0.0, + "step": 820 + }, + { + "epoch": 0.48, + "grad_norm": 18.5, + "learning_rate": 3.078919487604661e-06, + "logits/chosen": -0.43081316351890564, + "logits/rejected": -0.09694792330265045, + "logps/chosen": -211.90707397460938, + "logps/rejected": -447.4751892089844, + "loss": 0.1913, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 1.855574369430542, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.7887474298477173, + "rewards/student_margin": 3.644321918487549, + "rewards/teacher_margin": 0.0, + "step": 830 + }, + { + "epoch": 0.49, + "grad_norm": 12.75, + "learning_rate": 3.029382273652544e-06, + "logits/chosen": -0.4358023703098297, + "logits/rejected": 0.1673169583082199, + "logps/chosen": -241.4596710205078, + "logps/rejected": -499.11956787109375, + "loss": 0.1966, + "rewards/accuracies": 0.8666666746139526, + "rewards/chosen": 1.5806983709335327, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.8262617588043213, + "rewards/student_margin": 3.4069600105285645, + "rewards/teacher_margin": 0.0, + "step": 840 + }, + { + "epoch": 0.49, + "grad_norm": 16.875, + "learning_rate": 2.9796264672208535e-06, + "logits/chosen": -0.44691354036331177, + "logits/rejected": 0.14132077991962433, + "logps/chosen": -217.2314453125, + "logps/rejected": -476.49859619140625, + "loss": 0.1794, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.7376991510391235, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.2412829399108887, + "rewards/student_margin": 3.9789822101593018, + "rewards/teacher_margin": 0.0, + "step": 850 + }, + { + "epoch": 0.5, + "grad_norm": 12.375, + "learning_rate": 2.929672613472589e-06, + "logits/chosen": -0.6897366642951965, + "logits/rejected": 0.006776002235710621, + "logps/chosen": -216.3780517578125, + "logps/rejected": -620.7020263671875, + "loss": 0.1916, + "rewards/accuracies": 0.8666666150093079, + "rewards/chosen": 1.432846188545227, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -3.6632869243621826, + "rewards/student_margin": 5.096132755279541, + "rewards/teacher_margin": 0.0, + "step": 860 + }, + { + "epoch": 0.51, + "grad_norm": 13.375, + "learning_rate": 2.879541339348429e-06, + "logits/chosen": -0.5285438895225525, + "logits/rejected": 0.00928882509469986, + "logps/chosen": -242.85098266601562, + "logps/rejected": -481.8531188964844, + "loss": 0.1813, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.5595648288726807, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.5762245655059814, + "rewards/student_margin": 4.135788440704346, + "rewards/teacher_margin": 0.0, + "step": 870 + }, + { + "epoch": 0.51, + "grad_norm": 24.25, + "learning_rate": 2.829253345049458e-06, + "logits/chosen": -0.31855055689811707, + "logits/rejected": 0.0941370353102684, + "logps/chosen": -240.16650390625, + "logps/rejected": -531.5496826171875, + "loss": 0.2001, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 2.016098976135254, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.6434473991394043, + "rewards/student_margin": 4.659546375274658, + "rewards/teacher_margin": 0.0, + "step": 880 + }, + { + "epoch": 0.52, + "grad_norm": 5.5, + "learning_rate": 2.77882939548964e-06, + "logits/chosen": -0.4587094187736511, + "logits/rejected": 0.0055079348385334015, + "logps/chosen": -223.0956573486328, + "logps/rejected": -570.2337036132812, + "loss": 0.2067, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 1.6644607782363892, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.6545493602752686, + "rewards/student_margin": 4.3190107345581055, + "rewards/teacher_margin": 0.0, + "step": 890 + }, + { + "epoch": 0.52, + "grad_norm": 20.625, + "learning_rate": 2.7282903117215724e-06, + "logits/chosen": -0.26384082436561584, + "logits/rejected": 0.112058125436306, + "logps/chosen": -246.3653106689453, + "logps/rejected": -487.1790466308594, + "loss": 0.1951, + "rewards/accuracies": 0.9333332777023315, + "rewards/chosen": 1.8497905731201172, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.262843608856201, + "rewards/student_margin": 4.112634658813477, + "rewards/teacher_margin": 0.0, + "step": 900 + }, + { + "epoch": 0.53, + "grad_norm": 17.75, + "learning_rate": 2.6776569623390615e-06, + "logits/chosen": -0.3648647367954254, + "logits/rejected": 0.17572906613349915, + "logps/chosen": -262.3744201660156, + "logps/rejected": -555.7116088867188, + "loss": 0.171, + "rewards/accuracies": 0.8666666746139526, + "rewards/chosen": 1.8194630146026611, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.9717178344726562, + "rewards/student_margin": 4.791180610656738, + "rewards/teacher_margin": 0.0, + "step": 910 + }, + { + "epoch": 0.54, + "grad_norm": 12.9375, + "learning_rate": 2.626950254860059e-06, + "logits/chosen": -0.5600097179412842, + "logits/rejected": 0.1498565971851349, + "logps/chosen": -180.4383544921875, + "logps/rejected": -550.03857421875, + "loss": 0.2459, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 2.4329237937927246, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.4012227058410645, + "rewards/student_margin": 4.834146499633789, + "rewards/teacher_margin": 0.0, + "step": 920 + }, + { + "epoch": 0.54, + "grad_norm": 15.875, + "learning_rate": 2.576191127093538e-06, + "logits/chosen": -0.3777901232242584, + "logits/rejected": 0.046824149787425995, + "logps/chosen": -214.46310424804688, + "logps/rejected": -476.24627685546875, + "loss": 0.1634, + "rewards/accuracies": 0.8333333134651184, + "rewards/chosen": 1.7608407735824585, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.415520191192627, + "rewards/student_margin": 4.176361083984375, + "rewards/teacher_margin": 0.0, + "step": 930 + }, + { + "epoch": 0.55, + "grad_norm": 26.0, + "learning_rate": 2.5254005384938495e-06, + "logits/chosen": -0.49377965927124023, + "logits/rejected": 0.03530919924378395, + "logps/chosen": -214.46923828125, + "logps/rejected": -443.4176330566406, + "loss": 0.231, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 1.699275255203247, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.4718739986419678, + "rewards/student_margin": 4.171149253845215, + "rewards/teacher_margin": 0.0, + "step": 940 + }, + { + "epoch": 0.55, + "grad_norm": 9.5, + "learning_rate": 2.474599461506151e-06, + "logits/chosen": -0.6275911331176758, + "logits/rejected": 0.08756984770298004, + "logps/chosen": -198.088623046875, + "logps/rejected": -556.3299560546875, + "loss": 0.1402, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.5414060354232788, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -3.8458423614501953, + "rewards/student_margin": 5.387248516082764, + "rewards/teacher_margin": 0.0, + "step": 950 + }, + { + "epoch": 0.56, + "grad_norm": 5.71875, + "learning_rate": 2.4238088729064625e-06, + "logits/chosen": -0.031927138566970825, + "logits/rejected": -0.28637224435806274, + "logps/chosen": -263.4320373535156, + "logps/rejected": -391.6306457519531, + "loss": 0.1843, + "rewards/accuracies": 0.966666579246521, + "rewards/chosen": 1.5371907949447632, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.255502939224243, + "rewards/student_margin": 3.7926928997039795, + "rewards/teacher_margin": 0.0, + "step": 960 + }, + { + "epoch": 0.56, + "grad_norm": 62.25, + "learning_rate": 2.3730497451399416e-06, + "logits/chosen": -0.39691585302352905, + "logits/rejected": 0.12687797844409943, + "logps/chosen": -187.50283813476562, + "logps/rejected": -645.21533203125, + "loss": 0.2059, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.4940803050994873, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -4.447799205780029, + "rewards/student_margin": 5.941879749298096, + "rewards/teacher_margin": 0.0, + "step": 970 + }, + { + "epoch": 0.57, + "grad_norm": 17.0, + "learning_rate": 2.3223430376609397e-06, + "logits/chosen": -0.31980419158935547, + "logits/rejected": 0.05356071516871452, + "logps/chosen": -222.4664764404297, + "logps/rejected": -434.98681640625, + "loss": 0.2345, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 1.9460868835449219, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.792879104614258, + "rewards/student_margin": 4.7389655113220215, + "rewards/teacher_margin": 0.0, + "step": 980 + }, + { + "epoch": 0.58, + "grad_norm": 8.9375, + "learning_rate": 2.2717096882784275e-06, + "logits/chosen": -0.5998212099075317, + "logits/rejected": 0.12909957766532898, + "logps/chosen": -176.48876953125, + "logps/rejected": -488.4345703125, + "loss": 0.2036, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 2.0280308723449707, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.9598499536514282, + "rewards/student_margin": 3.9878811836242676, + "rewards/teacher_margin": 0.0, + "step": 990 + }, + { + "epoch": 0.58, + "grad_norm": 14.6875, + "learning_rate": 2.2211706045103607e-06, + "logits/chosen": -0.49985045194625854, + "logits/rejected": 0.06370660662651062, + "logps/chosen": -219.44418334960938, + "logps/rejected": -586.0706787109375, + "loss": 0.1749, + "rewards/accuracies": 1.0, + "rewards/chosen": 2.355344295501709, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.49812650680542, + "rewards/student_margin": 4.853470802307129, + "rewards/teacher_margin": 0.0, + "step": 1000 + }, + { + "epoch": 0.59, + "grad_norm": 12.5625, + "learning_rate": 2.1707466549505428e-06, + "logits/chosen": -0.5008007287979126, + "logits/rejected": -0.06781232357025146, + "logps/chosen": -212.249267578125, + "logps/rejected": -466.44964599609375, + "loss": 0.1597, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 1.9009497165679932, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.340704917907715, + "rewards/student_margin": 4.241654872894287, + "rewards/teacher_margin": 0.0, + "step": 1010 + }, + { + "epoch": 0.59, + "grad_norm": 7.78125, + "learning_rate": 2.120458660651572e-06, + "logits/chosen": -0.5323157906532288, + "logits/rejected": -0.03192181885242462, + "logps/chosen": -211.4275360107422, + "logps/rejected": -644.5025024414062, + "loss": 0.1576, + "rewards/accuracies": 1.0, + "rewards/chosen": 2.192241668701172, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -3.6768863201141357, + "rewards/student_margin": 5.869128227233887, + "rewards/teacher_margin": 0.0, + "step": 1020 + }, + { + "epoch": 0.6, + "grad_norm": 27.375, + "learning_rate": 2.070327386527412e-06, + "logits/chosen": -0.3082345128059387, + "logits/rejected": -0.09356869012117386, + "logps/chosen": -276.2046203613281, + "logps/rejected": -543.6260375976562, + "loss": 0.2272, + "rewards/accuracies": 0.8666666746139526, + "rewards/chosen": 1.5081729888916016, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.5921823978424072, + "rewards/student_margin": 4.100355625152588, + "rewards/teacher_margin": 0.0, + "step": 1030 + }, + { + "epoch": 0.61, + "grad_norm": 13.25, + "learning_rate": 2.020373532779147e-06, + "logits/chosen": -0.44038814306259155, + "logits/rejected": 0.11050190776586533, + "logps/chosen": -222.35556030273438, + "logps/rejected": -402.10443115234375, + "loss": 0.2101, + "rewards/accuracies": 0.966666579246521, + "rewards/chosen": 2.202882766723633, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.7409263849258423, + "rewards/student_margin": 3.9438090324401855, + "rewards/teacher_margin": 0.0, + "step": 1040 + }, + { + "epoch": 0.61, + "grad_norm": 11.1875, + "learning_rate": 1.9706177263474567e-06, + "logits/chosen": -0.4345908761024475, + "logits/rejected": 0.14452026784420013, + "logps/chosen": -266.3000793457031, + "logps/rejected": -518.9310913085938, + "loss": 0.1659, + "rewards/accuracies": 0.966666579246521, + "rewards/chosen": 2.2457222938537598, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.8994420766830444, + "rewards/student_margin": 4.145164489746094, + "rewards/teacher_margin": 0.0, + "step": 1050 + }, + { + "epoch": 0.62, + "grad_norm": 26.0, + "learning_rate": 1.9210805123953397e-06, + "logits/chosen": -0.47684893012046814, + "logits/rejected": 0.010342525318264961, + "logps/chosen": -221.49465942382812, + "logps/rejected": -640.6463623046875, + "loss": 0.1841, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.7234092950820923, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.9107723236083984, + "rewards/student_margin": 4.634181499481201, + "rewards/teacher_margin": 0.0, + "step": 1060 + }, + { + "epoch": 0.62, + "grad_norm": 19.25, + "learning_rate": 1.8717823458246082e-06, + "logits/chosen": -0.4791850447654724, + "logits/rejected": 0.2243192493915558, + "logps/chosen": -210.71337890625, + "logps/rejected": -536.1077880859375, + "loss": 0.1652, + "rewards/accuracies": 0.9333332777023315, + "rewards/chosen": 1.8451074361801147, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.979658842086792, + "rewards/student_margin": 4.824766159057617, + "rewards/teacher_margin": 0.0, + "step": 1070 + }, + { + "epoch": 0.63, + "grad_norm": 22.625, + "learning_rate": 1.8227435828296525e-06, + "logits/chosen": -0.7112604379653931, + "logits/rejected": 0.23266398906707764, + "logps/chosen": -200.66549682617188, + "logps/rejected": -564.5877685546875, + "loss": 0.1927, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 2.2678866386413574, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -3.094907283782959, + "rewards/student_margin": 5.362793922424316, + "rewards/teacher_margin": 0.0, + "step": 1080 + }, + { + "epoch": 0.63, + "grad_norm": 17.625, + "learning_rate": 1.7739844724919636e-06, + "logits/chosen": -0.18364885449409485, + "logits/rejected": -0.07437516748905182, + "logps/chosen": -236.36068725585938, + "logps/rejected": -423.6537170410156, + "loss": 0.1789, + "rewards/accuracies": 0.8666666150093079, + "rewards/chosen": 2.1759696006774902, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.076977014541626, + "rewards/student_margin": 4.252946376800537, + "rewards/teacher_margin": 0.0, + "step": 1090 + }, + { + "epoch": 0.64, + "grad_norm": 18.375, + "learning_rate": 1.72552514841888e-06, + "logits/chosen": -0.5487465858459473, + "logits/rejected": -0.06100478023290634, + "logps/chosen": -208.61264038085938, + "logps/rejected": -503.5950622558594, + "loss": 0.1366, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 1.9314453601837158, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.7049808502197266, + "rewards/student_margin": 4.636425971984863, + "rewards/teacher_margin": 0.0, + "step": 1100 + }, + { + "epoch": 0.65, + "grad_norm": 14.125, + "learning_rate": 1.6773856204300259e-06, + "logits/chosen": -0.48537248373031616, + "logits/rejected": 0.20526167750358582, + "logps/chosen": -204.2689208984375, + "logps/rejected": -438.7137145996094, + "loss": 0.1781, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.7250961065292358, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.3652234077453613, + "rewards/student_margin": 4.090319633483887, + "rewards/teacher_margin": 0.0, + "step": 1110 + }, + { + "epoch": 0.65, + "grad_norm": 20.75, + "learning_rate": 1.6295857662948535e-06, + "logits/chosen": -0.41923952102661133, + "logits/rejected": 0.08341416716575623, + "logps/chosen": -190.0411376953125, + "logps/rejected": -548.9283447265625, + "loss": 0.19, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.4958313703536987, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.9623820781707764, + "rewards/student_margin": 4.458213806152344, + "rewards/teacher_margin": 0.0, + "step": 1120 + }, + { + "epoch": 0.66, + "grad_norm": 20.75, + "learning_rate": 1.5821453235247217e-06, + "logits/chosen": -0.053016532212495804, + "logits/rejected": 0.08401618152856827, + "logps/chosen": -312.3536682128906, + "logps/rejected": -552.4027099609375, + "loss": 0.198, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 1.9978468418121338, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.899606227874756, + "rewards/student_margin": 4.897453784942627, + "rewards/teacher_margin": 0.0, + "step": 1130 + }, + { + "epoch": 0.66, + "grad_norm": 11.375, + "learning_rate": 1.5350838812228796e-06, + "logits/chosen": -0.5477980375289917, + "logits/rejected": 0.18308475613594055, + "logps/chosen": -194.05624389648438, + "logps/rejected": -481.322265625, + "loss": 0.1794, + "rewards/accuracies": 0.8666666150093079, + "rewards/chosen": 2.268648624420166, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.6922143697738647, + "rewards/student_margin": 3.9608635902404785, + "rewards/teacher_margin": 0.0, + "step": 1140 + }, + { + "epoch": 0.67, + "grad_norm": 11.5625, + "learning_rate": 1.4884208719957396e-06, + "logits/chosen": -0.4524906277656555, + "logits/rejected": 0.048188693821430206, + "logps/chosen": -240.8212890625, + "logps/rejected": -457.342041015625, + "loss": 0.1424, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 2.082411527633667, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.1976187229156494, + "rewards/student_margin": 4.280030727386475, + "rewards/teacher_margin": 0.0, + "step": 1150 + }, + { + "epoch": 0.68, + "grad_norm": 23.125, + "learning_rate": 1.4421755639287633e-06, + "logits/chosen": -0.735596239566803, + "logits/rejected": 0.0022054375149309635, + "logps/chosen": -182.4052734375, + "logps/rejected": -511.3103942871094, + "loss": 0.1992, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 2.068901538848877, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.3740670680999756, + "rewards/student_margin": 4.442968845367432, + "rewards/teacher_margin": 0.0, + "step": 1160 + }, + { + "epoch": 0.68, + "grad_norm": 7.28125, + "learning_rate": 1.39636705263029e-06, + "logits/chosen": -0.09597094357013702, + "logits/rejected": -0.051341939717531204, + "logps/chosen": -257.4917297363281, + "logps/rejected": -411.2127990722656, + "loss": 0.1143, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 2.1275782585144043, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.4862334728240967, + "rewards/student_margin": 3.613811492919922, + "rewards/teacher_margin": 0.0, + "step": 1170 + }, + { + "epoch": 0.69, + "grad_norm": 4.875, + "learning_rate": 1.3510142533465764e-06, + "logits/chosen": -0.35563239455223083, + "logits/rejected": 0.13452188670635223, + "logps/chosen": -267.82806396484375, + "logps/rejected": -522.6319580078125, + "loss": 0.1118, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 2.2255101203918457, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.5478758811950684, + "rewards/student_margin": 4.773386478424072, + "rewards/teacher_margin": 0.0, + "step": 1180 + }, + { + "epoch": 0.69, + "grad_norm": 28.5, + "learning_rate": 1.306135893151318e-06, + "logits/chosen": -0.37067538499832153, + "logits/rejected": 0.1407340168952942, + "logps/chosen": -292.832763671875, + "logps/rejected": -464.6031799316406, + "loss": 0.2175, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 2.117187976837158, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.6492514610290527, + "rewards/student_margin": 4.766438961029053, + "rewards/teacher_margin": 0.0, + "step": 1190 + }, + { + "epoch": 0.7, + "grad_norm": 11.9375, + "learning_rate": 1.2617505032128618e-06, + "logits/chosen": -0.2240295708179474, + "logits/rejected": 0.03423571586608887, + "logps/chosen": -241.01223754882812, + "logps/rejected": -511.5538635253906, + "loss": 0.1503, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 1.904809594154358, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.1720213890075684, + "rewards/student_margin": 4.076830863952637, + "rewards/teacher_margin": 0.0, + "step": 1200 + }, + { + "epoch": 0.7, + "grad_norm": 22.5, + "learning_rate": 1.2178764111423258e-06, + "logits/chosen": -0.453156054019928, + "logits/rejected": -0.015243321657180786, + "logps/chosen": -261.2116394042969, + "logps/rejected": -541.9698486328125, + "loss": 0.1609, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 2.2009761333465576, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.577160120010376, + "rewards/student_margin": 4.778136253356934, + "rewards/teacher_margin": 0.0, + "step": 1210 + }, + { + "epoch": 0.71, + "grad_norm": 11.625, + "learning_rate": 1.1745317334257577e-06, + "logits/chosen": -0.3408849537372589, + "logits/rejected": 0.10163626819849014, + "logps/chosen": -180.9221649169922, + "logps/rejected": -456.72412109375, + "loss": 0.1747, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.4770768880844116, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.7443153858184814, + "rewards/student_margin": 4.221392631530762, + "rewards/teacher_margin": 0.0, + "step": 1220 + }, + { + "epoch": 0.72, + "grad_norm": 18.0, + "learning_rate": 1.1317343679434797e-06, + "logits/chosen": -0.39155808091163635, + "logits/rejected": 0.03318040445446968, + "logps/chosen": -243.98617553710938, + "logps/rejected": -480.6688537597656, + "loss": 0.1238, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 1.888183355331421, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.6680731773376465, + "rewards/student_margin": 4.556256294250488, + "rewards/teacher_margin": 0.0, + "step": 1230 + }, + { + "epoch": 0.72, + "grad_norm": 10.1875, + "learning_rate": 1.0895019865796974e-06, + "logits/chosen": -0.35370463132858276, + "logits/rejected": -0.06857095658779144, + "logps/chosen": -214.5222930908203, + "logps/rejected": -411.405517578125, + "loss": 0.2088, + "rewards/accuracies": 0.8666666150093079, + "rewards/chosen": 1.605942726135254, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.65793776512146, + "rewards/student_margin": 3.263880968093872, + "rewards/teacher_margin": 0.0, + "step": 1240 + }, + { + "epoch": 0.73, + "grad_norm": 7.96875, + "learning_rate": 1.0478520279254232e-06, + "logits/chosen": -0.4441911578178406, + "logits/rejected": 0.0003550305846147239, + "logps/chosen": -173.74652099609375, + "logps/rejected": -404.5919494628906, + "loss": 0.1417, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.5375311374664307, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.0653061866760254, + "rewards/student_margin": 3.602837324142456, + "rewards/teacher_margin": 0.0, + "step": 1250 + }, + { + "epoch": 0.73, + "grad_norm": 11.0, + "learning_rate": 1.0068016900777411e-06, + "logits/chosen": -0.14587707817554474, + "logits/rejected": 0.1403418779373169, + "logps/chosen": -265.76385498046875, + "logps/rejected": -488.88360595703125, + "loss": 0.1327, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 1.8325278759002686, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.28810453414917, + "rewards/student_margin": 4.120633125305176, + "rewards/teacher_margin": 0.0, + "step": 1260 + }, + { + "epoch": 0.74, + "grad_norm": 15.875, + "learning_rate": 9.663679235383663e-07, + "logits/chosen": -0.4032668173313141, + "logits/rejected": 0.15200474858283997, + "logps/chosen": -253.056396484375, + "logps/rejected": -567.1189575195312, + "loss": 0.1393, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 1.7767219543457031, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.897172451019287, + "rewards/student_margin": 4.67389440536499, + "rewards/teacher_margin": 0.0, + "step": 1270 + }, + { + "epoch": 0.75, + "grad_norm": 13.25, + "learning_rate": 9.265674242144462e-07, + "logits/chosen": -0.5168894529342651, + "logits/rejected": 0.1268787384033203, + "logps/chosen": -241.34536743164062, + "logps/rejected": -535.0253295898438, + "loss": 0.167, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 2.2057933807373047, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.5597376823425293, + "rewards/student_margin": 4.765531539916992, + "rewards/teacher_margin": 0.0, + "step": 1280 + }, + { + "epoch": 0.75, + "grad_norm": 11.375, + "learning_rate": 8.874166265244891e-07, + "logits/chosen": -0.2563822865486145, + "logits/rejected": -0.007131014950573444, + "logps/chosen": -241.32510375976562, + "logps/rejected": -438.27783203125, + "loss": 0.1496, + "rewards/accuracies": 0.966666579246521, + "rewards/chosen": 2.1465556621551514, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.53509783744812, + "rewards/student_margin": 4.681653022766113, + "rewards/teacher_margin": 0.0, + "step": 1290 + }, + { + "epoch": 0.76, + "grad_norm": 5.71875, + "learning_rate": 8.48931696612261e-07, + "logits/chosen": -0.03185834363102913, + "logits/rejected": 0.0335637666285038, + "logps/chosen": -292.47320556640625, + "logps/rejected": -396.5122375488281, + "loss": 0.1623, + "rewards/accuracies": 1.0, + "rewards/chosen": 2.144620656967163, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.0947091579437256, + "rewards/student_margin": 4.239329814910889, + "rewards/teacher_margin": 0.0, + "step": 1300 + }, + { + "epoch": 0.76, + "grad_norm": 11.1875, + "learning_rate": 8.111285256714702e-07, + "logits/chosen": -0.5409881472587585, + "logits/rejected": -0.020885199308395386, + "logps/chosen": -259.0535583496094, + "logps/rejected": -619.1828002929688, + "loss": 0.1752, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 2.047285795211792, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -3.6237800121307373, + "rewards/student_margin": 5.6710662841796875, + "rewards/teacher_margin": 0.0, + "step": 1310 + }, + { + "epoch": 0.77, + "grad_norm": 11.3125, + "learning_rate": 7.740227233839726e-07, + "logits/chosen": -0.23395772278308868, + "logits/rejected": 0.0602710135281086, + "logps/chosen": -236.9947967529297, + "logps/rejected": -399.87115478515625, + "loss": 0.1768, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 1.6983884572982788, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.7023701667785645, + "rewards/student_margin": 4.400758743286133, + "rewards/teacher_margin": 0.0, + "step": 1320 + }, + { + "epoch": 0.77, + "grad_norm": 8.5, + "learning_rate": 7.376296114742231e-07, + "logits/chosen": -0.5401469469070435, + "logits/rejected": 0.14081577956676483, + "logps/chosen": -179.28860473632812, + "logps/rejected": -485.43218994140625, + "loss": 0.1542, + "rewards/accuracies": 1.0, + "rewards/chosen": 2.387885808944702, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.904003620147705, + "rewards/student_margin": 5.291889190673828, + "rewards/teacher_margin": 0.0, + "step": 1330 + }, + { + "epoch": 0.78, + "grad_norm": 8.3125, + "learning_rate": 7.019642173826335e-07, + "logits/chosen": -0.42929285764694214, + "logits/rejected": 0.14102640748023987, + "logps/chosen": -236.9450225830078, + "logps/rejected": -510.2264099121094, + "loss": 0.1831, + "rewards/accuracies": 0.966666579246521, + "rewards/chosen": 2.448880672454834, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.9888415336608887, + "rewards/student_margin": 5.437722206115723, + "rewards/teacher_margin": 0.0, + "step": 1340 + }, + { + "epoch": 0.79, + "grad_norm": 12.125, + "learning_rate": 6.670412680604379e-07, + "logits/chosen": -0.25430089235305786, + "logits/rejected": 0.12993714213371277, + "logps/chosen": -257.48284912109375, + "logps/rejected": -497.19110107421875, + "loss": 0.1467, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 2.0826573371887207, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.528546094894409, + "rewards/student_margin": 4.611203193664551, + "rewards/teacher_margin": 0.0, + "step": 1350 + }, + { + "epoch": 0.79, + "grad_norm": 22.875, + "learning_rate": 6.328751838886496e-07, + "logits/chosen": -0.2653670907020569, + "logits/rejected": 0.04784447327256203, + "logps/chosen": -231.7043914794922, + "logps/rejected": -405.35430908203125, + "loss": 0.1555, + "rewards/accuracies": 1.0, + "rewards/chosen": 2.0900282859802246, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.961517572402954, + "rewards/student_margin": 4.0515456199646, + "rewards/teacher_margin": 0.0, + "step": 1360 + }, + { + "epoch": 0.8, + "grad_norm": 12.0625, + "learning_rate": 5.994800727235894e-07, + "logits/chosen": -0.6459251642227173, + "logits/rejected": -0.028898591175675392, + "logps/chosen": -167.8853759765625, + "logps/rejected": -490.18621826171875, + "loss": 0.1421, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.746824026107788, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.833235502243042, + "rewards/student_margin": 4.580060005187988, + "rewards/teacher_margin": 0.0, + "step": 1370 + }, + { + "epoch": 0.8, + "grad_norm": 11.4375, + "learning_rate": 5.668697240714782e-07, + "logits/chosen": -0.736299991607666, + "logits/rejected": -0.027988219633698463, + "logps/chosen": -166.20730590820312, + "logps/rejected": -544.7757568359375, + "loss": 0.1785, + "rewards/accuracies": 1.0, + "rewards/chosen": 2.0296387672424316, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.3135805130004883, + "rewards/student_margin": 4.34321928024292, + "rewards/teacher_margin": 0.0, + "step": 1380 + }, + { + "epoch": 0.81, + "grad_norm": 26.0, + "learning_rate": 5.350576033944705e-07, + "logits/chosen": -0.4334324896335602, + "logits/rejected": 0.09274639934301376, + "logps/chosen": -197.7583465576172, + "logps/rejected": -443.3536071777344, + "loss": 0.1698, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 2.787088632583618, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.1186861991882324, + "rewards/student_margin": 4.9057745933532715, + "rewards/teacher_margin": 0.0, + "step": 1390 + }, + { + "epoch": 0.81, + "grad_norm": 12.5, + "learning_rate": 5.040568465504991e-07, + "logits/chosen": -0.23930701613426208, + "logits/rejected": 0.13585281372070312, + "logps/chosen": -247.615478515625, + "logps/rejected": -547.8809814453125, + "loss": 0.2305, + "rewards/accuracies": 0.966666579246521, + "rewards/chosen": 2.4914164543151855, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.0706675052642822, + "rewards/student_margin": 4.562083721160889, + "rewards/teacher_margin": 0.0, + "step": 1400 + }, + { + "epoch": 0.82, + "grad_norm": 6.71875, + "learning_rate": 4.738802543692214e-07, + "logits/chosen": -0.33991652727127075, + "logits/rejected": 0.23731207847595215, + "logps/chosen": -219.88528442382812, + "logps/rejected": -498.65618896484375, + "loss": 0.1918, + "rewards/accuracies": 0.966666579246521, + "rewards/chosen": 2.3377609252929688, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.0855703353881836, + "rewards/student_margin": 4.423331260681152, + "rewards/teacher_margin": 0.0, + "step": 1410 + }, + { + "epoch": 0.83, + "grad_norm": 17.625, + "learning_rate": 4.445402873662932e-07, + "logits/chosen": -0.33060532808303833, + "logits/rejected": 0.07881384342908859, + "logps/chosen": -235.1278839111328, + "logps/rejected": -671.2401123046875, + "loss": 0.1594, + "rewards/accuracies": 0.8666666746139526, + "rewards/chosen": 1.9622802734375, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.477834701538086, + "rewards/student_margin": 4.440115451812744, + "rewards/teacher_margin": 0.0, + "step": 1420 + }, + { + "epoch": 0.83, + "grad_norm": 18.125, + "learning_rate": 4.1604906059818234e-07, + "logits/chosen": -0.5972268581390381, + "logits/rejected": -0.00915351789444685, + "logps/chosen": -217.2305450439453, + "logps/rejected": -514.9454956054688, + "loss": 0.1546, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 1.6480636596679688, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.5352678298950195, + "rewards/student_margin": 4.183331489562988, + "rewards/teacher_margin": 0.0, + "step": 1430 + }, + { + "epoch": 0.84, + "grad_norm": 9.0625, + "learning_rate": 3.8841833865961435e-07, + "logits/chosen": -0.19409717619419098, + "logits/rejected": 0.05707187205553055, + "logps/chosen": -290.650390625, + "logps/rejected": -460.62548828125, + "loss": 0.1939, + "rewards/accuracies": 0.9333332777023315, + "rewards/chosen": 1.422797441482544, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.440934896469116, + "rewards/student_margin": 3.8637325763702393, + "rewards/teacher_margin": 0.0, + "step": 1440 + }, + { + "epoch": 0.84, + "grad_norm": 11.5, + "learning_rate": 3.6165953082574453e-07, + "logits/chosen": -0.34082725644111633, + "logits/rejected": 0.08756278455257416, + "logps/chosen": -258.0528259277344, + "logps/rejected": -627.5899047851562, + "loss": 0.1415, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 1.9923893213272095, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.4780945777893066, + "rewards/student_margin": 4.470483303070068, + "rewards/teacher_margin": 0.0, + "step": 1450 + }, + { + "epoch": 0.85, + "grad_norm": 14.25, + "learning_rate": 3.3578368634103133e-07, + "logits/chosen": -0.5575006008148193, + "logits/rejected": 0.1742563545703888, + "logps/chosen": -224.60537719726562, + "logps/rejected": -587.3863525390625, + "loss": 0.1389, + "rewards/accuracies": 0.8666666746139526, + "rewards/chosen": 1.9285333156585693, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -3.4504916667938232, + "rewards/student_margin": 5.379024982452393, + "rewards/teacher_margin": 0.0, + "step": 1460 + }, + { + "epoch": 0.86, + "grad_norm": 15.1875, + "learning_rate": 3.108014898567921e-07, + "logits/chosen": -0.32235223054885864, + "logits/rejected": 0.23691675066947937, + "logps/chosen": -289.0404052734375, + "logps/rejected": -518.9627685546875, + "loss": 0.1884, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 1.5728108882904053, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.671419143676758, + "rewards/student_margin": 4.244229793548584, + "rewards/teacher_margin": 0.0, + "step": 1470 + }, + { + "epoch": 0.86, + "grad_norm": 12.0, + "learning_rate": 2.867232570192932e-07, + "logits/chosen": -0.4675559103488922, + "logits/rejected": 0.2386462688446045, + "logps/chosen": -238.5657501220703, + "logps/rejected": -556.7369384765625, + "loss": 0.172, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 2.0571231842041016, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.1405348777770996, + "rewards/student_margin": 4.197658061981201, + "rewards/teacher_margin": 0.0, + "step": 1480 + }, + { + "epoch": 0.87, + "grad_norm": 9.6875, + "learning_rate": 2.635589302102143e-07, + "logits/chosen": -0.5631746053695679, + "logits/rejected": 0.1842821091413498, + "logps/chosen": -202.35910034179688, + "logps/rejected": -496.7588806152344, + "loss": 0.1308, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 1.9014427661895752, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.115988254547119, + "rewards/student_margin": 4.017430782318115, + "rewards/teacher_margin": 0.0, + "step": 1490 + }, + { + "epoch": 0.87, + "grad_norm": 12.875, + "learning_rate": 2.413180744412449e-07, + "logits/chosen": -0.45950451493263245, + "logits/rejected": -0.005202120635658503, + "logps/chosen": -199.29319763183594, + "logps/rejected": -393.0441589355469, + "loss": 0.1418, + "rewards/accuracies": 1.0, + "rewards/chosen": 2.295112371444702, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.899126648902893, + "rewards/student_margin": 4.194239139556885, + "rewards/teacher_margin": 0.0, + "step": 1500 + }, + { + "epoch": 0.88, + "grad_norm": 7.25, + "learning_rate": 2.200098734044953e-07, + "logits/chosen": -0.3245978355407715, + "logits/rejected": 0.15231186151504517, + "logps/chosen": -282.319091796875, + "logps/rejected": -564.8612060546875, + "loss": 0.1391, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 1.9999994039535522, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.712496757507324, + "rewards/student_margin": 4.712496757507324, + "rewards/teacher_margin": 0.0, + "step": 1510 + }, + { + "epoch": 0.88, + "grad_norm": 20.5, + "learning_rate": 1.9964312568037102e-07, + "logits/chosen": -0.43997129797935486, + "logits/rejected": 0.11200802028179169, + "logps/chosen": -215.91390991210938, + "logps/rejected": -454.8565979003906, + "loss": 0.2001, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.7922544479370117, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.544246196746826, + "rewards/student_margin": 4.336500644683838, + "rewards/teacher_margin": 0.0, + "step": 1520 + }, + { + "epoch": 0.89, + "grad_norm": 7.9375, + "learning_rate": 1.8022624110446223e-07, + "logits/chosen": -0.23303954303264618, + "logits/rejected": -0.25525903701782227, + "logps/chosen": -233.7492218017578, + "logps/rejected": -434.97943115234375, + "loss": 0.1737, + "rewards/accuracies": 0.9000000953674316, + "rewards/chosen": 2.0419070720672607, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.659355878829956, + "rewards/student_margin": 3.701263427734375, + "rewards/teacher_margin": 0.0, + "step": 1530 + }, + { + "epoch": 0.9, + "grad_norm": 11.3125, + "learning_rate": 1.617672372949558e-07, + "logits/chosen": -0.4891526699066162, + "logits/rejected": 0.0062074740417301655, + "logps/chosen": -184.78028869628906, + "logps/rejected": -495.7847595214844, + "loss": 0.1925, + "rewards/accuracies": 0.8666666150093079, + "rewards/chosen": 1.5070412158966064, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.09743070602417, + "rewards/student_margin": 3.6044719219207764, + "rewards/teacher_margin": 0.0, + "step": 1540 + }, + { + "epoch": 0.9, + "grad_norm": 8.8125, + "learning_rate": 1.442737363420016e-07, + "logits/chosen": -0.47575002908706665, + "logits/rejected": 0.09344211965799332, + "logps/chosen": -215.34390258789062, + "logps/rejected": -447.3370056152344, + "loss": 0.1637, + "rewards/accuracies": 0.9333332777023315, + "rewards/chosen": 1.5486562252044678, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.476374864578247, + "rewards/student_margin": 4.025030612945557, + "rewards/teacher_margin": 0.0, + "step": 1550 + }, + { + "epoch": 0.91, + "grad_norm": 5.09375, + "learning_rate": 1.2775296166039759e-07, + "logits/chosen": -0.34430769085884094, + "logits/rejected": 0.11993454396724701, + "logps/chosen": -198.77247619628906, + "logps/rejected": -414.20648193359375, + "loss": 0.1491, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 2.136995553970337, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.7208058834075928, + "rewards/student_margin": 3.8578014373779297, + "rewards/teacher_margin": 0.0, + "step": 1560 + }, + { + "epoch": 0.91, + "grad_norm": 9.125, + "learning_rate": 1.1221173500690074e-07, + "logits/chosen": -0.43167194724082947, + "logits/rejected": 0.05056245997548103, + "logps/chosen": -202.82408142089844, + "logps/rejected": -457.1270446777344, + "loss": 0.1328, + "rewards/accuracies": 0.966666579246521, + "rewards/chosen": 1.7842384576797485, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -1.8982183933258057, + "rewards/student_margin": 3.6824564933776855, + "rewards/teacher_margin": 0.0, + "step": 1570 + }, + { + "epoch": 0.92, + "grad_norm": 13.4375, + "learning_rate": 9.765647366338615e-08, + "logits/chosen": -0.617282509803772, + "logits/rejected": 0.20323173701763153, + "logps/chosen": -215.5322723388672, + "logps/rejected": -516.0162353515625, + "loss": 0.1082, + "rewards/accuracies": 1.0, + "rewards/chosen": 2.109098196029663, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.6933035850524902, + "rewards/student_margin": 4.802401542663574, + "rewards/teacher_margin": 0.0, + "step": 1580 + }, + { + "epoch": 0.93, + "grad_norm": 13.1875, + "learning_rate": 8.409318778702436e-08, + "logits/chosen": -0.24509771168231964, + "logits/rejected": 0.05450732633471489, + "logps/chosen": -244.3855438232422, + "logps/rejected": -399.8455505371094, + "loss": 0.1269, + "rewards/accuracies": 0.966666579246521, + "rewards/chosen": 2.09488582611084, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.1163201332092285, + "rewards/student_margin": 4.21120548248291, + "rewards/teacher_margin": 0.0, + "step": 1590 + }, + { + "epoch": 0.93, + "grad_norm": 12.625, + "learning_rate": 7.152747792856795e-08, + "logits/chosen": -0.5267202854156494, + "logits/rejected": 0.10546509921550751, + "logps/chosen": -246.46163940429688, + "logps/rejected": -488.86358642578125, + "loss": 0.1929, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 2.5057289600372314, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.4191794395446777, + "rewards/student_margin": 4.924908638000488, + "rewards/teacher_margin": 0.0, + "step": 1600 + }, + { + "epoch": 0.94, + "grad_norm": 8.8125, + "learning_rate": 5.99645327197712e-08, + "logits/chosen": -0.5035473108291626, + "logits/rejected": 0.06472502648830414, + "logps/chosen": -175.3248748779297, + "logps/rejected": -475.020263671875, + "loss": 0.1096, + "rewards/accuracies": 0.9333332777023315, + "rewards/chosen": 2.2377569675445557, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.6387767791748047, + "rewards/student_margin": 4.8765339851379395, + "rewards/teacher_margin": 0.0, + "step": 1610 + }, + { + "epoch": 0.94, + "grad_norm": 14.625, + "learning_rate": 4.94091267309027e-08, + "logits/chosen": -0.5968758463859558, + "logits/rejected": 0.08792127668857574, + "logps/chosen": -183.85728454589844, + "logps/rejected": -632.9024658203125, + "loss": 0.1953, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 2.25777530670166, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.826479434967041, + "rewards/student_margin": 5.084254741668701, + "rewards/teacher_margin": 0.0, + "step": 1620 + }, + { + "epoch": 0.95, + "grad_norm": 14.125, + "learning_rate": 3.9865618499229094e-08, + "logits/chosen": -0.41523703932762146, + "logits/rejected": -0.0010458737378939986, + "logps/chosen": -246.58242797851562, + "logps/rejected": -524.6209716796875, + "loss": 0.1439, + "rewards/accuracies": 0.9333332777023315, + "rewards/chosen": 1.9719102382659912, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.955763339996338, + "rewards/student_margin": 4.927673816680908, + "rewards/teacher_margin": 0.0, + "step": 1630 + }, + { + "epoch": 0.95, + "grad_norm": 7.59375, + "learning_rate": 3.1337948729288633e-08, + "logits/chosen": -0.45982232689857483, + "logits/rejected": 0.08918474614620209, + "logps/chosen": -246.1114501953125, + "logps/rejected": -527.5103149414062, + "loss": 0.1411, + "rewards/accuracies": 1.0, + "rewards/chosen": 2.207768440246582, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.371577501296997, + "rewards/student_margin": 4.579345703125, + "rewards/teacher_margin": 0.0, + "step": 1640 + }, + { + "epoch": 0.96, + "grad_norm": 12.5625, + "learning_rate": 2.3829638665695544e-08, + "logits/chosen": -0.5832090377807617, + "logits/rejected": 0.012857144698500633, + "logps/chosen": -177.62669372558594, + "logps/rejected": -534.47314453125, + "loss": 0.1282, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 1.428046703338623, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.5331838130950928, + "rewards/student_margin": 3.961230516433716, + "rewards/teacher_margin": 0.0, + "step": 1650 + }, + { + "epoch": 0.97, + "grad_norm": 9.6875, + "learning_rate": 1.734378863914782e-08, + "logits/chosen": -0.4249907433986664, + "logits/rejected": 0.0664374977350235, + "logps/chosen": -245.0904083251953, + "logps/rejected": -566.2081909179688, + "loss": 0.1449, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 2.0259430408477783, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.927861452102661, + "rewards/student_margin": 4.953804969787598, + "rewards/teacher_margin": 0.0, + "step": 1660 + }, + { + "epoch": 0.97, + "grad_norm": 15.75, + "learning_rate": 1.1883076786240177e-08, + "logits/chosen": -0.5717610716819763, + "logits/rejected": 0.050547920167446136, + "logps/chosen": -175.99942016601562, + "logps/rejected": -478.18524169921875, + "loss": 0.1616, + "rewards/accuracies": 0.9333333969116211, + "rewards/chosen": 1.8787494897842407, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.2613539695739746, + "rewards/student_margin": 4.140103340148926, + "rewards/teacher_margin": 0.0, + "step": 1670 + }, + { + "epoch": 0.98, + "grad_norm": 25.25, + "learning_rate": 7.449757943606972e-09, + "logits/chosen": -0.5395152568817139, + "logits/rejected": 0.1687006652355194, + "logps/chosen": -226.3690643310547, + "logps/rejected": -453.1436462402344, + "loss": 0.1588, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.7440685033798218, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.6785857677459717, + "rewards/student_margin": 4.422654151916504, + "rewards/teacher_margin": 0.0, + "step": 1680 + }, + { + "epoch": 0.98, + "grad_norm": 16.0, + "learning_rate": 4.045662716855303e-09, + "logits/chosen": -0.2058514803647995, + "logits/rejected": 0.07757274806499481, + "logps/chosen": -237.3645477294922, + "logps/rejected": -580.2776489257812, + "loss": 0.1531, + "rewards/accuracies": 0.8999999761581421, + "rewards/chosen": 1.9641157388687134, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.1610915660858154, + "rewards/student_margin": 4.12520694732666, + "rewards/teacher_margin": 0.0, + "step": 1690 + }, + { + "epoch": 0.99, + "grad_norm": 12.625, + "learning_rate": 1.6721967246721704e-09, + "logits/chosen": -0.36075156927108765, + "logits/rejected": 0.044158972799777985, + "logps/chosen": -210.068359375, + "logps/rejected": -480.38726806640625, + "loss": 0.1368, + "rewards/accuracies": 0.9333332777023315, + "rewards/chosen": 1.983689308166504, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.8466620445251465, + "rewards/student_margin": 4.830351829528809, + "rewards/teacher_margin": 0.0, + "step": 1700 + }, + { + "epoch": 1.0, + "grad_norm": 10.6875, + "learning_rate": 3.30340018415698e-10, + "logits/chosen": -0.4800366461277008, + "logits/rejected": 0.04889531061053276, + "logps/chosen": -193.82672119140625, + "logps/rejected": -558.3578491210938, + "loss": 0.1185, + "rewards/accuracies": 0.9666666984558105, + "rewards/chosen": 2.046447515487671, + "rewards/diff": 0.0, + "rewards/diff_abs": 0.0, + "rewards/rejected": -2.4438791275024414, + "rewards/student_margin": 4.490326881408691, + "rewards/teacher_margin": 0.0, + "step": 1710 + }, + { + "epoch": 1.0, + "step": 1718, + "total_flos": 0.0, + "train_loss": 0.24084537111144683, + "train_runtime": 1647.7299, + "train_samples_per_second": 25.016, + "train_steps_per_second": 1.043 + } + ], + "logging_steps": 10, + "max_steps": 1718, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 400, + "total_flos": 0.0, + "train_batch_size": 3, + "trial_name": null, + "trial_params": null +}