{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 2421, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 61.79818261705088, "learning_rate": 2.05761316872428e-09, "logits/chosen": -3.5, "logits/rejected": -1.4140625, "logps/chosen": -262.0, "logps/rejected": -788.0, "loss": 0.6914, "rewards/accuracies": 0.0, "rewards/chosen": 0.0, "rewards/margins": 0.0, "rewards/rejected": 0.0, "step": 1 }, { "epoch": 0.01, "grad_norm": 71.93619140446947, "learning_rate": 2.0576131687242796e-08, "logits/chosen": -1.9140625, "logits/rejected": -3.0625, "logps/chosen": -648.0, "logps/rejected": -760.0, "loss": 0.6928, "rewards/accuracies": 0.0, "rewards/chosen": -0.033447265625, "rewards/margins": -0.033447265625, "rewards/rejected": 0.0, "step": 10 }, { "epoch": 0.02, "grad_norm": 95.51795034006538, "learning_rate": 4.115226337448559e-08, "logits/chosen": -1.828125, "logits/rejected": -2.671875, "logps/chosen": -484.0, "logps/rejected": -352.0, "loss": 0.69, "rewards/accuracies": 0.20000000298023224, "rewards/chosen": 0.02001953125, "rewards/margins": 0.0, "rewards/rejected": 0.02001953125, "step": 20 }, { "epoch": 0.04, "grad_norm": 83.15780879075649, "learning_rate": 6.172839506172839e-08, "logits/chosen": -1.8671875, "logits/rejected": -1.75, "logps/chosen": -404.0, "logps/rejected": -540.0, "loss": 0.6933, "rewards/accuracies": 0.10000000149011612, "rewards/chosen": -0.0150146484375, "rewards/margins": -0.02001953125, "rewards/rejected": 0.0050048828125, "step": 30 }, { "epoch": 0.05, "grad_norm": 78.74493611311031, "learning_rate": 8.230452674897118e-08, "logits/chosen": -2.015625, "logits/rejected": -2.265625, "logps/chosen": -448.0, "logps/rejected": -600.0, "loss": 0.692, "rewards/accuracies": 0.5, "rewards/chosen": 0.0250244140625, "rewards/margins": 0.06005859375, "rewards/rejected": -0.03515625, "step": 40 }, { "epoch": 0.06, "grad_norm": 84.22871977974292, "learning_rate": 1.02880658436214e-07, "logits/chosen": -1.6640625, "logits/rejected": -2.421875, "logps/chosen": -592.0, "logps/rejected": -580.0, "loss": 0.6922, "rewards/accuracies": 0.10000000149011612, "rewards/chosen": -0.0400390625, "rewards/margins": -0.030029296875, "rewards/rejected": -0.010009765625, "step": 50 }, { "epoch": 0.07, "grad_norm": 57.8386192430028, "learning_rate": 1.2345679012345677e-07, "logits/chosen": -2.03125, "logits/rejected": -2.9375, "logps/chosen": -362.0, "logps/rejected": -368.0, "loss": 0.6872, "rewards/accuracies": 0.10000000149011612, "rewards/chosen": -0.0050048828125, "rewards/margins": -0.044921875, "rewards/rejected": 0.0400390625, "step": 60 }, { "epoch": 0.09, "grad_norm": 108.85645337456651, "learning_rate": 1.4403292181069958e-07, "logits/chosen": -2.140625, "logits/rejected": -2.4375, "logps/chosen": -460.0, "logps/rejected": -468.0, "loss": 0.6888, "rewards/accuracies": 0.5, "rewards/chosen": 0.05517578125, "rewards/margins": 0.0751953125, "rewards/rejected": -0.02001953125, "step": 70 }, { "epoch": 0.1, "grad_norm": 70.2187244330304, "learning_rate": 1.6460905349794237e-07, "logits/chosen": -1.8984375, "logits/rejected": -2.28125, "logps/chosen": -516.0, "logps/rejected": -506.0, "loss": 0.687, "rewards/accuracies": 0.20000000298023224, "rewards/chosen": -0.050048828125, "rewards/margins": -0.03515625, "rewards/rejected": -0.0150146484375, "step": 80 }, { "epoch": 0.11, "grad_norm": 54.744834672425526, "learning_rate": 1.8518518518518516e-07, "logits/chosen": -1.7578125, "logits/rejected": -2.078125, "logps/chosen": -532.0, "logps/rejected": -660.0, "loss": 0.677, "rewards/accuracies": 0.4000000059604645, "rewards/chosen": 0.050048828125, "rewards/margins": 0.06494140625, "rewards/rejected": -0.0150146484375, "step": 90 }, { "epoch": 0.12, "grad_norm": 98.04764626666417, "learning_rate": 2.05761316872428e-07, "logits/chosen": -1.6015625, "logits/rejected": -2.234375, "logps/chosen": -560.0, "logps/rejected": -660.0, "loss": 0.6657, "rewards/accuracies": 0.30000001192092896, "rewards/chosen": 0.03515625, "rewards/margins": -0.0301513671875, "rewards/rejected": 0.06494140625, "step": 100 }, { "epoch": 0.14, "grad_norm": 78.67800061099335, "learning_rate": 2.2633744855967078e-07, "logits/chosen": -2.03125, "logits/rejected": -2.703125, "logps/chosen": -484.0, "logps/rejected": -428.0, "loss": 0.6618, "rewards/accuracies": 0.6000000238418579, "rewards/chosen": 0.054931640625, "rewards/margins": 0.06982421875, "rewards/rejected": -0.01507568359375, "step": 110 }, { "epoch": 0.15, "grad_norm": 98.90783538853094, "learning_rate": 2.4691358024691354e-07, "logits/chosen": -1.8984375, "logits/rejected": -1.71875, "logps/chosen": -668.0, "logps/rejected": -752.0, "loss": 0.6576, "rewards/accuracies": 0.5, "rewards/chosen": 0.25, "rewards/margins": 0.205078125, "rewards/rejected": 0.045166015625, "step": 120 }, { "epoch": 0.16, "grad_norm": 183.377786164052, "learning_rate": 2.6748971193415635e-07, "logits/chosen": -2.140625, "logits/rejected": -2.78125, "logps/chosen": -408.0, "logps/rejected": -358.0, "loss": 0.6568, "rewards/accuracies": 0.30000001192092896, "rewards/chosen": -0.09521484375, "rewards/margins": 0.02001953125, "rewards/rejected": -0.115234375, "step": 130 }, { "epoch": 0.17, "grad_norm": 75.16806808811697, "learning_rate": 2.8806584362139917e-07, "logits/chosen": -1.359375, "logits/rejected": -2.21875, "logps/chosen": -668.0, "logps/rejected": -556.0, "loss": 0.6513, "rewards/accuracies": 0.4000000059604645, "rewards/chosen": -0.0703125, "rewards/margins": 0.0400390625, "rewards/rejected": -0.1103515625, "step": 140 }, { "epoch": 0.19, "grad_norm": 59.26706314390388, "learning_rate": 3.086419753086419e-07, "logits/chosen": -2.390625, "logits/rejected": -2.1875, "logps/chosen": -356.0, "logps/rejected": -382.0, "loss": 0.6289, "rewards/accuracies": 0.5, "rewards/chosen": 0.125, "rewards/margins": 0.1357421875, "rewards/rejected": -0.010009765625, "step": 150 }, { "epoch": 0.2, "grad_norm": 103.75204061254972, "learning_rate": 3.2921810699588474e-07, "logits/chosen": -2.46875, "logits/rejected": -1.6171875, "logps/chosen": -398.0, "logps/rejected": -486.0, "loss": 0.6381, "rewards/accuracies": 0.5, "rewards/chosen": -0.0703125, "rewards/margins": 0.08056640625, "rewards/rejected": -0.150390625, "step": 160 }, { "epoch": 0.21, "grad_norm": 113.86192442502461, "learning_rate": 3.4979423868312755e-07, "logits/chosen": -2.3125, "logits/rejected": -1.9140625, "logps/chosen": -450.0, "logps/rejected": -464.0, "loss": 0.6266, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -0.205078125, "rewards/margins": 0.1455078125, "rewards/rejected": -0.3515625, "step": 170 }, { "epoch": 0.22, "grad_norm": 63.20529708150695, "learning_rate": 3.703703703703703e-07, "logits/chosen": -1.8046875, "logits/rejected": -2.8125, "logps/chosen": -744.0, "logps/rejected": -584.0, "loss": 0.6384, "rewards/accuracies": 0.6000000238418579, "rewards/chosen": -0.0849609375, "rewards/margins": 0.1806640625, "rewards/rejected": -0.265625, "step": 180 }, { "epoch": 0.24, "grad_norm": 83.60084870315862, "learning_rate": 3.909465020576131e-07, "logits/chosen": -1.7578125, "logits/rejected": -1.828125, "logps/chosen": -716.0, "logps/rejected": -848.0, "loss": 0.5983, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.1201171875, "rewards/margins": 0.400390625, "rewards/rejected": -0.51953125, "step": 190 }, { "epoch": 0.25, "grad_norm": 62.24290659916015, "learning_rate": 4.11522633744856e-07, "logits/chosen": -2.125, "logits/rejected": -2.28125, "logps/chosen": -536.0, "logps/rejected": -576.0, "loss": 0.5939, "rewards/accuracies": 0.699999988079071, "rewards/chosen": 0.23046875, "rewards/margins": 0.49609375, "rewards/rejected": -0.265625, "step": 200 }, { "epoch": 0.26, "grad_norm": 54.79737482658103, "learning_rate": 4.320987654320987e-07, "logits/chosen": -1.9765625, "logits/rejected": -2.203125, "logps/chosen": -494.0, "logps/rejected": -648.0, "loss": 0.6176, "rewards/accuracies": 1.0, "rewards/chosen": 0.1103515625, "rewards/margins": 0.56640625, "rewards/rejected": -0.455078125, "step": 210 }, { "epoch": 0.27, "grad_norm": 75.36398911732266, "learning_rate": 4.5267489711934156e-07, "logits/chosen": -2.046875, "logits/rejected": -2.03125, "logps/chosen": -512.0, "logps/rejected": -390.0, "loss": 0.6214, "rewards/accuracies": 0.6000000238418579, "rewards/chosen": -0.1806640625, "rewards/margins": -0.0103759765625, "rewards/rejected": -0.169921875, "step": 220 }, { "epoch": 0.29, "grad_norm": 65.42502577887264, "learning_rate": 4.732510288065844e-07, "logits/chosen": -2.40625, "logits/rejected": -2.3125, "logps/chosen": -652.0, "logps/rejected": -504.0, "loss": 0.5634, "rewards/accuracies": 0.5, "rewards/chosen": -0.134765625, "rewards/margins": 0.359375, "rewards/rejected": -0.49609375, "step": 230 }, { "epoch": 0.3, "grad_norm": 101.59685524071844, "learning_rate": 4.938271604938271e-07, "logits/chosen": -1.6953125, "logits/rejected": -1.6640625, "logps/chosen": -704.0, "logps/rejected": -548.0, "loss": 0.5768, "rewards/accuracies": 0.5, "rewards/chosen": -0.37109375, "rewards/margins": 0.1103515625, "rewards/rejected": -0.48046875, "step": 240 }, { "epoch": 0.31, "grad_norm": 56.985215974106104, "learning_rate": 4.999872565682321e-07, "logits/chosen": -1.4921875, "logits/rejected": -2.109375, "logps/chosen": -732.0, "logps/rejected": -496.0, "loss": 0.5665, "rewards/accuracies": 0.5, "rewards/chosen": 0.205078125, "rewards/margins": 0.6328125, "rewards/rejected": -0.42578125, "step": 250 }, { "epoch": 0.32, "grad_norm": 60.61672615933723, "learning_rate": 4.999248428870611e-07, "logits/chosen": -1.890625, "logits/rejected": -3.453125, "logps/chosen": -588.0, "logps/rejected": -430.0, "loss": 0.5575, "rewards/accuracies": 0.4000000059604645, "rewards/chosen": -0.1396484375, "rewards/margins": 0.37109375, "rewards/rejected": -0.51171875, "step": 260 }, { "epoch": 0.33, "grad_norm": 171.89078996410038, "learning_rate": 4.99810431295357e-07, "logits/chosen": -1.640625, "logits/rejected": -2.140625, "logps/chosen": -688.0, "logps/rejected": -488.0, "loss": 0.5472, "rewards/accuracies": 0.4000000059604645, "rewards/chosen": -0.828125, "rewards/margins": 0.005462646484375, "rewards/rejected": -0.8359375, "step": 270 }, { "epoch": 0.35, "grad_norm": 132.22262423790198, "learning_rate": 4.99644045596931e-07, "logits/chosen": -1.9375, "logits/rejected": -2.875, "logps/chosen": -462.0, "logps/rejected": -356.0, "loss": 0.5785, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -0.2451171875, "rewards/margins": 0.474609375, "rewards/rejected": -0.71875, "step": 280 }, { "epoch": 0.36, "grad_norm": 123.26679640601527, "learning_rate": 4.994257204090243e-07, "logits/chosen": -1.609375, "logits/rejected": -3.0, "logps/chosen": -580.0, "logps/rejected": -476.0, "loss": 0.5198, "rewards/accuracies": 0.6000000238418579, "rewards/chosen": -0.42578125, "rewards/margins": 0.5703125, "rewards/rejected": -0.99609375, "step": 290 }, { "epoch": 0.37, "grad_norm": 47.55857001285771, "learning_rate": 4.991555011551073e-07, "logits/chosen": -1.53125, "logits/rejected": -2.203125, "logps/chosen": -624.0, "logps/rejected": -620.0, "loss": 0.5123, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.07470703125, "rewards/margins": 0.88671875, "rewards/rejected": -0.9609375, "step": 300 }, { "epoch": 0.38, "grad_norm": 57.83726580735741, "learning_rate": 4.988334440554274e-07, "logits/chosen": -1.6796875, "logits/rejected": -2.0, "logps/chosen": -800.0, "logps/rejected": -632.0, "loss": 0.5449, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -0.10498046875, "rewards/margins": 0.5234375, "rewards/rejected": -0.6328125, "step": 310 }, { "epoch": 0.4, "grad_norm": 63.9187247476568, "learning_rate": 4.984596161153135e-07, "logits/chosen": -1.96875, "logits/rejected": -1.6015625, "logps/chosen": -612.0, "logps/rejected": -664.0, "loss": 0.5372, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -0.8515625, "rewards/margins": 0.33984375, "rewards/rejected": -1.1875, "step": 320 }, { "epoch": 0.41, "grad_norm": 59.66658958585498, "learning_rate": 4.980340951112345e-07, "logits/chosen": -1.59375, "logits/rejected": -2.4375, "logps/chosen": -652.0, "logps/rejected": -612.0, "loss": 0.5528, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -0.65625, "rewards/margins": 0.58984375, "rewards/rejected": -1.25, "step": 330 }, { "epoch": 0.42, "grad_norm": 51.91101244482548, "learning_rate": 4.975569695746179e-07, "logits/chosen": -1.4140625, "logits/rejected": -2.703125, "logps/chosen": -716.0, "logps/rejected": -496.0, "loss": 0.5169, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.259765625, "rewards/margins": 0.4140625, "rewards/rejected": -0.671875, "step": 340 }, { "epoch": 0.43, "grad_norm": 72.46042088468953, "learning_rate": 4.970283387734303e-07, "logits/chosen": -2.34375, "logits/rejected": -1.96875, "logps/chosen": -552.0, "logps/rejected": -494.0, "loss": 0.5549, "rewards/accuracies": 0.5, "rewards/chosen": -0.546875, "rewards/margins": 0.1943359375, "rewards/rejected": -0.7421875, "step": 350 }, { "epoch": 0.45, "grad_norm": 98.36392274253446, "learning_rate": 4.964483126915245e-07, "logits/chosen": -1.9375, "logits/rejected": -1.6796875, "logps/chosen": -552.0, "logps/rejected": -524.0, "loss": 0.5237, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.375, "rewards/margins": 0.62109375, "rewards/rejected": -0.99609375, "step": 360 }, { "epoch": 0.46, "grad_norm": 35.92451209539994, "learning_rate": 4.958170120057565e-07, "logits/chosen": -1.875, "logits/rejected": -3.46875, "logps/chosen": -438.0, "logps/rejected": -400.0, "loss": 0.514, "rewards/accuracies": 0.6000000238418579, "rewards/chosen": -0.39453125, "rewards/margins": 0.75390625, "rewards/rejected": -1.1484375, "step": 370 }, { "epoch": 0.47, "grad_norm": 43.212042789239504, "learning_rate": 4.951345680608787e-07, "logits/chosen": -1.3984375, "logits/rejected": -2.65625, "logps/chosen": -660.0, "logps/rejected": -616.0, "loss": 0.5045, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -1.140625, "rewards/margins": 0.201171875, "rewards/rejected": -1.34375, "step": 380 }, { "epoch": 0.48, "grad_norm": 55.92901896058768, "learning_rate": 4.944011228422125e-07, "logits/chosen": -1.6875, "logits/rejected": -2.03125, "logps/chosen": -536.0, "logps/rejected": -508.0, "loss": 0.5521, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -0.6953125, "rewards/margins": 0.427734375, "rewards/rejected": -1.125, "step": 390 }, { "epoch": 0.5, "grad_norm": 48.27467634078171, "learning_rate": 4.936168289461084e-07, "logits/chosen": -2.25, "logits/rejected": -2.046875, "logps/chosen": -390.0, "logps/rejected": -452.0, "loss": 0.5116, "rewards/accuracies": 0.6000000238418579, "rewards/chosen": -0.76953125, "rewards/margins": 0.2353515625, "rewards/rejected": -1.0078125, "step": 400 }, { "epoch": 0.51, "grad_norm": 43.39520482895247, "learning_rate": 4.92781849548197e-07, "logits/chosen": -1.6640625, "logits/rejected": -1.6796875, "logps/chosen": -580.0, "logps/rejected": -548.0, "loss": 0.5259, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -1.0234375, "rewards/margins": 0.5859375, "rewards/rejected": -1.609375, "step": 410 }, { "epoch": 0.52, "grad_norm": 53.58876818695228, "learning_rate": 4.918963583694396e-07, "logits/chosen": -1.765625, "logits/rejected": -1.5234375, "logps/chosen": -548.0, "logps/rejected": -596.0, "loss": 0.516, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.93359375, "rewards/margins": 0.59765625, "rewards/rejected": -1.53125, "step": 420 }, { "epoch": 0.53, "grad_norm": 195.21265227218643, "learning_rate": 4.909605396399855e-07, "logits/chosen": -1.9375, "logits/rejected": -1.625, "logps/chosen": -414.0, "logps/rejected": -756.0, "loss": 0.5513, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -0.64453125, "rewards/margins": 1.6640625, "rewards/rejected": -2.3125, "step": 430 }, { "epoch": 0.55, "grad_norm": 46.48860995329084, "learning_rate": 4.899745880608417e-07, "logits/chosen": -1.5, "logits/rejected": -1.921875, "logps/chosen": -668.0, "logps/rejected": -664.0, "loss": 0.7219, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.921875, "rewards/margins": 0.462890625, "rewards/rejected": -1.3828125, "step": 440 }, { "epoch": 0.56, "grad_norm": 42.31847046504421, "learning_rate": 4.889387087633647e-07, "logits/chosen": -1.4296875, "logits/rejected": -2.84375, "logps/chosen": -612.0, "logps/rejected": -600.0, "loss": 0.5171, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.69921875, "rewards/margins": 0.6328125, "rewards/rejected": -1.328125, "step": 450 }, { "epoch": 0.57, "grad_norm": 43.44689835885094, "learning_rate": 4.878531172665815e-07, "logits/chosen": -1.890625, "logits/rejected": -2.75, "logps/chosen": -608.0, "logps/rejected": -544.0, "loss": 0.4904, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.169921875, "rewards/margins": 0.9453125, "rewards/rejected": -1.1171875, "step": 460 }, { "epoch": 0.58, "grad_norm": 47.14528734842805, "learning_rate": 4.867180394323509e-07, "logits/chosen": -1.5546875, "logits/rejected": -2.1875, "logps/chosen": -700.0, "logps/rejected": -732.0, "loss": 0.5632, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.0, "rewards/margins": 1.3359375, "rewards/rejected": -2.328125, "step": 470 }, { "epoch": 0.59, "grad_norm": 43.11262095050651, "learning_rate": 4.855337114183711e-07, "logits/chosen": -1.421875, "logits/rejected": -1.5859375, "logps/chosen": -644.0, "logps/rejected": -692.0, "loss": 0.4951, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.330078125, "rewards/margins": 1.53125, "rewards/rejected": -1.859375, "step": 480 }, { "epoch": 0.61, "grad_norm": 45.38075851262632, "learning_rate": 4.843003796290469e-07, "logits/chosen": -1.8046875, "logits/rejected": -2.15625, "logps/chosen": -772.0, "logps/rejected": -576.0, "loss": 0.5214, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.375, "rewards/margins": 0.71484375, "rewards/rejected": -1.09375, "step": 490 }, { "epoch": 0.62, "grad_norm": 50.08648685280834, "learning_rate": 4.830183006642236e-07, "logits/chosen": -2.40625, "logits/rejected": -1.9765625, "logps/chosen": -444.0, "logps/rejected": -536.0, "loss": 0.5388, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.83203125, "rewards/margins": 0.34765625, "rewards/rejected": -1.1796875, "step": 500 }, { "epoch": 0.62, "eval_logits/chosen": -1.0859375, "eval_logits/rejected": -1.0078125, "eval_logps/chosen": -540.0, "eval_logps/rejected": -620.0, "eval_loss": 0.49058592319488525, "eval_rewards/accuracies": 0.8055555820465088, "eval_rewards/chosen": -0.7578125, "eval_rewards/margins": 0.86328125, "eval_rewards/rejected": -1.625, "eval_runtime": 50.5034, "eval_samples_per_second": 20.791, "eval_steps_per_second": 0.178, "step": 500 }, { "epoch": 0.63, "grad_norm": 45.830329324613906, "learning_rate": 4.816877412658007e-07, "logits/chosen": -1.5546875, "logits/rejected": -1.84375, "logps/chosen": -772.0, "logps/rejected": -616.0, "loss": 0.4923, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.6171875, "rewards/margins": 0.921875, "rewards/rejected": -1.5390625, "step": 510 }, { "epoch": 0.64, "grad_norm": 53.190960939965166, "learning_rate": 4.80308978262235e-07, "logits/chosen": -1.515625, "logits/rejected": -2.609375, "logps/chosen": -548.0, "logps/rejected": -502.0, "loss": 0.4409, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -0.88671875, "rewards/margins": 0.84375, "rewards/rejected": -1.7265625, "step": 520 }, { "epoch": 0.66, "grad_norm": 45.04181332411229, "learning_rate": 4.788822985109449e-07, "logits/chosen": -1.875, "logits/rejected": -1.9140625, "logps/chosen": -432.0, "logps/rejected": -448.0, "loss": 0.5368, "rewards/accuracies": 0.6000000238418579, "rewards/chosen": -0.78515625, "rewards/margins": 0.50390625, "rewards/rejected": -1.2890625, "step": 530 }, { "epoch": 0.67, "grad_norm": 40.85074772821745, "learning_rate": 4.774079988386296e-07, "logits/chosen": -1.9765625, "logits/rejected": -1.640625, "logps/chosen": -436.0, "logps/rejected": -464.0, "loss": 0.5083, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.94140625, "rewards/margins": 0.2197265625, "rewards/rejected": -1.15625, "step": 540 }, { "epoch": 0.68, "grad_norm": 45.70832159940248, "learning_rate": 4.7588638597951173e-07, "logits/chosen": -1.1171875, "logits/rejected": -2.265625, "logps/chosen": -748.0, "logps/rejected": -524.0, "loss": 0.4754, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 0.287109375, "rewards/margins": 1.640625, "rewards/rejected": -1.3515625, "step": 550 }, { "epoch": 0.69, "grad_norm": 59.530448462467945, "learning_rate": 4.7431777651152103e-07, "logits/chosen": -1.4765625, "logits/rejected": -2.09375, "logps/chosen": -464.0, "logps/rejected": -452.0, "loss": 0.5154, "rewards/accuracies": 0.6000000238418579, "rewards/chosen": -0.890625, "rewards/margins": 0.380859375, "rewards/rejected": -1.2734375, "step": 560 }, { "epoch": 0.71, "grad_norm": 60.197735477558666, "learning_rate": 4.727024967904284e-07, "logits/chosen": -1.890625, "logits/rejected": -1.4375, "logps/chosen": -414.0, "logps/rejected": -564.0, "loss": 0.5178, "rewards/accuracies": 0.5, "rewards/chosen": -0.7421875, "rewards/margins": -0.025634765625, "rewards/rejected": -0.71484375, "step": 570 }, { "epoch": 0.72, "grad_norm": 42.26765729774977, "learning_rate": 4.710408828819463e-07, "logits/chosen": -1.609375, "logits/rejected": -2.359375, "logps/chosen": -552.0, "logps/rejected": -524.0, "loss": 0.5026, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.2890625, "rewards/margins": 0.6484375, "rewards/rejected": -1.9375, "step": 580 }, { "epoch": 0.73, "grad_norm": 50.89266004351078, "learning_rate": 4.6933328049180937e-07, "logits/chosen": -1.890625, "logits/rejected": -2.140625, "logps/chosen": -496.0, "logps/rejected": -580.0, "loss": 0.4856, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -0.734375, "rewards/margins": 0.421875, "rewards/rejected": -1.15625, "step": 590 }, { "epoch": 0.74, "grad_norm": 47.93987624657981, "learning_rate": 4.6758004489384815e-07, "logits/chosen": -1.7734375, "logits/rejected": -2.046875, "logps/chosen": -408.0, "logps/rejected": -544.0, "loss": 0.4713, "rewards/accuracies": 0.6000000238418579, "rewards/chosen": -1.2109375, "rewards/margins": 0.23828125, "rewards/rejected": -1.4453125, "step": 600 }, { "epoch": 0.76, "grad_norm": 48.72108104567496, "learning_rate": 4.6578154085607323e-07, "logits/chosen": -2.8125, "logits/rejected": -2.21875, "logps/chosen": -580.0, "logps/rejected": -824.0, "loss": 0.488, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.51953125, "rewards/margins": 1.4609375, "rewards/rejected": -1.984375, "step": 610 }, { "epoch": 0.77, "grad_norm": 55.85530255600864, "learning_rate": 4.639381425647841e-07, "logits/chosen": -1.3984375, "logits/rejected": -1.5703125, "logps/chosen": -502.0, "logps/rejected": -576.0, "loss": 0.5025, "rewards/accuracies": 0.6000000238418579, "rewards/chosen": -0.64453125, "rewards/margins": 1.0, "rewards/rejected": -1.6484375, "step": 620 }, { "epoch": 0.78, "grad_norm": 59.68466050286291, "learning_rate": 4.6205023354671735e-07, "logits/chosen": -1.5390625, "logits/rejected": -2.03125, "logps/chosen": -460.0, "logps/rejected": -446.0, "loss": 0.4818, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.77734375, "rewards/margins": 0.875, "rewards/rejected": -1.6484375, "step": 630 }, { "epoch": 0.79, "grad_norm": 44.610719658440864, "learning_rate": 4.601182065892529e-07, "logits/chosen": -1.5625, "logits/rejected": -1.453125, "logps/chosen": -556.0, "logps/rejected": -460.0, "loss": 0.476, "rewards/accuracies": 0.6000000238418579, "rewards/chosen": -0.91015625, "rewards/margins": 0.53515625, "rewards/rejected": -1.4453125, "step": 640 }, { "epoch": 0.81, "grad_norm": 49.65621035961898, "learning_rate": 4.581424636586928e-07, "logits/chosen": -1.671875, "logits/rejected": -2.203125, "logps/chosen": -456.0, "logps/rejected": -536.0, "loss": 0.5334, "rewards/accuracies": 0.5, "rewards/chosen": -0.83984375, "rewards/margins": 0.2060546875, "rewards/rejected": -1.046875, "step": 650 }, { "epoch": 0.82, "grad_norm": 37.22285340163814, "learning_rate": 4.561234158166305e-07, "logits/chosen": -1.265625, "logits/rejected": -1.3046875, "logps/chosen": -540.0, "logps/rejected": -524.0, "loss": 0.4929, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -0.87109375, "rewards/margins": 0.75, "rewards/rejected": -1.6171875, "step": 660 }, { "epoch": 0.83, "grad_norm": 72.63982166912706, "learning_rate": 4.5406148313442753e-07, "logits/chosen": -2.140625, "logits/rejected": -2.34375, "logps/chosen": -354.0, "logps/rejected": -372.0, "loss": 0.4682, "rewards/accuracies": 0.6000000238418579, "rewards/chosen": -0.98046875, "rewards/margins": 0.7109375, "rewards/rejected": -1.6875, "step": 670 }, { "epoch": 0.84, "grad_norm": 49.88245474539256, "learning_rate": 4.519570946058162e-07, "logits/chosen": -1.3046875, "logits/rejected": -1.6796875, "logps/chosen": -482.0, "logps/rejected": -556.0, "loss": 0.4673, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.84375, "rewards/margins": 1.0234375, "rewards/rejected": -1.8671875, "step": 680 }, { "epoch": 0.86, "grad_norm": 36.17988110201782, "learning_rate": 4.4981068805764545e-07, "logits/chosen": -1.2109375, "logits/rejected": -2.125, "logps/chosen": -700.0, "logps/rejected": -668.0, "loss": 0.4494, "rewards/accuracies": 0.5, "rewards/chosen": -1.4765625, "rewards/margins": -0.234375, "rewards/rejected": -1.2421875, "step": 690 }, { "epoch": 0.87, "grad_norm": 61.21171348805659, "learning_rate": 4.4762271005878913e-07, "logits/chosen": -1.765625, "logits/rejected": -1.9765625, "logps/chosen": -388.0, "logps/rejected": -446.0, "loss": 0.4665, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -1.2578125, "rewards/margins": 0.55859375, "rewards/rejected": -1.8203125, "step": 700 }, { "epoch": 0.88, "grad_norm": 52.27627457515467, "learning_rate": 4.4539361582723586e-07, "logits/chosen": -1.3515625, "logits/rejected": -1.453125, "logps/chosen": -668.0, "logps/rejected": -892.0, "loss": 0.4806, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.8828125, "rewards/margins": 1.46875, "rewards/rejected": -3.34375, "step": 710 }, { "epoch": 0.89, "grad_norm": 65.21759338991524, "learning_rate": 4.431238691353784e-07, "logits/chosen": -1.7265625, "logits/rejected": -1.9921875, "logps/chosen": -486.0, "logps/rejected": -536.0, "loss": 0.4816, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.125, "rewards/margins": 1.75, "rewards/rejected": -2.875, "step": 720 }, { "epoch": 0.9, "grad_norm": 51.27261353179044, "learning_rate": 4.408139422135241e-07, "logits/chosen": -1.4140625, "logits/rejected": -2.078125, "logps/chosen": -462.0, "logps/rejected": -428.0, "loss": 0.4583, "rewards/accuracies": 1.0, "rewards/chosen": -1.2890625, "rewards/margins": 1.859375, "rewards/rejected": -3.15625, "step": 730 }, { "epoch": 0.92, "grad_norm": 64.31719063334275, "learning_rate": 4.3846431565164596e-07, "logits/chosen": -1.390625, "logits/rejected": -1.3515625, "logps/chosen": -584.0, "logps/rejected": -608.0, "loss": 1.897, "rewards/accuracies": 1.0, "rewards/chosen": -1.828125, "rewards/margins": 1.9453125, "rewards/rejected": -3.78125, "step": 740 }, { "epoch": 0.93, "grad_norm": 46.19657074878751, "learning_rate": 4.360754782993929e-07, "logits/chosen": -1.4765625, "logits/rejected": -2.25, "logps/chosen": -484.0, "logps/rejected": -434.0, "loss": 0.465, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.328125, "rewards/margins": 0.7578125, "rewards/rejected": -2.078125, "step": 750 }, { "epoch": 0.94, "grad_norm": 55.55089778313834, "learning_rate": 4.336479271643833e-07, "logits/chosen": -1.4765625, "logits/rejected": -1.4609375, "logps/chosen": -488.0, "logps/rejected": -462.0, "loss": 0.49, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.2265625, "rewards/margins": 0.5390625, "rewards/rejected": -1.765625, "step": 760 }, { "epoch": 0.95, "grad_norm": 77.27127183474478, "learning_rate": 4.3118216730880015e-07, "logits/chosen": -1.5859375, "logits/rejected": -2.515625, "logps/chosen": -552.0, "logps/rejected": -520.0, "loss": 0.4607, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.0703125, "rewards/margins": 1.0859375, "rewards/rejected": -2.15625, "step": 770 }, { "epoch": 0.97, "grad_norm": 61.141458979311125, "learning_rate": 4.286787117443108e-07, "logits/chosen": -1.203125, "logits/rejected": -1.3984375, "logps/chosen": -588.0, "logps/rejected": -580.0, "loss": 0.4711, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.1796875, "rewards/margins": 1.1796875, "rewards/rejected": -2.359375, "step": 780 }, { "epoch": 0.98, "grad_norm": 48.22890637233016, "learning_rate": 4.261380813253328e-07, "logits/chosen": -1.828125, "logits/rejected": -2.578125, "logps/chosen": -448.0, "logps/rejected": -418.0, "loss": 0.4414, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -1.1640625, "rewards/margins": 0.4375, "rewards/rejected": -1.6015625, "step": 790 }, { "epoch": 0.99, "grad_norm": 34.09372820587962, "learning_rate": 4.2356080464066784e-07, "logits/chosen": -1.3671875, "logits/rejected": -3.171875, "logps/chosen": -536.0, "logps/rejected": -464.0, "loss": 0.4354, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.140625, "rewards/margins": 1.5390625, "rewards/rejected": -2.6875, "step": 800 }, { "epoch": 1.0, "grad_norm": 40.13298242319205, "learning_rate": 4.2094741790352673e-07, "logits/chosen": -1.859375, "logits/rejected": -2.203125, "logps/chosen": -516.0, "logps/rejected": -624.0, "loss": 0.3855, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.03125, "rewards/margins": 1.203125, "rewards/rejected": -2.234375, "step": 810 }, { "epoch": 1.02, "grad_norm": 24.15795246510711, "learning_rate": 4.1829846483996813e-07, "logits/chosen": -1.265625, "logits/rejected": -2.015625, "logps/chosen": -884.0, "logps/rejected": -696.0, "loss": 0.2332, "rewards/accuracies": 1.0, "rewards/chosen": -0.154296875, "rewards/margins": 2.125, "rewards/rejected": -2.28125, "step": 820 }, { "epoch": 1.03, "grad_norm": 40.12239181322821, "learning_rate": 4.156144965757735e-07, "logits/chosen": -2.828125, "logits/rejected": -2.40625, "logps/chosen": -416.0, "logps/rejected": -464.0, "loss": 0.264, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.484375, "rewards/margins": 1.7734375, "rewards/rejected": -3.25, "step": 830 }, { "epoch": 1.04, "grad_norm": 22.033239175244603, "learning_rate": 4.128960715217839e-07, "logits/chosen": -1.6015625, "logits/rejected": -2.375, "logps/chosen": -636.0, "logps/rejected": -572.0, "loss": 0.2482, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.7109375, "rewards/margins": 2.125, "rewards/rejected": -2.828125, "step": 840 }, { "epoch": 1.05, "grad_norm": 24.492268202299986, "learning_rate": 4.1014375525771963e-07, "logits/chosen": -1.21875, "logits/rejected": -1.875, "logps/chosen": -636.0, "logps/rejected": -600.0, "loss": 0.2447, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 0.0361328125, "rewards/margins": 2.453125, "rewards/rejected": -2.421875, "step": 850 }, { "epoch": 1.07, "grad_norm": 31.641742017850387, "learning_rate": 4.0735812041450926e-07, "logits/chosen": -1.578125, "logits/rejected": -1.6484375, "logps/chosen": -596.0, "logps/rejected": -596.0, "loss": 0.2428, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.5390625, "rewards/margins": 2.5625, "rewards/rejected": -3.09375, "step": 860 }, { "epoch": 1.08, "grad_norm": 44.35625008242433, "learning_rate": 4.045397465551513e-07, "logits/chosen": -1.5234375, "logits/rejected": -1.796875, "logps/chosen": -520.0, "logps/rejected": -596.0, "loss": 0.2393, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.0859375, "rewards/margins": 1.171875, "rewards/rejected": -2.265625, "step": 870 }, { "epoch": 1.09, "grad_norm": 47.415350341880725, "learning_rate": 4.0168922005413384e-07, "logits/chosen": -1.1171875, "logits/rejected": -1.7109375, "logps/chosen": -692.0, "logps/rejected": -612.0, "loss": 0.2804, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.65625, "rewards/margins": 1.5, "rewards/rejected": -2.15625, "step": 880 }, { "epoch": 1.1, "grad_norm": 31.910462711815494, "learning_rate": 3.988071339754366e-07, "logits/chosen": -1.296875, "logits/rejected": -1.765625, "logps/chosen": -632.0, "logps/rejected": -736.0, "loss": 0.2293, "rewards/accuracies": 1.0, "rewards/chosen": -0.21875, "rewards/margins": 2.296875, "rewards/rejected": -2.515625, "step": 890 }, { "epoch": 1.12, "grad_norm": 27.7996029933054, "learning_rate": 3.958940879491418e-07, "logits/chosen": -1.4296875, "logits/rejected": -1.984375, "logps/chosen": -568.0, "logps/rejected": -498.0, "loss": 0.2693, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.84765625, "rewards/margins": 1.734375, "rewards/rejected": -2.578125, "step": 900 }, { "epoch": 1.13, "grad_norm": 24.76648851906884, "learning_rate": 3.9295068804667823e-07, "logits/chosen": -1.3515625, "logits/rejected": -2.015625, "logps/chosen": -498.0, "logps/rejected": -494.0, "loss": 0.2089, "rewards/accuracies": 1.0, "rewards/chosen": -0.73046875, "rewards/margins": 2.203125, "rewards/rejected": -2.9375, "step": 910 }, { "epoch": 1.14, "grad_norm": 23.922743861511947, "learning_rate": 3.899775466547261e-07, "logits/chosen": -1.625, "logits/rejected": -1.9609375, "logps/chosen": -556.0, "logps/rejected": -532.0, "loss": 0.231, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.86328125, "rewards/margins": 1.8046875, "rewards/rejected": -2.671875, "step": 920 }, { "epoch": 1.15, "grad_norm": 51.66225882393323, "learning_rate": 3.8697528234780674e-07, "logits/chosen": -2.015625, "logits/rejected": -1.6640625, "logps/chosen": -400.0, "logps/rejected": -612.0, "loss": 0.9229, "rewards/accuracies": 1.0, "rewards/chosen": -1.515625, "rewards/margins": 2.140625, "rewards/rejected": -3.640625, "step": 930 }, { "epoch": 1.16, "grad_norm": 38.82610683477862, "learning_rate": 3.839445197595863e-07, "logits/chosen": -1.8671875, "logits/rejected": -2.046875, "logps/chosen": -418.0, "logps/rejected": -378.0, "loss": 0.231, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.71484375, "rewards/margins": 2.03125, "rewards/rejected": -2.75, "step": 940 }, { "epoch": 1.18, "grad_norm": 22.389795986878887, "learning_rate": 3.8088588945291734e-07, "logits/chosen": -1.3515625, "logits/rejected": -2.453125, "logps/chosen": -568.0, "logps/rejected": -496.0, "loss": 0.2115, "rewards/accuracies": 1.0, "rewards/chosen": -0.90625, "rewards/margins": 1.8203125, "rewards/rejected": -2.734375, "step": 950 }, { "epoch": 1.19, "grad_norm": 22.583198913209376, "learning_rate": 3.778000277886483e-07, "logits/chosen": -1.3984375, "logits/rejected": -2.046875, "logps/chosen": -668.0, "logps/rejected": -676.0, "loss": 0.2358, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.326171875, "rewards/margins": 2.78125, "rewards/rejected": -3.109375, "step": 960 }, { "epoch": 1.2, "grad_norm": 21.507764548315773, "learning_rate": 3.746875767932255e-07, "logits/chosen": -1.609375, "logits/rejected": -2.234375, "logps/chosen": -588.0, "logps/rejected": -668.0, "loss": 0.2342, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.453125, "rewards/margins": 2.078125, "rewards/rejected": -3.53125, "step": 970 }, { "epoch": 1.21, "grad_norm": 23.64514805939643, "learning_rate": 3.7154918402511714e-07, "logits/chosen": -1.5078125, "logits/rejected": -3.703125, "logps/chosen": -536.0, "logps/rejected": -436.0, "loss": 0.2427, "rewards/accuracies": 1.0, "rewards/chosen": -0.66015625, "rewards/margins": 2.46875, "rewards/rejected": -3.125, "step": 980 }, { "epoch": 1.23, "grad_norm": 25.279762808321895, "learning_rate": 3.6838550244008573e-07, "logits/chosen": -1.9765625, "logits/rejected": -1.5234375, "logps/chosen": -556.0, "logps/rejected": -592.0, "loss": 0.2198, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.875, "rewards/margins": 1.6015625, "rewards/rejected": -2.484375, "step": 990 }, { "epoch": 1.24, "grad_norm": 23.158055403708982, "learning_rate": 3.651971902553381e-07, "logits/chosen": -1.3984375, "logits/rejected": -2.609375, "logps/chosen": -564.0, "logps/rejected": -486.0, "loss": 0.2367, "rewards/accuracies": 1.0, "rewards/chosen": -0.63671875, "rewards/margins": 2.515625, "rewards/rejected": -3.140625, "step": 1000 }, { "epoch": 1.24, "eval_logits/chosen": -1.0859375, "eval_logits/rejected": -0.9921875, "eval_logps/chosen": -564.0, "eval_logps/rejected": -656.0, "eval_loss": 0.4058724045753479, "eval_rewards/accuracies": 0.8888888955116272, "eval_rewards/chosen": -1.90625, "eval_rewards/margins": 1.5625, "eval_rewards/rejected": -3.453125, "eval_runtime": 50.8495, "eval_samples_per_second": 20.649, "eval_steps_per_second": 0.177, "step": 1000 }, { "epoch": 1.25, "grad_norm": 39.231752515284, "learning_rate": 3.6198491081258066e-07, "logits/chosen": -1.1484375, "logits/rejected": -1.8828125, "logps/chosen": -604.0, "logps/rejected": -494.0, "loss": 0.2308, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -1.265625, "rewards/margins": 1.734375, "rewards/rejected": -3.0, "step": 1010 }, { "epoch": 1.26, "grad_norm": 29.796507962550486, "learning_rate": 3.58749332440008e-07, "logits/chosen": -1.359375, "logits/rejected": -1.6015625, "logps/chosen": -584.0, "logps/rejected": -556.0, "loss": 0.2362, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.2421875, "rewards/margins": 1.640625, "rewards/rejected": -2.875, "step": 1020 }, { "epoch": 1.28, "grad_norm": 21.63597550992159, "learning_rate": 3.55491128313255e-07, "logits/chosen": -2.3125, "logits/rejected": -1.7734375, "logps/chosen": -496.0, "logps/rejected": -548.0, "loss": 0.2107, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.5703125, "rewards/margins": 2.5, "rewards/rejected": -4.0625, "step": 1030 }, { "epoch": 1.29, "grad_norm": 14.702295094535923, "learning_rate": 3.522109763153392e-07, "logits/chosen": -1.546875, "logits/rejected": -1.6328125, "logps/chosen": -442.0, "logps/rejected": -520.0, "loss": 0.2199, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.3984375, "rewards/margins": 1.796875, "rewards/rejected": -3.203125, "step": 1040 }, { "epoch": 1.3, "grad_norm": 21.29596218246951, "learning_rate": 3.489095588956249e-07, "logits/chosen": -1.21875, "logits/rejected": -1.890625, "logps/chosen": -648.0, "logps/rejected": -668.0, "loss": 0.2325, "rewards/accuracies": 1.0, "rewards/chosen": -0.7890625, "rewards/margins": 2.53125, "rewards/rejected": -3.328125, "step": 1050 }, { "epoch": 1.31, "grad_norm": 52.09643118168795, "learning_rate": 3.455875629278363e-07, "logits/chosen": -1.5390625, "logits/rejected": -2.140625, "logps/chosen": -540.0, "logps/rejected": -536.0, "loss": 0.2484, "rewards/accuracies": 1.0, "rewards/chosen": -1.2578125, "rewards/margins": 2.265625, "rewards/rejected": -3.515625, "step": 1060 }, { "epoch": 1.33, "grad_norm": 35.57618950847242, "learning_rate": 3.4224567956715085e-07, "logits/chosen": -1.671875, "logits/rejected": -1.8515625, "logps/chosen": -536.0, "logps/rejected": -564.0, "loss": 0.2062, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.703125, "rewards/margins": 1.8203125, "rewards/rejected": -2.515625, "step": 1070 }, { "epoch": 1.34, "grad_norm": 20.490288705612254, "learning_rate": 3.388846041064012e-07, "logits/chosen": -1.671875, "logits/rejected": -2.25, "logps/chosen": -796.0, "logps/rejected": -604.0, "loss": 0.2027, "rewards/accuracies": 1.0, "rewards/chosen": -1.0078125, "rewards/margins": 1.8125, "rewards/rejected": -2.828125, "step": 1080 }, { "epoch": 1.35, "grad_norm": 25.587126609494646, "learning_rate": 3.355050358314172e-07, "logits/chosen": -1.125, "logits/rejected": -1.078125, "logps/chosen": -728.0, "logps/rejected": -732.0, "loss": 0.2549, "rewards/accuracies": 1.0, "rewards/chosen": -1.3203125, "rewards/margins": 2.125, "rewards/rejected": -3.4375, "step": 1090 }, { "epoch": 1.36, "grad_norm": 32.538109074719216, "learning_rate": 3.321076778755358e-07, "logits/chosen": -1.8359375, "logits/rejected": -1.3125, "logps/chosen": -572.0, "logps/rejected": -568.0, "loss": 0.2117, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.359375, "rewards/margins": 1.90625, "rewards/rejected": -3.265625, "step": 1100 }, { "epoch": 1.38, "grad_norm": 19.634746831169892, "learning_rate": 3.2869323707331176e-07, "logits/chosen": -1.453125, "logits/rejected": -1.96875, "logps/chosen": -490.0, "logps/rejected": -536.0, "loss": 0.217, "rewards/accuracies": 1.0, "rewards/chosen": -1.390625, "rewards/margins": 2.171875, "rewards/rejected": -3.5625, "step": 1110 }, { "epoch": 1.39, "grad_norm": 52.58964934069639, "learning_rate": 3.2526242381345766e-07, "logits/chosen": -2.21875, "logits/rejected": -2.390625, "logps/chosen": -484.0, "logps/rejected": -652.0, "loss": 0.2209, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.1875, "rewards/margins": 3.359375, "rewards/rejected": -4.53125, "step": 1120 }, { "epoch": 1.4, "grad_norm": 26.692041535833823, "learning_rate": 3.218159518910443e-07, "logits/chosen": -1.65625, "logits/rejected": -1.7265625, "logps/chosen": -540.0, "logps/rejected": -580.0, "loss": 0.2288, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.6328125, "rewards/margins": 2.015625, "rewards/rejected": -3.640625, "step": 1130 }, { "epoch": 1.41, "grad_norm": 36.803960598284185, "learning_rate": 3.183545383589927e-07, "logits/chosen": -1.1484375, "logits/rejected": -1.5078125, "logps/chosen": -628.0, "logps/rejected": -620.0, "loss": 0.2016, "rewards/accuracies": 1.0, "rewards/chosen": -1.4375, "rewards/margins": 2.0625, "rewards/rejected": -3.5, "step": 1140 }, { "epoch": 1.43, "grad_norm": 62.57539343068982, "learning_rate": 3.148789033788889e-07, "logits/chosen": -1.3359375, "logits/rejected": -1.5, "logps/chosen": -504.0, "logps/rejected": -524.0, "loss": 0.1919, "rewards/accuracies": 1.0, "rewards/chosen": -1.0546875, "rewards/margins": 2.234375, "rewards/rejected": -3.28125, "step": 1150 }, { "epoch": 1.44, "grad_norm": 27.545095580039987, "learning_rate": 3.113897700711502e-07, "logits/chosen": -0.9375, "logits/rejected": -1.7578125, "logps/chosen": -412.0, "logps/rejected": -728.0, "loss": 0.2339, "rewards/accuracies": 1.0, "rewards/chosen": -1.484375, "rewards/margins": 2.71875, "rewards/rejected": -4.1875, "step": 1160 }, { "epoch": 1.45, "grad_norm": 20.69956429305614, "learning_rate": 3.078878643645778e-07, "logits/chosen": -1.7109375, "logits/rejected": -1.4296875, "logps/chosen": -564.0, "logps/rejected": -628.0, "loss": 0.2147, "rewards/accuracies": 1.0, "rewards/chosen": -1.921875, "rewards/margins": 2.25, "rewards/rejected": -4.1875, "step": 1170 }, { "epoch": 1.46, "grad_norm": 30.998852642526103, "learning_rate": 3.0437391484532403e-07, "logits/chosen": -1.1328125, "logits/rejected": -1.9921875, "logps/chosen": -804.0, "logps/rejected": -600.0, "loss": 0.2046, "rewards/accuracies": 1.0, "rewards/chosen": -1.203125, "rewards/margins": 2.390625, "rewards/rejected": -3.59375, "step": 1180 }, { "epoch": 1.47, "grad_norm": 20.204732181232373, "learning_rate": 3.0084865260530666e-07, "logits/chosen": -1.6328125, "logits/rejected": -1.1875, "logps/chosen": -552.0, "logps/rejected": -780.0, "loss": 0.2196, "rewards/accuracies": 1.0, "rewards/chosen": -0.92578125, "rewards/margins": 3.203125, "rewards/rejected": -4.125, "step": 1190 }, { "epoch": 1.49, "grad_norm": 38.16020262352693, "learning_rate": 2.9731281109010253e-07, "logits/chosen": -1.8203125, "logits/rejected": -1.8359375, "logps/chosen": -446.0, "logps/rejected": -484.0, "loss": 0.233, "rewards/accuracies": 1.0, "rewards/chosen": -1.640625, "rewards/margins": 2.40625, "rewards/rejected": -4.0625, "step": 1200 }, { "epoch": 1.5, "grad_norm": 29.06302627643078, "learning_rate": 2.937671259463512e-07, "logits/chosen": -1.8515625, "logits/rejected": -1.625, "logps/chosen": -536.0, "logps/rejected": -648.0, "loss": 0.2302, "rewards/accuracies": 1.0, "rewards/chosen": -1.296875, "rewards/margins": 3.484375, "rewards/rejected": -4.78125, "step": 1210 }, { "epoch": 1.51, "grad_norm": 91.98272670818265, "learning_rate": 2.9021233486869994e-07, "logits/chosen": -1.4375, "logits/rejected": -1.7421875, "logps/chosen": -672.0, "logps/rejected": -560.0, "loss": 0.2402, "rewards/accuracies": 1.0, "rewards/chosen": -1.484375, "rewards/margins": 1.6953125, "rewards/rejected": -3.1875, "step": 1220 }, { "epoch": 1.52, "grad_norm": 30.373071126803815, "learning_rate": 2.8664917744632423e-07, "logits/chosen": -1.6484375, "logits/rejected": -1.875, "logps/chosen": -544.0, "logps/rejected": -640.0, "loss": 0.2211, "rewards/accuracies": 1.0, "rewards/chosen": -1.90625, "rewards/margins": 2.40625, "rewards/rejected": -4.3125, "step": 1230 }, { "epoch": 1.54, "grad_norm": 24.064215386247714, "learning_rate": 2.830783950090522e-07, "logits/chosen": -1.2265625, "logits/rejected": -1.734375, "logps/chosen": -528.0, "logps/rejected": -462.0, "loss": 0.2278, "rewards/accuracies": 1.0, "rewards/chosen": -1.875, "rewards/margins": 1.96875, "rewards/rejected": -3.828125, "step": 1240 }, { "epoch": 1.55, "grad_norm": 38.77040858523863, "learning_rate": 2.7950073047312855e-07, "logits/chosen": -1.2109375, "logits/rejected": -1.7109375, "logps/chosen": -640.0, "logps/rejected": -752.0, "loss": 0.2113, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.8125, "rewards/margins": 2.40625, "rewards/rejected": -4.21875, "step": 1250 }, { "epoch": 1.56, "grad_norm": 36.115163088163186, "learning_rate": 2.759169281866472e-07, "logits/chosen": -1.671875, "logits/rejected": -2.34375, "logps/chosen": -460.0, "logps/rejected": -430.0, "loss": 0.2048, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.953125, "rewards/margins": 1.7734375, "rewards/rejected": -3.71875, "step": 1260 }, { "epoch": 1.57, "grad_norm": 21.74767552298878, "learning_rate": 2.72327733774687e-07, "logits/chosen": -1.28125, "logits/rejected": -1.875, "logps/chosen": -752.0, "logps/rejected": -588.0, "loss": 0.2159, "rewards/accuracies": 1.0, "rewards/chosen": -0.8125, "rewards/margins": 2.40625, "rewards/rejected": -3.21875, "step": 1270 }, { "epoch": 1.59, "grad_norm": 29.90116775368825, "learning_rate": 2.6873389398418085e-07, "logits/chosen": -1.4453125, "logits/rejected": -1.9921875, "logps/chosen": -420.0, "logps/rejected": -490.0, "loss": 0.2191, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -2.390625, "rewards/margins": 1.59375, "rewards/rejected": -3.984375, "step": 1280 }, { "epoch": 1.6, "grad_norm": 23.793756365026436, "learning_rate": 2.6513615652855246e-07, "logits/chosen": -1.5546875, "logits/rejected": -1.3046875, "logps/chosen": -572.0, "logps/rejected": -600.0, "loss": 0.2011, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.84375, "rewards/margins": 1.53125, "rewards/rejected": -3.375, "step": 1290 }, { "epoch": 1.61, "grad_norm": 43.491509201941334, "learning_rate": 2.6153526993215085e-07, "logits/chosen": -1.65625, "logits/rejected": -1.6171875, "logps/chosen": -500.0, "logps/rejected": -548.0, "loss": 0.1999, "rewards/accuracies": 1.0, "rewards/chosen": -2.375, "rewards/margins": 2.34375, "rewards/rejected": -4.71875, "step": 1300 }, { "epoch": 1.62, "grad_norm": 43.19768159653058, "learning_rate": 2.579319833745169e-07, "logits/chosen": -1.546875, "logits/rejected": -1.859375, "logps/chosen": -460.0, "logps/rejected": -406.0, "loss": 0.2023, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.65625, "rewards/margins": 2.171875, "rewards/rejected": -3.828125, "step": 1310 }, { "epoch": 1.64, "grad_norm": 35.267489669162956, "learning_rate": 2.5432704653451374e-07, "logits/chosen": -1.3515625, "logits/rejected": -1.234375, "logps/chosen": -532.0, "logps/rejected": -728.0, "loss": 0.1962, "rewards/accuracies": 1.0, "rewards/chosen": -2.046875, "rewards/margins": 3.109375, "rewards/rejected": -5.15625, "step": 1320 }, { "epoch": 1.65, "grad_norm": 17.971641900070907, "learning_rate": 2.5072120943435246e-07, "logits/chosen": -1.1953125, "logits/rejected": -1.4765625, "logps/chosen": -784.0, "logps/rejected": -808.0, "loss": 0.1927, "rewards/accuracies": 1.0, "rewards/chosen": -1.453125, "rewards/margins": 2.765625, "rewards/rejected": -4.21875, "step": 1330 }, { "epoch": 1.66, "grad_norm": 26.236832847315178, "learning_rate": 2.471152222835471e-07, "logits/chosen": -2.1875, "logits/rejected": -1.5859375, "logps/chosen": -684.0, "logps/rejected": -624.0, "loss": 0.2161, "rewards/accuracies": 1.0, "rewards/chosen": -1.078125, "rewards/margins": 3.125, "rewards/rejected": -4.1875, "step": 1340 }, { "epoch": 1.67, "grad_norm": 25.147006316460192, "learning_rate": 2.4350983532283043e-07, "logits/chosen": -1.4296875, "logits/rejected": -1.1484375, "logps/chosen": -472.0, "logps/rejected": -592.0, "loss": 0.189, "rewards/accuracies": 0.699999988079071, "rewards/chosen": -1.34375, "rewards/margins": 2.125, "rewards/rejected": -3.46875, "step": 1350 }, { "epoch": 1.69, "grad_norm": 54.14713597288178, "learning_rate": 2.39905798668063e-07, "logits/chosen": -1.34375, "logits/rejected": -1.6875, "logps/chosen": -510.0, "logps/rejected": -568.0, "loss": 0.2336, "rewards/accuracies": 1.0, "rewards/chosen": -1.8359375, "rewards/margins": 2.484375, "rewards/rejected": -4.3125, "step": 1360 }, { "epoch": 1.7, "grad_norm": 21.942609458104677, "learning_rate": 2.3630386215416878e-07, "logits/chosen": -1.5390625, "logits/rejected": -1.8671875, "logps/chosen": -620.0, "logps/rejected": -620.0, "loss": 0.1966, "rewards/accuracies": 1.0, "rewards/chosen": -1.1484375, "rewards/margins": 3.0, "rewards/rejected": -4.15625, "step": 1370 }, { "epoch": 1.71, "grad_norm": 72.72707544492208, "learning_rate": 2.3270477517912835e-07, "logits/chosen": -1.1953125, "logits/rejected": -1.1171875, "logps/chosen": -716.0, "logps/rejected": -708.0, "loss": 0.2211, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -2.1875, "rewards/margins": 1.84375, "rewards/rejected": -4.03125, "step": 1380 }, { "epoch": 1.72, "grad_norm": 22.411175858752106, "learning_rate": 2.291092865480641e-07, "logits/chosen": -1.4375, "logits/rejected": -1.7890625, "logps/chosen": -568.0, "logps/rejected": -680.0, "loss": 0.2027, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.7109375, "rewards/margins": 3.09375, "rewards/rejected": -4.8125, "step": 1390 }, { "epoch": 1.73, "grad_norm": 124.47455963689552, "learning_rate": 2.2551814431744758e-07, "logits/chosen": -1.4765625, "logits/rejected": -1.5, "logps/chosen": -592.0, "logps/rejected": -600.0, "loss": 0.3274, "rewards/accuracies": 1.0, "rewards/chosen": -1.9765625, "rewards/margins": 2.40625, "rewards/rejected": -4.375, "step": 1400 }, { "epoch": 1.75, "grad_norm": 40.10464388387894, "learning_rate": 2.2193209563946382e-07, "logits/chosen": -1.1875, "logits/rejected": -1.859375, "logps/chosen": -740.0, "logps/rejected": -604.0, "loss": 0.2608, "rewards/accuracies": 1.0, "rewards/chosen": -2.34375, "rewards/margins": 2.984375, "rewards/rejected": -5.3125, "step": 1410 }, { "epoch": 1.76, "grad_norm": 139.91739653913288, "learning_rate": 2.1835188660656265e-07, "logits/chosen": -1.25, "logits/rejected": -1.4921875, "logps/chosen": -600.0, "logps/rejected": -580.0, "loss": 0.1985, "rewards/accuracies": 1.0, "rewards/chosen": -1.546875, "rewards/margins": 2.125, "rewards/rejected": -3.671875, "step": 1420 }, { "epoch": 1.77, "grad_norm": 45.52923168908793, "learning_rate": 2.147782620962314e-07, "logits/chosen": -1.5, "logits/rejected": -1.7109375, "logps/chosen": -544.0, "logps/rejected": -540.0, "loss": 0.2292, "rewards/accuracies": 1.0, "rewards/chosen": -1.4375, "rewards/margins": 2.140625, "rewards/rejected": -3.578125, "step": 1430 }, { "epoch": 1.78, "grad_norm": 55.0384600231688, "learning_rate": 2.112119656160199e-07, "logits/chosen": -1.15625, "logits/rejected": -1.25, "logps/chosen": -620.0, "logps/rejected": -700.0, "loss": 0.2028, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.8046875, "rewards/margins": 1.6953125, "rewards/rejected": -3.484375, "step": 1440 }, { "epoch": 1.8, "grad_norm": 18.297415510730342, "learning_rate": 2.0765373914885047e-07, "logits/chosen": -1.8203125, "logits/rejected": -1.5234375, "logps/chosen": -418.0, "logps/rejected": -508.0, "loss": 0.2187, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.3828125, "rewards/margins": 1.5234375, "rewards/rejected": -2.90625, "step": 1450 }, { "epoch": 1.81, "grad_norm": 25.14760440570739, "learning_rate": 2.0410432299864556e-07, "logits/chosen": -1.46875, "logits/rejected": -1.3828125, "logps/chosen": -584.0, "logps/rejected": -816.0, "loss": 0.2162, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -2.125, "rewards/margins": 2.9375, "rewards/rejected": -5.0625, "step": 1460 }, { "epoch": 1.82, "grad_norm": 25.279928980019168, "learning_rate": 2.0056445563630423e-07, "logits/chosen": -1.9609375, "logits/rejected": -1.5625, "logps/chosen": -532.0, "logps/rejected": -604.0, "loss": 0.217, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.40625, "rewards/margins": 2.046875, "rewards/rejected": -3.46875, "step": 1470 }, { "epoch": 1.83, "grad_norm": 31.005511813027777, "learning_rate": 1.9703487354606018e-07, "logits/chosen": -2.640625, "logits/rejected": -1.96875, "logps/chosen": -564.0, "logps/rejected": -628.0, "loss": 0.2051, "rewards/accuracies": 1.0, "rewards/chosen": -1.25, "rewards/margins": 3.265625, "rewards/rejected": -4.53125, "step": 1480 }, { "epoch": 1.85, "grad_norm": 17.907410467346242, "learning_rate": 1.935163110722533e-07, "logits/chosen": -1.578125, "logits/rejected": -1.8515625, "logps/chosen": -672.0, "logps/rejected": -528.0, "loss": 0.2019, "rewards/accuracies": 1.0, "rewards/chosen": -0.82421875, "rewards/margins": 2.359375, "rewards/rejected": -3.171875, "step": 1490 }, { "epoch": 1.86, "grad_norm": 37.56302682379733, "learning_rate": 1.900095002665459e-07, "logits/chosen": -1.375, "logits/rejected": -1.5859375, "logps/chosen": -544.0, "logps/rejected": -668.0, "loss": 0.2247, "rewards/accuracies": 1.0, "rewards/chosen": -1.4609375, "rewards/margins": 2.109375, "rewards/rejected": -3.5625, "step": 1500 }, { "epoch": 1.86, "eval_logits/chosen": -1.0546875, "eval_logits/rejected": -0.9765625, "eval_logps/chosen": -568.0, "eval_logps/rejected": -668.0, "eval_loss": 0.3828948140144348, "eval_rewards/accuracies": 0.7777777910232544, "eval_rewards/chosen": -2.203125, "eval_rewards/margins": 1.8125, "eval_rewards/rejected": -4.03125, "eval_runtime": 49.0823, "eval_samples_per_second": 21.393, "eval_steps_per_second": 0.183, "step": 1500 }, { "epoch": 1.87, "grad_norm": 142.03485499088688, "learning_rate": 1.8651517073561673e-07, "logits/chosen": -1.9140625, "logits/rejected": -1.84375, "logps/chosen": -516.0, "logps/rejected": -444.0, "loss": 0.2354, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.0546875, "rewards/margins": 1.6640625, "rewards/rejected": -2.71875, "step": 1510 }, { "epoch": 1.88, "grad_norm": 20.904872000144213, "learning_rate": 1.8303404948936285e-07, "logits/chosen": -1.5625, "logits/rejected": -1.3828125, "logps/chosen": -466.0, "logps/rejected": -492.0, "loss": 0.2063, "rewards/accuracies": 1.0, "rewards/chosen": -0.98046875, "rewards/margins": 2.203125, "rewards/rejected": -3.171875, "step": 1520 }, { "epoch": 1.9, "grad_norm": 23.082974237096174, "learning_rate": 1.7956686078964255e-07, "logits/chosen": -1.375, "logits/rejected": -1.4375, "logps/chosen": -528.0, "logps/rejected": -656.0, "loss": 0.2083, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.421875, "rewards/margins": 3.09375, "rewards/rejected": -4.5, "step": 1530 }, { "epoch": 1.91, "grad_norm": 65.20385510486626, "learning_rate": 1.7611432599958924e-07, "logits/chosen": -1.9140625, "logits/rejected": -2.234375, "logps/chosen": -352.0, "logps/rejected": -392.0, "loss": 0.2083, "rewards/accuracies": 1.0, "rewards/chosen": -1.3515625, "rewards/margins": 1.9609375, "rewards/rejected": -3.3125, "step": 1540 }, { "epoch": 1.92, "grad_norm": 26.23633528972487, "learning_rate": 1.726771634335293e-07, "logits/chosen": -1.4609375, "logits/rejected": -2.0625, "logps/chosen": -492.0, "logps/rejected": -456.0, "loss": 0.2321, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.0703125, "rewards/margins": 1.6015625, "rewards/rejected": -2.671875, "step": 1550 }, { "epoch": 1.93, "grad_norm": 49.10798542538174, "learning_rate": 1.6925608820753325e-07, "logits/chosen": -0.83203125, "logits/rejected": -1.1953125, "logps/chosen": -708.0, "logps/rejected": -880.0, "loss": 0.2232, "rewards/accuracies": 1.0, "rewards/chosen": -0.86328125, "rewards/margins": 2.625, "rewards/rejected": -3.484375, "step": 1560 }, { "epoch": 1.95, "grad_norm": 52.60647313486629, "learning_rate": 1.6585181209063321e-07, "logits/chosen": -1.71875, "logits/rejected": -1.4921875, "logps/chosen": -472.0, "logps/rejected": -704.0, "loss": 0.1907, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.5546875, "rewards/margins": 2.578125, "rewards/rejected": -4.125, "step": 1570 }, { "epoch": 1.96, "grad_norm": 24.213468426964006, "learning_rate": 1.6246504335673625e-07, "logits/chosen": -1.0390625, "logits/rejected": -1.4453125, "logps/chosen": -668.0, "logps/rejected": -856.0, "loss": 0.2086, "rewards/accuracies": 1.0, "rewards/chosen": -0.9375, "rewards/margins": 2.15625, "rewards/rejected": -3.09375, "step": 1580 }, { "epoch": 1.97, "grad_norm": 47.733763616697836, "learning_rate": 1.590964866372652e-07, "logits/chosen": -1.09375, "logits/rejected": -1.2734375, "logps/chosen": -636.0, "logps/rejected": -784.0, "loss": 0.2083, "rewards/accuracies": 1.0, "rewards/chosen": -1.3515625, "rewards/margins": 2.828125, "rewards/rejected": -4.1875, "step": 1590 }, { "epoch": 1.98, "grad_norm": 37.63668160123638, "learning_rate": 1.5574684277455685e-07, "logits/chosen": -1.765625, "logits/rejected": -1.1953125, "logps/chosen": -464.0, "logps/rejected": -640.0, "loss": 0.22, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -1.5234375, "rewards/margins": 1.765625, "rewards/rejected": -3.28125, "step": 1600 }, { "epoch": 2.0, "grad_norm": 20.84423028894674, "learning_rate": 1.5241680867604905e-07, "logits/chosen": -1.0078125, "logits/rejected": -2.34375, "logps/chosen": -660.0, "logps/rejected": -624.0, "loss": 0.2062, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.73046875, "rewards/margins": 2.265625, "rewards/rejected": -3.0, "step": 1610 }, { "epoch": 2.01, "grad_norm": 15.238180752697565, "learning_rate": 1.4910707716928586e-07, "logits/chosen": -1.75, "logits/rejected": -2.375, "logps/chosen": -568.0, "logps/rejected": -696.0, "loss": 0.1306, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.828125, "rewards/margins": 3.359375, "rewards/rejected": -5.1875, "step": 1620 }, { "epoch": 2.02, "grad_norm": 13.583277201205796, "learning_rate": 1.4581833685777228e-07, "logits/chosen": -1.34375, "logits/rejected": -1.578125, "logps/chosen": -552.0, "logps/rejected": -640.0, "loss": 0.1173, "rewards/accuracies": 1.0, "rewards/chosen": -1.5234375, "rewards/margins": 2.65625, "rewards/rejected": -4.1875, "step": 1630 }, { "epoch": 2.03, "grad_norm": 14.86440122341942, "learning_rate": 1.4255127197770707e-07, "logits/chosen": -1.4609375, "logits/rejected": -1.3828125, "logps/chosen": -434.0, "logps/rejected": -552.0, "loss": 0.1149, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.65625, "rewards/margins": 2.515625, "rewards/rejected": -4.1875, "step": 1640 }, { "epoch": 2.04, "grad_norm": 15.578800057924948, "learning_rate": 1.3930656225562474e-07, "logits/chosen": -1.6640625, "logits/rejected": -1.515625, "logps/chosen": -540.0, "logps/rejected": -620.0, "loss": 0.1074, "rewards/accuracies": 1.0, "rewards/chosen": -1.90625, "rewards/margins": 3.28125, "rewards/rejected": -5.1875, "step": 1650 }, { "epoch": 2.06, "grad_norm": 13.991553452696552, "learning_rate": 1.360848827669756e-07, "logits/chosen": -1.421875, "logits/rejected": -1.2265625, "logps/chosen": -524.0, "logps/rejected": -520.0, "loss": 0.1255, "rewards/accuracies": 1.0, "rewards/chosen": -1.7578125, "rewards/margins": 2.734375, "rewards/rejected": -4.5, "step": 1660 }, { "epoch": 2.07, "grad_norm": 21.745298822673373, "learning_rate": 1.3288690379567314e-07, "logits/chosen": -1.4140625, "logits/rejected": -1.84375, "logps/chosen": -506.0, "logps/rejected": -544.0, "loss": 0.123, "rewards/accuracies": 1.0, "rewards/chosen": -1.78125, "rewards/margins": 2.171875, "rewards/rejected": -3.953125, "step": 1670 }, { "epoch": 2.08, "grad_norm": 15.334862616251963, "learning_rate": 1.2971329069463932e-07, "logits/chosen": -1.328125, "logits/rejected": -1.8984375, "logps/chosen": -632.0, "logps/rejected": -672.0, "loss": 0.1169, "rewards/accuracies": 1.0, "rewards/chosen": -1.7421875, "rewards/margins": 3.03125, "rewards/rejected": -4.78125, "step": 1680 }, { "epoch": 2.09, "grad_norm": 36.5802518789977, "learning_rate": 1.2656470374737434e-07, "logits/chosen": -1.1875, "logits/rejected": -1.3671875, "logps/chosen": -716.0, "logps/rejected": -1024.0, "loss": 0.1232, "rewards/accuracies": 1.0, "rewards/chosen": -1.859375, "rewards/margins": 4.25, "rewards/rejected": -6.125, "step": 1690 }, { "epoch": 2.11, "grad_norm": 35.20242961161644, "learning_rate": 1.2344179803058264e-07, "logits/chosen": -1.2578125, "logits/rejected": -1.9921875, "logps/chosen": -528.0, "logps/rejected": -624.0, "loss": 0.1247, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -2.1875, "rewards/margins": 2.25, "rewards/rejected": -4.4375, "step": 1700 }, { "epoch": 2.12, "grad_norm": 20.682912146389263, "learning_rate": 1.203452232778807e-07, "logits/chosen": -1.4375, "logits/rejected": -1.6015625, "logps/chosen": -748.0, "logps/rejected": -824.0, "loss": 0.1213, "rewards/accuracies": 1.0, "rewards/chosen": -1.78125, "rewards/margins": 3.921875, "rewards/rejected": -5.71875, "step": 1710 }, { "epoch": 2.13, "grad_norm": 21.01075482943445, "learning_rate": 1.1727562374461788e-07, "logits/chosen": -1.9765625, "logits/rejected": -1.515625, "logps/chosen": -532.0, "logps/rejected": -620.0, "loss": 0.1279, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.5859375, "rewards/margins": 3.53125, "rewards/rejected": -5.125, "step": 1720 }, { "epoch": 2.14, "grad_norm": 29.30233670676864, "learning_rate": 1.142336380738361e-07, "logits/chosen": -1.3203125, "logits/rejected": -1.109375, "logps/chosen": -564.0, "logps/rejected": -544.0, "loss": 0.1133, "rewards/accuracies": 1.0, "rewards/chosen": -2.3125, "rewards/margins": 2.90625, "rewards/rejected": -5.21875, "step": 1730 }, { "epoch": 2.16, "grad_norm": 16.664591107532367, "learning_rate": 1.1121989916339756e-07, "logits/chosen": -1.203125, "logits/rejected": -2.9375, "logps/chosen": -732.0, "logps/rejected": -624.0, "loss": 0.1121, "rewards/accuracies": 1.0, "rewards/chosen": -2.46875, "rewards/margins": 3.109375, "rewards/rejected": -5.59375, "step": 1740 }, { "epoch": 2.17, "grad_norm": 83.44488397290417, "learning_rate": 1.0823503403430734e-07, "logits/chosen": -1.25, "logits/rejected": -1.5546875, "logps/chosen": -648.0, "logps/rejected": -508.0, "loss": 0.1218, "rewards/accuracies": 1.0, "rewards/chosen": -1.859375, "rewards/margins": 2.90625, "rewards/rejected": -4.78125, "step": 1750 }, { "epoch": 2.18, "grad_norm": 15.585689114051172, "learning_rate": 1.0527966370025964e-07, "logits/chosen": -1.125, "logits/rejected": -1.7578125, "logps/chosen": -716.0, "logps/rejected": -692.0, "loss": 0.1205, "rewards/accuracies": 1.0, "rewards/chosen": -2.015625, "rewards/margins": 2.875, "rewards/rejected": -4.875, "step": 1760 }, { "epoch": 2.19, "grad_norm": 10.765461249613185, "learning_rate": 1.0235440303843302e-07, "logits/chosen": -1.2109375, "logits/rejected": -1.796875, "logps/chosen": -500.0, "logps/rejected": -636.0, "loss": 0.1099, "rewards/accuracies": 1.0, "rewards/chosen": -2.59375, "rewards/margins": 2.515625, "rewards/rejected": -5.125, "step": 1770 }, { "epoch": 2.21, "grad_norm": 12.386913795936541, "learning_rate": 9.945986066156248e-08, "logits/chosen": -1.59375, "logits/rejected": -1.8828125, "logps/chosen": -498.0, "logps/rejected": -576.0, "loss": 0.108, "rewards/accuracies": 1.0, "rewards/chosen": -1.5703125, "rewards/margins": 3.5625, "rewards/rejected": -5.125, "step": 1780 }, { "epoch": 2.22, "grad_norm": 16.61091563337375, "learning_rate": 9.659663879131503e-08, "logits/chosen": -1.265625, "logits/rejected": -1.3125, "logps/chosen": -560.0, "logps/rejected": -528.0, "loss": 0.125, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -2.40625, "rewards/margins": 2.71875, "rewards/rejected": -5.125, "step": 1790 }, { "epoch": 2.23, "grad_norm": 24.411403141380244, "learning_rate": 9.376533313299542e-08, "logits/chosen": -1.2265625, "logits/rejected": -2.21875, "logps/chosen": -772.0, "logps/rejected": -660.0, "loss": 0.111, "rewards/accuracies": 1.0, "rewards/chosen": -2.078125, "rewards/margins": 2.9375, "rewards/rejected": -5.0, "step": 1800 }, { "epoch": 2.24, "grad_norm": 15.505538034971874, "learning_rate": 9.096653275160641e-08, "logits/chosen": -1.5390625, "logits/rejected": -1.59375, "logps/chosen": -492.0, "logps/rejected": -576.0, "loss": 0.1229, "rewards/accuracies": 1.0, "rewards/chosen": -1.7109375, "rewards/margins": 2.765625, "rewards/rejected": -4.46875, "step": 1810 }, { "epoch": 2.26, "grad_norm": 15.710939806805685, "learning_rate": 8.820081994929207e-08, "logits/chosen": -1.7421875, "logits/rejected": -2.03125, "logps/chosen": -724.0, "logps/rejected": -656.0, "loss": 0.1194, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -2.4375, "rewards/margins": 2.4375, "rewards/rejected": -4.875, "step": 1820 }, { "epoch": 2.27, "grad_norm": 15.364495322714388, "learning_rate": 8.546877014418671e-08, "logits/chosen": -1.9296875, "logits/rejected": -2.15625, "logps/chosen": -496.0, "logps/rejected": -532.0, "loss": 0.1282, "rewards/accuracies": 1.0, "rewards/chosen": -2.046875, "rewards/margins": 2.484375, "rewards/rejected": -4.53125, "step": 1830 }, { "epoch": 2.28, "grad_norm": 13.756725707803474, "learning_rate": 8.277095175069738e-08, "logits/chosen": -1.3984375, "logits/rejected": -1.4609375, "logps/chosen": -532.0, "logps/rejected": -552.0, "loss": 0.1072, "rewards/accuracies": 1.0, "rewards/chosen": -2.390625, "rewards/margins": 2.9375, "rewards/rejected": -5.3125, "step": 1840 }, { "epoch": 2.29, "grad_norm": 23.80983088717624, "learning_rate": 8.010792606124228e-08, "logits/chosen": -1.0703125, "logits/rejected": -1.0546875, "logps/chosen": -672.0, "logps/rejected": -680.0, "loss": 0.1081, "rewards/accuracies": 1.0, "rewards/chosen": -1.890625, "rewards/margins": 3.53125, "rewards/rejected": -5.40625, "step": 1850 }, { "epoch": 2.3, "grad_norm": 17.774551031970322, "learning_rate": 7.748024712947204e-08, "logits/chosen": -1.3984375, "logits/rejected": -1.2421875, "logps/chosen": -636.0, "logps/rejected": -652.0, "loss": 0.1291, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -2.28125, "rewards/margins": 2.34375, "rewards/rejected": -4.625, "step": 1860 }, { "epoch": 2.32, "grad_norm": 14.943619749566544, "learning_rate": 7.488846165499596e-08, "logits/chosen": -1.3984375, "logits/rejected": -1.9609375, "logps/chosen": -572.0, "logps/rejected": -684.0, "loss": 0.1282, "rewards/accuracies": 1.0, "rewards/chosen": -2.734375, "rewards/margins": 3.515625, "rewards/rejected": -6.25, "step": 1870 }, { "epoch": 2.33, "grad_norm": 16.562697765445648, "learning_rate": 7.233310886963942e-08, "logits/chosen": -1.375, "logits/rejected": -1.3984375, "logps/chosen": -474.0, "logps/rejected": -544.0, "loss": 0.1229, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -2.28125, "rewards/margins": 2.5, "rewards/rejected": -4.78125, "step": 1880 }, { "epoch": 2.34, "grad_norm": 16.237370125481036, "learning_rate": 6.981472042525416e-08, "logits/chosen": -1.515625, "logits/rejected": -1.765625, "logps/chosen": -640.0, "logps/rejected": -588.0, "loss": 0.1077, "rewards/accuracies": 1.0, "rewards/chosen": -2.078125, "rewards/margins": 2.890625, "rewards/rejected": -4.96875, "step": 1890 }, { "epoch": 2.35, "grad_norm": 12.54574310017106, "learning_rate": 6.7333820283106e-08, "logits/chosen": -0.94921875, "logits/rejected": -1.3828125, "logps/chosen": -696.0, "logps/rejected": -840.0, "loss": 0.1192, "rewards/accuracies": 1.0, "rewards/chosen": -1.578125, "rewards/margins": 4.125, "rewards/rejected": -5.71875, "step": 1900 }, { "epoch": 2.37, "grad_norm": 17.067698975214256, "learning_rate": 6.48909246048622e-08, "logits/chosen": -1.6953125, "logits/rejected": -1.6015625, "logps/chosen": -490.0, "logps/rejected": -560.0, "loss": 0.1259, "rewards/accuracies": 1.0, "rewards/chosen": -2.078125, "rewards/margins": 3.03125, "rewards/rejected": -5.09375, "step": 1910 }, { "epoch": 2.38, "grad_norm": 17.315279196446202, "learning_rate": 6.248654164520237e-08, "logits/chosen": -1.2890625, "logits/rejected": -1.4609375, "logps/chosen": -458.0, "logps/rejected": -426.0, "loss": 0.1221, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -2.515625, "rewards/margins": 2.09375, "rewards/rejected": -4.625, "step": 1920 }, { "epoch": 2.39, "grad_norm": 15.397715588828959, "learning_rate": 6.012117164607347e-08, "logits/chosen": -0.90625, "logits/rejected": -1.4921875, "logps/chosen": -796.0, "logps/rejected": -708.0, "loss": 0.109, "rewards/accuracies": 1.0, "rewards/chosen": -1.5078125, "rewards/margins": 3.78125, "rewards/rejected": -5.3125, "step": 1930 }, { "epoch": 2.4, "grad_norm": 45.120668890615434, "learning_rate": 5.779530673261279e-08, "logits/chosen": -1.0703125, "logits/rejected": -1.8125, "logps/chosen": -612.0, "logps/rejected": -820.0, "loss": 0.0907, "rewards/accuracies": 1.0, "rewards/chosen": -1.7265625, "rewards/margins": 3.5625, "rewards/rejected": -5.3125, "step": 1940 }, { "epoch": 2.42, "grad_norm": 12.405391170841106, "learning_rate": 5.5509430810758817e-08, "logits/chosen": -1.0234375, "logits/rejected": -1.5703125, "logps/chosen": -800.0, "logps/rejected": -848.0, "loss": 0.1012, "rewards/accuracies": 1.0, "rewards/chosen": -2.28125, "rewards/margins": 3.375, "rewards/rejected": -5.65625, "step": 1950 }, { "epoch": 2.43, "grad_norm": 13.29830997717489, "learning_rate": 5.3264019466573053e-08, "logits/chosen": -1.03125, "logits/rejected": -1.8203125, "logps/chosen": -660.0, "logps/rejected": -588.0, "loss": 0.0953, "rewards/accuracies": 1.0, "rewards/chosen": -1.859375, "rewards/margins": 3.640625, "rewards/rejected": -5.5, "step": 1960 }, { "epoch": 2.44, "grad_norm": 15.306039857091942, "learning_rate": 5.105953986729195e-08, "logits/chosen": -1.6171875, "logits/rejected": -1.2109375, "logps/chosen": -576.0, "logps/rejected": -732.0, "loss": 0.1057, "rewards/accuracies": 1.0, "rewards/chosen": -2.4375, "rewards/margins": 3.140625, "rewards/rejected": -5.5625, "step": 1970 }, { "epoch": 2.45, "grad_norm": 17.90216325537495, "learning_rate": 4.889645066413112e-08, "logits/chosen": -1.125, "logits/rejected": -1.5546875, "logps/chosen": -568.0, "logps/rejected": -612.0, "loss": 0.1165, "rewards/accuracies": 1.0, "rewards/chosen": -2.203125, "rewards/margins": 2.4375, "rewards/rejected": -4.625, "step": 1980 }, { "epoch": 2.47, "grad_norm": 13.665308623510128, "learning_rate": 4.67752018968606e-08, "logits/chosen": -1.1328125, "logits/rejected": -1.4765625, "logps/chosen": -624.0, "logps/rejected": -592.0, "loss": 0.0942, "rewards/accuracies": 1.0, "rewards/chosen": -2.1875, "rewards/margins": 3.3125, "rewards/rejected": -5.5, "step": 1990 }, { "epoch": 2.48, "grad_norm": 14.773793343841884, "learning_rate": 4.4696234900172744e-08, "logits/chosen": -1.5703125, "logits/rejected": -1.0546875, "logps/chosen": -540.0, "logps/rejected": -824.0, "loss": 0.1132, "rewards/accuracies": 1.0, "rewards/chosen": -2.4375, "rewards/margins": 3.46875, "rewards/rejected": -5.90625, "step": 2000 }, { "epoch": 2.48, "eval_logits/chosen": -1.0625, "eval_logits/rejected": -0.96484375, "eval_logps/chosen": -592.0, "eval_logps/rejected": -696.0, "eval_loss": 0.37350770831108093, "eval_rewards/accuracies": 0.75, "eval_rewards/chosen": -3.3125, "eval_rewards/margins": 2.09375, "eval_rewards/rejected": -5.40625, "eval_runtime": 49.8427, "eval_samples_per_second": 21.066, "eval_steps_per_second": 0.181, "step": 2000 }, { "epoch": 2.49, "grad_norm": 15.779889705391266, "learning_rate": 4.265998221186023e-08, "logits/chosen": -1.2421875, "logits/rejected": -1.1640625, "logps/chosen": -592.0, "logps/rejected": -560.0, "loss": 0.1073, "rewards/accuracies": 1.0, "rewards/chosen": -2.546875, "rewards/margins": 3.03125, "rewards/rejected": -5.5625, "step": 2010 }, { "epoch": 2.5, "grad_norm": 27.097026355265392, "learning_rate": 4.0666867482825135e-08, "logits/chosen": -1.0859375, "logits/rejected": -1.0859375, "logps/chosen": -668.0, "logps/rejected": -548.0, "loss": 0.117, "rewards/accuracies": 1.0, "rewards/chosen": -1.953125, "rewards/margins": 3.359375, "rewards/rejected": -5.3125, "step": 2020 }, { "epoch": 2.52, "grad_norm": 13.442767252473281, "learning_rate": 3.871730538893611e-08, "logits/chosen": -1.3515625, "logits/rejected": -1.9609375, "logps/chosen": -736.0, "logps/rejected": -740.0, "loss": 0.1108, "rewards/accuracies": 1.0, "rewards/chosen": -1.421875, "rewards/margins": 3.328125, "rewards/rejected": -4.75, "step": 2030 }, { "epoch": 2.53, "grad_norm": 17.818418586870905, "learning_rate": 3.681170154475391e-08, "logits/chosen": -1.625, "logits/rejected": -1.546875, "logps/chosen": -442.0, "logps/rejected": -620.0, "loss": 0.1236, "rewards/accuracies": 1.0, "rewards/chosen": -1.90625, "rewards/margins": 2.5625, "rewards/rejected": -4.46875, "step": 2040 }, { "epoch": 2.54, "grad_norm": 16.487242563455915, "learning_rate": 3.495045241914105e-08, "logits/chosen": -1.09375, "logits/rejected": -2.546875, "logps/chosen": -584.0, "logps/rejected": -672.0, "loss": 0.1079, "rewards/accuracies": 1.0, "rewards/chosen": -2.25, "rewards/margins": 4.09375, "rewards/rejected": -6.34375, "step": 2050 }, { "epoch": 2.55, "grad_norm": 15.802897986414916, "learning_rate": 3.313394525277527e-08, "logits/chosen": -1.4609375, "logits/rejected": -1.3125, "logps/chosen": -482.0, "logps/rejected": -572.0, "loss": 0.0979, "rewards/accuracies": 1.0, "rewards/chosen": -2.265625, "rewards/margins": 3.078125, "rewards/rejected": -5.34375, "step": 2060 }, { "epoch": 2.57, "grad_norm": 11.010895733144364, "learning_rate": 3.1362557977582e-08, "logits/chosen": -1.1953125, "logits/rejected": -1.2109375, "logps/chosen": -482.0, "logps/rejected": -506.0, "loss": 0.1009, "rewards/accuracies": 1.0, "rewards/chosen": -2.65625, "rewards/margins": 1.8515625, "rewards/rejected": -4.5, "step": 2070 }, { "epoch": 2.58, "grad_norm": 19.922568800825218, "learning_rate": 2.963665913810451e-08, "logits/chosen": -1.0078125, "logits/rejected": -2.40625, "logps/chosen": -712.0, "logps/rejected": -588.0, "loss": 0.1016, "rewards/accuracies": 1.0, "rewards/chosen": -2.34375, "rewards/margins": 2.796875, "rewards/rejected": -5.125, "step": 2080 }, { "epoch": 2.59, "grad_norm": 14.614675732714645, "learning_rate": 2.7956607814826366e-08, "logits/chosen": -1.2109375, "logits/rejected": -1.078125, "logps/chosen": -732.0, "logps/rejected": -712.0, "loss": 0.1244, "rewards/accuracies": 1.0, "rewards/chosen": -2.15625, "rewards/margins": 3.171875, "rewards/rejected": -5.34375, "step": 2090 }, { "epoch": 2.6, "grad_norm": 13.459903874571827, "learning_rate": 2.632275354946342e-08, "logits/chosen": -0.9375, "logits/rejected": -2.328125, "logps/chosen": -470.0, "logps/rejected": -386.0, "loss": 0.1195, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.8359375, "rewards/margins": 2.0625, "rewards/rejected": -3.90625, "step": 2100 }, { "epoch": 2.61, "grad_norm": 11.287103509304053, "learning_rate": 2.4735436272239922e-08, "logits/chosen": -1.7109375, "logits/rejected": -2.484375, "logps/chosen": -500.0, "logps/rejected": -612.0, "loss": 0.096, "rewards/accuracies": 1.0, "rewards/chosen": -2.03125, "rewards/margins": 3.71875, "rewards/rejected": -5.75, "step": 2110 }, { "epoch": 2.63, "grad_norm": 9.246071212037243, "learning_rate": 2.319498623116492e-08, "logits/chosen": -2.203125, "logits/rejected": -1.578125, "logps/chosen": -564.0, "logps/rejected": -796.0, "loss": 0.1098, "rewards/accuracies": 1.0, "rewards/chosen": -3.125, "rewards/margins": 4.15625, "rewards/rejected": -7.28125, "step": 2120 }, { "epoch": 2.64, "grad_norm": 25.87859572275051, "learning_rate": 2.1701723923322673e-08, "logits/chosen": -1.7265625, "logits/rejected": -1.8984375, "logps/chosen": -516.0, "logps/rejected": -644.0, "loss": 0.1225, "rewards/accuracies": 1.0, "rewards/chosen": -2.484375, "rewards/margins": 3.4375, "rewards/rejected": -5.9375, "step": 2130 }, { "epoch": 2.65, "grad_norm": 22.24550487415097, "learning_rate": 2.0255960028191798e-08, "logits/chosen": -1.5703125, "logits/rejected": -1.703125, "logps/chosen": -502.0, "logps/rejected": -552.0, "loss": 0.1103, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -2.375, "rewards/margins": 3.0625, "rewards/rejected": -5.4375, "step": 2140 }, { "epoch": 2.66, "grad_norm": 22.983007966397018, "learning_rate": 1.8857995343007167e-08, "logits/chosen": -1.8125, "logits/rejected": -1.3828125, "logps/chosen": -728.0, "logps/rejected": -900.0, "loss": 0.1094, "rewards/accuracies": 1.0, "rewards/chosen": -1.96875, "rewards/margins": 3.96875, "rewards/rejected": -5.9375, "step": 2150 }, { "epoch": 2.68, "grad_norm": 25.097623669642626, "learning_rate": 1.7508120720177795e-08, "logits/chosen": -1.1796875, "logits/rejected": -1.03125, "logps/chosen": -568.0, "logps/rejected": -604.0, "loss": 0.1184, "rewards/accuracies": 1.0, "rewards/chosen": -2.78125, "rewards/margins": 2.078125, "rewards/rejected": -4.875, "step": 2160 }, { "epoch": 2.69, "grad_norm": 32.20516783572916, "learning_rate": 1.6206617006773753e-08, "logits/chosen": -0.78515625, "logits/rejected": -2.078125, "logps/chosen": -736.0, "logps/rejected": -556.0, "loss": 0.1038, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -1.5078125, "rewards/margins": 2.828125, "rewards/rejected": -4.34375, "step": 2170 }, { "epoch": 2.7, "grad_norm": 11.93984865185982, "learning_rate": 1.4953754986094886e-08, "logits/chosen": -1.5859375, "logits/rejected": -1.6328125, "logps/chosen": -568.0, "logps/rejected": -580.0, "loss": 0.1095, "rewards/accuracies": 1.0, "rewards/chosen": -1.7109375, "rewards/margins": 2.640625, "rewards/rejected": -4.34375, "step": 2180 }, { "epoch": 2.71, "grad_norm": 30.015474257847476, "learning_rate": 1.3749795321332885e-08, "logits/chosen": -1.265625, "logits/rejected": -1.5390625, "logps/chosen": -664.0, "logps/rejected": -804.0, "loss": 0.1274, "rewards/accuracies": 1.0, "rewards/chosen": -1.96875, "rewards/margins": 3.421875, "rewards/rejected": -5.40625, "step": 2190 }, { "epoch": 2.73, "grad_norm": 20.828265625017977, "learning_rate": 1.2594988501339665e-08, "logits/chosen": -1.1796875, "logits/rejected": -1.796875, "logps/chosen": -628.0, "logps/rejected": -684.0, "loss": 0.1094, "rewards/accuracies": 1.0, "rewards/chosen": -2.71875, "rewards/margins": 3.109375, "rewards/rejected": -5.8125, "step": 2200 }, { "epoch": 2.74, "grad_norm": 22.34873903931498, "learning_rate": 1.148957478851173e-08, "logits/chosen": -1.515625, "logits/rejected": -1.375, "logps/chosen": -604.0, "logps/rejected": -572.0, "loss": 0.1136, "rewards/accuracies": 1.0, "rewards/chosen": -2.328125, "rewards/margins": 2.515625, "rewards/rejected": -4.84375, "step": 2210 }, { "epoch": 2.75, "grad_norm": 15.346930767785905, "learning_rate": 1.0433784168802805e-08, "logits/chosen": -1.3125, "logits/rejected": -1.578125, "logps/chosen": -624.0, "logps/rejected": -820.0, "loss": 0.1239, "rewards/accuracies": 1.0, "rewards/chosen": -3.03125, "rewards/margins": 3.015625, "rewards/rejected": -6.0625, "step": 2220 }, { "epoch": 2.76, "grad_norm": 16.847862725204944, "learning_rate": 9.427836303874115e-09, "logits/chosen": -1.1640625, "logits/rejected": -2.03125, "logps/chosen": -568.0, "logps/rejected": -648.0, "loss": 0.1139, "rewards/accuracies": 1.0, "rewards/chosen": -2.921875, "rewards/margins": 4.0, "rewards/rejected": -6.9375, "step": 2230 }, { "epoch": 2.78, "grad_norm": 11.648381674685714, "learning_rate": 8.47194048539307e-09, "logits/chosen": -1.015625, "logits/rejected": -1.6875, "logps/chosen": -880.0, "logps/rejected": -708.0, "loss": 0.114, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -2.515625, "rewards/margins": 3.515625, "rewards/rejected": -6.03125, "step": 2240 }, { "epoch": 2.79, "grad_norm": 18.35912272257459, "learning_rate": 7.566295591489052e-09, "logits/chosen": -1.4765625, "logits/rejected": -1.421875, "logps/chosen": -604.0, "logps/rejected": -676.0, "loss": 0.1191, "rewards/accuracies": 1.0, "rewards/chosen": -1.9140625, "rewards/margins": 3.015625, "rewards/rejected": -4.9375, "step": 2250 }, { "epoch": 2.8, "grad_norm": 13.681536454598728, "learning_rate": 6.71109004537615e-09, "logits/chosen": -1.125, "logits/rejected": -1.1875, "logps/chosen": -604.0, "logps/rejected": -664.0, "loss": 0.0951, "rewards/accuracies": 1.0, "rewards/chosen": -2.46875, "rewards/margins": 3.0625, "rewards/rejected": -5.53125, "step": 2260 }, { "epoch": 2.81, "grad_norm": 9.950566904078942, "learning_rate": 5.906501776150763e-09, "logits/chosen": -1.0, "logits/rejected": -2.921875, "logps/chosen": -712.0, "logps/rejected": -600.0, "loss": 0.1083, "rewards/accuracies": 1.0, "rewards/chosen": -1.8515625, "rewards/margins": 3.109375, "rewards/rejected": -4.96875, "step": 2270 }, { "epoch": 2.83, "grad_norm": 22.085690818869246, "learning_rate": 5.152698181772857e-09, "logits/chosen": -1.140625, "logits/rejected": -1.546875, "logps/chosen": -572.0, "logps/rejected": -760.0, "loss": 0.1072, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -2.125, "rewards/margins": 2.71875, "rewards/rejected": -4.84375, "step": 2280 }, { "epoch": 2.84, "grad_norm": 21.353543333604414, "learning_rate": 4.449836094238019e-09, "logits/chosen": -1.171875, "logits/rejected": -1.890625, "logps/chosen": -620.0, "logps/rejected": -486.0, "loss": 0.0966, "rewards/accuracies": 1.0, "rewards/chosen": -1.9453125, "rewards/margins": 3.21875, "rewards/rejected": -5.15625, "step": 2290 }, { "epoch": 2.85, "grad_norm": 11.392655664834834, "learning_rate": 3.798061746947995e-09, "logits/chosen": -1.09375, "logits/rejected": -1.546875, "logps/chosen": -500.0, "logps/rejected": -568.0, "loss": 0.1108, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -2.40625, "rewards/margins": 2.078125, "rewards/rejected": -4.5, "step": 2300 }, { "epoch": 2.86, "grad_norm": 16.899400188846528, "learning_rate": 3.1975107442860637e-09, "logits/chosen": -1.953125, "logits/rejected": -1.453125, "logps/chosen": -494.0, "logps/rejected": -928.0, "loss": 0.6379, "rewards/accuracies": 1.0, "rewards/chosen": -2.453125, "rewards/margins": 4.59375, "rewards/rejected": -7.0625, "step": 2310 }, { "epoch": 2.87, "grad_norm": 18.0993042603957, "learning_rate": 2.6483080334041287e-09, "logits/chosen": -1.328125, "logits/rejected": -1.453125, "logps/chosen": -652.0, "logps/rejected": -816.0, "loss": 0.0927, "rewards/accuracies": 1.0, "rewards/chosen": -2.078125, "rewards/margins": 3.625, "rewards/rejected": -5.6875, "step": 2320 }, { "epoch": 2.89, "grad_norm": 18.326061508758997, "learning_rate": 2.1505678782269e-09, "logits/chosen": -1.3125, "logits/rejected": -2.1875, "logps/chosen": -668.0, "logps/rejected": -572.0, "loss": 0.1068, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -2.125, "rewards/margins": 3.1875, "rewards/rejected": -5.3125, "step": 2330 }, { "epoch": 2.9, "grad_norm": 20.00896083752618, "learning_rate": 1.7043938356787467e-09, "logits/chosen": -1.4921875, "logits/rejected": -1.3125, "logps/chosen": -354.0, "logps/rejected": -556.0, "loss": 0.1215, "rewards/accuracies": 1.0, "rewards/chosen": -2.109375, "rewards/margins": 2.765625, "rewards/rejected": -4.875, "step": 2340 }, { "epoch": 2.91, "grad_norm": 33.844584510123674, "learning_rate": 1.30987873413832e-09, "logits/chosen": -0.94140625, "logits/rejected": -1.109375, "logps/chosen": -588.0, "logps/rejected": -660.0, "loss": 0.1078, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -2.375, "rewards/margins": 2.859375, "rewards/rejected": -5.21875, "step": 2350 }, { "epoch": 2.92, "grad_norm": 13.322271542703083, "learning_rate": 9.671046541251393e-10, "logits/chosen": -1.09375, "logits/rejected": -2.046875, "logps/chosen": -672.0, "logps/rejected": -576.0, "loss": 0.1159, "rewards/accuracies": 1.0, "rewards/chosen": -1.7890625, "rewards/margins": 3.765625, "rewards/rejected": -5.5625, "step": 2360 }, { "epoch": 2.94, "grad_norm": 28.085054792653725, "learning_rate": 6.761429112225326e-10, "logits/chosen": -1.1484375, "logits/rejected": -0.7734375, "logps/chosen": -688.0, "logps/rejected": -908.0, "loss": 0.0952, "rewards/accuracies": 1.0, "rewards/chosen": -1.8203125, "rewards/margins": 3.71875, "rewards/rejected": -5.53125, "step": 2370 }, { "epoch": 2.95, "grad_norm": 20.97872609812394, "learning_rate": 4.370540412399759e-10, "logits/chosen": -1.859375, "logits/rejected": -1.8125, "logps/chosen": -564.0, "logps/rejected": -684.0, "loss": 0.1135, "rewards/accuracies": 1.0, "rewards/chosen": -2.328125, "rewards/margins": 3.609375, "rewards/rejected": -5.9375, "step": 2380 }, { "epoch": 2.96, "grad_norm": 19.210485937449167, "learning_rate": 2.498877876184191e-10, "logits/chosen": -1.4765625, "logits/rejected": -1.046875, "logps/chosen": -688.0, "logps/rejected": -664.0, "loss": 0.1025, "rewards/accuracies": 1.0, "rewards/chosen": -1.6015625, "rewards/margins": 3.125, "rewards/rejected": -4.71875, "step": 2390 }, { "epoch": 2.97, "grad_norm": 12.682964834422256, "learning_rate": 1.1468309108100816e-10, "logits/chosen": -1.1171875, "logits/rejected": -1.3125, "logps/chosen": -456.0, "logps/rejected": -540.0, "loss": 0.1049, "rewards/accuracies": 1.0, "rewards/chosen": -3.0, "rewards/margins": 3.109375, "rewards/rejected": -6.125, "step": 2400 }, { "epoch": 2.99, "grad_norm": 16.080027717160323, "learning_rate": 3.146808153123293e-11, "logits/chosen": -1.234375, "logits/rejected": -2.03125, "logps/chosen": -492.0, "logps/rejected": -532.0, "loss": 0.1196, "rewards/accuracies": 1.0, "rewards/chosen": -2.109375, "rewards/margins": 2.046875, "rewards/rejected": -4.15625, "step": 2410 }, { "epoch": 3.0, "grad_norm": 13.95546245482141, "learning_rate": 2.60072200469752e-13, "logits/chosen": -1.2265625, "logits/rejected": -2.421875, "logps/chosen": -624.0, "logps/rejected": -482.0, "loss": 0.1097, "rewards/accuracies": 1.0, "rewards/chosen": -2.34375, "rewards/margins": 2.84375, "rewards/rejected": -5.1875, "step": 2420 } ], "logging_steps": 10, "max_steps": 2421, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "total_flos": 0.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }