|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9996020692399522, |
|
"eval_steps": 1000, |
|
"global_step": 628, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0015917230401910067, |
|
"grad_norm": 1.083263489032715, |
|
"learning_rate": 7.936507936507936e-09, |
|
"logits/chosen": -1.870174765586853, |
|
"logits/rejected": -1.8014098405838013, |
|
"logps/chosen": -158.94805908203125, |
|
"logps/rejected": -194.90118408203125, |
|
"loss": 0.6931, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.0, |
|
"rewards/margins": 0.0, |
|
"rewards/rejected": 0.0, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.01591723040191007, |
|
"grad_norm": 1.7486374056371567, |
|
"learning_rate": 7.936507936507936e-08, |
|
"logits/chosen": -1.8175890445709229, |
|
"logits/rejected": -1.8403409719467163, |
|
"logps/chosen": -188.36404418945312, |
|
"logps/rejected": -179.35687255859375, |
|
"loss": 0.6932, |
|
"rewards/accuracies": 0.3576388955116272, |
|
"rewards/chosen": -0.0004040856729261577, |
|
"rewards/margins": -0.0008076421800069511, |
|
"rewards/rejected": 0.00040355639066547155, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.03183446080382014, |
|
"grad_norm": 1.2288534081464715, |
|
"learning_rate": 1.5873015873015872e-07, |
|
"logits/chosen": -1.9721990823745728, |
|
"logits/rejected": -1.980015516281128, |
|
"logps/chosen": -197.92391967773438, |
|
"logps/rejected": -207.22412109375, |
|
"loss": 0.6932, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 8.275737855001353e-06, |
|
"rewards/margins": -0.0001316081325057894, |
|
"rewards/rejected": 0.00013988386490382254, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0477516912057302, |
|
"grad_norm": 0.9705773937715004, |
|
"learning_rate": 2.3809523809523806e-07, |
|
"logits/chosen": -1.8532848358154297, |
|
"logits/rejected": -1.8753681182861328, |
|
"logps/chosen": -192.031982421875, |
|
"logps/rejected": -195.5606231689453, |
|
"loss": 0.6931, |
|
"rewards/accuracies": 0.5375000238418579, |
|
"rewards/chosen": 0.0007524025277234614, |
|
"rewards/margins": 0.000668005901388824, |
|
"rewards/rejected": 8.439658995484933e-05, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.06366892160764027, |
|
"grad_norm": 1.3879715707434512, |
|
"learning_rate": 3.1746031746031743e-07, |
|
"logits/chosen": -1.9114805459976196, |
|
"logits/rejected": -1.9526256322860718, |
|
"logps/chosen": -198.54586791992188, |
|
"logps/rejected": -190.65158081054688, |
|
"loss": 0.6931, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": 0.0004263836017344147, |
|
"rewards/margins": 0.00022551305301021785, |
|
"rewards/rejected": 0.0002008705196203664, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.07958615200955034, |
|
"grad_norm": 0.9223665834941889, |
|
"learning_rate": 3.968253968253968e-07, |
|
"logits/chosen": -1.9061291217803955, |
|
"logits/rejected": -1.963629126548767, |
|
"logps/chosen": -190.56007385253906, |
|
"logps/rejected": -193.8275146484375, |
|
"loss": 0.693, |
|
"rewards/accuracies": 0.512499988079071, |
|
"rewards/chosen": 0.0014020069502294064, |
|
"rewards/margins": 0.00029031833400949836, |
|
"rewards/rejected": 0.001111688674427569, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0955033824114604, |
|
"grad_norm": 1.6880319723067283, |
|
"learning_rate": 4.761904761904761e-07, |
|
"logits/chosen": -1.9793567657470703, |
|
"logits/rejected": -2.001311779022217, |
|
"logps/chosen": -196.06846618652344, |
|
"logps/rejected": -188.52914428710938, |
|
"loss": 0.6929, |
|
"rewards/accuracies": 0.528124988079071, |
|
"rewards/chosen": 0.00374850956723094, |
|
"rewards/margins": 0.0002347224799450487, |
|
"rewards/rejected": 0.003513787407428026, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.11142061281337047, |
|
"grad_norm": 1.2950341728538266, |
|
"learning_rate": 4.998106548810311e-07, |
|
"logits/chosen": -1.9034782648086548, |
|
"logits/rejected": -1.9826198816299438, |
|
"logps/chosen": -196.86940002441406, |
|
"logps/rejected": -177.41644287109375, |
|
"loss": 0.6926, |
|
"rewards/accuracies": 0.5718749761581421, |
|
"rewards/chosen": 0.005933411885052919, |
|
"rewards/margins": 0.001384641625918448, |
|
"rewards/rejected": 0.004548769909888506, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.12733784321528055, |
|
"grad_norm": 2.34883059858935, |
|
"learning_rate": 4.988839406031596e-07, |
|
"logits/chosen": -1.9306480884552002, |
|
"logits/rejected": -1.9404945373535156, |
|
"logps/chosen": -176.39120483398438, |
|
"logps/rejected": -193.53347778320312, |
|
"loss": 0.6922, |
|
"rewards/accuracies": 0.550000011920929, |
|
"rewards/chosen": 0.010212738066911697, |
|
"rewards/margins": 0.0014647066127508879, |
|
"rewards/rejected": 0.008748031221330166, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.14325507361719061, |
|
"grad_norm": 1.8215899265671025, |
|
"learning_rate": 4.971879403278432e-07, |
|
"logits/chosen": -1.8801124095916748, |
|
"logits/rejected": -1.9323303699493408, |
|
"logps/chosen": -189.4450225830078, |
|
"logps/rejected": -190.84414672851562, |
|
"loss": 0.6922, |
|
"rewards/accuracies": 0.543749988079071, |
|
"rewards/chosen": 0.011725414544343948, |
|
"rewards/margins": 0.0013827414950355887, |
|
"rewards/rejected": 0.01034267246723175, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.15917230401910068, |
|
"grad_norm": 1.208023660425821, |
|
"learning_rate": 4.947278962947386e-07, |
|
"logits/chosen": -1.8471660614013672, |
|
"logits/rejected": -1.8932346105575562, |
|
"logps/chosen": -193.34718322753906, |
|
"logps/rejected": -196.87457275390625, |
|
"loss": 0.6915, |
|
"rewards/accuracies": 0.581250011920929, |
|
"rewards/chosen": 0.015522779896855354, |
|
"rewards/margins": 0.002660988597199321, |
|
"rewards/rejected": 0.01286179106682539, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.17508953442101075, |
|
"grad_norm": 1.1987469741900727, |
|
"learning_rate": 4.915114123589732e-07, |
|
"logits/chosen": -1.8427928686141968, |
|
"logits/rejected": -1.9131923913955688, |
|
"logps/chosen": -177.29983520507812, |
|
"logps/rejected": -178.62013244628906, |
|
"loss": 0.6911, |
|
"rewards/accuracies": 0.5562499761581421, |
|
"rewards/chosen": 0.019433852285146713, |
|
"rewards/margins": 0.003971750847995281, |
|
"rewards/rejected": 0.015462102368474007, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.1910067648229208, |
|
"grad_norm": 1.3513440295821657, |
|
"learning_rate": 4.875484304880629e-07, |
|
"logits/chosen": -1.852002501487732, |
|
"logits/rejected": -1.8977711200714111, |
|
"logps/chosen": -196.73301696777344, |
|
"logps/rejected": -198.54873657226562, |
|
"loss": 0.6905, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": 0.02165614813566208, |
|
"rewards/margins": 0.006273286882787943, |
|
"rewards/rejected": 0.015382861718535423, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.20692399522483088, |
|
"grad_norm": 0.919308465194322, |
|
"learning_rate": 4.828512000318616e-07, |
|
"logits/chosen": -1.8836482763290405, |
|
"logits/rejected": -1.9497747421264648, |
|
"logps/chosen": -197.17617797851562, |
|
"logps/rejected": -199.19927978515625, |
|
"loss": 0.6895, |
|
"rewards/accuracies": 0.5874999761581421, |
|
"rewards/chosen": 0.026965487748384476, |
|
"rewards/margins": 0.008914025500416756, |
|
"rewards/rejected": 0.01805146224796772, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.22284122562674094, |
|
"grad_norm": 1.3141295336399639, |
|
"learning_rate": 4.774342398605221e-07, |
|
"logits/chosen": -1.9415794610977173, |
|
"logits/rejected": -1.9936542510986328, |
|
"logps/chosen": -191.3373565673828, |
|
"logps/rejected": -182.86744689941406, |
|
"loss": 0.6894, |
|
"rewards/accuracies": 0.565625011920929, |
|
"rewards/chosen": 0.02900194190442562, |
|
"rewards/margins": 0.007812701165676117, |
|
"rewards/rejected": 0.021189238876104355, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.238758456028651, |
|
"grad_norm": 1.0426885097254153, |
|
"learning_rate": 4.713142934875005e-07, |
|
"logits/chosen": -1.9385217428207397, |
|
"logits/rejected": -2.025717258453369, |
|
"logps/chosen": -201.0162811279297, |
|
"logps/rejected": -189.43167114257812, |
|
"loss": 0.6882, |
|
"rewards/accuracies": 0.606249988079071, |
|
"rewards/chosen": 0.030758211389183998, |
|
"rewards/margins": 0.014848137274384499, |
|
"rewards/rejected": 0.01591007225215435, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.2546756864305611, |
|
"grad_norm": 0.9993599317482367, |
|
"learning_rate": 4.64510277316316e-07, |
|
"logits/chosen": -1.883013367652893, |
|
"logits/rejected": -1.9087369441986084, |
|
"logps/chosen": -177.09432983398438, |
|
"logps/rejected": -181.0247802734375, |
|
"loss": 0.6875, |
|
"rewards/accuracies": 0.581250011920929, |
|
"rewards/chosen": 0.019768565893173218, |
|
"rewards/margins": 0.01374894194304943, |
|
"rewards/rejected": 0.006019626744091511, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.27059291683247116, |
|
"grad_norm": 6.045830822367488, |
|
"learning_rate": 4.570432221710314e-07, |
|
"logits/chosen": -1.8615553379058838, |
|
"logits/rejected": -1.9155362844467163, |
|
"logps/chosen": -191.3380126953125, |
|
"logps/rejected": -196.19644165039062, |
|
"loss": 0.6866, |
|
"rewards/accuracies": 0.574999988079071, |
|
"rewards/chosen": 0.01975223794579506, |
|
"rewards/margins": 0.01370079256594181, |
|
"rewards/rejected": 0.006051443982869387, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.28651014723438123, |
|
"grad_norm": 3.6863233560878217, |
|
"learning_rate": 4.4893620829118124e-07, |
|
"logits/chosen": -1.8656437397003174, |
|
"logits/rejected": -1.9271419048309326, |
|
"logps/chosen": -197.45355224609375, |
|
"logps/rejected": -187.43008422851562, |
|
"loss": 0.6861, |
|
"rewards/accuracies": 0.621874988079071, |
|
"rewards/chosen": 0.011210992000997066, |
|
"rewards/margins": 0.020882535725831985, |
|
"rewards/rejected": -0.009671543724834919, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.3024273776362913, |
|
"grad_norm": 1.0448028144351018, |
|
"learning_rate": 4.40214293992074e-07, |
|
"logits/chosen": -1.9013402462005615, |
|
"logits/rejected": -1.9314658641815186, |
|
"logps/chosen": -189.4186553955078, |
|
"logps/rejected": -181.21237182617188, |
|
"loss": 0.685, |
|
"rewards/accuracies": 0.6031249761581421, |
|
"rewards/chosen": -0.005181929562240839, |
|
"rewards/margins": 0.016439586877822876, |
|
"rewards/rejected": -0.021621517837047577, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.31834460803820136, |
|
"grad_norm": 2.39968390976015, |
|
"learning_rate": 4.3090443821097566e-07, |
|
"logits/chosen": -1.7716095447540283, |
|
"logits/rejected": -1.8057435750961304, |
|
"logps/chosen": -191.8214111328125, |
|
"logps/rejected": -202.6089630126953, |
|
"loss": 0.6834, |
|
"rewards/accuracies": 0.612500011920929, |
|
"rewards/chosen": -0.008436818607151508, |
|
"rewards/margins": 0.02256094664335251, |
|
"rewards/rejected": -0.030997764319181442, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.3342618384401114, |
|
"grad_norm": 1.2956793625076484, |
|
"learning_rate": 4.210354171785795e-07, |
|
"logits/chosen": -1.7898409366607666, |
|
"logits/rejected": -1.8095554113388062, |
|
"logps/chosen": -187.91336059570312, |
|
"logps/rejected": -191.72274780273438, |
|
"loss": 0.6822, |
|
"rewards/accuracies": 0.6031249761581421, |
|
"rewards/chosen": -0.006862046662718058, |
|
"rewards/margins": 0.026659756898880005, |
|
"rewards/rejected": -0.033521804958581924, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.3501790688420215, |
|
"grad_norm": 2.0041098363033623, |
|
"learning_rate": 4.1063773547332584e-07, |
|
"logits/chosen": -1.7102569341659546, |
|
"logits/rejected": -1.7631139755249023, |
|
"logps/chosen": -192.54238891601562, |
|
"logps/rejected": -196.4750213623047, |
|
"loss": 0.6811, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -0.02755718305706978, |
|
"rewards/margins": 0.021966466680169106, |
|
"rewards/rejected": -0.04952365159988403, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.36609629924393156, |
|
"grad_norm": 1.9332381756111658, |
|
"learning_rate": 3.997435317334988e-07, |
|
"logits/chosen": -1.7369827032089233, |
|
"logits/rejected": -1.7770426273345947, |
|
"logps/chosen": -191.87298583984375, |
|
"logps/rejected": -193.9591522216797, |
|
"loss": 0.6817, |
|
"rewards/accuracies": 0.5375000238418579, |
|
"rewards/chosen": -0.043014999479055405, |
|
"rewards/margins": 0.020200857892632484, |
|
"rewards/rejected": -0.06321585923433304, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.3820135296458416, |
|
"grad_norm": 1.8903913985972478, |
|
"learning_rate": 3.8838647931853684e-07, |
|
"logits/chosen": -1.7956278324127197, |
|
"logits/rejected": -1.7797701358795166, |
|
"logps/chosen": -191.063720703125, |
|
"logps/rejected": -194.3412322998047, |
|
"loss": 0.6807, |
|
"rewards/accuracies": 0.596875011920929, |
|
"rewards/chosen": -0.04468019679188728, |
|
"rewards/margins": 0.024412112310528755, |
|
"rewards/rejected": -0.06909231096506119, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.3979307600477517, |
|
"grad_norm": 2.209986978975659, |
|
"learning_rate": 3.7660168222660824e-07, |
|
"logits/chosen": -1.7686593532562256, |
|
"logits/rejected": -1.784973382949829, |
|
"logps/chosen": -210.4496612548828, |
|
"logps/rejected": -210.5813751220703, |
|
"loss": 0.6764, |
|
"rewards/accuracies": 0.6187499761581421, |
|
"rewards/chosen": -0.06448944658041, |
|
"rewards/margins": 0.04242020100355148, |
|
"rewards/rejected": -0.10690965503454208, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.41384799044966175, |
|
"grad_norm": 2.7492197743164577, |
|
"learning_rate": 3.6442556659016475e-07, |
|
"logits/chosen": -1.6692039966583252, |
|
"logits/rejected": -1.7762839794158936, |
|
"logps/chosen": -204.4394073486328, |
|
"logps/rejected": -207.07705688476562, |
|
"loss": 0.6753, |
|
"rewards/accuracies": 0.590624988079071, |
|
"rewards/chosen": -0.07524862885475159, |
|
"rewards/margins": 0.02850642427802086, |
|
"rewards/rejected": -0.10375505685806274, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.4297652208515718, |
|
"grad_norm": 2.266511648349345, |
|
"learning_rate": 3.5189576808485404e-07, |
|
"logits/chosen": -1.7136598825454712, |
|
"logits/rejected": -1.710269570350647, |
|
"logps/chosen": -208.0347137451172, |
|
"logps/rejected": -209.95751953125, |
|
"loss": 0.6737, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.08503381162881851, |
|
"rewards/margins": 0.0269694272428751, |
|
"rewards/rejected": -0.11200324445962906, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.4456824512534819, |
|
"grad_norm": 2.5291958520679354, |
|
"learning_rate": 3.390510155998023e-07, |
|
"logits/chosen": -1.628941297531128, |
|
"logits/rejected": -1.6508426666259766, |
|
"logps/chosen": -214.1870574951172, |
|
"logps/rejected": -212.94284057617188, |
|
"loss": 0.6707, |
|
"rewards/accuracies": 0.6468750238418579, |
|
"rewards/chosen": -0.12327191978693008, |
|
"rewards/margins": 0.07120023667812347, |
|
"rewards/rejected": -0.19447214901447296, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.46159968165539195, |
|
"grad_norm": 2.574405785518012, |
|
"learning_rate": 3.2593101152883795e-07, |
|
"logits/chosen": -1.6216236352920532, |
|
"logits/rejected": -1.6766021251678467, |
|
"logps/chosen": -216.7034149169922, |
|
"logps/rejected": -219.9063262939453, |
|
"loss": 0.6732, |
|
"rewards/accuracies": 0.621874988079071, |
|
"rewards/chosen": -0.16276732087135315, |
|
"rewards/margins": 0.05662418156862259, |
|
"rewards/rejected": -0.21939150989055634, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.477516912057302, |
|
"grad_norm": 2.6453554421569514, |
|
"learning_rate": 3.125763090526674e-07, |
|
"logits/chosen": -1.5658096075057983, |
|
"logits/rejected": -1.609926462173462, |
|
"logps/chosen": -212.8961944580078, |
|
"logps/rejected": -213.0232696533203, |
|
"loss": 0.6673, |
|
"rewards/accuracies": 0.596875011920929, |
|
"rewards/chosen": -0.14756593108177185, |
|
"rewards/margins": 0.06441614031791687, |
|
"rewards/rejected": -0.21198205649852753, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.4934341424592121, |
|
"grad_norm": 4.170174585407687, |
|
"learning_rate": 2.9902818679131775e-07, |
|
"logits/chosen": -1.5784204006195068, |
|
"logits/rejected": -1.5990747213363647, |
|
"logps/chosen": -216.91799926757812, |
|
"logps/rejected": -221.1397705078125, |
|
"loss": 0.665, |
|
"rewards/accuracies": 0.590624988079071, |
|
"rewards/chosen": -0.20333552360534668, |
|
"rewards/margins": 0.05764692276716232, |
|
"rewards/rejected": -0.2609824538230896, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.5093513728611222, |
|
"grad_norm": 2.5979901565172714, |
|
"learning_rate": 2.8532852121428733e-07, |
|
"logits/chosen": -1.5708476305007935, |
|
"logits/rejected": -1.610568642616272, |
|
"logps/chosen": -212.8919219970703, |
|
"logps/rejected": -209.1988983154297, |
|
"loss": 0.6602, |
|
"rewards/accuracies": 0.6031249761581421, |
|
"rewards/chosen": -0.21002519130706787, |
|
"rewards/margins": 0.07976146787405014, |
|
"rewards/rejected": -0.2897866368293762, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.5252686032630323, |
|
"grad_norm": 3.0268583107868676, |
|
"learning_rate": 2.715196572027789e-07, |
|
"logits/chosen": -1.5595386028289795, |
|
"logits/rejected": -1.579420804977417, |
|
"logps/chosen": -208.55126953125, |
|
"logps/rejected": -228.11337280273438, |
|
"loss": 0.6561, |
|
"rewards/accuracies": 0.6343749761581421, |
|
"rewards/chosen": -0.19552794098854065, |
|
"rewards/margins": 0.11572189629077911, |
|
"rewards/rejected": -0.31124982237815857, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.5411858336649423, |
|
"grad_norm": 2.608251205411858, |
|
"learning_rate": 2.5764427716409815e-07, |
|
"logits/chosen": -1.5596858263015747, |
|
"logits/rejected": -1.5745502710342407, |
|
"logps/chosen": -222.6412811279297, |
|
"logps/rejected": -227.70651245117188, |
|
"loss": 0.6544, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.21032941341400146, |
|
"rewards/margins": 0.09801909327507019, |
|
"rewards/rejected": -0.30834850668907166, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.5571030640668524, |
|
"grad_norm": 2.3596476657059666, |
|
"learning_rate": 2.4374526910277886e-07, |
|
"logits/chosen": -1.4910805225372314, |
|
"logits/rejected": -1.4995691776275635, |
|
"logps/chosen": -219.46536254882812, |
|
"logps/rejected": -235.89620971679688, |
|
"loss": 0.6613, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.2539146840572357, |
|
"rewards/margins": 0.09661410748958588, |
|
"rewards/rejected": -0.3505287766456604, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.5730202944687625, |
|
"grad_norm": 2.721512353744307, |
|
"learning_rate": 2.2986559405621886e-07, |
|
"logits/chosen": -1.5120307207107544, |
|
"logits/rejected": -1.5819119215011597, |
|
"logps/chosen": -236.832763671875, |
|
"logps/rejected": -235.5454559326172, |
|
"loss": 0.6668, |
|
"rewards/accuracies": 0.565625011920929, |
|
"rewards/chosen": -0.3000979721546173, |
|
"rewards/margins": 0.07621388137340546, |
|
"rewards/rejected": -0.37631186842918396, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.5889375248706725, |
|
"grad_norm": 3.020330440823815, |
|
"learning_rate": 2.160481533045751e-07, |
|
"logits/chosen": -1.4668768644332886, |
|
"logits/rejected": -1.5363471508026123, |
|
"logps/chosen": -218.35946655273438, |
|
"logps/rejected": -224.10208129882812, |
|
"loss": 0.6492, |
|
"rewards/accuracies": 0.6156250238418579, |
|
"rewards/chosen": -0.2719998359680176, |
|
"rewards/margins": 0.12549051642417908, |
|
"rewards/rejected": -0.39749038219451904, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.6048547552725826, |
|
"grad_norm": 2.9446946400840974, |
|
"learning_rate": 2.0233565576536564e-07, |
|
"logits/chosen": -1.485190749168396, |
|
"logits/rejected": -1.5274041891098022, |
|
"logps/chosen": -222.81765747070312, |
|
"logps/rejected": -222.657470703125, |
|
"loss": 0.6538, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.26167741417884827, |
|
"rewards/margins": 0.08558806031942368, |
|
"rewards/rejected": -0.34726545214653015, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.6207719856744927, |
|
"grad_norm": 3.1883509652873223, |
|
"learning_rate": 1.887704859826528e-07, |
|
"logits/chosen": -1.4792619943618774, |
|
"logits/rejected": -1.5055263042449951, |
|
"logps/chosen": -221.5879364013672, |
|
"logps/rejected": -239.82571411132812, |
|
"loss": 0.6495, |
|
"rewards/accuracies": 0.6187499761581421, |
|
"rewards/chosen": -0.2849220633506775, |
|
"rewards/margins": 0.13855159282684326, |
|
"rewards/rejected": -0.42347365617752075, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.6366892160764027, |
|
"grad_norm": 2.8954609378442657, |
|
"learning_rate": 1.7539457311884675e-07, |
|
"logits/chosen": -1.4238282442092896, |
|
"logits/rejected": -1.423081398010254, |
|
"logps/chosen": -229.7063446044922, |
|
"logps/rejected": -235.14236450195312, |
|
"loss": 0.6522, |
|
"rewards/accuracies": 0.6312500238418579, |
|
"rewards/chosen": -0.2554658353328705, |
|
"rewards/margins": 0.14818701148033142, |
|
"rewards/rejected": -0.4036528468132019, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.6526064464783128, |
|
"grad_norm": 3.203997856966496, |
|
"learning_rate": 1.6224926135406693e-07, |
|
"logits/chosen": -1.460460901260376, |
|
"logits/rejected": -1.4777477979660034, |
|
"logps/chosen": -236.0177764892578, |
|
"logps/rejected": -240.2278289794922, |
|
"loss": 0.6525, |
|
"rewards/accuracies": 0.596875011920929, |
|
"rewards/chosen": -0.32773512601852417, |
|
"rewards/margins": 0.08220280706882477, |
|
"rewards/rejected": -0.4099380075931549, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.6685236768802229, |
|
"grad_norm": 2.8800206835573094, |
|
"learning_rate": 1.4937518209365108e-07, |
|
"logits/chosen": -1.5116090774536133, |
|
"logits/rejected": -1.5359185934066772, |
|
"logps/chosen": -248.60263061523438, |
|
"logps/rejected": -245.75967407226562, |
|
"loss": 0.6527, |
|
"rewards/accuracies": 0.6187499761581421, |
|
"rewards/chosen": -0.3213188052177429, |
|
"rewards/margins": 0.10633780807256699, |
|
"rewards/rejected": -0.4276565909385681, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.6844409072821329, |
|
"grad_norm": 3.29734083048723, |
|
"learning_rate": 1.3681212837880977e-07, |
|
"logits/chosen": -1.4253356456756592, |
|
"logits/rejected": -1.444923758506775, |
|
"logps/chosen": -220.8928985595703, |
|
"logps/rejected": -238.1382293701172, |
|
"loss": 0.6502, |
|
"rewards/accuracies": 0.621874988079071, |
|
"rewards/chosen": -0.3292413353919983, |
|
"rewards/margins": 0.10124459117650986, |
|
"rewards/rejected": -0.43048587441444397, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.700358137684043, |
|
"grad_norm": 3.2357462634957908, |
|
"learning_rate": 1.2459893188861613e-07, |
|
"logits/chosen": -1.3886950016021729, |
|
"logits/rejected": -1.4186146259307861, |
|
"logps/chosen": -229.12069702148438, |
|
"logps/rejected": -241.5669708251953, |
|
"loss": 0.65, |
|
"rewards/accuracies": 0.6656249761581421, |
|
"rewards/chosen": -0.31468960642814636, |
|
"rewards/margins": 0.14469651877880096, |
|
"rewards/rejected": -0.4593861699104309, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.716275368085953, |
|
"grad_norm": 3.664386493120931, |
|
"learning_rate": 1.1277334291351145e-07, |
|
"logits/chosen": -1.4633034467697144, |
|
"logits/rejected": -1.4953584671020508, |
|
"logps/chosen": -229.84976196289062, |
|
"logps/rejected": -245.46762084960938, |
|
"loss": 0.6451, |
|
"rewards/accuracies": 0.643750011920929, |
|
"rewards/chosen": -0.35235971212387085, |
|
"rewards/margins": 0.1600349247455597, |
|
"rewards/rejected": -0.5123946666717529, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.7321925984878631, |
|
"grad_norm": 3.054029038543547, |
|
"learning_rate": 1.0137191367132078e-07, |
|
"logits/chosen": -1.466225504875183, |
|
"logits/rejected": -1.5026957988739014, |
|
"logps/chosen": -227.568359375, |
|
"logps/rejected": -251.8888397216797, |
|
"loss": 0.6449, |
|
"rewards/accuracies": 0.6468750238418579, |
|
"rewards/chosen": -0.2990425229072571, |
|
"rewards/margins": 0.16379448771476746, |
|
"rewards/rejected": -0.46283698081970215, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.7481098288897732, |
|
"grad_norm": 3.604867452453406, |
|
"learning_rate": 9.042988532644249e-08, |
|
"logits/chosen": -1.447538137435913, |
|
"logits/rejected": -1.5032737255096436, |
|
"logps/chosen": -240.51644897460938, |
|
"logps/rejected": -250.5240020751953, |
|
"loss": 0.6527, |
|
"rewards/accuracies": 0.6000000238418579, |
|
"rewards/chosen": -0.36844930052757263, |
|
"rewards/margins": 0.1114746555685997, |
|
"rewards/rejected": -0.47992390394210815, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.7640270592916832, |
|
"grad_norm": 3.100922028441707, |
|
"learning_rate": 7.998107906142839e-08, |
|
"logits/chosen": -1.4335956573486328, |
|
"logits/rejected": -1.468783974647522, |
|
"logps/chosen": -220.85147094726562, |
|
"logps/rejected": -234.76596069335938, |
|
"loss": 0.6462, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.3198398947715759, |
|
"rewards/margins": 0.13547857105731964, |
|
"rewards/rejected": -0.45531851053237915, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.7799442896935933, |
|
"grad_norm": 3.0897060153169447, |
|
"learning_rate": 7.005779153764682e-08, |
|
"logits/chosen": -1.460636854171753, |
|
"logits/rejected": -1.481178641319275, |
|
"logps/chosen": -217.1626434326172, |
|
"logps/rejected": -241.839111328125, |
|
"loss": 0.6465, |
|
"rewards/accuracies": 0.628125011920929, |
|
"rewards/chosen": -0.33777886629104614, |
|
"rewards/margins": 0.1417325884103775, |
|
"rewards/rejected": -0.47951143980026245, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.7958615200955034, |
|
"grad_norm": 3.515426969892236, |
|
"learning_rate": 6.069069506815325e-08, |
|
"logits/chosen": -1.423098087310791, |
|
"logits/rejected": -1.4111202955245972, |
|
"logps/chosen": -228.85360717773438, |
|
"logps/rejected": -247.3131561279297, |
|
"loss": 0.6459, |
|
"rewards/accuracies": 0.6468750238418579, |
|
"rewards/chosen": -0.3209570348262787, |
|
"rewards/margins": 0.18556544184684753, |
|
"rewards/rejected": -0.506522536277771, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.8117787504974134, |
|
"grad_norm": 3.168259746643005, |
|
"learning_rate": 5.190874281132851e-08, |
|
"logits/chosen": -1.3531607389450073, |
|
"logits/rejected": -1.4315484762191772, |
|
"logps/chosen": -228.17135620117188, |
|
"logps/rejected": -237.56796264648438, |
|
"loss": 0.6446, |
|
"rewards/accuracies": 0.6187499761581421, |
|
"rewards/chosen": -0.3634713888168335, |
|
"rewards/margins": 0.11672137677669525, |
|
"rewards/rejected": -0.48019275069236755, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.8276959808993235, |
|
"grad_norm": 3.547783875081257, |
|
"learning_rate": 4.373907927832513e-08, |
|
"logits/chosen": -1.4724462032318115, |
|
"logits/rejected": -1.4711410999298096, |
|
"logps/chosen": -216.9486541748047, |
|
"logps/rejected": -229.16531372070312, |
|
"loss": 0.6497, |
|
"rewards/accuracies": 0.574999988079071, |
|
"rewards/chosen": -0.34610506892204285, |
|
"rewards/margins": 0.10790134966373444, |
|
"rewards/rejected": -0.4540063738822937, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.8436132113012336, |
|
"grad_norm": 3.0690519001773744, |
|
"learning_rate": 3.620695643093924e-08, |
|
"logits/chosen": -1.436292290687561, |
|
"logits/rejected": -1.4620921611785889, |
|
"logps/chosen": -220.2271728515625, |
|
"logps/rejected": -242.4230499267578, |
|
"loss": 0.6444, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.2857658565044403, |
|
"rewards/margins": 0.16418084502220154, |
|
"rewards/rejected": -0.44994670152664185, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.8595304417031436, |
|
"grad_norm": 3.5137530293056676, |
|
"learning_rate": 2.9335655629243645e-08, |
|
"logits/chosen": -1.4339349269866943, |
|
"logits/rejected": -1.4756128787994385, |
|
"logps/chosen": -225.1971893310547, |
|
"logps/rejected": -240.7228546142578, |
|
"loss": 0.6474, |
|
"rewards/accuracies": 0.628125011920929, |
|
"rewards/chosen": -0.32935741543769836, |
|
"rewards/margins": 0.12228095531463623, |
|
"rewards/rejected": -0.4516383707523346, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.8754476721050537, |
|
"grad_norm": 3.303113230393075, |
|
"learning_rate": 2.31464156702382e-08, |
|
"logits/chosen": -1.4354156255722046, |
|
"logits/rejected": -1.4688793420791626, |
|
"logps/chosen": -228.87588500976562, |
|
"logps/rejected": -227.47872924804688, |
|
"loss": 0.6503, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.31748342514038086, |
|
"rewards/margins": 0.12047611176967621, |
|
"rewards/rejected": -0.4379595220088959, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.8913649025069638, |
|
"grad_norm": 3.6187583437530138, |
|
"learning_rate": 1.7658367139945228e-08, |
|
"logits/chosen": -1.484116554260254, |
|
"logits/rejected": -1.4669333696365356, |
|
"logps/chosen": -220.8059844970703, |
|
"logps/rejected": -235.5103759765625, |
|
"loss": 0.6422, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.34964197874069214, |
|
"rewards/margins": 0.1103706806898117, |
|
"rewards/rejected": -0.46001267433166504, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.9072821329088738, |
|
"grad_norm": 2.946862169915585, |
|
"learning_rate": 1.2888473281864597e-08, |
|
"logits/chosen": -1.3717347383499146, |
|
"logits/rejected": -1.4307266473770142, |
|
"logps/chosen": -230.9296112060547, |
|
"logps/rejected": -248.7758026123047, |
|
"loss": 0.6373, |
|
"rewards/accuracies": 0.6468750238418579, |
|
"rewards/chosen": -0.3366944193840027, |
|
"rewards/margins": 0.13169880211353302, |
|
"rewards/rejected": -0.4683932363986969, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.9231993633107839, |
|
"grad_norm": 3.986364311332544, |
|
"learning_rate": 8.851477564560061e-09, |
|
"logits/chosen": -1.4669787883758545, |
|
"logits/rejected": -1.5305414199829102, |
|
"logps/chosen": -227.6433563232422, |
|
"logps/rejected": -243.227294921875, |
|
"loss": 0.6434, |
|
"rewards/accuracies": 0.6468750238418579, |
|
"rewards/chosen": -0.32894089818000793, |
|
"rewards/margins": 0.15033873915672302, |
|
"rewards/rejected": -0.47927966713905334, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.939116593712694, |
|
"grad_norm": 3.4745032456527394, |
|
"learning_rate": 5.559858110443016e-09, |
|
"logits/chosen": -1.418752908706665, |
|
"logits/rejected": -1.463607668876648, |
|
"logps/chosen": -237.24252319335938, |
|
"logps/rejected": -252.40145874023438, |
|
"loss": 0.6384, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.3376900553703308, |
|
"rewards/margins": 0.15840700268745422, |
|
"rewards/rejected": -0.4960970878601074, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.955033824114604, |
|
"grad_norm": 6.105907684391068, |
|
"learning_rate": 3.023789126611137e-09, |
|
"logits/chosen": -1.4318501949310303, |
|
"logits/rejected": -1.4236361980438232, |
|
"logps/chosen": -224.8287811279297, |
|
"logps/rejected": -240.1597442626953, |
|
"loss": 0.6398, |
|
"rewards/accuracies": 0.699999988079071, |
|
"rewards/chosen": -0.32297495007514954, |
|
"rewards/margins": 0.1879177987575531, |
|
"rewards/rejected": -0.5108927488327026, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.9709510545165141, |
|
"grad_norm": 7.921761427791219, |
|
"learning_rate": 1.2511094569571668e-09, |
|
"logits/chosen": -1.4697223901748657, |
|
"logits/rejected": -1.5091297626495361, |
|
"logps/chosen": -227.2742156982422, |
|
"logps/rejected": -238.57974243164062, |
|
"loss": 0.6414, |
|
"rewards/accuracies": 0.596875011920929, |
|
"rewards/chosen": -0.3365384042263031, |
|
"rewards/margins": 0.1458342969417572, |
|
"rewards/rejected": -0.4823727011680603, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.9868682849184242, |
|
"grad_norm": 3.1578814752663407, |
|
"learning_rate": 2.4729835275189016e-10, |
|
"logits/chosen": -1.4319922924041748, |
|
"logits/rejected": -1.4167181253433228, |
|
"logps/chosen": -224.8016357421875, |
|
"logps/rejected": -247.70339965820312, |
|
"loss": 0.6404, |
|
"rewards/accuracies": 0.6343749761581421, |
|
"rewards/chosen": -0.36974093317985535, |
|
"rewards/margins": 0.14402151107788086, |
|
"rewards/rejected": -0.5137624740600586, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.9996020692399522, |
|
"step": 628, |
|
"total_flos": 0.0, |
|
"train_loss": 0.6662861807331159, |
|
"train_runtime": 15158.4407, |
|
"train_samples_per_second": 10.608, |
|
"train_steps_per_second": 0.041 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 628, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|