|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.0, |
|
"eval_steps": 100, |
|
"global_step": 574, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.003484320557491289, |
|
"grad_norm": 306.63621441123377, |
|
"learning_rate": 1.0344827586206896e-08, |
|
"logits/chosen": -2.5345611572265625, |
|
"logits/rejected": -2.581700563430786, |
|
"logps/chosen": -60.002105712890625, |
|
"logps/rejected": -99.98374938964844, |
|
"loss": 0.6931, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.0, |
|
"rewards/margins": 0.0, |
|
"rewards/rejected": 0.0, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.03484320557491289, |
|
"grad_norm": 287.10820765444424, |
|
"learning_rate": 1.0344827586206897e-07, |
|
"logits/chosen": -2.5633163452148438, |
|
"logits/rejected": -2.562026023864746, |
|
"logps/chosen": -59.65489196777344, |
|
"logps/rejected": -73.39691925048828, |
|
"loss": 0.6954, |
|
"rewards/accuracies": 0.2152777761220932, |
|
"rewards/chosen": 0.0025260683614760637, |
|
"rewards/margins": 0.011007179506123066, |
|
"rewards/rejected": -0.008481111377477646, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.06968641114982578, |
|
"grad_norm": 361.11744918130063, |
|
"learning_rate": 2.0689655172413793e-07, |
|
"logits/chosen": -2.60577654838562, |
|
"logits/rejected": -2.5645222663879395, |
|
"logps/chosen": -104.05818939208984, |
|
"logps/rejected": -94.88358306884766, |
|
"loss": 0.6868, |
|
"rewards/accuracies": 0.3062500059604645, |
|
"rewards/chosen": 0.0365118607878685, |
|
"rewards/margins": 0.03092603012919426, |
|
"rewards/rejected": 0.005585831124335527, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.10452961672473868, |
|
"grad_norm": 362.0188620910404, |
|
"learning_rate": 3.103448275862069e-07, |
|
"logits/chosen": -2.593327045440674, |
|
"logits/rejected": -2.573579788208008, |
|
"logps/chosen": -82.2002944946289, |
|
"logps/rejected": -91.45396423339844, |
|
"loss": 0.6711, |
|
"rewards/accuracies": 0.29374998807907104, |
|
"rewards/chosen": 0.15702371299266815, |
|
"rewards/margins": 0.1113327145576477, |
|
"rewards/rejected": 0.04569100961089134, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.13937282229965156, |
|
"grad_norm": 283.91931715411215, |
|
"learning_rate": 4.1379310344827586e-07, |
|
"logits/chosen": -2.4966464042663574, |
|
"logits/rejected": -2.4948105812072754, |
|
"logps/chosen": -77.7404556274414, |
|
"logps/rejected": -73.33540344238281, |
|
"loss": 0.6457, |
|
"rewards/accuracies": 0.3125, |
|
"rewards/chosen": 0.09431798756122589, |
|
"rewards/margins": 0.27138587832450867, |
|
"rewards/rejected": -0.17706790566444397, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.17421602787456447, |
|
"grad_norm": 251.26626326645746, |
|
"learning_rate": 5.172413793103448e-07, |
|
"logits/chosen": -2.5199971199035645, |
|
"logits/rejected": -2.5240330696105957, |
|
"logps/chosen": -62.982337951660156, |
|
"logps/rejected": -75.54759216308594, |
|
"loss": 0.6705, |
|
"rewards/accuracies": 0.2750000059604645, |
|
"rewards/chosen": 0.5071097612380981, |
|
"rewards/margins": 0.21250581741333008, |
|
"rewards/rejected": 0.2946038842201233, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.20905923344947736, |
|
"grad_norm": 245.4785949332324, |
|
"learning_rate": 5.98062015503876e-07, |
|
"logits/chosen": -2.473501682281494, |
|
"logits/rejected": -2.4677951335906982, |
|
"logps/chosen": -70.63660430908203, |
|
"logps/rejected": -66.41564178466797, |
|
"loss": 0.6582, |
|
"rewards/accuracies": 0.32499998807907104, |
|
"rewards/chosen": 1.3620513677597046, |
|
"rewards/margins": 0.30580419301986694, |
|
"rewards/rejected": 1.0562469959259033, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.24390243902439024, |
|
"grad_norm": 268.1673884309266, |
|
"learning_rate": 5.883720930232558e-07, |
|
"logits/chosen": -2.48606538772583, |
|
"logits/rejected": -2.4808874130249023, |
|
"logps/chosen": -60.53791046142578, |
|
"logps/rejected": -65.51335906982422, |
|
"loss": 0.669, |
|
"rewards/accuracies": 0.3187499940395355, |
|
"rewards/chosen": 1.9600938558578491, |
|
"rewards/margins": 0.41606950759887695, |
|
"rewards/rejected": 1.5440242290496826, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.2787456445993031, |
|
"grad_norm": 266.9849201621109, |
|
"learning_rate": 5.786821705426356e-07, |
|
"logits/chosen": -2.44217586517334, |
|
"logits/rejected": -2.432021379470825, |
|
"logps/chosen": -71.77671813964844, |
|
"logps/rejected": -74.41423797607422, |
|
"loss": 0.6931, |
|
"rewards/accuracies": 0.3062500059604645, |
|
"rewards/chosen": 2.0607221126556396, |
|
"rewards/margins": 0.417347252368927, |
|
"rewards/rejected": 1.643375039100647, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.313588850174216, |
|
"grad_norm": 299.0911881932943, |
|
"learning_rate": 5.689922480620155e-07, |
|
"logits/chosen": -2.486508846282959, |
|
"logits/rejected": -2.5009188652038574, |
|
"logps/chosen": -62.32392120361328, |
|
"logps/rejected": -67.05072784423828, |
|
"loss": 0.745, |
|
"rewards/accuracies": 0.29374998807907104, |
|
"rewards/chosen": 1.938929557800293, |
|
"rewards/margins": 0.3153690993785858, |
|
"rewards/rejected": 1.6235605478286743, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.34843205574912894, |
|
"grad_norm": 316.56572577244293, |
|
"learning_rate": 5.593023255813953e-07, |
|
"logits/chosen": -2.4774889945983887, |
|
"logits/rejected": -2.4782986640930176, |
|
"logps/chosen": -71.84193420410156, |
|
"logps/rejected": -78.91864013671875, |
|
"loss": 0.7136, |
|
"rewards/accuracies": 0.32499998807907104, |
|
"rewards/chosen": 1.8174957036972046, |
|
"rewards/margins": 0.686999499797821, |
|
"rewards/rejected": 1.1304961442947388, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.34843205574912894, |
|
"eval_logits/chosen": -2.561777114868164, |
|
"eval_logits/rejected": -2.545793294906616, |
|
"eval_logps/chosen": -71.84760284423828, |
|
"eval_logps/rejected": -79.50099182128906, |
|
"eval_loss": 0.7108728885650635, |
|
"eval_rewards/accuracies": 0.3511904776096344, |
|
"eval_rewards/chosen": 1.4011276960372925, |
|
"eval_rewards/margins": 0.5386245846748352, |
|
"eval_rewards/rejected": 0.8625030517578125, |
|
"eval_runtime": 113.5305, |
|
"eval_samples_per_second": 17.616, |
|
"eval_steps_per_second": 0.555, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.3832752613240418, |
|
"grad_norm": 345.3192427581726, |
|
"learning_rate": 5.496124031007752e-07, |
|
"logits/chosen": -2.500793933868408, |
|
"logits/rejected": -2.4660115242004395, |
|
"logps/chosen": -72.03620910644531, |
|
"logps/rejected": -62.499176025390625, |
|
"loss": 0.7594, |
|
"rewards/accuracies": 0.28125, |
|
"rewards/chosen": 0.6009193658828735, |
|
"rewards/margins": 0.3298734426498413, |
|
"rewards/rejected": 0.2710459232330322, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.4181184668989547, |
|
"grad_norm": 243.7872401824864, |
|
"learning_rate": 5.399224806201551e-07, |
|
"logits/chosen": -2.5411460399627686, |
|
"logits/rejected": -2.510791778564453, |
|
"logps/chosen": -76.94505310058594, |
|
"logps/rejected": -67.3199462890625, |
|
"loss": 0.704, |
|
"rewards/accuracies": 0.29374998807907104, |
|
"rewards/chosen": 0.27375704050064087, |
|
"rewards/margins": 0.6173890829086304, |
|
"rewards/rejected": -0.3436321020126343, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.4529616724738676, |
|
"grad_norm": 432.4880968925023, |
|
"learning_rate": 5.302325581395349e-07, |
|
"logits/chosen": -2.5748209953308105, |
|
"logits/rejected": -2.5566792488098145, |
|
"logps/chosen": -83.82911682128906, |
|
"logps/rejected": -89.06166076660156, |
|
"loss": 0.7703, |
|
"rewards/accuracies": 0.35624998807907104, |
|
"rewards/chosen": 0.027380788698792458, |
|
"rewards/margins": 0.8505465388298035, |
|
"rewards/rejected": -0.823165774345398, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.4878048780487805, |
|
"grad_norm": 213.43504772096108, |
|
"learning_rate": 5.205426356589147e-07, |
|
"logits/chosen": -2.4663920402526855, |
|
"logits/rejected": -2.456141233444214, |
|
"logps/chosen": -79.90180969238281, |
|
"logps/rejected": -71.20616149902344, |
|
"loss": 0.7043, |
|
"rewards/accuracies": 0.3499999940395355, |
|
"rewards/chosen": 0.6591196060180664, |
|
"rewards/margins": 0.893332302570343, |
|
"rewards/rejected": -0.23421280086040497, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.5226480836236934, |
|
"grad_norm": 314.59295726808995, |
|
"learning_rate": 5.108527131782946e-07, |
|
"logits/chosen": -2.5461785793304443, |
|
"logits/rejected": -2.5047733783721924, |
|
"logps/chosen": -77.75875091552734, |
|
"logps/rejected": -79.21798706054688, |
|
"loss": 0.7534, |
|
"rewards/accuracies": 0.30000001192092896, |
|
"rewards/chosen": 0.9145609140396118, |
|
"rewards/margins": 0.7381798624992371, |
|
"rewards/rejected": 0.17638106644153595, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.5574912891986062, |
|
"grad_norm": 288.3107962146333, |
|
"learning_rate": 5.011627906976744e-07, |
|
"logits/chosen": -2.5273547172546387, |
|
"logits/rejected": -2.5466020107269287, |
|
"logps/chosen": -62.79814910888672, |
|
"logps/rejected": -71.2327880859375, |
|
"loss": 0.7312, |
|
"rewards/accuracies": 0.26249998807907104, |
|
"rewards/chosen": 0.9151515960693359, |
|
"rewards/margins": 0.5228849649429321, |
|
"rewards/rejected": 0.3922666311264038, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.5923344947735192, |
|
"grad_norm": 324.81976956861786, |
|
"learning_rate": 4.914728682170542e-07, |
|
"logits/chosen": -2.5521976947784424, |
|
"logits/rejected": -2.5384509563446045, |
|
"logps/chosen": -66.97964477539062, |
|
"logps/rejected": -75.58006286621094, |
|
"loss": 0.7191, |
|
"rewards/accuracies": 0.29374998807907104, |
|
"rewards/chosen": 1.0535058975219727, |
|
"rewards/margins": 0.664495587348938, |
|
"rewards/rejected": 0.3890102505683899, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.627177700348432, |
|
"grad_norm": 292.82490566453185, |
|
"learning_rate": 4.817829457364341e-07, |
|
"logits/chosen": -2.5870025157928467, |
|
"logits/rejected": -2.5761828422546387, |
|
"logps/chosen": -88.94615936279297, |
|
"logps/rejected": -84.6361083984375, |
|
"loss": 0.7825, |
|
"rewards/accuracies": 0.33125001192092896, |
|
"rewards/chosen": 1.5428403615951538, |
|
"rewards/margins": 0.5494336485862732, |
|
"rewards/rejected": 0.9934068918228149, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.662020905923345, |
|
"grad_norm": 234.8773688372816, |
|
"learning_rate": 4.7209302325581395e-07, |
|
"logits/chosen": -2.5796515941619873, |
|
"logits/rejected": -2.57136607170105, |
|
"logps/chosen": -68.72258758544922, |
|
"logps/rejected": -79.8985595703125, |
|
"loss": 0.7347, |
|
"rewards/accuracies": 0.2874999940395355, |
|
"rewards/chosen": 1.2343099117279053, |
|
"rewards/margins": 0.37060627341270447, |
|
"rewards/rejected": 0.8637038469314575, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.6968641114982579, |
|
"grad_norm": 439.89262521049136, |
|
"learning_rate": 4.6240310077519373e-07, |
|
"logits/chosen": -2.605844259262085, |
|
"logits/rejected": -2.612717390060425, |
|
"logps/chosen": -87.73692321777344, |
|
"logps/rejected": -90.65494537353516, |
|
"loss": 0.7461, |
|
"rewards/accuracies": 0.36250001192092896, |
|
"rewards/chosen": 1.5472986698150635, |
|
"rewards/margins": 0.8075531125068665, |
|
"rewards/rejected": 0.739745557308197, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.6968641114982579, |
|
"eval_logits/chosen": -2.5759267807006836, |
|
"eval_logits/rejected": -2.5600757598876953, |
|
"eval_logps/chosen": -72.48377227783203, |
|
"eval_logps/rejected": -80.43267822265625, |
|
"eval_loss": 0.7642679214477539, |
|
"eval_rewards/accuracies": 0.3273809552192688, |
|
"eval_rewards/chosen": 1.063955545425415, |
|
"eval_rewards/margins": 0.6952447295188904, |
|
"eval_rewards/rejected": 0.36871081590652466, |
|
"eval_runtime": 113.5756, |
|
"eval_samples_per_second": 17.609, |
|
"eval_steps_per_second": 0.555, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.7317073170731707, |
|
"grad_norm": 437.59995338780175, |
|
"learning_rate": 4.527131782945735e-07, |
|
"logits/chosen": -2.5814366340637207, |
|
"logits/rejected": -2.556798219680786, |
|
"logps/chosen": -67.51020812988281, |
|
"logps/rejected": -63.342933654785156, |
|
"loss": 0.7354, |
|
"rewards/accuracies": 0.35624998807907104, |
|
"rewards/chosen": 0.9586971402168274, |
|
"rewards/margins": 0.9261114001274109, |
|
"rewards/rejected": 0.03258571773767471, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.7665505226480837, |
|
"grad_norm": 270.9450930053244, |
|
"learning_rate": 4.4302325581395346e-07, |
|
"logits/chosen": -2.6260294914245605, |
|
"logits/rejected": -2.6077115535736084, |
|
"logps/chosen": -71.33647155761719, |
|
"logps/rejected": -70.29251861572266, |
|
"loss": 0.7602, |
|
"rewards/accuracies": 0.23125000298023224, |
|
"rewards/chosen": 1.1245003938674927, |
|
"rewards/margins": 0.3980458974838257, |
|
"rewards/rejected": 0.7264544367790222, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.8013937282229965, |
|
"grad_norm": 365.5474496656497, |
|
"learning_rate": 4.3333333333333335e-07, |
|
"logits/chosen": -2.6431806087493896, |
|
"logits/rejected": -2.623713493347168, |
|
"logps/chosen": -86.96244812011719, |
|
"logps/rejected": -87.527587890625, |
|
"loss": 0.7878, |
|
"rewards/accuracies": 0.375, |
|
"rewards/chosen": 1.6102378368377686, |
|
"rewards/margins": 1.2809720039367676, |
|
"rewards/rejected": 0.329265832901001, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.8362369337979094, |
|
"grad_norm": 311.80252726932434, |
|
"learning_rate": 4.2364341085271313e-07, |
|
"logits/chosen": -2.643277406692505, |
|
"logits/rejected": -2.609691619873047, |
|
"logps/chosen": -83.11528015136719, |
|
"logps/rejected": -77.69151306152344, |
|
"loss": 0.8033, |
|
"rewards/accuracies": 0.36250001192092896, |
|
"rewards/chosen": 1.9219558238983154, |
|
"rewards/margins": 0.6263306736946106, |
|
"rewards/rejected": 1.2956254482269287, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.8710801393728222, |
|
"grad_norm": 352.8980579572314, |
|
"learning_rate": 4.13953488372093e-07, |
|
"logits/chosen": -2.65216064453125, |
|
"logits/rejected": -2.617506504058838, |
|
"logps/chosen": -91.78350067138672, |
|
"logps/rejected": -87.70478820800781, |
|
"loss": 0.6512, |
|
"rewards/accuracies": 0.38749998807907104, |
|
"rewards/chosen": 1.942857027053833, |
|
"rewards/margins": 0.7549124956130981, |
|
"rewards/rejected": 1.1879446506500244, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.9059233449477352, |
|
"grad_norm": 265.7817814419997, |
|
"learning_rate": 4.0426356589147286e-07, |
|
"logits/chosen": -2.5617775917053223, |
|
"logits/rejected": -2.5759947299957275, |
|
"logps/chosen": -56.67157745361328, |
|
"logps/rejected": -64.38258361816406, |
|
"loss": 0.7944, |
|
"rewards/accuracies": 0.30000001192092896, |
|
"rewards/chosen": 1.26022469997406, |
|
"rewards/margins": 0.6000041961669922, |
|
"rewards/rejected": 0.6602205038070679, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.9407665505226481, |
|
"grad_norm": 513.5414604532187, |
|
"learning_rate": 3.9457364341085264e-07, |
|
"logits/chosen": -2.6499624252319336, |
|
"logits/rejected": -2.650109052658081, |
|
"logps/chosen": -66.84712219238281, |
|
"logps/rejected": -82.05715942382812, |
|
"loss": 0.7273, |
|
"rewards/accuracies": 0.3375000059604645, |
|
"rewards/chosen": 1.0775161981582642, |
|
"rewards/margins": 0.8259714841842651, |
|
"rewards/rejected": 0.2515445053577423, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.975609756097561, |
|
"grad_norm": 338.03485816068525, |
|
"learning_rate": 3.848837209302326e-07, |
|
"logits/chosen": -2.550506591796875, |
|
"logits/rejected": -2.529942512512207, |
|
"logps/chosen": -65.82142639160156, |
|
"logps/rejected": -70.44654846191406, |
|
"loss": 0.6677, |
|
"rewards/accuracies": 0.3187499940395355, |
|
"rewards/chosen": 1.0668643712997437, |
|
"rewards/margins": 0.8612996339797974, |
|
"rewards/rejected": 0.20556476712226868, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.0104529616724738, |
|
"grad_norm": 43.71804076421765, |
|
"learning_rate": 3.7519379844961237e-07, |
|
"logits/chosen": -2.5674736499786377, |
|
"logits/rejected": -2.54020357131958, |
|
"logps/chosen": -68.04539489746094, |
|
"logps/rejected": -65.61439514160156, |
|
"loss": 0.5817, |
|
"rewards/accuracies": 0.3812499940395355, |
|
"rewards/chosen": 2.1318130493164062, |
|
"rewards/margins": 2.503080368041992, |
|
"rewards/rejected": -0.37126731872558594, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.0452961672473868, |
|
"grad_norm": 14.731335253505609, |
|
"learning_rate": 3.6550387596899226e-07, |
|
"logits/chosen": -2.6065874099731445, |
|
"logits/rejected": -2.5927734375, |
|
"logps/chosen": -59.9798698425293, |
|
"logps/rejected": -77.5359115600586, |
|
"loss": 0.3949, |
|
"rewards/accuracies": 0.42500001192092896, |
|
"rewards/chosen": 3.481590747833252, |
|
"rewards/margins": 7.915454864501953, |
|
"rewards/rejected": -4.433863639831543, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.0452961672473868, |
|
"eval_logits/chosen": -2.6291775703430176, |
|
"eval_logits/rejected": -2.6134917736053467, |
|
"eval_logps/chosen": -74.10063934326172, |
|
"eval_logps/rejected": -82.3265380859375, |
|
"eval_loss": 0.7874619364738464, |
|
"eval_rewards/accuracies": 0.3472222089767456, |
|
"eval_rewards/chosen": 0.20701055228710175, |
|
"eval_rewards/margins": 0.8420494794845581, |
|
"eval_rewards/rejected": -0.6350388526916504, |
|
"eval_runtime": 113.6067, |
|
"eval_samples_per_second": 17.605, |
|
"eval_steps_per_second": 0.555, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.0801393728222997, |
|
"grad_norm": 4.834789646446964, |
|
"learning_rate": 3.558139534883721e-07, |
|
"logits/chosen": -2.578672409057617, |
|
"logits/rejected": -2.5800602436065674, |
|
"logps/chosen": -61.31939697265625, |
|
"logps/rejected": -86.94207000732422, |
|
"loss": 0.4017, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": 3.34814715385437, |
|
"rewards/margins": 9.546571731567383, |
|
"rewards/rejected": -6.198423862457275, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.1149825783972125, |
|
"grad_norm": 99.9815974317529, |
|
"learning_rate": 3.46124031007752e-07, |
|
"logits/chosen": -2.6197047233581543, |
|
"logits/rejected": -2.6059112548828125, |
|
"logps/chosen": -68.153076171875, |
|
"logps/rejected": -89.3502426147461, |
|
"loss": 0.3832, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": 3.07761812210083, |
|
"rewards/margins": 8.979107856750488, |
|
"rewards/rejected": -5.9014892578125, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.1498257839721253, |
|
"grad_norm": 65.16657207151003, |
|
"learning_rate": 3.3643410852713177e-07, |
|
"logits/chosen": -2.6080145835876465, |
|
"logits/rejected": -2.584524631500244, |
|
"logps/chosen": -77.22286224365234, |
|
"logps/rejected": -88.59306335449219, |
|
"loss": 0.3571, |
|
"rewards/accuracies": 0.5249999761581421, |
|
"rewards/chosen": 3.6562671661376953, |
|
"rewards/margins": 8.938417434692383, |
|
"rewards/rejected": -5.282149791717529, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.1846689895470384, |
|
"grad_norm": 20.064441334453125, |
|
"learning_rate": 3.267441860465116e-07, |
|
"logits/chosen": -2.58263897895813, |
|
"logits/rejected": -2.5866923332214355, |
|
"logps/chosen": -76.30878448486328, |
|
"logps/rejected": -106.47459411621094, |
|
"loss": 0.3758, |
|
"rewards/accuracies": 0.5562499761581421, |
|
"rewards/chosen": 3.9426658153533936, |
|
"rewards/margins": 9.874300003051758, |
|
"rewards/rejected": -5.931633949279785, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.2195121951219512, |
|
"grad_norm": 71.91053721699969, |
|
"learning_rate": 3.170542635658915e-07, |
|
"logits/chosen": -2.5975940227508545, |
|
"logits/rejected": -2.565171003341675, |
|
"logps/chosen": -61.74261474609375, |
|
"logps/rejected": -74.67215728759766, |
|
"loss": 0.3855, |
|
"rewards/accuracies": 0.48750001192092896, |
|
"rewards/chosen": 3.632841110229492, |
|
"rewards/margins": 8.399141311645508, |
|
"rewards/rejected": -4.766300201416016, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.254355400696864, |
|
"grad_norm": 78.70073770358177, |
|
"learning_rate": 3.073643410852713e-07, |
|
"logits/chosen": -2.59340238571167, |
|
"logits/rejected": -2.566377639770508, |
|
"logps/chosen": -65.64441680908203, |
|
"logps/rejected": -72.45703125, |
|
"loss": 0.3676, |
|
"rewards/accuracies": 0.44999998807907104, |
|
"rewards/chosen": 3.2890095710754395, |
|
"rewards/margins": 6.889220237731934, |
|
"rewards/rejected": -3.6002116203308105, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.289198606271777, |
|
"grad_norm": 42.412792949129724, |
|
"learning_rate": 2.9767441860465116e-07, |
|
"logits/chosen": -2.5649404525756836, |
|
"logits/rejected": -2.5833497047424316, |
|
"logps/chosen": -64.13664245605469, |
|
"logps/rejected": -83.73193359375, |
|
"loss": 0.399, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": 3.6725857257843018, |
|
"rewards/margins": 8.653525352478027, |
|
"rewards/rejected": -4.980940818786621, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.32404181184669, |
|
"grad_norm": 89.8718675271705, |
|
"learning_rate": 2.87984496124031e-07, |
|
"logits/chosen": -2.576333999633789, |
|
"logits/rejected": -2.5752272605895996, |
|
"logps/chosen": -77.921875, |
|
"logps/rejected": -98.92057800292969, |
|
"loss": 0.3699, |
|
"rewards/accuracies": 0.5562499761581421, |
|
"rewards/chosen": 4.744899749755859, |
|
"rewards/margins": 11.345683097839355, |
|
"rewards/rejected": -6.600783348083496, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.3588850174216027, |
|
"grad_norm": 53.88569450482403, |
|
"learning_rate": 2.7829457364341084e-07, |
|
"logits/chosen": -2.6578681468963623, |
|
"logits/rejected": -2.6431689262390137, |
|
"logps/chosen": -59.8434944152832, |
|
"logps/rejected": -80.90740203857422, |
|
"loss": 0.3745, |
|
"rewards/accuracies": 0.4437499940395355, |
|
"rewards/chosen": 3.790491819381714, |
|
"rewards/margins": 8.903260231018066, |
|
"rewards/rejected": -5.112768650054932, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.3937282229965158, |
|
"grad_norm": 103.59808023860636, |
|
"learning_rate": 2.686046511627907e-07, |
|
"logits/chosen": -2.638619899749756, |
|
"logits/rejected": -2.609290361404419, |
|
"logps/chosen": -78.6235122680664, |
|
"logps/rejected": -108.10832214355469, |
|
"loss": 0.3838, |
|
"rewards/accuracies": 0.5249999761581421, |
|
"rewards/chosen": 3.987473249435425, |
|
"rewards/margins": 9.130887985229492, |
|
"rewards/rejected": -5.1434149742126465, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.3937282229965158, |
|
"eval_logits/chosen": -2.6421802043914795, |
|
"eval_logits/rejected": -2.626558542251587, |
|
"eval_logps/chosen": -73.66181945800781, |
|
"eval_logps/rejected": -82.45710754394531, |
|
"eval_loss": 0.871368408203125, |
|
"eval_rewards/accuracies": 0.329365074634552, |
|
"eval_rewards/chosen": 0.4395846724510193, |
|
"eval_rewards/margins": 1.1438220739364624, |
|
"eval_rewards/rejected": -0.7042374610900879, |
|
"eval_runtime": 113.5108, |
|
"eval_samples_per_second": 17.619, |
|
"eval_steps_per_second": 0.555, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.4285714285714286, |
|
"grad_norm": 0.4652582217397593, |
|
"learning_rate": 2.589147286821705e-07, |
|
"logits/chosen": -2.625115156173706, |
|
"logits/rejected": -2.6144914627075195, |
|
"logps/chosen": -74.08203887939453, |
|
"logps/rejected": -89.65892791748047, |
|
"loss": 0.5043, |
|
"rewards/accuracies": 0.48750001192092896, |
|
"rewards/chosen": 4.521576881408691, |
|
"rewards/margins": 9.441361427307129, |
|
"rewards/rejected": -4.919784069061279, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.4634146341463414, |
|
"grad_norm": 14.542963331044218, |
|
"learning_rate": 2.492248062015504e-07, |
|
"logits/chosen": -2.6700663566589355, |
|
"logits/rejected": -2.671051502227783, |
|
"logps/chosen": -70.02223205566406, |
|
"logps/rejected": -93.91789245605469, |
|
"loss": 0.3962, |
|
"rewards/accuracies": 0.4437499940395355, |
|
"rewards/chosen": 3.217597484588623, |
|
"rewards/margins": 8.012170791625977, |
|
"rewards/rejected": -4.794573783874512, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.4982578397212545, |
|
"grad_norm": 7.616609761831896, |
|
"learning_rate": 2.3953488372093024e-07, |
|
"logits/chosen": -2.6445345878601074, |
|
"logits/rejected": -2.630586624145508, |
|
"logps/chosen": -61.7302360534668, |
|
"logps/rejected": -81.66353607177734, |
|
"loss": 0.3812, |
|
"rewards/accuracies": 0.45625001192092896, |
|
"rewards/chosen": 3.6740145683288574, |
|
"rewards/margins": 8.53648567199707, |
|
"rewards/rejected": -4.862471580505371, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.533101045296167, |
|
"grad_norm": 19.40227524240624, |
|
"learning_rate": 2.2984496124031007e-07, |
|
"logits/chosen": -2.639118194580078, |
|
"logits/rejected": -2.648719310760498, |
|
"logps/chosen": -61.67310333251953, |
|
"logps/rejected": -81.26171875, |
|
"loss": 0.3814, |
|
"rewards/accuracies": 0.4312500059604645, |
|
"rewards/chosen": 4.401946067810059, |
|
"rewards/margins": 8.5157470703125, |
|
"rewards/rejected": -4.113801002502441, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.5679442508710801, |
|
"grad_norm": 11.176099808562324, |
|
"learning_rate": 2.201550387596899e-07, |
|
"logits/chosen": -2.7013635635375977, |
|
"logits/rejected": -2.6666698455810547, |
|
"logps/chosen": -81.5969009399414, |
|
"logps/rejected": -96.87474060058594, |
|
"loss": 0.3926, |
|
"rewards/accuracies": 0.518750011920929, |
|
"rewards/chosen": 4.899897575378418, |
|
"rewards/margins": 10.683164596557617, |
|
"rewards/rejected": -5.783267021179199, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.6027874564459932, |
|
"grad_norm": 123.87369972777346, |
|
"learning_rate": 2.1046511627906974e-07, |
|
"logits/chosen": -2.6888933181762695, |
|
"logits/rejected": -2.6685612201690674, |
|
"logps/chosen": -66.90978240966797, |
|
"logps/rejected": -88.9155044555664, |
|
"loss": 0.4047, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 4.9129533767700195, |
|
"rewards/margins": 11.225677490234375, |
|
"rewards/rejected": -6.312723636627197, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.6376306620209058, |
|
"grad_norm": 3.43139199879915, |
|
"learning_rate": 2.0077519379844966e-07, |
|
"logits/chosen": -2.6837058067321777, |
|
"logits/rejected": -2.674848794937134, |
|
"logps/chosen": -54.61207962036133, |
|
"logps/rejected": -79.80362701416016, |
|
"loss": 0.3979, |
|
"rewards/accuracies": 0.40625, |
|
"rewards/chosen": 3.5458691120147705, |
|
"rewards/margins": 7.8562331199646, |
|
"rewards/rejected": -4.310364246368408, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.6724738675958188, |
|
"grad_norm": 12.439142544759255, |
|
"learning_rate": 1.9108527131782944e-07, |
|
"logits/chosen": -2.695263385772705, |
|
"logits/rejected": -2.6781885623931885, |
|
"logps/chosen": -48.82516860961914, |
|
"logps/rejected": -57.3747444152832, |
|
"loss": 0.4033, |
|
"rewards/accuracies": 0.3687500059604645, |
|
"rewards/chosen": 2.775264024734497, |
|
"rewards/margins": 5.633866786956787, |
|
"rewards/rejected": -2.8586020469665527, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.7073170731707317, |
|
"grad_norm": 40.080099001713826, |
|
"learning_rate": 1.8139534883720925e-07, |
|
"logits/chosen": -2.6758790016174316, |
|
"logits/rejected": -2.6650328636169434, |
|
"logps/chosen": -66.64186096191406, |
|
"logps/rejected": -78.38505554199219, |
|
"loss": 0.4696, |
|
"rewards/accuracies": 0.38749998807907104, |
|
"rewards/chosen": 3.4873664379119873, |
|
"rewards/margins": 7.898496150970459, |
|
"rewards/rejected": -4.411130428314209, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.7421602787456445, |
|
"grad_norm": 47.95492405495267, |
|
"learning_rate": 1.7170542635658914e-07, |
|
"logits/chosen": -2.593526601791382, |
|
"logits/rejected": -2.5879125595092773, |
|
"logps/chosen": -68.72013854980469, |
|
"logps/rejected": -92.43871307373047, |
|
"loss": 0.371, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": 4.010916709899902, |
|
"rewards/margins": 8.630084991455078, |
|
"rewards/rejected": -4.619168281555176, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.7421602787456445, |
|
"eval_logits/chosen": -2.7067761421203613, |
|
"eval_logits/rejected": -2.691006898880005, |
|
"eval_logps/chosen": -73.18505859375, |
|
"eval_logps/rejected": -82.15363311767578, |
|
"eval_loss": 0.863908588886261, |
|
"eval_rewards/accuracies": 0.3392857015132904, |
|
"eval_rewards/chosen": 0.6922710537910461, |
|
"eval_rewards/margins": 1.2356635332107544, |
|
"eval_rewards/rejected": -0.5433923602104187, |
|
"eval_runtime": 113.7753, |
|
"eval_samples_per_second": 17.579, |
|
"eval_steps_per_second": 0.554, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.7770034843205575, |
|
"grad_norm": 43.89133813372592, |
|
"learning_rate": 1.6201550387596898e-07, |
|
"logits/chosen": -2.665417432785034, |
|
"logits/rejected": -2.647148609161377, |
|
"logps/chosen": -63.23058319091797, |
|
"logps/rejected": -77.56340026855469, |
|
"loss": 0.3821, |
|
"rewards/accuracies": 0.45625001192092896, |
|
"rewards/chosen": 3.9886889457702637, |
|
"rewards/margins": 8.217727661132812, |
|
"rewards/rejected": -4.229039192199707, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.8118466898954704, |
|
"grad_norm": 26.582330424497826, |
|
"learning_rate": 1.523255813953488e-07, |
|
"logits/chosen": -2.671867847442627, |
|
"logits/rejected": -2.666949510574341, |
|
"logps/chosen": -66.01771545410156, |
|
"logps/rejected": -86.70332336425781, |
|
"loss": 0.382, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": 4.374355316162109, |
|
"rewards/margins": 9.404329299926758, |
|
"rewards/rejected": -5.029973983764648, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.8466898954703832, |
|
"grad_norm": 16.01608311357976, |
|
"learning_rate": 1.426356589147287e-07, |
|
"logits/chosen": -2.6440272331237793, |
|
"logits/rejected": -2.6383345127105713, |
|
"logps/chosen": -63.14166259765625, |
|
"logps/rejected": -80.24067687988281, |
|
"loss": 0.3916, |
|
"rewards/accuracies": 0.4625000059604645, |
|
"rewards/chosen": 4.34307336807251, |
|
"rewards/margins": 9.047937393188477, |
|
"rewards/rejected": -4.704863548278809, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.8815331010452963, |
|
"grad_norm": 503.79684025145957, |
|
"learning_rate": 1.3294573643410851e-07, |
|
"logits/chosen": -2.6588096618652344, |
|
"logits/rejected": -2.6732683181762695, |
|
"logps/chosen": -53.71875, |
|
"logps/rejected": -84.29718780517578, |
|
"loss": 0.46, |
|
"rewards/accuracies": 0.44999998807907104, |
|
"rewards/chosen": 5.3063740730285645, |
|
"rewards/margins": 11.112980842590332, |
|
"rewards/rejected": -5.806607246398926, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.916376306620209, |
|
"grad_norm": 1.5289785290366227, |
|
"learning_rate": 1.2325581395348835e-07, |
|
"logits/chosen": -2.636369228363037, |
|
"logits/rejected": -2.6182668209075928, |
|
"logps/chosen": -79.08727264404297, |
|
"logps/rejected": -97.51543426513672, |
|
"loss": 0.3689, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 6.226182460784912, |
|
"rewards/margins": 11.859712600708008, |
|
"rewards/rejected": -5.633530139923096, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.951219512195122, |
|
"grad_norm": 96.76132190493311, |
|
"learning_rate": 1.1356589147286824e-07, |
|
"logits/chosen": -2.64457368850708, |
|
"logits/rejected": -2.651458740234375, |
|
"logps/chosen": -55.0025520324707, |
|
"logps/rejected": -76.64137268066406, |
|
"loss": 0.3985, |
|
"rewards/accuracies": 0.4437499940395355, |
|
"rewards/chosen": 4.620251655578613, |
|
"rewards/margins": 7.983218193054199, |
|
"rewards/rejected": -3.362966537475586, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.986062717770035, |
|
"grad_norm": 13.582049528683124, |
|
"learning_rate": 1.0387596899224806e-07, |
|
"logits/chosen": -2.7313754558563232, |
|
"logits/rejected": -2.702322483062744, |
|
"logps/chosen": -56.71698760986328, |
|
"logps/rejected": -67.86329650878906, |
|
"loss": 0.3789, |
|
"rewards/accuracies": 0.4437499940395355, |
|
"rewards/chosen": 4.503007411956787, |
|
"rewards/margins": 7.31237268447876, |
|
"rewards/rejected": -2.8093647956848145, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"step": 574, |
|
"total_flos": 0.0, |
|
"train_loss": 0.5567006942287139, |
|
"train_runtime": 6469.9642, |
|
"train_samples_per_second": 5.669, |
|
"train_steps_per_second": 0.089 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 574, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|