rlced-conifer-zephyr-7b-dpo-full / trainer_state.json
NicholasCorrado's picture
Model save
020d6ef verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9963674104826155,
"eval_steps": 1000,
"global_step": 240,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0041515308770108976,
"grad_norm": 6.2112270362028115,
"learning_rate": 2.083333333333333e-08,
"logits/chosen": -2.6958627700805664,
"logits/rejected": -2.650068759918213,
"logps/chosen": -403.2443542480469,
"logps/rejected": -397.0637512207031,
"loss": 0.6931,
"rewards/accuracies": 0.0,
"rewards/chosen": 0.0,
"rewards/margins": 0.0,
"rewards/rejected": 0.0,
"step": 1
},
{
"epoch": 0.04151530877010898,
"grad_norm": 6.539288745411817,
"learning_rate": 2.0833333333333333e-07,
"logits/chosen": -2.690112352371216,
"logits/rejected": -2.6779024600982666,
"logps/chosen": -365.9743957519531,
"logps/rejected": -393.799560546875,
"loss": 0.6925,
"rewards/accuracies": 0.5303819179534912,
"rewards/chosen": 0.0005260682082735002,
"rewards/margins": 0.0014649044023826718,
"rewards/rejected": -0.0009388362523168325,
"step": 10
},
{
"epoch": 0.08303061754021795,
"grad_norm": 6.876570939434779,
"learning_rate": 4.1666666666666667e-07,
"logits/chosen": -2.6719064712524414,
"logits/rejected": -2.6511917114257812,
"logps/chosen": -357.5186462402344,
"logps/rejected": -380.5492858886719,
"loss": 0.6749,
"rewards/accuracies": 0.848437488079071,
"rewards/chosen": 0.018601149320602417,
"rewards/margins": 0.03796255216002464,
"rewards/rejected": -0.019361400976777077,
"step": 20
},
{
"epoch": 0.12454592631032693,
"grad_norm": 8.242981852125217,
"learning_rate": 4.990486745229364e-07,
"logits/chosen": -2.653783082962036,
"logits/rejected": -2.639296293258667,
"logps/chosen": -355.69207763671875,
"logps/rejected": -415.68768310546875,
"loss": 0.5772,
"rewards/accuracies": 0.867968738079071,
"rewards/chosen": 0.0034460537135601044,
"rewards/margins": 0.2814808487892151,
"rewards/rejected": -0.2780347764492035,
"step": 30
},
{
"epoch": 0.1660612350804359,
"grad_norm": 14.409074768095572,
"learning_rate": 4.932612176449559e-07,
"logits/chosen": -2.6512644290924072,
"logits/rejected": -2.613352060317993,
"logps/chosen": -496.20770263671875,
"logps/rejected": -619.66943359375,
"loss": 0.4041,
"rewards/accuracies": 0.848437488079071,
"rewards/chosen": -1.1875219345092773,
"rewards/margins": 1.0682841539382935,
"rewards/rejected": -2.2558062076568604,
"step": 40
},
{
"epoch": 0.2075765438505449,
"grad_norm": 14.911382681454745,
"learning_rate": 4.823368810567056e-07,
"logits/chosen": -2.64001202583313,
"logits/rejected": -2.6261935234069824,
"logps/chosen": -600.037109375,
"logps/rejected": -856.2901611328125,
"loss": 0.3072,
"rewards/accuracies": 0.86328125,
"rewards/chosen": -2.279493570327759,
"rewards/margins": 2.3056600093841553,
"rewards/rejected": -4.5851545333862305,
"step": 50
},
{
"epoch": 0.24909185262065386,
"grad_norm": 18.472917689948186,
"learning_rate": 4.6650635094610966e-07,
"logits/chosen": -2.5924713611602783,
"logits/rejected": -2.579529285430908,
"logps/chosen": -587.2591552734375,
"logps/rejected": -904.9329833984375,
"loss": 0.2658,
"rewards/accuracies": 0.88671875,
"rewards/chosen": -2.1298751831054688,
"rewards/margins": 2.945924758911133,
"rewards/rejected": -5.07580041885376,
"step": 60
},
{
"epoch": 0.29060716139076287,
"grad_norm": 12.678464521005063,
"learning_rate": 4.461039162298939e-07,
"logits/chosen": -2.2207727432250977,
"logits/rejected": -1.98198664188385,
"logps/chosen": -591.1898803710938,
"logps/rejected": -989.9195556640625,
"loss": 0.2276,
"rewards/accuracies": 0.9046875238418579,
"rewards/chosen": -2.155735969543457,
"rewards/margins": 3.7345290184020996,
"rewards/rejected": -5.890264987945557,
"step": 70
},
{
"epoch": 0.3321224701608718,
"grad_norm": 14.772965954120863,
"learning_rate": 4.2156040946718343e-07,
"logits/chosen": -1.5045950412750244,
"logits/rejected": -0.8392450213432312,
"logps/chosen": -616.548828125,
"logps/rejected": -1064.140869140625,
"loss": 0.1988,
"rewards/accuracies": 0.914843738079071,
"rewards/chosen": -2.4090540409088135,
"rewards/margins": 4.233697891235352,
"rewards/rejected": -6.642751216888428,
"step": 80
},
{
"epoch": 0.3736377789309808,
"grad_norm": 16.72422342704544,
"learning_rate": 3.933941090877615e-07,
"logits/chosen": -1.3608930110931396,
"logits/rejected": -0.5576863288879395,
"logps/chosen": -595.0042724609375,
"logps/rejected": -1064.654052734375,
"loss": 0.2026,
"rewards/accuracies": 0.91015625,
"rewards/chosen": -2.351076126098633,
"rewards/margins": 4.467952728271484,
"rewards/rejected": -6.819028377532959,
"step": 90
},
{
"epoch": 0.4151530877010898,
"grad_norm": 9.743291843330928,
"learning_rate": 3.6219979505011555e-07,
"logits/chosen": -0.8342965245246887,
"logits/rejected": -0.012925502844154835,
"logps/chosen": -646.4923706054688,
"logps/rejected": -1148.498779296875,
"loss": 0.1867,
"rewards/accuracies": 0.9195312261581421,
"rewards/chosen": -2.8086471557617188,
"rewards/margins": 4.717282295227051,
"rewards/rejected": -7.525929927825928,
"step": 100
},
{
"epoch": 0.4566683964711988,
"grad_norm": 11.81658995353539,
"learning_rate": 3.286361890379034e-07,
"logits/chosen": -0.1246568113565445,
"logits/rejected": 0.6437393426895142,
"logps/chosen": -648.837890625,
"logps/rejected": -1166.3304443359375,
"loss": 0.1785,
"rewards/accuracies": 0.917187511920929,
"rewards/chosen": -2.7460227012634277,
"rewards/margins": 4.996233940124512,
"rewards/rejected": -7.742256164550781,
"step": 110
},
{
"epoch": 0.49818370524130773,
"grad_norm": 9.286297855632633,
"learning_rate": 2.934120444167326e-07,
"logits/chosen": -0.19480545818805695,
"logits/rejected": 0.7439680695533752,
"logps/chosen": -654.71875,
"logps/rejected": -1212.785888671875,
"loss": 0.1786,
"rewards/accuracies": 0.91796875,
"rewards/chosen": -2.892089366912842,
"rewards/margins": 5.356635093688965,
"rewards/rejected": -8.248723983764648,
"step": 120
},
{
"epoch": 0.5396990140114167,
"grad_norm": 12.2921471616094,
"learning_rate": 2.5727117968577785e-07,
"logits/chosen": -0.4111465513706207,
"logits/rejected": 0.8093876838684082,
"logps/chosen": -645.12109375,
"logps/rejected": -1192.375244140625,
"loss": 0.1676,
"rewards/accuracies": 0.9234374761581421,
"rewards/chosen": -2.742527961730957,
"rewards/margins": 5.263998031616211,
"rewards/rejected": -8.006526947021484,
"step": 130
},
{
"epoch": 0.5812143227815257,
"grad_norm": 17.430656596120464,
"learning_rate": 2.209767714686924e-07,
"logits/chosen": -0.029587041586637497,
"logits/rejected": 1.3743274211883545,
"logps/chosen": -648.7828979492188,
"logps/rejected": -1265.0289306640625,
"loss": 0.1632,
"rewards/accuracies": 0.925000011920929,
"rewards/chosen": -2.7970809936523438,
"rewards/margins": 5.923464775085449,
"rewards/rejected": -8.720545768737793,
"step": 140
},
{
"epoch": 0.6227296315516346,
"grad_norm": 9.425394796778923,
"learning_rate": 1.8529523872436977e-07,
"logits/chosen": -0.18061885237693787,
"logits/rejected": 1.3279445171356201,
"logps/chosen": -643.9744873046875,
"logps/rejected": -1251.9317626953125,
"loss": 0.1642,
"rewards/accuracies": 0.9273437261581421,
"rewards/chosen": -2.7806482315063477,
"rewards/margins": 5.8400492668151855,
"rewards/rejected": -8.620697975158691,
"step": 150
},
{
"epoch": 0.6642449403217436,
"grad_norm": 8.084677706157798,
"learning_rate": 1.5098005849021078e-07,
"logits/chosen": -0.20541512966156006,
"logits/rejected": 1.3262333869934082,
"logps/chosen": -645.2593994140625,
"logps/rejected": -1234.405517578125,
"loss": 0.1618,
"rewards/accuracies": 0.907031238079071,
"rewards/chosen": -2.804506301879883,
"rewards/margins": 5.688388347625732,
"rewards/rejected": -8.492895126342773,
"step": 160
},
{
"epoch": 0.7057602490918526,
"grad_norm": 10.460249704683227,
"learning_rate": 1.1875585491635998e-07,
"logits/chosen": -0.6104969382286072,
"logits/rejected": 1.1572777032852173,
"logps/chosen": -623.3721923828125,
"logps/rejected": -1259.294189453125,
"loss": 0.1573,
"rewards/accuracies": 0.9203125238418579,
"rewards/chosen": -2.5536274909973145,
"rewards/margins": 6.065129280090332,
"rewards/rejected": -8.618757247924805,
"step": 170
},
{
"epoch": 0.7472755578619616,
"grad_norm": 12.223705856186251,
"learning_rate": 8.930309757836516e-08,
"logits/chosen": -0.717302680015564,
"logits/rejected": 1.1799119710922241,
"logps/chosen": -645.1275634765625,
"logps/rejected": -1283.7489013671875,
"loss": 0.1616,
"rewards/accuracies": 0.934374988079071,
"rewards/chosen": -2.7811636924743652,
"rewards/margins": 6.2163591384887695,
"rewards/rejected": -8.997522354125977,
"step": 180
},
{
"epoch": 0.7887908666320705,
"grad_norm": 9.676735610238229,
"learning_rate": 6.324373218975104e-08,
"logits/chosen": -0.8794542551040649,
"logits/rejected": 0.99409419298172,
"logps/chosen": -619.7286987304688,
"logps/rejected": -1276.9617919921875,
"loss": 0.1608,
"rewards/accuracies": 0.925000011920929,
"rewards/chosen": -2.615447521209717,
"rewards/margins": 6.301668643951416,
"rewards/rejected": -8.917116165161133,
"step": 190
},
{
"epoch": 0.8303061754021795,
"grad_norm": 11.57113935464218,
"learning_rate": 4.112804714676593e-08,
"logits/chosen": -0.7840622663497925,
"logits/rejected": 1.0056122541427612,
"logps/chosen": -637.8081665039062,
"logps/rejected": -1261.2838134765625,
"loss": 0.162,
"rewards/accuracies": 0.9281250238418579,
"rewards/chosen": -2.6473495960235596,
"rewards/margins": 6.003285884857178,
"rewards/rejected": -8.650635719299316,
"step": 200
},
{
"epoch": 0.8718214841722886,
"grad_norm": 13.665084975047886,
"learning_rate": 2.3423053240837514e-08,
"logits/chosen": -0.6102296710014343,
"logits/rejected": 1.163267731666565,
"logps/chosen": -638.1444091796875,
"logps/rejected": -1243.947265625,
"loss": 0.1576,
"rewards/accuracies": 0.918749988079071,
"rewards/chosen": -2.7065281867980957,
"rewards/margins": 5.835700988769531,
"rewards/rejected": -8.542229652404785,
"step": 210
},
{
"epoch": 0.9133367929423976,
"grad_norm": 10.802886491160189,
"learning_rate": 1.0502621921127774e-08,
"logits/chosen": -0.5309673547744751,
"logits/rejected": 1.326080560684204,
"logps/chosen": -637.7506713867188,
"logps/rejected": -1279.3565673828125,
"loss": 0.1579,
"rewards/accuracies": 0.922656238079071,
"rewards/chosen": -2.6862006187438965,
"rewards/margins": 6.1917243003845215,
"rewards/rejected": -8.877924919128418,
"step": 220
},
{
"epoch": 0.9548521017125065,
"grad_norm": 10.752955125049585,
"learning_rate": 2.639590354763882e-09,
"logits/chosen": -0.5724472403526306,
"logits/rejected": 1.2889587879180908,
"logps/chosen": -629.98779296875,
"logps/rejected": -1255.991943359375,
"loss": 0.1579,
"rewards/accuracies": 0.94140625,
"rewards/chosen": -2.661499500274658,
"rewards/margins": 6.085452556610107,
"rewards/rejected": -8.746953010559082,
"step": 230
},
{
"epoch": 0.9963674104826155,
"grad_norm": 9.420585318893654,
"learning_rate": 0.0,
"logits/chosen": -0.5639629364013672,
"logits/rejected": 1.3717336654663086,
"logps/chosen": -631.6014404296875,
"logps/rejected": -1273.350341796875,
"loss": 0.1514,
"rewards/accuracies": 0.934374988079071,
"rewards/chosen": -2.7341790199279785,
"rewards/margins": 6.176712989807129,
"rewards/rejected": -8.91089153289795,
"step": 240
},
{
"epoch": 0.9963674104826155,
"step": 240,
"total_flos": 0.0,
"train_loss": 0.250737202167511,
"train_runtime": 29028.2374,
"train_samples_per_second": 4.248,
"train_steps_per_second": 0.008
}
],
"logging_steps": 10,
"max_steps": 240,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}