SELM-Llama-3-8B-Instruct-iter-1 / trainer_state.json
ZhangShenao's picture
Model save
f44f70b verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9984301412872841,
"eval_steps": 500,
"global_step": 159,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.006279434850863423,
"grad_norm": 5.557445139643298,
"learning_rate": 3.125e-08,
"logits/chosen": 0.18015038967132568,
"logits/rejected": 0.2519298493862152,
"logps/chosen": -297.10906982421875,
"logps/pi_response": -130.58929443359375,
"logps/ref_response": -130.58929443359375,
"logps/rejected": -316.44769287109375,
"loss": 0.6931,
"rewards/accuracies": 0.0,
"rewards/chosen": 0.0,
"rewards/margins": 0.0,
"rewards/rejected": 0.0,
"step": 1
},
{
"epoch": 0.06279434850863422,
"grad_norm": 5.880023460230945,
"learning_rate": 3.1249999999999997e-07,
"logits/chosen": 0.16638809442520142,
"logits/rejected": 0.3159521222114563,
"logps/chosen": -243.77159118652344,
"logps/pi_response": -120.18633270263672,
"logps/ref_response": -120.15902709960938,
"logps/rejected": -281.09716796875,
"loss": 0.6928,
"rewards/accuracies": 0.4444444477558136,
"rewards/chosen": -0.0007160517852753401,
"rewards/margins": 0.00040519109461456537,
"rewards/rejected": -0.0011212429963052273,
"step": 10
},
{
"epoch": 0.12558869701726844,
"grad_norm": 6.011032264913819,
"learning_rate": 4.990353313429303e-07,
"logits/chosen": 0.1316775530576706,
"logits/rejected": 0.32217010855674744,
"logps/chosen": -244.0759735107422,
"logps/pi_response": -121.6043701171875,
"logps/ref_response": -121.85536193847656,
"logps/rejected": -266.6847229003906,
"loss": 0.6884,
"rewards/accuracies": 0.6937500238418579,
"rewards/chosen": -0.011644470505416393,
"rewards/margins": 0.011915634386241436,
"rewards/rejected": -0.02356010302901268,
"step": 20
},
{
"epoch": 0.18838304552590268,
"grad_norm": 6.296068063682766,
"learning_rate": 4.882681251368548e-07,
"logits/chosen": 0.17213600873947144,
"logits/rejected": 0.3042981028556824,
"logps/chosen": -244.4438018798828,
"logps/pi_response": -109.73341369628906,
"logps/ref_response": -110.8894271850586,
"logps/rejected": -290.1441650390625,
"loss": 0.6685,
"rewards/accuracies": 0.6812499761581421,
"rewards/chosen": -0.043184880167245865,
"rewards/margins": 0.054663728922605515,
"rewards/rejected": -0.09784860908985138,
"step": 30
},
{
"epoch": 0.25117739403453687,
"grad_norm": 5.762840753548806,
"learning_rate": 4.6604720940421207e-07,
"logits/chosen": 0.2107941210269928,
"logits/rejected": 0.39838385581970215,
"logps/chosen": -287.46002197265625,
"logps/pi_response": -125.36665344238281,
"logps/ref_response": -129.86325073242188,
"logps/rejected": -316.40423583984375,
"loss": 0.6349,
"rewards/accuracies": 0.7124999761581421,
"rewards/chosen": -0.08985555917024612,
"rewards/margins": 0.1335289627313614,
"rewards/rejected": -0.22338449954986572,
"step": 40
},
{
"epoch": 0.3139717425431711,
"grad_norm": 5.876442100853928,
"learning_rate": 4.3344075855595097e-07,
"logits/chosen": 0.3732234835624695,
"logits/rejected": 0.5105798840522766,
"logps/chosen": -247.15914916992188,
"logps/pi_response": -109.07597351074219,
"logps/ref_response": -116.5090560913086,
"logps/rejected": -310.7102966308594,
"loss": 0.606,
"rewards/accuracies": 0.737500011920929,
"rewards/chosen": -0.11287806183099747,
"rewards/margins": 0.24187734723091125,
"rewards/rejected": -0.3547554314136505,
"step": 50
},
{
"epoch": 0.37676609105180536,
"grad_norm": 9.315356312481537,
"learning_rate": 3.920161866827889e-07,
"logits/chosen": 0.5162631273269653,
"logits/rejected": 0.6798213124275208,
"logps/chosen": -268.0201721191406,
"logps/pi_response": -116.46971130371094,
"logps/ref_response": -119.4989242553711,
"logps/rejected": -347.75079345703125,
"loss": 0.5814,
"rewards/accuracies": 0.75,
"rewards/chosen": -0.20100387930870056,
"rewards/margins": 0.41021671891212463,
"rewards/rejected": -0.6112205386161804,
"step": 60
},
{
"epoch": 0.43956043956043955,
"grad_norm": 7.144146562974543,
"learning_rate": 3.4376480090239047e-07,
"logits/chosen": 0.6009117960929871,
"logits/rejected": 0.767359733581543,
"logps/chosen": -236.7908172607422,
"logps/pi_response": -114.82955169677734,
"logps/ref_response": -116.70068359375,
"logps/rejected": -368.27850341796875,
"loss": 0.5568,
"rewards/accuracies": 0.699999988079071,
"rewards/chosen": -0.2643741965293884,
"rewards/margins": 0.5534776449203491,
"rewards/rejected": -0.8178518414497375,
"step": 70
},
{
"epoch": 0.5023547880690737,
"grad_norm": 7.105190568030611,
"learning_rate": 2.910060778827554e-07,
"logits/chosen": 0.47289925813674927,
"logits/rejected": 0.8142975568771362,
"logps/chosen": -325.69622802734375,
"logps/pi_response": -129.21812438964844,
"logps/ref_response": -127.53900146484375,
"logps/rejected": -348.5531311035156,
"loss": 0.5849,
"rewards/accuracies": 0.6499999761581421,
"rewards/chosen": -0.3693317174911499,
"rewards/margins": 0.37722334265708923,
"rewards/rejected": -0.7465550303459167,
"step": 80
},
{
"epoch": 0.565149136577708,
"grad_norm": 6.295924145315453,
"learning_rate": 2.3627616503391812e-07,
"logits/chosen": 0.43639689683914185,
"logits/rejected": 0.8138043284416199,
"logps/chosen": -301.4584045410156,
"logps/pi_response": -135.66107177734375,
"logps/ref_response": -129.38760375976562,
"logps/rejected": -421.0269470214844,
"loss": 0.5432,
"rewards/accuracies": 0.7562500238418579,
"rewards/chosen": -0.29649025201797485,
"rewards/margins": 0.6172370910644531,
"rewards/rejected": -0.9137271642684937,
"step": 90
},
{
"epoch": 0.6279434850863422,
"grad_norm": 7.528550263483296,
"learning_rate": 1.8220596619089573e-07,
"logits/chosen": 0.6732058525085449,
"logits/rejected": 0.8459190130233765,
"logps/chosen": -283.48883056640625,
"logps/pi_response": -121.66845703125,
"logps/ref_response": -114.0061264038086,
"logps/rejected": -304.6356201171875,
"loss": 0.5426,
"rewards/accuracies": 0.6875,
"rewards/chosen": -0.30830463767051697,
"rewards/margins": 0.3349376320838928,
"rewards/rejected": -0.6432422995567322,
"step": 100
},
{
"epoch": 0.6907378335949764,
"grad_norm": 7.109656663434607,
"learning_rate": 1.3139467229135998e-07,
"logits/chosen": 0.6394161581993103,
"logits/rejected": 0.9435567855834961,
"logps/chosen": -298.015869140625,
"logps/pi_response": -136.70449829101562,
"logps/ref_response": -125.8144760131836,
"logps/rejected": -383.9209899902344,
"loss": 0.5314,
"rewards/accuracies": 0.7562500238418579,
"rewards/chosen": -0.33838868141174316,
"rewards/margins": 0.5101521015167236,
"rewards/rejected": -0.8485407829284668,
"step": 110
},
{
"epoch": 0.7535321821036107,
"grad_norm": 7.158165163184529,
"learning_rate": 8.628481651367875e-08,
"logits/chosen": 0.5716279745101929,
"logits/rejected": 0.8841035962104797,
"logps/chosen": -320.6312255859375,
"logps/pi_response": -131.89306640625,
"logps/ref_response": -120.58707427978516,
"logps/rejected": -369.5261535644531,
"loss": 0.5541,
"rewards/accuracies": 0.75,
"rewards/chosen": -0.4013099670410156,
"rewards/margins": 0.5214625597000122,
"rewards/rejected": -0.9227724075317383,
"step": 120
},
{
"epoch": 0.8163265306122449,
"grad_norm": 6.942114883986432,
"learning_rate": 4.904486005914027e-08,
"logits/chosen": 0.5996646881103516,
"logits/rejected": 0.8270283937454224,
"logps/chosen": -277.0096130371094,
"logps/pi_response": -136.15554809570312,
"logps/ref_response": -123.1449966430664,
"logps/rejected": -372.5554504394531,
"loss": 0.5271,
"rewards/accuracies": 0.737500011920929,
"rewards/chosen": -0.3302404284477234,
"rewards/margins": 0.5928428173065186,
"rewards/rejected": -0.9230831861495972,
"step": 130
},
{
"epoch": 0.8791208791208791,
"grad_norm": 7.301927243964412,
"learning_rate": 2.1464952759020856e-08,
"logits/chosen": 0.5675501823425293,
"logits/rejected": 0.8693594932556152,
"logps/chosen": -284.0587158203125,
"logps/pi_response": -131.0335235595703,
"logps/ref_response": -121.63087463378906,
"logps/rejected": -404.29278564453125,
"loss": 0.53,
"rewards/accuracies": 0.78125,
"rewards/chosen": -0.3401293456554413,
"rewards/margins": 0.748315155506134,
"rewards/rejected": -1.088444471359253,
"step": 140
},
{
"epoch": 0.9419152276295133,
"grad_norm": 8.46328336009499,
"learning_rate": 4.8708793644441086e-09,
"logits/chosen": 0.5060345530509949,
"logits/rejected": 0.8951950073242188,
"logps/chosen": -317.99163818359375,
"logps/pi_response": -145.62229919433594,
"logps/ref_response": -132.86119079589844,
"logps/rejected": -398.7874755859375,
"loss": 0.5193,
"rewards/accuracies": 0.762499988079071,
"rewards/chosen": -0.39088305830955505,
"rewards/margins": 0.5848211050033569,
"rewards/rejected": -0.9757040739059448,
"step": 150
},
{
"epoch": 0.9984301412872841,
"step": 159,
"total_flos": 0.0,
"train_loss": 0.5789602717513558,
"train_runtime": 4365.2801,
"train_samples_per_second": 4.668,
"train_steps_per_second": 0.036
}
],
"logging_steps": 10,
"max_steps": 159,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"total_flos": 0.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}