SELM-Llama-3-8B-Instruct-iter-2 / trainer_state.json
ZhangShenao's picture
Model save
fbfa663 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9984301412872841,
"eval_steps": 500,
"global_step": 159,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.006279434850863423,
"grad_norm": 17.524188433318294,
"learning_rate": 1.875e-08,
"logits/chosen": 0.050171270966529846,
"logits/rejected": 0.7975481748580933,
"logps/chosen": -213.99826049804688,
"logps/pi_response": -122.33531951904297,
"logps/ref_response": -122.33531951904297,
"logps/rejected": -327.0420227050781,
"loss": 0.6931,
"rewards/accuracies": 0.0,
"rewards/chosen": 0.0,
"rewards/margins": 0.0,
"rewards/rejected": 0.0,
"step": 1
},
{
"epoch": 0.06279434850863422,
"grad_norm": 23.45193685304271,
"learning_rate": 1.875e-07,
"logits/chosen": 0.6251156330108643,
"logits/rejected": 0.9096591472625732,
"logps/chosen": -268.11822509765625,
"logps/pi_response": -114.67009735107422,
"logps/ref_response": -114.69570922851562,
"logps/rejected": -390.1163330078125,
"loss": 0.6924,
"rewards/accuracies": 0.5416666865348816,
"rewards/chosen": -0.000606474990490824,
"rewards/margins": 0.0019376208074390888,
"rewards/rejected": -0.0025440959725528955,
"step": 10
},
{
"epoch": 0.12558869701726844,
"grad_norm": 19.551172670207706,
"learning_rate": 2.9942119880575817e-07,
"logits/chosen": 0.4843064248561859,
"logits/rejected": 0.8785603642463684,
"logps/chosen": -269.538330078125,
"logps/pi_response": -121.35018157958984,
"logps/ref_response": -121.34931945800781,
"logps/rejected": -417.247314453125,
"loss": 0.6747,
"rewards/accuracies": 0.75,
"rewards/chosen": -0.0313839390873909,
"rewards/margins": 0.04394400864839554,
"rewards/rejected": -0.07532794773578644,
"step": 20
},
{
"epoch": 0.18838304552590268,
"grad_norm": 13.701658129429845,
"learning_rate": 2.929608750821129e-07,
"logits/chosen": 0.5343824625015259,
"logits/rejected": 0.9908145666122437,
"logps/chosen": -310.638671875,
"logps/pi_response": -118.42143249511719,
"logps/ref_response": -117.36665344238281,
"logps/rejected": -428.20166015625,
"loss": 0.6175,
"rewards/accuracies": 0.7250000238418579,
"rewards/chosen": -0.18561159074306488,
"rewards/margins": 0.17133654654026031,
"rewards/rejected": -0.3569481372833252,
"step": 30
},
{
"epoch": 0.25117739403453687,
"grad_norm": 14.693007375551543,
"learning_rate": 2.7962832564252725e-07,
"logits/chosen": 0.5080328583717346,
"logits/rejected": 0.9311200976371765,
"logps/chosen": -303.4456481933594,
"logps/pi_response": -121.04774475097656,
"logps/ref_response": -117.7525634765625,
"logps/rejected": -484.09552001953125,
"loss": 0.5785,
"rewards/accuracies": 0.78125,
"rewards/chosen": -0.3230215311050415,
"rewards/margins": 0.4571772515773773,
"rewards/rejected": -0.7801988124847412,
"step": 40
},
{
"epoch": 0.3139717425431711,
"grad_norm": 18.087271424489508,
"learning_rate": 2.6006445513357056e-07,
"logits/chosen": 0.6870445013046265,
"logits/rejected": 0.9769166111946106,
"logps/chosen": -342.7091369628906,
"logps/pi_response": -124.7760009765625,
"logps/ref_response": -115.9255599975586,
"logps/rejected": -495.4518127441406,
"loss": 0.5714,
"rewards/accuracies": 0.643750011920929,
"rewards/chosen": -0.712685227394104,
"rewards/margins": 0.5671727657318115,
"rewards/rejected": -1.2798579931259155,
"step": 50
},
{
"epoch": 0.37676609105180536,
"grad_norm": 11.166951031052399,
"learning_rate": 2.3520971200967334e-07,
"logits/chosen": 0.5992540121078491,
"logits/rejected": 1.020711898803711,
"logps/chosen": -348.0439758300781,
"logps/pi_response": -127.2256088256836,
"logps/ref_response": -119.3854751586914,
"logps/rejected": -496.8860778808594,
"loss": 0.5649,
"rewards/accuracies": 0.71875,
"rewards/chosen": -0.5891093015670776,
"rewards/margins": 0.5816227793693542,
"rewards/rejected": -1.1707321405410767,
"step": 60
},
{
"epoch": 0.43956043956043955,
"grad_norm": 11.470597805963651,
"learning_rate": 2.0625888054143427e-07,
"logits/chosen": 0.6002563238143921,
"logits/rejected": 0.9772939682006836,
"logps/chosen": -272.6213073730469,
"logps/pi_response": -127.3548583984375,
"logps/ref_response": -120.81624603271484,
"logps/rejected": -524.6642456054688,
"loss": 0.5372,
"rewards/accuracies": 0.7749999761581421,
"rewards/chosen": -0.36886435747146606,
"rewards/margins": 0.7560935020446777,
"rewards/rejected": -1.124957799911499,
"step": 70
},
{
"epoch": 0.5023547880690737,
"grad_norm": 11.00220562972207,
"learning_rate": 1.7460364672965327e-07,
"logits/chosen": 0.6907710433006287,
"logits/rejected": 1.0923134088516235,
"logps/chosen": -269.1040954589844,
"logps/pi_response": -112.93983459472656,
"logps/ref_response": -106.69105529785156,
"logps/rejected": -492.8182678222656,
"loss": 0.5271,
"rewards/accuracies": 0.7749999761581421,
"rewards/chosen": -0.4200579524040222,
"rewards/margins": 0.6648958921432495,
"rewards/rejected": -1.084953784942627,
"step": 80
},
{
"epoch": 0.565149136577708,
"grad_norm": 11.111343223576297,
"learning_rate": 1.4176569902035086e-07,
"logits/chosen": 0.6897394061088562,
"logits/rejected": 1.039529800415039,
"logps/chosen": -341.63275146484375,
"logps/pi_response": -120.40811920166016,
"logps/ref_response": -111.32686614990234,
"logps/rejected": -511.60565185546875,
"loss": 0.5055,
"rewards/accuracies": 0.699999988079071,
"rewards/chosen": -0.5421563982963562,
"rewards/margins": 0.7095355987548828,
"rewards/rejected": -1.2516921758651733,
"step": 90
},
{
"epoch": 0.6279434850863422,
"grad_norm": 12.249034688194289,
"learning_rate": 1.0932357971453743e-07,
"logits/chosen": 0.7736852765083313,
"logits/rejected": 1.0484180450439453,
"logps/chosen": -290.811279296875,
"logps/pi_response": -116.5997543334961,
"logps/ref_response": -107.03324127197266,
"logps/rejected": -533.04931640625,
"loss": 0.539,
"rewards/accuracies": 0.7562500238418579,
"rewards/chosen": -0.49749264121055603,
"rewards/margins": 0.8334757089614868,
"rewards/rejected": -1.3309683799743652,
"step": 100
},
{
"epoch": 0.6907378335949764,
"grad_norm": 10.864663994027794,
"learning_rate": 7.883680337481599e-08,
"logits/chosen": 0.7670079469680786,
"logits/rejected": 0.9956636428833008,
"logps/chosen": -305.14593505859375,
"logps/pi_response": -127.0176773071289,
"logps/ref_response": -116.74520111083984,
"logps/rejected": -521.2225341796875,
"loss": 0.5335,
"rewards/accuracies": 0.75,
"rewards/chosen": -0.5058807134628296,
"rewards/margins": 0.7484409809112549,
"rewards/rejected": -1.2543216943740845,
"step": 110
},
{
"epoch": 0.7535321821036107,
"grad_norm": 11.061608119367962,
"learning_rate": 5.177088990820725e-08,
"logits/chosen": 0.5085455775260925,
"logits/rejected": 0.8686326146125793,
"logps/chosen": -308.0654602050781,
"logps/pi_response": -132.87335205078125,
"logps/ref_response": -123.3708267211914,
"logps/rejected": -556.9476928710938,
"loss": 0.5259,
"rewards/accuracies": 0.7562500238418579,
"rewards/chosen": -0.4727388918399811,
"rewards/margins": 0.8706821203231812,
"rewards/rejected": -1.3434208631515503,
"step": 120
},
{
"epoch": 0.8163265306122449,
"grad_norm": 13.028590054698721,
"learning_rate": 2.942691603548416e-08,
"logits/chosen": 0.5405811071395874,
"logits/rejected": 1.0355885028839111,
"logps/chosen": -320.9105224609375,
"logps/pi_response": -135.04757690429688,
"logps/ref_response": -127.4443588256836,
"logps/rejected": -553.2946166992188,
"loss": 0.5056,
"rewards/accuracies": 0.7875000238418579,
"rewards/chosen": -0.4617777466773987,
"rewards/margins": 0.8675910234451294,
"rewards/rejected": -1.3293688297271729,
"step": 130
},
{
"epoch": 0.8791208791208791,
"grad_norm": 11.32561508697925,
"learning_rate": 1.2878971655412513e-08,
"logits/chosen": 0.5967448353767395,
"logits/rejected": 0.9229636192321777,
"logps/chosen": -296.9305725097656,
"logps/pi_response": -139.7073974609375,
"logps/ref_response": -129.02000427246094,
"logps/rejected": -566.9810791015625,
"loss": 0.5117,
"rewards/accuracies": 0.800000011920929,
"rewards/chosen": -0.4712587893009186,
"rewards/margins": 0.9010453224182129,
"rewards/rejected": -1.3723042011260986,
"step": 140
},
{
"epoch": 0.9419152276295133,
"grad_norm": 11.652819230057418,
"learning_rate": 2.922527618666465e-09,
"logits/chosen": 0.5724425315856934,
"logits/rejected": 0.9267145395278931,
"logps/chosen": -312.5321350097656,
"logps/pi_response": -124.59476470947266,
"logps/ref_response": -114.87628173828125,
"logps/rejected": -503.21697998046875,
"loss": 0.5304,
"rewards/accuracies": 0.7437499761581421,
"rewards/chosen": -0.5373696088790894,
"rewards/margins": 0.6669692993164062,
"rewards/rejected": -1.204338788986206,
"step": 150
},
{
"epoch": 0.9984301412872841,
"step": 159,
"total_flos": 0.0,
"train_loss": 0.557155561147246,
"train_runtime": 4462.021,
"train_samples_per_second": 4.567,
"train_steps_per_second": 0.036
}
],
"logging_steps": 10,
"max_steps": 159,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"total_flos": 0.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}