zephyr-7b-gemma-v0.1 / trainer_state.json
lewtun's picture
lewtun HF staff
Model save
87a8563 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.971563981042654,
"eval_steps": 100,
"global_step": 104,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.02,
"grad_norm": 139.638709617328,
"learning_rate": 4.545454545454545e-08,
"logits/chosen": 111.16130065917969,
"logits/rejected": 86.8372802734375,
"logps/chosen": -326.8536071777344,
"logps/rejected": -329.15960693359375,
"loss": 0.6931,
"rewards/accuracies": 0.0,
"rewards/chosen": 0.0,
"rewards/margins": 0.0,
"rewards/rejected": 0.0,
"step": 1
},
{
"epoch": 0.19,
"grad_norm": 141.5345140695996,
"learning_rate": 4.545454545454545e-07,
"logits/chosen": 110.37065124511719,
"logits/rejected": 133.2639923095703,
"logps/chosen": -350.3541259765625,
"logps/rejected": -434.3558349609375,
"loss": 0.7191,
"rewards/accuracies": 0.4722222089767456,
"rewards/chosen": 0.13274627923965454,
"rewards/margins": 0.07573667168617249,
"rewards/rejected": 0.05700961872935295,
"step": 10
},
{
"epoch": 0.38,
"grad_norm": 123.71909837085582,
"learning_rate": 4.885348141000122e-07,
"logits/chosen": 117.74342346191406,
"logits/rejected": 128.52548217773438,
"logps/chosen": -333.21240234375,
"logps/rejected": -410.2923889160156,
"loss": 0.6097,
"rewards/accuracies": 0.7124999761581421,
"rewards/chosen": 0.11470325291156769,
"rewards/margins": 0.7479402422904968,
"rewards/rejected": -0.6332370042800903,
"step": 20
},
{
"epoch": 0.57,
"grad_norm": 111.89651526533274,
"learning_rate": 4.5025027361734613e-07,
"logits/chosen": 114.44095611572266,
"logits/rejected": 119.11683654785156,
"logps/chosen": -399.1412048339844,
"logps/rejected": -474.2645568847656,
"loss": 0.596,
"rewards/accuracies": 0.7250000238418579,
"rewards/chosen": -1.7276217937469482,
"rewards/margins": 1.0803521871566772,
"rewards/rejected": -2.807974100112915,
"step": 30
},
{
"epoch": 0.76,
"grad_norm": 102.67088507130228,
"learning_rate": 3.893311157806091e-07,
"logits/chosen": 116.33101654052734,
"logits/rejected": 111.0595703125,
"logps/chosen": -428.7275390625,
"logps/rejected": -464.0934143066406,
"loss": 0.5343,
"rewards/accuracies": 0.7250000238418579,
"rewards/chosen": -2.2770252227783203,
"rewards/margins": 0.9522085189819336,
"rewards/rejected": -3.229233503341675,
"step": 40
},
{
"epoch": 0.95,
"grad_norm": 130.9996197198566,
"learning_rate": 3.126631330646801e-07,
"logits/chosen": 123.2393569946289,
"logits/rejected": 124.50789642333984,
"logps/chosen": -438.548095703125,
"logps/rejected": -474.1234436035156,
"loss": 0.5138,
"rewards/accuracies": 0.762499988079071,
"rewards/chosen": -2.3258581161499023,
"rewards/margins": 1.3220884799957275,
"rewards/rejected": -3.647946834564209,
"step": 50
},
{
"epoch": 1.14,
"grad_norm": 56.950942870641875,
"learning_rate": 2.2891223348923882e-07,
"logits/chosen": 122.619384765625,
"logits/rejected": 126.1447525024414,
"logps/chosen": -414.3634338378906,
"logps/rejected": -468.19586181640625,
"loss": 0.2724,
"rewards/accuracies": 0.893750011920929,
"rewards/chosen": -2.3773388862609863,
"rewards/margins": 2.358515501022339,
"rewards/rejected": -4.735854148864746,
"step": 60
},
{
"epoch": 1.33,
"grad_norm": 52.820355390804025,
"learning_rate": 1.4754491880085317e-07,
"logits/chosen": 117.16709899902344,
"logits/rejected": 118.9737319946289,
"logps/chosen": -387.70526123046875,
"logps/rejected": -511.97503662109375,
"loss": 0.1936,
"rewards/accuracies": 0.9437500238418579,
"rewards/chosen": -2.4186935424804688,
"rewards/margins": 2.5914835929870605,
"rewards/rejected": -5.010177135467529,
"step": 70
},
{
"epoch": 1.52,
"grad_norm": 51.657826972971314,
"learning_rate": 7.775827023107834e-08,
"logits/chosen": 124.15473937988281,
"logits/rejected": 125.7086181640625,
"logps/chosen": -446.75421142578125,
"logps/rejected": -543.6109619140625,
"loss": 0.1779,
"rewards/accuracies": 0.981249988079071,
"rewards/chosen": -2.316882848739624,
"rewards/margins": 2.962496757507324,
"rewards/rejected": -5.279379844665527,
"step": 80
},
{
"epoch": 1.71,
"grad_norm": 86.34373603352554,
"learning_rate": 2.7440387297912122e-08,
"logits/chosen": 107.07579040527344,
"logits/rejected": 111.74522399902344,
"logps/chosen": -425.4237365722656,
"logps/rejected": -509.67718505859375,
"loss": 0.1765,
"rewards/accuracies": 0.9437500238418579,
"rewards/chosen": -2.749206066131592,
"rewards/margins": 3.0597147941589355,
"rewards/rejected": -5.8089213371276855,
"step": 90
},
{
"epoch": 1.9,
"grad_norm": 51.66215546933828,
"learning_rate": 2.27878296044029e-09,
"logits/chosen": 123.38490295410156,
"logits/rejected": 113.675537109375,
"logps/chosen": -439.7268981933594,
"logps/rejected": -550.8162841796875,
"loss": 0.1923,
"rewards/accuracies": 0.9624999761581421,
"rewards/chosen": -2.560769557952881,
"rewards/margins": 3.2135703563690186,
"rewards/rejected": -5.77433967590332,
"step": 100
},
{
"epoch": 1.9,
"eval_logits/chosen": 92.73604583740234,
"eval_logits/rejected": 86.38631439208984,
"eval_logps/chosen": -431.5707092285156,
"eval_logps/rejected": -459.1661682128906,
"eval_loss": 0.4735770523548126,
"eval_rewards/accuracies": 0.75,
"eval_rewards/chosen": -3.4575202465057373,
"eval_rewards/margins": 1.4980329275131226,
"eval_rewards/rejected": -4.9555535316467285,
"eval_runtime": 50.3064,
"eval_samples_per_second": 14.909,
"eval_steps_per_second": 0.477,
"step": 100
},
{
"epoch": 1.97,
"step": 104,
"total_flos": 0.0,
"train_loss": 0.38887147261546207,
"train_runtime": 1183.8142,
"train_samples_per_second": 11.404,
"train_steps_per_second": 0.088
}
],
"logging_steps": 10,
"max_steps": 104,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 500,
"total_flos": 0.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}