lewtun's picture
lewtun HF staff
Model save
58e897c verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.971563981042654,
"eval_steps": 100,
"global_step": 104,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.02,
"grad_norm": 133.0057671141645,
"learning_rate": 4.545454545454545e-08,
"logits/chosen": 123.11854553222656,
"logits/rejected": 97.00198364257812,
"logps/chosen": -425.18585205078125,
"logps/rejected": -424.1869201660156,
"loss": 0.6931,
"rewards/accuracies": 0.0,
"rewards/chosen": 0.0,
"rewards/margins": 0.0,
"rewards/rejected": 0.0,
"step": 1
},
{
"epoch": 0.19,
"grad_norm": 182.25849293095231,
"learning_rate": 4.545454545454545e-07,
"logits/chosen": 117.33037567138672,
"logits/rejected": 136.23989868164062,
"logps/chosen": -442.517333984375,
"logps/rejected": -524.100830078125,
"loss": 0.7193,
"rewards/accuracies": 0.4791666567325592,
"rewards/chosen": 0.04355551674962044,
"rewards/margins": 0.04285159707069397,
"rewards/rejected": 0.0007039175252430141,
"step": 10
},
{
"epoch": 0.38,
"grad_norm": 98.47580983421109,
"learning_rate": 4.885348141000122e-07,
"logits/chosen": 124.34559631347656,
"logits/rejected": 131.87884521484375,
"logps/chosen": -425.8033142089844,
"logps/rejected": -496.58966064453125,
"loss": 0.5981,
"rewards/accuracies": 0.625,
"rewards/chosen": 0.02306445501744747,
"rewards/margins": 0.4770180284976959,
"rewards/rejected": -0.4539535939693451,
"step": 20
},
{
"epoch": 0.57,
"grad_norm": 114.087847259367,
"learning_rate": 4.5025027361734613e-07,
"logits/chosen": 121.997314453125,
"logits/rejected": 125.30668640136719,
"logps/chosen": -473.92950439453125,
"logps/rejected": -546.4301147460938,
"loss": 0.5828,
"rewards/accuracies": 0.731249988079071,
"rewards/chosen": -0.9404676556587219,
"rewards/margins": 1.0203006267547607,
"rewards/rejected": -1.9607683420181274,
"step": 30
},
{
"epoch": 0.76,
"grad_norm": 102.23106994021863,
"learning_rate": 3.893311157806091e-07,
"logits/chosen": 123.2125473022461,
"logits/rejected": 118.11393737792969,
"logps/chosen": -498.98480224609375,
"logps/rejected": -529.0015869140625,
"loss": 0.5432,
"rewards/accuracies": 0.731249988079071,
"rewards/chosen": -1.29017174243927,
"rewards/margins": 0.921674907207489,
"rewards/rejected": -2.2118465900421143,
"step": 40
},
{
"epoch": 0.95,
"grad_norm": 116.89709640403532,
"learning_rate": 3.126631330646801e-07,
"logits/chosen": 131.3624267578125,
"logits/rejected": 131.5041046142578,
"logps/chosen": -517.221923828125,
"logps/rejected": -547.076904296875,
"loss": 0.5007,
"rewards/accuracies": 0.7250000238418579,
"rewards/chosen": -1.7199589014053345,
"rewards/margins": 1.0714657306671143,
"rewards/rejected": -2.791424512863159,
"step": 50
},
{
"epoch": 1.14,
"grad_norm": 53.67728401735822,
"learning_rate": 2.2891223348923882e-07,
"logits/chosen": 130.25340270996094,
"logits/rejected": 133.3688507080078,
"logps/chosen": -482.68798828125,
"logps/rejected": -531.1278076171875,
"loss": 0.2897,
"rewards/accuracies": 0.8687499761581421,
"rewards/chosen": -1.2961251735687256,
"rewards/margins": 2.0452933311462402,
"rewards/rejected": -3.341418504714966,
"step": 60
},
{
"epoch": 1.33,
"grad_norm": 46.49732419269788,
"learning_rate": 1.4754491880085317e-07,
"logits/chosen": 125.44151306152344,
"logits/rejected": 126.0068588256836,
"logps/chosen": -451.8477478027344,
"logps/rejected": -573.8129272460938,
"loss": 0.1884,
"rewards/accuracies": 0.918749988079071,
"rewards/chosen": -1.1466975212097168,
"rewards/margins": 2.4370884895324707,
"rewards/rejected": -3.5837860107421875,
"step": 70
},
{
"epoch": 1.52,
"grad_norm": 48.86628168194261,
"learning_rate": 7.775827023107834e-08,
"logits/chosen": 129.2606201171875,
"logits/rejected": 130.1272735595703,
"logps/chosen": -510.95819091796875,
"logps/rejected": -606.6770629882812,
"loss": 0.1877,
"rewards/accuracies": 0.956250011920929,
"rewards/chosen": -1.0929005146026611,
"rewards/margins": 2.9041507244110107,
"rewards/rejected": -3.9970507621765137,
"step": 80
},
{
"epoch": 1.71,
"grad_norm": 51.734548215896965,
"learning_rate": 2.7440387297912122e-08,
"logits/chosen": 116.0444564819336,
"logits/rejected": 118.35429382324219,
"logps/chosen": -488.314697265625,
"logps/rejected": -571.8872680664062,
"loss": 0.1774,
"rewards/accuracies": 0.96875,
"rewards/chosen": -1.3073575496673584,
"rewards/margins": 3.0684092044830322,
"rewards/rejected": -4.375766277313232,
"step": 90
},
{
"epoch": 1.9,
"grad_norm": 58.09774424636419,
"learning_rate": 2.27878296044029e-09,
"logits/chosen": 129.8248748779297,
"logits/rejected": 120.5040512084961,
"logps/chosen": -505.01513671875,
"logps/rejected": -610.0936279296875,
"loss": 0.193,
"rewards/accuracies": 0.9750000238418579,
"rewards/chosen": -1.331251859664917,
"rewards/margins": 3.034428596496582,
"rewards/rejected": -4.36568021774292,
"step": 100
},
{
"epoch": 1.9,
"eval_logits/chosen": 99.29187774658203,
"eval_logits/rejected": 93.2354736328125,
"eval_logps/chosen": -493.443359375,
"eval_logps/rejected": -520.8292236328125,
"eval_loss": 0.4772518575191498,
"eval_rewards/accuracies": 0.7291666865348816,
"eval_rewards/chosen": -2.0019426345825195,
"eval_rewards/margins": 1.4620394706726074,
"eval_rewards/rejected": -3.463982343673706,
"eval_runtime": 49.3044,
"eval_samples_per_second": 15.212,
"eval_steps_per_second": 0.487,
"step": 100
},
{
"epoch": 1.97,
"step": 104,
"total_flos": 0.0,
"train_loss": 0.3880494168171516,
"train_runtime": 1239.5887,
"train_samples_per_second": 10.891,
"train_steps_per_second": 0.084
}
],
"logging_steps": 10,
"max_steps": 104,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 100,
"total_flos": 0.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}