PeptideESM2_650M / trainer_state.json
jjoecclark's picture
Upload 9 files
8a1c5bd verified
{
"best_metric": 2.4635186195373535,
"best_model_checkpoint": "./drive/MyDrive/peptide_esm/checkpoint-10488",
"epoch": 2.0,
"eval_steps": 500,
"global_step": 10488,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.1,
"grad_norm": 0.7004187107086182,
"learning_rate": 7.5e-07,
"loss": 2.5303,
"step": 500
},
{
"epoch": 0.19,
"grad_norm": 0.7539237141609192,
"learning_rate": 1.5e-06,
"loss": 2.5119,
"step": 1000
},
{
"epoch": 0.29,
"grad_norm": 0.9745526313781738,
"learning_rate": 2.25e-06,
"loss": 2.4899,
"step": 1500
},
{
"epoch": 0.38,
"grad_norm": 0.6661661863327026,
"learning_rate": 3e-06,
"loss": 2.4903,
"step": 2000
},
{
"epoch": 0.48,
"grad_norm": 0.7792529463768005,
"learning_rate": 2.9380677126341868e-06,
"loss": 2.4831,
"step": 2500
},
{
"epoch": 0.57,
"grad_norm": 2.211763381958008,
"learning_rate": 2.8762592898431047e-06,
"loss": 2.4785,
"step": 3000
},
{
"epoch": 0.67,
"grad_norm": 0.7369110584259033,
"learning_rate": 2.814450867052023e-06,
"loss": 2.4763,
"step": 3500
},
{
"epoch": 0.76,
"grad_norm": 0.8026860952377319,
"learning_rate": 2.75251857968621e-06,
"loss": 2.4702,
"step": 4000
},
{
"epoch": 0.86,
"grad_norm": 0.5967105627059937,
"learning_rate": 2.6905862923203965e-06,
"loss": 2.4703,
"step": 4500
},
{
"epoch": 0.95,
"grad_norm": 0.68107670545578,
"learning_rate": 2.6286540049545828e-06,
"loss": 2.4699,
"step": 5000
},
{
"epoch": 1.0,
"eval_loss": 2.46760630607605,
"eval_runtime": 181.2222,
"eval_samples_per_second": 823.095,
"eval_steps_per_second": 6.434,
"step": 5244
},
{
"epoch": 1.05,
"grad_norm": 1.064468502998352,
"learning_rate": 2.56672171758877e-06,
"loss": 2.4671,
"step": 5500
},
{
"epoch": 1.14,
"grad_norm": 1.5440330505371094,
"learning_rate": 2.5047894302229566e-06,
"loss": 2.4663,
"step": 6000
},
{
"epoch": 1.24,
"grad_norm": 0.913020133972168,
"learning_rate": 2.442857142857143e-06,
"loss": 2.4654,
"step": 6500
},
{
"epoch": 1.33,
"grad_norm": 0.8316154479980469,
"learning_rate": 2.3810487200660612e-06,
"loss": 2.4694,
"step": 7000
},
{
"epoch": 1.43,
"grad_norm": 0.6185809969902039,
"learning_rate": 2.3191164327002475e-06,
"loss": 2.4668,
"step": 7500
},
{
"epoch": 1.53,
"grad_norm": 1.1215540170669556,
"learning_rate": 2.2571841453344346e-06,
"loss": 2.4668,
"step": 8000
},
{
"epoch": 1.62,
"grad_norm": 0.6679463982582092,
"learning_rate": 2.195251857968621e-06,
"loss": 2.4681,
"step": 8500
},
{
"epoch": 1.72,
"grad_norm": 1.1260002851486206,
"learning_rate": 2.1333195706028075e-06,
"loss": 2.4639,
"step": 9000
},
{
"epoch": 1.81,
"grad_norm": 0.6827268004417419,
"learning_rate": 2.0713872832369942e-06,
"loss": 2.4631,
"step": 9500
},
{
"epoch": 1.91,
"grad_norm": 0.7196531891822815,
"learning_rate": 2.0095788604459126e-06,
"loss": 2.4644,
"step": 10000
},
{
"epoch": 2.0,
"eval_loss": 2.4635186195373535,
"eval_runtime": 180.8599,
"eval_samples_per_second": 824.744,
"eval_steps_per_second": 6.447,
"step": 10488
}
],
"logging_steps": 500,
"max_steps": 26220,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"total_flos": 4.1949317916310464e+17,
"train_batch_size": 256,
"trial_name": null,
"trial_params": null
}