lectura's picture
Training in progress, epoch 12, checkpoint
e435291
raw
history blame contribute delete
No virus
12.2 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 12.923076923076923,
"eval_steps": 500,
"global_step": 105,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.12,
"learning_rate": 9.975000000000002e-06,
"loss": 12.7837,
"step": 1
},
{
"epoch": 0.25,
"learning_rate": 9.950000000000001e-06,
"loss": 12.4595,
"step": 2
},
{
"epoch": 0.37,
"learning_rate": 9.925e-06,
"loss": 11.9117,
"step": 3
},
{
"epoch": 0.49,
"learning_rate": 9.9e-06,
"loss": 11.662,
"step": 4
},
{
"epoch": 0.62,
"learning_rate": 9.875000000000001e-06,
"loss": 11.4278,
"step": 5
},
{
"epoch": 0.74,
"learning_rate": 9.85e-06,
"loss": 11.1398,
"step": 6
},
{
"epoch": 0.86,
"learning_rate": 9.825000000000002e-06,
"loss": 10.9135,
"step": 7
},
{
"epoch": 0.98,
"learning_rate": 9.800000000000001e-06,
"loss": 10.8274,
"step": 8
},
{
"epoch": 1.11,
"learning_rate": 9.775e-06,
"loss": 10.5068,
"step": 9
},
{
"epoch": 1.23,
"learning_rate": 9.75e-06,
"loss": 10.3551,
"step": 10
},
{
"epoch": 1.35,
"learning_rate": 9.725000000000001e-06,
"loss": 10.1776,
"step": 11
},
{
"epoch": 1.48,
"learning_rate": 9.7e-06,
"loss": 10.0708,
"step": 12
},
{
"epoch": 1.6,
"learning_rate": 9.675000000000001e-06,
"loss": 9.8703,
"step": 13
},
{
"epoch": 1.72,
"learning_rate": 9.65e-06,
"loss": 9.7534,
"step": 14
},
{
"epoch": 1.85,
"learning_rate": 9.625e-06,
"loss": 9.6094,
"step": 15
},
{
"epoch": 1.97,
"learning_rate": 9.600000000000001e-06,
"loss": 9.3989,
"step": 16
},
{
"epoch": 2.09,
"learning_rate": 9.575e-06,
"loss": 9.2753,
"step": 17
},
{
"epoch": 2.22,
"learning_rate": 9.55e-06,
"loss": 9.138,
"step": 18
},
{
"epoch": 2.34,
"learning_rate": 9.525000000000001e-06,
"loss": 9.0598,
"step": 19
},
{
"epoch": 2.46,
"learning_rate": 9.5e-06,
"loss": 8.8535,
"step": 20
},
{
"epoch": 2.58,
"learning_rate": 9.475000000000002e-06,
"loss": 8.7378,
"step": 21
},
{
"epoch": 2.71,
"learning_rate": 9.450000000000001e-06,
"loss": 8.6082,
"step": 22
},
{
"epoch": 2.83,
"learning_rate": 9.425e-06,
"loss": 8.4904,
"step": 23
},
{
"epoch": 2.95,
"learning_rate": 9.4e-06,
"loss": 8.2918,
"step": 24
},
{
"epoch": 3.08,
"learning_rate": 9.375000000000001e-06,
"loss": 8.2553,
"step": 25
},
{
"epoch": 3.2,
"learning_rate": 9.350000000000002e-06,
"loss": 8.0724,
"step": 26
},
{
"epoch": 3.32,
"learning_rate": 9.325000000000001e-06,
"loss": 8.0029,
"step": 27
},
{
"epoch": 3.45,
"learning_rate": 9.3e-06,
"loss": 7.8227,
"step": 28
},
{
"epoch": 3.57,
"learning_rate": 9.275e-06,
"loss": 7.7003,
"step": 29
},
{
"epoch": 3.69,
"learning_rate": 9.250000000000001e-06,
"loss": 7.5943,
"step": 30
},
{
"epoch": 3.82,
"learning_rate": 9.225e-06,
"loss": 7.5119,
"step": 31
},
{
"epoch": 3.94,
"learning_rate": 9.200000000000002e-06,
"loss": 7.3673,
"step": 32
},
{
"epoch": 4.06,
"learning_rate": 9.175000000000001e-06,
"loss": 7.3051,
"step": 33
},
{
"epoch": 4.18,
"learning_rate": 9.15e-06,
"loss": 7.224,
"step": 34
},
{
"epoch": 4.31,
"learning_rate": 9.125e-06,
"loss": 7.0517,
"step": 35
},
{
"epoch": 4.43,
"learning_rate": 9.100000000000001e-06,
"loss": 7.0021,
"step": 36
},
{
"epoch": 4.55,
"learning_rate": 9.075e-06,
"loss": 6.9407,
"step": 37
},
{
"epoch": 4.68,
"learning_rate": 9.050000000000001e-06,
"loss": 6.7762,
"step": 38
},
{
"epoch": 4.8,
"learning_rate": 9.025e-06,
"loss": 6.7675,
"step": 39
},
{
"epoch": 4.92,
"learning_rate": 9e-06,
"loss": 6.6573,
"step": 40
},
{
"epoch": 5.05,
"learning_rate": 8.975e-06,
"loss": 6.6093,
"step": 41
},
{
"epoch": 5.17,
"learning_rate": 8.95e-06,
"loss": 6.5268,
"step": 42
},
{
"epoch": 5.29,
"learning_rate": 8.925e-06,
"loss": 6.4456,
"step": 43
},
{
"epoch": 5.42,
"learning_rate": 8.900000000000001e-06,
"loss": 6.4134,
"step": 44
},
{
"epoch": 5.54,
"learning_rate": 8.875e-06,
"loss": 6.3414,
"step": 45
},
{
"epoch": 5.66,
"learning_rate": 8.85e-06,
"loss": 6.2924,
"step": 46
},
{
"epoch": 5.78,
"learning_rate": 8.825000000000001e-06,
"loss": 6.2176,
"step": 47
},
{
"epoch": 5.91,
"learning_rate": 8.8e-06,
"loss": 6.1821,
"step": 48
},
{
"epoch": 6.03,
"learning_rate": 8.775e-06,
"loss": 6.1344,
"step": 49
},
{
"epoch": 6.15,
"learning_rate": 8.750000000000001e-06,
"loss": 6.0681,
"step": 50
},
{
"epoch": 6.28,
"learning_rate": 8.725000000000002e-06,
"loss": 6.0286,
"step": 51
},
{
"epoch": 6.4,
"learning_rate": 8.700000000000001e-06,
"loss": 5.9494,
"step": 52
},
{
"epoch": 6.52,
"learning_rate": 8.675e-06,
"loss": 5.9192,
"step": 53
},
{
"epoch": 6.65,
"learning_rate": 8.65e-06,
"loss": 5.8799,
"step": 54
},
{
"epoch": 6.77,
"learning_rate": 8.625000000000001e-06,
"loss": 5.8367,
"step": 55
},
{
"epoch": 6.89,
"learning_rate": 8.6e-06,
"loss": 5.7919,
"step": 56
},
{
"epoch": 7.02,
"learning_rate": 8.575000000000002e-06,
"loss": 5.7692,
"step": 57
},
{
"epoch": 7.14,
"learning_rate": 8.550000000000001e-06,
"loss": 5.7151,
"step": 58
},
{
"epoch": 7.26,
"learning_rate": 8.525e-06,
"loss": 5.6839,
"step": 59
},
{
"epoch": 7.38,
"learning_rate": 8.5e-06,
"loss": 5.6504,
"step": 60
},
{
"epoch": 7.51,
"learning_rate": 8.475000000000001e-06,
"loss": 5.6222,
"step": 61
},
{
"epoch": 7.63,
"learning_rate": 8.45e-06,
"loss": 5.5806,
"step": 62
},
{
"epoch": 7.75,
"learning_rate": 8.425000000000001e-06,
"loss": 5.5354,
"step": 63
},
{
"epoch": 7.88,
"learning_rate": 8.400000000000001e-06,
"loss": 5.4984,
"step": 64
},
{
"epoch": 8.0,
"learning_rate": 8.375e-06,
"loss": 5.4788,
"step": 65
},
{
"epoch": 8.12,
"learning_rate": 8.35e-06,
"loss": 5.4293,
"step": 66
},
{
"epoch": 8.25,
"learning_rate": 8.325e-06,
"loss": 5.4256,
"step": 67
},
{
"epoch": 8.37,
"learning_rate": 8.3e-06,
"loss": 5.3499,
"step": 68
},
{
"epoch": 8.49,
"learning_rate": 8.275000000000001e-06,
"loss": 5.3758,
"step": 69
},
{
"epoch": 8.62,
"learning_rate": 8.25e-06,
"loss": 5.3384,
"step": 70
},
{
"epoch": 8.74,
"learning_rate": 8.225e-06,
"loss": 5.3189,
"step": 71
},
{
"epoch": 8.86,
"learning_rate": 8.2e-06,
"loss": 5.2841,
"step": 72
},
{
"epoch": 8.98,
"learning_rate": 8.175e-06,
"loss": 5.2387,
"step": 73
},
{
"epoch": 9.11,
"learning_rate": 8.15e-06,
"loss": 5.229,
"step": 74
},
{
"epoch": 9.23,
"learning_rate": 8.125000000000001e-06,
"loss": 5.1936,
"step": 75
},
{
"epoch": 9.35,
"learning_rate": 8.1e-06,
"loss": 5.1758,
"step": 76
},
{
"epoch": 9.48,
"learning_rate": 8.075000000000001e-06,
"loss": 5.1688,
"step": 77
},
{
"epoch": 9.6,
"learning_rate": 8.050000000000001e-06,
"loss": 5.1475,
"step": 78
},
{
"epoch": 9.72,
"learning_rate": 8.025e-06,
"loss": 5.1028,
"step": 79
},
{
"epoch": 9.85,
"learning_rate": 8.000000000000001e-06,
"loss": 5.0747,
"step": 80
},
{
"epoch": 9.97,
"learning_rate": 7.975e-06,
"loss": 5.0689,
"step": 81
},
{
"epoch": 10.09,
"learning_rate": 7.950000000000002e-06,
"loss": 5.045,
"step": 82
},
{
"epoch": 10.22,
"learning_rate": 7.925000000000001e-06,
"loss": 5.0361,
"step": 83
},
{
"epoch": 10.34,
"learning_rate": 7.9e-06,
"loss": 5.0273,
"step": 84
},
{
"epoch": 10.46,
"learning_rate": 7.875e-06,
"loss": 4.983,
"step": 85
},
{
"epoch": 10.58,
"learning_rate": 7.850000000000001e-06,
"loss": 4.9598,
"step": 86
},
{
"epoch": 10.71,
"learning_rate": 7.825e-06,
"loss": 4.9646,
"step": 87
},
{
"epoch": 10.83,
"learning_rate": 7.800000000000002e-06,
"loss": 4.9083,
"step": 88
},
{
"epoch": 10.95,
"learning_rate": 7.775000000000001e-06,
"loss": 4.906,
"step": 89
},
{
"epoch": 11.08,
"learning_rate": 7.75e-06,
"loss": 4.9062,
"step": 90
},
{
"epoch": 11.2,
"learning_rate": 7.725e-06,
"loss": 4.8742,
"step": 91
},
{
"epoch": 11.32,
"learning_rate": 7.7e-06,
"loss": 4.8425,
"step": 92
},
{
"epoch": 11.45,
"learning_rate": 7.675e-06,
"loss": 4.8332,
"step": 93
},
{
"epoch": 11.57,
"learning_rate": 7.650000000000001e-06,
"loss": 4.8399,
"step": 94
},
{
"epoch": 11.69,
"learning_rate": 7.625e-06,
"loss": 4.8392,
"step": 95
},
{
"epoch": 11.82,
"learning_rate": 7.600000000000001e-06,
"loss": 4.8131,
"step": 96
},
{
"epoch": 11.94,
"learning_rate": 7.575e-06,
"loss": 4.7643,
"step": 97
},
{
"epoch": 12.06,
"learning_rate": 7.5500000000000006e-06,
"loss": 4.7836,
"step": 98
},
{
"epoch": 12.18,
"learning_rate": 7.525e-06,
"loss": 4.7206,
"step": 99
},
{
"epoch": 12.31,
"learning_rate": 7.500000000000001e-06,
"loss": 4.7083,
"step": 100
},
{
"epoch": 12.43,
"learning_rate": 7.475000000000001e-06,
"loss": 4.7409,
"step": 101
},
{
"epoch": 12.55,
"learning_rate": 7.450000000000001e-06,
"loss": 4.7328,
"step": 102
},
{
"epoch": 12.68,
"learning_rate": 7.425000000000001e-06,
"loss": 4.7065,
"step": 103
},
{
"epoch": 12.8,
"learning_rate": 7.4e-06,
"loss": 4.7223,
"step": 104
},
{
"epoch": 12.92,
"learning_rate": 7.375000000000001e-06,
"loss": 4.6878,
"step": 105
}
],
"logging_steps": 1,
"max_steps": 400,
"num_train_epochs": 50,
"save_steps": 500,
"total_flos": 7016139155570688.0,
"trial_name": null,
"trial_params": null
}