colerobertson's picture
Training in progress, epoch 1
bac01b2 verified
raw
history blame
9.4 kB
{
"best_metric": 0.7326732673267327,
"best_model_checkpoint": "distilhubert-finetuned-not-a-word2/run-4/checkpoint-96",
"epoch": 5.0,
"eval_steps": 500,
"global_step": 240,
"is_hyper_param_search": true,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.1,
"grad_norm": 1.2733114957809448,
"learning_rate": 2.2702186710865246e-07,
"loss": 0.7025,
"step": 5
},
{
"epoch": 0.21,
"grad_norm": 1.243804931640625,
"learning_rate": 4.5404373421730493e-07,
"loss": 0.6974,
"step": 10
},
{
"epoch": 0.31,
"grad_norm": 1.7711552381515503,
"learning_rate": 6.810656013259573e-07,
"loss": 0.696,
"step": 15
},
{
"epoch": 0.42,
"grad_norm": 1.1453403234481812,
"learning_rate": 9.080874684346099e-07,
"loss": 0.6989,
"step": 20
},
{
"epoch": 0.52,
"grad_norm": 1.2729355096817017,
"learning_rate": 1.1351093355432624e-06,
"loss": 0.6968,
"step": 25
},
{
"epoch": 0.62,
"grad_norm": 1.1592165231704712,
"learning_rate": 1.3621312026519146e-06,
"loss": 0.6959,
"step": 30
},
{
"epoch": 0.73,
"grad_norm": 1.1798148155212402,
"learning_rate": 1.589153069760567e-06,
"loss": 0.6952,
"step": 35
},
{
"epoch": 0.83,
"grad_norm": 2.1216671466827393,
"learning_rate": 1.8161749368692197e-06,
"loss": 0.6886,
"step": 40
},
{
"epoch": 0.94,
"grad_norm": 1.3416370153427124,
"learning_rate": 2.043196803977872e-06,
"loss": 0.6864,
"step": 45
},
{
"epoch": 1.0,
"eval_f1": 0.72,
"eval_loss": 0.688262939453125,
"eval_runtime": 1.3468,
"eval_samples_per_second": 47.521,
"eval_steps_per_second": 5.94,
"step": 48
},
{
"epoch": 1.04,
"grad_norm": 2.1856281757354736,
"learning_rate": 2.169320063482679e-06,
"loss": 0.6917,
"step": 50
},
{
"epoch": 1.15,
"grad_norm": 1.4077153205871582,
"learning_rate": 2.1440954115817176e-06,
"loss": 0.6884,
"step": 55
},
{
"epoch": 1.25,
"grad_norm": 2.1792664527893066,
"learning_rate": 2.1188707596807562e-06,
"loss": 0.6668,
"step": 60
},
{
"epoch": 1.35,
"grad_norm": 1.0386197566986084,
"learning_rate": 2.093646107779795e-06,
"loss": 0.6694,
"step": 65
},
{
"epoch": 1.46,
"grad_norm": 2.0565919876098633,
"learning_rate": 2.0684214558788335e-06,
"loss": 0.6561,
"step": 70
},
{
"epoch": 1.56,
"grad_norm": 1.2978509664535522,
"learning_rate": 2.043196803977872e-06,
"loss": 0.6789,
"step": 75
},
{
"epoch": 1.67,
"grad_norm": 2.058328628540039,
"learning_rate": 2.0179721520769108e-06,
"loss": 0.6633,
"step": 80
},
{
"epoch": 1.77,
"grad_norm": 0.6023226976394653,
"learning_rate": 1.9927475001759494e-06,
"loss": 0.6655,
"step": 85
},
{
"epoch": 1.88,
"grad_norm": 0.5510762929916382,
"learning_rate": 1.967522848274988e-06,
"loss": 0.6622,
"step": 90
},
{
"epoch": 1.98,
"grad_norm": 1.098602533340454,
"learning_rate": 1.9422981963740267e-06,
"loss": 0.6633,
"step": 95
},
{
"epoch": 2.0,
"eval_f1": 0.7326732673267327,
"eval_loss": 0.6816024780273438,
"eval_runtime": 1.3765,
"eval_samples_per_second": 46.493,
"eval_steps_per_second": 5.812,
"step": 96
},
{
"epoch": 2.08,
"grad_norm": 0.9589098691940308,
"learning_rate": 1.9170735444730654e-06,
"loss": 0.659,
"step": 100
},
{
"epoch": 2.19,
"grad_norm": 1.070695161819458,
"learning_rate": 1.8918488925721038e-06,
"loss": 0.6313,
"step": 105
},
{
"epoch": 2.29,
"grad_norm": 0.9913639426231384,
"learning_rate": 1.8666242406711424e-06,
"loss": 0.6652,
"step": 110
},
{
"epoch": 2.4,
"grad_norm": 1.0632878541946411,
"learning_rate": 1.841399588770181e-06,
"loss": 0.673,
"step": 115
},
{
"epoch": 2.5,
"grad_norm": 2.1036579608917236,
"learning_rate": 1.8161749368692197e-06,
"loss": 0.6451,
"step": 120
},
{
"epoch": 2.6,
"grad_norm": 1.08384108543396,
"learning_rate": 1.7909502849682583e-06,
"loss": 0.6322,
"step": 125
},
{
"epoch": 2.71,
"grad_norm": 0.9407000541687012,
"learning_rate": 1.765725633067297e-06,
"loss": 0.6755,
"step": 130
},
{
"epoch": 2.81,
"grad_norm": 0.9016568660736084,
"learning_rate": 1.7405009811663356e-06,
"loss": 0.5985,
"step": 135
},
{
"epoch": 2.92,
"grad_norm": 1.1134448051452637,
"learning_rate": 1.7152763292653743e-06,
"loss": 0.603,
"step": 140
},
{
"epoch": 3.0,
"eval_f1": 0.7326732673267327,
"eval_loss": 0.6800689697265625,
"eval_runtime": 1.3861,
"eval_samples_per_second": 46.173,
"eval_steps_per_second": 5.772,
"step": 144
},
{
"epoch": 3.02,
"grad_norm": 0.7627719640731812,
"learning_rate": 1.6900516773644127e-06,
"loss": 0.6557,
"step": 145
},
{
"epoch": 3.12,
"grad_norm": 0.9291415214538574,
"learning_rate": 1.6648270254634511e-06,
"loss": 0.6219,
"step": 150
},
{
"epoch": 3.23,
"grad_norm": 0.9248765707015991,
"learning_rate": 1.6396023735624898e-06,
"loss": 0.6325,
"step": 155
},
{
"epoch": 3.33,
"grad_norm": 0.9842573404312134,
"learning_rate": 1.6143777216615284e-06,
"loss": 0.6521,
"step": 160
},
{
"epoch": 3.44,
"grad_norm": 0.8689214587211609,
"learning_rate": 1.589153069760567e-06,
"loss": 0.5929,
"step": 165
},
{
"epoch": 3.54,
"grad_norm": 1.0012000799179077,
"learning_rate": 1.5639284178596057e-06,
"loss": 0.584,
"step": 170
},
{
"epoch": 3.65,
"grad_norm": 0.7438368797302246,
"learning_rate": 1.5387037659586443e-06,
"loss": 0.6813,
"step": 175
},
{
"epoch": 3.75,
"grad_norm": 1.8603870868682861,
"learning_rate": 1.513479114057683e-06,
"loss": 0.6099,
"step": 180
},
{
"epoch": 3.85,
"grad_norm": 0.9918416738510132,
"learning_rate": 1.4882544621567216e-06,
"loss": 0.6192,
"step": 185
},
{
"epoch": 3.96,
"grad_norm": 1.9146322011947632,
"learning_rate": 1.4630298102557603e-06,
"loss": 0.6472,
"step": 190
},
{
"epoch": 4.0,
"eval_f1": 0.7326732673267327,
"eval_loss": 0.6818161010742188,
"eval_runtime": 1.3841,
"eval_samples_per_second": 46.239,
"eval_steps_per_second": 5.78,
"step": 192
},
{
"epoch": 4.06,
"grad_norm": 0.9502781629562378,
"learning_rate": 1.437805158354799e-06,
"loss": 0.6447,
"step": 195
},
{
"epoch": 4.17,
"grad_norm": 0.8570067286491394,
"learning_rate": 1.4125805064538375e-06,
"loss": 0.5306,
"step": 200
},
{
"epoch": 4.27,
"grad_norm": 0.8097484111785889,
"learning_rate": 1.3873558545528762e-06,
"loss": 0.6202,
"step": 205
},
{
"epoch": 4.38,
"grad_norm": 2.0106472969055176,
"learning_rate": 1.3621312026519146e-06,
"loss": 0.6705,
"step": 210
},
{
"epoch": 4.48,
"grad_norm": 1.090775489807129,
"learning_rate": 1.3369065507509533e-06,
"loss": 0.6297,
"step": 215
},
{
"epoch": 4.58,
"grad_norm": 0.8988145589828491,
"learning_rate": 1.311681898849992e-06,
"loss": 0.5896,
"step": 220
},
{
"epoch": 4.69,
"grad_norm": 0.9149978756904602,
"learning_rate": 1.2864572469490305e-06,
"loss": 0.6156,
"step": 225
},
{
"epoch": 4.79,
"grad_norm": 1.9398412704467773,
"learning_rate": 1.2612325950480692e-06,
"loss": 0.6305,
"step": 230
},
{
"epoch": 4.9,
"grad_norm": 0.9217966794967651,
"learning_rate": 1.2360079431471078e-06,
"loss": 0.5943,
"step": 235
},
{
"epoch": 5.0,
"grad_norm": 0.9083653688430786,
"learning_rate": 1.2107832912461465e-06,
"loss": 0.6386,
"step": 240
},
{
"epoch": 5.0,
"eval_f1": 0.7326732673267327,
"eval_loss": 0.6846389770507812,
"eval_runtime": 1.4094,
"eval_samples_per_second": 45.409,
"eval_steps_per_second": 5.676,
"step": 240
}
],
"logging_steps": 5,
"max_steps": 480,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 500,
"total_flos": 3654362860415712.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": {
"learning_rate": 2.1794099242430636e-06,
"per_device_train_batch_size": 4
}
}