TianyiQ's picture
Upload ./trainer_state.json with huggingface_hub
6f78d0b verified
raw
history blame
6.2 kB
{
"best_metric": 2.2968251705169678,
"best_model_checkpoint": "./output/training_results/C018_Meta-Llama-3-8B_pretrain_20240726_033210/checkpoint-4230",
"epoch": 4.0,
"eval_steps": 470,
"global_step": 4696,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0008517887563884157,
"grad_norm": 0.0,
"learning_rate": 0.0,
"loss": 2.4637,
"step": 1
},
{
"epoch": 0.20017035775127767,
"grad_norm": 2.3409501850123138,
"learning_rate": 1.9546742209631728e-06,
"loss": 2.4082,
"step": 235
},
{
"epoch": 0.40034071550255534,
"grad_norm": 2.1722293582224492,
"learning_rate": 2.2631312554186003e-06,
"loss": 2.3529,
"step": 470
},
{
"epoch": 0.40034071550255534,
"eval_loss": 2.341965675354004,
"eval_runtime": 41.4873,
"eval_samples_per_second": 201.218,
"eval_steps_per_second": 1.591,
"step": 470
},
{
"epoch": 0.600511073253833,
"grad_norm": 2.185798119323745,
"learning_rate": 1.2303591421466819e-06,
"loss": 2.3196,
"step": 705
},
{
"epoch": 0.8006814310051107,
"grad_norm": 1.9581500824387317,
"learning_rate": 6.559971206312988e-07,
"loss": 2.3053,
"step": 940
},
{
"epoch": 0.8006814310051107,
"eval_loss": 2.307070732116699,
"eval_runtime": 41.2244,
"eval_samples_per_second": 202.501,
"eval_steps_per_second": 1.601,
"step": 940
},
{
"epoch": 1.0008517887563884,
"grad_norm": 2.0030962313242693,
"learning_rate": 3.4801579366796346e-07,
"loss": 2.2905,
"step": 1175
},
{
"epoch": 1.201022146507666,
"grad_norm": 2.1511768973070087,
"learning_rate": 1.8955345667471282e-07,
"loss": 2.2195,
"step": 1410
},
{
"epoch": 1.201022146507666,
"eval_loss": 2.3012468814849854,
"eval_runtime": 41.1845,
"eval_samples_per_second": 202.698,
"eval_steps_per_second": 1.603,
"step": 1410
},
{
"epoch": 1.4011925042589437,
"grad_norm": 2.038172613706977,
"learning_rate": 1.1177613622113936e-07,
"loss": 2.217,
"step": 1645
},
{
"epoch": 1.6013628620102214,
"grad_norm": 2.6056027737544087,
"learning_rate": 7.561933429867634e-08,
"loss": 2.2134,
"step": 1880
},
{
"epoch": 1.6013628620102214,
"eval_loss": 2.2990095615386963,
"eval_runtime": 41.2014,
"eval_samples_per_second": 202.615,
"eval_steps_per_second": 1.602,
"step": 1880
},
{
"epoch": 1.8015332197614993,
"grad_norm": 2.3428592690655385,
"learning_rate": 5.984119005303602e-08,
"loss": 2.2142,
"step": 2115
},
{
"epoch": 2.0017035775127767,
"grad_norm": 1.9738830418944124,
"learning_rate": 5.345074457083591e-08,
"loss": 2.2183,
"step": 2350
},
{
"epoch": 2.0017035775127767,
"eval_loss": 2.2979490756988525,
"eval_runtime": 41.2381,
"eval_samples_per_second": 202.434,
"eval_steps_per_second": 1.6,
"step": 2350
},
{
"epoch": 2.2018739352640546,
"grad_norm": 1.9766056487629942,
"learning_rate": 5.108344330433012e-08,
"loss": 2.1996,
"step": 2585
},
{
"epoch": 2.402044293015332,
"grad_norm": 1.9886986639549535,
"learning_rate": 5.0296763609045817e-08,
"loss": 2.2069,
"step": 2820
},
{
"epoch": 2.402044293015332,
"eval_loss": 2.2981810569763184,
"eval_runtime": 41.2271,
"eval_samples_per_second": 202.488,
"eval_steps_per_second": 1.601,
"step": 2820
},
{
"epoch": 2.60221465076661,
"grad_norm": 2.0358858583570556,
"learning_rate": 5.006836944156395e-08,
"loss": 2.2071,
"step": 3055
},
{
"epoch": 2.8023850085178874,
"grad_norm": 2.0667439556087315,
"learning_rate": 5.001265655634458e-08,
"loss": 2.205,
"step": 3290
},
{
"epoch": 2.8023850085178874,
"eval_loss": 2.2976646423339844,
"eval_runtime": 41.1148,
"eval_samples_per_second": 203.041,
"eval_steps_per_second": 1.605,
"step": 3290
},
{
"epoch": 3.0025553662691653,
"grad_norm": 2.048334597597571,
"learning_rate": 5.000170873605877e-08,
"loss": 2.2038,
"step": 3525
},
{
"epoch": 3.2027257240204428,
"grad_norm": 2.0830071658637626,
"learning_rate": 5.000014746665313e-08,
"loss": 2.1934,
"step": 3760
},
{
"epoch": 3.2027257240204428,
"eval_loss": 2.2974419593811035,
"eval_runtime": 41.1847,
"eval_samples_per_second": 202.697,
"eval_steps_per_second": 1.603,
"step": 3760
},
{
"epoch": 3.4028960817717206,
"grad_norm": 2.067015754816994,
"learning_rate": 5.000000637528681e-08,
"loss": 2.193,
"step": 3995
},
{
"epoch": 3.6030664395229985,
"grad_norm": 2.0937714004243673,
"learning_rate": 5.000000007544082e-08,
"loss": 2.2047,
"step": 4230
},
{
"epoch": 3.6030664395229985,
"eval_loss": 2.2968251705169678,
"eval_runtime": 41.1318,
"eval_samples_per_second": 202.957,
"eval_steps_per_second": 1.605,
"step": 4230
},
{
"epoch": 3.803236797274276,
"grad_norm": 2.076604473202926,
"learning_rate": 5.000000000003948e-08,
"loss": 2.2036,
"step": 4465
},
{
"epoch": 4.0,
"step": 4696,
"total_flos": 490890602741760.0,
"train_loss": 2.2402087043862937,
"train_runtime": 7138.091,
"train_samples_per_second": 42.101,
"train_steps_per_second": 0.658
}
],
"logging_steps": 235,
"max_steps": 4696,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 470,
"total_flos": 490890602741760.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}