output_model / checkpoint-500 /trainer_state.json
tktung's picture
Training in progress, step 500, checkpoint
e99e81a verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.3480985118788617,
"eval_steps": 500,
"global_step": 500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0006961970237577234,
"grad_norm": 74.93397158603199,
"learning_rate": 6.944444444444445e-07,
"loss": 2.6763,
"step": 1
},
{
"epoch": 0.006961970237577234,
"grad_norm": 42.51151469703164,
"learning_rate": 6.944444444444445e-06,
"loss": 2.5014,
"step": 10
},
{
"epoch": 0.013923940475154469,
"grad_norm": 3.9060174491324595,
"learning_rate": 1.388888888888889e-05,
"loss": 2.1436,
"step": 20
},
{
"epoch": 0.020885910712731704,
"grad_norm": 2.4653356578163312,
"learning_rate": 2.0833333333333336e-05,
"loss": 2.1299,
"step": 30
},
{
"epoch": 0.027847880950308938,
"grad_norm": 1.9758176342444926,
"learning_rate": 2.777777777777778e-05,
"loss": 2.0629,
"step": 40
},
{
"epoch": 0.034809851187886175,
"grad_norm": 2.24304363639927,
"learning_rate": 3.472222222222222e-05,
"loss": 2.0022,
"step": 50
},
{
"epoch": 0.04177182142546341,
"grad_norm": 2.137058600643756,
"learning_rate": 4.166666666666667e-05,
"loss": 1.9324,
"step": 60
},
{
"epoch": 0.04873379166304064,
"grad_norm": 1.8905774575553251,
"learning_rate": 4.8611111111111115e-05,
"loss": 1.8735,
"step": 70
},
{
"epoch": 0.055695761900617875,
"grad_norm": 2.277356627020573,
"learning_rate": 5.555555555555556e-05,
"loss": 1.8459,
"step": 80
},
{
"epoch": 0.06265773213819512,
"grad_norm": 2.1850479551940736,
"learning_rate": 6.25e-05,
"loss": 1.8657,
"step": 90
},
{
"epoch": 0.06961970237577235,
"grad_norm": 2.2821298235140066,
"learning_rate": 6.944444444444444e-05,
"loss": 1.8105,
"step": 100
},
{
"epoch": 0.07658167261334958,
"grad_norm": 1.7409612334315907,
"learning_rate": 7.638888888888889e-05,
"loss": 1.7874,
"step": 110
},
{
"epoch": 0.08354364285092682,
"grad_norm": 1.9712592588798674,
"learning_rate": 8.333333333333334e-05,
"loss": 1.7892,
"step": 120
},
{
"epoch": 0.09050561308850405,
"grad_norm": 2.509099517180512,
"learning_rate": 9.027777777777779e-05,
"loss": 1.7808,
"step": 130
},
{
"epoch": 0.09746758332608128,
"grad_norm": 2.0842685380596038,
"learning_rate": 9.722222222222223e-05,
"loss": 1.803,
"step": 140
},
{
"epoch": 0.10442955356365852,
"grad_norm": 1.8429004845744987,
"learning_rate": 0.00010416666666666667,
"loss": 1.7588,
"step": 150
},
{
"epoch": 0.11139152380123575,
"grad_norm": 2.8327524674330524,
"learning_rate": 0.00011111111111111112,
"loss": 1.7772,
"step": 160
},
{
"epoch": 0.11835349403881298,
"grad_norm": 2.0729891765954154,
"learning_rate": 0.00011805555555555556,
"loss": 1.7641,
"step": 170
},
{
"epoch": 0.12531546427639023,
"grad_norm": 7.58810403295457,
"learning_rate": 0.000125,
"loss": 1.833,
"step": 180
},
{
"epoch": 0.13227743451396745,
"grad_norm": 19.362231963576104,
"learning_rate": 0.00013194444444444446,
"loss": 1.7874,
"step": 190
},
{
"epoch": 0.1392394047515447,
"grad_norm": 3.96368325072897,
"learning_rate": 0.0001388888888888889,
"loss": 1.846,
"step": 200
},
{
"epoch": 0.14620137498912192,
"grad_norm": 2.7386120910604945,
"learning_rate": 0.00014583333333333335,
"loss": 1.8197,
"step": 210
},
{
"epoch": 0.15316334522669917,
"grad_norm": 2.110144128803464,
"learning_rate": 0.00015277777777777777,
"loss": 1.7561,
"step": 220
},
{
"epoch": 0.16012531546427639,
"grad_norm": 2.5725492351480352,
"learning_rate": 0.00015972222222222223,
"loss": 1.7749,
"step": 230
},
{
"epoch": 0.16708728570185363,
"grad_norm": 1.121359657360383,
"learning_rate": 0.0001666666666666667,
"loss": 1.7614,
"step": 240
},
{
"epoch": 0.17404925593943085,
"grad_norm": 2.155597643370339,
"learning_rate": 0.00017361111111111112,
"loss": 1.7587,
"step": 250
},
{
"epoch": 0.1810112261770081,
"grad_norm": 2.5773482568660393,
"learning_rate": 0.00018055555555555557,
"loss": 1.7934,
"step": 260
},
{
"epoch": 0.18797319641458532,
"grad_norm": 1.6740595013333657,
"learning_rate": 0.0001875,
"loss": 1.7504,
"step": 270
},
{
"epoch": 0.19493516665216257,
"grad_norm": 2.2858392200215696,
"learning_rate": 0.00019444444444444446,
"loss": 1.7499,
"step": 280
},
{
"epoch": 0.2018971368897398,
"grad_norm": 1.7434751943536266,
"learning_rate": 0.0001999999336897035,
"loss": 1.7543,
"step": 290
},
{
"epoch": 0.20885910712731703,
"grad_norm": 1.3619486244331678,
"learning_rate": 0.00019999761283856016,
"loss": 1.7123,
"step": 300
},
{
"epoch": 0.21582107736489425,
"grad_norm": 1.6373278228861707,
"learning_rate": 0.00019999197656053288,
"loss": 1.7195,
"step": 310
},
{
"epoch": 0.2227830476024715,
"grad_norm": 1.281241393641898,
"learning_rate": 0.00019998302504249278,
"loss": 1.7336,
"step": 320
},
{
"epoch": 0.22974501784004872,
"grad_norm": 1.4989445469875955,
"learning_rate": 0.000199970758581228,
"loss": 1.7219,
"step": 330
},
{
"epoch": 0.23670698807762597,
"grad_norm": 1.321518251513136,
"learning_rate": 0.00019995517758343386,
"loss": 1.7302,
"step": 340
},
{
"epoch": 0.24366895831520322,
"grad_norm": 1.2671299687318527,
"learning_rate": 0.0001999362825656992,
"loss": 1.7159,
"step": 350
},
{
"epoch": 0.25063092855278046,
"grad_norm": 1.0577637068118013,
"learning_rate": 0.00019991407415448947,
"loss": 1.7203,
"step": 360
},
{
"epoch": 0.2575928987903577,
"grad_norm": 1.134207084144878,
"learning_rate": 0.00019988855308612595,
"loss": 1.7114,
"step": 370
},
{
"epoch": 0.2645548690279349,
"grad_norm": 1.2266368964891576,
"learning_rate": 0.00019985972020676116,
"loss": 1.7238,
"step": 380
},
{
"epoch": 0.2715168392655121,
"grad_norm": 1.1990301126734546,
"learning_rate": 0.00019982757647235094,
"loss": 1.704,
"step": 390
},
{
"epoch": 0.2784788095030894,
"grad_norm": 0.7719332777254844,
"learning_rate": 0.0001997921229486228,
"loss": 1.7127,
"step": 400
},
{
"epoch": 0.2854407797406666,
"grad_norm": 0.9054141958817113,
"learning_rate": 0.00019975336081104038,
"loss": 1.6892,
"step": 410
},
{
"epoch": 0.29240274997824384,
"grad_norm": 0.7201089451408423,
"learning_rate": 0.00019971129134476473,
"loss": 1.689,
"step": 420
},
{
"epoch": 0.29936472021582106,
"grad_norm": 0.6596449276890678,
"learning_rate": 0.00019966591594461157,
"loss": 1.7046,
"step": 430
},
{
"epoch": 0.30632669045339833,
"grad_norm": 0.9894358407915632,
"learning_rate": 0.000199617236115005,
"loss": 1.7063,
"step": 440
},
{
"epoch": 0.31328866069097555,
"grad_norm": 0.6354863462129365,
"learning_rate": 0.00019956525346992768,
"loss": 1.6896,
"step": 450
},
{
"epoch": 0.32025063092855277,
"grad_norm": 1.004907546559211,
"learning_rate": 0.0001995099697328674,
"loss": 1.6662,
"step": 460
},
{
"epoch": 0.32721260116613,
"grad_norm": 0.5536272096656565,
"learning_rate": 0.00019945138673675973,
"loss": 1.7043,
"step": 470
},
{
"epoch": 0.33417457140370727,
"grad_norm": 1.0951977004377467,
"learning_rate": 0.00019938950642392746,
"loss": 1.6751,
"step": 480
},
{
"epoch": 0.3411365416412845,
"grad_norm": 2.3040787232248343,
"learning_rate": 0.00019932433084601613,
"loss": 1.7194,
"step": 490
},
{
"epoch": 0.3480985118788617,
"grad_norm": 1.484568533076407,
"learning_rate": 0.00019925586216392596,
"loss": 1.6743,
"step": 500
}
],
"logging_steps": 10,
"max_steps": 5744,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 418654621532160.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}