diaenra's picture
Training in progress, step 100, checkpoint
a3a46f4 verified
{
"best_metric": 1.62236487865448,
"best_model_checkpoint": "miner_id_24/checkpoint-100",
"epoch": 0.0017317741593318816,
"eval_steps": 25,
"global_step": 100,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 1.7317741593318815e-05,
"grad_norm": 1.3246665000915527,
"learning_rate": 4e-05,
"loss": 1.6209,
"step": 1
},
{
"epoch": 1.7317741593318815e-05,
"eval_loss": 2.6184682846069336,
"eval_runtime": 4891.6495,
"eval_samples_per_second": 4.971,
"eval_steps_per_second": 2.485,
"step": 1
},
{
"epoch": 3.463548318663763e-05,
"grad_norm": 1.7559422254562378,
"learning_rate": 8e-05,
"loss": 2.6328,
"step": 2
},
{
"epoch": 5.1953224779956444e-05,
"grad_norm": 1.3109291791915894,
"learning_rate": 0.00012,
"loss": 1.8847,
"step": 3
},
{
"epoch": 6.927096637327526e-05,
"grad_norm": 1.6149224042892456,
"learning_rate": 0.00016,
"loss": 2.4206,
"step": 4
},
{
"epoch": 8.658870796659407e-05,
"grad_norm": 1.6735433340072632,
"learning_rate": 0.0002,
"loss": 2.4259,
"step": 5
},
{
"epoch": 0.00010390644955991289,
"grad_norm": 1.4547559022903442,
"learning_rate": 0.00019994532573409262,
"loss": 1.7482,
"step": 6
},
{
"epoch": 0.0001212241911532317,
"grad_norm": 1.628714919090271,
"learning_rate": 0.00019978136272187747,
"loss": 2.2129,
"step": 7
},
{
"epoch": 0.00013854193274655052,
"grad_norm": 1.8027219772338867,
"learning_rate": 0.00019950829025450114,
"loss": 1.8543,
"step": 8
},
{
"epoch": 0.00015585967433986934,
"grad_norm": 2.021082639694214,
"learning_rate": 0.00019912640693269752,
"loss": 2.0592,
"step": 9
},
{
"epoch": 0.00017317741593318815,
"grad_norm": 2.312199115753174,
"learning_rate": 0.00019863613034027224,
"loss": 2.055,
"step": 10
},
{
"epoch": 0.00019049515752650697,
"grad_norm": 1.8753467798233032,
"learning_rate": 0.00019803799658748094,
"loss": 1.7057,
"step": 11
},
{
"epoch": 0.00020781289911982577,
"grad_norm": 3.184194803237915,
"learning_rate": 0.0001973326597248006,
"loss": 2.1195,
"step": 12
},
{
"epoch": 0.0002251306407131446,
"grad_norm": 2.2157609462738037,
"learning_rate": 0.00019652089102773488,
"loss": 1.6142,
"step": 13
},
{
"epoch": 0.0002424483823064634,
"grad_norm": 1.8846286535263062,
"learning_rate": 0.00019560357815343577,
"loss": 1.6909,
"step": 14
},
{
"epoch": 0.00025976612389978223,
"grad_norm": 1.8501865863800049,
"learning_rate": 0.00019458172417006347,
"loss": 1.5635,
"step": 15
},
{
"epoch": 0.00027708386549310103,
"grad_norm": 1.9967668056488037,
"learning_rate": 0.0001934564464599461,
"loss": 1.7519,
"step": 16
},
{
"epoch": 0.00029440160708641983,
"grad_norm": 1.8826665878295898,
"learning_rate": 0.00019222897549773848,
"loss": 1.6652,
"step": 17
},
{
"epoch": 0.0003117193486797387,
"grad_norm": 2.2450180053710938,
"learning_rate": 0.00019090065350491626,
"loss": 1.8715,
"step": 18
},
{
"epoch": 0.0003290370902730575,
"grad_norm": 2.0638725757598877,
"learning_rate": 0.00018947293298207635,
"loss": 1.7423,
"step": 19
},
{
"epoch": 0.0003463548318663763,
"grad_norm": 3.073434829711914,
"learning_rate": 0.0001879473751206489,
"loss": 2.1458,
"step": 20
},
{
"epoch": 0.00036367257345969515,
"grad_norm": 1.9750235080718994,
"learning_rate": 0.00018632564809575742,
"loss": 1.6923,
"step": 21
},
{
"epoch": 0.00038099031505301395,
"grad_norm": 2.1666667461395264,
"learning_rate": 0.00018460952524209355,
"loss": 1.6162,
"step": 22
},
{
"epoch": 0.00039830805664633275,
"grad_norm": 1.9802206754684448,
"learning_rate": 0.00018280088311480201,
"loss": 1.6125,
"step": 23
},
{
"epoch": 0.00041562579823965155,
"grad_norm": 1.9258074760437012,
"learning_rate": 0.00018090169943749476,
"loss": 1.6174,
"step": 24
},
{
"epoch": 0.0004329435398329704,
"grad_norm": 2.670494556427002,
"learning_rate": 0.00017891405093963938,
"loss": 2.1219,
"step": 25
},
{
"epoch": 0.0004329435398329704,
"eval_loss": 1.7579680681228638,
"eval_runtime": 4921.5904,
"eval_samples_per_second": 4.94,
"eval_steps_per_second": 2.47,
"step": 25
},
{
"epoch": 0.0004502612814262892,
"grad_norm": 2.315001964569092,
"learning_rate": 0.00017684011108568592,
"loss": 1.9471,
"step": 26
},
{
"epoch": 0.000467579023019608,
"grad_norm": 4.8373613357543945,
"learning_rate": 0.0001746821476984154,
"loss": 1.4014,
"step": 27
},
{
"epoch": 0.0004848967646129268,
"grad_norm": 2.2539186477661133,
"learning_rate": 0.00017244252047910892,
"loss": 1.8659,
"step": 28
},
{
"epoch": 0.0005022145062062456,
"grad_norm": 2.21065616607666,
"learning_rate": 0.00017012367842724887,
"loss": 1.8552,
"step": 29
},
{
"epoch": 0.0005195322477995645,
"grad_norm": 2.3214337825775146,
"learning_rate": 0.00016772815716257412,
"loss": 2.0893,
"step": 30
},
{
"epoch": 0.0005368499893928833,
"grad_norm": 2.827920436859131,
"learning_rate": 0.00016525857615241687,
"loss": 1.7269,
"step": 31
},
{
"epoch": 0.0005541677309862021,
"grad_norm": 2.3385794162750244,
"learning_rate": 0.0001627176358473537,
"loss": 1.4844,
"step": 32
},
{
"epoch": 0.0005714854725795209,
"grad_norm": 2.1411702632904053,
"learning_rate": 0.00016010811472830252,
"loss": 1.6424,
"step": 33
},
{
"epoch": 0.0005888032141728397,
"grad_norm": 1.8949635028839111,
"learning_rate": 0.00015743286626829437,
"loss": 1.8416,
"step": 34
},
{
"epoch": 0.0006061209557661585,
"grad_norm": 2.325913667678833,
"learning_rate": 0.00015469481581224272,
"loss": 1.9681,
"step": 35
},
{
"epoch": 0.0006234386973594774,
"grad_norm": 2.5775182247161865,
"learning_rate": 0.00015189695737812152,
"loss": 2.0605,
"step": 36
},
{
"epoch": 0.0006407564389527961,
"grad_norm": 2.1500771045684814,
"learning_rate": 0.00014904235038305083,
"loss": 1.7005,
"step": 37
},
{
"epoch": 0.000658074180546115,
"grad_norm": 1.9176790714263916,
"learning_rate": 0.0001461341162978688,
"loss": 1.544,
"step": 38
},
{
"epoch": 0.0006753919221394338,
"grad_norm": 5.9821248054504395,
"learning_rate": 0.00014317543523384928,
"loss": 1.5349,
"step": 39
},
{
"epoch": 0.0006927096637327526,
"grad_norm": 1.723404049873352,
"learning_rate": 0.00014016954246529696,
"loss": 1.4033,
"step": 40
},
{
"epoch": 0.0007100274053260714,
"grad_norm": 2.131654739379883,
"learning_rate": 0.00013711972489182208,
"loss": 1.8816,
"step": 41
},
{
"epoch": 0.0007273451469193903,
"grad_norm": 2.1149301528930664,
"learning_rate": 0.00013402931744416433,
"loss": 1.7547,
"step": 42
},
{
"epoch": 0.000744662888512709,
"grad_norm": 2.1402599811553955,
"learning_rate": 0.00013090169943749476,
"loss": 1.4692,
"step": 43
},
{
"epoch": 0.0007619806301060279,
"grad_norm": 2.189884662628174,
"learning_rate": 0.00012774029087618446,
"loss": 1.6887,
"step": 44
},
{
"epoch": 0.0007792983716993466,
"grad_norm": 2.3986988067626953,
"learning_rate": 0.00012454854871407994,
"loss": 1.5637,
"step": 45
},
{
"epoch": 0.0007966161132926655,
"grad_norm": 2.0756654739379883,
"learning_rate": 0.0001213299630743747,
"loss": 1.3412,
"step": 46
},
{
"epoch": 0.0008139338548859844,
"grad_norm": 2.2220942974090576,
"learning_rate": 0.000118088053433211,
"loss": 1.9958,
"step": 47
},
{
"epoch": 0.0008312515964793031,
"grad_norm": 2.2260262966156006,
"learning_rate": 0.0001148263647711842,
"loss": 1.8054,
"step": 48
},
{
"epoch": 0.000848569338072622,
"grad_norm": 2.96877121925354,
"learning_rate": 0.00011154846369695863,
"loss": 1.7525,
"step": 49
},
{
"epoch": 0.0008658870796659408,
"grad_norm": 2.6540238857269287,
"learning_rate": 0.00010825793454723325,
"loss": 1.731,
"step": 50
},
{
"epoch": 0.0008658870796659408,
"eval_loss": 1.665256142616272,
"eval_runtime": 4923.0054,
"eval_samples_per_second": 4.939,
"eval_steps_per_second": 2.469,
"step": 50
},
{
"epoch": 0.0008832048212592596,
"grad_norm": 1.3105987310409546,
"learning_rate": 0.00010495837546732224,
"loss": 1.4402,
"step": 51
},
{
"epoch": 0.0009005225628525784,
"grad_norm": 1.4101057052612305,
"learning_rate": 0.00010165339447663587,
"loss": 1.4313,
"step": 52
},
{
"epoch": 0.0009178403044458972,
"grad_norm": 1.5237852334976196,
"learning_rate": 9.834660552336415e-05,
"loss": 1.7096,
"step": 53
},
{
"epoch": 0.000935158046039216,
"grad_norm": 1.6052296161651611,
"learning_rate": 9.504162453267777e-05,
"loss": 1.4336,
"step": 54
},
{
"epoch": 0.0009524757876325349,
"grad_norm": 1.7279136180877686,
"learning_rate": 9.174206545276677e-05,
"loss": 1.7388,
"step": 55
},
{
"epoch": 0.0009697935292258536,
"grad_norm": 1.3830236196517944,
"learning_rate": 8.845153630304139e-05,
"loss": 1.2797,
"step": 56
},
{
"epoch": 0.0009871112708191726,
"grad_norm": 2.031076669692993,
"learning_rate": 8.517363522881579e-05,
"loss": 1.9524,
"step": 57
},
{
"epoch": 0.0010044290124124912,
"grad_norm": 1.437910556793213,
"learning_rate": 8.191194656678904e-05,
"loss": 1.4775,
"step": 58
},
{
"epoch": 0.00102174675400581,
"grad_norm": 1.4608979225158691,
"learning_rate": 7.867003692562534e-05,
"loss": 1.4998,
"step": 59
},
{
"epoch": 0.001039064495599129,
"grad_norm": 1.6951931715011597,
"learning_rate": 7.54514512859201e-05,
"loss": 1.8601,
"step": 60
},
{
"epoch": 0.0010563822371924478,
"grad_norm": 1.654209852218628,
"learning_rate": 7.225970912381556e-05,
"loss": 1.8643,
"step": 61
},
{
"epoch": 0.0010736999787857666,
"grad_norm": 1.837178349494934,
"learning_rate": 6.909830056250527e-05,
"loss": 1.7543,
"step": 62
},
{
"epoch": 0.0010910177203790853,
"grad_norm": 1.8673418760299683,
"learning_rate": 6.59706825558357e-05,
"loss": 1.408,
"step": 63
},
{
"epoch": 0.0011083354619724041,
"grad_norm": 1.660733938217163,
"learning_rate": 6.28802751081779e-05,
"loss": 1.6586,
"step": 64
},
{
"epoch": 0.001125653203565723,
"grad_norm": 1.6716222763061523,
"learning_rate": 5.983045753470308e-05,
"loss": 1.7785,
"step": 65
},
{
"epoch": 0.0011429709451590418,
"grad_norm": 1.4793193340301514,
"learning_rate": 5.6824564766150726e-05,
"loss": 1.524,
"step": 66
},
{
"epoch": 0.0011602886867523607,
"grad_norm": 1.4535232782363892,
"learning_rate": 5.386588370213124e-05,
"loss": 1.3191,
"step": 67
},
{
"epoch": 0.0011776064283456793,
"grad_norm": 1.9624608755111694,
"learning_rate": 5.095764961694922e-05,
"loss": 1.5482,
"step": 68
},
{
"epoch": 0.0011949241699389982,
"grad_norm": 1.7577117681503296,
"learning_rate": 4.810304262187852e-05,
"loss": 1.5874,
"step": 69
},
{
"epoch": 0.001212241911532317,
"grad_norm": 1.5105639696121216,
"learning_rate": 4.530518418775733e-05,
"loss": 1.5703,
"step": 70
},
{
"epoch": 0.001229559653125636,
"grad_norm": 1.529349446296692,
"learning_rate": 4.256713373170564e-05,
"loss": 1.1905,
"step": 71
},
{
"epoch": 0.0012468773947189548,
"grad_norm": 1.8508775234222412,
"learning_rate": 3.9891885271697496e-05,
"loss": 1.8208,
"step": 72
},
{
"epoch": 0.0012641951363122736,
"grad_norm": 1.593199610710144,
"learning_rate": 3.7282364152646297e-05,
"loss": 1.3999,
"step": 73
},
{
"epoch": 0.0012815128779055922,
"grad_norm": 1.6738864183425903,
"learning_rate": 3.4741423847583134e-05,
"loss": 1.4956,
"step": 74
},
{
"epoch": 0.001298830619498911,
"grad_norm": 1.758865475654602,
"learning_rate": 3.227184283742591e-05,
"loss": 1.631,
"step": 75
},
{
"epoch": 0.001298830619498911,
"eval_loss": 1.6363804340362549,
"eval_runtime": 4921.7849,
"eval_samples_per_second": 4.94,
"eval_steps_per_second": 2.47,
"step": 75
},
{
"epoch": 0.00131614836109223,
"grad_norm": 2.4752883911132812,
"learning_rate": 2.9876321572751144e-05,
"loss": 2.0203,
"step": 76
},
{
"epoch": 0.0013334661026855488,
"grad_norm": 2.1483538150787354,
"learning_rate": 2.7557479520891104e-05,
"loss": 1.5806,
"step": 77
},
{
"epoch": 0.0013507838442788677,
"grad_norm": 1.9293116331100464,
"learning_rate": 2.5317852301584643e-05,
"loss": 2.0804,
"step": 78
},
{
"epoch": 0.0013681015858721863,
"grad_norm": 1.519399881362915,
"learning_rate": 2.315988891431412e-05,
"loss": 1.4393,
"step": 79
},
{
"epoch": 0.0013854193274655052,
"grad_norm": 1.9969561100006104,
"learning_rate": 2.1085949060360654e-05,
"loss": 1.5912,
"step": 80
},
{
"epoch": 0.001402737069058824,
"grad_norm": 1.9334006309509277,
"learning_rate": 1.9098300562505266e-05,
"loss": 1.5095,
"step": 81
},
{
"epoch": 0.0014200548106521429,
"grad_norm": 1.8917192220687866,
"learning_rate": 1.7199116885197995e-05,
"loss": 1.8712,
"step": 82
},
{
"epoch": 0.0014373725522454617,
"grad_norm": 1.6807430982589722,
"learning_rate": 1.5390474757906446e-05,
"loss": 1.6938,
"step": 83
},
{
"epoch": 0.0014546902938387806,
"grad_norm": 1.641627550125122,
"learning_rate": 1.3674351904242611e-05,
"loss": 1.5532,
"step": 84
},
{
"epoch": 0.0014720080354320992,
"grad_norm": 2.6051013469696045,
"learning_rate": 1.2052624879351104e-05,
"loss": 1.8125,
"step": 85
},
{
"epoch": 0.001489325777025418,
"grad_norm": 1.6829463243484497,
"learning_rate": 1.0527067017923654e-05,
"loss": 1.5045,
"step": 86
},
{
"epoch": 0.001506643518618737,
"grad_norm": 1.6704628467559814,
"learning_rate": 9.09934649508375e-06,
"loss": 1.5594,
"step": 87
},
{
"epoch": 0.0015239612602120558,
"grad_norm": 2.3541147708892822,
"learning_rate": 7.771024502261526e-06,
"loss": 2.2284,
"step": 88
},
{
"epoch": 0.0015412790018053746,
"grad_norm": 2.744718074798584,
"learning_rate": 6.543553540053926e-06,
"loss": 1.2474,
"step": 89
},
{
"epoch": 0.0015585967433986933,
"grad_norm": 2.3282785415649414,
"learning_rate": 5.418275829936537e-06,
"loss": 1.9081,
"step": 90
},
{
"epoch": 0.0015759144849920121,
"grad_norm": 1.7221317291259766,
"learning_rate": 4.3964218465642355e-06,
"loss": 1.5467,
"step": 91
},
{
"epoch": 0.001593232226585331,
"grad_norm": 2.2455320358276367,
"learning_rate": 3.4791089722651436e-06,
"loss": 1.8693,
"step": 92
},
{
"epoch": 0.0016105499681786498,
"grad_norm": 1.896304726600647,
"learning_rate": 2.667340275199426e-06,
"loss": 1.3848,
"step": 93
},
{
"epoch": 0.0016278677097719687,
"grad_norm": 2.8605098724365234,
"learning_rate": 1.9620034125190644e-06,
"loss": 1.6698,
"step": 94
},
{
"epoch": 0.0016451854513652876,
"grad_norm": 1.6910480260849,
"learning_rate": 1.3638696597277679e-06,
"loss": 1.4249,
"step": 95
},
{
"epoch": 0.0016625031929586062,
"grad_norm": 3.1670644283294678,
"learning_rate": 8.735930673024806e-07,
"loss": 1.7864,
"step": 96
},
{
"epoch": 0.001679820934551925,
"grad_norm": 2.209742784500122,
"learning_rate": 4.917097454988584e-07,
"loss": 1.7708,
"step": 97
},
{
"epoch": 0.001697138676145244,
"grad_norm": 2.015029191970825,
"learning_rate": 2.1863727812254653e-07,
"loss": 1.471,
"step": 98
},
{
"epoch": 0.0017144564177385628,
"grad_norm": 2.405651569366455,
"learning_rate": 5.467426590739511e-08,
"loss": 1.6094,
"step": 99
},
{
"epoch": 0.0017317741593318816,
"grad_norm": 4.6894025802612305,
"learning_rate": 0.0,
"loss": 2.0444,
"step": 100
},
{
"epoch": 0.0017317741593318816,
"eval_loss": 1.62236487865448,
"eval_runtime": 4922.5195,
"eval_samples_per_second": 4.939,
"eval_steps_per_second": 2.47,
"step": 100
}
],
"logging_steps": 1,
"max_steps": 100,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 25,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 1,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.51427898343424e+16,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}