vinhnq29's picture
Upload folder using huggingface_hub
b9d9807 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.986078886310905,
"eval_steps": 500,
"global_step": 107,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.02,
"grad_norm": 28.83467674255371,
"learning_rate": 2e-05,
"loss": 2.7878,
"step": 1
},
{
"epoch": 0.04,
"grad_norm": 28.955718994140625,
"learning_rate": 4e-05,
"loss": 2.7826,
"step": 2
},
{
"epoch": 0.06,
"grad_norm": 17.856950759887695,
"learning_rate": 6e-05,
"loss": 2.2388,
"step": 3
},
{
"epoch": 0.07,
"grad_norm": 16.615158081054688,
"learning_rate": 8e-05,
"loss": 1.7809,
"step": 4
},
{
"epoch": 0.09,
"grad_norm": 34.79571533203125,
"learning_rate": 0.0001,
"loss": 1.9714,
"step": 5
},
{
"epoch": 0.11,
"grad_norm": 6.3113694190979,
"learning_rate": 9.935064935064936e-05,
"loss": 1.4032,
"step": 6
},
{
"epoch": 0.13,
"grad_norm": 8.230330467224121,
"learning_rate": 9.870129870129871e-05,
"loss": 1.241,
"step": 7
},
{
"epoch": 0.15,
"grad_norm": 17.19231605529785,
"learning_rate": 9.805194805194806e-05,
"loss": 1.2388,
"step": 8
},
{
"epoch": 0.17,
"grad_norm": 6.900284767150879,
"learning_rate": 9.74025974025974e-05,
"loss": 1.0881,
"step": 9
},
{
"epoch": 0.19,
"grad_norm": 2.7498185634613037,
"learning_rate": 9.675324675324677e-05,
"loss": 1.0071,
"step": 10
},
{
"epoch": 0.2,
"grad_norm": 4.016067028045654,
"learning_rate": 9.610389610389611e-05,
"loss": 0.9962,
"step": 11
},
{
"epoch": 0.22,
"grad_norm": 2.365187883377075,
"learning_rate": 9.545454545454546e-05,
"loss": 0.9928,
"step": 12
},
{
"epoch": 0.24,
"grad_norm": 1.4348371028900146,
"learning_rate": 9.480519480519481e-05,
"loss": 0.9107,
"step": 13
},
{
"epoch": 0.26,
"grad_norm": 1.8148932456970215,
"learning_rate": 9.415584415584417e-05,
"loss": 0.853,
"step": 14
},
{
"epoch": 0.28,
"grad_norm": 1.9855575561523438,
"learning_rate": 9.35064935064935e-05,
"loss": 0.8598,
"step": 15
},
{
"epoch": 0.3,
"grad_norm": 1.9960970878601074,
"learning_rate": 9.285714285714286e-05,
"loss": 0.8384,
"step": 16
},
{
"epoch": 0.32,
"grad_norm": 1.4565762281417847,
"learning_rate": 9.220779220779221e-05,
"loss": 0.7895,
"step": 17
},
{
"epoch": 0.33,
"grad_norm": 1.419858694076538,
"learning_rate": 9.155844155844156e-05,
"loss": 0.8162,
"step": 18
},
{
"epoch": 0.35,
"grad_norm": 1.3651608228683472,
"learning_rate": 9.090909090909092e-05,
"loss": 0.7924,
"step": 19
},
{
"epoch": 0.37,
"grad_norm": 1.1524626016616821,
"learning_rate": 9.025974025974027e-05,
"loss": 0.7484,
"step": 20
},
{
"epoch": 0.39,
"grad_norm": 0.9174069166183472,
"learning_rate": 8.961038961038961e-05,
"loss": 0.7276,
"step": 21
},
{
"epoch": 0.41,
"grad_norm": 1.1832919120788574,
"learning_rate": 8.896103896103896e-05,
"loss": 0.6658,
"step": 22
},
{
"epoch": 0.43,
"grad_norm": 1.3024309873580933,
"learning_rate": 8.831168831168831e-05,
"loss": 0.7379,
"step": 23
},
{
"epoch": 0.45,
"grad_norm": 1.1416062116622925,
"learning_rate": 8.766233766233767e-05,
"loss": 0.7202,
"step": 24
},
{
"epoch": 0.46,
"grad_norm": 1.2058277130126953,
"learning_rate": 8.701298701298701e-05,
"loss": 0.6777,
"step": 25
},
{
"epoch": 0.48,
"grad_norm": 1.0915583372116089,
"learning_rate": 8.636363636363637e-05,
"loss": 0.7063,
"step": 26
},
{
"epoch": 0.5,
"grad_norm": 1.0384303331375122,
"learning_rate": 8.571428571428571e-05,
"loss": 0.6447,
"step": 27
},
{
"epoch": 0.52,
"grad_norm": 1.131259560585022,
"learning_rate": 8.506493506493507e-05,
"loss": 0.6911,
"step": 28
},
{
"epoch": 0.54,
"grad_norm": 1.1505099534988403,
"learning_rate": 8.441558441558442e-05,
"loss": 0.6142,
"step": 29
},
{
"epoch": 0.56,
"grad_norm": 2.6675026416778564,
"learning_rate": 8.376623376623377e-05,
"loss": 0.6181,
"step": 30
},
{
"epoch": 0.58,
"grad_norm": 1.2021816968917847,
"learning_rate": 8.311688311688312e-05,
"loss": 0.6542,
"step": 31
},
{
"epoch": 0.59,
"grad_norm": 3.363269090652466,
"learning_rate": 8.246753246753248e-05,
"loss": 0.5968,
"step": 32
},
{
"epoch": 0.61,
"grad_norm": 3.8007972240448,
"learning_rate": 8.181818181818183e-05,
"loss": 0.5983,
"step": 33
},
{
"epoch": 0.63,
"grad_norm": 3.570556163787842,
"learning_rate": 8.116883116883117e-05,
"loss": 0.6375,
"step": 34
},
{
"epoch": 0.65,
"grad_norm": 4.428549766540527,
"learning_rate": 8.051948051948052e-05,
"loss": 0.6519,
"step": 35
},
{
"epoch": 0.67,
"grad_norm": 3.7240254878997803,
"learning_rate": 7.987012987012987e-05,
"loss": 0.5935,
"step": 36
},
{
"epoch": 0.69,
"grad_norm": 2.453350782394409,
"learning_rate": 7.922077922077923e-05,
"loss": 0.604,
"step": 37
},
{
"epoch": 0.71,
"grad_norm": 3.341071844100952,
"learning_rate": 7.857142857142858e-05,
"loss": 0.5994,
"step": 38
},
{
"epoch": 0.72,
"grad_norm": 3.4222161769866943,
"learning_rate": 7.792207792207793e-05,
"loss": 0.5731,
"step": 39
},
{
"epoch": 0.74,
"grad_norm": 2.9063804149627686,
"learning_rate": 7.727272727272727e-05,
"loss": 0.5566,
"step": 40
},
{
"epoch": 0.76,
"grad_norm": 3.170330762863159,
"learning_rate": 7.662337662337662e-05,
"loss": 0.5684,
"step": 41
},
{
"epoch": 0.78,
"grad_norm": 2.8363170623779297,
"learning_rate": 7.597402597402598e-05,
"loss": 0.5936,
"step": 42
},
{
"epoch": 0.8,
"grad_norm": 2.3944592475891113,
"learning_rate": 7.532467532467533e-05,
"loss": 0.5571,
"step": 43
},
{
"epoch": 0.82,
"grad_norm": 2.816237688064575,
"learning_rate": 7.467532467532467e-05,
"loss": 0.5598,
"step": 44
},
{
"epoch": 0.84,
"grad_norm": 2.57438325881958,
"learning_rate": 7.402597402597404e-05,
"loss": 0.5615,
"step": 45
},
{
"epoch": 0.85,
"grad_norm": 2.3164243698120117,
"learning_rate": 7.337662337662338e-05,
"loss": 0.5381,
"step": 46
},
{
"epoch": 0.87,
"grad_norm": 2.4226622581481934,
"learning_rate": 7.272727272727273e-05,
"loss": 0.5592,
"step": 47
},
{
"epoch": 0.89,
"grad_norm": 1.5568475723266602,
"learning_rate": 7.207792207792208e-05,
"loss": 0.5729,
"step": 48
},
{
"epoch": 0.91,
"grad_norm": 1.1320440769195557,
"learning_rate": 7.142857142857143e-05,
"loss": 0.5546,
"step": 49
},
{
"epoch": 0.93,
"grad_norm": 2.220273733139038,
"learning_rate": 7.077922077922077e-05,
"loss": 0.5587,
"step": 50
},
{
"epoch": 0.95,
"grad_norm": 1.8099664449691772,
"learning_rate": 7.012987012987014e-05,
"loss": 0.5403,
"step": 51
},
{
"epoch": 0.97,
"grad_norm": 1.7992054224014282,
"learning_rate": 6.948051948051948e-05,
"loss": 0.5374,
"step": 52
},
{
"epoch": 0.98,
"grad_norm": 1.5891600847244263,
"learning_rate": 6.883116883116883e-05,
"loss": 0.5188,
"step": 53
},
{
"epoch": 1.0,
"grad_norm": 1.713174819946289,
"learning_rate": 6.818181818181818e-05,
"loss": 0.5632,
"step": 54
},
{
"epoch": 1.02,
"grad_norm": 1.529587984085083,
"learning_rate": 6.753246753246754e-05,
"loss": 0.4607,
"step": 55
},
{
"epoch": 1.04,
"grad_norm": 1.243905782699585,
"learning_rate": 6.688311688311688e-05,
"loss": 0.4807,
"step": 56
},
{
"epoch": 1.06,
"grad_norm": 0.92779141664505,
"learning_rate": 6.623376623376624e-05,
"loss": 0.4515,
"step": 57
},
{
"epoch": 1.08,
"grad_norm": 1.8945612907409668,
"learning_rate": 6.55844155844156e-05,
"loss": 0.4788,
"step": 58
},
{
"epoch": 1.1,
"grad_norm": 1.6997075080871582,
"learning_rate": 6.493506493506494e-05,
"loss": 0.4411,
"step": 59
},
{
"epoch": 1.11,
"grad_norm": 1.4320181608200073,
"learning_rate": 6.428571428571429e-05,
"loss": 0.4474,
"step": 60
},
{
"epoch": 1.13,
"grad_norm": 1.5225821733474731,
"learning_rate": 6.363636363636364e-05,
"loss": 0.4578,
"step": 61
},
{
"epoch": 1.15,
"grad_norm": 1.4867883920669556,
"learning_rate": 6.2987012987013e-05,
"loss": 0.4838,
"step": 62
},
{
"epoch": 1.17,
"grad_norm": 1.3935158252716064,
"learning_rate": 6.233766233766233e-05,
"loss": 0.4383,
"step": 63
},
{
"epoch": 1.19,
"grad_norm": 1.3193836212158203,
"learning_rate": 6.16883116883117e-05,
"loss": 0.4369,
"step": 64
},
{
"epoch": 1.21,
"grad_norm": 1.0565460920333862,
"learning_rate": 6.103896103896104e-05,
"loss": 0.4618,
"step": 65
},
{
"epoch": 1.23,
"grad_norm": 1.9069595336914062,
"learning_rate": 6.03896103896104e-05,
"loss": 0.4636,
"step": 66
},
{
"epoch": 1.24,
"grad_norm": 1.592588186264038,
"learning_rate": 5.9740259740259744e-05,
"loss": 0.4363,
"step": 67
},
{
"epoch": 1.26,
"grad_norm": 1.392088770866394,
"learning_rate": 5.90909090909091e-05,
"loss": 0.4573,
"step": 68
},
{
"epoch": 1.28,
"grad_norm": 1.4312165975570679,
"learning_rate": 5.844155844155844e-05,
"loss": 0.4404,
"step": 69
},
{
"epoch": 1.3,
"grad_norm": 1.4288945198059082,
"learning_rate": 5.7792207792207796e-05,
"loss": 0.4517,
"step": 70
},
{
"epoch": 1.32,
"grad_norm": 1.247237205505371,
"learning_rate": 5.714285714285714e-05,
"loss": 0.4731,
"step": 71
},
{
"epoch": 1.34,
"grad_norm": 1.4292420148849487,
"learning_rate": 5.64935064935065e-05,
"loss": 0.4561,
"step": 72
},
{
"epoch": 1.35,
"grad_norm": 1.1647123098373413,
"learning_rate": 5.584415584415584e-05,
"loss": 0.4702,
"step": 73
},
{
"epoch": 1.37,
"grad_norm": 1.6123448610305786,
"learning_rate": 5.51948051948052e-05,
"loss": 0.4681,
"step": 74
},
{
"epoch": 1.39,
"grad_norm": 1.3807860612869263,
"learning_rate": 5.4545454545454546e-05,
"loss": 0.4217,
"step": 75
},
{
"epoch": 1.41,
"grad_norm": 1.4928926229476929,
"learning_rate": 5.38961038961039e-05,
"loss": 0.4344,
"step": 76
},
{
"epoch": 1.43,
"grad_norm": 1.3118643760681152,
"learning_rate": 5.3246753246753245e-05,
"loss": 0.4509,
"step": 77
},
{
"epoch": 1.45,
"grad_norm": 1.1260207891464233,
"learning_rate": 5.25974025974026e-05,
"loss": 0.4423,
"step": 78
},
{
"epoch": 1.47,
"grad_norm": 1.1223907470703125,
"learning_rate": 5.1948051948051944e-05,
"loss": 0.4238,
"step": 79
},
{
"epoch": 1.48,
"grad_norm": 1.123289704322815,
"learning_rate": 5.1298701298701304e-05,
"loss": 0.4415,
"step": 80
},
{
"epoch": 1.5,
"grad_norm": 0.8493527770042419,
"learning_rate": 5.064935064935065e-05,
"loss": 0.4278,
"step": 81
},
{
"epoch": 1.52,
"grad_norm": 1.3558621406555176,
"learning_rate": 5e-05,
"loss": 0.4326,
"step": 82
},
{
"epoch": 1.54,
"grad_norm": 1.198364496231079,
"learning_rate": 4.9350649350649355e-05,
"loss": 0.4504,
"step": 83
},
{
"epoch": 1.56,
"grad_norm": 1.318117380142212,
"learning_rate": 4.87012987012987e-05,
"loss": 0.4389,
"step": 84
},
{
"epoch": 1.58,
"grad_norm": 1.323292851448059,
"learning_rate": 4.8051948051948054e-05,
"loss": 0.4559,
"step": 85
},
{
"epoch": 1.6,
"grad_norm": 1.0558339357376099,
"learning_rate": 4.740259740259741e-05,
"loss": 0.4613,
"step": 86
},
{
"epoch": 1.61,
"grad_norm": 1.0301963090896606,
"learning_rate": 4.675324675324675e-05,
"loss": 0.4194,
"step": 87
},
{
"epoch": 1.63,
"grad_norm": 0.9772946834564209,
"learning_rate": 4.6103896103896106e-05,
"loss": 0.4278,
"step": 88
},
{
"epoch": 1.65,
"grad_norm": 0.93486487865448,
"learning_rate": 4.545454545454546e-05,
"loss": 0.454,
"step": 89
},
{
"epoch": 1.67,
"grad_norm": 1.1458832025527954,
"learning_rate": 4.4805194805194805e-05,
"loss": 0.4363,
"step": 90
},
{
"epoch": 1.69,
"grad_norm": 0.9339586496353149,
"learning_rate": 4.415584415584416e-05,
"loss": 0.434,
"step": 91
},
{
"epoch": 1.71,
"grad_norm": 1.3206239938735962,
"learning_rate": 4.3506493506493503e-05,
"loss": 0.4431,
"step": 92
},
{
"epoch": 1.73,
"grad_norm": 1.1199604272842407,
"learning_rate": 4.2857142857142856e-05,
"loss": 0.4472,
"step": 93
},
{
"epoch": 1.74,
"grad_norm": 1.0672180652618408,
"learning_rate": 4.220779220779221e-05,
"loss": 0.4091,
"step": 94
},
{
"epoch": 1.76,
"grad_norm": 1.0920594930648804,
"learning_rate": 4.155844155844156e-05,
"loss": 0.417,
"step": 95
},
{
"epoch": 1.78,
"grad_norm": 0.9002637267112732,
"learning_rate": 4.0909090909090915e-05,
"loss": 0.4103,
"step": 96
},
{
"epoch": 1.8,
"grad_norm": 0.8468071222305298,
"learning_rate": 4.025974025974026e-05,
"loss": 0.4167,
"step": 97
},
{
"epoch": 1.82,
"grad_norm": 0.9398940205574036,
"learning_rate": 3.9610389610389614e-05,
"loss": 0.4201,
"step": 98
},
{
"epoch": 1.84,
"grad_norm": 0.7842636108398438,
"learning_rate": 3.8961038961038966e-05,
"loss": 0.4064,
"step": 99
},
{
"epoch": 1.86,
"grad_norm": 1.0805550813674927,
"learning_rate": 3.831168831168831e-05,
"loss": 0.4271,
"step": 100
},
{
"epoch": 1.87,
"grad_norm": 0.8424351215362549,
"learning_rate": 3.7662337662337665e-05,
"loss": 0.4087,
"step": 101
},
{
"epoch": 1.89,
"grad_norm": 1.0850481986999512,
"learning_rate": 3.701298701298702e-05,
"loss": 0.409,
"step": 102
},
{
"epoch": 1.91,
"grad_norm": 0.9894604086875916,
"learning_rate": 3.6363636363636364e-05,
"loss": 0.4139,
"step": 103
},
{
"epoch": 1.93,
"grad_norm": 0.97883540391922,
"learning_rate": 3.571428571428572e-05,
"loss": 0.4281,
"step": 104
},
{
"epoch": 1.95,
"grad_norm": 0.9291779398918152,
"learning_rate": 3.506493506493507e-05,
"loss": 0.4376,
"step": 105
},
{
"epoch": 1.97,
"grad_norm": 0.8272209167480469,
"learning_rate": 3.4415584415584416e-05,
"loss": 0.42,
"step": 106
},
{
"epoch": 1.99,
"grad_norm": 0.8428053855895996,
"learning_rate": 3.376623376623377e-05,
"loss": 0.4131,
"step": 107
}
],
"logging_steps": 1,
"max_steps": 159,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 5,
"total_flos": 1.6171847189151744e+17,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}