PEFT
Safetensors
M2WEB-HTML-WM / trainer_state.json
Gwanwoo's picture
Upload folder using huggingface_hub
664374f verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.9751552795031055,
"eval_steps": 41,
"global_step": 322,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.006211180124223602,
"grad_norm": 0.3814874589443207,
"learning_rate": 1.0000000000000002e-06,
"loss": 1.538,
"step": 1
},
{
"epoch": 0.006211180124223602,
"eval_loss": 1.570106863975525,
"eval_runtime": 26.7523,
"eval_samples_per_second": 2.953,
"eval_steps_per_second": 0.374,
"step": 1
},
{
"epoch": 0.012422360248447204,
"grad_norm": 0.34790050983428955,
"learning_rate": 2.0000000000000003e-06,
"loss": 1.5515,
"step": 2
},
{
"epoch": 0.018633540372670808,
"grad_norm": 0.315510630607605,
"learning_rate": 3e-06,
"loss": 1.6394,
"step": 3
},
{
"epoch": 0.024844720496894408,
"grad_norm": 0.333831787109375,
"learning_rate": 4.000000000000001e-06,
"loss": 1.5686,
"step": 4
},
{
"epoch": 0.031055900621118012,
"grad_norm": 0.3356265723705292,
"learning_rate": 5e-06,
"loss": 1.6442,
"step": 5
},
{
"epoch": 0.037267080745341616,
"grad_norm": 0.3302271366119385,
"learning_rate": 6e-06,
"loss": 1.6543,
"step": 6
},
{
"epoch": 0.043478260869565216,
"grad_norm": 0.3428245484828949,
"learning_rate": 7e-06,
"loss": 1.583,
"step": 7
},
{
"epoch": 0.049689440993788817,
"grad_norm": 0.3571615517139435,
"learning_rate": 8.000000000000001e-06,
"loss": 1.5863,
"step": 8
},
{
"epoch": 0.055900621118012424,
"grad_norm": 0.3609354496002197,
"learning_rate": 9e-06,
"loss": 1.5861,
"step": 9
},
{
"epoch": 0.062111801242236024,
"grad_norm": 0.38313618302345276,
"learning_rate": 1e-05,
"loss": 1.6357,
"step": 10
},
{
"epoch": 0.06832298136645963,
"grad_norm": 0.3678596317768097,
"learning_rate": 9.99974652980635e-06,
"loss": 1.6452,
"step": 11
},
{
"epoch": 0.07453416149068323,
"grad_norm": 0.36209872364997864,
"learning_rate": 9.998986144924253e-06,
"loss": 1.6364,
"step": 12
},
{
"epoch": 0.08074534161490683,
"grad_norm": 0.3838407099246979,
"learning_rate": 9.997718922447669e-06,
"loss": 1.6456,
"step": 13
},
{
"epoch": 0.08695652173913043,
"grad_norm": 0.3796548545360565,
"learning_rate": 9.995944990857848e-06,
"loss": 1.5417,
"step": 14
},
{
"epoch": 0.09316770186335403,
"grad_norm": 0.393232524394989,
"learning_rate": 9.993664530010308e-06,
"loss": 1.5553,
"step": 15
},
{
"epoch": 0.09937888198757763,
"grad_norm": 0.40382516384124756,
"learning_rate": 9.990877771116588e-06,
"loss": 1.4974,
"step": 16
},
{
"epoch": 0.10559006211180125,
"grad_norm": 0.410980224609375,
"learning_rate": 9.987584996720813e-06,
"loss": 1.5143,
"step": 17
},
{
"epoch": 0.11180124223602485,
"grad_norm": 0.43435031175613403,
"learning_rate": 9.983786540671052e-06,
"loss": 1.7078,
"step": 18
},
{
"epoch": 0.11801242236024845,
"grad_norm": 0.46852847933769226,
"learning_rate": 9.979482788085455e-06,
"loss": 1.5725,
"step": 19
},
{
"epoch": 0.12422360248447205,
"grad_norm": 0.44078120589256287,
"learning_rate": 9.974674175313228e-06,
"loss": 1.5146,
"step": 20
},
{
"epoch": 0.13043478260869565,
"grad_norm": 0.4520568251609802,
"learning_rate": 9.969361189890373e-06,
"loss": 1.551,
"step": 21
},
{
"epoch": 0.13664596273291926,
"grad_norm": 0.44940102100372314,
"learning_rate": 9.96354437049027e-06,
"loss": 1.5154,
"step": 22
},
{
"epoch": 0.14285714285714285,
"grad_norm": 0.42880144715309143,
"learning_rate": 9.957224306869053e-06,
"loss": 1.5664,
"step": 23
},
{
"epoch": 0.14906832298136646,
"grad_norm": 0.4296427071094513,
"learning_rate": 9.950401639805822e-06,
"loss": 1.5391,
"step": 24
},
{
"epoch": 0.15527950310559005,
"grad_norm": 0.4148462116718292,
"learning_rate": 9.943077061037672e-06,
"loss": 1.559,
"step": 25
},
{
"epoch": 0.16149068322981366,
"grad_norm": 0.45027458667755127,
"learning_rate": 9.935251313189564e-06,
"loss": 1.5797,
"step": 26
},
{
"epoch": 0.16770186335403728,
"grad_norm": 0.43325772881507874,
"learning_rate": 9.92692518969903e-06,
"loss": 1.5651,
"step": 27
},
{
"epoch": 0.17391304347826086,
"grad_norm": 0.4363733232021332,
"learning_rate": 9.91809953473572e-06,
"loss": 1.4171,
"step": 28
},
{
"epoch": 0.18012422360248448,
"grad_norm": 0.4001283347606659,
"learning_rate": 9.908775243115822e-06,
"loss": 1.5172,
"step": 29
},
{
"epoch": 0.18633540372670807,
"grad_norm": 0.42918145656585693,
"learning_rate": 9.89895326021134e-06,
"loss": 1.4501,
"step": 30
},
{
"epoch": 0.19254658385093168,
"grad_norm": 0.3902086019515991,
"learning_rate": 9.888634581854235e-06,
"loss": 1.4396,
"step": 31
},
{
"epoch": 0.19875776397515527,
"grad_norm": 0.4069195091724396,
"learning_rate": 9.87782025423547e-06,
"loss": 1.4399,
"step": 32
},
{
"epoch": 0.20496894409937888,
"grad_norm": 0.40982964634895325,
"learning_rate": 9.86651137379893e-06,
"loss": 1.4445,
"step": 33
},
{
"epoch": 0.2111801242236025,
"grad_norm": 0.4249035716056824,
"learning_rate": 9.854709087130261e-06,
"loss": 1.4402,
"step": 34
},
{
"epoch": 0.21739130434782608,
"grad_norm": 0.4334266781806946,
"learning_rate": 9.842414590840618e-06,
"loss": 1.4633,
"step": 35
},
{
"epoch": 0.2236024844720497,
"grad_norm": 0.4241773188114166,
"learning_rate": 9.829629131445342e-06,
"loss": 1.5354,
"step": 36
},
{
"epoch": 0.22981366459627328,
"grad_norm": 0.45813798904418945,
"learning_rate": 9.816354005237583e-06,
"loss": 1.4568,
"step": 37
},
{
"epoch": 0.2360248447204969,
"grad_norm": 0.4499090909957886,
"learning_rate": 9.802590558156863e-06,
"loss": 1.4483,
"step": 38
},
{
"epoch": 0.2422360248447205,
"grad_norm": 0.44764548540115356,
"learning_rate": 9.78834018565262e-06,
"loss": 1.4769,
"step": 39
},
{
"epoch": 0.2484472049689441,
"grad_norm": 0.4358612298965454,
"learning_rate": 9.77360433254273e-06,
"loss": 1.5545,
"step": 40
},
{
"epoch": 0.2546583850931677,
"grad_norm": 0.44274434447288513,
"learning_rate": 9.758384492867004e-06,
"loss": 1.3615,
"step": 41
},
{
"epoch": 0.2546583850931677,
"eval_loss": 1.3996479511260986,
"eval_runtime": 27.0472,
"eval_samples_per_second": 2.921,
"eval_steps_per_second": 0.37,
"step": 41
},
{
"epoch": 0.2608695652173913,
"grad_norm": 0.42516258358955383,
"learning_rate": 9.742682209735727e-06,
"loss": 1.4802,
"step": 42
},
{
"epoch": 0.2670807453416149,
"grad_norm": 0.40934988856315613,
"learning_rate": 9.726499075173201e-06,
"loss": 1.3283,
"step": 43
},
{
"epoch": 0.2732919254658385,
"grad_norm": 0.40728411078453064,
"learning_rate": 9.709836729956326e-06,
"loss": 1.4724,
"step": 44
},
{
"epoch": 0.2795031055900621,
"grad_norm": 0.405512273311615,
"learning_rate": 9.692696863448246e-06,
"loss": 1.3943,
"step": 45
},
{
"epoch": 0.2857142857142857,
"grad_norm": 0.4585839807987213,
"learning_rate": 9.675081213427076e-06,
"loss": 1.4038,
"step": 46
},
{
"epoch": 0.2919254658385093,
"grad_norm": 0.40490880608558655,
"learning_rate": 9.656991565909703e-06,
"loss": 1.3807,
"step": 47
},
{
"epoch": 0.2981366459627329,
"grad_norm": 0.4017808735370636,
"learning_rate": 9.638429754970715e-06,
"loss": 1.3532,
"step": 48
},
{
"epoch": 0.30434782608695654,
"grad_norm": 0.41484227776527405,
"learning_rate": 9.619397662556434e-06,
"loss": 1.3379,
"step": 49
},
{
"epoch": 0.3105590062111801,
"grad_norm": 0.36943531036376953,
"learning_rate": 9.599897218294122e-06,
"loss": 1.393,
"step": 50
},
{
"epoch": 0.3167701863354037,
"grad_norm": 0.39943498373031616,
"learning_rate": 9.57993039929633e-06,
"loss": 1.3039,
"step": 51
},
{
"epoch": 0.32298136645962733,
"grad_norm": 0.374943345785141,
"learning_rate": 9.55949922996045e-06,
"loss": 1.355,
"step": 52
},
{
"epoch": 0.32919254658385094,
"grad_norm": 0.4175715446472168,
"learning_rate": 9.538605781763464e-06,
"loss": 1.3716,
"step": 53
},
{
"epoch": 0.33540372670807456,
"grad_norm": 0.37441951036453247,
"learning_rate": 9.517252173051912e-06,
"loss": 1.3081,
"step": 54
},
{
"epoch": 0.3416149068322981,
"grad_norm": 0.39863765239715576,
"learning_rate": 9.49544056882713e-06,
"loss": 1.3942,
"step": 55
},
{
"epoch": 0.34782608695652173,
"grad_norm": 0.3854600787162781,
"learning_rate": 9.473173180525737e-06,
"loss": 1.2902,
"step": 56
},
{
"epoch": 0.35403726708074534,
"grad_norm": 0.42524421215057373,
"learning_rate": 9.450452265795423e-06,
"loss": 1.2544,
"step": 57
},
{
"epoch": 0.36024844720496896,
"grad_norm": 0.41367658972740173,
"learning_rate": 9.427280128266049e-06,
"loss": 1.3194,
"step": 58
},
{
"epoch": 0.36645962732919257,
"grad_norm": 0.3787434697151184,
"learning_rate": 9.403659117316093e-06,
"loss": 1.3077,
"step": 59
},
{
"epoch": 0.37267080745341613,
"grad_norm": 0.4197174608707428,
"learning_rate": 9.37959162783444e-06,
"loss": 1.2356,
"step": 60
},
{
"epoch": 0.37888198757763975,
"grad_norm": 0.39209407567977905,
"learning_rate": 9.355080099977579e-06,
"loss": 1.2168,
"step": 61
},
{
"epoch": 0.38509316770186336,
"grad_norm": 0.41228196024894714,
"learning_rate": 9.330127018922195e-06,
"loss": 1.2237,
"step": 62
},
{
"epoch": 0.391304347826087,
"grad_norm": 0.4241032302379608,
"learning_rate": 9.3047349146132e-06,
"loss": 1.3146,
"step": 63
},
{
"epoch": 0.39751552795031053,
"grad_norm": 0.4294179677963257,
"learning_rate": 9.278906361507238e-06,
"loss": 1.2716,
"step": 64
},
{
"epoch": 0.40372670807453415,
"grad_norm": 0.4411190450191498,
"learning_rate": 9.252643978311649e-06,
"loss": 1.2735,
"step": 65
},
{
"epoch": 0.40993788819875776,
"grad_norm": 0.46513664722442627,
"learning_rate": 9.225950427718974e-06,
"loss": 1.2337,
"step": 66
},
{
"epoch": 0.4161490683229814,
"grad_norm": 0.36713191866874695,
"learning_rate": 9.198828416136991e-06,
"loss": 1.3339,
"step": 67
},
{
"epoch": 0.422360248447205,
"grad_norm": 0.351909875869751,
"learning_rate": 9.171280693414307e-06,
"loss": 1.358,
"step": 68
},
{
"epoch": 0.42857142857142855,
"grad_norm": 0.3643895983695984,
"learning_rate": 9.14331005256157e-06,
"loss": 1.2482,
"step": 69
},
{
"epoch": 0.43478260869565216,
"grad_norm": 0.4360922873020172,
"learning_rate": 9.114919329468283e-06,
"loss": 1.3115,
"step": 70
},
{
"epoch": 0.4409937888198758,
"grad_norm": 0.4254709482192993,
"learning_rate": 9.086111402615274e-06,
"loss": 1.2351,
"step": 71
},
{
"epoch": 0.4472049689440994,
"grad_norm": 0.3786623179912567,
"learning_rate": 9.056889192782865e-06,
"loss": 1.3746,
"step": 72
},
{
"epoch": 0.453416149068323,
"grad_norm": 0.4647882878780365,
"learning_rate": 9.02725566275473e-06,
"loss": 1.2395,
"step": 73
},
{
"epoch": 0.45962732919254656,
"grad_norm": 0.4040616452693939,
"learning_rate": 8.997213817017508e-06,
"loss": 1.2998,
"step": 74
},
{
"epoch": 0.4658385093167702,
"grad_norm": 0.41700279712677,
"learning_rate": 8.966766701456177e-06,
"loss": 1.2105,
"step": 75
},
{
"epoch": 0.4720496894409938,
"grad_norm": 0.3908318877220154,
"learning_rate": 8.935917403045251e-06,
"loss": 1.2011,
"step": 76
},
{
"epoch": 0.4782608695652174,
"grad_norm": 0.38885021209716797,
"learning_rate": 8.90466904953579e-06,
"loss": 1.2471,
"step": 77
},
{
"epoch": 0.484472049689441,
"grad_norm": 0.42335832118988037,
"learning_rate": 8.873024809138272e-06,
"loss": 1.2683,
"step": 78
},
{
"epoch": 0.4906832298136646,
"grad_norm": 0.39654576778411865,
"learning_rate": 8.840987890201404e-06,
"loss": 1.2587,
"step": 79
},
{
"epoch": 0.4968944099378882,
"grad_norm": 0.4123787581920624,
"learning_rate": 8.808561540886796e-06,
"loss": 1.1794,
"step": 80
},
{
"epoch": 0.5031055900621118,
"grad_norm": 0.42227044701576233,
"learning_rate": 8.77574904883967e-06,
"loss": 1.2339,
"step": 81
},
{
"epoch": 0.5093167701863354,
"grad_norm": 0.4471132159233093,
"learning_rate": 8.742553740855507e-06,
"loss": 1.1954,
"step": 82
},
{
"epoch": 0.5093167701863354,
"eval_loss": 1.2136852741241455,
"eval_runtime": 26.7767,
"eval_samples_per_second": 2.95,
"eval_steps_per_second": 0.373,
"step": 82
},
{
"epoch": 0.515527950310559,
"grad_norm": 0.42590177059173584,
"learning_rate": 8.708978982542766e-06,
"loss": 1.2426,
"step": 83
},
{
"epoch": 0.5217391304347826,
"grad_norm": 0.39879706501960754,
"learning_rate": 8.675028177981643e-06,
"loss": 1.2618,
"step": 84
},
{
"epoch": 0.5279503105590062,
"grad_norm": 0.4137211740016937,
"learning_rate": 8.640704769378943e-06,
"loss": 1.1735,
"step": 85
},
{
"epoch": 0.5341614906832298,
"grad_norm": 0.42463183403015137,
"learning_rate": 8.606012236719073e-06,
"loss": 1.1433,
"step": 86
},
{
"epoch": 0.5403726708074534,
"grad_norm": 0.42853713035583496,
"learning_rate": 8.570954097411224e-06,
"loss": 1.2307,
"step": 87
},
{
"epoch": 0.546583850931677,
"grad_norm": 0.408589243888855,
"learning_rate": 8.535533905932739e-06,
"loss": 1.1523,
"step": 88
},
{
"epoch": 0.5527950310559007,
"grad_norm": 0.42884618043899536,
"learning_rate": 8.499755253468732e-06,
"loss": 1.2494,
"step": 89
},
{
"epoch": 0.5590062111801242,
"grad_norm": 0.4813741445541382,
"learning_rate": 8.463621767547998e-06,
"loss": 1.2417,
"step": 90
},
{
"epoch": 0.5652173913043478,
"grad_norm": 0.4021080732345581,
"learning_rate": 8.4271371116752e-06,
"loss": 1.2715,
"step": 91
},
{
"epoch": 0.5714285714285714,
"grad_norm": 0.3832411468029022,
"learning_rate": 8.390304984959455e-06,
"loss": 1.1586,
"step": 92
},
{
"epoch": 0.577639751552795,
"grad_norm": 0.4360566735267639,
"learning_rate": 8.35312912173928e-06,
"loss": 1.217,
"step": 93
},
{
"epoch": 0.5838509316770186,
"grad_norm": 0.44515809416770935,
"learning_rate": 8.315613291203977e-06,
"loss": 1.264,
"step": 94
},
{
"epoch": 0.5900621118012422,
"grad_norm": 0.41051459312438965,
"learning_rate": 8.277761297011475e-06,
"loss": 1.1578,
"step": 95
},
{
"epoch": 0.5962732919254659,
"grad_norm": 0.45815277099609375,
"learning_rate": 8.239576976902694e-06,
"loss": 1.2084,
"step": 96
},
{
"epoch": 0.6024844720496895,
"grad_norm": 0.4373926520347595,
"learning_rate": 8.20106420231244e-06,
"loss": 1.1925,
"step": 97
},
{
"epoch": 0.6086956521739131,
"grad_norm": 0.444323867559433,
"learning_rate": 8.162226877976886e-06,
"loss": 1.2183,
"step": 98
},
{
"epoch": 0.6149068322981367,
"grad_norm": 0.4503950774669647,
"learning_rate": 8.123068941537681e-06,
"loss": 1.1156,
"step": 99
},
{
"epoch": 0.6211180124223602,
"grad_norm": 0.4149700105190277,
"learning_rate": 8.083594363142717e-06,
"loss": 1.2379,
"step": 100
},
{
"epoch": 0.6273291925465838,
"grad_norm": 0.472960501909256,
"learning_rate": 8.043807145043604e-06,
"loss": 1.14,
"step": 101
},
{
"epoch": 0.6335403726708074,
"grad_norm": 0.40744665265083313,
"learning_rate": 8.003711321189895e-06,
"loss": 1.2446,
"step": 102
},
{
"epoch": 0.639751552795031,
"grad_norm": 0.3995897173881531,
"learning_rate": 7.963310956820085e-06,
"loss": 1.179,
"step": 103
},
{
"epoch": 0.6459627329192547,
"grad_norm": 0.4376934766769409,
"learning_rate": 7.922610148049445e-06,
"loss": 1.2818,
"step": 104
},
{
"epoch": 0.6521739130434783,
"grad_norm": 0.4294185936450958,
"learning_rate": 7.881613021454728e-06,
"loss": 1.1463,
"step": 105
},
{
"epoch": 0.6583850931677019,
"grad_norm": 0.46748086810112,
"learning_rate": 7.84032373365578e-06,
"loss": 1.1702,
"step": 106
},
{
"epoch": 0.6645962732919255,
"grad_norm": 0.47421255707740784,
"learning_rate": 7.798746470894113e-06,
"loss": 1.2922,
"step": 107
},
{
"epoch": 0.6708074534161491,
"grad_norm": 0.37985584139823914,
"learning_rate": 7.75688544860846e-06,
"loss": 1.1581,
"step": 108
},
{
"epoch": 0.6770186335403726,
"grad_norm": 0.4336451590061188,
"learning_rate": 7.714744911007395e-06,
"loss": 1.1532,
"step": 109
},
{
"epoch": 0.6832298136645962,
"grad_norm": 0.4210762679576874,
"learning_rate": 7.672329130639007e-06,
"loss": 1.2047,
"step": 110
},
{
"epoch": 0.6894409937888198,
"grad_norm": 0.42413201928138733,
"learning_rate": 7.62964240795772e-06,
"loss": 1.125,
"step": 111
},
{
"epoch": 0.6956521739130435,
"grad_norm": 0.47521328926086426,
"learning_rate": 7.586689070888284e-06,
"loss": 1.2143,
"step": 112
},
{
"epoch": 0.7018633540372671,
"grad_norm": 0.41499438881874084,
"learning_rate": 7.543473474386962e-06,
"loss": 1.07,
"step": 113
},
{
"epoch": 0.7080745341614907,
"grad_norm": 0.4027252495288849,
"learning_rate": 7.500000000000001e-06,
"loss": 1.262,
"step": 114
},
{
"epoch": 0.7142857142857143,
"grad_norm": 0.4009539484977722,
"learning_rate": 7.4562730554193875e-06,
"loss": 1.1742,
"step": 115
},
{
"epoch": 0.7204968944099379,
"grad_norm": 0.46428173780441284,
"learning_rate": 7.412297074035968e-06,
"loss": 1.1629,
"step": 116
},
{
"epoch": 0.7267080745341615,
"grad_norm": 0.4374319016933441,
"learning_rate": 7.368076514489947e-06,
"loss": 1.1065,
"step": 117
},
{
"epoch": 0.7329192546583851,
"grad_norm": 0.42879152297973633,
"learning_rate": 7.323615860218844e-06,
"loss": 1.193,
"step": 118
},
{
"epoch": 0.7391304347826086,
"grad_norm": 0.4889649748802185,
"learning_rate": 7.2789196190029155e-06,
"loss": 1.2618,
"step": 119
},
{
"epoch": 0.7453416149068323,
"grad_norm": 0.4490755796432495,
"learning_rate": 7.2339923225081296e-06,
"loss": 1.1857,
"step": 120
},
{
"epoch": 0.7515527950310559,
"grad_norm": 0.4516308009624481,
"learning_rate": 7.188838525826702e-06,
"loss": 1.1076,
"step": 121
},
{
"epoch": 0.7577639751552795,
"grad_norm": 0.5366808176040649,
"learning_rate": 7.143462807015271e-06,
"loss": 1.1523,
"step": 122
},
{
"epoch": 0.7639751552795031,
"grad_norm": 0.4146997332572937,
"learning_rate": 7.09786976663073e-06,
"loss": 1.1766,
"step": 123
},
{
"epoch": 0.7639751552795031,
"eval_loss": 1.1551172733306885,
"eval_runtime": 26.7368,
"eval_samples_per_second": 2.955,
"eval_steps_per_second": 0.374,
"step": 123
},
{
"epoch": 0.7701863354037267,
"grad_norm": 0.4490116536617279,
"learning_rate": 7.052064027263785e-06,
"loss": 1.1971,
"step": 124
},
{
"epoch": 0.7763975155279503,
"grad_norm": 0.48434358835220337,
"learning_rate": 7.006050233070289e-06,
"loss": 1.239,
"step": 125
},
{
"epoch": 0.782608695652174,
"grad_norm": 0.4746357798576355,
"learning_rate": 6.959833049300376e-06,
"loss": 1.1586,
"step": 126
},
{
"epoch": 0.7888198757763976,
"grad_norm": 0.39503100514411926,
"learning_rate": 6.913417161825449e-06,
"loss": 1.0894,
"step": 127
},
{
"epoch": 0.7950310559006211,
"grad_norm": 0.43696898221969604,
"learning_rate": 6.8668072766631054e-06,
"loss": 1.0384,
"step": 128
},
{
"epoch": 0.8012422360248447,
"grad_norm": 0.4739917814731598,
"learning_rate": 6.820008119499992e-06,
"loss": 1.1741,
"step": 129
},
{
"epoch": 0.8074534161490683,
"grad_norm": 0.4144044518470764,
"learning_rate": 6.773024435212678e-06,
"loss": 1.1911,
"step": 130
},
{
"epoch": 0.8136645962732919,
"grad_norm": 0.4699600636959076,
"learning_rate": 6.7258609873865895e-06,
"loss": 1.2041,
"step": 131
},
{
"epoch": 0.8198757763975155,
"grad_norm": 0.40221303701400757,
"learning_rate": 6.678522557833025e-06,
"loss": 1.1077,
"step": 132
},
{
"epoch": 0.8260869565217391,
"grad_norm": 0.4720539152622223,
"learning_rate": 6.631013946104348e-06,
"loss": 1.0987,
"step": 133
},
{
"epoch": 0.8322981366459627,
"grad_norm": 0.4860423505306244,
"learning_rate": 6.583339969007364e-06,
"loss": 1.1455,
"step": 134
},
{
"epoch": 0.8385093167701864,
"grad_norm": 0.5083861947059631,
"learning_rate": 6.5355054601149545e-06,
"loss": 1.1314,
"step": 135
},
{
"epoch": 0.84472049689441,
"grad_norm": 0.5092173218727112,
"learning_rate": 6.487515269276015e-06,
"loss": 1.2228,
"step": 136
},
{
"epoch": 0.8509316770186336,
"grad_norm": 0.4544651508331299,
"learning_rate": 6.439374262123731e-06,
"loss": 1.1177,
"step": 137
},
{
"epoch": 0.8571428571428571,
"grad_norm": 0.45371013879776,
"learning_rate": 6.391087319582264e-06,
"loss": 1.1782,
"step": 138
},
{
"epoch": 0.8633540372670807,
"grad_norm": 0.5025550723075867,
"learning_rate": 6.342659337371884e-06,
"loss": 1.1594,
"step": 139
},
{
"epoch": 0.8695652173913043,
"grad_norm": 0.48699814081192017,
"learning_rate": 6.294095225512604e-06,
"loss": 1.1246,
"step": 140
},
{
"epoch": 0.8757763975155279,
"grad_norm": 0.442331999540329,
"learning_rate": 6.2453999078263596e-06,
"loss": 1.1265,
"step": 141
},
{
"epoch": 0.8819875776397516,
"grad_norm": 0.4538658857345581,
"learning_rate": 6.1965783214377895e-06,
"loss": 1.2054,
"step": 142
},
{
"epoch": 0.8881987577639752,
"grad_norm": 0.4796921908855438,
"learning_rate": 6.147635416273679e-06,
"loss": 1.1621,
"step": 143
},
{
"epoch": 0.8944099378881988,
"grad_norm": 0.4758358895778656,
"learning_rate": 6.0985761545610865e-06,
"loss": 1.117,
"step": 144
},
{
"epoch": 0.9006211180124224,
"grad_norm": 0.44177138805389404,
"learning_rate": 6.049405510324237e-06,
"loss": 1.1016,
"step": 145
},
{
"epoch": 0.906832298136646,
"grad_norm": 0.48750707507133484,
"learning_rate": 6.000128468880223e-06,
"loss": 1.2238,
"step": 146
},
{
"epoch": 0.9130434782608695,
"grad_norm": 0.53321373462677,
"learning_rate": 5.950750026333534e-06,
"loss": 1.0768,
"step": 147
},
{
"epoch": 0.9192546583850931,
"grad_norm": 0.4829699993133545,
"learning_rate": 5.90127518906953e-06,
"loss": 1.1229,
"step": 148
},
{
"epoch": 0.9254658385093167,
"grad_norm": 0.5378725528717041,
"learning_rate": 5.851708973246841e-06,
"loss": 1.076,
"step": 149
},
{
"epoch": 0.9316770186335404,
"grad_norm": 0.44234123826026917,
"learning_rate": 5.8020564042888015e-06,
"loss": 1.1543,
"step": 150
},
{
"epoch": 0.937888198757764,
"grad_norm": 0.41885337233543396,
"learning_rate": 5.752322516373916e-06,
"loss": 1.1918,
"step": 151
},
{
"epoch": 0.9440993788819876,
"grad_norm": 0.4357737600803375,
"learning_rate": 5.7025123519254644e-06,
"loss": 1.1822,
"step": 152
},
{
"epoch": 0.9503105590062112,
"grad_norm": 0.48525384068489075,
"learning_rate": 5.65263096110026e-06,
"loss": 1.1569,
"step": 153
},
{
"epoch": 0.9565217391304348,
"grad_norm": 0.4788309335708618,
"learning_rate": 5.6026834012766155e-06,
"loss": 1.1826,
"step": 154
},
{
"epoch": 0.9627329192546584,
"grad_norm": 0.4695623815059662,
"learning_rate": 5.5526747365416e-06,
"loss": 1.196,
"step": 155
},
{
"epoch": 0.968944099378882,
"grad_norm": 0.38793641328811646,
"learning_rate": 5.502610037177586e-06,
"loss": 1.0959,
"step": 156
},
{
"epoch": 0.9751552795031055,
"grad_norm": 0.4816080629825592,
"learning_rate": 5.45249437914819e-06,
"loss": 1.1359,
"step": 157
},
{
"epoch": 0.9813664596273292,
"grad_norm": 0.4505300223827362,
"learning_rate": 5.402332843583631e-06,
"loss": 1.0793,
"step": 158
},
{
"epoch": 0.9875776397515528,
"grad_norm": 0.5171789526939392,
"learning_rate": 5.35213051626556e-06,
"loss": 1.1582,
"step": 159
},
{
"epoch": 0.9937888198757764,
"grad_norm": 0.4746135473251343,
"learning_rate": 5.301892487111431e-06,
"loss": 1.13,
"step": 160
},
{
"epoch": 1.0,
"grad_norm": 0.5038459897041321,
"learning_rate": 5.251623849658434e-06,
"loss": 1.2033,
"step": 161
},
{
"epoch": 1.0062111801242235,
"grad_norm": 0.4865545332431793,
"learning_rate": 5.201329700547077e-06,
"loss": 1.2215,
"step": 162
},
{
"epoch": 1.0124223602484472,
"grad_norm": 0.44650915265083313,
"learning_rate": 5.151015139004445e-06,
"loss": 1.1592,
"step": 163
},
{
"epoch": 1.0186335403726707,
"grad_norm": 0.4772714376449585,
"learning_rate": 5.100685266327202e-06,
"loss": 1.1361,
"step": 164
},
{
"epoch": 1.0186335403726707,
"eval_loss": 1.1316967010498047,
"eval_runtime": 26.8815,
"eval_samples_per_second": 2.939,
"eval_steps_per_second": 0.372,
"step": 164
},
{
"epoch": 1.0248447204968945,
"grad_norm": 0.4834270179271698,
"learning_rate": 5.050345185364378e-06,
"loss": 1.1859,
"step": 165
},
{
"epoch": 1.0062111801242235,
"grad_norm": 0.48713916540145874,
"learning_rate": 5e-06,
"loss": 1.0621,
"step": 166
},
{
"epoch": 1.0124223602484472,
"grad_norm": 0.4767020642757416,
"learning_rate": 4.949654814635623e-06,
"loss": 1.0793,
"step": 167
},
{
"epoch": 1.0186335403726707,
"grad_norm": 0.4696958065032959,
"learning_rate": 4.8993147336728e-06,
"loss": 1.2284,
"step": 168
},
{
"epoch": 1.0248447204968945,
"grad_norm": 0.47341403365135193,
"learning_rate": 4.848984860995557e-06,
"loss": 1.1235,
"step": 169
},
{
"epoch": 1.031055900621118,
"grad_norm": 0.5151461958885193,
"learning_rate": 4.798670299452926e-06,
"loss": 1.1104,
"step": 170
},
{
"epoch": 1.0372670807453417,
"grad_norm": 0.4465959668159485,
"learning_rate": 4.748376150341566e-06,
"loss": 1.106,
"step": 171
},
{
"epoch": 1.0434782608695652,
"grad_norm": 0.4610389769077301,
"learning_rate": 4.69810751288857e-06,
"loss": 1.2657,
"step": 172
},
{
"epoch": 1.049689440993789,
"grad_norm": 0.4607885479927063,
"learning_rate": 4.647869483734441e-06,
"loss": 1.1618,
"step": 173
},
{
"epoch": 1.0559006211180124,
"grad_norm": 0.4866524934768677,
"learning_rate": 4.597667156416371e-06,
"loss": 1.1848,
"step": 174
},
{
"epoch": 1.062111801242236,
"grad_norm": 0.4338420033454895,
"learning_rate": 4.547505620851812e-06,
"loss": 1.1223,
"step": 175
},
{
"epoch": 1.0683229813664596,
"grad_norm": 0.5134140849113464,
"learning_rate": 4.497389962822416e-06,
"loss": 1.1493,
"step": 176
},
{
"epoch": 1.0745341614906831,
"grad_norm": 0.5529913902282715,
"learning_rate": 4.447325263458401e-06,
"loss": 1.1222,
"step": 177
},
{
"epoch": 1.0807453416149069,
"grad_norm": 0.48266178369522095,
"learning_rate": 4.397316598723385e-06,
"loss": 1.2108,
"step": 178
},
{
"epoch": 1.0869565217391304,
"grad_norm": 0.42973291873931885,
"learning_rate": 4.347369038899744e-06,
"loss": 1.1305,
"step": 179
},
{
"epoch": 1.093167701863354,
"grad_norm": 0.4676703214645386,
"learning_rate": 4.297487648074538e-06,
"loss": 1.2059,
"step": 180
},
{
"epoch": 1.0993788819875776,
"grad_norm": 0.4928361773490906,
"learning_rate": 4.247677483626085e-06,
"loss": 1.1674,
"step": 181
},
{
"epoch": 1.1055900621118013,
"grad_norm": 0.526463508605957,
"learning_rate": 4.1979435957111984e-06,
"loss": 1.0983,
"step": 182
},
{
"epoch": 1.1118012422360248,
"grad_norm": 0.44327542185783386,
"learning_rate": 4.148291026753159e-06,
"loss": 1.1574,
"step": 183
},
{
"epoch": 1.1180124223602483,
"grad_norm": 0.4704626798629761,
"learning_rate": 4.098724810930472e-06,
"loss": 1.1953,
"step": 184
},
{
"epoch": 1.124223602484472,
"grad_norm": 0.43922385573387146,
"learning_rate": 4.049249973666468e-06,
"loss": 1.1376,
"step": 185
},
{
"epoch": 1.1304347826086956,
"grad_norm": 0.462429404258728,
"learning_rate": 3.999871531119779e-06,
"loss": 1.0687,
"step": 186
},
{
"epoch": 1.1366459627329193,
"grad_norm": 0.47191864252090454,
"learning_rate": 3.9505944896757635e-06,
"loss": 1.139,
"step": 187
},
{
"epoch": 1.1428571428571428,
"grad_norm": 0.46805888414382935,
"learning_rate": 3.901423845438916e-06,
"loss": 1.1135,
"step": 188
},
{
"epoch": 1.1490683229813665,
"grad_norm": 0.4984455704689026,
"learning_rate": 3.852364583726324e-06,
"loss": 1.0954,
"step": 189
},
{
"epoch": 1.15527950310559,
"grad_norm": 0.5003888010978699,
"learning_rate": 3.803421678562213e-06,
"loss": 1.1432,
"step": 190
},
{
"epoch": 1.1614906832298137,
"grad_norm": 0.5040602087974548,
"learning_rate": 3.7546000921736413e-06,
"loss": 1.1675,
"step": 191
},
{
"epoch": 1.1677018633540373,
"grad_norm": 0.45909303426742554,
"learning_rate": 3.705904774487396e-06,
"loss": 1.1865,
"step": 192
},
{
"epoch": 1.1739130434782608,
"grad_norm": 0.4990633428096771,
"learning_rate": 3.657340662628116e-06,
"loss": 1.1393,
"step": 193
},
{
"epoch": 1.1801242236024845,
"grad_norm": 0.4696108400821686,
"learning_rate": 3.6089126804177373e-06,
"loss": 1.1307,
"step": 194
},
{
"epoch": 1.186335403726708,
"grad_norm": 0.5014775395393372,
"learning_rate": 3.56062573787627e-06,
"loss": 1.137,
"step": 195
},
{
"epoch": 1.1925465838509317,
"grad_norm": 0.46463677287101746,
"learning_rate": 3.5124847307239863e-06,
"loss": 1.1561,
"step": 196
},
{
"epoch": 1.1987577639751552,
"grad_norm": 0.43704545497894287,
"learning_rate": 3.464494539885047e-06,
"loss": 1.1201,
"step": 197
},
{
"epoch": 1.204968944099379,
"grad_norm": 0.445740669965744,
"learning_rate": 3.416660030992639e-06,
"loss": 1.1516,
"step": 198
},
{
"epoch": 1.2111801242236024,
"grad_norm": 0.5268468856811523,
"learning_rate": 3.3689860538956547e-06,
"loss": 1.1288,
"step": 199
},
{
"epoch": 1.2173913043478262,
"grad_norm": 0.4934462606906891,
"learning_rate": 3.3214774421669777e-06,
"loss": 1.1642,
"step": 200
},
{
"epoch": 1.2236024844720497,
"grad_norm": 0.5016915202140808,
"learning_rate": 3.274139012613411e-06,
"loss": 1.0885,
"step": 201
},
{
"epoch": 1.2298136645962732,
"grad_norm": 0.4747258722782135,
"learning_rate": 3.226975564787322e-06,
"loss": 1.1381,
"step": 202
},
{
"epoch": 1.236024844720497,
"grad_norm": 0.522182285785675,
"learning_rate": 3.1799918805000097e-06,
"loss": 1.0668,
"step": 203
},
{
"epoch": 1.2422360248447206,
"grad_norm": 0.5038187503814697,
"learning_rate": 3.1331927233368954e-06,
"loss": 1.0818,
"step": 204
},
{
"epoch": 1.2484472049689441,
"grad_norm": 0.46231281757354736,
"learning_rate": 3.0865828381745515e-06,
"loss": 1.0431,
"step": 205
},
{
"epoch": 1.2484472049689441,
"eval_loss": 1.1210699081420898,
"eval_runtime": 26.9725,
"eval_samples_per_second": 2.929,
"eval_steps_per_second": 0.371,
"step": 205
},
{
"epoch": 1.2546583850931676,
"grad_norm": 0.5070775747299194,
"learning_rate": 3.040166950699626e-06,
"loss": 1.1643,
"step": 206
},
{
"epoch": 1.2608695652173914,
"grad_norm": 0.5011193752288818,
"learning_rate": 2.993949766929711e-06,
"loss": 1.0472,
"step": 207
},
{
"epoch": 1.2670807453416149,
"grad_norm": 0.4393374025821686,
"learning_rate": 2.947935972736217e-06,
"loss": 1.0875,
"step": 208
},
{
"epoch": 1.2732919254658386,
"grad_norm": 0.46959808468818665,
"learning_rate": 2.9021302333692738e-06,
"loss": 1.1078,
"step": 209
},
{
"epoch": 1.279503105590062,
"grad_norm": 0.4769798219203949,
"learning_rate": 2.8565371929847286e-06,
"loss": 1.16,
"step": 210
},
{
"epoch": 1.2857142857142856,
"grad_norm": 0.4200010895729065,
"learning_rate": 2.8111614741732975e-06,
"loss": 1.1258,
"step": 211
},
{
"epoch": 1.2919254658385093,
"grad_norm": 0.502759575843811,
"learning_rate": 2.766007677491871e-06,
"loss": 1.1263,
"step": 212
},
{
"epoch": 1.298136645962733,
"grad_norm": 0.48976272344589233,
"learning_rate": 2.7210803809970853e-06,
"loss": 1.174,
"step": 213
},
{
"epoch": 1.3043478260869565,
"grad_norm": 0.582251787185669,
"learning_rate": 2.6763841397811576e-06,
"loss": 1.1542,
"step": 214
},
{
"epoch": 1.31055900621118,
"grad_norm": 0.525684654712677,
"learning_rate": 2.631923485510054e-06,
"loss": 1.1594,
"step": 215
},
{
"epoch": 1.3167701863354038,
"grad_norm": 0.5200883150100708,
"learning_rate": 2.587702925964034e-06,
"loss": 1.1374,
"step": 216
},
{
"epoch": 1.3229813664596273,
"grad_norm": 0.4962819218635559,
"learning_rate": 2.5437269445806146e-06,
"loss": 1.0839,
"step": 217
},
{
"epoch": 1.329192546583851,
"grad_norm": 0.5223807692527771,
"learning_rate": 2.5000000000000015e-06,
"loss": 1.0708,
"step": 218
},
{
"epoch": 1.3354037267080745,
"grad_norm": 0.434766560792923,
"learning_rate": 2.4565265256130393e-06,
"loss": 1.1387,
"step": 219
},
{
"epoch": 1.341614906832298,
"grad_norm": 0.4618338346481323,
"learning_rate": 2.4133109291117156e-06,
"loss": 1.0728,
"step": 220
},
{
"epoch": 1.3478260869565217,
"grad_norm": 0.5078674554824829,
"learning_rate": 2.3703575920422793e-06,
"loss": 1.1649,
"step": 221
},
{
"epoch": 1.3540372670807455,
"grad_norm": 0.5160240530967712,
"learning_rate": 2.3276708693609947e-06,
"loss": 1.1226,
"step": 222
},
{
"epoch": 1.360248447204969,
"grad_norm": 0.44181424379348755,
"learning_rate": 2.2852550889926067e-06,
"loss": 1.0961,
"step": 223
},
{
"epoch": 1.3664596273291925,
"grad_norm": 0.4639110863208771,
"learning_rate": 2.243114551391542e-06,
"loss": 1.0937,
"step": 224
},
{
"epoch": 1.3726708074534162,
"grad_norm": 0.44535526633262634,
"learning_rate": 2.20125352910589e-06,
"loss": 1.1514,
"step": 225
},
{
"epoch": 1.3788819875776397,
"grad_norm": 0.463021844625473,
"learning_rate": 2.159676266344222e-06,
"loss": 1.1019,
"step": 226
},
{
"epoch": 1.3850931677018634,
"grad_norm": 0.494054913520813,
"learning_rate": 2.1183869785452744e-06,
"loss": 1.1175,
"step": 227
},
{
"epoch": 1.391304347826087,
"grad_norm": 0.5025504231452942,
"learning_rate": 2.077389851950557e-06,
"loss": 1.0575,
"step": 228
},
{
"epoch": 1.3975155279503104,
"grad_norm": 0.46806544065475464,
"learning_rate": 2.036689043179917e-06,
"loss": 1.1058,
"step": 229
},
{
"epoch": 1.4037267080745341,
"grad_norm": 0.4613620638847351,
"learning_rate": 1.996288678810105e-06,
"loss": 1.1508,
"step": 230
},
{
"epoch": 1.4099378881987579,
"grad_norm": 0.5427222847938538,
"learning_rate": 1.956192854956397e-06,
"loss": 1.2126,
"step": 231
},
{
"epoch": 1.4161490683229814,
"grad_norm": 0.4144895374774933,
"learning_rate": 1.9164056368572847e-06,
"loss": 1.0932,
"step": 232
},
{
"epoch": 1.4223602484472049,
"grad_norm": 0.47324222326278687,
"learning_rate": 1.87693105846232e-06,
"loss": 1.075,
"step": 233
},
{
"epoch": 1.4285714285714286,
"grad_norm": 0.4890088737010956,
"learning_rate": 1.8377731220231144e-06,
"loss": 1.1909,
"step": 234
},
{
"epoch": 1.434782608695652,
"grad_norm": 0.5050180554389954,
"learning_rate": 1.7989357976875603e-06,
"loss": 1.0406,
"step": 235
},
{
"epoch": 1.4409937888198758,
"grad_norm": 0.5166496634483337,
"learning_rate": 1.7604230230973068e-06,
"loss": 1.1579,
"step": 236
},
{
"epoch": 1.4472049689440993,
"grad_norm": 0.5063875913619995,
"learning_rate": 1.7222387029885268e-06,
"loss": 1.0532,
"step": 237
},
{
"epoch": 1.453416149068323,
"grad_norm": 0.43411558866500854,
"learning_rate": 1.6843867087960252e-06,
"loss": 1.1798,
"step": 238
},
{
"epoch": 1.4596273291925466,
"grad_norm": 0.5085105895996094,
"learning_rate": 1.6468708782607213e-06,
"loss": 1.0451,
"step": 239
},
{
"epoch": 1.4658385093167703,
"grad_norm": 0.5479459166526794,
"learning_rate": 1.6096950150405454e-06,
"loss": 1.1467,
"step": 240
},
{
"epoch": 1.4720496894409938,
"grad_norm": 0.5029129981994629,
"learning_rate": 1.572862888324801e-06,
"loss": 1.1512,
"step": 241
},
{
"epoch": 1.4782608695652173,
"grad_norm": 0.4606218934059143,
"learning_rate": 1.5363782324520033e-06,
"loss": 1.1006,
"step": 242
},
{
"epoch": 1.484472049689441,
"grad_norm": 0.4794401526451111,
"learning_rate": 1.5002447465312675e-06,
"loss": 1.0808,
"step": 243
},
{
"epoch": 1.4906832298136645,
"grad_norm": 0.49627459049224854,
"learning_rate": 1.4644660940672628e-06,
"loss": 1.1399,
"step": 244
},
{
"epoch": 1.4968944099378882,
"grad_norm": 0.5150241851806641,
"learning_rate": 1.4290459025887771e-06,
"loss": 1.1315,
"step": 245
},
{
"epoch": 1.5031055900621118,
"grad_norm": 0.5009984374046326,
"learning_rate": 1.3939877632809279e-06,
"loss": 1.1501,
"step": 246
},
{
"epoch": 1.5031055900621118,
"eval_loss": 1.1153790950775146,
"eval_runtime": 26.9095,
"eval_samples_per_second": 2.936,
"eval_steps_per_second": 0.372,
"step": 246
},
{
"epoch": 1.5093167701863353,
"grad_norm": 0.5049424767494202,
"learning_rate": 1.3592952306210589e-06,
"loss": 1.1243,
"step": 247
},
{
"epoch": 1.515527950310559,
"grad_norm": 0.5412471890449524,
"learning_rate": 1.3249718220183583e-06,
"loss": 1.1214,
"step": 248
},
{
"epoch": 1.5217391304347827,
"grad_norm": 0.5261203050613403,
"learning_rate": 1.2910210174572346e-06,
"loss": 1.0369,
"step": 249
},
{
"epoch": 1.5279503105590062,
"grad_norm": 0.47771474719047546,
"learning_rate": 1.257446259144494e-06,
"loss": 1.0598,
"step": 250
},
{
"epoch": 1.5341614906832297,
"grad_norm": 0.4789620339870453,
"learning_rate": 1.2242509511603318e-06,
"loss": 1.0653,
"step": 251
},
{
"epoch": 1.5403726708074534,
"grad_norm": 0.5012646317481995,
"learning_rate": 1.1914384591132045e-06,
"loss": 1.0909,
"step": 252
},
{
"epoch": 1.5465838509316772,
"grad_norm": 0.49552974104881287,
"learning_rate": 1.159012109798598e-06,
"loss": 1.1051,
"step": 253
},
{
"epoch": 1.5527950310559007,
"grad_norm": 0.48969194293022156,
"learning_rate": 1.1269751908617277e-06,
"loss": 1.2151,
"step": 254
},
{
"epoch": 1.5590062111801242,
"grad_norm": 0.5141896605491638,
"learning_rate": 1.095330950464213e-06,
"loss": 1.1452,
"step": 255
},
{
"epoch": 1.5652173913043477,
"grad_norm": 0.5273729562759399,
"learning_rate": 1.0640825969547498e-06,
"loss": 1.1325,
"step": 256
},
{
"epoch": 1.5714285714285714,
"grad_norm": 0.5029155015945435,
"learning_rate": 1.0332332985438248e-06,
"loss": 1.0992,
"step": 257
},
{
"epoch": 1.5776397515527951,
"grad_norm": 0.47117024660110474,
"learning_rate": 1.0027861829824953e-06,
"loss": 1.1461,
"step": 258
},
{
"epoch": 1.5838509316770186,
"grad_norm": 0.4335295855998993,
"learning_rate": 9.7274433724527e-07,
"loss": 1.0592,
"step": 259
},
{
"epoch": 1.5900621118012421,
"grad_norm": 0.49043720960617065,
"learning_rate": 9.431108072171346e-07,
"loss": 1.1516,
"step": 260
},
{
"epoch": 1.5962732919254659,
"grad_norm": 0.5176280736923218,
"learning_rate": 9.138885973847262e-07,
"loss": 1.168,
"step": 261
},
{
"epoch": 1.6024844720496896,
"grad_norm": 0.4628904461860657,
"learning_rate": 8.850806705317183e-07,
"loss": 1.0914,
"step": 262
},
{
"epoch": 1.608695652173913,
"grad_norm": 0.4468037784099579,
"learning_rate": 8.566899474384299e-07,
"loss": 1.0685,
"step": 263
},
{
"epoch": 1.6149068322981366,
"grad_norm": 0.5194827318191528,
"learning_rate": 8.287193065856936e-07,
"loss": 1.1787,
"step": 264
},
{
"epoch": 1.62111801242236,
"grad_norm": 0.48394885659217834,
"learning_rate": 8.011715838630107e-07,
"loss": 1.2196,
"step": 265
},
{
"epoch": 1.6273291925465838,
"grad_norm": 0.486788809299469,
"learning_rate": 7.740495722810271e-07,
"loss": 1.1354,
"step": 266
},
{
"epoch": 1.6335403726708075,
"grad_norm": 0.5033959150314331,
"learning_rate": 7.473560216883524e-07,
"loss": 1.13,
"step": 267
},
{
"epoch": 1.639751552795031,
"grad_norm": 0.4517529606819153,
"learning_rate": 7.210936384927631e-07,
"loss": 1.1115,
"step": 268
},
{
"epoch": 1.6459627329192545,
"grad_norm": 0.5163289904594421,
"learning_rate": 6.952650853867993e-07,
"loss": 1.1724,
"step": 269
},
{
"epoch": 1.6521739130434783,
"grad_norm": 0.5138719081878662,
"learning_rate": 6.698729810778065e-07,
"loss": 1.1498,
"step": 270
},
{
"epoch": 1.658385093167702,
"grad_norm": 0.5368619561195374,
"learning_rate": 6.449199000224221e-07,
"loss": 1.1271,
"step": 271
},
{
"epoch": 1.6645962732919255,
"grad_norm": 0.46004772186279297,
"learning_rate": 6.204083721655607e-07,
"loss": 1.0344,
"step": 272
},
{
"epoch": 1.670807453416149,
"grad_norm": 0.4616802930831909,
"learning_rate": 5.963408826839079e-07,
"loss": 1.1346,
"step": 273
},
{
"epoch": 1.6770186335403725,
"grad_norm": 0.514466404914856,
"learning_rate": 5.727198717339511e-07,
"loss": 1.1606,
"step": 274
},
{
"epoch": 1.6832298136645962,
"grad_norm": 0.4969620406627655,
"learning_rate": 5.49547734204578e-07,
"loss": 1.0281,
"step": 275
},
{
"epoch": 1.68944099378882,
"grad_norm": 0.5189189314842224,
"learning_rate": 5.268268194742638e-07,
"loss": 1.1554,
"step": 276
},
{
"epoch": 1.6956521739130435,
"grad_norm": 0.49855127930641174,
"learning_rate": 5.045594311728708e-07,
"loss": 1.1224,
"step": 277
},
{
"epoch": 1.701863354037267,
"grad_norm": 0.4984905421733856,
"learning_rate": 4.827478269480895e-07,
"loss": 1.0951,
"step": 278
},
{
"epoch": 1.7080745341614907,
"grad_norm": 0.5486929416656494,
"learning_rate": 4.6139421823653716e-07,
"loss": 1.1237,
"step": 279
},
{
"epoch": 1.7142857142857144,
"grad_norm": 0.4911029636859894,
"learning_rate": 4.405007700395497e-07,
"loss": 1.0601,
"step": 280
},
{
"epoch": 1.720496894409938,
"grad_norm": 0.4640960991382599,
"learning_rate": 4.200696007036703e-07,
"loss": 1.059,
"step": 281
},
{
"epoch": 1.7267080745341614,
"grad_norm": 0.48524951934814453,
"learning_rate": 4.001027817058789e-07,
"loss": 1.1238,
"step": 282
},
{
"epoch": 1.7329192546583851,
"grad_norm": 0.46812015771865845,
"learning_rate": 3.8060233744356634e-07,
"loss": 1.2098,
"step": 283
},
{
"epoch": 1.7391304347826086,
"grad_norm": 0.4788132309913635,
"learning_rate": 3.615702450292857e-07,
"loss": 1.1629,
"step": 284
},
{
"epoch": 1.7453416149068324,
"grad_norm": 0.47269925475120544,
"learning_rate": 3.430084340902973e-07,
"loss": 1.1628,
"step": 285
},
{
"epoch": 1.7515527950310559,
"grad_norm": 0.5449272394180298,
"learning_rate": 3.2491878657292643e-07,
"loss": 1.1587,
"step": 286
},
{
"epoch": 1.7577639751552794,
"grad_norm": 0.46690839529037476,
"learning_rate": 3.0730313655175647e-07,
"loss": 1.2015,
"step": 287
},
{
"epoch": 1.7577639751552794,
"eval_loss": 1.1139311790466309,
"eval_runtime": 27.0876,
"eval_samples_per_second": 2.916,
"eval_steps_per_second": 0.369,
"step": 287
},
{
"epoch": 1.763975155279503,
"grad_norm": 0.570745050907135,
"learning_rate": 2.901632700436757e-07,
"loss": 1.1747,
"step": 288
},
{
"epoch": 1.7701863354037268,
"grad_norm": 0.5287693738937378,
"learning_rate": 2.7350092482679836e-07,
"loss": 1.0633,
"step": 289
},
{
"epoch": 1.7763975155279503,
"grad_norm": 0.4709635376930237,
"learning_rate": 2.573177902642726e-07,
"loss": 1.1393,
"step": 290
},
{
"epoch": 1.7826086956521738,
"grad_norm": 0.49392980337142944,
"learning_rate": 2.416155071329973e-07,
"loss": 1.0933,
"step": 291
},
{
"epoch": 1.7888198757763976,
"grad_norm": 0.5224959850311279,
"learning_rate": 2.2639566745727203e-07,
"loss": 1.1316,
"step": 292
},
{
"epoch": 1.795031055900621,
"grad_norm": 0.4887482821941376,
"learning_rate": 2.1165981434738025e-07,
"loss": 1.1066,
"step": 293
},
{
"epoch": 1.8012422360248448,
"grad_norm": 0.4881725013256073,
"learning_rate": 1.9740944184313882e-07,
"loss": 1.032,
"step": 294
},
{
"epoch": 1.8074534161490683,
"grad_norm": 0.47561758756637573,
"learning_rate": 1.8364599476241862e-07,
"loss": 1.1154,
"step": 295
},
{
"epoch": 1.8136645962732918,
"grad_norm": 0.4982871413230896,
"learning_rate": 1.7037086855465902e-07,
"loss": 1.1942,
"step": 296
},
{
"epoch": 1.8198757763975155,
"grad_norm": 0.7219922542572021,
"learning_rate": 1.575854091593837e-07,
"loss": 1.0776,
"step": 297
},
{
"epoch": 1.8260869565217392,
"grad_norm": 0.5696266293525696,
"learning_rate": 1.4529091286973994e-07,
"loss": 1.1663,
"step": 298
},
{
"epoch": 1.8322981366459627,
"grad_norm": 0.45875245332717896,
"learning_rate": 1.3348862620107038e-07,
"loss": 1.1009,
"step": 299
},
{
"epoch": 1.8385093167701863,
"grad_norm": 0.4918197989463806,
"learning_rate": 1.2217974576453072e-07,
"loss": 1.1357,
"step": 300
},
{
"epoch": 1.84472049689441,
"grad_norm": 0.5131499171257019,
"learning_rate": 1.1136541814576574e-07,
"loss": 1.1376,
"step": 301
},
{
"epoch": 1.8509316770186337,
"grad_norm": 0.49699532985687256,
"learning_rate": 1.0104673978866164e-07,
"loss": 1.231,
"step": 302
},
{
"epoch": 1.8571428571428572,
"grad_norm": 0.5739893913269043,
"learning_rate": 9.122475688417954e-08,
"loss": 1.1645,
"step": 303
},
{
"epoch": 1.8633540372670807,
"grad_norm": 0.4959266781806946,
"learning_rate": 8.190046526428241e-08,
"loss": 1.1419,
"step": 304
},
{
"epoch": 1.8695652173913042,
"grad_norm": 0.5045937299728394,
"learning_rate": 7.307481030097152e-08,
"loss": 1.1859,
"step": 305
},
{
"epoch": 1.875776397515528,
"grad_norm": 0.47639361023902893,
"learning_rate": 6.474868681043578e-08,
"loss": 1.1026,
"step": 306
},
{
"epoch": 1.8819875776397517,
"grad_norm": 0.5153087973594666,
"learning_rate": 5.6922938962329364e-08,
"loss": 1.1612,
"step": 307
},
{
"epoch": 1.8881987577639752,
"grad_norm": 0.4941299855709076,
"learning_rate": 4.959836019417963e-08,
"loss": 1.0876,
"step": 308
},
{
"epoch": 1.8944099378881987,
"grad_norm": 0.4527961015701294,
"learning_rate": 4.2775693130948094e-08,
"loss": 1.1657,
"step": 309
},
{
"epoch": 1.9006211180124224,
"grad_norm": 0.49893486499786377,
"learning_rate": 3.645562950973014e-08,
"loss": 1.1261,
"step": 310
},
{
"epoch": 1.9068322981366461,
"grad_norm": 0.498981773853302,
"learning_rate": 3.063881010962611e-08,
"loss": 1.1179,
"step": 311
},
{
"epoch": 1.9130434782608696,
"grad_norm": 0.5112736225128174,
"learning_rate": 2.5325824686772138e-08,
"loss": 1.1863,
"step": 312
},
{
"epoch": 1.9192546583850931,
"grad_norm": 0.539107620716095,
"learning_rate": 2.0517211914545254e-08,
"loss": 1.0501,
"step": 313
},
{
"epoch": 1.9254658385093166,
"grad_norm": 0.514134407043457,
"learning_rate": 1.6213459328950355e-08,
"loss": 1.1264,
"step": 314
},
{
"epoch": 1.9316770186335404,
"grad_norm": 0.5193644165992737,
"learning_rate": 1.2415003279186988e-08,
"loss": 1.1484,
"step": 315
},
{
"epoch": 1.937888198757764,
"grad_norm": 0.4869459867477417,
"learning_rate": 9.12222888341252e-09,
"loss": 1.1454,
"step": 316
},
{
"epoch": 1.9440993788819876,
"grad_norm": 0.4767916798591614,
"learning_rate": 6.335469989692255e-09,
"loss": 1.2203,
"step": 317
},
{
"epoch": 1.950310559006211,
"grad_norm": 0.5365714430809021,
"learning_rate": 4.055009142152066e-09,
"loss": 1.252,
"step": 318
},
{
"epoch": 1.9565217391304348,
"grad_norm": 0.4672715961933136,
"learning_rate": 2.2810775523329775e-09,
"loss": 1.134,
"step": 319
},
{
"epoch": 1.9627329192546585,
"grad_norm": 0.45685145258903503,
"learning_rate": 1.0138550757493592e-09,
"loss": 1.1381,
"step": 320
},
{
"epoch": 1.968944099378882,
"grad_norm": 0.49931174516677856,
"learning_rate": 2.534701936512951e-10,
"loss": 1.1142,
"step": 321
},
{
"epoch": 1.9751552795031055,
"grad_norm": 0.4903966188430786,
"learning_rate": 0.0,
"loss": 1.0961,
"step": 322
}
],
"logging_steps": 1,
"max_steps": 322,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 81,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.7811992452614062e+18,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}