IeBoytsov's picture
Model save
fce79fe verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.0,
"eval_steps": 500,
"global_step": 4200,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0004761904761904762,
"grad_norm": 40.45813118239569,
"learning_rate": 4.7619047619047627e-08,
"loss": 2.1387,
"step": 1
},
{
"epoch": 0.002380952380952381,
"grad_norm": 39.63780477662549,
"learning_rate": 2.3809523809523811e-07,
"loss": 2.0871,
"step": 5
},
{
"epoch": 0.004761904761904762,
"grad_norm": 43.499844517947736,
"learning_rate": 4.7619047619047623e-07,
"loss": 2.0652,
"step": 10
},
{
"epoch": 0.007142857142857143,
"grad_norm": 10.896090871742048,
"learning_rate": 7.142857142857143e-07,
"loss": 1.9004,
"step": 15
},
{
"epoch": 0.009523809523809525,
"grad_norm": 6.2002952611415845,
"learning_rate": 9.523809523809525e-07,
"loss": 1.7859,
"step": 20
},
{
"epoch": 0.011904761904761904,
"grad_norm": 4.97841465233645,
"learning_rate": 1.1904761904761906e-06,
"loss": 1.6739,
"step": 25
},
{
"epoch": 0.014285714285714285,
"grad_norm": 4.293746378401411,
"learning_rate": 1.4285714285714286e-06,
"loss": 1.4907,
"step": 30
},
{
"epoch": 0.016666666666666666,
"grad_norm": 2.299524492927895,
"learning_rate": 1.6666666666666667e-06,
"loss": 1.3909,
"step": 35
},
{
"epoch": 0.01904761904761905,
"grad_norm": 2.3149107855519064,
"learning_rate": 1.904761904761905e-06,
"loss": 1.2509,
"step": 40
},
{
"epoch": 0.02142857142857143,
"grad_norm": 2.251875691331579,
"learning_rate": 2.1428571428571427e-06,
"loss": 1.1418,
"step": 45
},
{
"epoch": 0.023809523809523808,
"grad_norm": 1.293593518327437,
"learning_rate": 2.380952380952381e-06,
"loss": 1.0812,
"step": 50
},
{
"epoch": 0.02619047619047619,
"grad_norm": 1.1337840513174344,
"learning_rate": 2.6190476190476192e-06,
"loss": 1.0113,
"step": 55
},
{
"epoch": 0.02857142857142857,
"grad_norm": 1.1685369096623774,
"learning_rate": 2.8571428571428573e-06,
"loss": 0.9634,
"step": 60
},
{
"epoch": 0.030952380952380953,
"grad_norm": 1.070331563673874,
"learning_rate": 3.0952380952380957e-06,
"loss": 0.9291,
"step": 65
},
{
"epoch": 0.03333333333333333,
"grad_norm": 1.059586065170695,
"learning_rate": 3.3333333333333333e-06,
"loss": 0.91,
"step": 70
},
{
"epoch": 0.03571428571428571,
"grad_norm": 1.131151685372533,
"learning_rate": 3.5714285714285718e-06,
"loss": 0.8976,
"step": 75
},
{
"epoch": 0.0380952380952381,
"grad_norm": 1.193443615687326,
"learning_rate": 3.80952380952381e-06,
"loss": 0.8781,
"step": 80
},
{
"epoch": 0.04047619047619048,
"grad_norm": 1.0875818440706093,
"learning_rate": 4.047619047619048e-06,
"loss": 0.8631,
"step": 85
},
{
"epoch": 0.04285714285714286,
"grad_norm": 1.1503140744906168,
"learning_rate": 4.2857142857142855e-06,
"loss": 0.8646,
"step": 90
},
{
"epoch": 0.04523809523809524,
"grad_norm": 1.7469535134849759,
"learning_rate": 4.523809523809524e-06,
"loss": 0.8426,
"step": 95
},
{
"epoch": 0.047619047619047616,
"grad_norm": 1.299723199427286,
"learning_rate": 4.761904761904762e-06,
"loss": 0.8305,
"step": 100
},
{
"epoch": 0.05,
"grad_norm": 1.40840933352402,
"learning_rate": 5e-06,
"loss": 0.8316,
"step": 105
},
{
"epoch": 0.05238095238095238,
"grad_norm": 1.0810746876875799,
"learning_rate": 5.2380952380952384e-06,
"loss": 0.8225,
"step": 110
},
{
"epoch": 0.05476190476190476,
"grad_norm": 1.1168753096844857,
"learning_rate": 5.476190476190477e-06,
"loss": 0.8141,
"step": 115
},
{
"epoch": 0.05714285714285714,
"grad_norm": 1.117741427127342,
"learning_rate": 5.7142857142857145e-06,
"loss": 0.8128,
"step": 120
},
{
"epoch": 0.05952380952380952,
"grad_norm": 1.117824418032963,
"learning_rate": 5.9523809523809525e-06,
"loss": 0.8104,
"step": 125
},
{
"epoch": 0.06190476190476191,
"grad_norm": 1.1765989152315592,
"learning_rate": 6.1904761904761914e-06,
"loss": 0.8186,
"step": 130
},
{
"epoch": 0.06428571428571428,
"grad_norm": 1.14607344297389,
"learning_rate": 6.4285714285714295e-06,
"loss": 0.8094,
"step": 135
},
{
"epoch": 0.06666666666666667,
"grad_norm": 1.0689364794376695,
"learning_rate": 6.666666666666667e-06,
"loss": 0.7889,
"step": 140
},
{
"epoch": 0.06904761904761905,
"grad_norm": 1.0802475624680903,
"learning_rate": 6.9047619047619055e-06,
"loss": 0.8072,
"step": 145
},
{
"epoch": 0.07142857142857142,
"grad_norm": 1.1977014078487442,
"learning_rate": 7.1428571428571436e-06,
"loss": 0.7951,
"step": 150
},
{
"epoch": 0.07380952380952381,
"grad_norm": 1.0956749755790958,
"learning_rate": 7.380952380952382e-06,
"loss": 0.7961,
"step": 155
},
{
"epoch": 0.0761904761904762,
"grad_norm": 1.0269565395277456,
"learning_rate": 7.61904761904762e-06,
"loss": 0.7883,
"step": 160
},
{
"epoch": 0.07857142857142857,
"grad_norm": 1.2693412905779187,
"learning_rate": 7.857142857142858e-06,
"loss": 0.7941,
"step": 165
},
{
"epoch": 0.08095238095238096,
"grad_norm": 1.4270103378714147,
"learning_rate": 8.095238095238097e-06,
"loss": 0.7976,
"step": 170
},
{
"epoch": 0.08333333333333333,
"grad_norm": 1.3965101517157619,
"learning_rate": 8.333333333333334e-06,
"loss": 0.7907,
"step": 175
},
{
"epoch": 0.08571428571428572,
"grad_norm": 1.1737877049103937,
"learning_rate": 8.571428571428571e-06,
"loss": 0.7772,
"step": 180
},
{
"epoch": 0.0880952380952381,
"grad_norm": 0.9620328652517623,
"learning_rate": 8.80952380952381e-06,
"loss": 0.7895,
"step": 185
},
{
"epoch": 0.09047619047619047,
"grad_norm": 0.9954694445859312,
"learning_rate": 9.047619047619049e-06,
"loss": 0.7728,
"step": 190
},
{
"epoch": 0.09285714285714286,
"grad_norm": 0.9933982715674028,
"learning_rate": 9.285714285714288e-06,
"loss": 0.7784,
"step": 195
},
{
"epoch": 0.09523809523809523,
"grad_norm": 1.2380225457608798,
"learning_rate": 9.523809523809525e-06,
"loss": 0.762,
"step": 200
},
{
"epoch": 0.09761904761904762,
"grad_norm": 1.1645854540151788,
"learning_rate": 9.761904761904762e-06,
"loss": 0.7708,
"step": 205
},
{
"epoch": 0.1,
"grad_norm": 1.0347109854779557,
"learning_rate": 1e-05,
"loss": 0.7777,
"step": 210
},
{
"epoch": 0.10238095238095238,
"grad_norm": 1.007801398524834,
"learning_rate": 1.0238095238095238e-05,
"loss": 0.7696,
"step": 215
},
{
"epoch": 0.10476190476190476,
"grad_norm": 1.1084444980598738,
"learning_rate": 1.0476190476190477e-05,
"loss": 0.7667,
"step": 220
},
{
"epoch": 0.10714285714285714,
"grad_norm": 0.9232282482964133,
"learning_rate": 1.0714285714285714e-05,
"loss": 0.7682,
"step": 225
},
{
"epoch": 0.10952380952380952,
"grad_norm": 1.0266526312590731,
"learning_rate": 1.0952380952380955e-05,
"loss": 0.76,
"step": 230
},
{
"epoch": 0.11190476190476191,
"grad_norm": 0.9800538664923668,
"learning_rate": 1.1190476190476192e-05,
"loss": 0.7787,
"step": 235
},
{
"epoch": 0.11428571428571428,
"grad_norm": 1.0852968890165064,
"learning_rate": 1.1428571428571429e-05,
"loss": 0.7584,
"step": 240
},
{
"epoch": 0.11666666666666667,
"grad_norm": 1.101079132960328,
"learning_rate": 1.1666666666666668e-05,
"loss": 0.7532,
"step": 245
},
{
"epoch": 0.11904761904761904,
"grad_norm": 1.0434818420770975,
"learning_rate": 1.1904761904761905e-05,
"loss": 0.766,
"step": 250
},
{
"epoch": 0.12142857142857143,
"grad_norm": 0.8949846922479692,
"learning_rate": 1.2142857142857142e-05,
"loss": 0.764,
"step": 255
},
{
"epoch": 0.12380952380952381,
"grad_norm": 0.9144371936075102,
"learning_rate": 1.2380952380952383e-05,
"loss": 0.7518,
"step": 260
},
{
"epoch": 0.1261904761904762,
"grad_norm": 1.098930660048903,
"learning_rate": 1.261904761904762e-05,
"loss": 0.7549,
"step": 265
},
{
"epoch": 0.12857142857142856,
"grad_norm": 1.0213006883333104,
"learning_rate": 1.2857142857142859e-05,
"loss": 0.7514,
"step": 270
},
{
"epoch": 0.13095238095238096,
"grad_norm": 0.9116166496696912,
"learning_rate": 1.3095238095238096e-05,
"loss": 0.7515,
"step": 275
},
{
"epoch": 0.13333333333333333,
"grad_norm": 1.0474312553231273,
"learning_rate": 1.3333333333333333e-05,
"loss": 0.7663,
"step": 280
},
{
"epoch": 0.1357142857142857,
"grad_norm": 0.8710985201179252,
"learning_rate": 1.3571428571428574e-05,
"loss": 0.7608,
"step": 285
},
{
"epoch": 0.1380952380952381,
"grad_norm": 1.3401884340352725,
"learning_rate": 1.3809523809523811e-05,
"loss": 0.7554,
"step": 290
},
{
"epoch": 0.14047619047619048,
"grad_norm": 1.0740872828901855,
"learning_rate": 1.4047619047619048e-05,
"loss": 0.7582,
"step": 295
},
{
"epoch": 0.14285714285714285,
"grad_norm": 0.9126548421949322,
"learning_rate": 1.4285714285714287e-05,
"loss": 0.7434,
"step": 300
},
{
"epoch": 0.14523809523809525,
"grad_norm": 1.1120651258567102,
"learning_rate": 1.4523809523809524e-05,
"loss": 0.7388,
"step": 305
},
{
"epoch": 0.14761904761904762,
"grad_norm": 0.9950599559753186,
"learning_rate": 1.4761904761904763e-05,
"loss": 0.7585,
"step": 310
},
{
"epoch": 0.15,
"grad_norm": 0.8107934316232559,
"learning_rate": 1.5000000000000002e-05,
"loss": 0.7355,
"step": 315
},
{
"epoch": 0.1523809523809524,
"grad_norm": 0.808619303284316,
"learning_rate": 1.523809523809524e-05,
"loss": 0.7504,
"step": 320
},
{
"epoch": 0.15476190476190477,
"grad_norm": 0.8677410171473149,
"learning_rate": 1.5476190476190476e-05,
"loss": 0.7607,
"step": 325
},
{
"epoch": 0.15714285714285714,
"grad_norm": 0.8160199476236252,
"learning_rate": 1.5714285714285715e-05,
"loss": 0.7464,
"step": 330
},
{
"epoch": 0.1595238095238095,
"grad_norm": 0.8264103314451071,
"learning_rate": 1.5952380952380954e-05,
"loss": 0.7416,
"step": 335
},
{
"epoch": 0.1619047619047619,
"grad_norm": 0.9284321857373765,
"learning_rate": 1.6190476190476193e-05,
"loss": 0.7649,
"step": 340
},
{
"epoch": 0.16428571428571428,
"grad_norm": 0.9711426336534976,
"learning_rate": 1.642857142857143e-05,
"loss": 0.7388,
"step": 345
},
{
"epoch": 0.16666666666666666,
"grad_norm": 1.0029532403604327,
"learning_rate": 1.6666666666666667e-05,
"loss": 0.7495,
"step": 350
},
{
"epoch": 0.16904761904761906,
"grad_norm": 0.8655401816761288,
"learning_rate": 1.6904761904761906e-05,
"loss": 0.7431,
"step": 355
},
{
"epoch": 0.17142857142857143,
"grad_norm": 0.811823602075138,
"learning_rate": 1.7142857142857142e-05,
"loss": 0.7409,
"step": 360
},
{
"epoch": 0.1738095238095238,
"grad_norm": 0.7850380081345122,
"learning_rate": 1.7380952380952384e-05,
"loss": 0.7496,
"step": 365
},
{
"epoch": 0.1761904761904762,
"grad_norm": 0.8871490424351564,
"learning_rate": 1.761904761904762e-05,
"loss": 0.7502,
"step": 370
},
{
"epoch": 0.17857142857142858,
"grad_norm": 0.857041601013508,
"learning_rate": 1.785714285714286e-05,
"loss": 0.7344,
"step": 375
},
{
"epoch": 0.18095238095238095,
"grad_norm": 0.8798387422023826,
"learning_rate": 1.8095238095238097e-05,
"loss": 0.7451,
"step": 380
},
{
"epoch": 0.18333333333333332,
"grad_norm": 0.8208318602053298,
"learning_rate": 1.8333333333333333e-05,
"loss": 0.7519,
"step": 385
},
{
"epoch": 0.18571428571428572,
"grad_norm": 0.9816478125661463,
"learning_rate": 1.8571428571428575e-05,
"loss": 0.7438,
"step": 390
},
{
"epoch": 0.1880952380952381,
"grad_norm": 0.8587183984289388,
"learning_rate": 1.880952380952381e-05,
"loss": 0.74,
"step": 395
},
{
"epoch": 0.19047619047619047,
"grad_norm": 0.8766655919062369,
"learning_rate": 1.904761904761905e-05,
"loss": 0.7506,
"step": 400
},
{
"epoch": 0.19285714285714287,
"grad_norm": 0.8309437973338848,
"learning_rate": 1.928571428571429e-05,
"loss": 0.745,
"step": 405
},
{
"epoch": 0.19523809523809524,
"grad_norm": 0.8106184766065911,
"learning_rate": 1.9523809523809524e-05,
"loss": 0.7475,
"step": 410
},
{
"epoch": 0.1976190476190476,
"grad_norm": 0.809891965368119,
"learning_rate": 1.9761904761904763e-05,
"loss": 0.7335,
"step": 415
},
{
"epoch": 0.2,
"grad_norm": 0.7555036664771403,
"learning_rate": 2e-05,
"loss": 0.7424,
"step": 420
},
{
"epoch": 0.20238095238095238,
"grad_norm": 0.7650413836527804,
"learning_rate": 1.999991365731819e-05,
"loss": 0.7501,
"step": 425
},
{
"epoch": 0.20476190476190476,
"grad_norm": 0.7312208748548112,
"learning_rate": 1.999965463076377e-05,
"loss": 0.7334,
"step": 430
},
{
"epoch": 0.20714285714285716,
"grad_norm": 0.821378907440853,
"learning_rate": 1.999922292480975e-05,
"loss": 0.7398,
"step": 435
},
{
"epoch": 0.20952380952380953,
"grad_norm": 0.7961564796311429,
"learning_rate": 1.999861854691106e-05,
"loss": 0.7383,
"step": 440
},
{
"epoch": 0.2119047619047619,
"grad_norm": 0.864941128195607,
"learning_rate": 1.999784150750442e-05,
"loss": 0.7453,
"step": 445
},
{
"epoch": 0.21428571428571427,
"grad_norm": 0.9432973729407097,
"learning_rate": 1.9996891820008165e-05,
"loss": 0.7428,
"step": 450
},
{
"epoch": 0.21666666666666667,
"grad_norm": 0.744201840220319,
"learning_rate": 1.9995769500822007e-05,
"loss": 0.732,
"step": 455
},
{
"epoch": 0.21904761904761905,
"grad_norm": 0.8051356121636954,
"learning_rate": 1.999447456932676e-05,
"loss": 0.7452,
"step": 460
},
{
"epoch": 0.22142857142857142,
"grad_norm": 0.7446879893905328,
"learning_rate": 1.9993007047883988e-05,
"loss": 0.7406,
"step": 465
},
{
"epoch": 0.22380952380952382,
"grad_norm": 0.8200085500457187,
"learning_rate": 1.9991366961835643e-05,
"loss": 0.7342,
"step": 470
},
{
"epoch": 0.2261904761904762,
"grad_norm": 0.6879793430211231,
"learning_rate": 1.9989554339503612e-05,
"loss": 0.737,
"step": 475
},
{
"epoch": 0.22857142857142856,
"grad_norm": 0.6767209429381076,
"learning_rate": 1.9987569212189224e-05,
"loss": 0.7365,
"step": 480
},
{
"epoch": 0.23095238095238096,
"grad_norm": 0.7417349665434714,
"learning_rate": 1.9985411614172728e-05,
"loss": 0.7173,
"step": 485
},
{
"epoch": 0.23333333333333334,
"grad_norm": 0.7446106009527734,
"learning_rate": 1.9983081582712684e-05,
"loss": 0.7423,
"step": 490
},
{
"epoch": 0.2357142857142857,
"grad_norm": 0.7442452133783487,
"learning_rate": 1.9980579158045322e-05,
"loss": 0.7165,
"step": 495
},
{
"epoch": 0.23809523809523808,
"grad_norm": 0.7769424686532669,
"learning_rate": 1.997790438338385e-05,
"loss": 0.749,
"step": 500
},
{
"epoch": 0.24047619047619048,
"grad_norm": 0.8034839265796356,
"learning_rate": 1.997505730491772e-05,
"loss": 0.7385,
"step": 505
},
{
"epoch": 0.24285714285714285,
"grad_norm": 0.7676614166297908,
"learning_rate": 1.9972037971811802e-05,
"loss": 0.7383,
"step": 510
},
{
"epoch": 0.24523809523809523,
"grad_norm": 0.9359149520411593,
"learning_rate": 1.9968846436205566e-05,
"loss": 0.7498,
"step": 515
},
{
"epoch": 0.24761904761904763,
"grad_norm": 0.7783563085530937,
"learning_rate": 1.9965482753212154e-05,
"loss": 0.7466,
"step": 520
},
{
"epoch": 0.25,
"grad_norm": 0.7058199644697823,
"learning_rate": 1.9961946980917457e-05,
"loss": 0.7327,
"step": 525
},
{
"epoch": 0.2523809523809524,
"grad_norm": 0.6768276735212928,
"learning_rate": 1.995823918037908e-05,
"loss": 0.7396,
"step": 530
},
{
"epoch": 0.25476190476190474,
"grad_norm": 0.7530681690630258,
"learning_rate": 1.9954359415625313e-05,
"loss": 0.736,
"step": 535
},
{
"epoch": 0.2571428571428571,
"grad_norm": 0.790729948461572,
"learning_rate": 1.9950307753654016e-05,
"loss": 0.7367,
"step": 540
},
{
"epoch": 0.25952380952380955,
"grad_norm": 0.8098472943164579,
"learning_rate": 1.994608426443146e-05,
"loss": 0.7484,
"step": 545
},
{
"epoch": 0.2619047619047619,
"grad_norm": 0.6594040663378262,
"learning_rate": 1.994168902089112e-05,
"loss": 0.7256,
"step": 550
},
{
"epoch": 0.2642857142857143,
"grad_norm": 0.7061644477917397,
"learning_rate": 1.9937122098932428e-05,
"loss": 0.7293,
"step": 555
},
{
"epoch": 0.26666666666666666,
"grad_norm": 0.6552145622334516,
"learning_rate": 1.9932383577419432e-05,
"loss": 0.7235,
"step": 560
},
{
"epoch": 0.26904761904761904,
"grad_norm": 0.7850549934752342,
"learning_rate": 1.9927473538179467e-05,
"loss": 0.7242,
"step": 565
},
{
"epoch": 0.2714285714285714,
"grad_norm": 0.6946715346854551,
"learning_rate": 1.9922392066001724e-05,
"loss": 0.7352,
"step": 570
},
{
"epoch": 0.27380952380952384,
"grad_norm": 0.6701958153898289,
"learning_rate": 1.9917139248635788e-05,
"loss": 0.7358,
"step": 575
},
{
"epoch": 0.2761904761904762,
"grad_norm": 0.6957543084857959,
"learning_rate": 1.991171517679013e-05,
"loss": 0.7287,
"step": 580
},
{
"epoch": 0.2785714285714286,
"grad_norm": 0.7412462140647854,
"learning_rate": 1.9906119944130527e-05,
"loss": 0.7314,
"step": 585
},
{
"epoch": 0.28095238095238095,
"grad_norm": 0.7180123780939238,
"learning_rate": 1.9900353647278466e-05,
"loss": 0.7049,
"step": 590
},
{
"epoch": 0.2833333333333333,
"grad_norm": 0.6346311822511415,
"learning_rate": 1.9894416385809444e-05,
"loss": 0.7371,
"step": 595
},
{
"epoch": 0.2857142857142857,
"grad_norm": 0.9181856171310567,
"learning_rate": 1.9888308262251286e-05,
"loss": 0.7537,
"step": 600
},
{
"epoch": 0.28809523809523807,
"grad_norm": 0.7646478108344241,
"learning_rate": 1.9882029382082342e-05,
"loss": 0.724,
"step": 605
},
{
"epoch": 0.2904761904761905,
"grad_norm": 0.7332726331013449,
"learning_rate": 1.9875579853729677e-05,
"loss": 0.7273,
"step": 610
},
{
"epoch": 0.29285714285714287,
"grad_norm": 0.8337211626177949,
"learning_rate": 1.9868959788567213e-05,
"loss": 0.7359,
"step": 615
},
{
"epoch": 0.29523809523809524,
"grad_norm": 0.72291541850261,
"learning_rate": 1.9862169300913784e-05,
"loss": 0.7246,
"step": 620
},
{
"epoch": 0.2976190476190476,
"grad_norm": 0.6775385452101815,
"learning_rate": 1.9855208508031173e-05,
"loss": 0.719,
"step": 625
},
{
"epoch": 0.3,
"grad_norm": 0.6899659929159091,
"learning_rate": 1.9848077530122083e-05,
"loss": 0.7311,
"step": 630
},
{
"epoch": 0.30238095238095236,
"grad_norm": 0.6092353332947169,
"learning_rate": 1.9840776490328067e-05,
"loss": 0.7283,
"step": 635
},
{
"epoch": 0.3047619047619048,
"grad_norm": 0.6170832705569598,
"learning_rate": 1.9833305514727396e-05,
"loss": 0.7341,
"step": 640
},
{
"epoch": 0.30714285714285716,
"grad_norm": 0.6299433950672038,
"learning_rate": 1.9825664732332886e-05,
"loss": 0.7224,
"step": 645
},
{
"epoch": 0.30952380952380953,
"grad_norm": 0.620304145431242,
"learning_rate": 1.981785427508966e-05,
"loss": 0.7325,
"step": 650
},
{
"epoch": 0.3119047619047619,
"grad_norm": 0.6551690352072297,
"learning_rate": 1.9809874277872886e-05,
"loss": 0.7042,
"step": 655
},
{
"epoch": 0.3142857142857143,
"grad_norm": 0.6965123962977487,
"learning_rate": 1.9801724878485438e-05,
"loss": 0.7142,
"step": 660
},
{
"epoch": 0.31666666666666665,
"grad_norm": 0.6522958895192698,
"learning_rate": 1.9793406217655516e-05,
"loss": 0.7188,
"step": 665
},
{
"epoch": 0.319047619047619,
"grad_norm": 0.6710935693138489,
"learning_rate": 1.9784918439034216e-05,
"loss": 0.7127,
"step": 670
},
{
"epoch": 0.32142857142857145,
"grad_norm": 0.6394392122127832,
"learning_rate": 1.977626168919305e-05,
"loss": 0.7301,
"step": 675
},
{
"epoch": 0.3238095238095238,
"grad_norm": 0.7388339763922518,
"learning_rate": 1.9767436117621416e-05,
"loss": 0.7217,
"step": 680
},
{
"epoch": 0.3261904761904762,
"grad_norm": 0.6684391122808601,
"learning_rate": 1.975844187672402e-05,
"loss": 0.6975,
"step": 685
},
{
"epoch": 0.32857142857142857,
"grad_norm": 0.6933743953258287,
"learning_rate": 1.9749279121818235e-05,
"loss": 0.712,
"step": 690
},
{
"epoch": 0.33095238095238094,
"grad_norm": 0.6999429639549631,
"learning_rate": 1.9739948011131438e-05,
"loss": 0.7275,
"step": 695
},
{
"epoch": 0.3333333333333333,
"grad_norm": 0.6381718516637829,
"learning_rate": 1.973044870579824e-05,
"loss": 0.7269,
"step": 700
},
{
"epoch": 0.3357142857142857,
"grad_norm": 0.6640845851267309,
"learning_rate": 1.9720781369857747e-05,
"loss": 0.7254,
"step": 705
},
{
"epoch": 0.3380952380952381,
"grad_norm": 0.6592163354528279,
"learning_rate": 1.9710946170250702e-05,
"loss": 0.7071,
"step": 710
},
{
"epoch": 0.3404761904761905,
"grad_norm": 0.7655182235424155,
"learning_rate": 1.9700943276816602e-05,
"loss": 0.7241,
"step": 715
},
{
"epoch": 0.34285714285714286,
"grad_norm": 0.643273448107574,
"learning_rate": 1.969077286229078e-05,
"loss": 0.7286,
"step": 720
},
{
"epoch": 0.34523809523809523,
"grad_norm": 0.6857609061991075,
"learning_rate": 1.9680435102301412e-05,
"loss": 0.7079,
"step": 725
},
{
"epoch": 0.3476190476190476,
"grad_norm": 0.8672880237155081,
"learning_rate": 1.9669930175366474e-05,
"loss": 0.7112,
"step": 730
},
{
"epoch": 0.35,
"grad_norm": 0.6654257223083666,
"learning_rate": 1.9659258262890683e-05,
"loss": 0.726,
"step": 735
},
{
"epoch": 0.3523809523809524,
"grad_norm": 0.6635633178878109,
"learning_rate": 1.964841954916235e-05,
"loss": 0.7095,
"step": 740
},
{
"epoch": 0.3547619047619048,
"grad_norm": 0.7692774931160247,
"learning_rate": 1.9637414221350198e-05,
"loss": 0.7078,
"step": 745
},
{
"epoch": 0.35714285714285715,
"grad_norm": 0.6290937699685107,
"learning_rate": 1.962624246950012e-05,
"loss": 0.7228,
"step": 750
},
{
"epoch": 0.3595238095238095,
"grad_norm": 0.645738278667474,
"learning_rate": 1.9614904486531935e-05,
"loss": 0.7134,
"step": 755
},
{
"epoch": 0.3619047619047619,
"grad_norm": 0.6234306399886239,
"learning_rate": 1.9603400468236e-05,
"loss": 0.7132,
"step": 760
},
{
"epoch": 0.36428571428571427,
"grad_norm": 0.6027074648266817,
"learning_rate": 1.9591730613269878e-05,
"loss": 0.7019,
"step": 765
},
{
"epoch": 0.36666666666666664,
"grad_norm": 0.5952054087646503,
"learning_rate": 1.957989512315489e-05,
"loss": 0.7134,
"step": 770
},
{
"epoch": 0.36904761904761907,
"grad_norm": 0.5954424676547249,
"learning_rate": 1.9567894202272623e-05,
"loss": 0.7102,
"step": 775
},
{
"epoch": 0.37142857142857144,
"grad_norm": 0.6146466190073,
"learning_rate": 1.955572805786141e-05,
"loss": 0.7054,
"step": 780
},
{
"epoch": 0.3738095238095238,
"grad_norm": 0.6196401686238979,
"learning_rate": 1.9543396900012763e-05,
"loss": 0.7223,
"step": 785
},
{
"epoch": 0.3761904761904762,
"grad_norm": 0.6376202895536227,
"learning_rate": 1.9530900941667733e-05,
"loss": 0.7227,
"step": 790
},
{
"epoch": 0.37857142857142856,
"grad_norm": 0.6600915780771018,
"learning_rate": 1.9518240398613226e-05,
"loss": 0.6997,
"step": 795
},
{
"epoch": 0.38095238095238093,
"grad_norm": 0.5882182414716866,
"learning_rate": 1.9505415489478293e-05,
"loss": 0.7268,
"step": 800
},
{
"epoch": 0.38333333333333336,
"grad_norm": 0.6384408692158889,
"learning_rate": 1.949242643573034e-05,
"loss": 0.712,
"step": 805
},
{
"epoch": 0.38571428571428573,
"grad_norm": 0.6036619986439133,
"learning_rate": 1.947927346167132e-05,
"loss": 0.7175,
"step": 810
},
{
"epoch": 0.3880952380952381,
"grad_norm": 0.7029032826216544,
"learning_rate": 1.9465956794433837e-05,
"loss": 0.707,
"step": 815
},
{
"epoch": 0.3904761904761905,
"grad_norm": 0.6457510326949268,
"learning_rate": 1.945247666397725e-05,
"loss": 0.7036,
"step": 820
},
{
"epoch": 0.39285714285714285,
"grad_norm": 0.6878440257755464,
"learning_rate": 1.9438833303083677e-05,
"loss": 0.6989,
"step": 825
},
{
"epoch": 0.3952380952380952,
"grad_norm": 0.6188605951379363,
"learning_rate": 1.9425026947353994e-05,
"loss": 0.7105,
"step": 830
},
{
"epoch": 0.3976190476190476,
"grad_norm": 0.5840120067803837,
"learning_rate": 1.9411057835203756e-05,
"loss": 0.7058,
"step": 835
},
{
"epoch": 0.4,
"grad_norm": 0.5937069559439255,
"learning_rate": 1.9396926207859085e-05,
"loss": 0.695,
"step": 840
},
{
"epoch": 0.4023809523809524,
"grad_norm": 0.7110875706843621,
"learning_rate": 1.9382632309352503e-05,
"loss": 0.7227,
"step": 845
},
{
"epoch": 0.40476190476190477,
"grad_norm": 0.6311314888265864,
"learning_rate": 1.936817638651871e-05,
"loss": 0.7011,
"step": 850
},
{
"epoch": 0.40714285714285714,
"grad_norm": 0.6286300385263309,
"learning_rate": 1.935355868899034e-05,
"loss": 0.6979,
"step": 855
},
{
"epoch": 0.4095238095238095,
"grad_norm": 0.6209403160626118,
"learning_rate": 1.9338779469193638e-05,
"loss": 0.6996,
"step": 860
},
{
"epoch": 0.4119047619047619,
"grad_norm": 0.5727631219976955,
"learning_rate": 1.9323838982344092e-05,
"loss": 0.7114,
"step": 865
},
{
"epoch": 0.4142857142857143,
"grad_norm": 0.6417775935264546,
"learning_rate": 1.9308737486442045e-05,
"loss": 0.702,
"step": 870
},
{
"epoch": 0.4166666666666667,
"grad_norm": 0.6620199236263371,
"learning_rate": 1.9293475242268224e-05,
"loss": 0.7226,
"step": 875
},
{
"epoch": 0.41904761904761906,
"grad_norm": 0.6399880589166196,
"learning_rate": 1.9278052513379256e-05,
"loss": 0.6925,
"step": 880
},
{
"epoch": 0.42142857142857143,
"grad_norm": 0.6138453722448008,
"learning_rate": 1.926246956610309e-05,
"loss": 0.7118,
"step": 885
},
{
"epoch": 0.4238095238095238,
"grad_norm": 0.597989080002243,
"learning_rate": 1.9246726669534416e-05,
"loss": 0.73,
"step": 890
},
{
"epoch": 0.4261904761904762,
"grad_norm": 0.5901338899915719,
"learning_rate": 1.923082409553002e-05,
"loss": 0.7067,
"step": 895
},
{
"epoch": 0.42857142857142855,
"grad_norm": 0.6016114619725854,
"learning_rate": 1.921476211870408e-05,
"loss": 0.7103,
"step": 900
},
{
"epoch": 0.430952380952381,
"grad_norm": 0.5950314108298304,
"learning_rate": 1.919854101642342e-05,
"loss": 0.7052,
"step": 905
},
{
"epoch": 0.43333333333333335,
"grad_norm": 0.5879370937942204,
"learning_rate": 1.9182161068802742e-05,
"loss": 0.6995,
"step": 910
},
{
"epoch": 0.4357142857142857,
"grad_norm": 0.6266968125559099,
"learning_rate": 1.9165622558699763e-05,
"loss": 0.7097,
"step": 915
},
{
"epoch": 0.4380952380952381,
"grad_norm": 0.6639659528952871,
"learning_rate": 1.9148925771710347e-05,
"loss": 0.7009,
"step": 920
},
{
"epoch": 0.44047619047619047,
"grad_norm": 0.6321493695976013,
"learning_rate": 1.913207099616357e-05,
"loss": 0.7053,
"step": 925
},
{
"epoch": 0.44285714285714284,
"grad_norm": 0.6782903937955044,
"learning_rate": 1.9115058523116734e-05,
"loss": 0.7004,
"step": 930
},
{
"epoch": 0.4452380952380952,
"grad_norm": 0.6716356876066842,
"learning_rate": 1.9097888646350347e-05,
"loss": 0.7058,
"step": 935
},
{
"epoch": 0.44761904761904764,
"grad_norm": 0.5994163783124282,
"learning_rate": 1.908056166236305e-05,
"loss": 0.6965,
"step": 940
},
{
"epoch": 0.45,
"grad_norm": 0.7011305510255346,
"learning_rate": 1.9063077870366504e-05,
"loss": 0.6973,
"step": 945
},
{
"epoch": 0.4523809523809524,
"grad_norm": 0.6382034265421508,
"learning_rate": 1.9045437572280193e-05,
"loss": 0.7122,
"step": 950
},
{
"epoch": 0.45476190476190476,
"grad_norm": 0.652835650762035,
"learning_rate": 1.902764107272626e-05,
"loss": 0.7154,
"step": 955
},
{
"epoch": 0.45714285714285713,
"grad_norm": 0.5737401205474331,
"learning_rate": 1.900968867902419e-05,
"loss": 0.7094,
"step": 960
},
{
"epoch": 0.4595238095238095,
"grad_norm": 0.6510894452271442,
"learning_rate": 1.8991580701185564e-05,
"loss": 0.7126,
"step": 965
},
{
"epoch": 0.46190476190476193,
"grad_norm": 0.7386017741154395,
"learning_rate": 1.8973317451908642e-05,
"loss": 0.7007,
"step": 970
},
{
"epoch": 0.4642857142857143,
"grad_norm": 0.6752036021573017,
"learning_rate": 1.895489924657301e-05,
"loss": 0.7041,
"step": 975
},
{
"epoch": 0.4666666666666667,
"grad_norm": 0.597408051286458,
"learning_rate": 1.8936326403234125e-05,
"loss": 0.7212,
"step": 980
},
{
"epoch": 0.46904761904761905,
"grad_norm": 0.5796714555019369,
"learning_rate": 1.8917599242617796e-05,
"loss": 0.7159,
"step": 985
},
{
"epoch": 0.4714285714285714,
"grad_norm": 0.5688480490870661,
"learning_rate": 1.8898718088114688e-05,
"loss": 0.6997,
"step": 990
},
{
"epoch": 0.4738095238095238,
"grad_norm": 0.6090212361351263,
"learning_rate": 1.8879683265774695e-05,
"loss": 0.7152,
"step": 995
},
{
"epoch": 0.47619047619047616,
"grad_norm": 0.5638933903277263,
"learning_rate": 1.8860495104301346e-05,
"loss": 0.689,
"step": 1000
},
{
"epoch": 0.4785714285714286,
"grad_norm": 0.5775879905939127,
"learning_rate": 1.8841153935046098e-05,
"loss": 0.7071,
"step": 1005
},
{
"epoch": 0.48095238095238096,
"grad_norm": 0.5955667805199039,
"learning_rate": 1.8821660092002642e-05,
"loss": 0.718,
"step": 1010
},
{
"epoch": 0.48333333333333334,
"grad_norm": 0.6578844006654452,
"learning_rate": 1.880201391180111e-05,
"loss": 0.7203,
"step": 1015
},
{
"epoch": 0.4857142857142857,
"grad_norm": 0.6288999781322553,
"learning_rate": 1.8782215733702286e-05,
"loss": 0.7005,
"step": 1020
},
{
"epoch": 0.4880952380952381,
"grad_norm": 0.6154852254510217,
"learning_rate": 1.8762265899591724e-05,
"loss": 0.7112,
"step": 1025
},
{
"epoch": 0.49047619047619045,
"grad_norm": 0.5946678081221346,
"learning_rate": 1.874216475397386e-05,
"loss": 0.7072,
"step": 1030
},
{
"epoch": 0.4928571428571429,
"grad_norm": 0.5542144431999712,
"learning_rate": 1.8721912643966055e-05,
"loss": 0.7234,
"step": 1035
},
{
"epoch": 0.49523809523809526,
"grad_norm": 0.6327874278600627,
"learning_rate": 1.870150991929261e-05,
"loss": 0.7199,
"step": 1040
},
{
"epoch": 0.4976190476190476,
"grad_norm": 0.6050382016312242,
"learning_rate": 1.868095693227872e-05,
"loss": 0.6989,
"step": 1045
},
{
"epoch": 0.5,
"grad_norm": 0.6060484644475487,
"learning_rate": 1.866025403784439e-05,
"loss": 0.7094,
"step": 1050
},
{
"epoch": 0.5023809523809524,
"grad_norm": 0.5578204676405346,
"learning_rate": 1.86394015934983e-05,
"loss": 0.6998,
"step": 1055
},
{
"epoch": 0.5047619047619047,
"grad_norm": 0.5912489553089566,
"learning_rate": 1.8618399959331642e-05,
"loss": 0.704,
"step": 1060
},
{
"epoch": 0.5071428571428571,
"grad_norm": 0.6302255111379419,
"learning_rate": 1.8597249498011906e-05,
"loss": 0.7273,
"step": 1065
},
{
"epoch": 0.5095238095238095,
"grad_norm": 0.6377098555634427,
"learning_rate": 1.8575950574776595e-05,
"loss": 0.7118,
"step": 1070
},
{
"epoch": 0.5119047619047619,
"grad_norm": 0.6306289974408017,
"learning_rate": 1.855450355742695e-05,
"loss": 0.6975,
"step": 1075
},
{
"epoch": 0.5142857142857142,
"grad_norm": 0.6118126960406806,
"learning_rate": 1.8532908816321557e-05,
"loss": 0.7089,
"step": 1080
},
{
"epoch": 0.5166666666666667,
"grad_norm": 0.5631635531801148,
"learning_rate": 1.8511166724369997e-05,
"loss": 0.7027,
"step": 1085
},
{
"epoch": 0.5190476190476191,
"grad_norm": 0.5712800757429177,
"learning_rate": 1.8489277657026377e-05,
"loss": 0.7068,
"step": 1090
},
{
"epoch": 0.5214285714285715,
"grad_norm": 0.560805851835222,
"learning_rate": 1.8467241992282842e-05,
"loss": 0.7012,
"step": 1095
},
{
"epoch": 0.5238095238095238,
"grad_norm": 0.6607109669389528,
"learning_rate": 1.844506011066308e-05,
"loss": 0.6919,
"step": 1100
},
{
"epoch": 0.5261904761904762,
"grad_norm": 0.5900866880976501,
"learning_rate": 1.8422732395215717e-05,
"loss": 0.6903,
"step": 1105
},
{
"epoch": 0.5285714285714286,
"grad_norm": 0.5897342260968936,
"learning_rate": 1.8400259231507716e-05,
"loss": 0.6908,
"step": 1110
},
{
"epoch": 0.530952380952381,
"grad_norm": 0.5723672657261824,
"learning_rate": 1.8377641007617724e-05,
"loss": 0.6954,
"step": 1115
},
{
"epoch": 0.5333333333333333,
"grad_norm": 0.611058093387684,
"learning_rate": 1.8354878114129368e-05,
"loss": 0.7017,
"step": 1120
},
{
"epoch": 0.5357142857142857,
"grad_norm": 0.5318902267208334,
"learning_rate": 1.833197094412449e-05,
"loss": 0.6953,
"step": 1125
},
{
"epoch": 0.5380952380952381,
"grad_norm": 0.5570664080835552,
"learning_rate": 1.8308919893176397e-05,
"loss": 0.7066,
"step": 1130
},
{
"epoch": 0.5404761904761904,
"grad_norm": 0.5885337969920328,
"learning_rate": 1.8285725359343e-05,
"loss": 0.706,
"step": 1135
},
{
"epoch": 0.5428571428571428,
"grad_norm": 0.5855314615908318,
"learning_rate": 1.826238774315995e-05,
"loss": 0.701,
"step": 1140
},
{
"epoch": 0.5452380952380952,
"grad_norm": 0.5851946825105788,
"learning_rate": 1.8238907447633716e-05,
"loss": 0.692,
"step": 1145
},
{
"epoch": 0.5476190476190477,
"grad_norm": 0.5927970007531321,
"learning_rate": 1.8215284878234644e-05,
"loss": 0.716,
"step": 1150
},
{
"epoch": 0.55,
"grad_norm": 0.5263874307074303,
"learning_rate": 1.819152044288992e-05,
"loss": 0.6904,
"step": 1155
},
{
"epoch": 0.5523809523809524,
"grad_norm": 0.6124793068078697,
"learning_rate": 1.816761455197657e-05,
"loss": 0.6864,
"step": 1160
},
{
"epoch": 0.5547619047619048,
"grad_norm": 0.5828934954007933,
"learning_rate": 1.8143567618314336e-05,
"loss": 0.6904,
"step": 1165
},
{
"epoch": 0.5571428571428572,
"grad_norm": 0.5831980381697426,
"learning_rate": 1.811938005715857e-05,
"loss": 0.7086,
"step": 1170
},
{
"epoch": 0.5595238095238095,
"grad_norm": 0.6435462278644365,
"learning_rate": 1.8095052286193044e-05,
"loss": 0.6976,
"step": 1175
},
{
"epoch": 0.5619047619047619,
"grad_norm": 0.5815432160069361,
"learning_rate": 1.8070584725522763e-05,
"loss": 0.6925,
"step": 1180
},
{
"epoch": 0.5642857142857143,
"grad_norm": 0.5401285846482883,
"learning_rate": 1.8045977797666685e-05,
"loss": 0.6999,
"step": 1185
},
{
"epoch": 0.5666666666666667,
"grad_norm": 0.5677019239930577,
"learning_rate": 1.802123192755044e-05,
"loss": 0.6923,
"step": 1190
},
{
"epoch": 0.569047619047619,
"grad_norm": 0.5683306800833698,
"learning_rate": 1.7996347542498983e-05,
"loss": 0.7017,
"step": 1195
},
{
"epoch": 0.5714285714285714,
"grad_norm": 0.5113013942929165,
"learning_rate": 1.7971325072229227e-05,
"loss": 0.6921,
"step": 1200
},
{
"epoch": 0.5738095238095238,
"grad_norm": 0.5841499935734249,
"learning_rate": 1.7946164948842604e-05,
"loss": 0.6915,
"step": 1205
},
{
"epoch": 0.5761904761904761,
"grad_norm": 0.6279630088060286,
"learning_rate": 1.7920867606817625e-05,
"loss": 0.7061,
"step": 1210
},
{
"epoch": 0.5785714285714286,
"grad_norm": 0.5599484344769574,
"learning_rate": 1.7895433483002356e-05,
"loss": 0.715,
"step": 1215
},
{
"epoch": 0.580952380952381,
"grad_norm": 0.5661003553169632,
"learning_rate": 1.7869863016606893e-05,
"loss": 0.6986,
"step": 1220
},
{
"epoch": 0.5833333333333334,
"grad_norm": 0.5716166266125108,
"learning_rate": 1.784415664919576e-05,
"loss": 0.7031,
"step": 1225
},
{
"epoch": 0.5857142857142857,
"grad_norm": 0.5567338032160342,
"learning_rate": 1.78183148246803e-05,
"loss": 0.7068,
"step": 1230
},
{
"epoch": 0.5880952380952381,
"grad_norm": 0.5683883961420689,
"learning_rate": 1.7792337989311e-05,
"loss": 0.689,
"step": 1235
},
{
"epoch": 0.5904761904761905,
"grad_norm": 0.5477282163005394,
"learning_rate": 1.7766226591669787e-05,
"loss": 0.689,
"step": 1240
},
{
"epoch": 0.5928571428571429,
"grad_norm": 0.5311691713580732,
"learning_rate": 1.7739981082662275e-05,
"loss": 0.7215,
"step": 1245
},
{
"epoch": 0.5952380952380952,
"grad_norm": 0.598618505069467,
"learning_rate": 1.771360191551e-05,
"loss": 0.7151,
"step": 1250
},
{
"epoch": 0.5976190476190476,
"grad_norm": 0.5543663594273839,
"learning_rate": 1.768708954574256e-05,
"loss": 0.701,
"step": 1255
},
{
"epoch": 0.6,
"grad_norm": 0.5321278870334822,
"learning_rate": 1.766044443118978e-05,
"loss": 0.6935,
"step": 1260
},
{
"epoch": 0.6023809523809524,
"grad_norm": 0.5564736107581606,
"learning_rate": 1.7633667031973793e-05,
"loss": 0.6925,
"step": 1265
},
{
"epoch": 0.6047619047619047,
"grad_norm": 0.623003830977377,
"learning_rate": 1.760675781050109e-05,
"loss": 0.6904,
"step": 1270
},
{
"epoch": 0.6071428571428571,
"grad_norm": 0.5425831672559595,
"learning_rate": 1.757971723145453e-05,
"loss": 0.6941,
"step": 1275
},
{
"epoch": 0.6095238095238096,
"grad_norm": 0.5691326789130872,
"learning_rate": 1.755254576178535e-05,
"loss": 0.6943,
"step": 1280
},
{
"epoch": 0.611904761904762,
"grad_norm": 0.5472161245788499,
"learning_rate": 1.7525243870705052e-05,
"loss": 0.7019,
"step": 1285
},
{
"epoch": 0.6142857142857143,
"grad_norm": 0.5126814767906358,
"learning_rate": 1.7497812029677344e-05,
"loss": 0.6979,
"step": 1290
},
{
"epoch": 0.6166666666666667,
"grad_norm": 0.5808568867050167,
"learning_rate": 1.7470250712409963e-05,
"loss": 0.7017,
"step": 1295
},
{
"epoch": 0.6190476190476191,
"grad_norm": 0.5731714166045258,
"learning_rate": 1.7442560394846518e-05,
"loss": 0.6896,
"step": 1300
},
{
"epoch": 0.6214285714285714,
"grad_norm": 0.5416282781910543,
"learning_rate": 1.741474155515827e-05,
"loss": 0.6881,
"step": 1305
},
{
"epoch": 0.6238095238095238,
"grad_norm": 0.5989028487744685,
"learning_rate": 1.738679467373586e-05,
"loss": 0.6853,
"step": 1310
},
{
"epoch": 0.6261904761904762,
"grad_norm": 0.5928143358170742,
"learning_rate": 1.7358720233181023e-05,
"loss": 0.6945,
"step": 1315
},
{
"epoch": 0.6285714285714286,
"grad_norm": 0.5917504797279695,
"learning_rate": 1.7330518718298263e-05,
"loss": 0.708,
"step": 1320
},
{
"epoch": 0.6309523809523809,
"grad_norm": 0.5735050094781514,
"learning_rate": 1.7302190616086464e-05,
"loss": 0.7048,
"step": 1325
},
{
"epoch": 0.6333333333333333,
"grad_norm": 0.57827020923432,
"learning_rate": 1.7273736415730488e-05,
"loss": 0.6793,
"step": 1330
},
{
"epoch": 0.6357142857142857,
"grad_norm": 0.5880932620123437,
"learning_rate": 1.7245156608592727e-05,
"loss": 0.6915,
"step": 1335
},
{
"epoch": 0.638095238095238,
"grad_norm": 0.5890769914796811,
"learning_rate": 1.7216451688204623e-05,
"loss": 0.69,
"step": 1340
},
{
"epoch": 0.6404761904761904,
"grad_norm": 0.527942811449849,
"learning_rate": 1.718762215025813e-05,
"loss": 0.682,
"step": 1345
},
{
"epoch": 0.6428571428571429,
"grad_norm": 0.6301973133824051,
"learning_rate": 1.7158668492597186e-05,
"loss": 0.6934,
"step": 1350
},
{
"epoch": 0.6452380952380953,
"grad_norm": 0.5410524060403492,
"learning_rate": 1.712959121520907e-05,
"loss": 0.6892,
"step": 1355
},
{
"epoch": 0.6476190476190476,
"grad_norm": 0.7801149059212953,
"learning_rate": 1.7100390820215805e-05,
"loss": 0.6999,
"step": 1360
},
{
"epoch": 0.65,
"grad_norm": 0.5546640495559166,
"learning_rate": 1.7071067811865477e-05,
"loss": 0.6742,
"step": 1365
},
{
"epoch": 0.6523809523809524,
"grad_norm": 0.5722087188570253,
"learning_rate": 1.704162269652352e-05,
"loss": 0.7179,
"step": 1370
},
{
"epoch": 0.6547619047619048,
"grad_norm": 0.5249847693462395,
"learning_rate": 1.701205598266398e-05,
"loss": 0.6945,
"step": 1375
},
{
"epoch": 0.6571428571428571,
"grad_norm": 0.5639053989400045,
"learning_rate": 1.698236818086073e-05,
"loss": 0.6925,
"step": 1380
},
{
"epoch": 0.6595238095238095,
"grad_norm": 0.5897582949134731,
"learning_rate": 1.6952559803778656e-05,
"loss": 0.6958,
"step": 1385
},
{
"epoch": 0.6619047619047619,
"grad_norm": 0.5506537272905166,
"learning_rate": 1.6922631366164795e-05,
"loss": 0.6744,
"step": 1390
},
{
"epoch": 0.6642857142857143,
"grad_norm": 0.5613120451869654,
"learning_rate": 1.689258338483947e-05,
"loss": 0.687,
"step": 1395
},
{
"epoch": 0.6666666666666666,
"grad_norm": 0.5687059713184212,
"learning_rate": 1.686241637868734e-05,
"loss": 0.7011,
"step": 1400
},
{
"epoch": 0.669047619047619,
"grad_norm": 0.5267403463210103,
"learning_rate": 1.683213086864843e-05,
"loss": 0.6903,
"step": 1405
},
{
"epoch": 0.6714285714285714,
"grad_norm": 0.5655713255912641,
"learning_rate": 1.6801727377709195e-05,
"loss": 0.6896,
"step": 1410
},
{
"epoch": 0.6738095238095239,
"grad_norm": 0.5285552000123795,
"learning_rate": 1.6771206430893408e-05,
"loss": 0.6946,
"step": 1415
},
{
"epoch": 0.6761904761904762,
"grad_norm": 0.5761004876016846,
"learning_rate": 1.6740568555253153e-05,
"loss": 0.6895,
"step": 1420
},
{
"epoch": 0.6785714285714286,
"grad_norm": 0.5390507685841307,
"learning_rate": 1.67098142798597e-05,
"loss": 0.6942,
"step": 1425
},
{
"epoch": 0.680952380952381,
"grad_norm": 0.5132897821784491,
"learning_rate": 1.6678944135794375e-05,
"loss": 0.6948,
"step": 1430
},
{
"epoch": 0.6833333333333333,
"grad_norm": 0.5469471858786907,
"learning_rate": 1.6647958656139377e-05,
"loss": 0.6944,
"step": 1435
},
{
"epoch": 0.6857142857142857,
"grad_norm": 0.553201497920686,
"learning_rate": 1.6616858375968596e-05,
"loss": 0.6788,
"step": 1440
},
{
"epoch": 0.6880952380952381,
"grad_norm": 0.5623963206097778,
"learning_rate": 1.6585643832338342e-05,
"loss": 0.6815,
"step": 1445
},
{
"epoch": 0.6904761904761905,
"grad_norm": 0.5778391136516833,
"learning_rate": 1.6554315564278102e-05,
"loss": 0.6843,
"step": 1450
},
{
"epoch": 0.6928571428571428,
"grad_norm": 0.5551499013662371,
"learning_rate": 1.6522874112781213e-05,
"loss": 0.6872,
"step": 1455
},
{
"epoch": 0.6952380952380952,
"grad_norm": 0.5309204279264464,
"learning_rate": 1.649132002079552e-05,
"loss": 0.6873,
"step": 1460
},
{
"epoch": 0.6976190476190476,
"grad_norm": 0.508876104576006,
"learning_rate": 1.645965383321401e-05,
"loss": 0.6792,
"step": 1465
},
{
"epoch": 0.7,
"grad_norm": 0.5802541953819853,
"learning_rate": 1.6427876096865394e-05,
"loss": 0.6881,
"step": 1470
},
{
"epoch": 0.7023809523809523,
"grad_norm": 0.5641741696009784,
"learning_rate": 1.6395987360504667e-05,
"loss": 0.6789,
"step": 1475
},
{
"epoch": 0.7047619047619048,
"grad_norm": 0.5366027622698923,
"learning_rate": 1.6363988174803638e-05,
"loss": 0.686,
"step": 1480
},
{
"epoch": 0.7071428571428572,
"grad_norm": 0.5234128778556285,
"learning_rate": 1.6331879092341402e-05,
"loss": 0.7134,
"step": 1485
},
{
"epoch": 0.7095238095238096,
"grad_norm": 0.5759728577947257,
"learning_rate": 1.6299660667594814e-05,
"loss": 0.701,
"step": 1490
},
{
"epoch": 0.7119047619047619,
"grad_norm": 0.5756551282837324,
"learning_rate": 1.626733345692892e-05,
"loss": 0.6843,
"step": 1495
},
{
"epoch": 0.7142857142857143,
"grad_norm": 0.5426722958339261,
"learning_rate": 1.6234898018587336e-05,
"loss": 0.6906,
"step": 1500
},
{
"epoch": 0.7166666666666667,
"grad_norm": 0.48554563250860505,
"learning_rate": 1.6202354912682602e-05,
"loss": 0.6824,
"step": 1505
},
{
"epoch": 0.719047619047619,
"grad_norm": 0.5744429442854848,
"learning_rate": 1.6169704701186528e-05,
"loss": 0.6979,
"step": 1510
},
{
"epoch": 0.7214285714285714,
"grad_norm": 0.5277461906754672,
"learning_rate": 1.6136947947920477e-05,
"loss": 0.682,
"step": 1515
},
{
"epoch": 0.7238095238095238,
"grad_norm": 0.5681961605460194,
"learning_rate": 1.6104085218545633e-05,
"loss": 0.6893,
"step": 1520
},
{
"epoch": 0.7261904761904762,
"grad_norm": 0.58856327424583,
"learning_rate": 1.6071117080553236e-05,
"loss": 0.694,
"step": 1525
},
{
"epoch": 0.7285714285714285,
"grad_norm": 0.5585907608313626,
"learning_rate": 1.6038044103254775e-05,
"loss": 0.6807,
"step": 1530
},
{
"epoch": 0.7309523809523809,
"grad_norm": 0.4975987871579692,
"learning_rate": 1.600486685777216e-05,
"loss": 0.6839,
"step": 1535
},
{
"epoch": 0.7333333333333333,
"grad_norm": 0.5683357922839659,
"learning_rate": 1.5971585917027864e-05,
"loss": 0.6924,
"step": 1540
},
{
"epoch": 0.7357142857142858,
"grad_norm": 0.5507550407504584,
"learning_rate": 1.5938201855735017e-05,
"loss": 0.6952,
"step": 1545
},
{
"epoch": 0.7380952380952381,
"grad_norm": 0.5184672194625306,
"learning_rate": 1.5904715250387498e-05,
"loss": 0.6864,
"step": 1550
},
{
"epoch": 0.7404761904761905,
"grad_norm": 0.5315141830604754,
"learning_rate": 1.5871126679249977e-05,
"loss": 0.686,
"step": 1555
},
{
"epoch": 0.7428571428571429,
"grad_norm": 0.5292902163348473,
"learning_rate": 1.5837436722347902e-05,
"loss": 0.6826,
"step": 1560
},
{
"epoch": 0.7452380952380953,
"grad_norm": 0.570110548449842,
"learning_rate": 1.5803645961457522e-05,
"loss": 0.6792,
"step": 1565
},
{
"epoch": 0.7476190476190476,
"grad_norm": 0.5543717733657283,
"learning_rate": 1.576975498009583e-05,
"loss": 0.6754,
"step": 1570
},
{
"epoch": 0.75,
"grad_norm": 0.5222973089292835,
"learning_rate": 1.573576436351046e-05,
"loss": 0.6902,
"step": 1575
},
{
"epoch": 0.7523809523809524,
"grad_norm": 0.5112696664324252,
"learning_rate": 1.570167469866962e-05,
"loss": 0.6972,
"step": 1580
},
{
"epoch": 0.7547619047619047,
"grad_norm": 0.5807388645411898,
"learning_rate": 1.5667486574251916e-05,
"loss": 0.6863,
"step": 1585
},
{
"epoch": 0.7571428571428571,
"grad_norm": 0.5522196746267701,
"learning_rate": 1.563320058063622e-05,
"loss": 0.6705,
"step": 1590
},
{
"epoch": 0.7595238095238095,
"grad_norm": 0.5302665914230218,
"learning_rate": 1.5598817309891466e-05,
"loss": 0.6842,
"step": 1595
},
{
"epoch": 0.7619047619047619,
"grad_norm": 0.5286809702848861,
"learning_rate": 1.5564337355766412e-05,
"loss": 0.6976,
"step": 1600
},
{
"epoch": 0.7642857142857142,
"grad_norm": 0.5393439786479364,
"learning_rate": 1.5529761313679396e-05,
"loss": 0.7063,
"step": 1605
},
{
"epoch": 0.7666666666666667,
"grad_norm": 0.479895247668522,
"learning_rate": 1.5495089780708062e-05,
"loss": 0.6737,
"step": 1610
},
{
"epoch": 0.7690476190476191,
"grad_norm": 0.5401818745358772,
"learning_rate": 1.5460323355579035e-05,
"loss": 0.6808,
"step": 1615
},
{
"epoch": 0.7714285714285715,
"grad_norm": 0.6132409197814304,
"learning_rate": 1.5425462638657597e-05,
"loss": 0.6912,
"step": 1620
},
{
"epoch": 0.7738095238095238,
"grad_norm": 0.5605359051745877,
"learning_rate": 1.53905082319373e-05,
"loss": 0.6912,
"step": 1625
},
{
"epoch": 0.7761904761904762,
"grad_norm": 0.5095591675298641,
"learning_rate": 1.5355460739029585e-05,
"loss": 0.6854,
"step": 1630
},
{
"epoch": 0.7785714285714286,
"grad_norm": 0.5124981468976912,
"learning_rate": 1.5320320765153367e-05,
"loss": 0.6805,
"step": 1635
},
{
"epoch": 0.780952380952381,
"grad_norm": 0.5543805272032538,
"learning_rate": 1.5285088917124555e-05,
"loss": 0.6872,
"step": 1640
},
{
"epoch": 0.7833333333333333,
"grad_norm": 0.5240863656017781,
"learning_rate": 1.5249765803345602e-05,
"loss": 0.6812,
"step": 1645
},
{
"epoch": 0.7857142857142857,
"grad_norm": 0.5597831529794077,
"learning_rate": 1.5214352033794981e-05,
"loss": 0.6838,
"step": 1650
},
{
"epoch": 0.7880952380952381,
"grad_norm": 0.5239372373145886,
"learning_rate": 1.517884822001666e-05,
"loss": 0.6884,
"step": 1655
},
{
"epoch": 0.7904761904761904,
"grad_norm": 0.5776231529708008,
"learning_rate": 1.5143254975109538e-05,
"loss": 0.6728,
"step": 1660
},
{
"epoch": 0.7928571428571428,
"grad_norm": 0.5215371586345835,
"learning_rate": 1.5107572913716859e-05,
"loss": 0.6748,
"step": 1665
},
{
"epoch": 0.7952380952380952,
"grad_norm": 0.5282942517891518,
"learning_rate": 1.5071802652015592e-05,
"loss": 0.6713,
"step": 1670
},
{
"epoch": 0.7976190476190477,
"grad_norm": 0.5541374243554975,
"learning_rate": 1.503594480770581e-05,
"loss": 0.6597,
"step": 1675
},
{
"epoch": 0.8,
"grad_norm": 0.5521158876521693,
"learning_rate": 1.5000000000000002e-05,
"loss": 0.686,
"step": 1680
},
{
"epoch": 0.8023809523809524,
"grad_norm": 0.5683308647861297,
"learning_rate": 1.496396884961238e-05,
"loss": 0.6693,
"step": 1685
},
{
"epoch": 0.8047619047619048,
"grad_norm": 0.573228337305753,
"learning_rate": 1.4927851978748177e-05,
"loss": 0.6926,
"step": 1690
},
{
"epoch": 0.8071428571428572,
"grad_norm": 0.521492632745778,
"learning_rate": 1.4891650011092896e-05,
"loss": 0.6947,
"step": 1695
},
{
"epoch": 0.8095238095238095,
"grad_norm": 0.5236798235457268,
"learning_rate": 1.4855363571801523e-05,
"loss": 0.679,
"step": 1700
},
{
"epoch": 0.8119047619047619,
"grad_norm": 0.5183500676344015,
"learning_rate": 1.481899328748776e-05,
"loss": 0.6657,
"step": 1705
},
{
"epoch": 0.8142857142857143,
"grad_norm": 0.5018602569821992,
"learning_rate": 1.4782539786213184e-05,
"loss": 0.6836,
"step": 1710
},
{
"epoch": 0.8166666666666667,
"grad_norm": 0.5469580528625957,
"learning_rate": 1.4746003697476406e-05,
"loss": 0.6841,
"step": 1715
},
{
"epoch": 0.819047619047619,
"grad_norm": 0.5610880961409519,
"learning_rate": 1.4709385652202204e-05,
"loss": 0.6864,
"step": 1720
},
{
"epoch": 0.8214285714285714,
"grad_norm": 0.5376175612680147,
"learning_rate": 1.4672686282730622e-05,
"loss": 0.6709,
"step": 1725
},
{
"epoch": 0.8238095238095238,
"grad_norm": 0.49458630831613376,
"learning_rate": 1.4635906222806058e-05,
"loss": 0.6832,
"step": 1730
},
{
"epoch": 0.8261904761904761,
"grad_norm": 0.5502214264070843,
"learning_rate": 1.4599046107566314e-05,
"loss": 0.68,
"step": 1735
},
{
"epoch": 0.8285714285714286,
"grad_norm": 0.5611151303400349,
"learning_rate": 1.4562106573531632e-05,
"loss": 0.698,
"step": 1740
},
{
"epoch": 0.830952380952381,
"grad_norm": 0.5735560840092161,
"learning_rate": 1.4525088258593695e-05,
"loss": 0.6839,
"step": 1745
},
{
"epoch": 0.8333333333333334,
"grad_norm": 0.5510543712119496,
"learning_rate": 1.4487991802004625e-05,
"loss": 0.6911,
"step": 1750
},
{
"epoch": 0.8357142857142857,
"grad_norm": 0.5018246480371149,
"learning_rate": 1.4450817844365924e-05,
"loss": 0.6752,
"step": 1755
},
{
"epoch": 0.8380952380952381,
"grad_norm": 0.5016576788767521,
"learning_rate": 1.4413567027617442e-05,
"loss": 0.683,
"step": 1760
},
{
"epoch": 0.8404761904761905,
"grad_norm": 0.5384327340092925,
"learning_rate": 1.4376239995026254e-05,
"loss": 0.6941,
"step": 1765
},
{
"epoch": 0.8428571428571429,
"grad_norm": 0.5344821729910321,
"learning_rate": 1.4338837391175582e-05,
"loss": 0.6754,
"step": 1770
},
{
"epoch": 0.8452380952380952,
"grad_norm": 0.529015386728389,
"learning_rate": 1.4301359861953652e-05,
"loss": 0.6718,
"step": 1775
},
{
"epoch": 0.8476190476190476,
"grad_norm": 0.5467267886434187,
"learning_rate": 1.4263808054542541e-05,
"loss": 0.6885,
"step": 1780
},
{
"epoch": 0.85,
"grad_norm": 0.5209380288088742,
"learning_rate": 1.4226182617406996e-05,
"loss": 0.6725,
"step": 1785
},
{
"epoch": 0.8523809523809524,
"grad_norm": 0.5120625589068878,
"learning_rate": 1.418848420028325e-05,
"loss": 0.678,
"step": 1790
},
{
"epoch": 0.8547619047619047,
"grad_norm": 0.5177286158188428,
"learning_rate": 1.4150713454167788e-05,
"loss": 0.6795,
"step": 1795
},
{
"epoch": 0.8571428571428571,
"grad_norm": 0.5180241473157509,
"learning_rate": 1.4112871031306118e-05,
"loss": 0.685,
"step": 1800
},
{
"epoch": 0.8595238095238096,
"grad_norm": 0.5376670786030856,
"learning_rate": 1.4074957585181488e-05,
"loss": 0.6823,
"step": 1805
},
{
"epoch": 0.861904761904762,
"grad_norm": 0.5159889497243569,
"learning_rate": 1.4036973770503623e-05,
"loss": 0.6684,
"step": 1810
},
{
"epoch": 0.8642857142857143,
"grad_norm": 0.5328385205015794,
"learning_rate": 1.3998920243197408e-05,
"loss": 0.6785,
"step": 1815
},
{
"epoch": 0.8666666666666667,
"grad_norm": 0.5596052359171065,
"learning_rate": 1.396079766039157e-05,
"loss": 0.6971,
"step": 1820
},
{
"epoch": 0.8690476190476191,
"grad_norm": 0.5476844927900769,
"learning_rate": 1.3922606680407307e-05,
"loss": 0.6843,
"step": 1825
},
{
"epoch": 0.8714285714285714,
"grad_norm": 0.49918261945442965,
"learning_rate": 1.3884347962746949e-05,
"loss": 0.6761,
"step": 1830
},
{
"epoch": 0.8738095238095238,
"grad_norm": 0.5116149514821221,
"learning_rate": 1.3846022168082553e-05,
"loss": 0.6763,
"step": 1835
},
{
"epoch": 0.8761904761904762,
"grad_norm": 0.5280902250134181,
"learning_rate": 1.3807629958244498e-05,
"loss": 0.667,
"step": 1840
},
{
"epoch": 0.8785714285714286,
"grad_norm": 0.48857909294423746,
"learning_rate": 1.3769171996210053e-05,
"loss": 0.6707,
"step": 1845
},
{
"epoch": 0.8809523809523809,
"grad_norm": 0.4975153014854871,
"learning_rate": 1.373064894609194e-05,
"loss": 0.688,
"step": 1850
},
{
"epoch": 0.8833333333333333,
"grad_norm": 0.5206570618857133,
"learning_rate": 1.3692061473126845e-05,
"loss": 0.687,
"step": 1855
},
{
"epoch": 0.8857142857142857,
"grad_norm": 0.47088196958418377,
"learning_rate": 1.3653410243663953e-05,
"loss": 0.6659,
"step": 1860
},
{
"epoch": 0.888095238095238,
"grad_norm": 0.5098589712869,
"learning_rate": 1.361469592515342e-05,
"loss": 0.6784,
"step": 1865
},
{
"epoch": 0.8904761904761904,
"grad_norm": 0.47648371702756814,
"learning_rate": 1.3575919186134862e-05,
"loss": 0.682,
"step": 1870
},
{
"epoch": 0.8928571428571429,
"grad_norm": 0.5003536067341952,
"learning_rate": 1.3537080696225815e-05,
"loss": 0.6671,
"step": 1875
},
{
"epoch": 0.8952380952380953,
"grad_norm": 0.519430135596028,
"learning_rate": 1.349818112611015e-05,
"loss": 0.6885,
"step": 1880
},
{
"epoch": 0.8976190476190476,
"grad_norm": 0.49741488661881644,
"learning_rate": 1.3459221147526504e-05,
"loss": 0.6754,
"step": 1885
},
{
"epoch": 0.9,
"grad_norm": 0.5408088216797241,
"learning_rate": 1.342020143325669e-05,
"loss": 0.6757,
"step": 1890
},
{
"epoch": 0.9023809523809524,
"grad_norm": 0.502061784820922,
"learning_rate": 1.3381122657114059e-05,
"loss": 0.669,
"step": 1895
},
{
"epoch": 0.9047619047619048,
"grad_norm": 0.5031558920583495,
"learning_rate": 1.3341985493931877e-05,
"loss": 0.6756,
"step": 1900
},
{
"epoch": 0.9071428571428571,
"grad_norm": 0.45651178296082845,
"learning_rate": 1.3302790619551673e-05,
"loss": 0.6629,
"step": 1905
},
{
"epoch": 0.9095238095238095,
"grad_norm": 0.5111760063303825,
"learning_rate": 1.3263538710811559e-05,
"loss": 0.6787,
"step": 1910
},
{
"epoch": 0.9119047619047619,
"grad_norm": 0.5038700388575827,
"learning_rate": 1.3224230445534544e-05,
"loss": 0.6814,
"step": 1915
},
{
"epoch": 0.9142857142857143,
"grad_norm": 0.5023414779956755,
"learning_rate": 1.3184866502516846e-05,
"loss": 0.6859,
"step": 1920
},
{
"epoch": 0.9166666666666666,
"grad_norm": 0.5154019438212942,
"learning_rate": 1.3145447561516138e-05,
"loss": 0.6717,
"step": 1925
},
{
"epoch": 0.919047619047619,
"grad_norm": 0.5621900790890961,
"learning_rate": 1.3105974303239838e-05,
"loss": 0.7009,
"step": 1930
},
{
"epoch": 0.9214285714285714,
"grad_norm": 0.5137612757351608,
"learning_rate": 1.3066447409333345e-05,
"loss": 0.6957,
"step": 1935
},
{
"epoch": 0.9238095238095239,
"grad_norm": 0.509045076292474,
"learning_rate": 1.3026867562368262e-05,
"loss": 0.6804,
"step": 1940
},
{
"epoch": 0.9261904761904762,
"grad_norm": 0.5067359253184963,
"learning_rate": 1.2987235445830612e-05,
"loss": 0.6707,
"step": 1945
},
{
"epoch": 0.9285714285714286,
"grad_norm": 0.5488540810447187,
"learning_rate": 1.2947551744109044e-05,
"loss": 0.6784,
"step": 1950
},
{
"epoch": 0.930952380952381,
"grad_norm": 0.49916198703229403,
"learning_rate": 1.2907817142483002e-05,
"loss": 0.6533,
"step": 1955
},
{
"epoch": 0.9333333333333333,
"grad_norm": 0.5306836509898268,
"learning_rate": 1.2868032327110904e-05,
"loss": 0.6694,
"step": 1960
},
{
"epoch": 0.9357142857142857,
"grad_norm": 0.5192454887104005,
"learning_rate": 1.2828197985018276e-05,
"loss": 0.6814,
"step": 1965
},
{
"epoch": 0.9380952380952381,
"grad_norm": 0.48011848457133666,
"learning_rate": 1.2788314804085904e-05,
"loss": 0.6656,
"step": 1970
},
{
"epoch": 0.9404761904761905,
"grad_norm": 0.501551598406461,
"learning_rate": 1.2748383473037948e-05,
"loss": 0.6709,
"step": 1975
},
{
"epoch": 0.9428571428571428,
"grad_norm": 0.4798488372067313,
"learning_rate": 1.2708404681430054e-05,
"loss": 0.6884,
"step": 1980
},
{
"epoch": 0.9452380952380952,
"grad_norm": 0.4954430271705226,
"learning_rate": 1.266837911963743e-05,
"loss": 0.6789,
"step": 1985
},
{
"epoch": 0.9476190476190476,
"grad_norm": 0.5134775709739979,
"learning_rate": 1.2628307478842955e-05,
"loss": 0.6743,
"step": 1990
},
{
"epoch": 0.95,
"grad_norm": 0.4932466173893802,
"learning_rate": 1.2588190451025209e-05,
"loss": 0.6741,
"step": 1995
},
{
"epoch": 0.9523809523809523,
"grad_norm": 0.49375841486555283,
"learning_rate": 1.2548028728946548e-05,
"loss": 0.682,
"step": 2000
},
{
"epoch": 0.9547619047619048,
"grad_norm": 0.553381199639126,
"learning_rate": 1.2507823006141128e-05,
"loss": 0.704,
"step": 2005
},
{
"epoch": 0.9571428571428572,
"grad_norm": 0.5286463574396035,
"learning_rate": 1.2467573976902936e-05,
"loss": 0.6642,
"step": 2010
},
{
"epoch": 0.9595238095238096,
"grad_norm": 0.5330908591194152,
"learning_rate": 1.24272823362738e-05,
"loss": 0.6973,
"step": 2015
},
{
"epoch": 0.9619047619047619,
"grad_norm": 0.5338462915299015,
"learning_rate": 1.238694878003138e-05,
"loss": 0.6608,
"step": 2020
},
{
"epoch": 0.9642857142857143,
"grad_norm": 0.5626548093456667,
"learning_rate": 1.2346574004677154e-05,
"loss": 0.6807,
"step": 2025
},
{
"epoch": 0.9666666666666667,
"grad_norm": 0.5363264212587037,
"learning_rate": 1.2306158707424402e-05,
"loss": 0.6742,
"step": 2030
},
{
"epoch": 0.969047619047619,
"grad_norm": 0.5223504672217075,
"learning_rate": 1.2265703586186158e-05,
"loss": 0.6756,
"step": 2035
},
{
"epoch": 0.9714285714285714,
"grad_norm": 0.5026492923180158,
"learning_rate": 1.2225209339563144e-05,
"loss": 0.6857,
"step": 2040
},
{
"epoch": 0.9738095238095238,
"grad_norm": 0.4958319271212678,
"learning_rate": 1.2184676666831741e-05,
"loss": 0.6638,
"step": 2045
},
{
"epoch": 0.9761904761904762,
"grad_norm": 0.49120488962765346,
"learning_rate": 1.2144106267931877e-05,
"loss": 0.6767,
"step": 2050
},
{
"epoch": 0.9785714285714285,
"grad_norm": 0.56338802960644,
"learning_rate": 1.210349884345496e-05,
"loss": 0.6657,
"step": 2055
},
{
"epoch": 0.9809523809523809,
"grad_norm": 0.5506153013895527,
"learning_rate": 1.2062855094631777e-05,
"loss": 0.6708,
"step": 2060
},
{
"epoch": 0.9833333333333333,
"grad_norm": 0.4911898234196216,
"learning_rate": 1.2022175723320382e-05,
"loss": 0.6736,
"step": 2065
},
{
"epoch": 0.9857142857142858,
"grad_norm": 0.5037996502262528,
"learning_rate": 1.1981461431993978e-05,
"loss": 0.6613,
"step": 2070
},
{
"epoch": 0.9880952380952381,
"grad_norm": 0.48983503414857615,
"learning_rate": 1.1940712923728784e-05,
"loss": 0.6705,
"step": 2075
},
{
"epoch": 0.9904761904761905,
"grad_norm": 0.5089658881380381,
"learning_rate": 1.1899930902191904e-05,
"loss": 0.6554,
"step": 2080
},
{
"epoch": 0.9928571428571429,
"grad_norm": 0.5353432851904248,
"learning_rate": 1.1859116071629148e-05,
"loss": 0.6713,
"step": 2085
},
{
"epoch": 0.9952380952380953,
"grad_norm": 0.5083284406725272,
"learning_rate": 1.181826913685291e-05,
"loss": 0.6593,
"step": 2090
},
{
"epoch": 0.9976190476190476,
"grad_norm": 0.5338566827047485,
"learning_rate": 1.1777390803229964e-05,
"loss": 0.6771,
"step": 2095
},
{
"epoch": 1.0,
"grad_norm": 0.5259573856012386,
"learning_rate": 1.1736481776669307e-05,
"loss": 0.6603,
"step": 2100
},
{
"epoch": 1.0,
"eval_loss": 0.7571888566017151,
"eval_runtime": 110.7511,
"eval_samples_per_second": 77.299,
"eval_steps_per_second": 1.21,
"step": 2100
},
{
"epoch": 1.0023809523809524,
"grad_norm": 0.5901038052706317,
"learning_rate": 1.1695542763609944e-05,
"loss": 0.5957,
"step": 2105
},
{
"epoch": 1.0047619047619047,
"grad_norm": 0.5438749703677144,
"learning_rate": 1.1654574471008712e-05,
"loss": 0.5972,
"step": 2110
},
{
"epoch": 1.0071428571428571,
"grad_norm": 0.5432897226752162,
"learning_rate": 1.1613577606328068e-05,
"loss": 0.595,
"step": 2115
},
{
"epoch": 1.0095238095238095,
"grad_norm": 0.5692170924338006,
"learning_rate": 1.1572552877523855e-05,
"loss": 0.5869,
"step": 2120
},
{
"epoch": 1.0119047619047619,
"grad_norm": 0.5633601882768045,
"learning_rate": 1.1531500993033094e-05,
"loss": 0.59,
"step": 2125
},
{
"epoch": 1.0142857142857142,
"grad_norm": 0.5402220697166501,
"learning_rate": 1.1490422661761744e-05,
"loss": 0.5928,
"step": 2130
},
{
"epoch": 1.0166666666666666,
"grad_norm": 0.6014408487693759,
"learning_rate": 1.1449318593072468e-05,
"loss": 0.5943,
"step": 2135
},
{
"epoch": 1.019047619047619,
"grad_norm": 0.48300700922294626,
"learning_rate": 1.1408189496772369e-05,
"loss": 0.593,
"step": 2140
},
{
"epoch": 1.0214285714285714,
"grad_norm": 0.5461515536108149,
"learning_rate": 1.1367036083100735e-05,
"loss": 0.6118,
"step": 2145
},
{
"epoch": 1.0238095238095237,
"grad_norm": 0.5225774711520927,
"learning_rate": 1.1325859062716795e-05,
"loss": 0.5934,
"step": 2150
},
{
"epoch": 1.026190476190476,
"grad_norm": 0.5652770991249514,
"learning_rate": 1.1284659146687416e-05,
"loss": 0.5906,
"step": 2155
},
{
"epoch": 1.0285714285714285,
"grad_norm": 0.49282124739115296,
"learning_rate": 1.1243437046474854e-05,
"loss": 0.6029,
"step": 2160
},
{
"epoch": 1.0309523809523808,
"grad_norm": 0.517411950290186,
"learning_rate": 1.120219347392444e-05,
"loss": 0.6105,
"step": 2165
},
{
"epoch": 1.0333333333333334,
"grad_norm": 0.5547803815907365,
"learning_rate": 1.1160929141252303e-05,
"loss": 0.5888,
"step": 2170
},
{
"epoch": 1.0357142857142858,
"grad_norm": 0.5448805389067126,
"learning_rate": 1.1119644761033079e-05,
"loss": 0.606,
"step": 2175
},
{
"epoch": 1.0380952380952382,
"grad_norm": 0.5270587508515787,
"learning_rate": 1.1078341046187588e-05,
"loss": 0.5978,
"step": 2180
},
{
"epoch": 1.0404761904761906,
"grad_norm": 0.5470808129733004,
"learning_rate": 1.1037018709970528e-05,
"loss": 0.6079,
"step": 2185
},
{
"epoch": 1.042857142857143,
"grad_norm": 0.5185275354171291,
"learning_rate": 1.0995678465958168e-05,
"loss": 0.5998,
"step": 2190
},
{
"epoch": 1.0452380952380953,
"grad_norm": 0.5222569126782542,
"learning_rate": 1.0954321028036013e-05,
"loss": 0.6065,
"step": 2195
},
{
"epoch": 1.0476190476190477,
"grad_norm": 0.5902869820958465,
"learning_rate": 1.0912947110386484e-05,
"loss": 0.5952,
"step": 2200
},
{
"epoch": 1.05,
"grad_norm": 0.5049079899744102,
"learning_rate": 1.0871557427476585e-05,
"loss": 0.5984,
"step": 2205
},
{
"epoch": 1.0523809523809524,
"grad_norm": 0.5211587946973806,
"learning_rate": 1.0830152694045553e-05,
"loss": 0.5822,
"step": 2210
},
{
"epoch": 1.0547619047619048,
"grad_norm": 0.5411611115879803,
"learning_rate": 1.078873362509254e-05,
"loss": 0.5882,
"step": 2215
},
{
"epoch": 1.0571428571428572,
"grad_norm": 0.5442183494774795,
"learning_rate": 1.0747300935864245e-05,
"loss": 0.5996,
"step": 2220
},
{
"epoch": 1.0595238095238095,
"grad_norm": 0.5612089724718119,
"learning_rate": 1.0705855341842564e-05,
"loss": 0.6,
"step": 2225
},
{
"epoch": 1.061904761904762,
"grad_norm": 0.5293855485137681,
"learning_rate": 1.0664397558732245e-05,
"loss": 0.5918,
"step": 2230
},
{
"epoch": 1.0642857142857143,
"grad_norm": 0.5907188005151464,
"learning_rate": 1.0622928302448523e-05,
"loss": 0.608,
"step": 2235
},
{
"epoch": 1.0666666666666667,
"grad_norm": 0.5251701766583902,
"learning_rate": 1.0581448289104759e-05,
"loss": 0.6058,
"step": 2240
},
{
"epoch": 1.069047619047619,
"grad_norm": 0.5250031426883407,
"learning_rate": 1.0539958235000075e-05,
"loss": 0.6064,
"step": 2245
},
{
"epoch": 1.0714285714285714,
"grad_norm": 0.5217021022246795,
"learning_rate": 1.0498458856606972e-05,
"loss": 0.6098,
"step": 2250
},
{
"epoch": 1.0738095238095238,
"grad_norm": 0.5671230242631112,
"learning_rate": 1.0456950870558982e-05,
"loss": 0.5985,
"step": 2255
},
{
"epoch": 1.0761904761904761,
"grad_norm": 0.5640807879300845,
"learning_rate": 1.0415434993638269e-05,
"loss": 0.5992,
"step": 2260
},
{
"epoch": 1.0785714285714285,
"grad_norm": 0.5349112926063765,
"learning_rate": 1.037391194276326e-05,
"loss": 0.6061,
"step": 2265
},
{
"epoch": 1.0809523809523809,
"grad_norm": 0.5592160403982473,
"learning_rate": 1.0332382434976267e-05,
"loss": 0.5864,
"step": 2270
},
{
"epoch": 1.0833333333333333,
"grad_norm": 0.5235545437475504,
"learning_rate": 1.0290847187431115e-05,
"loss": 0.592,
"step": 2275
},
{
"epoch": 1.0857142857142856,
"grad_norm": 0.556896126496801,
"learning_rate": 1.0249306917380731e-05,
"loss": 0.6005,
"step": 2280
},
{
"epoch": 1.088095238095238,
"grad_norm": 0.522494265444706,
"learning_rate": 1.0207762342164778e-05,
"loss": 0.5958,
"step": 2285
},
{
"epoch": 1.0904761904761904,
"grad_norm": 0.5307880240675681,
"learning_rate": 1.0166214179197265e-05,
"loss": 0.5911,
"step": 2290
},
{
"epoch": 1.092857142857143,
"grad_norm": 0.5364461659449388,
"learning_rate": 1.0124663145954152e-05,
"loss": 0.6018,
"step": 2295
},
{
"epoch": 1.0952380952380953,
"grad_norm": 0.5520100652972162,
"learning_rate": 1.0083109959960974e-05,
"loss": 0.5955,
"step": 2300
},
{
"epoch": 1.0976190476190477,
"grad_norm": 0.5020623533166694,
"learning_rate": 1.0041555338780427e-05,
"loss": 0.5961,
"step": 2305
},
{
"epoch": 1.1,
"grad_norm": 0.52041207867341,
"learning_rate": 1e-05,
"loss": 0.5954,
"step": 2310
},
{
"epoch": 1.1023809523809525,
"grad_norm": 0.569101456193797,
"learning_rate": 9.958444661219578e-06,
"loss": 0.6046,
"step": 2315
},
{
"epoch": 1.1047619047619048,
"grad_norm": 0.5290911882364042,
"learning_rate": 9.916890040039031e-06,
"loss": 0.5891,
"step": 2320
},
{
"epoch": 1.1071428571428572,
"grad_norm": 0.5305645713582501,
"learning_rate": 9.87533685404585e-06,
"loss": 0.5988,
"step": 2325
},
{
"epoch": 1.1095238095238096,
"grad_norm": 0.5469543228815086,
"learning_rate": 9.833785820802739e-06,
"loss": 0.5957,
"step": 2330
},
{
"epoch": 1.111904761904762,
"grad_norm": 0.5324948019005903,
"learning_rate": 9.792237657835225e-06,
"loss": 0.5953,
"step": 2335
},
{
"epoch": 1.1142857142857143,
"grad_norm": 0.5304279403221299,
"learning_rate": 9.750693082619274e-06,
"loss": 0.5937,
"step": 2340
},
{
"epoch": 1.1166666666666667,
"grad_norm": 0.5485078647852204,
"learning_rate": 9.709152812568886e-06,
"loss": 0.5987,
"step": 2345
},
{
"epoch": 1.119047619047619,
"grad_norm": 0.552749009152491,
"learning_rate": 9.667617565023734e-06,
"loss": 0.5978,
"step": 2350
},
{
"epoch": 1.1214285714285714,
"grad_norm": 0.5177836960191152,
"learning_rate": 9.626088057236745e-06,
"loss": 0.5973,
"step": 2355
},
{
"epoch": 1.1238095238095238,
"grad_norm": 0.6109837673102856,
"learning_rate": 9.584565006361735e-06,
"loss": 0.595,
"step": 2360
},
{
"epoch": 1.1261904761904762,
"grad_norm": 0.5167360036746549,
"learning_rate": 9.543049129441021e-06,
"loss": 0.5891,
"step": 2365
},
{
"epoch": 1.1285714285714286,
"grad_norm": 0.5297524492676912,
"learning_rate": 9.501541143393028e-06,
"loss": 0.606,
"step": 2370
},
{
"epoch": 1.130952380952381,
"grad_norm": 0.5411957992836974,
"learning_rate": 9.460041764999929e-06,
"loss": 0.5973,
"step": 2375
},
{
"epoch": 1.1333333333333333,
"grad_norm": 0.5287203316976509,
"learning_rate": 9.418551710895243e-06,
"loss": 0.5848,
"step": 2380
},
{
"epoch": 1.1357142857142857,
"grad_norm": 0.6710026001993966,
"learning_rate": 9.377071697551479e-06,
"loss": 0.6086,
"step": 2385
},
{
"epoch": 1.138095238095238,
"grad_norm": 0.5503478041423361,
"learning_rate": 9.33560244126776e-06,
"loss": 0.5941,
"step": 2390
},
{
"epoch": 1.1404761904761904,
"grad_norm": 0.5347482917383127,
"learning_rate": 9.294144658157443e-06,
"loss": 0.5974,
"step": 2395
},
{
"epoch": 1.1428571428571428,
"grad_norm": 0.5497165323598144,
"learning_rate": 9.252699064135759e-06,
"loss": 0.6046,
"step": 2400
},
{
"epoch": 1.1452380952380952,
"grad_norm": 0.5190090017461467,
"learning_rate": 9.211266374907463e-06,
"loss": 0.5869,
"step": 2405
},
{
"epoch": 1.1476190476190475,
"grad_norm": 0.5719299143043652,
"learning_rate": 9.169847305954448e-06,
"loss": 0.6048,
"step": 2410
},
{
"epoch": 1.15,
"grad_norm": 0.5672551417094259,
"learning_rate": 9.128442572523418e-06,
"loss": 0.6068,
"step": 2415
},
{
"epoch": 1.1523809523809523,
"grad_norm": 0.5811171033851283,
"learning_rate": 9.087052889613519e-06,
"loss": 0.5982,
"step": 2420
},
{
"epoch": 1.1547619047619047,
"grad_norm": 0.5227291415650476,
"learning_rate": 9.045678971963988e-06,
"loss": 0.591,
"step": 2425
},
{
"epoch": 1.157142857142857,
"grad_norm": 0.5413206931558642,
"learning_rate": 9.004321534041836e-06,
"loss": 0.5916,
"step": 2430
},
{
"epoch": 1.1595238095238094,
"grad_norm": 0.5298903577250825,
"learning_rate": 8.962981290029475e-06,
"loss": 0.5883,
"step": 2435
},
{
"epoch": 1.161904761904762,
"grad_norm": 0.4876508870345226,
"learning_rate": 8.921658953812416e-06,
"loss": 0.5938,
"step": 2440
},
{
"epoch": 1.1642857142857144,
"grad_norm": 0.5198015834539063,
"learning_rate": 8.880355238966923e-06,
"loss": 0.5975,
"step": 2445
},
{
"epoch": 1.1666666666666667,
"grad_norm": 0.5097940796805619,
"learning_rate": 8.839070858747697e-06,
"loss": 0.5868,
"step": 2450
},
{
"epoch": 1.1690476190476191,
"grad_norm": 0.5663643871944489,
"learning_rate": 8.797806526075566e-06,
"loss": 0.5902,
"step": 2455
},
{
"epoch": 1.1714285714285715,
"grad_norm": 0.5396345415398657,
"learning_rate": 8.756562953525151e-06,
"loss": 0.5958,
"step": 2460
},
{
"epoch": 1.1738095238095239,
"grad_norm": 0.5577630837652126,
"learning_rate": 8.715340853312586e-06,
"loss": 0.5833,
"step": 2465
},
{
"epoch": 1.1761904761904762,
"grad_norm": 0.5518263276547959,
"learning_rate": 8.674140937283208e-06,
"loss": 0.5935,
"step": 2470
},
{
"epoch": 1.1785714285714286,
"grad_norm": 0.5084066290422407,
"learning_rate": 8.632963916899268e-06,
"loss": 0.5835,
"step": 2475
},
{
"epoch": 1.180952380952381,
"grad_norm": 0.5098303646711122,
"learning_rate": 8.591810503227634e-06,
"loss": 0.6024,
"step": 2480
},
{
"epoch": 1.1833333333333333,
"grad_norm": 0.5399045936688955,
"learning_rate": 8.550681406927534e-06,
"loss": 0.5916,
"step": 2485
},
{
"epoch": 1.1857142857142857,
"grad_norm": 0.5139645415633143,
"learning_rate": 8.509577338238255e-06,
"loss": 0.5867,
"step": 2490
},
{
"epoch": 1.188095238095238,
"grad_norm": 0.5670585522900756,
"learning_rate": 8.46849900696691e-06,
"loss": 0.6006,
"step": 2495
},
{
"epoch": 1.1904761904761905,
"grad_norm": 0.5457106741839642,
"learning_rate": 8.427447122476148e-06,
"loss": 0.597,
"step": 2500
},
{
"epoch": 1.1928571428571428,
"grad_norm": 0.5571733334835492,
"learning_rate": 8.386422393671934e-06,
"loss": 0.5913,
"step": 2505
},
{
"epoch": 1.1952380952380952,
"grad_norm": 0.529527429611943,
"learning_rate": 8.34542552899129e-06,
"loss": 0.6006,
"step": 2510
},
{
"epoch": 1.1976190476190476,
"grad_norm": 0.5235732448086102,
"learning_rate": 8.304457236390062e-06,
"loss": 0.5819,
"step": 2515
},
{
"epoch": 1.2,
"grad_norm": 0.5332920937716661,
"learning_rate": 8.263518223330698e-06,
"loss": 0.6024,
"step": 2520
},
{
"epoch": 1.2023809523809523,
"grad_norm": 0.5137113000682924,
"learning_rate": 8.222609196770037e-06,
"loss": 0.6001,
"step": 2525
},
{
"epoch": 1.2047619047619047,
"grad_norm": 0.5000093032244755,
"learning_rate": 8.181730863147094e-06,
"loss": 0.5839,
"step": 2530
},
{
"epoch": 1.207142857142857,
"grad_norm": 0.5084439515294099,
"learning_rate": 8.140883928370855e-06,
"loss": 0.5987,
"step": 2535
},
{
"epoch": 1.2095238095238094,
"grad_norm": 0.5420324353647344,
"learning_rate": 8.100069097808103e-06,
"loss": 0.5918,
"step": 2540
},
{
"epoch": 1.2119047619047618,
"grad_norm": 0.5051962888682799,
"learning_rate": 8.059287076271216e-06,
"loss": 0.5929,
"step": 2545
},
{
"epoch": 1.2142857142857142,
"grad_norm": 0.5112423923421076,
"learning_rate": 8.018538568006027e-06,
"loss": 0.5851,
"step": 2550
},
{
"epoch": 1.2166666666666668,
"grad_norm": 0.5404164861280687,
"learning_rate": 7.977824276679623e-06,
"loss": 0.6097,
"step": 2555
},
{
"epoch": 1.2190476190476192,
"grad_norm": 0.5318602860411683,
"learning_rate": 7.937144905368226e-06,
"loss": 0.5952,
"step": 2560
},
{
"epoch": 1.2214285714285715,
"grad_norm": 0.5521690395673133,
"learning_rate": 7.896501156545044e-06,
"loss": 0.594,
"step": 2565
},
{
"epoch": 1.223809523809524,
"grad_norm": 0.4999490189546094,
"learning_rate": 7.855893732068124e-06,
"loss": 0.5934,
"step": 2570
},
{
"epoch": 1.2261904761904763,
"grad_norm": 0.5048993618252382,
"learning_rate": 7.815323333168262e-06,
"loss": 0.5959,
"step": 2575
},
{
"epoch": 1.2285714285714286,
"grad_norm": 0.5109949262663191,
"learning_rate": 7.774790660436857e-06,
"loss": 0.5988,
"step": 2580
},
{
"epoch": 1.230952380952381,
"grad_norm": 0.5143824021233596,
"learning_rate": 7.734296413813847e-06,
"loss": 0.5844,
"step": 2585
},
{
"epoch": 1.2333333333333334,
"grad_norm": 0.5337082713078726,
"learning_rate": 7.6938412925756e-06,
"loss": 0.5898,
"step": 2590
},
{
"epoch": 1.2357142857142858,
"grad_norm": 0.5024058244542078,
"learning_rate": 7.653425995322852e-06,
"loss": 0.5959,
"step": 2595
},
{
"epoch": 1.2380952380952381,
"grad_norm": 0.5089373868097246,
"learning_rate": 7.613051219968624e-06,
"loss": 0.6007,
"step": 2600
},
{
"epoch": 1.2404761904761905,
"grad_norm": 0.5292691125658913,
"learning_rate": 7.5727176637262034e-06,
"loss": 0.5893,
"step": 2605
},
{
"epoch": 1.2428571428571429,
"grad_norm": 0.5105236474902384,
"learning_rate": 7.532426023097063e-06,
"loss": 0.5991,
"step": 2610
},
{
"epoch": 1.2452380952380953,
"grad_norm": 0.5289609790386512,
"learning_rate": 7.492176993858873e-06,
"loss": 0.5833,
"step": 2615
},
{
"epoch": 1.2476190476190476,
"grad_norm": 0.5289162389327398,
"learning_rate": 7.451971271053455e-06,
"loss": 0.5852,
"step": 2620
},
{
"epoch": 1.25,
"grad_norm": 0.570777487333234,
"learning_rate": 7.411809548974792e-06,
"loss": 0.5965,
"step": 2625
},
{
"epoch": 1.2523809523809524,
"grad_norm": 0.5208585348103933,
"learning_rate": 7.371692521157048e-06,
"loss": 0.6133,
"step": 2630
},
{
"epoch": 1.2547619047619047,
"grad_norm": 0.5279916869252734,
"learning_rate": 7.331620880362571e-06,
"loss": 0.5949,
"step": 2635
},
{
"epoch": 1.2571428571428571,
"grad_norm": 0.5269636791495603,
"learning_rate": 7.291595318569951e-06,
"loss": 0.6078,
"step": 2640
},
{
"epoch": 1.2595238095238095,
"grad_norm": 0.527352227349602,
"learning_rate": 7.2516165269620534e-06,
"loss": 0.5869,
"step": 2645
},
{
"epoch": 1.2619047619047619,
"grad_norm": 0.5144295584353445,
"learning_rate": 7.2116851959140965e-06,
"loss": 0.5918,
"step": 2650
},
{
"epoch": 1.2642857142857142,
"grad_norm": 0.5601926258663379,
"learning_rate": 7.171802014981726e-06,
"loss": 0.585,
"step": 2655
},
{
"epoch": 1.2666666666666666,
"grad_norm": 0.5379381373225919,
"learning_rate": 7.131967672889101e-06,
"loss": 0.595,
"step": 2660
},
{
"epoch": 1.269047619047619,
"grad_norm": 0.5251125215829732,
"learning_rate": 7.092182857516998e-06,
"loss": 0.5839,
"step": 2665
},
{
"epoch": 1.2714285714285714,
"grad_norm": 0.4965704448975545,
"learning_rate": 7.052448255890958e-06,
"loss": 0.5991,
"step": 2670
},
{
"epoch": 1.2738095238095237,
"grad_norm": 0.5450909049558729,
"learning_rate": 7.012764554169393e-06,
"loss": 0.6083,
"step": 2675
},
{
"epoch": 1.276190476190476,
"grad_norm": 0.5915810150427402,
"learning_rate": 6.973132437631743e-06,
"loss": 0.5899,
"step": 2680
},
{
"epoch": 1.2785714285714285,
"grad_norm": 0.5300752289151952,
"learning_rate": 6.933552590666659e-06,
"loss": 0.5876,
"step": 2685
},
{
"epoch": 1.2809523809523808,
"grad_norm": 0.5026232106359814,
"learning_rate": 6.8940256967601625e-06,
"loss": 0.5853,
"step": 2690
},
{
"epoch": 1.2833333333333332,
"grad_norm": 0.5197308254582468,
"learning_rate": 6.854552438483866e-06,
"loss": 0.5798,
"step": 2695
},
{
"epoch": 1.2857142857142856,
"grad_norm": 0.5211709556282136,
"learning_rate": 6.815133497483157e-06,
"loss": 0.5708,
"step": 2700
},
{
"epoch": 1.288095238095238,
"grad_norm": 0.4782326251158853,
"learning_rate": 6.775769554465455e-06,
"loss": 0.6023,
"step": 2705
},
{
"epoch": 1.2904761904761906,
"grad_norm": 0.5151978950364201,
"learning_rate": 6.736461289188445e-06,
"loss": 0.6054,
"step": 2710
},
{
"epoch": 1.292857142857143,
"grad_norm": 0.544144875536584,
"learning_rate": 6.697209380448333e-06,
"loss": 0.5983,
"step": 2715
},
{
"epoch": 1.2952380952380953,
"grad_norm": 0.525706847543693,
"learning_rate": 6.6580145060681255e-06,
"loss": 0.5909,
"step": 2720
},
{
"epoch": 1.2976190476190477,
"grad_norm": 0.5337499789796499,
"learning_rate": 6.618877342885945e-06,
"loss": 0.5932,
"step": 2725
},
{
"epoch": 1.3,
"grad_norm": 0.4979999689542144,
"learning_rate": 6.579798566743314e-06,
"loss": 0.5872,
"step": 2730
},
{
"epoch": 1.3023809523809524,
"grad_norm": 0.5293968795002583,
"learning_rate": 6.540778852473497e-06,
"loss": 0.595,
"step": 2735
},
{
"epoch": 1.3047619047619048,
"grad_norm": 0.513270923657658,
"learning_rate": 6.501818873889856e-06,
"loss": 0.5951,
"step": 2740
},
{
"epoch": 1.3071428571428572,
"grad_norm": 0.5284113230022897,
"learning_rate": 6.462919303774186e-06,
"loss": 0.5886,
"step": 2745
},
{
"epoch": 1.3095238095238095,
"grad_norm": 0.511386553250273,
"learning_rate": 6.424080813865139e-06,
"loss": 0.5785,
"step": 2750
},
{
"epoch": 1.311904761904762,
"grad_norm": 0.5612313704527047,
"learning_rate": 6.3853040748465855e-06,
"loss": 0.5934,
"step": 2755
},
{
"epoch": 1.3142857142857143,
"grad_norm": 0.49635536234133354,
"learning_rate": 6.34658975633605e-06,
"loss": 0.5804,
"step": 2760
},
{
"epoch": 1.3166666666666667,
"grad_norm": 0.5158067784575548,
"learning_rate": 6.3079385268731575e-06,
"loss": 0.5982,
"step": 2765
},
{
"epoch": 1.319047619047619,
"grad_norm": 0.5424701107493409,
"learning_rate": 6.269351053908061e-06,
"loss": 0.5873,
"step": 2770
},
{
"epoch": 1.3214285714285714,
"grad_norm": 0.5095803553683687,
"learning_rate": 6.230828003789949e-06,
"loss": 0.59,
"step": 2775
},
{
"epoch": 1.3238095238095238,
"grad_norm": 0.5004483050726994,
"learning_rate": 6.192370041755505e-06,
"loss": 0.588,
"step": 2780
},
{
"epoch": 1.3261904761904761,
"grad_norm": 0.5307962744979307,
"learning_rate": 6.153977831917451e-06,
"loss": 0.5879,
"step": 2785
},
{
"epoch": 1.3285714285714285,
"grad_norm": 0.5520488950744148,
"learning_rate": 6.115652037253054e-06,
"loss": 0.5967,
"step": 2790
},
{
"epoch": 1.3309523809523809,
"grad_norm": 0.4933954521187605,
"learning_rate": 6.077393319592697e-06,
"loss": 0.5922,
"step": 2795
},
{
"epoch": 1.3333333333333333,
"grad_norm": 0.5376192804239446,
"learning_rate": 6.039202339608432e-06,
"loss": 0.577,
"step": 2800
},
{
"epoch": 1.3357142857142856,
"grad_norm": 0.5213896422824574,
"learning_rate": 6.001079756802592e-06,
"loss": 0.58,
"step": 2805
},
{
"epoch": 1.3380952380952382,
"grad_norm": 0.5605679212505607,
"learning_rate": 5.963026229496378e-06,
"loss": 0.5801,
"step": 2810
},
{
"epoch": 1.3404761904761906,
"grad_norm": 0.5500894406128793,
"learning_rate": 5.925042414818514e-06,
"loss": 0.5893,
"step": 2815
},
{
"epoch": 1.342857142857143,
"grad_norm": 0.5051376225418218,
"learning_rate": 5.887128968693887e-06,
"loss": 0.5804,
"step": 2820
},
{
"epoch": 1.3452380952380953,
"grad_norm": 0.5241811747282762,
"learning_rate": 5.849286545832211e-06,
"loss": 0.576,
"step": 2825
},
{
"epoch": 1.3476190476190477,
"grad_norm": 0.5163613845049234,
"learning_rate": 5.811515799716754e-06,
"loss": 0.585,
"step": 2830
},
{
"epoch": 1.35,
"grad_norm": 0.5236385208944061,
"learning_rate": 5.773817382593008e-06,
"loss": 0.6006,
"step": 2835
},
{
"epoch": 1.3523809523809525,
"grad_norm": 0.532779005684654,
"learning_rate": 5.736191945457463e-06,
"loss": 0.5946,
"step": 2840
},
{
"epoch": 1.3547619047619048,
"grad_norm": 0.5259472612299654,
"learning_rate": 5.698640138046349e-06,
"loss": 0.5974,
"step": 2845
},
{
"epoch": 1.3571428571428572,
"grad_norm": 0.5109032533795757,
"learning_rate": 5.66116260882442e-06,
"loss": 0.5936,
"step": 2850
},
{
"epoch": 1.3595238095238096,
"grad_norm": 0.5122818131471542,
"learning_rate": 5.623760004973749e-06,
"loss": 0.5795,
"step": 2855
},
{
"epoch": 1.361904761904762,
"grad_norm": 0.4768696514947288,
"learning_rate": 5.586432972382561e-06,
"loss": 0.5758,
"step": 2860
},
{
"epoch": 1.3642857142857143,
"grad_norm": 0.53080481784541,
"learning_rate": 5.549182155634076e-06,
"loss": 0.5767,
"step": 2865
},
{
"epoch": 1.3666666666666667,
"grad_norm": 0.5038341903214718,
"learning_rate": 5.512008197995379e-06,
"loss": 0.5933,
"step": 2870
},
{
"epoch": 1.369047619047619,
"grad_norm": 0.5229825704537906,
"learning_rate": 5.47491174140631e-06,
"loss": 0.5824,
"step": 2875
},
{
"epoch": 1.3714285714285714,
"grad_norm": 0.49627233889726186,
"learning_rate": 5.43789342646837e-06,
"loss": 0.5988,
"step": 2880
},
{
"epoch": 1.3738095238095238,
"grad_norm": 0.514904186125751,
"learning_rate": 5.4009538924336864e-06,
"loss": 0.5876,
"step": 2885
},
{
"epoch": 1.3761904761904762,
"grad_norm": 0.5509458010595618,
"learning_rate": 5.364093777193944e-06,
"loss": 0.6045,
"step": 2890
},
{
"epoch": 1.3785714285714286,
"grad_norm": 0.5039270525726298,
"learning_rate": 5.32731371726938e-06,
"loss": 0.585,
"step": 2895
},
{
"epoch": 1.380952380952381,
"grad_norm": 0.5437109056015712,
"learning_rate": 5.290614347797802e-06,
"loss": 0.5924,
"step": 2900
},
{
"epoch": 1.3833333333333333,
"grad_norm": 0.5280438662821184,
"learning_rate": 5.253996302523596e-06,
"loss": 0.5808,
"step": 2905
},
{
"epoch": 1.3857142857142857,
"grad_norm": 0.5411985643744547,
"learning_rate": 5.217460213786822e-06,
"loss": 0.5909,
"step": 2910
},
{
"epoch": 1.388095238095238,
"grad_norm": 0.4928419336681969,
"learning_rate": 5.181006712512245e-06,
"loss": 0.5969,
"step": 2915
},
{
"epoch": 1.3904761904761904,
"grad_norm": 0.49801415174834524,
"learning_rate": 5.144636428198477e-06,
"loss": 0.5869,
"step": 2920
},
{
"epoch": 1.3928571428571428,
"grad_norm": 0.5060212093373364,
"learning_rate": 5.108349988907111e-06,
"loss": 0.5827,
"step": 2925
},
{
"epoch": 1.3952380952380952,
"grad_norm": 0.4959605026950924,
"learning_rate": 5.072148021251822e-06,
"loss": 0.5904,
"step": 2930
},
{
"epoch": 1.3976190476190475,
"grad_norm": 0.5256166273733636,
"learning_rate": 5.036031150387624e-06,
"loss": 0.582,
"step": 2935
},
{
"epoch": 1.4,
"grad_norm": 0.5644237371840894,
"learning_rate": 5.000000000000003e-06,
"loss": 0.5876,
"step": 2940
},
{
"epoch": 1.4023809523809523,
"grad_norm": 0.5544605233225762,
"learning_rate": 4.964055192294187e-06,
"loss": 0.5736,
"step": 2945
},
{
"epoch": 1.4047619047619047,
"grad_norm": 0.5111614229837703,
"learning_rate": 4.92819734798441e-06,
"loss": 0.5872,
"step": 2950
},
{
"epoch": 1.407142857142857,
"grad_norm": 0.4853589383206433,
"learning_rate": 4.892427086283147e-06,
"loss": 0.5899,
"step": 2955
},
{
"epoch": 1.4095238095238094,
"grad_norm": 0.5024366384928733,
"learning_rate": 4.856745024890466e-06,
"loss": 0.5739,
"step": 2960
},
{
"epoch": 1.4119047619047618,
"grad_norm": 0.5165017541670609,
"learning_rate": 4.821151779983343e-06,
"loss": 0.5889,
"step": 2965
},
{
"epoch": 1.4142857142857144,
"grad_norm": 0.5326057098439853,
"learning_rate": 4.78564796620502e-06,
"loss": 0.5984,
"step": 2970
},
{
"epoch": 1.4166666666666667,
"grad_norm": 0.5578793439842268,
"learning_rate": 4.7502341966544e-06,
"loss": 0.5919,
"step": 2975
},
{
"epoch": 1.4190476190476191,
"grad_norm": 0.5266644847974143,
"learning_rate": 4.714911082875446e-06,
"loss": 0.5811,
"step": 2980
},
{
"epoch": 1.4214285714285715,
"grad_norm": 0.5405356548584016,
"learning_rate": 4.679679234846636e-06,
"loss": 0.5852,
"step": 2985
},
{
"epoch": 1.4238095238095239,
"grad_norm": 0.5332823348361332,
"learning_rate": 4.644539260970417e-06,
"loss": 0.5904,
"step": 2990
},
{
"epoch": 1.4261904761904762,
"grad_norm": 0.5289204371473316,
"learning_rate": 4.609491768062705e-06,
"loss": 0.5715,
"step": 2995
},
{
"epoch": 1.4285714285714286,
"grad_norm": 0.49718290349750854,
"learning_rate": 4.5745373613424075e-06,
"loss": 0.5875,
"step": 3000
},
{
"epoch": 1.430952380952381,
"grad_norm": 0.5370577064296572,
"learning_rate": 4.539676644420966e-06,
"loss": 0.5838,
"step": 3005
},
{
"epoch": 1.4333333333333333,
"grad_norm": 0.5318499526470056,
"learning_rate": 4.504910219291941e-06,
"loss": 0.6081,
"step": 3010
},
{
"epoch": 1.4357142857142857,
"grad_norm": 0.5051255352622448,
"learning_rate": 4.470238686320606e-06,
"loss": 0.5796,
"step": 3015
},
{
"epoch": 1.438095238095238,
"grad_norm": 0.5318507665411399,
"learning_rate": 4.435662644233594e-06,
"loss": 0.5763,
"step": 3020
},
{
"epoch": 1.4404761904761905,
"grad_norm": 0.5337465779813616,
"learning_rate": 4.4011826901085346e-06,
"loss": 0.5856,
"step": 3025
},
{
"epoch": 1.4428571428571428,
"grad_norm": 0.5165852420124566,
"learning_rate": 4.3667994193637794e-06,
"loss": 0.5879,
"step": 3030
},
{
"epoch": 1.4452380952380952,
"grad_norm": 0.5155869722101444,
"learning_rate": 4.3325134257480905e-06,
"loss": 0.5831,
"step": 3035
},
{
"epoch": 1.4476190476190476,
"grad_norm": 0.4775386551110262,
"learning_rate": 4.298325301330383e-06,
"loss": 0.5866,
"step": 3040
},
{
"epoch": 1.45,
"grad_norm": 0.5054313057714112,
"learning_rate": 4.264235636489542e-06,
"loss": 0.5892,
"step": 3045
},
{
"epoch": 1.4523809523809523,
"grad_norm": 0.527424288981721,
"learning_rate": 4.23024501990417e-06,
"loss": 0.581,
"step": 3050
},
{
"epoch": 1.4547619047619047,
"grad_norm": 0.5089540127795982,
"learning_rate": 4.196354038542476e-06,
"loss": 0.5892,
"step": 3055
},
{
"epoch": 1.457142857142857,
"grad_norm": 0.5207863838757609,
"learning_rate": 4.162563277652104e-06,
"loss": 0.5865,
"step": 3060
},
{
"epoch": 1.4595238095238094,
"grad_norm": 0.5075766650861325,
"learning_rate": 4.128873320750027e-06,
"loss": 0.5756,
"step": 3065
},
{
"epoch": 1.461904761904762,
"grad_norm": 0.5025228767994807,
"learning_rate": 4.095284749612504e-06,
"loss": 0.5872,
"step": 3070
},
{
"epoch": 1.4642857142857144,
"grad_norm": 0.49216468636258304,
"learning_rate": 4.061798144264986e-06,
"loss": 0.5724,
"step": 3075
},
{
"epoch": 1.4666666666666668,
"grad_norm": 0.5053765390136462,
"learning_rate": 4.028414082972141e-06,
"loss": 0.5725,
"step": 3080
},
{
"epoch": 1.4690476190476192,
"grad_norm": 0.5698025824976778,
"learning_rate": 3.995133142227843e-06,
"loss": 0.5817,
"step": 3085
},
{
"epoch": 1.4714285714285715,
"grad_norm": 0.510204796748791,
"learning_rate": 3.961955896745224e-06,
"loss": 0.5777,
"step": 3090
},
{
"epoch": 1.473809523809524,
"grad_norm": 0.5091694662046772,
"learning_rate": 3.928882919446767e-06,
"loss": 0.5857,
"step": 3095
},
{
"epoch": 1.4761904761904763,
"grad_norm": 0.5151359736079936,
"learning_rate": 3.89591478145437e-06,
"loss": 0.5801,
"step": 3100
},
{
"epoch": 1.4785714285714286,
"grad_norm": 0.5182734669196073,
"learning_rate": 3.8630520520795275e-06,
"loss": 0.5693,
"step": 3105
},
{
"epoch": 1.480952380952381,
"grad_norm": 0.5129439390544596,
"learning_rate": 3.830295298813475e-06,
"loss": 0.586,
"step": 3110
},
{
"epoch": 1.4833333333333334,
"grad_norm": 0.5576172854485092,
"learning_rate": 3.797645087317401e-06,
"loss": 0.5753,
"step": 3115
},
{
"epoch": 1.4857142857142858,
"grad_norm": 0.5067485290590222,
"learning_rate": 3.7651019814126656e-06,
"loss": 0.5958,
"step": 3120
},
{
"epoch": 1.4880952380952381,
"grad_norm": 0.5319104622052202,
"learning_rate": 3.7326665430710798e-06,
"loss": 0.5868,
"step": 3125
},
{
"epoch": 1.4904761904761905,
"grad_norm": 0.49375049435832247,
"learning_rate": 3.7003393324051874e-06,
"loss": 0.5757,
"step": 3130
},
{
"epoch": 1.4928571428571429,
"grad_norm": 0.5074776283156045,
"learning_rate": 3.6681209076586035e-06,
"loss": 0.5845,
"step": 3135
},
{
"epoch": 1.4952380952380953,
"grad_norm": 0.5065655047968505,
"learning_rate": 3.636011825196365e-06,
"loss": 0.5836,
"step": 3140
},
{
"epoch": 1.4976190476190476,
"grad_norm": 0.517677275546858,
"learning_rate": 3.6040126394953334e-06,
"loss": 0.5794,
"step": 3145
},
{
"epoch": 1.5,
"grad_norm": 0.4918393882322778,
"learning_rate": 3.5721239031346067e-06,
"loss": 0.5894,
"step": 3150
},
{
"epoch": 1.5023809523809524,
"grad_norm": 0.5004981541944984,
"learning_rate": 3.540346166785994e-06,
"loss": 0.5731,
"step": 3155
},
{
"epoch": 1.5047619047619047,
"grad_norm": 0.5117157105186525,
"learning_rate": 3.5086799792044812e-06,
"loss": 0.5856,
"step": 3160
},
{
"epoch": 1.5071428571428571,
"grad_norm": 0.5239013800449769,
"learning_rate": 3.4771258872187917e-06,
"loss": 0.5784,
"step": 3165
},
{
"epoch": 1.5095238095238095,
"grad_norm": 0.499421090386738,
"learning_rate": 3.4456844357218977e-06,
"loss": 0.5853,
"step": 3170
},
{
"epoch": 1.5119047619047619,
"grad_norm": 0.49970944259407496,
"learning_rate": 3.414356167661658e-06,
"loss": 0.57,
"step": 3175
},
{
"epoch": 1.5142857142857142,
"grad_norm": 0.5258093843059876,
"learning_rate": 3.3831416240314085e-06,
"loss": 0.5761,
"step": 3180
},
{
"epoch": 1.5166666666666666,
"grad_norm": 0.5244353245311157,
"learning_rate": 3.3520413438606215e-06,
"loss": 0.5882,
"step": 3185
},
{
"epoch": 1.519047619047619,
"grad_norm": 0.4977521350042896,
"learning_rate": 3.3210558642056277e-06,
"loss": 0.5901,
"step": 3190
},
{
"epoch": 1.5214285714285714,
"grad_norm": 0.5122476763468953,
"learning_rate": 3.290185720140301e-06,
"loss": 0.5811,
"step": 3195
},
{
"epoch": 1.5238095238095237,
"grad_norm": 0.49340352912131785,
"learning_rate": 3.2594314447468457e-06,
"loss": 0.5754,
"step": 3200
},
{
"epoch": 1.526190476190476,
"grad_norm": 0.4874919671111681,
"learning_rate": 3.228793569106594e-06,
"loss": 0.5841,
"step": 3205
},
{
"epoch": 1.5285714285714285,
"grad_norm": 0.5324961393719349,
"learning_rate": 3.1982726222908046e-06,
"loss": 0.5727,
"step": 3210
},
{
"epoch": 1.5309523809523808,
"grad_norm": 0.4995381050758276,
"learning_rate": 3.1678691313515688e-06,
"loss": 0.579,
"step": 3215
},
{
"epoch": 1.5333333333333332,
"grad_norm": 0.49736016607701533,
"learning_rate": 3.1375836213126653e-06,
"loss": 0.5847,
"step": 3220
},
{
"epoch": 1.5357142857142856,
"grad_norm": 0.549492541961606,
"learning_rate": 3.10741661516053e-06,
"loss": 0.5814,
"step": 3225
},
{
"epoch": 1.538095238095238,
"grad_norm": 0.49389981419744866,
"learning_rate": 3.077368633835205e-06,
"loss": 0.5737,
"step": 3230
},
{
"epoch": 1.5404761904761903,
"grad_norm": 0.5564195244743101,
"learning_rate": 3.0474401962213483e-06,
"loss": 0.5769,
"step": 3235
},
{
"epoch": 1.5428571428571427,
"grad_norm": 0.5123987478065914,
"learning_rate": 3.017631819139273e-06,
"loss": 0.591,
"step": 3240
},
{
"epoch": 1.545238095238095,
"grad_norm": 0.5012262974081656,
"learning_rate": 2.987944017336023e-06,
"loss": 0.5888,
"step": 3245
},
{
"epoch": 1.5476190476190477,
"grad_norm": 0.5152768019089213,
"learning_rate": 2.958377303476483e-06,
"loss": 0.5702,
"step": 3250
},
{
"epoch": 1.55,
"grad_norm": 0.5066389868980488,
"learning_rate": 2.9289321881345257e-06,
"loss": 0.5868,
"step": 3255
},
{
"epoch": 1.5523809523809524,
"grad_norm": 0.5058336453434157,
"learning_rate": 2.8996091797841976e-06,
"loss": 0.5702,
"step": 3260
},
{
"epoch": 1.5547619047619048,
"grad_norm": 0.4984793009615145,
"learning_rate": 2.8704087847909333e-06,
"loss": 0.5662,
"step": 3265
},
{
"epoch": 1.5571428571428572,
"grad_norm": 0.5108314631589933,
"learning_rate": 2.8413315074028157e-06,
"loss": 0.5707,
"step": 3270
},
{
"epoch": 1.5595238095238095,
"grad_norm": 0.5031743279940161,
"learning_rate": 2.8123778497418687e-06,
"loss": 0.5872,
"step": 3275
},
{
"epoch": 1.561904761904762,
"grad_norm": 0.5140205415498605,
"learning_rate": 2.783548311795379e-06,
"loss": 0.5814,
"step": 3280
},
{
"epoch": 1.5642857142857143,
"grad_norm": 0.4951045700618797,
"learning_rate": 2.7548433914072736e-06,
"loss": 0.5688,
"step": 3285
},
{
"epoch": 1.5666666666666667,
"grad_norm": 0.4809510784828599,
"learning_rate": 2.726263584269513e-06,
"loss": 0.5822,
"step": 3290
},
{
"epoch": 1.569047619047619,
"grad_norm": 0.5144657250346224,
"learning_rate": 2.6978093839135365e-06,
"loss": 0.5924,
"step": 3295
},
{
"epoch": 1.5714285714285714,
"grad_norm": 0.5069069605136209,
"learning_rate": 2.669481281701739e-06,
"loss": 0.5808,
"step": 3300
},
{
"epoch": 1.5738095238095238,
"grad_norm": 0.5080243059955405,
"learning_rate": 2.641279766818977e-06,
"loss": 0.5826,
"step": 3305
},
{
"epoch": 1.5761904761904761,
"grad_norm": 0.5233910641374417,
"learning_rate": 2.6132053262641467e-06,
"loss": 0.5862,
"step": 3310
},
{
"epoch": 1.5785714285714287,
"grad_norm": 0.5114261441752712,
"learning_rate": 2.5852584448417327e-06,
"loss": 0.582,
"step": 3315
},
{
"epoch": 1.580952380952381,
"grad_norm": 0.5249361251608791,
"learning_rate": 2.5574396051534835e-06,
"loss": 0.5812,
"step": 3320
},
{
"epoch": 1.5833333333333335,
"grad_norm": 0.5014730717082388,
"learning_rate": 2.529749287590042e-06,
"loss": 0.5836,
"step": 3325
},
{
"epoch": 1.5857142857142859,
"grad_norm": 0.5031355239363027,
"learning_rate": 2.502187970322657e-06,
"loss": 0.5713,
"step": 3330
},
{
"epoch": 1.5880952380952382,
"grad_norm": 0.5267100809447453,
"learning_rate": 2.4747561292949496e-06,
"loss": 0.5871,
"step": 3335
},
{
"epoch": 1.5904761904761906,
"grad_norm": 0.5066033968112823,
"learning_rate": 2.447454238214654e-06,
"loss": 0.5777,
"step": 3340
},
{
"epoch": 1.592857142857143,
"grad_norm": 0.5249173659213104,
"learning_rate": 2.420282768545469e-06,
"loss": 0.5932,
"step": 3345
},
{
"epoch": 1.5952380952380953,
"grad_norm": 0.48114876503065274,
"learning_rate": 2.3932421894989167e-06,
"loss": 0.5733,
"step": 3350
},
{
"epoch": 1.5976190476190477,
"grad_norm": 0.48457566365078014,
"learning_rate": 2.366332968026207e-06,
"loss": 0.581,
"step": 3355
},
{
"epoch": 1.6,
"grad_norm": 0.49253983378100513,
"learning_rate": 2.339555568810221e-06,
"loss": 0.569,
"step": 3360
},
{
"epoch": 1.6023809523809525,
"grad_norm": 0.521391190119619,
"learning_rate": 2.3129104542574433e-06,
"loss": 0.575,
"step": 3365
},
{
"epoch": 1.6047619047619048,
"grad_norm": 0.4986688245603456,
"learning_rate": 2.2863980844900036e-06,
"loss": 0.5853,
"step": 3370
},
{
"epoch": 1.6071428571428572,
"grad_norm": 0.5270951766715323,
"learning_rate": 2.2600189173377263e-06,
"loss": 0.5898,
"step": 3375
},
{
"epoch": 1.6095238095238096,
"grad_norm": 0.46626058041434404,
"learning_rate": 2.2337734083302164e-06,
"loss": 0.584,
"step": 3380
},
{
"epoch": 1.611904761904762,
"grad_norm": 0.5313981173035822,
"learning_rate": 2.207662010689002e-06,
"loss": 0.5822,
"step": 3385
},
{
"epoch": 1.6142857142857143,
"grad_norm": 0.5153471281514648,
"learning_rate": 2.1816851753197023e-06,
"loss": 0.5797,
"step": 3390
},
{
"epoch": 1.6166666666666667,
"grad_norm": 0.5000763124278017,
"learning_rate": 2.155843350804243e-06,
"loss": 0.567,
"step": 3395
},
{
"epoch": 1.619047619047619,
"grad_norm": 0.49402952120811466,
"learning_rate": 2.130136983393112e-06,
"loss": 0.5697,
"step": 3400
},
{
"epoch": 1.6214285714285714,
"grad_norm": 0.48898166681541133,
"learning_rate": 2.104566516997647e-06,
"loss": 0.5779,
"step": 3405
},
{
"epoch": 1.6238095238095238,
"grad_norm": 0.5391774389924573,
"learning_rate": 2.0791323931823783e-06,
"loss": 0.5914,
"step": 3410
},
{
"epoch": 1.6261904761904762,
"grad_norm": 0.5080483298837424,
"learning_rate": 2.053835051157397e-06,
"loss": 0.5817,
"step": 3415
},
{
"epoch": 1.6285714285714286,
"grad_norm": 0.49302417797140397,
"learning_rate": 2.0286749277707783e-06,
"loss": 0.5647,
"step": 3420
},
{
"epoch": 1.630952380952381,
"grad_norm": 0.5241769643468105,
"learning_rate": 2.0036524575010176e-06,
"loss": 0.5853,
"step": 3425
},
{
"epoch": 1.6333333333333333,
"grad_norm": 0.5265692131855949,
"learning_rate": 1.9787680724495617e-06,
"loss": 0.5817,
"step": 3430
},
{
"epoch": 1.6357142857142857,
"grad_norm": 0.4902606126508354,
"learning_rate": 1.9540222023333165e-06,
"loss": 0.5807,
"step": 3435
},
{
"epoch": 1.638095238095238,
"grad_norm": 0.5237480696892753,
"learning_rate": 1.929415274477239e-06,
"loss": 0.5701,
"step": 3440
},
{
"epoch": 1.6404761904761904,
"grad_norm": 0.5357308687997984,
"learning_rate": 1.9049477138069606e-06,
"loss": 0.5804,
"step": 3445
},
{
"epoch": 1.6428571428571428,
"grad_norm": 0.5263899257687392,
"learning_rate": 1.880619942841435e-06,
"loss": 0.5883,
"step": 3450
},
{
"epoch": 1.6452380952380952,
"grad_norm": 0.5115574850941745,
"learning_rate": 1.856432381685669e-06,
"loss": 0.59,
"step": 3455
},
{
"epoch": 1.6476190476190475,
"grad_norm": 0.5173104409710929,
"learning_rate": 1.8323854480234348e-06,
"loss": 0.5805,
"step": 3460
},
{
"epoch": 1.65,
"grad_norm": 0.5011073792449935,
"learning_rate": 1.808479557110081e-06,
"loss": 0.5738,
"step": 3465
},
{
"epoch": 1.6523809523809523,
"grad_norm": 0.4933932876019805,
"learning_rate": 1.7847151217653624e-06,
"loss": 0.5722,
"step": 3470
},
{
"epoch": 1.6547619047619047,
"grad_norm": 0.5181601753202433,
"learning_rate": 1.7610925523662836e-06,
"loss": 0.5673,
"step": 3475
},
{
"epoch": 1.657142857142857,
"grad_norm": 0.5054495843864125,
"learning_rate": 1.7376122568400533e-06,
"loss": 0.5771,
"step": 3480
},
{
"epoch": 1.6595238095238094,
"grad_norm": 0.5277775744809468,
"learning_rate": 1.714274640657001e-06,
"loss": 0.5747,
"step": 3485
},
{
"epoch": 1.6619047619047618,
"grad_norm": 0.48735398585218814,
"learning_rate": 1.6910801068236015e-06,
"loss": 0.5847,
"step": 3490
},
{
"epoch": 1.6642857142857141,
"grad_norm": 0.5157125695595391,
"learning_rate": 1.6680290558755119e-06,
"loss": 0.5828,
"step": 3495
},
{
"epoch": 1.6666666666666665,
"grad_norm": 0.5089727599844109,
"learning_rate": 1.6451218858706374e-06,
"loss": 0.591,
"step": 3500
},
{
"epoch": 1.669047619047619,
"grad_norm": 0.5026463229681872,
"learning_rate": 1.6223589923822768e-06,
"loss": 0.5817,
"step": 3505
},
{
"epoch": 1.6714285714285713,
"grad_norm": 0.4931313873291854,
"learning_rate": 1.599740768492286e-06,
"loss": 0.5713,
"step": 3510
},
{
"epoch": 1.6738095238095239,
"grad_norm": 0.5181473464694085,
"learning_rate": 1.5772676047842862e-06,
"loss": 0.5759,
"step": 3515
},
{
"epoch": 1.6761904761904762,
"grad_norm": 0.4927232886574381,
"learning_rate": 1.5549398893369216e-06,
"loss": 0.5817,
"step": 3520
},
{
"epoch": 1.6785714285714286,
"grad_norm": 0.5220587931111694,
"learning_rate": 1.5327580077171589e-06,
"loss": 0.5782,
"step": 3525
},
{
"epoch": 1.680952380952381,
"grad_norm": 0.5333713690182812,
"learning_rate": 1.5107223429736273e-06,
"loss": 0.5883,
"step": 3530
},
{
"epoch": 1.6833333333333333,
"grad_norm": 0.5150308121762561,
"learning_rate": 1.4888332756300027e-06,
"loss": 0.586,
"step": 3535
},
{
"epoch": 1.6857142857142857,
"grad_norm": 0.4978412003341323,
"learning_rate": 1.467091183678444e-06,
"loss": 0.5872,
"step": 3540
},
{
"epoch": 1.688095238095238,
"grad_norm": 0.4880778638366708,
"learning_rate": 1.4454964425730533e-06,
"loss": 0.5708,
"step": 3545
},
{
"epoch": 1.6904761904761905,
"grad_norm": 0.5516107606480507,
"learning_rate": 1.424049425223405e-06,
"loss": 0.567,
"step": 3550
},
{
"epoch": 1.6928571428571428,
"grad_norm": 0.5693622851862525,
"learning_rate": 1.4027505019880972e-06,
"loss": 0.5776,
"step": 3555
},
{
"epoch": 1.6952380952380952,
"grad_norm": 0.5070794770825253,
"learning_rate": 1.3816000406683604e-06,
"loss": 0.5704,
"step": 3560
},
{
"epoch": 1.6976190476190476,
"grad_norm": 0.49091161185672627,
"learning_rate": 1.3605984065017074e-06,
"loss": 0.5645,
"step": 3565
},
{
"epoch": 1.7,
"grad_norm": 0.519545490092357,
"learning_rate": 1.339745962155613e-06,
"loss": 0.5866,
"step": 3570
},
{
"epoch": 1.7023809523809523,
"grad_norm": 0.5212093423730666,
"learning_rate": 1.3190430677212795e-06,
"loss": 0.5864,
"step": 3575
},
{
"epoch": 1.704761904761905,
"grad_norm": 0.531691412658331,
"learning_rate": 1.2984900807073919e-06,
"loss": 0.5789,
"step": 3580
},
{
"epoch": 1.7071428571428573,
"grad_norm": 0.5456532742307001,
"learning_rate": 1.278087356033947e-06,
"loss": 0.5722,
"step": 3585
},
{
"epoch": 1.7095238095238097,
"grad_norm": 0.5205925476159533,
"learning_rate": 1.2578352460261456e-06,
"loss": 0.5778,
"step": 3590
},
{
"epoch": 1.711904761904762,
"grad_norm": 0.5062569568474926,
"learning_rate": 1.2377341004082778e-06,
"loss": 0.5741,
"step": 3595
},
{
"epoch": 1.7142857142857144,
"grad_norm": 0.5175248723743612,
"learning_rate": 1.2177842662977136e-06,
"loss": 0.5785,
"step": 3600
},
{
"epoch": 1.7166666666666668,
"grad_norm": 0.5159262102883804,
"learning_rate": 1.1979860881988903e-06,
"loss": 0.5754,
"step": 3605
},
{
"epoch": 1.7190476190476192,
"grad_norm": 0.5083274127105409,
"learning_rate": 1.1783399079973578e-06,
"loss": 0.5842,
"step": 3610
},
{
"epoch": 1.7214285714285715,
"grad_norm": 0.5120999260869891,
"learning_rate": 1.1588460649539036e-06,
"loss": 0.5734,
"step": 3615
},
{
"epoch": 1.723809523809524,
"grad_norm": 0.49029779483860925,
"learning_rate": 1.1395048956986577e-06,
"loss": 0.5899,
"step": 3620
},
{
"epoch": 1.7261904761904763,
"grad_norm": 0.5065004286184338,
"learning_rate": 1.1203167342253063e-06,
"loss": 0.5728,
"step": 3625
},
{
"epoch": 1.7285714285714286,
"grad_norm": 0.5162711801738049,
"learning_rate": 1.1012819118853147e-06,
"loss": 0.57,
"step": 3630
},
{
"epoch": 1.730952380952381,
"grad_norm": 0.512398683483683,
"learning_rate": 1.0824007573822025e-06,
"loss": 0.573,
"step": 3635
},
{
"epoch": 1.7333333333333334,
"grad_norm": 0.5060773003944762,
"learning_rate": 1.0636735967658785e-06,
"loss": 0.5758,
"step": 3640
},
{
"epoch": 1.7357142857142858,
"grad_norm": 0.5719110243417397,
"learning_rate": 1.0451007534269908e-06,
"loss": 0.5706,
"step": 3645
},
{
"epoch": 1.7380952380952381,
"grad_norm": 0.5159140197103158,
"learning_rate": 1.026682548091361e-06,
"loss": 0.5787,
"step": 3650
},
{
"epoch": 1.7404761904761905,
"grad_norm": 0.5756689272489154,
"learning_rate": 1.0084192988144392e-06,
"loss": 0.589,
"step": 3655
},
{
"epoch": 1.7428571428571429,
"grad_norm": 0.4987087820127243,
"learning_rate": 9.903113209758098e-07,
"loss": 0.5713,
"step": 3660
},
{
"epoch": 1.7452380952380953,
"grad_norm": 0.49383352905457956,
"learning_rate": 9.723589272737443e-07,
"loss": 0.5692,
"step": 3665
},
{
"epoch": 1.7476190476190476,
"grad_norm": 0.5341671518005138,
"learning_rate": 9.545624277198085e-07,
"loss": 0.5653,
"step": 3670
},
{
"epoch": 1.75,
"grad_norm": 0.5264048094976569,
"learning_rate": 9.369221296335007e-07,
"loss": 0.5716,
"step": 3675
},
{
"epoch": 1.7523809523809524,
"grad_norm": 0.49117865624067225,
"learning_rate": 9.194383376369509e-07,
"loss": 0.5755,
"step": 3680
},
{
"epoch": 1.7547619047619047,
"grad_norm": 0.4802311797210391,
"learning_rate": 9.021113536496551e-07,
"loss": 0.5796,
"step": 3685
},
{
"epoch": 1.7571428571428571,
"grad_norm": 0.5170954646484791,
"learning_rate": 8.849414768832687e-07,
"loss": 0.5575,
"step": 3690
},
{
"epoch": 1.7595238095238095,
"grad_norm": 0.502792596764845,
"learning_rate": 8.679290038364319e-07,
"loss": 0.5833,
"step": 3695
},
{
"epoch": 1.7619047619047619,
"grad_norm": 0.5049868760014847,
"learning_rate": 8.510742282896545e-07,
"loss": 0.577,
"step": 3700
},
{
"epoch": 1.7642857142857142,
"grad_norm": 0.5037269478659716,
"learning_rate": 8.343774413002382e-07,
"loss": 0.5857,
"step": 3705
},
{
"epoch": 1.7666666666666666,
"grad_norm": 0.531756805530802,
"learning_rate": 8.178389311972612e-07,
"loss": 0.5706,
"step": 3710
},
{
"epoch": 1.769047619047619,
"grad_norm": 0.5354038647559776,
"learning_rate": 8.014589835765807e-07,
"loss": 0.5793,
"step": 3715
},
{
"epoch": 1.7714285714285714,
"grad_norm": 0.5165889882026443,
"learning_rate": 7.852378812959227e-07,
"loss": 0.565,
"step": 3720
},
{
"epoch": 1.7738095238095237,
"grad_norm": 0.5188950618246705,
"learning_rate": 7.69175904469982e-07,
"loss": 0.5681,
"step": 3725
},
{
"epoch": 1.776190476190476,
"grad_norm": 0.5219125713505919,
"learning_rate": 7.532733304655848e-07,
"loss": 0.5823,
"step": 3730
},
{
"epoch": 1.7785714285714285,
"grad_norm": 0.4842042459830578,
"learning_rate": 7.375304338969135e-07,
"loss": 0.5754,
"step": 3735
},
{
"epoch": 1.7809523809523808,
"grad_norm": 0.5028634955697662,
"learning_rate": 7.219474866207465e-07,
"loss": 0.5655,
"step": 3740
},
{
"epoch": 1.7833333333333332,
"grad_norm": 0.48467803772894236,
"learning_rate": 7.065247577317747e-07,
"loss": 0.5725,
"step": 3745
},
{
"epoch": 1.7857142857142856,
"grad_norm": 0.4918168156661026,
"learning_rate": 6.912625135579587e-07,
"loss": 0.5722,
"step": 3750
},
{
"epoch": 1.788095238095238,
"grad_norm": 0.48464787354503214,
"learning_rate": 6.761610176559086e-07,
"loss": 0.5589,
"step": 3755
},
{
"epoch": 1.7904761904761903,
"grad_norm": 0.5182220123762346,
"learning_rate": 6.612205308063646e-07,
"loss": 0.5684,
"step": 3760
},
{
"epoch": 1.7928571428571427,
"grad_norm": 0.48991260111946716,
"learning_rate": 6.464413110096601e-07,
"loss": 0.5778,
"step": 3765
},
{
"epoch": 1.795238095238095,
"grad_norm": 0.4667646370432281,
"learning_rate": 6.318236134812917e-07,
"loss": 0.5783,
"step": 3770
},
{
"epoch": 1.7976190476190477,
"grad_norm": 0.5127849630517298,
"learning_rate": 6.173676906475012e-07,
"loss": 0.5793,
"step": 3775
},
{
"epoch": 1.8,
"grad_norm": 0.5479174273289018,
"learning_rate": 6.030737921409169e-07,
"loss": 0.5789,
"step": 3780
},
{
"epoch": 1.8023809523809524,
"grad_norm": 0.5027002463972741,
"learning_rate": 5.889421647962456e-07,
"loss": 0.5844,
"step": 3785
},
{
"epoch": 1.8047619047619048,
"grad_norm": 0.47085642019543783,
"learning_rate": 5.749730526460073e-07,
"loss": 0.5675,
"step": 3790
},
{
"epoch": 1.8071428571428572,
"grad_norm": 0.5072363123110816,
"learning_rate": 5.611666969163243e-07,
"loss": 0.5775,
"step": 3795
},
{
"epoch": 1.8095238095238095,
"grad_norm": 0.4957008376046064,
"learning_rate": 5.475233360227516e-07,
"loss": 0.5811,
"step": 3800
},
{
"epoch": 1.811904761904762,
"grad_norm": 0.5212409865496715,
"learning_rate": 5.340432055661637e-07,
"loss": 0.5761,
"step": 3805
},
{
"epoch": 1.8142857142857143,
"grad_norm": 0.470997569919902,
"learning_rate": 5.207265383286831e-07,
"loss": 0.5727,
"step": 3810
},
{
"epoch": 1.8166666666666667,
"grad_norm": 0.5180739672868397,
"learning_rate": 5.075735642696611e-07,
"loss": 0.5655,
"step": 3815
},
{
"epoch": 1.819047619047619,
"grad_norm": 0.4888611854375848,
"learning_rate": 4.945845105217118e-07,
"loss": 0.5642,
"step": 3820
},
{
"epoch": 1.8214285714285714,
"grad_norm": 0.5177332953584531,
"learning_rate": 4.817596013867765e-07,
"loss": 0.5845,
"step": 3825
},
{
"epoch": 1.8238095238095238,
"grad_norm": 0.5279903730811234,
"learning_rate": 4.6909905833226965e-07,
"loss": 0.5775,
"step": 3830
},
{
"epoch": 1.8261904761904761,
"grad_norm": 0.5151201728696789,
"learning_rate": 4.566030999872384e-07,
"loss": 0.5863,
"step": 3835
},
{
"epoch": 1.8285714285714287,
"grad_norm": 0.5216946630712812,
"learning_rate": 4.4427194213859216e-07,
"loss": 0.5824,
"step": 3840
},
{
"epoch": 1.830952380952381,
"grad_norm": 0.48997634713435234,
"learning_rate": 4.3210579772738237e-07,
"loss": 0.5633,
"step": 3845
},
{
"epoch": 1.8333333333333335,
"grad_norm": 0.5266753228806161,
"learning_rate": 4.2010487684511105e-07,
"loss": 0.5861,
"step": 3850
},
{
"epoch": 1.8357142857142859,
"grad_norm": 0.4963390005103402,
"learning_rate": 4.082693867301224e-07,
"loss": 0.5866,
"step": 3855
},
{
"epoch": 1.8380952380952382,
"grad_norm": 0.5075963559582308,
"learning_rate": 3.965995317640026e-07,
"loss": 0.566,
"step": 3860
},
{
"epoch": 1.8404761904761906,
"grad_norm": 0.5210567644268166,
"learning_rate": 3.850955134680678e-07,
"loss": 0.576,
"step": 3865
},
{
"epoch": 1.842857142857143,
"grad_norm": 0.521115702642029,
"learning_rate": 3.7375753049987974e-07,
"loss": 0.5695,
"step": 3870
},
{
"epoch": 1.8452380952380953,
"grad_norm": 0.5060214448757876,
"learning_rate": 3.625857786498055e-07,
"loss": 0.5704,
"step": 3875
},
{
"epoch": 1.8476190476190477,
"grad_norm": 0.4875613223286847,
"learning_rate": 3.515804508376508e-07,
"loss": 0.5697,
"step": 3880
},
{
"epoch": 1.85,
"grad_norm": 0.4832852448893152,
"learning_rate": 3.4074173710931804e-07,
"loss": 0.5768,
"step": 3885
},
{
"epoch": 1.8523809523809525,
"grad_norm": 0.5331240685153525,
"learning_rate": 3.3006982463352764e-07,
"loss": 0.5737,
"step": 3890
},
{
"epoch": 1.8547619047619048,
"grad_norm": 0.5001592000344391,
"learning_rate": 3.1956489769859213e-07,
"loss": 0.572,
"step": 3895
},
{
"epoch": 1.8571428571428572,
"grad_norm": 0.48713909339424555,
"learning_rate": 3.0922713770922155e-07,
"loss": 0.5842,
"step": 3900
},
{
"epoch": 1.8595238095238096,
"grad_norm": 0.505832290948494,
"learning_rate": 2.9905672318339963e-07,
"loss": 0.5794,
"step": 3905
},
{
"epoch": 1.861904761904762,
"grad_norm": 0.5001487285158832,
"learning_rate": 2.8905382974930173e-07,
"loss": 0.5797,
"step": 3910
},
{
"epoch": 1.8642857142857143,
"grad_norm": 0.5134685274388742,
"learning_rate": 2.7921863014225504e-07,
"loss": 0.5938,
"step": 3915
},
{
"epoch": 1.8666666666666667,
"grad_norm": 0.48901171207964383,
"learning_rate": 2.6955129420176193e-07,
"loss": 0.5804,
"step": 3920
},
{
"epoch": 1.869047619047619,
"grad_norm": 0.4928766933172695,
"learning_rate": 2.6005198886856486e-07,
"loss": 0.5738,
"step": 3925
},
{
"epoch": 1.8714285714285714,
"grad_norm": 0.4929640596310142,
"learning_rate": 2.507208781817638e-07,
"loss": 0.5628,
"step": 3930
},
{
"epoch": 1.8738095238095238,
"grad_norm": 0.5172714992148169,
"learning_rate": 2.4155812327598337e-07,
"loss": 0.5649,
"step": 3935
},
{
"epoch": 1.8761904761904762,
"grad_norm": 0.5022968661364299,
"learning_rate": 2.3256388237858806e-07,
"loss": 0.5681,
"step": 3940
},
{
"epoch": 1.8785714285714286,
"grad_norm": 0.5232052126416001,
"learning_rate": 2.2373831080695463e-07,
"loss": 0.5745,
"step": 3945
},
{
"epoch": 1.880952380952381,
"grad_norm": 0.488579854673118,
"learning_rate": 2.1508156096578748e-07,
"loss": 0.5729,
"step": 3950
},
{
"epoch": 1.8833333333333333,
"grad_norm": 0.5155005458528391,
"learning_rate": 2.0659378234448524e-07,
"loss": 0.5732,
"step": 3955
},
{
"epoch": 1.8857142857142857,
"grad_norm": 0.5024985648134219,
"learning_rate": 1.9827512151456175e-07,
"loss": 0.5627,
"step": 3960
},
{
"epoch": 1.888095238095238,
"grad_norm": 0.5210350946419919,
"learning_rate": 1.9012572212711467e-07,
"loss": 0.5766,
"step": 3965
},
{
"epoch": 1.8904761904761904,
"grad_norm": 0.505351224359543,
"learning_rate": 1.82145724910342e-07,
"loss": 0.5787,
"step": 3970
},
{
"epoch": 1.8928571428571428,
"grad_norm": 0.5042030458059477,
"learning_rate": 1.7433526766711727e-07,
"loss": 0.5756,
"step": 3975
},
{
"epoch": 1.8952380952380952,
"grad_norm": 0.5196433352960352,
"learning_rate": 1.6669448527260602e-07,
"loss": 0.5787,
"step": 3980
},
{
"epoch": 1.8976190476190475,
"grad_norm": 0.5042961399463299,
"learning_rate": 1.5922350967193524e-07,
"loss": 0.5828,
"step": 3985
},
{
"epoch": 1.9,
"grad_norm": 0.49509327596749514,
"learning_rate": 1.519224698779198e-07,
"loss": 0.5666,
"step": 3990
},
{
"epoch": 1.9023809523809523,
"grad_norm": 0.4993392687429765,
"learning_rate": 1.447914919688298e-07,
"loss": 0.5862,
"step": 3995
},
{
"epoch": 1.9047619047619047,
"grad_norm": 0.5302407636844851,
"learning_rate": 1.3783069908621772e-07,
"loss": 0.5711,
"step": 4000
},
{
"epoch": 1.907142857142857,
"grad_norm": 0.49474811360911164,
"learning_rate": 1.3104021143278911e-07,
"loss": 0.5876,
"step": 4005
},
{
"epoch": 1.9095238095238094,
"grad_norm": 0.497576228240754,
"learning_rate": 1.2442014627032318e-07,
"loss": 0.5744,
"step": 4010
},
{
"epoch": 1.9119047619047618,
"grad_norm": 0.4871944438237385,
"learning_rate": 1.1797061791766207e-07,
"loss": 0.5714,
"step": 4015
},
{
"epoch": 1.9142857142857141,
"grad_norm": 0.4830245999351171,
"learning_rate": 1.1169173774871478e-07,
"loss": 0.5686,
"step": 4020
},
{
"epoch": 1.9166666666666665,
"grad_norm": 0.5413142563859302,
"learning_rate": 1.055836141905553e-07,
"loss": 0.5888,
"step": 4025
},
{
"epoch": 1.919047619047619,
"grad_norm": 0.46941478343612175,
"learning_rate": 9.964635272153633e-08,
"loss": 0.5786,
"step": 4030
},
{
"epoch": 1.9214285714285713,
"grad_norm": 0.5140729728882855,
"learning_rate": 9.388005586947191e-08,
"loss": 0.565,
"step": 4035
},
{
"epoch": 1.9238095238095239,
"grad_norm": 0.4991452996322733,
"learning_rate": 8.82848232098732e-08,
"loss": 0.5708,
"step": 4040
},
{
"epoch": 1.9261904761904762,
"grad_norm": 0.505925510646528,
"learning_rate": 8.286075136421435e-08,
"loss": 0.5814,
"step": 4045
},
{
"epoch": 1.9285714285714286,
"grad_norm": 0.5275663420688306,
"learning_rate": 7.760793399827937e-08,
"loss": 0.5635,
"step": 4050
},
{
"epoch": 1.930952380952381,
"grad_norm": 0.5049137853663558,
"learning_rate": 7.25264618205357e-08,
"loss": 0.5547,
"step": 4055
},
{
"epoch": 1.9333333333333333,
"grad_norm": 0.4878320213926508,
"learning_rate": 6.761642258056977e-08,
"loss": 0.5695,
"step": 4060
},
{
"epoch": 1.9357142857142857,
"grad_norm": 0.5081138801093147,
"learning_rate": 6.287790106757396e-08,
"loss": 0.5683,
"step": 4065
},
{
"epoch": 1.938095238095238,
"grad_norm": 0.4943589564870212,
"learning_rate": 5.831097910887873e-08,
"loss": 0.5658,
"step": 4070
},
{
"epoch": 1.9404761904761905,
"grad_norm": 0.5302217136380152,
"learning_rate": 5.391573556854157e-08,
"loss": 0.5678,
"step": 4075
},
{
"epoch": 1.9428571428571428,
"grad_norm": 0.5038390499507364,
"learning_rate": 4.9692246345985905e-08,
"loss": 0.5702,
"step": 4080
},
{
"epoch": 1.9452380952380952,
"grad_norm": 0.5151211498515915,
"learning_rate": 4.564058437468877e-08,
"loss": 0.5784,
"step": 4085
},
{
"epoch": 1.9476190476190476,
"grad_norm": 0.4789745893173457,
"learning_rate": 4.176081962092182e-08,
"loss": 0.5749,
"step": 4090
},
{
"epoch": 1.95,
"grad_norm": 0.4997223839030282,
"learning_rate": 3.805301908254455e-08,
"loss": 0.5871,
"step": 4095
},
{
"epoch": 1.9523809523809523,
"grad_norm": 0.49934046802623616,
"learning_rate": 3.451724678784518e-08,
"loss": 0.561,
"step": 4100
},
{
"epoch": 1.954761904761905,
"grad_norm": 0.4974229497151692,
"learning_rate": 3.115356379443601e-08,
"loss": 0.5661,
"step": 4105
},
{
"epoch": 1.9571428571428573,
"grad_norm": 0.5196515473348364,
"learning_rate": 2.796202818819871e-08,
"loss": 0.5856,
"step": 4110
},
{
"epoch": 1.9595238095238097,
"grad_norm": 0.49970636623212594,
"learning_rate": 2.4942695082281752e-08,
"loss": 0.5797,
"step": 4115
},
{
"epoch": 1.961904761904762,
"grad_norm": 0.5173329159698928,
"learning_rate": 2.2095616616150117e-08,
"loss": 0.5747,
"step": 4120
},
{
"epoch": 1.9642857142857144,
"grad_norm": 0.49959590807990745,
"learning_rate": 1.9420841954681525e-08,
"loss": 0.5646,
"step": 4125
},
{
"epoch": 1.9666666666666668,
"grad_norm": 0.48412036634259087,
"learning_rate": 1.6918417287318245e-08,
"loss": 0.5777,
"step": 4130
},
{
"epoch": 1.9690476190476192,
"grad_norm": 0.50340974957607,
"learning_rate": 1.4588385827272178e-08,
"loss": 0.5762,
"step": 4135
},
{
"epoch": 1.9714285714285715,
"grad_norm": 0.5095137173840767,
"learning_rate": 1.2430787810776556e-08,
"loss": 0.5779,
"step": 4140
},
{
"epoch": 1.973809523809524,
"grad_norm": 0.5201276330625599,
"learning_rate": 1.0445660496390952e-08,
"loss": 0.5829,
"step": 4145
},
{
"epoch": 1.9761904761904763,
"grad_norm": 0.4957082874285413,
"learning_rate": 8.633038164358454e-09,
"loss": 0.5782,
"step": 4150
},
{
"epoch": 1.9785714285714286,
"grad_norm": 0.49005045720787455,
"learning_rate": 6.992952116013918e-09,
"loss": 0.5769,
"step": 4155
},
{
"epoch": 1.980952380952381,
"grad_norm": 0.5163548919624951,
"learning_rate": 5.525430673244403e-09,
"loss": 0.574,
"step": 4160
},
{
"epoch": 1.9833333333333334,
"grad_norm": 0.5289724215651535,
"learning_rate": 4.230499177994007e-09,
"loss": 0.5773,
"step": 4165
},
{
"epoch": 1.9857142857142858,
"grad_norm": 0.485812450600075,
"learning_rate": 3.1081799918375454e-09,
"loss": 0.5835,
"step": 4170
},
{
"epoch": 1.9880952380952381,
"grad_norm": 0.48465992123370755,
"learning_rate": 2.1584924955819763e-09,
"loss": 0.5851,
"step": 4175
},
{
"epoch": 1.9904761904761905,
"grad_norm": 0.5073868341064405,
"learning_rate": 1.3814530889433298e-09,
"loss": 0.5656,
"step": 4180
},
{
"epoch": 1.9928571428571429,
"grad_norm": 0.4822167121104924,
"learning_rate": 7.770751902513862e-10,
"loss": 0.5766,
"step": 4185
},
{
"epoch": 1.9952380952380953,
"grad_norm": 0.5135567153768665,
"learning_rate": 3.4536923623096353e-10,
"loss": 0.5891,
"step": 4190
},
{
"epoch": 1.9976190476190476,
"grad_norm": 0.5058535961206515,
"learning_rate": 8.634268181095806e-11,
"loss": 0.5653,
"step": 4195
},
{
"epoch": 2.0,
"grad_norm": 0.5344236072743405,
"learning_rate": 0.0,
"loss": 0.56,
"step": 4200
},
{
"epoch": 2.0,
"eval_loss": 0.741725742816925,
"eval_runtime": 110.7755,
"eval_samples_per_second": 77.282,
"eval_steps_per_second": 1.21,
"step": 4200
},
{
"epoch": 2.0,
"step": 4200,
"total_flos": 879394553856000.0,
"train_loss": 0.6596568237599872,
"train_runtime": 12326.0183,
"train_samples_per_second": 21.801,
"train_steps_per_second": 0.341
}
],
"logging_steps": 5,
"max_steps": 4200,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": false,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 879394553856000.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}