cgihlstorf's picture
Upload 8 files
47c69f2 verified
{
"best_metric": 0.2655787765979767,
"best_model_checkpoint": "/scratch/czm5kz/llama2-7b_8_100_0.0003_sg_finetuned_combined/checkpoint-1680",
"epoch": 100.0,
"eval_steps": 20,
"global_step": 1700,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.29,
"grad_norm": 2.5130302906036377,
"learning_rate": 0.0002994705882352941,
"loss": 5.7234,
"step": 5
},
{
"epoch": 0.59,
"grad_norm": 2.1954381465911865,
"learning_rate": 0.0002985882352941176,
"loss": 4.7563,
"step": 10
},
{
"epoch": 0.88,
"grad_norm": 2.519700050354004,
"learning_rate": 0.0002978823529411764,
"loss": 4.2519,
"step": 15
},
{
"epoch": 1.18,
"grad_norm": 2.1008927822113037,
"learning_rate": 0.00029699999999999996,
"loss": 3.8694,
"step": 20
},
{
"epoch": 1.18,
"eval_loss": 3.4059982299804688,
"eval_runtime": 1.9717,
"eval_samples_per_second": 67.962,
"eval_steps_per_second": 8.622,
"step": 20
},
{
"epoch": 1.47,
"grad_norm": 1.9939683675765991,
"learning_rate": 0.0002961176470588235,
"loss": 3.0042,
"step": 25
},
{
"epoch": 1.76,
"grad_norm": 1.9956856966018677,
"learning_rate": 0.00029523529411764704,
"loss": 2.7303,
"step": 30
},
{
"epoch": 2.06,
"grad_norm": 2.552236318588257,
"learning_rate": 0.0002943529411764706,
"loss": 2.7002,
"step": 35
},
{
"epoch": 2.35,
"grad_norm": 2.5456089973449707,
"learning_rate": 0.00029347058823529413,
"loss": 2.3598,
"step": 40
},
{
"epoch": 2.35,
"eval_loss": 2.0730819702148438,
"eval_runtime": 1.9745,
"eval_samples_per_second": 67.864,
"eval_steps_per_second": 8.61,
"step": 40
},
{
"epoch": 2.65,
"grad_norm": 2.949130058288574,
"learning_rate": 0.0002925882352941176,
"loss": 1.9988,
"step": 45
},
{
"epoch": 2.94,
"grad_norm": 2.44789457321167,
"learning_rate": 0.00029170588235294116,
"loss": 1.9843,
"step": 50
},
{
"epoch": 3.24,
"grad_norm": 5.215721130371094,
"learning_rate": 0.00029082352941176465,
"loss": 1.5394,
"step": 55
},
{
"epoch": 3.53,
"grad_norm": 3.831580638885498,
"learning_rate": 0.0002899411764705882,
"loss": 1.5503,
"step": 60
},
{
"epoch": 3.53,
"eval_loss": 1.2807282209396362,
"eval_runtime": 1.9725,
"eval_samples_per_second": 67.933,
"eval_steps_per_second": 8.618,
"step": 60
},
{
"epoch": 3.82,
"grad_norm": 3.5273172855377197,
"learning_rate": 0.00028905882352941173,
"loss": 1.4236,
"step": 65
},
{
"epoch": 4.12,
"grad_norm": 4.674720287322998,
"learning_rate": 0.0002881764705882353,
"loss": 1.1729,
"step": 70
},
{
"epoch": 4.41,
"grad_norm": 5.43109655380249,
"learning_rate": 0.0002872941176470588,
"loss": 0.9213,
"step": 75
},
{
"epoch": 4.71,
"grad_norm": 6.3855814933776855,
"learning_rate": 0.0002864117647058823,
"loss": 0.9106,
"step": 80
},
{
"epoch": 4.71,
"eval_loss": 0.7538219094276428,
"eval_runtime": 1.9756,
"eval_samples_per_second": 67.827,
"eval_steps_per_second": 8.605,
"step": 80
},
{
"epoch": 5.0,
"grad_norm": 5.1275529861450195,
"learning_rate": 0.00028552941176470585,
"loss": 1.0219,
"step": 85
},
{
"epoch": 5.29,
"grad_norm": 3.2476627826690674,
"learning_rate": 0.0002846470588235294,
"loss": 0.6821,
"step": 90
},
{
"epoch": 5.59,
"grad_norm": 3.561427116394043,
"learning_rate": 0.00028376470588235294,
"loss": 0.6022,
"step": 95
},
{
"epoch": 5.88,
"grad_norm": 4.358293056488037,
"learning_rate": 0.0002828823529411765,
"loss": 0.6565,
"step": 100
},
{
"epoch": 5.88,
"eval_loss": 0.5009134411811829,
"eval_runtime": 1.9816,
"eval_samples_per_second": 67.622,
"eval_steps_per_second": 8.579,
"step": 100
},
{
"epoch": 6.18,
"grad_norm": 3.2790002822875977,
"learning_rate": 0.00028199999999999997,
"loss": 0.6397,
"step": 105
},
{
"epoch": 6.47,
"grad_norm": 2.5752627849578857,
"learning_rate": 0.0002811176470588235,
"loss": 0.4518,
"step": 110
},
{
"epoch": 6.76,
"grad_norm": 3.419771909713745,
"learning_rate": 0.000280235294117647,
"loss": 0.4741,
"step": 115
},
{
"epoch": 7.06,
"grad_norm": 2.2088027000427246,
"learning_rate": 0.00027935294117647054,
"loss": 0.5214,
"step": 120
},
{
"epoch": 7.06,
"eval_loss": 0.40768566727638245,
"eval_runtime": 1.9831,
"eval_samples_per_second": 67.57,
"eval_steps_per_second": 8.572,
"step": 120
},
{
"epoch": 7.35,
"grad_norm": 2.9267308712005615,
"learning_rate": 0.0002784705882352941,
"loss": 0.4434,
"step": 125
},
{
"epoch": 7.65,
"grad_norm": 3.3790996074676514,
"learning_rate": 0.00027758823529411763,
"loss": 0.4631,
"step": 130
},
{
"epoch": 7.94,
"grad_norm": 2.3158915042877197,
"learning_rate": 0.0002767058823529412,
"loss": 0.4464,
"step": 135
},
{
"epoch": 8.24,
"grad_norm": 2.231661081314087,
"learning_rate": 0.00027582352941176466,
"loss": 0.4275,
"step": 140
},
{
"epoch": 8.24,
"eval_loss": 0.3691316843032837,
"eval_runtime": 1.9948,
"eval_samples_per_second": 67.174,
"eval_steps_per_second": 8.522,
"step": 140
},
{
"epoch": 8.53,
"grad_norm": 2.341855525970459,
"learning_rate": 0.0002749411764705882,
"loss": 0.3938,
"step": 145
},
{
"epoch": 8.82,
"grad_norm": 1.6353213787078857,
"learning_rate": 0.00027405882352941175,
"loss": 0.4794,
"step": 150
},
{
"epoch": 9.12,
"grad_norm": 1.7900513410568237,
"learning_rate": 0.0002731764705882353,
"loss": 0.4073,
"step": 155
},
{
"epoch": 9.41,
"grad_norm": 2.082562208175659,
"learning_rate": 0.0002722941176470588,
"loss": 0.3793,
"step": 160
},
{
"epoch": 9.41,
"eval_loss": 0.3403892517089844,
"eval_runtime": 1.988,
"eval_samples_per_second": 67.405,
"eval_steps_per_second": 8.551,
"step": 160
},
{
"epoch": 9.71,
"grad_norm": 1.9803985357284546,
"learning_rate": 0.0002714117647058823,
"loss": 0.3943,
"step": 165
},
{
"epoch": 10.0,
"grad_norm": 1.6986087560653687,
"learning_rate": 0.00027052941176470587,
"loss": 0.4254,
"step": 170
},
{
"epoch": 10.29,
"grad_norm": 1.120859146118164,
"learning_rate": 0.0002696470588235294,
"loss": 0.3333,
"step": 175
},
{
"epoch": 10.59,
"grad_norm": 2.3081905841827393,
"learning_rate": 0.0002687647058823529,
"loss": 0.3671,
"step": 180
},
{
"epoch": 10.59,
"eval_loss": 0.32923150062561035,
"eval_runtime": 1.9871,
"eval_samples_per_second": 67.435,
"eval_steps_per_second": 8.555,
"step": 180
},
{
"epoch": 10.88,
"grad_norm": 3.328174352645874,
"learning_rate": 0.00026788235294117644,
"loss": 0.3888,
"step": 185
},
{
"epoch": 11.18,
"grad_norm": 2.408024549484253,
"learning_rate": 0.000267,
"loss": 0.3964,
"step": 190
},
{
"epoch": 11.47,
"grad_norm": 1.6264761686325073,
"learning_rate": 0.00026611764705882353,
"loss": 0.3579,
"step": 195
},
{
"epoch": 11.76,
"grad_norm": 1.857405662536621,
"learning_rate": 0.00026523529411764707,
"loss": 0.375,
"step": 200
},
{
"epoch": 11.76,
"eval_loss": 0.3141331374645233,
"eval_runtime": 1.9899,
"eval_samples_per_second": 67.341,
"eval_steps_per_second": 8.543,
"step": 200
},
{
"epoch": 12.06,
"grad_norm": 1.0782090425491333,
"learning_rate": 0.00026435294117647056,
"loss": 0.3808,
"step": 205
},
{
"epoch": 12.35,
"grad_norm": 1.4318054914474487,
"learning_rate": 0.0002634705882352941,
"loss": 0.3148,
"step": 210
},
{
"epoch": 12.65,
"grad_norm": 1.6774518489837646,
"learning_rate": 0.0002625882352941176,
"loss": 0.3729,
"step": 215
},
{
"epoch": 12.94,
"grad_norm": 1.2252236604690552,
"learning_rate": 0.00026170588235294113,
"loss": 0.379,
"step": 220
},
{
"epoch": 12.94,
"eval_loss": 0.3084084689617157,
"eval_runtime": 1.9898,
"eval_samples_per_second": 67.342,
"eval_steps_per_second": 8.543,
"step": 220
},
{
"epoch": 13.24,
"grad_norm": 0.9840493202209473,
"learning_rate": 0.0002608235294117647,
"loss": 0.3292,
"step": 225
},
{
"epoch": 13.53,
"grad_norm": 2.3673970699310303,
"learning_rate": 0.0002599411764705882,
"loss": 0.3803,
"step": 230
},
{
"epoch": 13.82,
"grad_norm": 1.2279037237167358,
"learning_rate": 0.00025905882352941176,
"loss": 0.3444,
"step": 235
},
{
"epoch": 14.12,
"grad_norm": 0.928492546081543,
"learning_rate": 0.00025817647058823525,
"loss": 0.334,
"step": 240
},
{
"epoch": 14.12,
"eval_loss": 0.30614903569221497,
"eval_runtime": 1.9906,
"eval_samples_per_second": 67.317,
"eval_steps_per_second": 8.54,
"step": 240
},
{
"epoch": 14.41,
"grad_norm": 1.6385815143585205,
"learning_rate": 0.0002572941176470588,
"loss": 0.309,
"step": 245
},
{
"epoch": 14.71,
"grad_norm": 0.9644414782524109,
"learning_rate": 0.00025641176470588234,
"loss": 0.3318,
"step": 250
},
{
"epoch": 15.0,
"grad_norm": 1.3966537714004517,
"learning_rate": 0.0002555294117647059,
"loss": 0.3974,
"step": 255
},
{
"epoch": 15.29,
"grad_norm": 1.0618793964385986,
"learning_rate": 0.0002546470588235294,
"loss": 0.3027,
"step": 260
},
{
"epoch": 15.29,
"eval_loss": 0.29958730936050415,
"eval_runtime": 1.9894,
"eval_samples_per_second": 67.357,
"eval_steps_per_second": 8.545,
"step": 260
},
{
"epoch": 15.59,
"grad_norm": 1.31105637550354,
"learning_rate": 0.0002537647058823529,
"loss": 0.3196,
"step": 265
},
{
"epoch": 15.88,
"grad_norm": 1.4833605289459229,
"learning_rate": 0.00025288235294117646,
"loss": 0.3485,
"step": 270
},
{
"epoch": 16.18,
"grad_norm": 0.8568434119224548,
"learning_rate": 0.00025199999999999995,
"loss": 0.3391,
"step": 275
},
{
"epoch": 16.47,
"grad_norm": 1.2687785625457764,
"learning_rate": 0.0002511176470588235,
"loss": 0.3179,
"step": 280
},
{
"epoch": 16.47,
"eval_loss": 0.29810982942581177,
"eval_runtime": 1.9921,
"eval_samples_per_second": 67.267,
"eval_steps_per_second": 8.534,
"step": 280
},
{
"epoch": 16.76,
"grad_norm": 1.125921368598938,
"learning_rate": 0.00025023529411764703,
"loss": 0.3231,
"step": 285
},
{
"epoch": 17.06,
"grad_norm": 0.751727819442749,
"learning_rate": 0.0002493529411764706,
"loss": 0.349,
"step": 290
},
{
"epoch": 17.35,
"grad_norm": 1.0114847421646118,
"learning_rate": 0.0002484705882352941,
"loss": 0.3027,
"step": 295
},
{
"epoch": 17.65,
"grad_norm": 2.5540404319763184,
"learning_rate": 0.0002475882352941176,
"loss": 0.3021,
"step": 300
},
{
"epoch": 17.65,
"eval_loss": 0.29670611023902893,
"eval_runtime": 2.0093,
"eval_samples_per_second": 66.689,
"eval_steps_per_second": 8.461,
"step": 300
},
{
"epoch": 17.94,
"grad_norm": 0.9592204689979553,
"learning_rate": 0.00024670588235294115,
"loss": 0.3553,
"step": 305
},
{
"epoch": 18.24,
"grad_norm": 0.6442326307296753,
"learning_rate": 0.0002458235294117647,
"loss": 0.3101,
"step": 310
},
{
"epoch": 18.53,
"grad_norm": 1.0425697565078735,
"learning_rate": 0.00024494117647058824,
"loss": 0.3155,
"step": 315
},
{
"epoch": 18.82,
"grad_norm": 3.3176205158233643,
"learning_rate": 0.00024405882352941172,
"loss": 0.3634,
"step": 320
},
{
"epoch": 18.82,
"eval_loss": 0.2934125065803528,
"eval_runtime": 1.9921,
"eval_samples_per_second": 67.267,
"eval_steps_per_second": 8.534,
"step": 320
},
{
"epoch": 19.12,
"grad_norm": 0.5256841778755188,
"learning_rate": 0.00024317647058823527,
"loss": 0.3197,
"step": 325
},
{
"epoch": 19.41,
"grad_norm": 0.7443175315856934,
"learning_rate": 0.00024229411764705878,
"loss": 0.3117,
"step": 330
},
{
"epoch": 19.71,
"grad_norm": 0.8534731864929199,
"learning_rate": 0.00024141176470588233,
"loss": 0.322,
"step": 335
},
{
"epoch": 20.0,
"grad_norm": 1.1688783168792725,
"learning_rate": 0.00024052941176470587,
"loss": 0.3293,
"step": 340
},
{
"epoch": 20.0,
"eval_loss": 0.2844869792461395,
"eval_runtime": 1.9933,
"eval_samples_per_second": 67.224,
"eval_steps_per_second": 8.528,
"step": 340
},
{
"epoch": 20.29,
"grad_norm": 0.8509626388549805,
"learning_rate": 0.00023964705882352939,
"loss": 0.2899,
"step": 345
},
{
"epoch": 20.59,
"grad_norm": 1.0683010816574097,
"learning_rate": 0.00023876470588235293,
"loss": 0.3368,
"step": 350
},
{
"epoch": 20.88,
"grad_norm": 0.8359882235527039,
"learning_rate": 0.00023788235294117644,
"loss": 0.3189,
"step": 355
},
{
"epoch": 21.18,
"grad_norm": 0.746294379234314,
"learning_rate": 0.000237,
"loss": 0.2907,
"step": 360
},
{
"epoch": 21.18,
"eval_loss": 0.28352469205856323,
"eval_runtime": 1.9943,
"eval_samples_per_second": 67.191,
"eval_steps_per_second": 8.524,
"step": 360
},
{
"epoch": 21.47,
"grad_norm": 0.6525380611419678,
"learning_rate": 0.00023611764705882353,
"loss": 0.3088,
"step": 365
},
{
"epoch": 21.76,
"grad_norm": 0.7996423840522766,
"learning_rate": 0.00023523529411764702,
"loss": 0.3241,
"step": 370
},
{
"epoch": 22.06,
"grad_norm": 0.6835091710090637,
"learning_rate": 0.00023435294117647056,
"loss": 0.3078,
"step": 375
},
{
"epoch": 22.35,
"grad_norm": 0.7048420310020447,
"learning_rate": 0.00023347058823529408,
"loss": 0.3098,
"step": 380
},
{
"epoch": 22.35,
"eval_loss": 0.28362563252449036,
"eval_runtime": 1.9954,
"eval_samples_per_second": 67.156,
"eval_steps_per_second": 8.52,
"step": 380
},
{
"epoch": 22.65,
"grad_norm": 0.7149125337600708,
"learning_rate": 0.00023258823529411762,
"loss": 0.2859,
"step": 385
},
{
"epoch": 22.94,
"grad_norm": 0.8284432291984558,
"learning_rate": 0.00023170588235294116,
"loss": 0.3193,
"step": 390
},
{
"epoch": 23.24,
"grad_norm": 0.7711451053619385,
"learning_rate": 0.00023082352941176468,
"loss": 0.2894,
"step": 395
},
{
"epoch": 23.53,
"grad_norm": 0.7395182847976685,
"learning_rate": 0.00022994117647058822,
"loss": 0.2911,
"step": 400
},
{
"epoch": 23.53,
"eval_loss": 0.28144964575767517,
"eval_runtime": 1.9967,
"eval_samples_per_second": 67.11,
"eval_steps_per_second": 8.514,
"step": 400
},
{
"epoch": 23.82,
"grad_norm": 0.694525957107544,
"learning_rate": 0.00022905882352941174,
"loss": 0.3048,
"step": 405
},
{
"epoch": 24.12,
"grad_norm": 0.6153783798217773,
"learning_rate": 0.00022817647058823528,
"loss": 0.2975,
"step": 410
},
{
"epoch": 24.41,
"grad_norm": 0.7303423285484314,
"learning_rate": 0.00022729411764705883,
"loss": 0.2838,
"step": 415
},
{
"epoch": 24.71,
"grad_norm": 0.6618799567222595,
"learning_rate": 0.00022641176470588234,
"loss": 0.326,
"step": 420
},
{
"epoch": 24.71,
"eval_loss": 0.28115326166152954,
"eval_runtime": 1.9929,
"eval_samples_per_second": 67.238,
"eval_steps_per_second": 8.53,
"step": 420
},
{
"epoch": 25.0,
"grad_norm": 0.8111241459846497,
"learning_rate": 0.00022552941176470586,
"loss": 0.3084,
"step": 425
},
{
"epoch": 25.29,
"grad_norm": 0.6205019354820251,
"learning_rate": 0.00022464705882352937,
"loss": 0.2873,
"step": 430
},
{
"epoch": 25.59,
"grad_norm": 0.8169652223587036,
"learning_rate": 0.00022376470588235292,
"loss": 0.2965,
"step": 435
},
{
"epoch": 25.88,
"grad_norm": 0.9366277456283569,
"learning_rate": 0.00022288235294117643,
"loss": 0.3002,
"step": 440
},
{
"epoch": 25.88,
"eval_loss": 0.27707499265670776,
"eval_runtime": 1.9949,
"eval_samples_per_second": 67.171,
"eval_steps_per_second": 8.522,
"step": 440
},
{
"epoch": 26.18,
"grad_norm": 0.7006655931472778,
"learning_rate": 0.00022199999999999998,
"loss": 0.3185,
"step": 445
},
{
"epoch": 26.47,
"grad_norm": 0.6538874506950378,
"learning_rate": 0.00022111764705882352,
"loss": 0.2813,
"step": 450
},
{
"epoch": 26.76,
"grad_norm": 0.8294863104820251,
"learning_rate": 0.00022023529411764703,
"loss": 0.2989,
"step": 455
},
{
"epoch": 27.06,
"grad_norm": 0.6035735011100769,
"learning_rate": 0.00021935294117647058,
"loss": 0.2996,
"step": 460
},
{
"epoch": 27.06,
"eval_loss": 0.27492547035217285,
"eval_runtime": 1.9946,
"eval_samples_per_second": 67.183,
"eval_steps_per_second": 8.523,
"step": 460
},
{
"epoch": 27.35,
"grad_norm": 0.44762280583381653,
"learning_rate": 0.00021847058823529412,
"loss": 0.2766,
"step": 465
},
{
"epoch": 27.65,
"grad_norm": 0.6129966974258423,
"learning_rate": 0.00021758823529411764,
"loss": 0.2849,
"step": 470
},
{
"epoch": 27.94,
"grad_norm": 1.0713130235671997,
"learning_rate": 0.00021670588235294118,
"loss": 0.3428,
"step": 475
},
{
"epoch": 28.24,
"grad_norm": 0.5866985321044922,
"learning_rate": 0.00021582352941176467,
"loss": 0.2759,
"step": 480
},
{
"epoch": 28.24,
"eval_loss": 0.27579402923583984,
"eval_runtime": 1.9947,
"eval_samples_per_second": 67.176,
"eval_steps_per_second": 8.522,
"step": 480
},
{
"epoch": 28.53,
"grad_norm": 0.5810605883598328,
"learning_rate": 0.0002149411764705882,
"loss": 0.2954,
"step": 485
},
{
"epoch": 28.82,
"grad_norm": 0.6972721815109253,
"learning_rate": 0.00021405882352941173,
"loss": 0.3118,
"step": 490
},
{
"epoch": 29.12,
"grad_norm": 0.5709840059280396,
"learning_rate": 0.00021317647058823527,
"loss": 0.2901,
"step": 495
},
{
"epoch": 29.41,
"grad_norm": 0.5941194295883179,
"learning_rate": 0.0002122941176470588,
"loss": 0.3009,
"step": 500
},
{
"epoch": 29.41,
"eval_loss": 0.27494505047798157,
"eval_runtime": 1.9954,
"eval_samples_per_second": 67.154,
"eval_steps_per_second": 8.52,
"step": 500
},
{
"epoch": 29.71,
"grad_norm": 0.5286316275596619,
"learning_rate": 0.00021141176470588233,
"loss": 0.2753,
"step": 505
},
{
"epoch": 30.0,
"grad_norm": 0.6919939517974854,
"learning_rate": 0.00021052941176470587,
"loss": 0.3041,
"step": 510
},
{
"epoch": 30.29,
"grad_norm": 0.7751806974411011,
"learning_rate": 0.0002096470588235294,
"loss": 0.2994,
"step": 515
},
{
"epoch": 30.59,
"grad_norm": 0.6728924512863159,
"learning_rate": 0.00020876470588235293,
"loss": 0.2758,
"step": 520
},
{
"epoch": 30.59,
"eval_loss": 0.2755880057811737,
"eval_runtime": 1.9941,
"eval_samples_per_second": 67.197,
"eval_steps_per_second": 8.525,
"step": 520
},
{
"epoch": 30.88,
"grad_norm": 0.6528185606002808,
"learning_rate": 0.00020788235294117647,
"loss": 0.3003,
"step": 525
},
{
"epoch": 31.18,
"grad_norm": 0.6421308517456055,
"learning_rate": 0.00020699999999999996,
"loss": 0.2887,
"step": 530
},
{
"epoch": 31.47,
"grad_norm": 0.8102923631668091,
"learning_rate": 0.0002061176470588235,
"loss": 0.2871,
"step": 535
},
{
"epoch": 31.76,
"grad_norm": 0.6277897357940674,
"learning_rate": 0.00020523529411764702,
"loss": 0.2909,
"step": 540
},
{
"epoch": 31.76,
"eval_loss": 0.27382102608680725,
"eval_runtime": 1.9939,
"eval_samples_per_second": 67.207,
"eval_steps_per_second": 8.526,
"step": 540
},
{
"epoch": 32.06,
"grad_norm": 0.6897189021110535,
"learning_rate": 0.00020435294117647056,
"loss": 0.3024,
"step": 545
},
{
"epoch": 32.35,
"grad_norm": 0.4933074712753296,
"learning_rate": 0.0002034705882352941,
"loss": 0.2617,
"step": 550
},
{
"epoch": 32.65,
"grad_norm": 0.6854521632194519,
"learning_rate": 0.00020258823529411762,
"loss": 0.2882,
"step": 555
},
{
"epoch": 32.94,
"grad_norm": 0.6881256103515625,
"learning_rate": 0.00020170588235294117,
"loss": 0.3115,
"step": 560
},
{
"epoch": 32.94,
"eval_loss": 0.274933785200119,
"eval_runtime": 1.9951,
"eval_samples_per_second": 67.166,
"eval_steps_per_second": 8.521,
"step": 560
},
{
"epoch": 33.24,
"grad_norm": 0.6826930642127991,
"learning_rate": 0.00020082352941176468,
"loss": 0.3037,
"step": 565
},
{
"epoch": 33.53,
"grad_norm": 0.624535322189331,
"learning_rate": 0.00019994117647058823,
"loss": 0.2821,
"step": 570
},
{
"epoch": 33.82,
"grad_norm": 0.6433701515197754,
"learning_rate": 0.00019905882352941177,
"loss": 0.2946,
"step": 575
},
{
"epoch": 34.12,
"grad_norm": 0.5257152318954468,
"learning_rate": 0.00019817647058823528,
"loss": 0.2963,
"step": 580
},
{
"epoch": 34.12,
"eval_loss": 0.2734356224536896,
"eval_runtime": 1.9971,
"eval_samples_per_second": 67.096,
"eval_steps_per_second": 8.512,
"step": 580
},
{
"epoch": 34.41,
"grad_norm": 0.5631464719772339,
"learning_rate": 0.0001972941176470588,
"loss": 0.2992,
"step": 585
},
{
"epoch": 34.71,
"grad_norm": 0.6825907230377197,
"learning_rate": 0.00019641176470588232,
"loss": 0.2819,
"step": 590
},
{
"epoch": 35.0,
"grad_norm": 0.7912545204162598,
"learning_rate": 0.00019552941176470586,
"loss": 0.3009,
"step": 595
},
{
"epoch": 35.29,
"grad_norm": 0.6068854928016663,
"learning_rate": 0.0001946470588235294,
"loss": 0.2938,
"step": 600
},
{
"epoch": 35.29,
"eval_loss": 0.2715306282043457,
"eval_runtime": 1.997,
"eval_samples_per_second": 67.101,
"eval_steps_per_second": 8.513,
"step": 600
},
{
"epoch": 35.59,
"grad_norm": 0.6175413131713867,
"learning_rate": 0.00019376470588235292,
"loss": 0.2853,
"step": 605
},
{
"epoch": 35.88,
"grad_norm": 0.6325289011001587,
"learning_rate": 0.00019288235294117646,
"loss": 0.2825,
"step": 610
},
{
"epoch": 36.18,
"grad_norm": 0.5814202427864075,
"learning_rate": 0.00019199999999999998,
"loss": 0.2785,
"step": 615
},
{
"epoch": 36.47,
"grad_norm": 0.6043010950088501,
"learning_rate": 0.00019111764705882352,
"loss": 0.2707,
"step": 620
},
{
"epoch": 36.47,
"eval_loss": 0.27390608191490173,
"eval_runtime": 1.9949,
"eval_samples_per_second": 67.173,
"eval_steps_per_second": 8.522,
"step": 620
},
{
"epoch": 36.76,
"grad_norm": 0.7618833780288696,
"learning_rate": 0.00019023529411764706,
"loss": 0.3051,
"step": 625
},
{
"epoch": 37.06,
"grad_norm": 0.49574750661849976,
"learning_rate": 0.00018935294117647058,
"loss": 0.3009,
"step": 630
},
{
"epoch": 37.35,
"grad_norm": 0.6526892185211182,
"learning_rate": 0.00018847058823529412,
"loss": 0.2764,
"step": 635
},
{
"epoch": 37.65,
"grad_norm": 0.674574613571167,
"learning_rate": 0.0001875882352941176,
"loss": 0.2903,
"step": 640
},
{
"epoch": 37.65,
"eval_loss": 0.27166232466697693,
"eval_runtime": 1.9963,
"eval_samples_per_second": 67.125,
"eval_steps_per_second": 8.516,
"step": 640
},
{
"epoch": 37.94,
"grad_norm": 0.7739295363426208,
"learning_rate": 0.00018670588235294115,
"loss": 0.295,
"step": 645
},
{
"epoch": 38.24,
"grad_norm": 0.5252580642700195,
"learning_rate": 0.00018582352941176467,
"loss": 0.2616,
"step": 650
},
{
"epoch": 38.53,
"grad_norm": 0.8815963268280029,
"learning_rate": 0.00018494117647058821,
"loss": 0.3151,
"step": 655
},
{
"epoch": 38.82,
"grad_norm": 0.6178475618362427,
"learning_rate": 0.00018405882352941176,
"loss": 0.2902,
"step": 660
},
{
"epoch": 38.82,
"eval_loss": 0.27082693576812744,
"eval_runtime": 1.995,
"eval_samples_per_second": 67.167,
"eval_steps_per_second": 8.521,
"step": 660
},
{
"epoch": 39.12,
"grad_norm": 0.6863218545913696,
"learning_rate": 0.00018317647058823527,
"loss": 0.2846,
"step": 665
},
{
"epoch": 39.41,
"grad_norm": 0.5531587600708008,
"learning_rate": 0.00018229411764705882,
"loss": 0.2857,
"step": 670
},
{
"epoch": 39.71,
"grad_norm": 0.580801248550415,
"learning_rate": 0.00018141176470588236,
"loss": 0.2876,
"step": 675
},
{
"epoch": 40.0,
"grad_norm": 0.559680163860321,
"learning_rate": 0.00018052941176470587,
"loss": 0.2825,
"step": 680
},
{
"epoch": 40.0,
"eval_loss": 0.27086344361305237,
"eval_runtime": 2.0106,
"eval_samples_per_second": 66.647,
"eval_steps_per_second": 8.455,
"step": 680
},
{
"epoch": 40.29,
"grad_norm": 0.4894808232784271,
"learning_rate": 0.00017964705882352942,
"loss": 0.2762,
"step": 685
},
{
"epoch": 40.59,
"grad_norm": 0.4080250859260559,
"learning_rate": 0.0001787647058823529,
"loss": 0.2552,
"step": 690
},
{
"epoch": 40.88,
"grad_norm": 0.7825730443000793,
"learning_rate": 0.00017788235294117645,
"loss": 0.3105,
"step": 695
},
{
"epoch": 41.18,
"grad_norm": 0.6776964068412781,
"learning_rate": 0.00017699999999999997,
"loss": 0.2815,
"step": 700
},
{
"epoch": 41.18,
"eval_loss": 0.2690856158733368,
"eval_runtime": 1.9978,
"eval_samples_per_second": 67.073,
"eval_steps_per_second": 8.509,
"step": 700
},
{
"epoch": 41.47,
"grad_norm": 0.7837787866592407,
"learning_rate": 0.0001761176470588235,
"loss": 0.2957,
"step": 705
},
{
"epoch": 41.76,
"grad_norm": 0.63068687915802,
"learning_rate": 0.00017523529411764705,
"loss": 0.2904,
"step": 710
},
{
"epoch": 42.06,
"grad_norm": 0.6573776602745056,
"learning_rate": 0.00017435294117647057,
"loss": 0.2979,
"step": 715
},
{
"epoch": 42.35,
"grad_norm": 0.6143296360969543,
"learning_rate": 0.0001734705882352941,
"loss": 0.2618,
"step": 720
},
{
"epoch": 42.35,
"eval_loss": 0.2710973918437958,
"eval_runtime": 1.995,
"eval_samples_per_second": 67.167,
"eval_steps_per_second": 8.521,
"step": 720
},
{
"epoch": 42.65,
"grad_norm": 0.5681264400482178,
"learning_rate": 0.00017258823529411763,
"loss": 0.2918,
"step": 725
},
{
"epoch": 42.94,
"grad_norm": 0.8199780583381653,
"learning_rate": 0.00017170588235294117,
"loss": 0.291,
"step": 730
},
{
"epoch": 43.24,
"grad_norm": 0.5809566378593445,
"learning_rate": 0.0001708235294117647,
"loss": 0.287,
"step": 735
},
{
"epoch": 43.53,
"grad_norm": 0.7246254682540894,
"learning_rate": 0.00016994117647058823,
"loss": 0.2792,
"step": 740
},
{
"epoch": 43.53,
"eval_loss": 0.2726564109325409,
"eval_runtime": 1.9952,
"eval_samples_per_second": 67.162,
"eval_steps_per_second": 8.521,
"step": 740
},
{
"epoch": 43.82,
"grad_norm": 0.6743197441101074,
"learning_rate": 0.00016905882352941174,
"loss": 0.2871,
"step": 745
},
{
"epoch": 44.12,
"grad_norm": 0.6329538822174072,
"learning_rate": 0.00016817647058823526,
"loss": 0.269,
"step": 750
},
{
"epoch": 44.41,
"grad_norm": 0.6352816820144653,
"learning_rate": 0.0001672941176470588,
"loss": 0.2671,
"step": 755
},
{
"epoch": 44.71,
"grad_norm": 0.6283583045005798,
"learning_rate": 0.00016641176470588235,
"loss": 0.2907,
"step": 760
},
{
"epoch": 44.71,
"eval_loss": 0.27082738280296326,
"eval_runtime": 1.9938,
"eval_samples_per_second": 67.208,
"eval_steps_per_second": 8.526,
"step": 760
},
{
"epoch": 45.0,
"grad_norm": 0.658597469329834,
"learning_rate": 0.00016552941176470586,
"loss": 0.3013,
"step": 765
},
{
"epoch": 45.29,
"grad_norm": 0.4039109945297241,
"learning_rate": 0.0001646470588235294,
"loss": 0.2666,
"step": 770
},
{
"epoch": 45.59,
"grad_norm": 0.5540700554847717,
"learning_rate": 0.00016376470588235292,
"loss": 0.3015,
"step": 775
},
{
"epoch": 45.88,
"grad_norm": 0.7095409035682678,
"learning_rate": 0.00016288235294117646,
"loss": 0.2822,
"step": 780
},
{
"epoch": 45.88,
"eval_loss": 0.26980897784233093,
"eval_runtime": 1.9956,
"eval_samples_per_second": 67.148,
"eval_steps_per_second": 8.519,
"step": 780
},
{
"epoch": 46.18,
"grad_norm": 0.5550056099891663,
"learning_rate": 0.000162,
"loss": 0.2906,
"step": 785
},
{
"epoch": 46.47,
"grad_norm": 0.5746665000915527,
"learning_rate": 0.00016111764705882352,
"loss": 0.2841,
"step": 790
},
{
"epoch": 46.76,
"grad_norm": 0.6271633505821228,
"learning_rate": 0.00016023529411764704,
"loss": 0.2783,
"step": 795
},
{
"epoch": 47.06,
"grad_norm": 0.6846187710762024,
"learning_rate": 0.00015935294117647056,
"loss": 0.2969,
"step": 800
},
{
"epoch": 47.06,
"eval_loss": 0.268819659948349,
"eval_runtime": 1.9933,
"eval_samples_per_second": 67.226,
"eval_steps_per_second": 8.529,
"step": 800
},
{
"epoch": 47.35,
"grad_norm": 0.4498227834701538,
"learning_rate": 0.0001584705882352941,
"loss": 0.2682,
"step": 805
},
{
"epoch": 47.65,
"grad_norm": 0.4453158676624298,
"learning_rate": 0.00015758823529411761,
"loss": 0.2835,
"step": 810
},
{
"epoch": 47.94,
"grad_norm": 0.5814605355262756,
"learning_rate": 0.00015670588235294116,
"loss": 0.2936,
"step": 815
},
{
"epoch": 48.24,
"grad_norm": 0.5212646722793579,
"learning_rate": 0.0001558235294117647,
"loss": 0.2578,
"step": 820
},
{
"epoch": 48.24,
"eval_loss": 0.26919475197792053,
"eval_runtime": 1.9948,
"eval_samples_per_second": 67.174,
"eval_steps_per_second": 8.522,
"step": 820
},
{
"epoch": 48.53,
"grad_norm": 0.5765751004219055,
"learning_rate": 0.00015494117647058822,
"loss": 0.2831,
"step": 825
},
{
"epoch": 48.82,
"grad_norm": 0.719187319278717,
"learning_rate": 0.00015405882352941176,
"loss": 0.3037,
"step": 830
},
{
"epoch": 49.12,
"grad_norm": 0.5635544061660767,
"learning_rate": 0.0001531764705882353,
"loss": 0.2855,
"step": 835
},
{
"epoch": 49.41,
"grad_norm": 0.6977357864379883,
"learning_rate": 0.00015229411764705882,
"loss": 0.2715,
"step": 840
},
{
"epoch": 49.41,
"eval_loss": 0.2697232961654663,
"eval_runtime": 1.9956,
"eval_samples_per_second": 67.147,
"eval_steps_per_second": 8.519,
"step": 840
},
{
"epoch": 49.71,
"grad_norm": 0.5384779572486877,
"learning_rate": 0.00015141176470588236,
"loss": 0.2872,
"step": 845
},
{
"epoch": 50.0,
"grad_norm": 0.820724606513977,
"learning_rate": 0.00015052941176470585,
"loss": 0.2855,
"step": 850
},
{
"epoch": 50.29,
"grad_norm": 0.6426348686218262,
"learning_rate": 0.0001496470588235294,
"loss": 0.2732,
"step": 855
},
{
"epoch": 50.59,
"grad_norm": 0.49072563648223877,
"learning_rate": 0.00014876470588235294,
"loss": 0.2791,
"step": 860
},
{
"epoch": 50.59,
"eval_loss": 0.27140939235687256,
"eval_runtime": 1.9964,
"eval_samples_per_second": 67.122,
"eval_steps_per_second": 8.516,
"step": 860
},
{
"epoch": 50.88,
"grad_norm": 0.6572851538658142,
"learning_rate": 0.00014788235294117645,
"loss": 0.2788,
"step": 865
},
{
"epoch": 51.18,
"grad_norm": 0.5061244368553162,
"learning_rate": 0.000147,
"loss": 0.2984,
"step": 870
},
{
"epoch": 51.47,
"grad_norm": 0.6135215163230896,
"learning_rate": 0.0001461176470588235,
"loss": 0.2914,
"step": 875
},
{
"epoch": 51.76,
"grad_norm": 0.556154191493988,
"learning_rate": 0.00014523529411764705,
"loss": 0.2746,
"step": 880
},
{
"epoch": 51.76,
"eval_loss": 0.2693510055541992,
"eval_runtime": 1.9947,
"eval_samples_per_second": 67.177,
"eval_steps_per_second": 8.522,
"step": 880
},
{
"epoch": 52.06,
"grad_norm": 0.4937208592891693,
"learning_rate": 0.00014435294117647057,
"loss": 0.2796,
"step": 885
},
{
"epoch": 52.35,
"grad_norm": 0.5362540483474731,
"learning_rate": 0.00014347058823529409,
"loss": 0.2741,
"step": 890
},
{
"epoch": 52.65,
"grad_norm": 0.4936700761318207,
"learning_rate": 0.00014258823529411763,
"loss": 0.2876,
"step": 895
},
{
"epoch": 52.94,
"grad_norm": 0.6930960416793823,
"learning_rate": 0.00014170588235294117,
"loss": 0.2916,
"step": 900
},
{
"epoch": 52.94,
"eval_loss": 0.2691669166088104,
"eval_runtime": 1.9969,
"eval_samples_per_second": 67.104,
"eval_steps_per_second": 8.513,
"step": 900
},
{
"epoch": 53.24,
"grad_norm": 0.5722822546958923,
"learning_rate": 0.0001408235294117647,
"loss": 0.2753,
"step": 905
},
{
"epoch": 53.53,
"grad_norm": 0.4354608952999115,
"learning_rate": 0.00013994117647058823,
"loss": 0.2508,
"step": 910
},
{
"epoch": 53.82,
"grad_norm": 0.7151679396629333,
"learning_rate": 0.00013905882352941175,
"loss": 0.3153,
"step": 915
},
{
"epoch": 54.12,
"grad_norm": 0.5432928204536438,
"learning_rate": 0.0001381764705882353,
"loss": 0.295,
"step": 920
},
{
"epoch": 54.12,
"eval_loss": 0.26740530133247375,
"eval_runtime": 1.9964,
"eval_samples_per_second": 67.121,
"eval_steps_per_second": 8.515,
"step": 920
},
{
"epoch": 54.41,
"grad_norm": 0.50511234998703,
"learning_rate": 0.0001372941176470588,
"loss": 0.2678,
"step": 925
},
{
"epoch": 54.71,
"grad_norm": 0.7931966781616211,
"learning_rate": 0.00013641176470588235,
"loss": 0.2921,
"step": 930
},
{
"epoch": 55.0,
"grad_norm": 1.1295318603515625,
"learning_rate": 0.00013552941176470587,
"loss": 0.2874,
"step": 935
},
{
"epoch": 55.29,
"grad_norm": 0.5579237341880798,
"learning_rate": 0.0001346470588235294,
"loss": 0.2781,
"step": 940
},
{
"epoch": 55.29,
"eval_loss": 0.2684433162212372,
"eval_runtime": 1.9968,
"eval_samples_per_second": 67.108,
"eval_steps_per_second": 8.514,
"step": 940
},
{
"epoch": 55.59,
"grad_norm": 0.516738772392273,
"learning_rate": 0.00013376470588235292,
"loss": 0.2622,
"step": 945
},
{
"epoch": 55.88,
"grad_norm": 0.6151651740074158,
"learning_rate": 0.00013288235294117647,
"loss": 0.2957,
"step": 950
},
{
"epoch": 56.18,
"grad_norm": 0.7047178149223328,
"learning_rate": 0.00013199999999999998,
"loss": 0.2893,
"step": 955
},
{
"epoch": 56.47,
"grad_norm": 0.5808413624763489,
"learning_rate": 0.00013111764705882353,
"loss": 0.2727,
"step": 960
},
{
"epoch": 56.47,
"eval_loss": 0.2685576379299164,
"eval_runtime": 1.994,
"eval_samples_per_second": 67.203,
"eval_steps_per_second": 8.526,
"step": 960
},
{
"epoch": 56.76,
"grad_norm": 0.6171568632125854,
"learning_rate": 0.00013023529411764704,
"loss": 0.2781,
"step": 965
},
{
"epoch": 57.06,
"grad_norm": 0.5795710682868958,
"learning_rate": 0.00012935294117647056,
"loss": 0.2989,
"step": 970
},
{
"epoch": 57.35,
"grad_norm": 0.4831099510192871,
"learning_rate": 0.0001284705882352941,
"loss": 0.2719,
"step": 975
},
{
"epoch": 57.65,
"grad_norm": 0.6539974808692932,
"learning_rate": 0.00012758823529411764,
"loss": 0.2783,
"step": 980
},
{
"epoch": 57.65,
"eval_loss": 0.2687915861606598,
"eval_runtime": 1.9961,
"eval_samples_per_second": 67.13,
"eval_steps_per_second": 8.516,
"step": 980
},
{
"epoch": 57.94,
"grad_norm": 0.6757897734642029,
"learning_rate": 0.00012670588235294116,
"loss": 0.2885,
"step": 985
},
{
"epoch": 58.24,
"grad_norm": 0.5477202534675598,
"learning_rate": 0.0001258235294117647,
"loss": 0.2931,
"step": 990
},
{
"epoch": 58.53,
"grad_norm": 0.6801700592041016,
"learning_rate": 0.00012494117647058822,
"loss": 0.2692,
"step": 995
},
{
"epoch": 58.82,
"grad_norm": 0.5633025765419006,
"learning_rate": 0.00012405882352941176,
"loss": 0.2747,
"step": 1000
},
{
"epoch": 58.82,
"eval_loss": 0.269478976726532,
"eval_runtime": 1.9946,
"eval_samples_per_second": 67.181,
"eval_steps_per_second": 8.523,
"step": 1000
},
{
"epoch": 59.12,
"grad_norm": 0.6582775115966797,
"learning_rate": 0.00012317647058823528,
"loss": 0.2863,
"step": 1005
},
{
"epoch": 59.41,
"grad_norm": 0.6495881676673889,
"learning_rate": 0.00012229411764705882,
"loss": 0.2797,
"step": 1010
},
{
"epoch": 59.71,
"grad_norm": 0.5010665655136108,
"learning_rate": 0.00012141176470588235,
"loss": 0.2832,
"step": 1015
},
{
"epoch": 60.0,
"grad_norm": 0.5570430755615234,
"learning_rate": 0.00012052941176470588,
"loss": 0.2755,
"step": 1020
},
{
"epoch": 60.0,
"eval_loss": 0.26759082078933716,
"eval_runtime": 1.9963,
"eval_samples_per_second": 67.125,
"eval_steps_per_second": 8.516,
"step": 1020
},
{
"epoch": 60.29,
"grad_norm": 0.5243686437606812,
"learning_rate": 0.0001196470588235294,
"loss": 0.2639,
"step": 1025
},
{
"epoch": 60.59,
"grad_norm": 0.598508358001709,
"learning_rate": 0.00011876470588235293,
"loss": 0.2876,
"step": 1030
},
{
"epoch": 60.88,
"grad_norm": 0.595809280872345,
"learning_rate": 0.00011788235294117645,
"loss": 0.2703,
"step": 1035
},
{
"epoch": 61.18,
"grad_norm": 0.6807280778884888,
"learning_rate": 0.000117,
"loss": 0.2889,
"step": 1040
},
{
"epoch": 61.18,
"eval_loss": 0.267448753118515,
"eval_runtime": 1.997,
"eval_samples_per_second": 67.102,
"eval_steps_per_second": 8.513,
"step": 1040
},
{
"epoch": 61.47,
"grad_norm": 0.4826294183731079,
"learning_rate": 0.00011611764705882353,
"loss": 0.2689,
"step": 1045
},
{
"epoch": 61.76,
"grad_norm": 0.5328644514083862,
"learning_rate": 0.00011523529411764704,
"loss": 0.2928,
"step": 1050
},
{
"epoch": 62.06,
"grad_norm": 0.6303868889808655,
"learning_rate": 0.00011435294117647057,
"loss": 0.2778,
"step": 1055
},
{
"epoch": 62.35,
"grad_norm": 0.6396406292915344,
"learning_rate": 0.0001134705882352941,
"loss": 0.2898,
"step": 1060
},
{
"epoch": 62.35,
"eval_loss": 0.26878321170806885,
"eval_runtime": 2.0027,
"eval_samples_per_second": 66.908,
"eval_steps_per_second": 8.488,
"step": 1060
},
{
"epoch": 62.65,
"grad_norm": 0.6628149151802063,
"learning_rate": 0.00011258823529411765,
"loss": 0.2777,
"step": 1065
},
{
"epoch": 62.94,
"grad_norm": 0.5486224293708801,
"learning_rate": 0.00011170588235294117,
"loss": 0.2738,
"step": 1070
},
{
"epoch": 63.24,
"grad_norm": 0.5262264609336853,
"learning_rate": 0.00011082352941176469,
"loss": 0.2782,
"step": 1075
},
{
"epoch": 63.53,
"grad_norm": 0.5544824600219727,
"learning_rate": 0.00010994117647058822,
"loss": 0.2603,
"step": 1080
},
{
"epoch": 63.53,
"eval_loss": 0.268815815448761,
"eval_runtime": 1.9965,
"eval_samples_per_second": 67.116,
"eval_steps_per_second": 8.515,
"step": 1080
},
{
"epoch": 63.82,
"grad_norm": 0.5544533729553223,
"learning_rate": 0.00010905882352941175,
"loss": 0.2955,
"step": 1085
},
{
"epoch": 64.12,
"grad_norm": 0.5809686183929443,
"learning_rate": 0.00010817647058823529,
"loss": 0.2811,
"step": 1090
},
{
"epoch": 64.41,
"grad_norm": 0.4934030771255493,
"learning_rate": 0.00010729411764705882,
"loss": 0.2832,
"step": 1095
},
{
"epoch": 64.71,
"grad_norm": 0.6920121312141418,
"learning_rate": 0.00010641176470588235,
"loss": 0.2655,
"step": 1100
},
{
"epoch": 64.71,
"eval_loss": 0.2684324383735657,
"eval_runtime": 1.996,
"eval_samples_per_second": 67.134,
"eval_steps_per_second": 8.517,
"step": 1100
},
{
"epoch": 65.0,
"grad_norm": 0.8359130620956421,
"learning_rate": 0.00010552941176470587,
"loss": 0.2858,
"step": 1105
},
{
"epoch": 65.29,
"grad_norm": 0.47680842876434326,
"learning_rate": 0.0001046470588235294,
"loss": 0.256,
"step": 1110
},
{
"epoch": 65.59,
"grad_norm": 0.5028588771820068,
"learning_rate": 0.00010376470588235293,
"loss": 0.2834,
"step": 1115
},
{
"epoch": 65.88,
"grad_norm": 0.7377525568008423,
"learning_rate": 0.00010288235294117647,
"loss": 0.3062,
"step": 1120
},
{
"epoch": 65.88,
"eval_loss": 0.2679174244403839,
"eval_runtime": 1.996,
"eval_samples_per_second": 67.134,
"eval_steps_per_second": 8.517,
"step": 1120
},
{
"epoch": 66.18,
"grad_norm": 0.5834652185440063,
"learning_rate": 0.000102,
"loss": 0.2754,
"step": 1125
},
{
"epoch": 66.47,
"grad_norm": 0.5451943874359131,
"learning_rate": 0.00010111764705882352,
"loss": 0.2775,
"step": 1130
},
{
"epoch": 66.76,
"grad_norm": 0.7877181768417358,
"learning_rate": 0.00010023529411764704,
"loss": 0.2968,
"step": 1135
},
{
"epoch": 67.06,
"grad_norm": 0.706478476524353,
"learning_rate": 9.935294117647057e-05,
"loss": 0.2679,
"step": 1140
},
{
"epoch": 67.06,
"eval_loss": 0.26736557483673096,
"eval_runtime": 1.9963,
"eval_samples_per_second": 67.124,
"eval_steps_per_second": 8.516,
"step": 1140
},
{
"epoch": 67.35,
"grad_norm": 0.5112646818161011,
"learning_rate": 9.847058823529412e-05,
"loss": 0.2893,
"step": 1145
},
{
"epoch": 67.65,
"grad_norm": 0.5195066332817078,
"learning_rate": 9.758823529411765e-05,
"loss": 0.26,
"step": 1150
},
{
"epoch": 67.94,
"grad_norm": 0.7363670468330383,
"learning_rate": 9.670588235294116e-05,
"loss": 0.2879,
"step": 1155
},
{
"epoch": 68.24,
"grad_norm": 0.5193796753883362,
"learning_rate": 9.582352941176469e-05,
"loss": 0.2792,
"step": 1160
},
{
"epoch": 68.24,
"eval_loss": 0.26693105697631836,
"eval_runtime": 1.9965,
"eval_samples_per_second": 67.117,
"eval_steps_per_second": 8.515,
"step": 1160
},
{
"epoch": 68.53,
"grad_norm": 0.5236190557479858,
"learning_rate": 9.494117647058822e-05,
"loss": 0.2693,
"step": 1165
},
{
"epoch": 68.82,
"grad_norm": 0.5888519287109375,
"learning_rate": 9.405882352941176e-05,
"loss": 0.3004,
"step": 1170
},
{
"epoch": 69.12,
"grad_norm": 0.45090892910957336,
"learning_rate": 9.31764705882353e-05,
"loss": 0.2667,
"step": 1175
},
{
"epoch": 69.41,
"grad_norm": 0.553806483745575,
"learning_rate": 9.229411764705881e-05,
"loss": 0.2606,
"step": 1180
},
{
"epoch": 69.41,
"eval_loss": 0.26794832944869995,
"eval_runtime": 1.9956,
"eval_samples_per_second": 67.149,
"eval_steps_per_second": 8.519,
"step": 1180
},
{
"epoch": 69.71,
"grad_norm": 0.5653892755508423,
"learning_rate": 9.141176470588234e-05,
"loss": 0.2884,
"step": 1185
},
{
"epoch": 70.0,
"grad_norm": 0.6985815167427063,
"learning_rate": 9.052941176470587e-05,
"loss": 0.2787,
"step": 1190
},
{
"epoch": 70.29,
"grad_norm": 0.6280454397201538,
"learning_rate": 8.96470588235294e-05,
"loss": 0.2744,
"step": 1195
},
{
"epoch": 70.59,
"grad_norm": 0.669928252696991,
"learning_rate": 8.876470588235294e-05,
"loss": 0.2762,
"step": 1200
},
{
"epoch": 70.59,
"eval_loss": 0.26804956793785095,
"eval_runtime": 1.9958,
"eval_samples_per_second": 67.141,
"eval_steps_per_second": 8.518,
"step": 1200
},
{
"epoch": 70.88,
"grad_norm": 0.5315139293670654,
"learning_rate": 8.788235294117647e-05,
"loss": 0.2856,
"step": 1205
},
{
"epoch": 71.18,
"grad_norm": 0.5895591378211975,
"learning_rate": 8.699999999999999e-05,
"loss": 0.2916,
"step": 1210
},
{
"epoch": 71.47,
"grad_norm": 0.484454870223999,
"learning_rate": 8.611764705882352e-05,
"loss": 0.255,
"step": 1215
},
{
"epoch": 71.76,
"grad_norm": 0.624661922454834,
"learning_rate": 8.523529411764705e-05,
"loss": 0.3034,
"step": 1220
},
{
"epoch": 71.76,
"eval_loss": 0.2677695155143738,
"eval_runtime": 1.9953,
"eval_samples_per_second": 67.159,
"eval_steps_per_second": 8.52,
"step": 1220
},
{
"epoch": 72.06,
"grad_norm": 0.46470680832862854,
"learning_rate": 8.435294117647059e-05,
"loss": 0.2518,
"step": 1225
},
{
"epoch": 72.35,
"grad_norm": 0.5925688147544861,
"learning_rate": 8.347058823529412e-05,
"loss": 0.2665,
"step": 1230
},
{
"epoch": 72.65,
"grad_norm": 0.5845850706100464,
"learning_rate": 8.258823529411763e-05,
"loss": 0.2842,
"step": 1235
},
{
"epoch": 72.94,
"grad_norm": 0.7143939137458801,
"learning_rate": 8.170588235294116e-05,
"loss": 0.2874,
"step": 1240
},
{
"epoch": 72.94,
"eval_loss": 0.26679080724716187,
"eval_runtime": 1.996,
"eval_samples_per_second": 67.134,
"eval_steps_per_second": 8.517,
"step": 1240
},
{
"epoch": 73.24,
"grad_norm": 0.7805367708206177,
"learning_rate": 8.08235294117647e-05,
"loss": 0.2572,
"step": 1245
},
{
"epoch": 73.53,
"grad_norm": 0.6749709844589233,
"learning_rate": 7.994117647058824e-05,
"loss": 0.2832,
"step": 1250
},
{
"epoch": 73.82,
"grad_norm": 0.618565559387207,
"learning_rate": 7.905882352941177e-05,
"loss": 0.2604,
"step": 1255
},
{
"epoch": 74.12,
"grad_norm": 0.5196065306663513,
"learning_rate": 7.817647058823528e-05,
"loss": 0.3091,
"step": 1260
},
{
"epoch": 74.12,
"eval_loss": 0.2668515741825104,
"eval_runtime": 1.996,
"eval_samples_per_second": 67.133,
"eval_steps_per_second": 8.517,
"step": 1260
},
{
"epoch": 74.41,
"grad_norm": 0.6182024478912354,
"learning_rate": 7.729411764705881e-05,
"loss": 0.2786,
"step": 1265
},
{
"epoch": 74.71,
"grad_norm": 0.7030547857284546,
"learning_rate": 7.641176470588234e-05,
"loss": 0.2864,
"step": 1270
},
{
"epoch": 75.0,
"grad_norm": 0.48524388670921326,
"learning_rate": 7.552941176470588e-05,
"loss": 0.2613,
"step": 1275
},
{
"epoch": 75.29,
"grad_norm": 0.5933300256729126,
"learning_rate": 7.46470588235294e-05,
"loss": 0.2954,
"step": 1280
},
{
"epoch": 75.29,
"eval_loss": 0.2671739161014557,
"eval_runtime": 1.9942,
"eval_samples_per_second": 67.196,
"eval_steps_per_second": 8.525,
"step": 1280
},
{
"epoch": 75.59,
"grad_norm": 0.6050323843955994,
"learning_rate": 7.376470588235293e-05,
"loss": 0.2572,
"step": 1285
},
{
"epoch": 75.88,
"grad_norm": 0.5846253633499146,
"learning_rate": 7.288235294117647e-05,
"loss": 0.268,
"step": 1290
},
{
"epoch": 76.18,
"grad_norm": 0.5491711497306824,
"learning_rate": 7.199999999999999e-05,
"loss": 0.2706,
"step": 1295
},
{
"epoch": 76.47,
"grad_norm": 0.5089874863624573,
"learning_rate": 7.111764705882352e-05,
"loss": 0.2591,
"step": 1300
},
{
"epoch": 76.47,
"eval_loss": 0.2666688561439514,
"eval_runtime": 1.9966,
"eval_samples_per_second": 67.114,
"eval_steps_per_second": 8.514,
"step": 1300
},
{
"epoch": 76.76,
"grad_norm": 0.7518558502197266,
"learning_rate": 7.023529411764706e-05,
"loss": 0.282,
"step": 1305
},
{
"epoch": 77.06,
"grad_norm": 0.624338686466217,
"learning_rate": 6.935294117647058e-05,
"loss": 0.2991,
"step": 1310
},
{
"epoch": 77.35,
"grad_norm": 0.5179735422134399,
"learning_rate": 6.847058823529412e-05,
"loss": 0.2667,
"step": 1315
},
{
"epoch": 77.65,
"grad_norm": 0.6224756240844727,
"learning_rate": 6.758823529411764e-05,
"loss": 0.28,
"step": 1320
},
{
"epoch": 77.65,
"eval_loss": 0.26699915528297424,
"eval_runtime": 1.9958,
"eval_samples_per_second": 67.139,
"eval_steps_per_second": 8.518,
"step": 1320
},
{
"epoch": 77.94,
"grad_norm": 0.6646649837493896,
"learning_rate": 6.670588235294117e-05,
"loss": 0.2726,
"step": 1325
},
{
"epoch": 78.24,
"grad_norm": 0.4936932325363159,
"learning_rate": 6.582352941176471e-05,
"loss": 0.2663,
"step": 1330
},
{
"epoch": 78.53,
"grad_norm": 0.5740850567817688,
"learning_rate": 6.494117647058822e-05,
"loss": 0.2642,
"step": 1335
},
{
"epoch": 78.82,
"grad_norm": 0.5269001722335815,
"learning_rate": 6.405882352941175e-05,
"loss": 0.2818,
"step": 1340
},
{
"epoch": 78.82,
"eval_loss": 0.2666696310043335,
"eval_runtime": 1.9974,
"eval_samples_per_second": 67.088,
"eval_steps_per_second": 8.511,
"step": 1340
},
{
"epoch": 79.12,
"grad_norm": 0.6438319087028503,
"learning_rate": 6.31764705882353e-05,
"loss": 0.3015,
"step": 1345
},
{
"epoch": 79.41,
"grad_norm": 0.5811217427253723,
"learning_rate": 6.229411764705881e-05,
"loss": 0.2438,
"step": 1350
},
{
"epoch": 79.71,
"grad_norm": 0.5257366895675659,
"learning_rate": 6.141176470588236e-05,
"loss": 0.2775,
"step": 1355
},
{
"epoch": 80.0,
"grad_norm": 0.67855304479599,
"learning_rate": 6.052941176470587e-05,
"loss": 0.2922,
"step": 1360
},
{
"epoch": 80.0,
"eval_loss": 0.2664375901222229,
"eval_runtime": 1.9959,
"eval_samples_per_second": 67.136,
"eval_steps_per_second": 8.517,
"step": 1360
},
{
"epoch": 80.29,
"grad_norm": 0.6666358113288879,
"learning_rate": 5.964705882352941e-05,
"loss": 0.2827,
"step": 1365
},
{
"epoch": 80.59,
"grad_norm": 0.5569036602973938,
"learning_rate": 5.876470588235294e-05,
"loss": 0.2741,
"step": 1370
},
{
"epoch": 80.88,
"grad_norm": 0.5340925455093384,
"learning_rate": 5.788235294117646e-05,
"loss": 0.2587,
"step": 1375
},
{
"epoch": 81.18,
"grad_norm": 0.6019856929779053,
"learning_rate": 5.6999999999999996e-05,
"loss": 0.2883,
"step": 1380
},
{
"epoch": 81.18,
"eval_loss": 0.266279011964798,
"eval_runtime": 1.9934,
"eval_samples_per_second": 67.222,
"eval_steps_per_second": 8.528,
"step": 1380
},
{
"epoch": 81.47,
"grad_norm": 0.4570218026638031,
"learning_rate": 5.6117647058823526e-05,
"loss": 0.2521,
"step": 1385
},
{
"epoch": 81.76,
"grad_norm": 0.6152669191360474,
"learning_rate": 5.5235294117647055e-05,
"loss": 0.2826,
"step": 1390
},
{
"epoch": 82.06,
"grad_norm": 0.537754237651825,
"learning_rate": 5.4352941176470585e-05,
"loss": 0.2782,
"step": 1395
},
{
"epoch": 82.35,
"grad_norm": 0.7443511486053467,
"learning_rate": 5.347058823529411e-05,
"loss": 0.2717,
"step": 1400
},
{
"epoch": 82.35,
"eval_loss": 0.2660791873931885,
"eval_runtime": 1.9952,
"eval_samples_per_second": 67.162,
"eval_steps_per_second": 8.521,
"step": 1400
},
{
"epoch": 82.65,
"grad_norm": 0.45767509937286377,
"learning_rate": 5.2588235294117644e-05,
"loss": 0.2595,
"step": 1405
},
{
"epoch": 82.94,
"grad_norm": 0.7017338871955872,
"learning_rate": 5.170588235294117e-05,
"loss": 0.2865,
"step": 1410
},
{
"epoch": 83.24,
"grad_norm": 0.4680289030075073,
"learning_rate": 5.0823529411764696e-05,
"loss": 0.2761,
"step": 1415
},
{
"epoch": 83.53,
"grad_norm": 0.5954050421714783,
"learning_rate": 4.994117647058823e-05,
"loss": 0.2874,
"step": 1420
},
{
"epoch": 83.53,
"eval_loss": 0.2670120894908905,
"eval_runtime": 1.9957,
"eval_samples_per_second": 67.146,
"eval_steps_per_second": 8.519,
"step": 1420
},
{
"epoch": 83.82,
"grad_norm": 0.5585493445396423,
"learning_rate": 4.905882352941176e-05,
"loss": 0.274,
"step": 1425
},
{
"epoch": 84.12,
"grad_norm": 0.6001076698303223,
"learning_rate": 4.817647058823529e-05,
"loss": 0.2579,
"step": 1430
},
{
"epoch": 84.41,
"grad_norm": 0.5703680515289307,
"learning_rate": 4.729411764705882e-05,
"loss": 0.2602,
"step": 1435
},
{
"epoch": 84.71,
"grad_norm": 0.5915176868438721,
"learning_rate": 4.6411764705882343e-05,
"loss": 0.2676,
"step": 1440
},
{
"epoch": 84.71,
"eval_loss": 0.2662909924983978,
"eval_runtime": 2.007,
"eval_samples_per_second": 66.766,
"eval_steps_per_second": 8.47,
"step": 1440
},
{
"epoch": 85.0,
"grad_norm": 0.659074068069458,
"learning_rate": 4.552941176470588e-05,
"loss": 0.2864,
"step": 1445
},
{
"epoch": 85.29,
"grad_norm": 0.6976066827774048,
"learning_rate": 4.464705882352941e-05,
"loss": 0.277,
"step": 1450
},
{
"epoch": 85.59,
"grad_norm": 0.6575342416763306,
"learning_rate": 4.376470588235293e-05,
"loss": 0.2886,
"step": 1455
},
{
"epoch": 85.88,
"grad_norm": 0.4784369170665741,
"learning_rate": 4.288235294117647e-05,
"loss": 0.2637,
"step": 1460
},
{
"epoch": 85.88,
"eval_loss": 0.2663717269897461,
"eval_runtime": 1.9942,
"eval_samples_per_second": 67.195,
"eval_steps_per_second": 8.525,
"step": 1460
},
{
"epoch": 86.18,
"grad_norm": 0.5686805844306946,
"learning_rate": 4.2e-05,
"loss": 0.2677,
"step": 1465
},
{
"epoch": 86.47,
"grad_norm": 0.6093593239784241,
"learning_rate": 4.111764705882353e-05,
"loss": 0.2594,
"step": 1470
},
{
"epoch": 86.76,
"grad_norm": 0.7324638366699219,
"learning_rate": 4.023529411764706e-05,
"loss": 0.2916,
"step": 1475
},
{
"epoch": 87.06,
"grad_norm": 0.6090161204338074,
"learning_rate": 3.935294117647058e-05,
"loss": 0.2539,
"step": 1480
},
{
"epoch": 87.06,
"eval_loss": 0.26603055000305176,
"eval_runtime": 1.9935,
"eval_samples_per_second": 67.218,
"eval_steps_per_second": 8.528,
"step": 1480
},
{
"epoch": 87.35,
"grad_norm": 0.7536705136299133,
"learning_rate": 3.8470588235294116e-05,
"loss": 0.2886,
"step": 1485
},
{
"epoch": 87.65,
"grad_norm": 0.5662975311279297,
"learning_rate": 3.7588235294117645e-05,
"loss": 0.2741,
"step": 1490
},
{
"epoch": 87.94,
"grad_norm": 0.5700023174285889,
"learning_rate": 3.6705882352941175e-05,
"loss": 0.2767,
"step": 1495
},
{
"epoch": 88.24,
"grad_norm": 0.5497795343399048,
"learning_rate": 3.5823529411764704e-05,
"loss": 0.2648,
"step": 1500
},
{
"epoch": 88.24,
"eval_loss": 0.2659338712692261,
"eval_runtime": 1.9949,
"eval_samples_per_second": 67.171,
"eval_steps_per_second": 8.522,
"step": 1500
},
{
"epoch": 88.53,
"grad_norm": 0.5747610926628113,
"learning_rate": 3.4941176470588234e-05,
"loss": 0.2627,
"step": 1505
},
{
"epoch": 88.82,
"grad_norm": 0.4665067195892334,
"learning_rate": 3.405882352941176e-05,
"loss": 0.2847,
"step": 1510
},
{
"epoch": 89.12,
"grad_norm": 0.7121643424034119,
"learning_rate": 3.317647058823529e-05,
"loss": 0.2676,
"step": 1515
},
{
"epoch": 89.41,
"grad_norm": 0.5986043214797974,
"learning_rate": 3.229411764705882e-05,
"loss": 0.2782,
"step": 1520
},
{
"epoch": 89.41,
"eval_loss": 0.2659454941749573,
"eval_runtime": 1.9957,
"eval_samples_per_second": 67.143,
"eval_steps_per_second": 8.518,
"step": 1520
},
{
"epoch": 89.71,
"grad_norm": 0.5361090898513794,
"learning_rate": 3.141176470588235e-05,
"loss": 0.2644,
"step": 1525
},
{
"epoch": 90.0,
"grad_norm": 0.7882270812988281,
"learning_rate": 3.052941176470588e-05,
"loss": 0.2778,
"step": 1530
},
{
"epoch": 90.29,
"grad_norm": 0.6407003402709961,
"learning_rate": 2.9647058823529407e-05,
"loss": 0.2635,
"step": 1535
},
{
"epoch": 90.59,
"grad_norm": 0.6531354188919067,
"learning_rate": 2.876470588235294e-05,
"loss": 0.275,
"step": 1540
},
{
"epoch": 90.59,
"eval_loss": 0.2659270465373993,
"eval_runtime": 1.9941,
"eval_samples_per_second": 67.198,
"eval_steps_per_second": 8.525,
"step": 1540
},
{
"epoch": 90.88,
"grad_norm": 0.6199626326560974,
"learning_rate": 2.788235294117647e-05,
"loss": 0.2681,
"step": 1545
},
{
"epoch": 91.18,
"grad_norm": 0.6273800134658813,
"learning_rate": 2.6999999999999996e-05,
"loss": 0.287,
"step": 1550
},
{
"epoch": 91.47,
"grad_norm": 0.6072767972946167,
"learning_rate": 2.6117647058823525e-05,
"loss": 0.2583,
"step": 1555
},
{
"epoch": 91.76,
"grad_norm": 0.5658175945281982,
"learning_rate": 2.5235294117647058e-05,
"loss": 0.2717,
"step": 1560
},
{
"epoch": 91.76,
"eval_loss": 0.26581788063049316,
"eval_runtime": 1.994,
"eval_samples_per_second": 67.201,
"eval_steps_per_second": 8.525,
"step": 1560
},
{
"epoch": 92.06,
"grad_norm": 0.6367130279541016,
"learning_rate": 2.4352941176470587e-05,
"loss": 0.2819,
"step": 1565
},
{
"epoch": 92.35,
"grad_norm": 0.6430366039276123,
"learning_rate": 2.3470588235294114e-05,
"loss": 0.2759,
"step": 1570
},
{
"epoch": 92.65,
"grad_norm": 0.6165040731430054,
"learning_rate": 2.2588235294117643e-05,
"loss": 0.2715,
"step": 1575
},
{
"epoch": 92.94,
"grad_norm": 0.5176894664764404,
"learning_rate": 2.1705882352941176e-05,
"loss": 0.2646,
"step": 1580
},
{
"epoch": 92.94,
"eval_loss": 0.2657198905944824,
"eval_runtime": 1.9962,
"eval_samples_per_second": 67.126,
"eval_steps_per_second": 8.516,
"step": 1580
},
{
"epoch": 93.24,
"grad_norm": 0.7515372037887573,
"learning_rate": 2.0823529411764705e-05,
"loss": 0.2631,
"step": 1585
},
{
"epoch": 93.53,
"grad_norm": 0.4853091835975647,
"learning_rate": 1.994117647058823e-05,
"loss": 0.2573,
"step": 1590
},
{
"epoch": 93.82,
"grad_norm": 0.4851018190383911,
"learning_rate": 1.905882352941176e-05,
"loss": 0.271,
"step": 1595
},
{
"epoch": 94.12,
"grad_norm": 0.503040075302124,
"learning_rate": 1.817647058823529e-05,
"loss": 0.2863,
"step": 1600
},
{
"epoch": 94.12,
"eval_loss": 0.26566752791404724,
"eval_runtime": 1.9956,
"eval_samples_per_second": 67.147,
"eval_steps_per_second": 8.519,
"step": 1600
},
{
"epoch": 94.41,
"grad_norm": 0.560333788394928,
"learning_rate": 1.7294117647058823e-05,
"loss": 0.2546,
"step": 1605
},
{
"epoch": 94.71,
"grad_norm": 0.5879210233688354,
"learning_rate": 1.641176470588235e-05,
"loss": 0.2733,
"step": 1610
},
{
"epoch": 95.0,
"grad_norm": 0.6487263441085815,
"learning_rate": 1.5529411764705882e-05,
"loss": 0.2894,
"step": 1615
},
{
"epoch": 95.29,
"grad_norm": 0.6337783336639404,
"learning_rate": 1.464705882352941e-05,
"loss": 0.2731,
"step": 1620
},
{
"epoch": 95.29,
"eval_loss": 0.2656712234020233,
"eval_runtime": 1.9955,
"eval_samples_per_second": 67.153,
"eval_steps_per_second": 8.519,
"step": 1620
},
{
"epoch": 95.59,
"grad_norm": 0.5507830381393433,
"learning_rate": 1.376470588235294e-05,
"loss": 0.2801,
"step": 1625
},
{
"epoch": 95.88,
"grad_norm": 0.7097357511520386,
"learning_rate": 1.2882352941176469e-05,
"loss": 0.274,
"step": 1630
},
{
"epoch": 96.18,
"grad_norm": 0.5612406730651855,
"learning_rate": 1.1999999999999999e-05,
"loss": 0.2571,
"step": 1635
},
{
"epoch": 96.47,
"grad_norm": 0.7559302449226379,
"learning_rate": 1.1117647058823528e-05,
"loss": 0.2795,
"step": 1640
},
{
"epoch": 96.47,
"eval_loss": 0.2656753361225128,
"eval_runtime": 1.9935,
"eval_samples_per_second": 67.217,
"eval_steps_per_second": 8.528,
"step": 1640
},
{
"epoch": 96.76,
"grad_norm": 0.542195737361908,
"learning_rate": 1.0235294117647058e-05,
"loss": 0.2549,
"step": 1645
},
{
"epoch": 97.06,
"grad_norm": 0.6926820874214172,
"learning_rate": 9.352941176470587e-06,
"loss": 0.2598,
"step": 1650
},
{
"epoch": 97.35,
"grad_norm": 0.5095303654670715,
"learning_rate": 8.470588235294117e-06,
"loss": 0.2753,
"step": 1655
},
{
"epoch": 97.65,
"grad_norm": 0.5196524262428284,
"learning_rate": 7.588235294117647e-06,
"loss": 0.2501,
"step": 1660
},
{
"epoch": 97.65,
"eval_loss": 0.26558738946914673,
"eval_runtime": 1.9934,
"eval_samples_per_second": 67.221,
"eval_steps_per_second": 8.528,
"step": 1660
},
{
"epoch": 97.94,
"grad_norm": 0.6086786985397339,
"learning_rate": 6.705882352941176e-06,
"loss": 0.2955,
"step": 1665
},
{
"epoch": 98.24,
"grad_norm": 0.41819947957992554,
"learning_rate": 5.823529411764706e-06,
"loss": 0.2543,
"step": 1670
},
{
"epoch": 98.53,
"grad_norm": 0.6898114681243896,
"learning_rate": 4.941176470588235e-06,
"loss": 0.288,
"step": 1675
},
{
"epoch": 98.82,
"grad_norm": 0.5956689119338989,
"learning_rate": 4.058823529411764e-06,
"loss": 0.2623,
"step": 1680
},
{
"epoch": 98.82,
"eval_loss": 0.2655787765979767,
"eval_runtime": 1.9955,
"eval_samples_per_second": 67.152,
"eval_steps_per_second": 8.519,
"step": 1680
},
{
"epoch": 99.12,
"grad_norm": 0.5556439757347107,
"learning_rate": 3.1764705882352935e-06,
"loss": 0.2614,
"step": 1685
},
{
"epoch": 99.41,
"grad_norm": 0.46832630038261414,
"learning_rate": 2.2941176470588234e-06,
"loss": 0.2647,
"step": 1690
},
{
"epoch": 99.71,
"grad_norm": 0.49263256788253784,
"learning_rate": 1.4117647058823527e-06,
"loss": 0.273,
"step": 1695
},
{
"epoch": 100.0,
"grad_norm": 0.6977939605712891,
"learning_rate": 5.294117647058823e-07,
"loss": 0.2655,
"step": 1700
},
{
"epoch": 100.0,
"eval_loss": 0.26560500264167786,
"eval_runtime": 1.9947,
"eval_samples_per_second": 67.178,
"eval_steps_per_second": 8.523,
"step": 1700
}
],
"logging_steps": 5,
"max_steps": 1700,
"num_input_tokens_seen": 0,
"num_train_epochs": 100,
"save_steps": 20,
"total_flos": 1.523488749060096e+16,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}