Korean_syllable_roberta_512 / trainer_state.json
Trofish's picture
Upload 10 files
010b40a verified
raw
history blame
246 kB
{
"best_metric": 0.833830714225769,
"best_model_checkpoint": "/home/nlplab12/Desktop/roberta-pretrain/ckpt/roberta/pretrain/medium/512/checkpoint-12420",
"epoch": 10.559424425487993,
"eval_steps": 90,
"global_step": 12420,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.008500045535958229,
"grad_norm": 514775.65625,
"learning_rate": 4.166666666666667e-06,
"loss": 7.2556,
"step": 10
},
{
"epoch": 0.017000091071916457,
"grad_norm": 399649.5,
"learning_rate": 8.333333333333334e-06,
"loss": 6.7984,
"step": 20
},
{
"epoch": 0.025500136607874684,
"grad_norm": 137177.0,
"learning_rate": 1.25e-05,
"loss": 6.2165,
"step": 30
},
{
"epoch": 0.034000182143832915,
"grad_norm": 151481.765625,
"learning_rate": 1.6666666666666667e-05,
"loss": 5.9275,
"step": 40
},
{
"epoch": 0.04250022767979114,
"grad_norm": 92555.8359375,
"learning_rate": 2.0833333333333333e-05,
"loss": 5.7498,
"step": 50
},
{
"epoch": 0.05100027321574937,
"grad_norm": 86679.46875,
"learning_rate": 2.5e-05,
"loss": 5.6096,
"step": 60
},
{
"epoch": 0.059500318751707595,
"grad_norm": 66003.234375,
"learning_rate": 2.9166666666666666e-05,
"loss": 5.4786,
"step": 70
},
{
"epoch": 0.06800036428766583,
"grad_norm": 57082.796875,
"learning_rate": 3.3333333333333335e-05,
"loss": 5.3565,
"step": 80
},
{
"epoch": 0.07650040982362405,
"grad_norm": 46905.7734375,
"learning_rate": 3.75e-05,
"loss": 5.2382,
"step": 90
},
{
"epoch": 0.07650040982362405,
"eval_accuracy": 0.2240983525108359,
"eval_loss": 5.1665496826171875,
"eval_runtime": 1330.3439,
"eval_samples_per_second": 375.375,
"eval_steps_per_second": 1.303,
"step": 90
},
{
"epoch": 0.08500045535958228,
"grad_norm": 36661.265625,
"learning_rate": 4.1666666666666665e-05,
"loss": 5.1301,
"step": 100
},
{
"epoch": 0.09350050089554052,
"grad_norm": 28157.064453125,
"learning_rate": 4.5833333333333334e-05,
"loss": 5.0237,
"step": 110
},
{
"epoch": 0.10200054643149874,
"grad_norm": 31035.537109375,
"learning_rate": 5e-05,
"loss": 4.9291,
"step": 120
},
{
"epoch": 0.11050059196745697,
"grad_norm": 41497.1640625,
"learning_rate": 5.416666666666667e-05,
"loss": 4.8501,
"step": 130
},
{
"epoch": 0.11900063750341519,
"grad_norm": 44447.87890625,
"learning_rate": 5.833333333333333e-05,
"loss": 4.7938,
"step": 140
},
{
"epoch": 0.12750068303937342,
"grad_norm": 36388.1015625,
"learning_rate": 6.25e-05,
"loss": 4.7491,
"step": 150
},
{
"epoch": 0.13600072857533166,
"grad_norm": 24360.1875,
"learning_rate": 6.666666666666667e-05,
"loss": 4.7147,
"step": 160
},
{
"epoch": 0.1445007741112899,
"grad_norm": 19699.2734375,
"learning_rate": 7.083333333333334e-05,
"loss": 4.6829,
"step": 170
},
{
"epoch": 0.1530008196472481,
"grad_norm": 11622.908203125,
"learning_rate": 7.5e-05,
"loss": 4.6545,
"step": 180
},
{
"epoch": 0.1530008196472481,
"eval_accuracy": 0.284967956348924,
"eval_loss": 4.628236293792725,
"eval_runtime": 1276.531,
"eval_samples_per_second": 391.199,
"eval_steps_per_second": 1.358,
"step": 180
},
{
"epoch": 0.16150086518320633,
"grad_norm": 16312.888671875,
"learning_rate": 7.916666666666666e-05,
"loss": 4.6377,
"step": 190
},
{
"epoch": 0.17000091071916457,
"grad_norm": 37823.2890625,
"learning_rate": 8.333333333333333e-05,
"loss": 4.6244,
"step": 200
},
{
"epoch": 0.1785009562551228,
"grad_norm": 18776.1796875,
"learning_rate": 8.75e-05,
"loss": 4.6141,
"step": 210
},
{
"epoch": 0.18700100179108103,
"grad_norm": 34368.3359375,
"learning_rate": 9.166666666666667e-05,
"loss": 4.6043,
"step": 220
},
{
"epoch": 0.19550104732703924,
"grad_norm": 38324.95703125,
"learning_rate": 9.583333333333334e-05,
"loss": 4.5989,
"step": 230
},
{
"epoch": 0.20400109286299747,
"grad_norm": 35916.953125,
"learning_rate": 0.0001,
"loss": 4.5924,
"step": 240
},
{
"epoch": 0.2125011383989557,
"grad_norm": 34779.6328125,
"learning_rate": 0.00010416666666666667,
"loss": 4.5855,
"step": 250
},
{
"epoch": 0.22100118393491394,
"grad_norm": 38605.4921875,
"learning_rate": 0.00010833333333333334,
"loss": 4.5771,
"step": 260
},
{
"epoch": 0.22950122947087218,
"grad_norm": 17166.369140625,
"learning_rate": 0.00011250000000000001,
"loss": 4.5704,
"step": 270
},
{
"epoch": 0.22950122947087218,
"eval_accuracy": 0.2895338322091062,
"eval_loss": 4.5623779296875,
"eval_runtime": 1273.631,
"eval_samples_per_second": 392.09,
"eval_steps_per_second": 1.361,
"step": 270
},
{
"epoch": 0.23800127500683038,
"grad_norm": 42357.046875,
"learning_rate": 0.00011666666666666667,
"loss": 4.5682,
"step": 280
},
{
"epoch": 0.24650132054278862,
"grad_norm": 20787.40625,
"learning_rate": 0.00012083333333333333,
"loss": 4.5646,
"step": 290
},
{
"epoch": 0.25500136607874685,
"grad_norm": 49082.19921875,
"learning_rate": 0.000125,
"loss": 4.559,
"step": 300
},
{
"epoch": 0.2635014116147051,
"grad_norm": 18796.533203125,
"learning_rate": 0.00012916666666666667,
"loss": 4.5534,
"step": 310
},
{
"epoch": 0.2720014571506633,
"grad_norm": 45862.125,
"learning_rate": 0.00013333333333333334,
"loss": 4.5513,
"step": 320
},
{
"epoch": 0.28050150268662155,
"grad_norm": 27371.748046875,
"learning_rate": 0.0001375,
"loss": 4.5489,
"step": 330
},
{
"epoch": 0.2890015482225798,
"grad_norm": 26154.533203125,
"learning_rate": 0.00014166666666666668,
"loss": 4.5471,
"step": 340
},
{
"epoch": 0.297501593758538,
"grad_norm": 46019.3671875,
"learning_rate": 0.00014583333333333335,
"loss": 4.5453,
"step": 350
},
{
"epoch": 0.3060016392944962,
"grad_norm": 54414.8828125,
"learning_rate": 0.00015,
"loss": 4.5435,
"step": 360
},
{
"epoch": 0.3060016392944962,
"eval_accuracy": 0.28996644531895915,
"eval_loss": 4.538370609283447,
"eval_runtime": 1274.2161,
"eval_samples_per_second": 391.91,
"eval_steps_per_second": 1.361,
"step": 360
},
{
"epoch": 0.31450168483045443,
"grad_norm": 61712.6640625,
"learning_rate": 0.00015416666666666668,
"loss": 4.5387,
"step": 370
},
{
"epoch": 0.32300173036641266,
"grad_norm": 35200.015625,
"learning_rate": 0.00015833333333333332,
"loss": 4.537,
"step": 380
},
{
"epoch": 0.3315017759023709,
"grad_norm": 40758.90234375,
"learning_rate": 0.00016250000000000002,
"loss": 4.5345,
"step": 390
},
{
"epoch": 0.34000182143832913,
"grad_norm": 38277.3515625,
"learning_rate": 0.00016666666666666666,
"loss": 4.5351,
"step": 400
},
{
"epoch": 0.34850186697428737,
"grad_norm": 28720.861328125,
"learning_rate": 0.00017083333333333333,
"loss": 4.534,
"step": 410
},
{
"epoch": 0.3570019125102456,
"grad_norm": 21332.2109375,
"learning_rate": 0.000175,
"loss": 4.5329,
"step": 420
},
{
"epoch": 0.36550195804620383,
"grad_norm": 46977.23828125,
"learning_rate": 0.00017916666666666667,
"loss": 4.5282,
"step": 430
},
{
"epoch": 0.37400200358216207,
"grad_norm": 54403.328125,
"learning_rate": 0.00018333333333333334,
"loss": 4.5273,
"step": 440
},
{
"epoch": 0.3825020491181203,
"grad_norm": 33038.54296875,
"learning_rate": 0.0001875,
"loss": 4.5272,
"step": 450
},
{
"epoch": 0.3825020491181203,
"eval_accuracy": 0.2900167539077763,
"eval_loss": 4.523510932922363,
"eval_runtime": 1270.6276,
"eval_samples_per_second": 393.017,
"eval_steps_per_second": 1.365,
"step": 450
},
{
"epoch": 0.3910020946540785,
"grad_norm": 38754.59765625,
"learning_rate": 0.00019166666666666667,
"loss": 4.5216,
"step": 460
},
{
"epoch": 0.3995021401900367,
"grad_norm": 36281.0078125,
"learning_rate": 0.00019583333333333334,
"loss": 4.5237,
"step": 470
},
{
"epoch": 0.40800218572599495,
"grad_norm": 35634.79296875,
"learning_rate": 0.0002,
"loss": 4.5186,
"step": 480
},
{
"epoch": 0.4165022312619532,
"grad_norm": 15257.716796875,
"learning_rate": 0.00020416666666666668,
"loss": 4.5173,
"step": 490
},
{
"epoch": 0.4250022767979114,
"grad_norm": 36190.26171875,
"learning_rate": 0.00020833333333333335,
"loss": 4.5174,
"step": 500
},
{
"epoch": 0.43350232233386965,
"grad_norm": 57666.21875,
"learning_rate": 0.0002125,
"loss": 4.5144,
"step": 510
},
{
"epoch": 0.4420023678698279,
"grad_norm": 33545.0078125,
"learning_rate": 0.00021666666666666668,
"loss": 4.5131,
"step": 520
},
{
"epoch": 0.4505024134057861,
"grad_norm": 75054.1328125,
"learning_rate": 0.00022083333333333333,
"loss": 4.5116,
"step": 530
},
{
"epoch": 0.45900245894174435,
"grad_norm": 27162.548828125,
"learning_rate": 0.00022500000000000002,
"loss": 4.5104,
"step": 540
},
{
"epoch": 0.45900245894174435,
"eval_accuracy": 0.2903866574093662,
"eval_loss": 4.506768703460693,
"eval_runtime": 1274.9615,
"eval_samples_per_second": 391.681,
"eval_steps_per_second": 1.36,
"step": 540
},
{
"epoch": 0.4675025044777026,
"grad_norm": 31226.064453125,
"learning_rate": 0.00022916666666666666,
"loss": 4.5064,
"step": 550
},
{
"epoch": 0.47600255001366076,
"grad_norm": 34113.46484375,
"learning_rate": 0.00023333333333333333,
"loss": 4.5085,
"step": 560
},
{
"epoch": 0.484502595549619,
"grad_norm": 26731.498046875,
"learning_rate": 0.0002375,
"loss": 4.5079,
"step": 570
},
{
"epoch": 0.49300264108557723,
"grad_norm": 18876.388671875,
"learning_rate": 0.00024166666666666667,
"loss": 4.5041,
"step": 580
},
{
"epoch": 0.5015026866215355,
"grad_norm": 30968.685546875,
"learning_rate": 0.0002458333333333333,
"loss": 4.5044,
"step": 590
},
{
"epoch": 0.5100027321574937,
"grad_norm": 33946.8125,
"learning_rate": 0.00025,
"loss": 4.5036,
"step": 600
},
{
"epoch": 0.5185027776934519,
"grad_norm": 27673.900390625,
"learning_rate": 0.00025416666666666665,
"loss": 4.5007,
"step": 610
},
{
"epoch": 0.5270028232294102,
"grad_norm": 17412.22265625,
"learning_rate": 0.00025833333333333334,
"loss": 4.5003,
"step": 620
},
{
"epoch": 0.5355028687653683,
"grad_norm": 38210.4140625,
"learning_rate": 0.00026250000000000004,
"loss": 4.4977,
"step": 630
},
{
"epoch": 0.5355028687653683,
"eval_accuracy": 0.2904019042861853,
"eval_loss": 4.49587869644165,
"eval_runtime": 1274.4439,
"eval_samples_per_second": 391.84,
"eval_steps_per_second": 1.361,
"step": 630
},
{
"epoch": 0.5440029143013266,
"grad_norm": 22160.921875,
"learning_rate": 0.0002666666666666667,
"loss": 4.498,
"step": 640
},
{
"epoch": 0.5525029598372848,
"grad_norm": 40694.95703125,
"learning_rate": 0.0002708333333333333,
"loss": 4.4947,
"step": 650
},
{
"epoch": 0.5610030053732431,
"grad_norm": 22046.97265625,
"learning_rate": 0.000275,
"loss": 4.495,
"step": 660
},
{
"epoch": 0.5695030509092013,
"grad_norm": 45871.296875,
"learning_rate": 0.00027916666666666666,
"loss": 4.4946,
"step": 670
},
{
"epoch": 0.5780030964451596,
"grad_norm": 32230.837890625,
"learning_rate": 0.00028333333333333335,
"loss": 4.4941,
"step": 680
},
{
"epoch": 0.5865031419811177,
"grad_norm": 31339.076171875,
"learning_rate": 0.0002875,
"loss": 4.4921,
"step": 690
},
{
"epoch": 0.595003187517076,
"grad_norm": 21844.26171875,
"learning_rate": 0.0002916666666666667,
"loss": 4.4912,
"step": 700
},
{
"epoch": 0.6035032330530342,
"grad_norm": 20966.12109375,
"learning_rate": 0.00029583333333333333,
"loss": 4.489,
"step": 710
},
{
"epoch": 0.6120032785889924,
"grad_norm": 19211.65625,
"learning_rate": 0.0003,
"loss": 4.4903,
"step": 720
},
{
"epoch": 0.6120032785889924,
"eval_accuracy": 0.2904662633238704,
"eval_loss": 4.487661361694336,
"eval_runtime": 1273.4031,
"eval_samples_per_second": 392.16,
"eval_steps_per_second": 1.362,
"step": 720
},
{
"epoch": 0.6205033241249507,
"grad_norm": 47674.4140625,
"learning_rate": 0.00030416666666666667,
"loss": 4.4886,
"step": 730
},
{
"epoch": 0.6290033696609089,
"grad_norm": 44354.7890625,
"learning_rate": 0.00030833333333333337,
"loss": 4.4865,
"step": 740
},
{
"epoch": 0.6375034151968672,
"grad_norm": 27939.216796875,
"learning_rate": 0.0003125,
"loss": 4.4868,
"step": 750
},
{
"epoch": 0.6460034607328253,
"grad_norm": 29380.662109375,
"learning_rate": 0.00031666666666666665,
"loss": 4.4858,
"step": 760
},
{
"epoch": 0.6545035062687836,
"grad_norm": 38158.73046875,
"learning_rate": 0.00032083333333333334,
"loss": 4.4834,
"step": 770
},
{
"epoch": 0.6630035518047418,
"grad_norm": 40741.23828125,
"learning_rate": 0.00032500000000000004,
"loss": 4.4826,
"step": 780
},
{
"epoch": 0.6715035973407001,
"grad_norm": 25295.03515625,
"learning_rate": 0.0003291666666666667,
"loss": 4.4845,
"step": 790
},
{
"epoch": 0.6800036428766583,
"grad_norm": 49988.22265625,
"learning_rate": 0.0003333333333333333,
"loss": 4.4814,
"step": 800
},
{
"epoch": 0.6885036884126164,
"grad_norm": 32892.44921875,
"learning_rate": 0.0003375,
"loss": 4.4826,
"step": 810
},
{
"epoch": 0.6885036884126164,
"eval_accuracy": 0.29046815360567907,
"eval_loss": 4.479104518890381,
"eval_runtime": 1276.8746,
"eval_samples_per_second": 391.094,
"eval_steps_per_second": 1.358,
"step": 810
},
{
"epoch": 0.6970037339485747,
"grad_norm": 32854.1015625,
"learning_rate": 0.00034166666666666666,
"loss": 4.4804,
"step": 820
},
{
"epoch": 0.7055037794845329,
"grad_norm": 31755.115234375,
"learning_rate": 0.00034583333333333335,
"loss": 4.4803,
"step": 830
},
{
"epoch": 0.7140038250204912,
"grad_norm": 26255.55078125,
"learning_rate": 0.00035,
"loss": 4.4767,
"step": 840
},
{
"epoch": 0.7225038705564494,
"grad_norm": 23372.48046875,
"learning_rate": 0.0003541666666666667,
"loss": 4.477,
"step": 850
},
{
"epoch": 0.7310039160924077,
"grad_norm": 37866.93359375,
"learning_rate": 0.00035833333333333333,
"loss": 4.4765,
"step": 860
},
{
"epoch": 0.7395039616283658,
"grad_norm": 26253.822265625,
"learning_rate": 0.0003625,
"loss": 4.4738,
"step": 870
},
{
"epoch": 0.7480040071643241,
"grad_norm": 25544.201171875,
"learning_rate": 0.00036666666666666667,
"loss": 4.4745,
"step": 880
},
{
"epoch": 0.7565040527002823,
"grad_norm": 25811.166015625,
"learning_rate": 0.00037083333333333337,
"loss": 4.4734,
"step": 890
},
{
"epoch": 0.7650040982362406,
"grad_norm": 27087.75390625,
"learning_rate": 0.000375,
"loss": 4.4737,
"step": 900
},
{
"epoch": 0.7650040982362406,
"eval_accuracy": 0.2906699089310573,
"eval_loss": 4.468409538269043,
"eval_runtime": 1272.8084,
"eval_samples_per_second": 392.343,
"eval_steps_per_second": 1.362,
"step": 900
},
{
"epoch": 0.7735041437721988,
"grad_norm": 34065.42578125,
"learning_rate": 0.00037916666666666665,
"loss": 4.4712,
"step": 910
},
{
"epoch": 0.782004189308157,
"grad_norm": 20391.37109375,
"learning_rate": 0.00038333333333333334,
"loss": 4.4627,
"step": 920
},
{
"epoch": 0.7905042348441152,
"grad_norm": 22009.44140625,
"learning_rate": 0.00038750000000000004,
"loss": 4.4501,
"step": 930
},
{
"epoch": 0.7990042803800734,
"grad_norm": 37206.39453125,
"learning_rate": 0.0003916666666666667,
"loss": 4.4388,
"step": 940
},
{
"epoch": 0.8075043259160317,
"grad_norm": 51573.41015625,
"learning_rate": 0.0003958333333333333,
"loss": 4.4256,
"step": 950
},
{
"epoch": 0.8160043714519899,
"grad_norm": 47200.921875,
"learning_rate": 0.0004,
"loss": 4.4164,
"step": 960
},
{
"epoch": 0.8245044169879482,
"grad_norm": 34512.5234375,
"learning_rate": 0.00040416666666666666,
"loss": 4.4093,
"step": 970
},
{
"epoch": 0.8330044625239064,
"grad_norm": 38512.90234375,
"learning_rate": 0.00040833333333333336,
"loss": 4.4011,
"step": 980
},
{
"epoch": 0.8415045080598647,
"grad_norm": 44749.2109375,
"learning_rate": 0.0004125,
"loss": 4.3852,
"step": 990
},
{
"epoch": 0.8415045080598647,
"eval_accuracy": 0.2906419834101783,
"eval_loss": 4.365363597869873,
"eval_runtime": 1274.1773,
"eval_samples_per_second": 391.922,
"eval_steps_per_second": 1.361,
"step": 990
},
{
"epoch": 0.8500045535958228,
"grad_norm": 43667.578125,
"learning_rate": 0.0004166666666666667,
"loss": 4.3701,
"step": 1000
},
{
"epoch": 0.858504599131781,
"grad_norm": 55209.21875,
"learning_rate": 0.00042083333333333333,
"loss": 4.3609,
"step": 1010
},
{
"epoch": 0.8670046446677393,
"grad_norm": 60726.68359375,
"learning_rate": 0.000425,
"loss": 4.3476,
"step": 1020
},
{
"epoch": 0.8755046902036975,
"grad_norm": 49279.09765625,
"learning_rate": 0.00042916666666666667,
"loss": 4.3382,
"step": 1030
},
{
"epoch": 0.8840047357396558,
"grad_norm": 44572.08203125,
"learning_rate": 0.00043333333333333337,
"loss": 4.3271,
"step": 1040
},
{
"epoch": 0.892504781275614,
"grad_norm": 70982.9140625,
"learning_rate": 0.0004375,
"loss": 4.3161,
"step": 1050
},
{
"epoch": 0.9010048268115722,
"grad_norm": 49678.1796875,
"learning_rate": 0.00044166666666666665,
"loss": 4.3032,
"step": 1060
},
{
"epoch": 0.9095048723475304,
"grad_norm": 42499.19140625,
"learning_rate": 0.00044583333333333335,
"loss": 4.2959,
"step": 1070
},
{
"epoch": 0.9180049178834887,
"grad_norm": 61131.90625,
"learning_rate": 0.00045000000000000004,
"loss": 4.2901,
"step": 1080
},
{
"epoch": 0.9180049178834887,
"eval_accuracy": 0.2915224046693052,
"eval_loss": 4.254197597503662,
"eval_runtime": 1275.3488,
"eval_samples_per_second": 391.562,
"eval_steps_per_second": 1.36,
"step": 1080
},
{
"epoch": 0.9265049634194469,
"grad_norm": 61474.8046875,
"learning_rate": 0.0004541666666666667,
"loss": 4.2782,
"step": 1090
},
{
"epoch": 0.9350050089554052,
"grad_norm": 81287.625,
"learning_rate": 0.0004583333333333333,
"loss": 4.2578,
"step": 1100
},
{
"epoch": 0.9435050544913633,
"grad_norm": 71441.078125,
"learning_rate": 0.0004625,
"loss": 4.2455,
"step": 1110
},
{
"epoch": 0.9520051000273215,
"grad_norm": 52111.4921875,
"learning_rate": 0.00046666666666666666,
"loss": 4.237,
"step": 1120
},
{
"epoch": 0.9605051455632798,
"grad_norm": 79908.65625,
"learning_rate": 0.00047083333333333336,
"loss": 4.2271,
"step": 1130
},
{
"epoch": 0.969005191099238,
"grad_norm": 57955.83203125,
"learning_rate": 0.000475,
"loss": 4.2194,
"step": 1140
},
{
"epoch": 0.9775052366351963,
"grad_norm": 52858.40234375,
"learning_rate": 0.0004791666666666667,
"loss": 4.2118,
"step": 1150
},
{
"epoch": 0.9860052821711545,
"grad_norm": 78292.59375,
"learning_rate": 0.00048333333333333334,
"loss": 4.2114,
"step": 1160
},
{
"epoch": 0.9945053277071128,
"grad_norm": 69718.5859375,
"learning_rate": 0.0004875,
"loss": 4.1945,
"step": 1170
},
{
"epoch": 0.9945053277071128,
"eval_accuracy": 0.29390537675129147,
"eval_loss": 4.130845069885254,
"eval_runtime": 1275.9587,
"eval_samples_per_second": 391.375,
"eval_steps_per_second": 1.359,
"step": 1170
},
{
"epoch": 1.003005373243071,
"grad_norm": 75537.1875,
"learning_rate": 0.0004916666666666666,
"loss": 4.1849,
"step": 1180
},
{
"epoch": 1.011505418779029,
"grad_norm": 81501.7734375,
"learning_rate": 0.0004958333333333334,
"loss": 4.1814,
"step": 1190
},
{
"epoch": 1.0200054643149874,
"grad_norm": 148886.71875,
"learning_rate": 0.0005,
"loss": 4.1642,
"step": 1200
},
{
"epoch": 1.0285055098509457,
"grad_norm": 78217.7109375,
"learning_rate": 0.0005041666666666667,
"loss": 4.1625,
"step": 1210
},
{
"epoch": 1.0370055553869038,
"grad_norm": 74189.0234375,
"learning_rate": 0.0005083333333333333,
"loss": 4.141,
"step": 1220
},
{
"epoch": 1.045505600922862,
"grad_norm": 105369.5546875,
"learning_rate": 0.0005124999999999999,
"loss": 4.1294,
"step": 1230
},
{
"epoch": 1.0540056464588203,
"grad_norm": 105365.2109375,
"learning_rate": 0.0005166666666666667,
"loss": 4.123,
"step": 1240
},
{
"epoch": 1.0625056919947786,
"grad_norm": 81112.6328125,
"learning_rate": 0.0005208333333333334,
"loss": 4.1089,
"step": 1250
},
{
"epoch": 1.0710057375307367,
"grad_norm": 115335.5859375,
"learning_rate": 0.0005250000000000001,
"loss": 4.1027,
"step": 1260
},
{
"epoch": 1.0710057375307367,
"eval_accuracy": 0.3011365978433179,
"eval_loss": 4.004420757293701,
"eval_runtime": 1276.3722,
"eval_samples_per_second": 391.248,
"eval_steps_per_second": 1.359,
"step": 1260
},
{
"epoch": 1.079505783066695,
"grad_norm": 108373.8359375,
"learning_rate": 0.0005291666666666667,
"loss": 4.0735,
"step": 1270
},
{
"epoch": 1.0880058286026533,
"grad_norm": 112014.7890625,
"learning_rate": 0.0005333333333333334,
"loss": 4.0363,
"step": 1280
},
{
"epoch": 1.0965058741386116,
"grad_norm": 153617.40625,
"learning_rate": 0.0005375,
"loss": 4.0039,
"step": 1290
},
{
"epoch": 1.1050059196745696,
"grad_norm": 124114.390625,
"learning_rate": 0.0005416666666666666,
"loss": 3.9605,
"step": 1300
},
{
"epoch": 1.113505965210528,
"grad_norm": 146727.78125,
"learning_rate": 0.0005458333333333333,
"loss": 3.8962,
"step": 1310
},
{
"epoch": 1.1220060107464862,
"grad_norm": 178636.21875,
"learning_rate": 0.00055,
"loss": 3.8446,
"step": 1320
},
{
"epoch": 1.1305060562824445,
"grad_norm": 177699.4375,
"learning_rate": 0.0005541666666666667,
"loss": 3.8045,
"step": 1330
},
{
"epoch": 1.1390061018184026,
"grad_norm": 121470.7890625,
"learning_rate": 0.0005583333333333333,
"loss": 3.7662,
"step": 1340
},
{
"epoch": 1.1475061473543609,
"grad_norm": 185655.078125,
"learning_rate": 0.0005625000000000001,
"loss": 3.7529,
"step": 1350
},
{
"epoch": 1.1475061473543609,
"eval_accuracy": 0.34186952633889467,
"eval_loss": 3.526193380355835,
"eval_runtime": 1274.9146,
"eval_samples_per_second": 391.695,
"eval_steps_per_second": 1.36,
"step": 1350
},
{
"epoch": 1.1560061928903191,
"grad_norm": 194355.203125,
"learning_rate": 0.0005666666666666667,
"loss": 3.6921,
"step": 1360
},
{
"epoch": 1.1645062384262772,
"grad_norm": 102708.9375,
"learning_rate": 0.0005708333333333333,
"loss": 3.6505,
"step": 1370
},
{
"epoch": 1.1730062839622355,
"grad_norm": 153301.15625,
"learning_rate": 0.000575,
"loss": 3.6163,
"step": 1380
},
{
"epoch": 1.1815063294981938,
"grad_norm": 158013.796875,
"learning_rate": 0.0005791666666666667,
"loss": 3.5602,
"step": 1390
},
{
"epoch": 1.190006375034152,
"grad_norm": 113413.78125,
"learning_rate": 0.0005833333333333334,
"loss": 3.5185,
"step": 1400
},
{
"epoch": 1.1985064205701101,
"grad_norm": 157948.4375,
"learning_rate": 0.0005875,
"loss": 3.4583,
"step": 1410
},
{
"epoch": 1.2070064661060684,
"grad_norm": 157980.484375,
"learning_rate": 0.0005916666666666667,
"loss": 3.3973,
"step": 1420
},
{
"epoch": 1.2155065116420267,
"grad_norm": 100785.5234375,
"learning_rate": 0.0005958333333333333,
"loss": 3.3448,
"step": 1430
},
{
"epoch": 1.2240065571779848,
"grad_norm": 146644.875,
"learning_rate": 0.0006,
"loss": 3.3135,
"step": 1440
},
{
"epoch": 1.2240065571779848,
"eval_accuracy": 0.40612111173593607,
"eval_loss": 3.1175479888916016,
"eval_runtime": 1275.6913,
"eval_samples_per_second": 391.457,
"eval_steps_per_second": 1.359,
"step": 1440
},
{
"epoch": 1.232506602713943,
"grad_norm": 169930.046875,
"learning_rate": 0.0006041666666666666,
"loss": 3.31,
"step": 1450
},
{
"epoch": 1.2410066482499014,
"grad_norm": 166900.796875,
"learning_rate": 0.0006083333333333333,
"loss": 3.2813,
"step": 1460
},
{
"epoch": 1.2495066937858597,
"grad_norm": 142726.859375,
"learning_rate": 0.0006125000000000001,
"loss": 3.2236,
"step": 1470
},
{
"epoch": 1.2580067393218177,
"grad_norm": 113215.296875,
"learning_rate": 0.0006166666666666667,
"loss": 3.1767,
"step": 1480
},
{
"epoch": 1.266506784857776,
"grad_norm": 150677.703125,
"learning_rate": 0.0006208333333333334,
"loss": 3.1548,
"step": 1490
},
{
"epoch": 1.2750068303937343,
"grad_norm": 134445.515625,
"learning_rate": 0.000625,
"loss": 3.0924,
"step": 1500
},
{
"epoch": 1.2835068759296924,
"grad_norm": 116583.6796875,
"learning_rate": 0.0006291666666666667,
"loss": 3.023,
"step": 1510
},
{
"epoch": 1.2920069214656507,
"grad_norm": 99715.9765625,
"learning_rate": 0.0006333333333333333,
"loss": 2.9354,
"step": 1520
},
{
"epoch": 1.300506967001609,
"grad_norm": 97103.8125,
"learning_rate": 0.0006374999999999999,
"loss": 2.8965,
"step": 1530
},
{
"epoch": 1.300506967001609,
"eval_accuracy": 0.46788244925851397,
"eval_loss": 2.698025941848755,
"eval_runtime": 1277.5104,
"eval_samples_per_second": 390.899,
"eval_steps_per_second": 1.357,
"step": 1530
},
{
"epoch": 1.3090070125375672,
"grad_norm": 87332.8515625,
"learning_rate": 0.0006416666666666667,
"loss": 2.8339,
"step": 1540
},
{
"epoch": 1.3175070580735255,
"grad_norm": 89923.7890625,
"learning_rate": 0.0006458333333333334,
"loss": 2.7669,
"step": 1550
},
{
"epoch": 1.3260071036094836,
"grad_norm": 75226.8515625,
"learning_rate": 0.0006500000000000001,
"loss": 2.7303,
"step": 1560
},
{
"epoch": 1.3345071491454419,
"grad_norm": 66864.1484375,
"learning_rate": 0.0006541666666666667,
"loss": 2.6759,
"step": 1570
},
{
"epoch": 1.3430071946814,
"grad_norm": 61924.27734375,
"learning_rate": 0.0006583333333333334,
"loss": 2.6349,
"step": 1580
},
{
"epoch": 1.3515072402173582,
"grad_norm": 54619.734375,
"learning_rate": 0.0006625,
"loss": 2.5945,
"step": 1590
},
{
"epoch": 1.3600072857533165,
"grad_norm": 74138.375,
"learning_rate": 0.0006666666666666666,
"loss": 2.5633,
"step": 1600
},
{
"epoch": 1.3685073312892748,
"grad_norm": 58250.90625,
"learning_rate": 0.0006708333333333333,
"loss": 2.5349,
"step": 1610
},
{
"epoch": 1.377007376825233,
"grad_norm": 76097.328125,
"learning_rate": 0.000675,
"loss": 2.4952,
"step": 1620
},
{
"epoch": 1.377007376825233,
"eval_accuracy": 0.5196907758785413,
"eval_loss": 2.3480491638183594,
"eval_runtime": 1280.8752,
"eval_samples_per_second": 389.872,
"eval_steps_per_second": 1.354,
"step": 1620
},
{
"epoch": 1.3855074223611912,
"grad_norm": 77774.96875,
"learning_rate": 0.0006791666666666667,
"loss": 2.4686,
"step": 1630
},
{
"epoch": 1.3940074678971495,
"grad_norm": 65427.421875,
"learning_rate": 0.0006833333333333333,
"loss": 2.4615,
"step": 1640
},
{
"epoch": 1.4025075134331078,
"grad_norm": 56960.09765625,
"learning_rate": 0.0006875,
"loss": 2.4228,
"step": 1650
},
{
"epoch": 1.4110075589690658,
"grad_norm": 74808.4296875,
"learning_rate": 0.0006916666666666667,
"loss": 2.391,
"step": 1660
},
{
"epoch": 1.4195076045050241,
"grad_norm": 47526.44140625,
"learning_rate": 0.0006958333333333334,
"loss": 2.3693,
"step": 1670
},
{
"epoch": 1.4280076500409824,
"grad_norm": 54673.7578125,
"learning_rate": 0.0007,
"loss": 2.3505,
"step": 1680
},
{
"epoch": 1.4365076955769407,
"grad_norm": 33447.546875,
"learning_rate": 0.0007041666666666667,
"loss": 2.3205,
"step": 1690
},
{
"epoch": 1.4450077411128988,
"grad_norm": 56293.46484375,
"learning_rate": 0.0007083333333333334,
"loss": 2.318,
"step": 1700
},
{
"epoch": 1.453507786648857,
"grad_norm": 41077.09375,
"learning_rate": 0.0007125,
"loss": 2.2838,
"step": 1710
},
{
"epoch": 1.453507786648857,
"eval_accuracy": 0.5489981207105539,
"eval_loss": 2.1566824913024902,
"eval_runtime": 1279.9042,
"eval_samples_per_second": 390.168,
"eval_steps_per_second": 1.355,
"step": 1710
},
{
"epoch": 1.4620078321848153,
"grad_norm": 51759.4296875,
"learning_rate": 0.0007166666666666667,
"loss": 2.264,
"step": 1720
},
{
"epoch": 1.4705078777207734,
"grad_norm": 50884.80078125,
"learning_rate": 0.0007208333333333333,
"loss": 2.2408,
"step": 1730
},
{
"epoch": 1.4790079232567317,
"grad_norm": 38072.73046875,
"learning_rate": 0.000725,
"loss": 2.2221,
"step": 1740
},
{
"epoch": 1.48750796879269,
"grad_norm": 45036.5,
"learning_rate": 0.0007291666666666666,
"loss": 2.2046,
"step": 1750
},
{
"epoch": 1.4960080143286483,
"grad_norm": 30498.197265625,
"learning_rate": 0.0007333333333333333,
"loss": 2.1925,
"step": 1760
},
{
"epoch": 1.5045080598646066,
"grad_norm": 52624.26953125,
"learning_rate": 0.0007375000000000001,
"loss": 2.1845,
"step": 1770
},
{
"epoch": 1.5130081054005646,
"grad_norm": 39472.8125,
"learning_rate": 0.0007416666666666667,
"loss": 2.1579,
"step": 1780
},
{
"epoch": 1.521508150936523,
"grad_norm": 43868.4375,
"learning_rate": 0.0007458333333333334,
"loss": 2.1517,
"step": 1790
},
{
"epoch": 1.530008196472481,
"grad_norm": 39694.20703125,
"learning_rate": 0.00075,
"loss": 2.1324,
"step": 1800
},
{
"epoch": 1.530008196472481,
"eval_accuracy": 0.571443558749772,
"eval_loss": 2.0166752338409424,
"eval_runtime": 1279.8418,
"eval_samples_per_second": 390.187,
"eval_steps_per_second": 1.355,
"step": 1800
},
{
"epoch": 1.5385082420084393,
"grad_norm": 41399.078125,
"learning_rate": 0.0007541666666666667,
"loss": 2.1129,
"step": 1810
},
{
"epoch": 1.5470082875443976,
"grad_norm": 39988.51171875,
"learning_rate": 0.0007583333333333333,
"loss": 2.0979,
"step": 1820
},
{
"epoch": 1.5555083330803559,
"grad_norm": 40106.96875,
"learning_rate": 0.0007624999999999999,
"loss": 2.0861,
"step": 1830
},
{
"epoch": 1.5640083786163141,
"grad_norm": 37416.859375,
"learning_rate": 0.0007666666666666667,
"loss": 2.0691,
"step": 1840
},
{
"epoch": 1.5725084241522722,
"grad_norm": 45897.46484375,
"learning_rate": 0.0007708333333333334,
"loss": 2.0541,
"step": 1850
},
{
"epoch": 1.5810084696882305,
"grad_norm": 43155.91796875,
"learning_rate": 0.0007750000000000001,
"loss": 2.0409,
"step": 1860
},
{
"epoch": 1.5895085152241886,
"grad_norm": 44230.0703125,
"learning_rate": 0.0007791666666666667,
"loss": 2.026,
"step": 1870
},
{
"epoch": 1.5980085607601469,
"grad_norm": 41950.99609375,
"learning_rate": 0.0007833333333333334,
"loss": 2.007,
"step": 1880
},
{
"epoch": 1.6065086062961051,
"grad_norm": 50098.26171875,
"learning_rate": 0.0007875,
"loss": 2.004,
"step": 1890
},
{
"epoch": 1.6065086062961051,
"eval_accuracy": 0.5934511080462407,
"eval_loss": 1.89019775390625,
"eval_runtime": 1279.3367,
"eval_samples_per_second": 390.341,
"eval_steps_per_second": 1.355,
"step": 1890
},
{
"epoch": 1.6150086518320634,
"grad_norm": 38339.48046875,
"learning_rate": 0.0007916666666666666,
"loss": 1.992,
"step": 1900
},
{
"epoch": 1.6235086973680217,
"grad_norm": 45457.52734375,
"learning_rate": 0.0007958333333333333,
"loss": 1.9782,
"step": 1910
},
{
"epoch": 1.6320087429039798,
"grad_norm": 44519.359375,
"learning_rate": 0.0008,
"loss": 1.9713,
"step": 1920
},
{
"epoch": 1.640508788439938,
"grad_norm": 32601.9921875,
"learning_rate": 0.0008041666666666667,
"loss": 1.9564,
"step": 1930
},
{
"epoch": 1.6490088339758961,
"grad_norm": 38710.42578125,
"learning_rate": 0.0008083333333333333,
"loss": 1.9461,
"step": 1940
},
{
"epoch": 1.6575088795118544,
"grad_norm": 36938.29296875,
"learning_rate": 0.0008125000000000001,
"loss": 1.9386,
"step": 1950
},
{
"epoch": 1.6660089250478127,
"grad_norm": 39862.9609375,
"learning_rate": 0.0008166666666666667,
"loss": 1.9229,
"step": 1960
},
{
"epoch": 1.674508970583771,
"grad_norm": 29285.794921875,
"learning_rate": 0.0008208333333333334,
"loss": 1.9096,
"step": 1970
},
{
"epoch": 1.6830090161197293,
"grad_norm": 40694.77734375,
"learning_rate": 0.000825,
"loss": 1.8992,
"step": 1980
},
{
"epoch": 1.6830090161197293,
"eval_accuracy": 0.6102293333175925,
"eval_loss": 1.7952998876571655,
"eval_runtime": 1281.484,
"eval_samples_per_second": 389.687,
"eval_steps_per_second": 1.353,
"step": 1980
},
{
"epoch": 1.6915090616556876,
"grad_norm": 52670.69140625,
"learning_rate": 0.0008291666666666667,
"loss": 1.8874,
"step": 1990
},
{
"epoch": 1.7000091071916457,
"grad_norm": 41263.34375,
"learning_rate": 0.0008333333333333334,
"loss": 1.8832,
"step": 2000
},
{
"epoch": 1.708509152727604,
"grad_norm": 48928.69921875,
"learning_rate": 0.0008375,
"loss": 1.8709,
"step": 2010
},
{
"epoch": 1.717009198263562,
"grad_norm": 34242.77734375,
"learning_rate": 0.0008416666666666667,
"loss": 1.8672,
"step": 2020
},
{
"epoch": 1.7255092437995203,
"grad_norm": 43822.64453125,
"learning_rate": 0.0008458333333333333,
"loss": 1.8542,
"step": 2030
},
{
"epoch": 1.7340092893354786,
"grad_norm": 30140.544921875,
"learning_rate": 0.00085,
"loss": 1.8418,
"step": 2040
},
{
"epoch": 1.7425093348714369,
"grad_norm": 32354.177734375,
"learning_rate": 0.0008541666666666666,
"loss": 1.833,
"step": 2050
},
{
"epoch": 1.7510093804073952,
"grad_norm": 37106.171875,
"learning_rate": 0.0008583333333333333,
"loss": 1.8285,
"step": 2060
},
{
"epoch": 1.7595094259433532,
"grad_norm": 35038.6875,
"learning_rate": 0.0008625000000000001,
"loss": 1.815,
"step": 2070
},
{
"epoch": 1.7595094259433532,
"eval_accuracy": 0.6255571282197189,
"eval_loss": 1.7101428508758545,
"eval_runtime": 1279.0851,
"eval_samples_per_second": 390.418,
"eval_steps_per_second": 1.356,
"step": 2070
},
{
"epoch": 1.7680094714793115,
"grad_norm": 33263.3046875,
"learning_rate": 0.0008666666666666667,
"loss": 1.8021,
"step": 2080
},
{
"epoch": 1.7765095170152696,
"grad_norm": 29019.279296875,
"learning_rate": 0.0008708333333333334,
"loss": 1.7913,
"step": 2090
},
{
"epoch": 1.785009562551228,
"grad_norm": 40720.78125,
"learning_rate": 0.000875,
"loss": 1.7823,
"step": 2100
},
{
"epoch": 1.7935096080871862,
"grad_norm": 40485.59765625,
"learning_rate": 0.0008791666666666667,
"loss": 1.7738,
"step": 2110
},
{
"epoch": 1.8020096536231445,
"grad_norm": 34836.8984375,
"learning_rate": 0.0008833333333333333,
"loss": 1.7628,
"step": 2120
},
{
"epoch": 1.8105096991591028,
"grad_norm": 37340.25,
"learning_rate": 0.0008874999999999999,
"loss": 1.7564,
"step": 2130
},
{
"epoch": 1.8190097446950608,
"grad_norm": 34712.98828125,
"learning_rate": 0.0008916666666666667,
"loss": 1.7454,
"step": 2140
},
{
"epoch": 1.8275097902310191,
"grad_norm": 36124.44921875,
"learning_rate": 0.0008958333333333334,
"loss": 1.7375,
"step": 2150
},
{
"epoch": 1.8360098357669772,
"grad_norm": 26800.90234375,
"learning_rate": 0.0009000000000000001,
"loss": 1.724,
"step": 2160
},
{
"epoch": 1.8360098357669772,
"eval_accuracy": 0.6417359150220479,
"eval_loss": 1.619744062423706,
"eval_runtime": 1279.7377,
"eval_samples_per_second": 390.219,
"eval_steps_per_second": 1.355,
"step": 2160
},
{
"epoch": 1.8445098813029355,
"grad_norm": 29157.935546875,
"learning_rate": 0.0009041666666666667,
"loss": 1.714,
"step": 2170
},
{
"epoch": 1.8530099268388938,
"grad_norm": 32622.201171875,
"learning_rate": 0.0009083333333333334,
"loss": 1.6972,
"step": 2180
},
{
"epoch": 1.861509972374852,
"grad_norm": 27361.466796875,
"learning_rate": 0.0009125,
"loss": 1.6807,
"step": 2190
},
{
"epoch": 1.8700100179108103,
"grad_norm": 30800.376953125,
"learning_rate": 0.0009166666666666666,
"loss": 1.6644,
"step": 2200
},
{
"epoch": 1.8785100634467686,
"grad_norm": 28547.041015625,
"learning_rate": 0.0009208333333333333,
"loss": 1.6546,
"step": 2210
},
{
"epoch": 1.8870101089827267,
"grad_norm": 32673.380859375,
"learning_rate": 0.000925,
"loss": 1.6407,
"step": 2220
},
{
"epoch": 1.8955101545186848,
"grad_norm": 38690.15625,
"learning_rate": 0.0009291666666666667,
"loss": 1.6327,
"step": 2230
},
{
"epoch": 1.904010200054643,
"grad_norm": 24017.017578125,
"learning_rate": 0.0009333333333333333,
"loss": 1.6185,
"step": 2240
},
{
"epoch": 1.9125102455906013,
"grad_norm": 33731.96484375,
"learning_rate": 0.0009375,
"loss": 1.6115,
"step": 2250
},
{
"epoch": 1.9125102455906013,
"eval_accuracy": 0.6670528637301408,
"eval_loss": 1.5023800134658813,
"eval_runtime": 1278.3406,
"eval_samples_per_second": 390.646,
"eval_steps_per_second": 1.356,
"step": 2250
},
{
"epoch": 1.9210102911265596,
"grad_norm": 35314.43359375,
"learning_rate": 0.0009416666666666667,
"loss": 1.5993,
"step": 2260
},
{
"epoch": 1.929510336662518,
"grad_norm": 36152.0078125,
"learning_rate": 0.0009458333333333334,
"loss": 1.5944,
"step": 2270
},
{
"epoch": 1.9380103821984762,
"grad_norm": 31576.48046875,
"learning_rate": 0.00095,
"loss": 1.5867,
"step": 2280
},
{
"epoch": 1.9465104277344343,
"grad_norm": 31250.708984375,
"learning_rate": 0.0009541666666666667,
"loss": 1.5779,
"step": 2290
},
{
"epoch": 1.9550104732703926,
"grad_norm": 39243.34375,
"learning_rate": 0.0009583333333333334,
"loss": 1.5693,
"step": 2300
},
{
"epoch": 1.9635105188063506,
"grad_norm": 26391.3046875,
"learning_rate": 0.0009625,
"loss": 1.5671,
"step": 2310
},
{
"epoch": 1.972010564342309,
"grad_norm": 34276.82421875,
"learning_rate": 0.0009666666666666667,
"loss": 1.56,
"step": 2320
},
{
"epoch": 1.9805106098782672,
"grad_norm": 28257.998046875,
"learning_rate": 0.0009708333333333333,
"loss": 1.5482,
"step": 2330
},
{
"epoch": 1.9890106554142255,
"grad_norm": 38320.5234375,
"learning_rate": 0.000975,
"loss": 1.5435,
"step": 2340
},
{
"epoch": 1.9890106554142255,
"eval_accuracy": 0.679645168221919,
"eval_loss": 1.4393101930618286,
"eval_runtime": 1277.7056,
"eval_samples_per_second": 390.84,
"eval_steps_per_second": 1.357,
"step": 2340
},
{
"epoch": 1.9975107009501838,
"grad_norm": 30497.1171875,
"learning_rate": 0.0009791666666666666,
"loss": 1.5364,
"step": 2350
},
{
"epoch": 2.006010746486142,
"grad_norm": 30113.25,
"learning_rate": 0.0009833333333333332,
"loss": 1.5333,
"step": 2360
},
{
"epoch": 2.0145107920221,
"grad_norm": 23521.29296875,
"learning_rate": 0.0009875,
"loss": 1.528,
"step": 2370
},
{
"epoch": 2.023010837558058,
"grad_norm": 33135.26953125,
"learning_rate": 0.0009916666666666667,
"loss": 1.5211,
"step": 2380
},
{
"epoch": 2.0315108830940165,
"grad_norm": 22258.33984375,
"learning_rate": 0.0009958333333333334,
"loss": 1.5173,
"step": 2390
},
{
"epoch": 2.040010928629975,
"grad_norm": 30941.345703125,
"learning_rate": 0.001,
"loss": 1.5097,
"step": 2400
},
{
"epoch": 2.048510974165933,
"grad_norm": 27459.111328125,
"learning_rate": 0.000999009900990099,
"loss": 1.5022,
"step": 2410
},
{
"epoch": 2.0570110197018914,
"grad_norm": 23237.234375,
"learning_rate": 0.0009980198019801981,
"loss": 1.4956,
"step": 2420
},
{
"epoch": 2.0655110652378497,
"grad_norm": 24305.986328125,
"learning_rate": 0.000997029702970297,
"loss": 1.4921,
"step": 2430
},
{
"epoch": 2.0655110652378497,
"eval_accuracy": 0.6886431014072092,
"eval_loss": 1.3913538455963135,
"eval_runtime": 1278.454,
"eval_samples_per_second": 390.611,
"eval_steps_per_second": 1.356,
"step": 2430
},
{
"epoch": 2.0740111107738075,
"grad_norm": 26497.21484375,
"learning_rate": 0.000996039603960396,
"loss": 1.4843,
"step": 2440
},
{
"epoch": 2.082511156309766,
"grad_norm": 24107.4375,
"learning_rate": 0.000995049504950495,
"loss": 1.4824,
"step": 2450
},
{
"epoch": 2.091011201845724,
"grad_norm": 26006.498046875,
"learning_rate": 0.0009940594059405941,
"loss": 1.4763,
"step": 2460
},
{
"epoch": 2.0995112473816824,
"grad_norm": 27216.076171875,
"learning_rate": 0.0009930693069306932,
"loss": 1.4704,
"step": 2470
},
{
"epoch": 2.1080112929176407,
"grad_norm": 21669.0546875,
"learning_rate": 0.000992079207920792,
"loss": 1.4636,
"step": 2480
},
{
"epoch": 2.116511338453599,
"grad_norm": 23455.484375,
"learning_rate": 0.000991089108910891,
"loss": 1.4566,
"step": 2490
},
{
"epoch": 2.1250113839895572,
"grad_norm": 20500.41015625,
"learning_rate": 0.0009900990099009901,
"loss": 1.4554,
"step": 2500
},
{
"epoch": 2.133511429525515,
"grad_norm": 26141.59765625,
"learning_rate": 0.0009891089108910892,
"loss": 1.4484,
"step": 2510
},
{
"epoch": 2.1420114750614734,
"grad_norm": 23275.765625,
"learning_rate": 0.0009881188118811882,
"loss": 1.4427,
"step": 2520
},
{
"epoch": 2.1420114750614734,
"eval_accuracy": 0.6980444623562028,
"eval_loss": 1.3418840169906616,
"eval_runtime": 1280.6296,
"eval_samples_per_second": 389.947,
"eval_steps_per_second": 1.354,
"step": 2520
},
{
"epoch": 2.1505115205974317,
"grad_norm": 26020.708984375,
"learning_rate": 0.000987128712871287,
"loss": 1.4374,
"step": 2530
},
{
"epoch": 2.15901156613339,
"grad_norm": 20408.48828125,
"learning_rate": 0.000986138613861386,
"loss": 1.4302,
"step": 2540
},
{
"epoch": 2.1675116116693482,
"grad_norm": 23640.2265625,
"learning_rate": 0.0009851485148514852,
"loss": 1.4277,
"step": 2550
},
{
"epoch": 2.1760116572053065,
"grad_norm": 22099.396484375,
"learning_rate": 0.0009841584158415842,
"loss": 1.4217,
"step": 2560
},
{
"epoch": 2.184511702741265,
"grad_norm": 23317.130859375,
"learning_rate": 0.0009831683168316833,
"loss": 1.4188,
"step": 2570
},
{
"epoch": 2.193011748277223,
"grad_norm": 24106.27734375,
"learning_rate": 0.000982178217821782,
"loss": 1.4147,
"step": 2580
},
{
"epoch": 2.201511793813181,
"grad_norm": 18410.43359375,
"learning_rate": 0.0009811881188118811,
"loss": 1.4107,
"step": 2590
},
{
"epoch": 2.2100118393491393,
"grad_norm": 22841.390625,
"learning_rate": 0.0009801980198019802,
"loss": 1.4075,
"step": 2600
},
{
"epoch": 2.2185118848850975,
"grad_norm": 24554.787109375,
"learning_rate": 0.0009792079207920793,
"loss": 1.404,
"step": 2610
},
{
"epoch": 2.2185118848850975,
"eval_accuracy": 0.7051629840505044,
"eval_loss": 1.3090746402740479,
"eval_runtime": 1279.6345,
"eval_samples_per_second": 390.25,
"eval_steps_per_second": 1.355,
"step": 2610
},
{
"epoch": 2.227011930421056,
"grad_norm": 25836.71484375,
"learning_rate": 0.0009782178217821783,
"loss": 1.4039,
"step": 2620
},
{
"epoch": 2.235511975957014,
"grad_norm": 21653.619140625,
"learning_rate": 0.0009772277227722771,
"loss": 1.3946,
"step": 2630
},
{
"epoch": 2.2440120214929724,
"grad_norm": 23728.3046875,
"learning_rate": 0.0009762376237623762,
"loss": 1.3929,
"step": 2640
},
{
"epoch": 2.2525120670289307,
"grad_norm": 21733.001953125,
"learning_rate": 0.0009752475247524752,
"loss": 1.3884,
"step": 2650
},
{
"epoch": 2.261012112564889,
"grad_norm": 21566.19921875,
"learning_rate": 0.0009742574257425743,
"loss": 1.3825,
"step": 2660
},
{
"epoch": 2.269512158100847,
"grad_norm": 19529.41015625,
"learning_rate": 0.0009732673267326732,
"loss": 1.3786,
"step": 2670
},
{
"epoch": 2.278012203636805,
"grad_norm": 20555.123046875,
"learning_rate": 0.0009722772277227723,
"loss": 1.3755,
"step": 2680
},
{
"epoch": 2.2865122491727634,
"grad_norm": 19776.90234375,
"learning_rate": 0.0009712871287128712,
"loss": 1.3706,
"step": 2690
},
{
"epoch": 2.2950122947087217,
"grad_norm": 18725.46875,
"learning_rate": 0.0009702970297029703,
"loss": 1.3703,
"step": 2700
},
{
"epoch": 2.2950122947087217,
"eval_accuracy": 0.7118258434856092,
"eval_loss": 1.2756892442703247,
"eval_runtime": 1278.3225,
"eval_samples_per_second": 390.651,
"eval_steps_per_second": 1.356,
"step": 2700
},
{
"epoch": 2.30351234024468,
"grad_norm": 21781.615234375,
"learning_rate": 0.0009693069306930693,
"loss": 1.369,
"step": 2710
},
{
"epoch": 2.3120123857806383,
"grad_norm": 21352.599609375,
"learning_rate": 0.0009683168316831683,
"loss": 1.3609,
"step": 2720
},
{
"epoch": 2.3205124313165966,
"grad_norm": 19621.919921875,
"learning_rate": 0.0009673267326732673,
"loss": 1.3579,
"step": 2730
},
{
"epoch": 2.3290124768525544,
"grad_norm": 15898.189453125,
"learning_rate": 0.0009663366336633663,
"loss": 1.3585,
"step": 2740
},
{
"epoch": 2.3375125223885127,
"grad_norm": 24640.2734375,
"learning_rate": 0.0009653465346534653,
"loss": 1.3521,
"step": 2750
},
{
"epoch": 2.346012567924471,
"grad_norm": 21775.046875,
"learning_rate": 0.0009643564356435644,
"loss": 1.3496,
"step": 2760
},
{
"epoch": 2.3545126134604293,
"grad_norm": 20199.751953125,
"learning_rate": 0.0009633663366336633,
"loss": 1.3477,
"step": 2770
},
{
"epoch": 2.3630126589963876,
"grad_norm": 23324.25,
"learning_rate": 0.0009623762376237624,
"loss": 1.344,
"step": 2780
},
{
"epoch": 2.371512704532346,
"grad_norm": 23789.677734375,
"learning_rate": 0.0009613861386138613,
"loss": 1.3423,
"step": 2790
},
{
"epoch": 2.371512704532346,
"eval_accuracy": 0.7178257874607675,
"eval_loss": 1.2434872388839722,
"eval_runtime": 1276.5226,
"eval_samples_per_second": 391.202,
"eval_steps_per_second": 1.358,
"step": 2790
},
{
"epoch": 2.380012750068304,
"grad_norm": 22512.796875,
"learning_rate": 0.0009603960396039604,
"loss": 1.3394,
"step": 2800
},
{
"epoch": 2.388512795604262,
"grad_norm": 24620.826171875,
"learning_rate": 0.0009594059405940594,
"loss": 1.3331,
"step": 2810
},
{
"epoch": 2.3970128411402203,
"grad_norm": 20756.77734375,
"learning_rate": 0.0009584158415841584,
"loss": 1.3293,
"step": 2820
},
{
"epoch": 2.4055128866761786,
"grad_norm": 17418.865234375,
"learning_rate": 0.0009574257425742574,
"loss": 1.3268,
"step": 2830
},
{
"epoch": 2.414012932212137,
"grad_norm": 19701.8984375,
"learning_rate": 0.0009564356435643564,
"loss": 1.3256,
"step": 2840
},
{
"epoch": 2.422512977748095,
"grad_norm": 22757.66015625,
"learning_rate": 0.0009554455445544554,
"loss": 1.3229,
"step": 2850
},
{
"epoch": 2.4310130232840534,
"grad_norm": 24491.03515625,
"learning_rate": 0.0009544554455445545,
"loss": 1.3187,
"step": 2860
},
{
"epoch": 2.4395130688200117,
"grad_norm": 19789.939453125,
"learning_rate": 0.0009534653465346534,
"loss": 1.3176,
"step": 2870
},
{
"epoch": 2.4480131143559696,
"grad_norm": 19331.78515625,
"learning_rate": 0.0009524752475247525,
"loss": 1.3165,
"step": 2880
},
{
"epoch": 2.4480131143559696,
"eval_accuracy": 0.7224858086201374,
"eval_loss": 1.2204521894454956,
"eval_runtime": 1278.1811,
"eval_samples_per_second": 390.694,
"eval_steps_per_second": 1.357,
"step": 2880
},
{
"epoch": 2.456513159891928,
"grad_norm": 19115.720703125,
"learning_rate": 0.0009514851485148514,
"loss": 1.3118,
"step": 2890
},
{
"epoch": 2.465013205427886,
"grad_norm": 18051.21875,
"learning_rate": 0.0009504950495049505,
"loss": 1.3087,
"step": 2900
},
{
"epoch": 2.4735132509638444,
"grad_norm": 17369.791015625,
"learning_rate": 0.0009495049504950495,
"loss": 1.3088,
"step": 2910
},
{
"epoch": 2.4820132964998027,
"grad_norm": 19926.080078125,
"learning_rate": 0.0009485148514851485,
"loss": 1.3067,
"step": 2920
},
{
"epoch": 2.490513342035761,
"grad_norm": 21108.0078125,
"learning_rate": 0.0009475247524752475,
"loss": 1.3034,
"step": 2930
},
{
"epoch": 2.4990133875717193,
"grad_norm": 20779.73828125,
"learning_rate": 0.0009465346534653465,
"loss": 1.2954,
"step": 2940
},
{
"epoch": 2.507513433107677,
"grad_norm": 19840.251953125,
"learning_rate": 0.0009455445544554455,
"loss": 1.2957,
"step": 2950
},
{
"epoch": 2.5160134786436354,
"grad_norm": 21993.7734375,
"learning_rate": 0.0009445544554455446,
"loss": 1.2948,
"step": 2960
},
{
"epoch": 2.5245135241795937,
"grad_norm": 23669.6171875,
"learning_rate": 0.0009435643564356435,
"loss": 1.2932,
"step": 2970
},
{
"epoch": 2.5245135241795937,
"eval_accuracy": 0.7274937619747395,
"eval_loss": 1.196637511253357,
"eval_runtime": 1278.0042,
"eval_samples_per_second": 390.748,
"eval_steps_per_second": 1.357,
"step": 2970
},
{
"epoch": 2.533013569715552,
"grad_norm": 16925.79296875,
"learning_rate": 0.0009425742574257426,
"loss": 1.2884,
"step": 2980
},
{
"epoch": 2.5415136152515103,
"grad_norm": 16753.69921875,
"learning_rate": 0.0009415841584158415,
"loss": 1.2862,
"step": 2990
},
{
"epoch": 2.5500136607874686,
"grad_norm": 19538.154296875,
"learning_rate": 0.0009405940594059406,
"loss": 1.2843,
"step": 3000
},
{
"epoch": 2.558513706323427,
"grad_norm": 18874.037109375,
"learning_rate": 0.0009396039603960396,
"loss": 1.2821,
"step": 3010
},
{
"epoch": 2.5670137518593847,
"grad_norm": 19626.3671875,
"learning_rate": 0.0009386138613861386,
"loss": 1.2793,
"step": 3020
},
{
"epoch": 2.575513797395343,
"grad_norm": 21052.349609375,
"learning_rate": 0.0009376237623762376,
"loss": 1.2789,
"step": 3030
},
{
"epoch": 2.5840138429313013,
"grad_norm": 19245.396484375,
"learning_rate": 0.0009366336633663367,
"loss": 1.2744,
"step": 3040
},
{
"epoch": 2.5925138884672596,
"grad_norm": 16579.640625,
"learning_rate": 0.0009356435643564357,
"loss": 1.2726,
"step": 3050
},
{
"epoch": 2.601013934003218,
"grad_norm": 17490.4609375,
"learning_rate": 0.0009346534653465348,
"loss": 1.2726,
"step": 3060
},
{
"epoch": 2.601013934003218,
"eval_accuracy": 0.7312536295087113,
"eval_loss": 1.1788941621780396,
"eval_runtime": 1282.3919,
"eval_samples_per_second": 389.411,
"eval_steps_per_second": 1.352,
"step": 3060
},
{
"epoch": 2.609513979539176,
"grad_norm": 21261.8125,
"learning_rate": 0.0009336633663366337,
"loss": 1.2683,
"step": 3070
},
{
"epoch": 2.6180140250751345,
"grad_norm": 21293.064453125,
"learning_rate": 0.0009326732673267328,
"loss": 1.2695,
"step": 3080
},
{
"epoch": 2.6265140706110923,
"grad_norm": 18312.087890625,
"learning_rate": 0.0009316831683168317,
"loss": 1.2651,
"step": 3090
},
{
"epoch": 2.635014116147051,
"grad_norm": 18225.01953125,
"learning_rate": 0.0009306930693069308,
"loss": 1.2608,
"step": 3100
},
{
"epoch": 2.643514161683009,
"grad_norm": 19320.2890625,
"learning_rate": 0.0009297029702970298,
"loss": 1.2628,
"step": 3110
},
{
"epoch": 2.652014207218967,
"grad_norm": 19210.060546875,
"learning_rate": 0.0009287128712871288,
"loss": 1.2602,
"step": 3120
},
{
"epoch": 2.6605142527549255,
"grad_norm": 20046.904296875,
"learning_rate": 0.0009277227722772278,
"loss": 1.2569,
"step": 3130
},
{
"epoch": 2.6690142982908838,
"grad_norm": 20291.744140625,
"learning_rate": 0.0009267326732673268,
"loss": 1.2546,
"step": 3140
},
{
"epoch": 2.677514343826842,
"grad_norm": 17438.1796875,
"learning_rate": 0.0009257425742574258,
"loss": 1.2553,
"step": 3150
},
{
"epoch": 2.677514343826842,
"eval_accuracy": 0.7340778616085779,
"eval_loss": 1.1635637283325195,
"eval_runtime": 1283.8261,
"eval_samples_per_second": 388.976,
"eval_steps_per_second": 1.351,
"step": 3150
},
{
"epoch": 2.6860143893628,
"grad_norm": 16703.447265625,
"learning_rate": 0.0009247524752475249,
"loss": 1.2526,
"step": 3160
},
{
"epoch": 2.6945144348987586,
"grad_norm": 16225.796875,
"learning_rate": 0.0009237623762376238,
"loss": 1.252,
"step": 3170
},
{
"epoch": 2.7030144804347165,
"grad_norm": 19865.849609375,
"learning_rate": 0.0009227722772277229,
"loss": 1.2472,
"step": 3180
},
{
"epoch": 2.7115145259706748,
"grad_norm": 18682.638671875,
"learning_rate": 0.0009217821782178218,
"loss": 1.2485,
"step": 3190
},
{
"epoch": 2.720014571506633,
"grad_norm": 24501.96875,
"learning_rate": 0.0009207920792079209,
"loss": 1.2447,
"step": 3200
},
{
"epoch": 2.7285146170425914,
"grad_norm": 16843.01953125,
"learning_rate": 0.0009198019801980199,
"loss": 1.2431,
"step": 3210
},
{
"epoch": 2.7370146625785496,
"grad_norm": 19249.625,
"learning_rate": 0.0009188118811881188,
"loss": 1.2404,
"step": 3220
},
{
"epoch": 2.7455147081145075,
"grad_norm": 16381.576171875,
"learning_rate": 0.0009178217821782179,
"loss": 1.2402,
"step": 3230
},
{
"epoch": 2.754014753650466,
"grad_norm": 17400.1015625,
"learning_rate": 0.0009168316831683168,
"loss": 1.2352,
"step": 3240
},
{
"epoch": 2.754014753650466,
"eval_accuracy": 0.738359822040415,
"eval_loss": 1.1412700414657593,
"eval_runtime": 1278.9235,
"eval_samples_per_second": 390.467,
"eval_steps_per_second": 1.356,
"step": 3240
},
{
"epoch": 2.762514799186424,
"grad_norm": 20167.6328125,
"learning_rate": 0.0009158415841584159,
"loss": 1.2357,
"step": 3250
},
{
"epoch": 2.7710148447223824,
"grad_norm": 17094.001953125,
"learning_rate": 0.000914851485148515,
"loss": 1.2327,
"step": 3260
},
{
"epoch": 2.7795148902583406,
"grad_norm": 19824.177734375,
"learning_rate": 0.0009138613861386139,
"loss": 1.2323,
"step": 3270
},
{
"epoch": 2.788014935794299,
"grad_norm": 20459.54296875,
"learning_rate": 0.0009128712871287129,
"loss": 1.2317,
"step": 3280
},
{
"epoch": 2.796514981330257,
"grad_norm": 19844.548828125,
"learning_rate": 0.0009118811881188119,
"loss": 1.2303,
"step": 3290
},
{
"epoch": 2.8050150268662155,
"grad_norm": 15563.7314453125,
"learning_rate": 0.0009108910891089109,
"loss": 1.2254,
"step": 3300
},
{
"epoch": 2.813515072402174,
"grad_norm": 19343.908203125,
"learning_rate": 0.00090990099009901,
"loss": 1.2253,
"step": 3310
},
{
"epoch": 2.8220151179381316,
"grad_norm": 15785.529296875,
"learning_rate": 0.0009089108910891089,
"loss": 1.2263,
"step": 3320
},
{
"epoch": 2.83051516347409,
"grad_norm": 21025.06640625,
"learning_rate": 0.000907920792079208,
"loss": 1.224,
"step": 3330
},
{
"epoch": 2.83051516347409,
"eval_accuracy": 0.7409290871538374,
"eval_loss": 1.1298062801361084,
"eval_runtime": 1281.4127,
"eval_samples_per_second": 389.709,
"eval_steps_per_second": 1.353,
"step": 3330
},
{
"epoch": 2.8390152090100482,
"grad_norm": 16538.94140625,
"learning_rate": 0.0009069306930693069,
"loss": 1.223,
"step": 3340
},
{
"epoch": 2.8475152545460065,
"grad_norm": 18982.095703125,
"learning_rate": 0.000905940594059406,
"loss": 1.2217,
"step": 3350
},
{
"epoch": 2.856015300081965,
"grad_norm": 18279.283203125,
"learning_rate": 0.000904950495049505,
"loss": 1.2178,
"step": 3360
},
{
"epoch": 2.864515345617923,
"grad_norm": 19073.97265625,
"learning_rate": 0.000903960396039604,
"loss": 1.2142,
"step": 3370
},
{
"epoch": 2.8730153911538814,
"grad_norm": 18313.646484375,
"learning_rate": 0.000902970297029703,
"loss": 1.2164,
"step": 3380
},
{
"epoch": 2.8815154366898392,
"grad_norm": 18153.08203125,
"learning_rate": 0.000901980198019802,
"loss": 1.2132,
"step": 3390
},
{
"epoch": 2.8900154822257975,
"grad_norm": 17545.95703125,
"learning_rate": 0.000900990099009901,
"loss": 1.211,
"step": 3400
},
{
"epoch": 2.898515527761756,
"grad_norm": 17355.42578125,
"learning_rate": 0.0009000000000000001,
"loss": 1.2137,
"step": 3410
},
{
"epoch": 2.907015573297714,
"grad_norm": 16659.796875,
"learning_rate": 0.000899009900990099,
"loss": 1.2094,
"step": 3420
},
{
"epoch": 2.907015573297714,
"eval_accuracy": 0.7434043563853827,
"eval_loss": 1.1177165508270264,
"eval_runtime": 1280.5435,
"eval_samples_per_second": 389.973,
"eval_steps_per_second": 1.354,
"step": 3420
},
{
"epoch": 2.9155156188336724,
"grad_norm": 20371.392578125,
"learning_rate": 0.0008980198019801981,
"loss": 1.2082,
"step": 3430
},
{
"epoch": 2.9240156643696307,
"grad_norm": 20884.041015625,
"learning_rate": 0.000897029702970297,
"loss": 1.2049,
"step": 3440
},
{
"epoch": 2.932515709905589,
"grad_norm": 18348.6328125,
"learning_rate": 0.0008960396039603961,
"loss": 1.2043,
"step": 3450
},
{
"epoch": 2.941015755441547,
"grad_norm": 17201.962890625,
"learning_rate": 0.0008950495049504951,
"loss": 1.2023,
"step": 3460
},
{
"epoch": 2.949515800977505,
"grad_norm": 18133.63671875,
"learning_rate": 0.0008940594059405941,
"loss": 1.2026,
"step": 3470
},
{
"epoch": 2.9580158465134634,
"grad_norm": 16873.494140625,
"learning_rate": 0.0008930693069306931,
"loss": 1.1983,
"step": 3480
},
{
"epoch": 2.9665158920494217,
"grad_norm": 18968.74609375,
"learning_rate": 0.0008920792079207921,
"loss": 1.2007,
"step": 3490
},
{
"epoch": 2.97501593758538,
"grad_norm": 16066.7578125,
"learning_rate": 0.0008910891089108911,
"loss": 1.1992,
"step": 3500
},
{
"epoch": 2.9835159831213383,
"grad_norm": 18147.33203125,
"learning_rate": 0.0008900990099009902,
"loss": 1.1955,
"step": 3510
},
{
"epoch": 2.9835159831213383,
"eval_accuracy": 0.7461886735137164,
"eval_loss": 1.101768970489502,
"eval_runtime": 1279.5411,
"eval_samples_per_second": 390.279,
"eval_steps_per_second": 1.355,
"step": 3510
},
{
"epoch": 2.9920160286572965,
"grad_norm": 16867.4296875,
"learning_rate": 0.0008891089108910891,
"loss": 1.1933,
"step": 3520
},
{
"epoch": 3.0005160741932544,
"grad_norm": 17814.7109375,
"learning_rate": 0.0008881188118811882,
"loss": 1.1949,
"step": 3530
},
{
"epoch": 3.0090161197292127,
"grad_norm": 17680.919921875,
"learning_rate": 0.0008871287128712871,
"loss": 1.1922,
"step": 3540
},
{
"epoch": 3.017516165265171,
"grad_norm": 20543.033203125,
"learning_rate": 0.0008861386138613862,
"loss": 1.1924,
"step": 3550
},
{
"epoch": 3.0260162108011293,
"grad_norm": 22293.501953125,
"learning_rate": 0.0008851485148514852,
"loss": 1.1918,
"step": 3560
},
{
"epoch": 3.0345162563370875,
"grad_norm": 18199.220703125,
"learning_rate": 0.0008841584158415842,
"loss": 1.1886,
"step": 3570
},
{
"epoch": 3.043016301873046,
"grad_norm": 17805.478515625,
"learning_rate": 0.0008831683168316832,
"loss": 1.1878,
"step": 3580
},
{
"epoch": 3.051516347409004,
"grad_norm": 17259.236328125,
"learning_rate": 0.0008821782178217822,
"loss": 1.1862,
"step": 3590
},
{
"epoch": 3.0600163929449624,
"grad_norm": 15020.2236328125,
"learning_rate": 0.0008811881188118812,
"loss": 1.1859,
"step": 3600
},
{
"epoch": 3.0600163929449624,
"eval_accuracy": 0.7483353625113318,
"eval_loss": 1.0916943550109863,
"eval_runtime": 1279.6266,
"eval_samples_per_second": 390.253,
"eval_steps_per_second": 1.355,
"step": 3600
},
{
"epoch": 3.0685164384809203,
"grad_norm": 17711.4765625,
"learning_rate": 0.0008801980198019803,
"loss": 1.1849,
"step": 3610
},
{
"epoch": 3.0770164840168786,
"grad_norm": 21087.078125,
"learning_rate": 0.0008792079207920792,
"loss": 1.1835,
"step": 3620
},
{
"epoch": 3.085516529552837,
"grad_norm": 15549.55859375,
"learning_rate": 0.0008782178217821783,
"loss": 1.1821,
"step": 3630
},
{
"epoch": 3.094016575088795,
"grad_norm": 17686.115234375,
"learning_rate": 0.0008772277227722772,
"loss": 1.1793,
"step": 3640
},
{
"epoch": 3.1025166206247534,
"grad_norm": 18682.033203125,
"learning_rate": 0.0008762376237623763,
"loss": 1.1787,
"step": 3650
},
{
"epoch": 3.1110166661607117,
"grad_norm": 18502.626953125,
"learning_rate": 0.0008752475247524753,
"loss": 1.1776,
"step": 3660
},
{
"epoch": 3.11951671169667,
"grad_norm": 14834.3359375,
"learning_rate": 0.0008742574257425743,
"loss": 1.1772,
"step": 3670
},
{
"epoch": 3.128016757232628,
"grad_norm": 17454.373046875,
"learning_rate": 0.0008732673267326733,
"loss": 1.1766,
"step": 3680
},
{
"epoch": 3.136516802768586,
"grad_norm": 17697.75,
"learning_rate": 0.0008722772277227722,
"loss": 1.1764,
"step": 3690
},
{
"epoch": 3.136516802768586,
"eval_accuracy": 0.7508950932738141,
"eval_loss": 1.0795581340789795,
"eval_runtime": 1281.2745,
"eval_samples_per_second": 389.751,
"eval_steps_per_second": 1.353,
"step": 3690
},
{
"epoch": 3.1450168483045444,
"grad_norm": 20430.919921875,
"learning_rate": 0.0008712871287128713,
"loss": 1.1738,
"step": 3700
},
{
"epoch": 3.1535168938405027,
"grad_norm": 16903.4453125,
"learning_rate": 0.0008702970297029704,
"loss": 1.1709,
"step": 3710
},
{
"epoch": 3.162016939376461,
"grad_norm": 17143.486328125,
"learning_rate": 0.0008693069306930693,
"loss": 1.1698,
"step": 3720
},
{
"epoch": 3.1705169849124193,
"grad_norm": 15930.9501953125,
"learning_rate": 0.0008683168316831684,
"loss": 1.1696,
"step": 3730
},
{
"epoch": 3.1790170304483776,
"grad_norm": 16211.982421875,
"learning_rate": 0.0008673267326732673,
"loss": 1.169,
"step": 3740
},
{
"epoch": 3.1875170759843354,
"grad_norm": 19266.4140625,
"learning_rate": 0.0008663366336633663,
"loss": 1.1673,
"step": 3750
},
{
"epoch": 3.1960171215202937,
"grad_norm": 18455.83203125,
"learning_rate": 0.0008653465346534654,
"loss": 1.1682,
"step": 3760
},
{
"epoch": 3.204517167056252,
"grad_norm": 17270.205078125,
"learning_rate": 0.0008643564356435643,
"loss": 1.1662,
"step": 3770
},
{
"epoch": 3.2130172125922103,
"grad_norm": 15437.96484375,
"learning_rate": 0.0008633663366336634,
"loss": 1.1654,
"step": 3780
},
{
"epoch": 3.2130172125922103,
"eval_accuracy": 0.7525794036416251,
"eval_loss": 1.0734323263168335,
"eval_runtime": 1278.3867,
"eval_samples_per_second": 390.631,
"eval_steps_per_second": 1.356,
"step": 3780
},
{
"epoch": 3.2215172581281686,
"grad_norm": 16193.529296875,
"learning_rate": 0.0008623762376237623,
"loss": 1.1628,
"step": 3790
},
{
"epoch": 3.230017303664127,
"grad_norm": 17204.466796875,
"learning_rate": 0.0008613861386138614,
"loss": 1.1631,
"step": 3800
},
{
"epoch": 3.238517349200085,
"grad_norm": 17406.2265625,
"learning_rate": 0.0008603960396039604,
"loss": 1.1621,
"step": 3810
},
{
"epoch": 3.2470173947360434,
"grad_norm": 17054.076171875,
"learning_rate": 0.0008594059405940594,
"loss": 1.1614,
"step": 3820
},
{
"epoch": 3.2555174402720013,
"grad_norm": 18476.68359375,
"learning_rate": 0.0008584158415841584,
"loss": 1.1606,
"step": 3830
},
{
"epoch": 3.2640174858079596,
"grad_norm": 19170.802734375,
"learning_rate": 0.0008574257425742574,
"loss": 1.1581,
"step": 3840
},
{
"epoch": 3.272517531343918,
"grad_norm": 17178.326171875,
"learning_rate": 0.0008564356435643564,
"loss": 1.1562,
"step": 3850
},
{
"epoch": 3.281017576879876,
"grad_norm": 17949.396484375,
"learning_rate": 0.0008554455445544555,
"loss": 1.1557,
"step": 3860
},
{
"epoch": 3.2895176224158345,
"grad_norm": 15490.787109375,
"learning_rate": 0.0008544554455445544,
"loss": 1.1571,
"step": 3870
},
{
"epoch": 3.2895176224158345,
"eval_accuracy": 0.754430037196134,
"eval_loss": 1.0631576776504517,
"eval_runtime": 1278.1739,
"eval_samples_per_second": 390.696,
"eval_steps_per_second": 1.357,
"step": 3870
},
{
"epoch": 3.2980176679517927,
"grad_norm": 17647.640625,
"learning_rate": 0.0008534653465346535,
"loss": 1.1552,
"step": 3880
},
{
"epoch": 3.306517713487751,
"grad_norm": 16109.29296875,
"learning_rate": 0.0008524752475247524,
"loss": 1.153,
"step": 3890
},
{
"epoch": 3.315017759023709,
"grad_norm": 19355.884765625,
"learning_rate": 0.0008514851485148515,
"loss": 1.1543,
"step": 3900
},
{
"epoch": 3.323517804559667,
"grad_norm": 15843.697265625,
"learning_rate": 0.0008504950495049505,
"loss": 1.1533,
"step": 3910
},
{
"epoch": 3.3320178500956255,
"grad_norm": 20818.416015625,
"learning_rate": 0.0008495049504950495,
"loss": 1.1497,
"step": 3920
},
{
"epoch": 3.3405178956315837,
"grad_norm": 17527.841796875,
"learning_rate": 0.0008485148514851485,
"loss": 1.1508,
"step": 3930
},
{
"epoch": 3.349017941167542,
"grad_norm": 20557.205078125,
"learning_rate": 0.0008475247524752475,
"loss": 1.1508,
"step": 3940
},
{
"epoch": 3.3575179867035003,
"grad_norm": 17984.611328125,
"learning_rate": 0.0008465346534653465,
"loss": 1.1462,
"step": 3950
},
{
"epoch": 3.3660180322394586,
"grad_norm": 15756.77734375,
"learning_rate": 0.0008455445544554456,
"loss": 1.1456,
"step": 3960
},
{
"epoch": 3.3660180322394586,
"eval_accuracy": 0.7564186600937407,
"eval_loss": 1.0549671649932861,
"eval_runtime": 1283.4177,
"eval_samples_per_second": 389.1,
"eval_steps_per_second": 1.351,
"step": 3960
},
{
"epoch": 3.3745180777754165,
"grad_norm": 15412.625,
"learning_rate": 0.0008445544554455445,
"loss": 1.1451,
"step": 3970
},
{
"epoch": 3.3830181233113747,
"grad_norm": 16771.29296875,
"learning_rate": 0.0008435643564356436,
"loss": 1.1465,
"step": 3980
},
{
"epoch": 3.391518168847333,
"grad_norm": 18703.853515625,
"learning_rate": 0.0008425742574257425,
"loss": 1.1454,
"step": 3990
},
{
"epoch": 3.4000182143832913,
"grad_norm": 15575.3798828125,
"learning_rate": 0.0008415841584158416,
"loss": 1.1453,
"step": 4000
},
{
"epoch": 3.4085182599192496,
"grad_norm": 18316.923828125,
"learning_rate": 0.0008405940594059406,
"loss": 1.1408,
"step": 4010
},
{
"epoch": 3.417018305455208,
"grad_norm": 18408.486328125,
"learning_rate": 0.0008396039603960396,
"loss": 1.1432,
"step": 4020
},
{
"epoch": 3.425518350991166,
"grad_norm": 18312.7734375,
"learning_rate": 0.0008386138613861386,
"loss": 1.1404,
"step": 4030
},
{
"epoch": 3.434018396527124,
"grad_norm": 15673.939453125,
"learning_rate": 0.0008376237623762376,
"loss": 1.1406,
"step": 4040
},
{
"epoch": 3.4425184420630823,
"grad_norm": 18445.818359375,
"learning_rate": 0.0008366336633663366,
"loss": 1.1391,
"step": 4050
},
{
"epoch": 3.4425184420630823,
"eval_accuracy": 0.7579978835836059,
"eval_loss": 1.0442180633544922,
"eval_runtime": 1278.59,
"eval_samples_per_second": 390.569,
"eval_steps_per_second": 1.356,
"step": 4050
},
{
"epoch": 3.4510184875990406,
"grad_norm": 21039.810546875,
"learning_rate": 0.0008356435643564357,
"loss": 1.1395,
"step": 4060
},
{
"epoch": 3.459518533134999,
"grad_norm": 15481.697265625,
"learning_rate": 0.0008346534653465346,
"loss": 1.1379,
"step": 4070
},
{
"epoch": 3.468018578670957,
"grad_norm": 17034.869140625,
"learning_rate": 0.0008336633663366337,
"loss": 1.1365,
"step": 4080
},
{
"epoch": 3.4765186242069155,
"grad_norm": 17020.0703125,
"learning_rate": 0.0008326732673267326,
"loss": 1.1331,
"step": 4090
},
{
"epoch": 3.4850186697428738,
"grad_norm": 19914.296875,
"learning_rate": 0.0008316831683168317,
"loss": 1.1352,
"step": 4100
},
{
"epoch": 3.4935187152788316,
"grad_norm": 19380.546875,
"learning_rate": 0.0008306930693069307,
"loss": 1.1334,
"step": 4110
},
{
"epoch": 3.5020187608147904,
"grad_norm": 16126.1982421875,
"learning_rate": 0.0008297029702970297,
"loss": 1.133,
"step": 4120
},
{
"epoch": 3.510518806350748,
"grad_norm": 16914.3046875,
"learning_rate": 0.0008287128712871287,
"loss": 1.1328,
"step": 4130
},
{
"epoch": 3.5190188518867065,
"grad_norm": 15453.5771484375,
"learning_rate": 0.0008277227722772277,
"loss": 1.1302,
"step": 4140
},
{
"epoch": 3.5190188518867065,
"eval_accuracy": 0.7592944501997413,
"eval_loss": 1.0387505292892456,
"eval_runtime": 1280.9395,
"eval_samples_per_second": 389.853,
"eval_steps_per_second": 1.354,
"step": 4140
},
{
"epoch": 3.527518897422665,
"grad_norm": 17636.76171875,
"learning_rate": 0.0008267326732673267,
"loss": 1.1313,
"step": 4150
},
{
"epoch": 3.536018942958623,
"grad_norm": 16709.236328125,
"learning_rate": 0.0008257425742574258,
"loss": 1.1307,
"step": 4160
},
{
"epoch": 3.5445189884945814,
"grad_norm": 18499.865234375,
"learning_rate": 0.0008247524752475247,
"loss": 1.1294,
"step": 4170
},
{
"epoch": 3.553019034030539,
"grad_norm": 16481.85546875,
"learning_rate": 0.0008237623762376238,
"loss": 1.1276,
"step": 4180
},
{
"epoch": 3.561519079566498,
"grad_norm": 16035.626953125,
"learning_rate": 0.0008227722772277227,
"loss": 1.1259,
"step": 4190
},
{
"epoch": 3.570019125102456,
"grad_norm": 17429.3046875,
"learning_rate": 0.0008217821782178218,
"loss": 1.1281,
"step": 4200
},
{
"epoch": 3.578519170638414,
"grad_norm": 17563.748046875,
"learning_rate": 0.0008207920792079208,
"loss": 1.1244,
"step": 4210
},
{
"epoch": 3.5870192161743724,
"grad_norm": 17140.78125,
"learning_rate": 0.0008198019801980197,
"loss": 1.1264,
"step": 4220
},
{
"epoch": 3.5955192617103306,
"grad_norm": 18145.185546875,
"learning_rate": 0.0008188118811881188,
"loss": 1.1229,
"step": 4230
},
{
"epoch": 3.5955192617103306,
"eval_accuracy": 0.7607957295247235,
"eval_loss": 1.0331238508224487,
"eval_runtime": 1281.4596,
"eval_samples_per_second": 389.695,
"eval_steps_per_second": 1.353,
"step": 4230
},
{
"epoch": 3.604019307246289,
"grad_norm": 14324.3125,
"learning_rate": 0.0008178217821782177,
"loss": 1.124,
"step": 4240
},
{
"epoch": 3.612519352782247,
"grad_norm": 19569.59375,
"learning_rate": 0.0008168316831683168,
"loss": 1.124,
"step": 4250
},
{
"epoch": 3.6210193983182055,
"grad_norm": 18639.96484375,
"learning_rate": 0.0008158415841584159,
"loss": 1.1221,
"step": 4260
},
{
"epoch": 3.6295194438541634,
"grad_norm": 17077.609375,
"learning_rate": 0.0008148514851485148,
"loss": 1.1177,
"step": 4270
},
{
"epoch": 3.6380194893901217,
"grad_norm": 19580.677734375,
"learning_rate": 0.0008138613861386138,
"loss": 1.1208,
"step": 4280
},
{
"epoch": 3.64651953492608,
"grad_norm": 14986.6796875,
"learning_rate": 0.0008128712871287128,
"loss": 1.1192,
"step": 4290
},
{
"epoch": 3.6550195804620382,
"grad_norm": 16046.3740234375,
"learning_rate": 0.000811881188118812,
"loss": 1.1193,
"step": 4300
},
{
"epoch": 3.6635196259979965,
"grad_norm": 14974.083984375,
"learning_rate": 0.000810891089108911,
"loss": 1.1162,
"step": 4310
},
{
"epoch": 3.672019671533955,
"grad_norm": 17986.744140625,
"learning_rate": 0.00080990099009901,
"loss": 1.1183,
"step": 4320
},
{
"epoch": 3.672019671533955,
"eval_accuracy": 0.7619108635382325,
"eval_loss": 1.0262423753738403,
"eval_runtime": 1281.4568,
"eval_samples_per_second": 389.696,
"eval_steps_per_second": 1.353,
"step": 4320
},
{
"epoch": 3.680519717069913,
"grad_norm": 19318.87109375,
"learning_rate": 0.000808910891089109,
"loss": 1.1153,
"step": 4330
},
{
"epoch": 3.689019762605871,
"grad_norm": 18132.083984375,
"learning_rate": 0.0008079207920792079,
"loss": 1.1155,
"step": 4340
},
{
"epoch": 3.6975198081418292,
"grad_norm": 15848.3271484375,
"learning_rate": 0.000806930693069307,
"loss": 1.1172,
"step": 4350
},
{
"epoch": 3.7060198536777875,
"grad_norm": 16213.193359375,
"learning_rate": 0.000805940594059406,
"loss": 1.1136,
"step": 4360
},
{
"epoch": 3.714519899213746,
"grad_norm": 18759.615234375,
"learning_rate": 0.000804950495049505,
"loss": 1.1129,
"step": 4370
},
{
"epoch": 3.723019944749704,
"grad_norm": 19659.138671875,
"learning_rate": 0.000803960396039604,
"loss": 1.115,
"step": 4380
},
{
"epoch": 3.7315199902856624,
"grad_norm": 19299.97265625,
"learning_rate": 0.000802970297029703,
"loss": 1.1124,
"step": 4390
},
{
"epoch": 3.7400200358216207,
"grad_norm": 15511.5537109375,
"learning_rate": 0.000801980198019802,
"loss": 1.1121,
"step": 4400
},
{
"epoch": 3.7485200813575785,
"grad_norm": 18630.70703125,
"learning_rate": 0.0008009900990099011,
"loss": 1.1106,
"step": 4410
},
{
"epoch": 3.7485200813575785,
"eval_accuracy": 0.763328105961178,
"eval_loss": 1.0201424360275269,
"eval_runtime": 1277.6344,
"eval_samples_per_second": 390.861,
"eval_steps_per_second": 1.357,
"step": 4410
},
{
"epoch": 3.757020126893537,
"grad_norm": 17692.544921875,
"learning_rate": 0.0008,
"loss": 1.1116,
"step": 4420
},
{
"epoch": 3.765520172429495,
"grad_norm": 14010.234375,
"learning_rate": 0.0007990099009900991,
"loss": 1.1089,
"step": 4430
},
{
"epoch": 3.7740202179654534,
"grad_norm": 18022.48828125,
"learning_rate": 0.000798019801980198,
"loss": 1.1082,
"step": 4440
},
{
"epoch": 3.7825202635014117,
"grad_norm": 16178.5849609375,
"learning_rate": 0.0007970297029702971,
"loss": 1.108,
"step": 4450
},
{
"epoch": 3.79102030903737,
"grad_norm": 16874.7734375,
"learning_rate": 0.0007960396039603961,
"loss": 1.1071,
"step": 4460
},
{
"epoch": 3.7995203545733283,
"grad_norm": 16226.3564453125,
"learning_rate": 0.0007950495049504951,
"loss": 1.1052,
"step": 4470
},
{
"epoch": 3.808020400109286,
"grad_norm": 21136.931640625,
"learning_rate": 0.0007940594059405941,
"loss": 1.1071,
"step": 4480
},
{
"epoch": 3.8165204456452444,
"grad_norm": 19177.98828125,
"learning_rate": 0.0007930693069306931,
"loss": 1.1048,
"step": 4490
},
{
"epoch": 3.8250204911812027,
"grad_norm": 16423.876953125,
"learning_rate": 0.0007920792079207921,
"loss": 1.1021,
"step": 4500
},
{
"epoch": 3.8250204911812027,
"eval_accuracy": 0.7649368216993604,
"eval_loss": 1.0112674236297607,
"eval_runtime": 1282.4987,
"eval_samples_per_second": 389.379,
"eval_steps_per_second": 1.352,
"step": 4500
},
{
"epoch": 3.833520536717161,
"grad_norm": 15636.1162109375,
"learning_rate": 0.0007910891089108912,
"loss": 1.1007,
"step": 4510
},
{
"epoch": 3.8420205822531193,
"grad_norm": 16542.201171875,
"learning_rate": 0.0007900990099009901,
"loss": 1.1032,
"step": 4520
},
{
"epoch": 3.8505206277890776,
"grad_norm": 17437.408203125,
"learning_rate": 0.0007891089108910892,
"loss": 1.1015,
"step": 4530
},
{
"epoch": 3.859020673325036,
"grad_norm": 19502.89453125,
"learning_rate": 0.0007881188118811881,
"loss": 1.1025,
"step": 4540
},
{
"epoch": 3.8675207188609937,
"grad_norm": 18670.546875,
"learning_rate": 0.0007871287128712872,
"loss": 1.0993,
"step": 4550
},
{
"epoch": 3.876020764396952,
"grad_norm": 18770.947265625,
"learning_rate": 0.0007861386138613862,
"loss": 1.1026,
"step": 4560
},
{
"epoch": 3.8845208099329103,
"grad_norm": 17557.958984375,
"learning_rate": 0.0007851485148514852,
"loss": 1.0983,
"step": 4570
},
{
"epoch": 3.8930208554688686,
"grad_norm": 18551.2265625,
"learning_rate": 0.0007841584158415842,
"loss": 1.1001,
"step": 4580
},
{
"epoch": 3.901520901004827,
"grad_norm": 17471.43359375,
"learning_rate": 0.0007831683168316832,
"loss": 1.097,
"step": 4590
},
{
"epoch": 3.901520901004827,
"eval_accuracy": 0.7659643920785019,
"eval_loss": 1.0095082521438599,
"eval_runtime": 1281.981,
"eval_samples_per_second": 389.536,
"eval_steps_per_second": 1.353,
"step": 4590
},
{
"epoch": 3.910020946540785,
"grad_norm": 17272.111328125,
"learning_rate": 0.0007821782178217822,
"loss": 1.0978,
"step": 4600
},
{
"epoch": 3.9185209920767434,
"grad_norm": 13404.1982421875,
"learning_rate": 0.0007811881188118813,
"loss": 1.0995,
"step": 4610
},
{
"epoch": 3.9270210376127013,
"grad_norm": 14501.0966796875,
"learning_rate": 0.0007801980198019802,
"loss": 1.0989,
"step": 4620
},
{
"epoch": 3.9355210831486596,
"grad_norm": 19314.373046875,
"learning_rate": 0.0007792079207920793,
"loss": 1.0949,
"step": 4630
},
{
"epoch": 3.944021128684618,
"grad_norm": 17886.853515625,
"learning_rate": 0.0007782178217821782,
"loss": 1.097,
"step": 4640
},
{
"epoch": 3.952521174220576,
"grad_norm": 16101.6513671875,
"learning_rate": 0.0007772277227722773,
"loss": 1.0922,
"step": 4650
},
{
"epoch": 3.9610212197565344,
"grad_norm": 16483.103515625,
"learning_rate": 0.0007762376237623763,
"loss": 1.094,
"step": 4660
},
{
"epoch": 3.9695212652924927,
"grad_norm": 17740.6640625,
"learning_rate": 0.0007752475247524753,
"loss": 1.0923,
"step": 4670
},
{
"epoch": 3.978021310828451,
"grad_norm": 18332.16015625,
"learning_rate": 0.0007742574257425743,
"loss": 1.0932,
"step": 4680
},
{
"epoch": 3.978021310828451,
"eval_accuracy": 0.7671802734913508,
"eval_loss": 1.0010571479797363,
"eval_runtime": 1282.4979,
"eval_samples_per_second": 389.379,
"eval_steps_per_second": 1.352,
"step": 4680
},
{
"epoch": 3.986521356364409,
"grad_norm": 15639.021484375,
"learning_rate": 0.0007732673267326733,
"loss": 1.0928,
"step": 4690
},
{
"epoch": 3.9950214019003676,
"grad_norm": 15639.115234375,
"learning_rate": 0.0007722772277227723,
"loss": 1.0929,
"step": 4700
},
{
"epoch": 4.003521447436325,
"grad_norm": 16150.44140625,
"learning_rate": 0.0007712871287128714,
"loss": 1.0907,
"step": 4710
},
{
"epoch": 4.012021492972284,
"grad_norm": 20656.818359375,
"learning_rate": 0.0007702970297029703,
"loss": 1.0894,
"step": 4720
},
{
"epoch": 4.020521538508242,
"grad_norm": 18745.37890625,
"learning_rate": 0.0007693069306930694,
"loss": 1.0919,
"step": 4730
},
{
"epoch": 4.0290215840442,
"grad_norm": 16387.470703125,
"learning_rate": 0.0007683168316831683,
"loss": 1.09,
"step": 4740
},
{
"epoch": 4.037521629580159,
"grad_norm": 18658.521484375,
"learning_rate": 0.0007673267326732674,
"loss": 1.0873,
"step": 4750
},
{
"epoch": 4.046021675116116,
"grad_norm": 19641.572265625,
"learning_rate": 0.0007663366336633664,
"loss": 1.0886,
"step": 4760
},
{
"epoch": 4.054521720652075,
"grad_norm": 16653.2578125,
"learning_rate": 0.0007653465346534654,
"loss": 1.0873,
"step": 4770
},
{
"epoch": 4.054521720652075,
"eval_accuracy": 0.7683563302945906,
"eval_loss": 0.9970803260803223,
"eval_runtime": 1282.3032,
"eval_samples_per_second": 389.438,
"eval_steps_per_second": 1.352,
"step": 4770
},
{
"epoch": 4.063021766188033,
"grad_norm": 17606.6015625,
"learning_rate": 0.0007643564356435644,
"loss": 1.0889,
"step": 4780
},
{
"epoch": 4.071521811723992,
"grad_norm": 17704.697265625,
"learning_rate": 0.0007633663366336634,
"loss": 1.0869,
"step": 4790
},
{
"epoch": 4.08002185725995,
"grad_norm": 20424.306640625,
"learning_rate": 0.0007623762376237624,
"loss": 1.0861,
"step": 4800
},
{
"epoch": 4.088521902795907,
"grad_norm": 15488.392578125,
"learning_rate": 0.0007613861386138615,
"loss": 1.0858,
"step": 4810
},
{
"epoch": 4.097021948331866,
"grad_norm": 18205.36328125,
"learning_rate": 0.0007603960396039604,
"loss": 1.0847,
"step": 4820
},
{
"epoch": 4.105521993867824,
"grad_norm": 15557.935546875,
"learning_rate": 0.0007594059405940595,
"loss": 1.0845,
"step": 4830
},
{
"epoch": 4.114022039403783,
"grad_norm": 17741.306640625,
"learning_rate": 0.0007584158415841584,
"loss": 1.0837,
"step": 4840
},
{
"epoch": 4.122522084939741,
"grad_norm": 21305.34765625,
"learning_rate": 0.0007574257425742574,
"loss": 1.0841,
"step": 4850
},
{
"epoch": 4.131022130475699,
"grad_norm": 20442.0390625,
"learning_rate": 0.0007564356435643565,
"loss": 1.0813,
"step": 4860
},
{
"epoch": 4.131022130475699,
"eval_accuracy": 0.769627324769287,
"eval_loss": 0.9902046322822571,
"eval_runtime": 1278.8399,
"eval_samples_per_second": 390.493,
"eval_steps_per_second": 1.356,
"step": 4860
},
{
"epoch": 4.139522176011657,
"grad_norm": 19991.08203125,
"learning_rate": 0.0007554455445544554,
"loss": 1.0809,
"step": 4870
},
{
"epoch": 4.148022221547615,
"grad_norm": 21327.169921875,
"learning_rate": 0.0007544554455445545,
"loss": 1.0822,
"step": 4880
},
{
"epoch": 4.156522267083574,
"grad_norm": 20165.03515625,
"learning_rate": 0.0007534653465346534,
"loss": 1.0824,
"step": 4890
},
{
"epoch": 4.165022312619532,
"grad_norm": 15693.8251953125,
"learning_rate": 0.0007524752475247525,
"loss": 1.0786,
"step": 4900
},
{
"epoch": 4.17352235815549,
"grad_norm": 18861.962890625,
"learning_rate": 0.0007514851485148515,
"loss": 1.0802,
"step": 4910
},
{
"epoch": 4.182022403691448,
"grad_norm": 18349.630859375,
"learning_rate": 0.0007504950495049505,
"loss": 1.0811,
"step": 4920
},
{
"epoch": 4.190522449227407,
"grad_norm": 16894.869140625,
"learning_rate": 0.0007495049504950495,
"loss": 1.0784,
"step": 4930
},
{
"epoch": 4.199022494763365,
"grad_norm": 15264.0703125,
"learning_rate": 0.0007485148514851485,
"loss": 1.0779,
"step": 4940
},
{
"epoch": 4.207522540299323,
"grad_norm": 18500.328125,
"learning_rate": 0.0007475247524752475,
"loss": 1.0769,
"step": 4950
},
{
"epoch": 4.207522540299323,
"eval_accuracy": 0.7703979549467554,
"eval_loss": 0.9842762351036072,
"eval_runtime": 1282.9458,
"eval_samples_per_second": 389.243,
"eval_steps_per_second": 1.352,
"step": 4950
},
{
"epoch": 4.216022585835281,
"grad_norm": 16738.04296875,
"learning_rate": 0.0007465346534653466,
"loss": 1.0763,
"step": 4960
},
{
"epoch": 4.224522631371239,
"grad_norm": 15980.9921875,
"learning_rate": 0.0007455445544554455,
"loss": 1.0779,
"step": 4970
},
{
"epoch": 4.233022676907198,
"grad_norm": 17264.802734375,
"learning_rate": 0.0007445544554455446,
"loss": 1.0752,
"step": 4980
},
{
"epoch": 4.241522722443156,
"grad_norm": 19698.16796875,
"learning_rate": 0.0007435643564356435,
"loss": 1.0726,
"step": 4990
},
{
"epoch": 4.2500227679791145,
"grad_norm": 16336.4052734375,
"learning_rate": 0.0007425742574257426,
"loss": 1.0749,
"step": 5000
},
{
"epoch": 4.258522813515072,
"grad_norm": 17774.734375,
"learning_rate": 0.0007415841584158416,
"loss": 1.073,
"step": 5010
},
{
"epoch": 4.26702285905103,
"grad_norm": 19482.98046875,
"learning_rate": 0.0007405940594059406,
"loss": 1.0754,
"step": 5020
},
{
"epoch": 4.275522904586989,
"grad_norm": 17985.568359375,
"learning_rate": 0.0007396039603960396,
"loss": 1.072,
"step": 5030
},
{
"epoch": 4.284022950122947,
"grad_norm": 16432.27734375,
"learning_rate": 0.0007386138613861386,
"loss": 1.0737,
"step": 5040
},
{
"epoch": 4.284022950122947,
"eval_accuracy": 0.7713623846481911,
"eval_loss": 0.982131838798523,
"eval_runtime": 1279.4531,
"eval_samples_per_second": 390.306,
"eval_steps_per_second": 1.355,
"step": 5040
},
{
"epoch": 4.2925229956589055,
"grad_norm": 18672.43359375,
"learning_rate": 0.0007376237623762376,
"loss": 1.0719,
"step": 5050
},
{
"epoch": 4.301023041194863,
"grad_norm": 15502.23828125,
"learning_rate": 0.0007366336633663367,
"loss": 1.0719,
"step": 5060
},
{
"epoch": 4.309523086730822,
"grad_norm": 17056.31640625,
"learning_rate": 0.0007356435643564356,
"loss": 1.0708,
"step": 5070
},
{
"epoch": 4.31802313226678,
"grad_norm": 17404.642578125,
"learning_rate": 0.0007346534653465347,
"loss": 1.0711,
"step": 5080
},
{
"epoch": 4.326523177802738,
"grad_norm": 16851.5390625,
"learning_rate": 0.0007336633663366336,
"loss": 1.0711,
"step": 5090
},
{
"epoch": 4.3350232233386965,
"grad_norm": 16218.2109375,
"learning_rate": 0.0007326732673267327,
"loss": 1.0705,
"step": 5100
},
{
"epoch": 4.343523268874654,
"grad_norm": 15998.1015625,
"learning_rate": 0.0007316831683168317,
"loss": 1.0706,
"step": 5110
},
{
"epoch": 4.352023314410613,
"grad_norm": 17217.677734375,
"learning_rate": 0.0007306930693069307,
"loss": 1.0688,
"step": 5120
},
{
"epoch": 4.360523359946571,
"grad_norm": 20126.892578125,
"learning_rate": 0.0007297029702970297,
"loss": 1.0687,
"step": 5130
},
{
"epoch": 4.360523359946571,
"eval_accuracy": 0.7720324660360576,
"eval_loss": 0.9775763154029846,
"eval_runtime": 1279.545,
"eval_samples_per_second": 390.278,
"eval_steps_per_second": 1.355,
"step": 5130
},
{
"epoch": 4.36902340548253,
"grad_norm": 18316.3203125,
"learning_rate": 0.0007287128712871287,
"loss": 1.0677,
"step": 5140
},
{
"epoch": 4.3775234510184875,
"grad_norm": 19130.66796875,
"learning_rate": 0.0007277227722772277,
"loss": 1.0673,
"step": 5150
},
{
"epoch": 4.386023496554446,
"grad_norm": 16458.451171875,
"learning_rate": 0.0007267326732673268,
"loss": 1.0662,
"step": 5160
},
{
"epoch": 4.394523542090404,
"grad_norm": 15268.9111328125,
"learning_rate": 0.0007257425742574257,
"loss": 1.0674,
"step": 5170
},
{
"epoch": 4.403023587626362,
"grad_norm": 15841.1025390625,
"learning_rate": 0.0007247524752475248,
"loss": 1.0661,
"step": 5180
},
{
"epoch": 4.411523633162321,
"grad_norm": 15076.5078125,
"learning_rate": 0.0007237623762376237,
"loss": 1.0656,
"step": 5190
},
{
"epoch": 4.4200236786982785,
"grad_norm": 17972.560546875,
"learning_rate": 0.0007227722772277228,
"loss": 1.0653,
"step": 5200
},
{
"epoch": 4.428523724234237,
"grad_norm": 17051.736328125,
"learning_rate": 0.0007217821782178218,
"loss": 1.0642,
"step": 5210
},
{
"epoch": 4.437023769770195,
"grad_norm": 18644.025390625,
"learning_rate": 0.0007207920792079208,
"loss": 1.0635,
"step": 5220
},
{
"epoch": 4.437023769770195,
"eval_accuracy": 0.7729909297402872,
"eval_loss": 0.9726957678794861,
"eval_runtime": 1279.9366,
"eval_samples_per_second": 390.158,
"eval_steps_per_second": 1.355,
"step": 5220
},
{
"epoch": 4.445523815306154,
"grad_norm": 17667.255859375,
"learning_rate": 0.0007198019801980198,
"loss": 1.0627,
"step": 5230
},
{
"epoch": 4.454023860842112,
"grad_norm": 16890.02734375,
"learning_rate": 0.0007188118811881188,
"loss": 1.0621,
"step": 5240
},
{
"epoch": 4.4625239063780695,
"grad_norm": 17795.29296875,
"learning_rate": 0.0007178217821782178,
"loss": 1.0609,
"step": 5250
},
{
"epoch": 4.471023951914028,
"grad_norm": 17139.46875,
"learning_rate": 0.0007168316831683169,
"loss": 1.0626,
"step": 5260
},
{
"epoch": 4.479523997449986,
"grad_norm": 16425.12109375,
"learning_rate": 0.0007158415841584158,
"loss": 1.0611,
"step": 5270
},
{
"epoch": 4.488024042985945,
"grad_norm": 20427.578125,
"learning_rate": 0.0007148514851485149,
"loss": 1.0575,
"step": 5280
},
{
"epoch": 4.496524088521903,
"grad_norm": 21385.232421875,
"learning_rate": 0.0007138613861386138,
"loss": 1.0609,
"step": 5290
},
{
"epoch": 4.505024134057861,
"grad_norm": 16601.791015625,
"learning_rate": 0.0007128712871287129,
"loss": 1.0594,
"step": 5300
},
{
"epoch": 4.513524179593819,
"grad_norm": 17113.41015625,
"learning_rate": 0.0007118811881188119,
"loss": 1.0583,
"step": 5310
},
{
"epoch": 4.513524179593819,
"eval_accuracy": 0.7740536654886605,
"eval_loss": 0.9671040773391724,
"eval_runtime": 1277.3232,
"eval_samples_per_second": 390.957,
"eval_steps_per_second": 1.358,
"step": 5310
},
{
"epoch": 4.522024225129778,
"grad_norm": 18315.3984375,
"learning_rate": 0.0007108910891089109,
"loss": 1.0581,
"step": 5320
},
{
"epoch": 4.530524270665736,
"grad_norm": 17246.2890625,
"learning_rate": 0.0007099009900990099,
"loss": 1.0594,
"step": 5330
},
{
"epoch": 4.539024316201694,
"grad_norm": 19144.822265625,
"learning_rate": 0.0007089108910891088,
"loss": 1.0577,
"step": 5340
},
{
"epoch": 4.547524361737652,
"grad_norm": 17691.115234375,
"learning_rate": 0.0007079207920792079,
"loss": 1.0581,
"step": 5350
},
{
"epoch": 4.55602440727361,
"grad_norm": 17498.75,
"learning_rate": 0.000706930693069307,
"loss": 1.0567,
"step": 5360
},
{
"epoch": 4.564524452809569,
"grad_norm": 19553.53515625,
"learning_rate": 0.0007059405940594059,
"loss": 1.0565,
"step": 5370
},
{
"epoch": 4.573024498345527,
"grad_norm": 17747.8359375,
"learning_rate": 0.000704950495049505,
"loss": 1.0603,
"step": 5380
},
{
"epoch": 4.581524543881486,
"grad_norm": 17742.814453125,
"learning_rate": 0.0007039603960396039,
"loss": 1.0567,
"step": 5390
},
{
"epoch": 4.590024589417443,
"grad_norm": 18410.59765625,
"learning_rate": 0.0007029702970297029,
"loss": 1.0561,
"step": 5400
},
{
"epoch": 4.590024589417443,
"eval_accuracy": 0.7741942335122652,
"eval_loss": 0.968053936958313,
"eval_runtime": 1278.1291,
"eval_samples_per_second": 390.71,
"eval_steps_per_second": 1.357,
"step": 5400
},
{
"epoch": 4.598524634953401,
"grad_norm": 14357.056640625,
"learning_rate": 0.000701980198019802,
"loss": 1.0544,
"step": 5410
},
{
"epoch": 4.60702468048936,
"grad_norm": 18278.533203125,
"learning_rate": 0.0007009900990099009,
"loss": 1.0531,
"step": 5420
},
{
"epoch": 4.615524726025318,
"grad_norm": 18012.599609375,
"learning_rate": 0.0007,
"loss": 1.0567,
"step": 5430
},
{
"epoch": 4.624024771561277,
"grad_norm": 18881.583984375,
"learning_rate": 0.0006990099009900989,
"loss": 1.0548,
"step": 5440
},
{
"epoch": 4.632524817097234,
"grad_norm": 19899.90234375,
"learning_rate": 0.000698019801980198,
"loss": 1.0541,
"step": 5450
},
{
"epoch": 4.641024862633193,
"grad_norm": 15934.2138671875,
"learning_rate": 0.000697029702970297,
"loss": 1.0536,
"step": 5460
},
{
"epoch": 4.649524908169151,
"grad_norm": 15930.8544921875,
"learning_rate": 0.000696039603960396,
"loss": 1.0529,
"step": 5470
},
{
"epoch": 4.658024953705109,
"grad_norm": 18346.009765625,
"learning_rate": 0.000695049504950495,
"loss": 1.0507,
"step": 5480
},
{
"epoch": 4.666524999241068,
"grad_norm": 17350.013671875,
"learning_rate": 0.000694059405940594,
"loss": 1.0528,
"step": 5490
},
{
"epoch": 4.666524999241068,
"eval_accuracy": 0.7752448834285515,
"eval_loss": 0.9624823927879333,
"eval_runtime": 1279.2263,
"eval_samples_per_second": 390.375,
"eval_steps_per_second": 1.356,
"step": 5490
},
{
"epoch": 4.675025044777025,
"grad_norm": 17104.44921875,
"learning_rate": 0.000693069306930693,
"loss": 1.0502,
"step": 5500
},
{
"epoch": 4.683525090312984,
"grad_norm": 16279.3994140625,
"learning_rate": 0.0006920792079207921,
"loss": 1.0525,
"step": 5510
},
{
"epoch": 4.692025135848942,
"grad_norm": 17741.423828125,
"learning_rate": 0.000691089108910891,
"loss": 1.0495,
"step": 5520
},
{
"epoch": 4.700525181384901,
"grad_norm": 18454.111328125,
"learning_rate": 0.0006900990099009901,
"loss": 1.0501,
"step": 5530
},
{
"epoch": 4.709025226920859,
"grad_norm": 20272.701171875,
"learning_rate": 0.000689108910891089,
"loss": 1.0518,
"step": 5540
},
{
"epoch": 4.717525272456816,
"grad_norm": 17187.880859375,
"learning_rate": 0.0006881188118811881,
"loss": 1.0498,
"step": 5550
},
{
"epoch": 4.726025317992775,
"grad_norm": 16848.552734375,
"learning_rate": 0.0006871287128712872,
"loss": 1.0475,
"step": 5560
},
{
"epoch": 4.734525363528733,
"grad_norm": 18908.748046875,
"learning_rate": 0.0006861386138613862,
"loss": 1.0477,
"step": 5570
},
{
"epoch": 4.743025409064692,
"grad_norm": 17731.822265625,
"learning_rate": 0.0006851485148514852,
"loss": 1.0496,
"step": 5580
},
{
"epoch": 4.743025409064692,
"eval_accuracy": 0.7761857833832465,
"eval_loss": 0.9594100117683411,
"eval_runtime": 1281.0251,
"eval_samples_per_second": 389.827,
"eval_steps_per_second": 1.354,
"step": 5580
},
{
"epoch": 4.75152545460065,
"grad_norm": 16594.486328125,
"learning_rate": 0.0006841584158415842,
"loss": 1.0479,
"step": 5590
},
{
"epoch": 4.760025500136608,
"grad_norm": 17834.193359375,
"learning_rate": 0.0006831683168316832,
"loss": 1.0465,
"step": 5600
},
{
"epoch": 4.768525545672566,
"grad_norm": 16431.404296875,
"learning_rate": 0.0006821782178217823,
"loss": 1.047,
"step": 5610
},
{
"epoch": 4.777025591208524,
"grad_norm": 18339.9921875,
"learning_rate": 0.0006811881188118812,
"loss": 1.0456,
"step": 5620
},
{
"epoch": 4.785525636744483,
"grad_norm": 15955.2890625,
"learning_rate": 0.0006801980198019803,
"loss": 1.0436,
"step": 5630
},
{
"epoch": 4.794025682280441,
"grad_norm": 18920.619140625,
"learning_rate": 0.0006792079207920792,
"loss": 1.0458,
"step": 5640
},
{
"epoch": 4.802525727816399,
"grad_norm": 16657.296875,
"learning_rate": 0.0006782178217821783,
"loss": 1.0452,
"step": 5650
},
{
"epoch": 4.811025773352357,
"grad_norm": 19700.88671875,
"learning_rate": 0.0006772277227722773,
"loss": 1.0473,
"step": 5660
},
{
"epoch": 4.819525818888316,
"grad_norm": 19387.7578125,
"learning_rate": 0.0006762376237623763,
"loss": 1.0458,
"step": 5670
},
{
"epoch": 4.819525818888316,
"eval_accuracy": 0.7771768462948865,
"eval_loss": 0.9521399736404419,
"eval_runtime": 1279.986,
"eval_samples_per_second": 390.143,
"eval_steps_per_second": 1.355,
"step": 5670
},
{
"epoch": 4.828025864424274,
"grad_norm": 18805.466796875,
"learning_rate": 0.0006752475247524753,
"loss": 1.0425,
"step": 5680
},
{
"epoch": 4.836525909960232,
"grad_norm": 17308.564453125,
"learning_rate": 0.0006742574257425743,
"loss": 1.043,
"step": 5690
},
{
"epoch": 4.84502595549619,
"grad_norm": 19598.796875,
"learning_rate": 0.0006732673267326733,
"loss": 1.0425,
"step": 5700
},
{
"epoch": 4.853526001032148,
"grad_norm": 17387.70703125,
"learning_rate": 0.0006722772277227724,
"loss": 1.0435,
"step": 5710
},
{
"epoch": 4.862026046568107,
"grad_norm": 18542.671875,
"learning_rate": 0.0006712871287128713,
"loss": 1.0429,
"step": 5720
},
{
"epoch": 4.870526092104065,
"grad_norm": 15447.505859375,
"learning_rate": 0.0006702970297029704,
"loss": 1.0428,
"step": 5730
},
{
"epoch": 4.8790261376400235,
"grad_norm": 19042.740234375,
"learning_rate": 0.0006693069306930693,
"loss": 1.0428,
"step": 5740
},
{
"epoch": 4.887526183175981,
"grad_norm": 19183.861328125,
"learning_rate": 0.0006683168316831684,
"loss": 1.042,
"step": 5750
},
{
"epoch": 4.896026228711939,
"grad_norm": 16694.583984375,
"learning_rate": 0.0006673267326732674,
"loss": 1.0407,
"step": 5760
},
{
"epoch": 4.896026228711939,
"eval_accuracy": 0.7776950393076092,
"eval_loss": 0.950414776802063,
"eval_runtime": 1280.3997,
"eval_samples_per_second": 390.017,
"eval_steps_per_second": 1.354,
"step": 5760
},
{
"epoch": 4.904526274247898,
"grad_norm": 16852.962890625,
"learning_rate": 0.0006663366336633664,
"loss": 1.0404,
"step": 5770
},
{
"epoch": 4.913026319783856,
"grad_norm": 20231.42578125,
"learning_rate": 0.0006653465346534654,
"loss": 1.0399,
"step": 5780
},
{
"epoch": 4.9215263653198145,
"grad_norm": 15834.216796875,
"learning_rate": 0.0006643564356435644,
"loss": 1.042,
"step": 5790
},
{
"epoch": 4.930026410855772,
"grad_norm": 15273.1630859375,
"learning_rate": 0.0006633663366336634,
"loss": 1.0397,
"step": 5800
},
{
"epoch": 4.938526456391731,
"grad_norm": 20834.515625,
"learning_rate": 0.0006623762376237625,
"loss": 1.0386,
"step": 5810
},
{
"epoch": 4.947026501927689,
"grad_norm": 21407.017578125,
"learning_rate": 0.0006613861386138614,
"loss": 1.0389,
"step": 5820
},
{
"epoch": 4.955526547463647,
"grad_norm": 19523.822265625,
"learning_rate": 0.0006603960396039605,
"loss": 1.038,
"step": 5830
},
{
"epoch": 4.9640265929996055,
"grad_norm": 18045.078125,
"learning_rate": 0.0006594059405940594,
"loss": 1.0359,
"step": 5840
},
{
"epoch": 4.972526638535563,
"grad_norm": 16661.830078125,
"learning_rate": 0.0006584158415841585,
"loss": 1.037,
"step": 5850
},
{
"epoch": 4.972526638535563,
"eval_accuracy": 0.7784280857816438,
"eval_loss": 0.9486609101295471,
"eval_runtime": 1279.799,
"eval_samples_per_second": 390.2,
"eval_steps_per_second": 1.355,
"step": 5850
},
{
"epoch": 4.981026684071522,
"grad_norm": 19275.583984375,
"learning_rate": 0.0006574257425742575,
"loss": 1.0376,
"step": 5860
},
{
"epoch": 4.98952672960748,
"grad_norm": 20361.98828125,
"learning_rate": 0.0006564356435643565,
"loss": 1.0386,
"step": 5870
},
{
"epoch": 4.998026775143439,
"grad_norm": 18276.357421875,
"learning_rate": 0.0006554455445544555,
"loss": 1.0364,
"step": 5880
},
{
"epoch": 5.0065268206793965,
"grad_norm": 18657.6484375,
"learning_rate": 0.0006544554455445545,
"loss": 1.0364,
"step": 5890
},
{
"epoch": 5.015026866215354,
"grad_norm": 17588.00390625,
"learning_rate": 0.0006534653465346535,
"loss": 1.0346,
"step": 5900
},
{
"epoch": 5.023526911751313,
"grad_norm": 20156.283203125,
"learning_rate": 0.0006524752475247526,
"loss": 1.0341,
"step": 5910
},
{
"epoch": 5.032026957287271,
"grad_norm": 18811.03125,
"learning_rate": 0.0006514851485148515,
"loss": 1.034,
"step": 5920
},
{
"epoch": 5.04052700282323,
"grad_norm": 18864.615234375,
"learning_rate": 0.0006504950495049506,
"loss": 1.0356,
"step": 5930
},
{
"epoch": 5.0490270483591875,
"grad_norm": 18199.67578125,
"learning_rate": 0.0006495049504950495,
"loss": 1.0339,
"step": 5940
},
{
"epoch": 5.0490270483591875,
"eval_accuracy": 0.7792297436274894,
"eval_loss": 0.9439004063606262,
"eval_runtime": 1279.6853,
"eval_samples_per_second": 390.235,
"eval_steps_per_second": 1.355,
"step": 5940
},
{
"epoch": 5.057527093895146,
"grad_norm": 19274.73828125,
"learning_rate": 0.0006485148514851485,
"loss": 1.0334,
"step": 5950
},
{
"epoch": 5.066027139431104,
"grad_norm": 20778.62109375,
"learning_rate": 0.0006475247524752476,
"loss": 1.0337,
"step": 5960
},
{
"epoch": 5.074527184967062,
"grad_norm": 19509.306640625,
"learning_rate": 0.0006465346534653465,
"loss": 1.0355,
"step": 5970
},
{
"epoch": 5.083027230503021,
"grad_norm": 18669.24609375,
"learning_rate": 0.0006455445544554456,
"loss": 1.0329,
"step": 5980
},
{
"epoch": 5.0915272760389785,
"grad_norm": 15668.9755859375,
"learning_rate": 0.0006445544554455445,
"loss": 1.0294,
"step": 5990
},
{
"epoch": 5.100027321574937,
"grad_norm": 18858.783203125,
"learning_rate": 0.0006435643564356436,
"loss": 1.0324,
"step": 6000
},
{
"epoch": 5.108527367110895,
"grad_norm": 16453.763671875,
"learning_rate": 0.0006425742574257426,
"loss": 1.033,
"step": 6010
},
{
"epoch": 5.117027412646854,
"grad_norm": 19821.6875,
"learning_rate": 0.0006415841584158416,
"loss": 1.0324,
"step": 6020
},
{
"epoch": 5.125527458182812,
"grad_norm": 18548.5078125,
"learning_rate": 0.0006405940594059406,
"loss": 1.0284,
"step": 6030
},
{
"epoch": 5.125527458182812,
"eval_accuracy": 0.7796817210719912,
"eval_loss": 0.9404940605163574,
"eval_runtime": 1281.8344,
"eval_samples_per_second": 389.581,
"eval_steps_per_second": 1.353,
"step": 6030
},
{
"epoch": 5.1340275037187695,
"grad_norm": 18221.74609375,
"learning_rate": 0.0006396039603960396,
"loss": 1.0309,
"step": 6040
},
{
"epoch": 5.142527549254728,
"grad_norm": 19257.5078125,
"learning_rate": 0.0006386138613861386,
"loss": 1.0316,
"step": 6050
},
{
"epoch": 5.151027594790686,
"grad_norm": 20660.0859375,
"learning_rate": 0.0006376237623762377,
"loss": 1.0281,
"step": 6060
},
{
"epoch": 5.159527640326645,
"grad_norm": 17646.05859375,
"learning_rate": 0.0006366336633663366,
"loss": 1.0298,
"step": 6070
},
{
"epoch": 5.168027685862603,
"grad_norm": 15755.4951171875,
"learning_rate": 0.0006356435643564357,
"loss": 1.0296,
"step": 6080
},
{
"epoch": 5.176527731398561,
"grad_norm": 19384.6484375,
"learning_rate": 0.0006346534653465346,
"loss": 1.0277,
"step": 6090
},
{
"epoch": 5.185027776934519,
"grad_norm": 16928.974609375,
"learning_rate": 0.0006336633663366337,
"loss": 1.0295,
"step": 6100
},
{
"epoch": 5.193527822470478,
"grad_norm": 15206.5625,
"learning_rate": 0.0006326732673267327,
"loss": 1.0288,
"step": 6110
},
{
"epoch": 5.202027868006436,
"grad_norm": 18075.7421875,
"learning_rate": 0.0006316831683168317,
"loss": 1.0288,
"step": 6120
},
{
"epoch": 5.202027868006436,
"eval_accuracy": 0.7805738833087461,
"eval_loss": 0.9371596574783325,
"eval_runtime": 1279.2071,
"eval_samples_per_second": 390.381,
"eval_steps_per_second": 1.356,
"step": 6120
},
{
"epoch": 5.210527913542394,
"grad_norm": 17331.333984375,
"learning_rate": 0.0006306930693069307,
"loss": 1.0245,
"step": 6130
},
{
"epoch": 5.219027959078352,
"grad_norm": 19757.05078125,
"learning_rate": 0.0006297029702970297,
"loss": 1.0266,
"step": 6140
},
{
"epoch": 5.22752800461431,
"grad_norm": 16021.1396484375,
"learning_rate": 0.0006287128712871287,
"loss": 1.0256,
"step": 6150
},
{
"epoch": 5.236028050150269,
"grad_norm": 16731.166015625,
"learning_rate": 0.0006277227722772278,
"loss": 1.0257,
"step": 6160
},
{
"epoch": 5.244528095686227,
"grad_norm": 17451.5390625,
"learning_rate": 0.0006267326732673267,
"loss": 1.0271,
"step": 6170
},
{
"epoch": 5.2530281412221855,
"grad_norm": 19212.619140625,
"learning_rate": 0.0006257425742574258,
"loss": 1.0255,
"step": 6180
},
{
"epoch": 5.261528186758143,
"grad_norm": 17397.1015625,
"learning_rate": 0.0006247524752475247,
"loss": 1.026,
"step": 6190
},
{
"epoch": 5.270028232294101,
"grad_norm": 18388.25,
"learning_rate": 0.0006237623762376238,
"loss": 1.0253,
"step": 6200
},
{
"epoch": 5.27852827783006,
"grad_norm": 18145.59765625,
"learning_rate": 0.0006227722772277228,
"loss": 1.0251,
"step": 6210
},
{
"epoch": 5.27852827783006,
"eval_accuracy": 0.7809789625861284,
"eval_loss": 0.9343125820159912,
"eval_runtime": 1280.9852,
"eval_samples_per_second": 389.839,
"eval_steps_per_second": 1.354,
"step": 6210
},
{
"epoch": 5.287028323366018,
"grad_norm": 18143.001953125,
"learning_rate": 0.0006217821782178218,
"loss": 1.0251,
"step": 6220
},
{
"epoch": 5.2955283689019765,
"grad_norm": 23030.259765625,
"learning_rate": 0.0006207920792079208,
"loss": 1.0248,
"step": 6230
},
{
"epoch": 5.304028414437934,
"grad_norm": 20499.400390625,
"learning_rate": 0.0006198019801980198,
"loss": 1.0239,
"step": 6240
},
{
"epoch": 5.312528459973893,
"grad_norm": 16773.736328125,
"learning_rate": 0.0006188118811881188,
"loss": 1.0251,
"step": 6250
},
{
"epoch": 5.321028505509851,
"grad_norm": 15036.5712890625,
"learning_rate": 0.0006178217821782179,
"loss": 1.0248,
"step": 6260
},
{
"epoch": 5.329528551045809,
"grad_norm": 18476.0703125,
"learning_rate": 0.0006168316831683168,
"loss": 1.0241,
"step": 6270
},
{
"epoch": 5.3380285965817675,
"grad_norm": 18741.55859375,
"learning_rate": 0.0006158415841584159,
"loss": 1.0238,
"step": 6280
},
{
"epoch": 5.346528642117725,
"grad_norm": 17602.275390625,
"learning_rate": 0.0006148514851485148,
"loss": 1.0222,
"step": 6290
},
{
"epoch": 5.355028687653684,
"grad_norm": 15750.1806640625,
"learning_rate": 0.0006138613861386139,
"loss": 1.0218,
"step": 6300
},
{
"epoch": 5.355028687653684,
"eval_accuracy": 0.7811827125583685,
"eval_loss": 0.9342101216316223,
"eval_runtime": 1278.6765,
"eval_samples_per_second": 390.543,
"eval_steps_per_second": 1.356,
"step": 6300
},
{
"epoch": 5.363528733189642,
"grad_norm": 15361.734375,
"learning_rate": 0.0006128712871287129,
"loss": 1.0212,
"step": 6310
},
{
"epoch": 5.372028778725601,
"grad_norm": 17438.470703125,
"learning_rate": 0.0006118811881188119,
"loss": 1.0217,
"step": 6320
},
{
"epoch": 5.3805288242615585,
"grad_norm": 16182.955078125,
"learning_rate": 0.0006108910891089109,
"loss": 1.0207,
"step": 6330
},
{
"epoch": 5.389028869797516,
"grad_norm": 15098.9970703125,
"learning_rate": 0.0006099009900990099,
"loss": 1.021,
"step": 6340
},
{
"epoch": 5.397528915333475,
"grad_norm": 16916.810546875,
"learning_rate": 0.0006089108910891089,
"loss": 1.0187,
"step": 6350
},
{
"epoch": 5.406028960869433,
"grad_norm": 15510.921875,
"learning_rate": 0.000607920792079208,
"loss": 1.0203,
"step": 6360
},
{
"epoch": 5.414529006405392,
"grad_norm": 18232.8671875,
"learning_rate": 0.0006069306930693069,
"loss": 1.0209,
"step": 6370
},
{
"epoch": 5.4230290519413495,
"grad_norm": 15943.5810546875,
"learning_rate": 0.000605940594059406,
"loss": 1.0189,
"step": 6380
},
{
"epoch": 5.431529097477308,
"grad_norm": 17515.39453125,
"learning_rate": 0.0006049504950495049,
"loss": 1.0185,
"step": 6390
},
{
"epoch": 5.431529097477308,
"eval_accuracy": 0.7821448982293225,
"eval_loss": 0.9290784597396851,
"eval_runtime": 1274.8156,
"eval_samples_per_second": 391.726,
"eval_steps_per_second": 1.36,
"step": 6390
},
{
"epoch": 5.4420023678698275,
"grad_norm": 17196.8828125,
"learning_rate": 0.000603960396039604,
"loss": 1.0209,
"step": 6400
},
{
"epoch": 5.450502413405786,
"grad_norm": 19976.57421875,
"learning_rate": 0.000602970297029703,
"loss": 1.0202,
"step": 6410
},
{
"epoch": 5.459002458941744,
"grad_norm": 16986.595703125,
"learning_rate": 0.000601980198019802,
"loss": 1.0194,
"step": 6420
},
{
"epoch": 5.467502504477703,
"grad_norm": 16322.2490234375,
"learning_rate": 0.000600990099009901,
"loss": 1.018,
"step": 6430
},
{
"epoch": 5.476002550013661,
"grad_norm": 16943.923828125,
"learning_rate": 0.0006,
"loss": 1.0173,
"step": 6440
},
{
"epoch": 5.484502595549619,
"grad_norm": 17515.416015625,
"learning_rate": 0.000599009900990099,
"loss": 1.0171,
"step": 6450
},
{
"epoch": 5.493002641085577,
"grad_norm": 20824.935546875,
"learning_rate": 0.000598019801980198,
"loss": 1.0167,
"step": 6460
},
{
"epoch": 5.501502686621535,
"grad_norm": 19812.29296875,
"learning_rate": 0.000597029702970297,
"loss": 1.017,
"step": 6470
},
{
"epoch": 5.510002732157494,
"grad_norm": 20733.58984375,
"learning_rate": 0.000596039603960396,
"loss": 1.0143,
"step": 6480
},
{
"epoch": 5.510002732157494,
"eval_accuracy": 0.7827784542349862,
"eval_loss": 0.926770031452179,
"eval_runtime": 1034.8114,
"eval_samples_per_second": 482.579,
"eval_steps_per_second": 0.821,
"step": 6480
},
{
"epoch": 5.518502777693452,
"grad_norm": 19389.359375,
"learning_rate": 0.000595049504950495,
"loss": 1.0156,
"step": 6490
},
{
"epoch": 5.52700282322941,
"grad_norm": 18111.1796875,
"learning_rate": 0.000594059405940594,
"loss": 1.015,
"step": 6500
},
{
"epoch": 5.535502868765368,
"grad_norm": 17627.548828125,
"learning_rate": 0.0005930693069306931,
"loss": 1.0149,
"step": 6510
},
{
"epoch": 5.544002914301327,
"grad_norm": 19444.2421875,
"learning_rate": 0.000592079207920792,
"loss": 1.0161,
"step": 6520
},
{
"epoch": 5.552502959837285,
"grad_norm": 20379.31640625,
"learning_rate": 0.0005910891089108911,
"loss": 1.0151,
"step": 6530
},
{
"epoch": 5.561003005373243,
"grad_norm": 17187.162109375,
"learning_rate": 0.00059009900990099,
"loss": 1.0164,
"step": 6540
},
{
"epoch": 5.569503050909201,
"grad_norm": 20134.74609375,
"learning_rate": 0.0005891089108910891,
"loss": 1.0125,
"step": 6550
},
{
"epoch": 5.578003096445159,
"grad_norm": 16345.7314453125,
"learning_rate": 0.0005881188118811881,
"loss": 1.0143,
"step": 6560
},
{
"epoch": 5.586503141981118,
"grad_norm": 16676.40625,
"learning_rate": 0.0005871287128712871,
"loss": 1.0146,
"step": 6570
},
{
"epoch": 5.586503141981118,
"eval_accuracy": 0.7830355090944983,
"eval_loss": 0.925011157989502,
"eval_runtime": 1032.366,
"eval_samples_per_second": 483.722,
"eval_steps_per_second": 0.823,
"step": 6570
},
{
"epoch": 5.595003187517076,
"grad_norm": 18004.888671875,
"learning_rate": 0.0005861386138613861,
"loss": 1.0134,
"step": 6580
},
{
"epoch": 5.6035032330530345,
"grad_norm": 17843.01171875,
"learning_rate": 0.0005851485148514851,
"loss": 1.0121,
"step": 6590
},
{
"epoch": 5.612003278588992,
"grad_norm": 19261.90625,
"learning_rate": 0.0005841584158415841,
"loss": 1.014,
"step": 6600
},
{
"epoch": 5.62050332412495,
"grad_norm": 19034.001953125,
"learning_rate": 0.0005831683168316832,
"loss": 1.014,
"step": 6610
},
{
"epoch": 5.629003369660909,
"grad_norm": 15280.501953125,
"learning_rate": 0.0005821782178217821,
"loss": 1.0127,
"step": 6620
},
{
"epoch": 5.637503415196867,
"grad_norm": 16361.3017578125,
"learning_rate": 0.0005811881188118812,
"loss": 1.0135,
"step": 6630
},
{
"epoch": 5.6460034607328256,
"grad_norm": 18691.2578125,
"learning_rate": 0.0005801980198019801,
"loss": 1.0116,
"step": 6640
},
{
"epoch": 5.654503506268783,
"grad_norm": 17126.857421875,
"learning_rate": 0.0005792079207920792,
"loss": 1.0123,
"step": 6650
},
{
"epoch": 5.663003551804742,
"grad_norm": 18155.701171875,
"learning_rate": 0.0005782178217821782,
"loss": 1.01,
"step": 6660
},
{
"epoch": 5.663003551804742,
"eval_accuracy": 0.7837371016336877,
"eval_loss": 0.9207693338394165,
"eval_runtime": 1032.8984,
"eval_samples_per_second": 483.473,
"eval_steps_per_second": 0.823,
"step": 6660
},
{
"epoch": 5.6715035973407,
"grad_norm": 19733.861328125,
"learning_rate": 0.0005772277227722772,
"loss": 1.0122,
"step": 6670
},
{
"epoch": 5.680003642876658,
"grad_norm": 16721.662109375,
"learning_rate": 0.0005762376237623762,
"loss": 1.013,
"step": 6680
},
{
"epoch": 5.6885036884126166,
"grad_norm": 19389.369140625,
"learning_rate": 0.0005752475247524752,
"loss": 1.012,
"step": 6690
},
{
"epoch": 5.697003733948574,
"grad_norm": 20522.015625,
"learning_rate": 0.0005742574257425742,
"loss": 1.0087,
"step": 6700
},
{
"epoch": 5.705503779484533,
"grad_norm": 19054.39453125,
"learning_rate": 0.0005732673267326733,
"loss": 1.0092,
"step": 6710
},
{
"epoch": 5.714003825020491,
"grad_norm": 16219.0595703125,
"learning_rate": 0.0005722772277227722,
"loss": 1.0105,
"step": 6720
},
{
"epoch": 5.72250387055645,
"grad_norm": 19937.716796875,
"learning_rate": 0.0005712871287128713,
"loss": 1.01,
"step": 6730
},
{
"epoch": 5.731003916092408,
"grad_norm": 18509.220703125,
"learning_rate": 0.0005702970297029702,
"loss": 1.0084,
"step": 6740
},
{
"epoch": 5.739503961628365,
"grad_norm": 18168.08203125,
"learning_rate": 0.0005693069306930693,
"loss": 1.0081,
"step": 6750
},
{
"epoch": 5.739503961628365,
"eval_accuracy": 0.7840970171619854,
"eval_loss": 0.9189032316207886,
"eval_runtime": 1032.6648,
"eval_samples_per_second": 483.582,
"eval_steps_per_second": 0.823,
"step": 6750
},
{
"epoch": 5.748004007164324,
"grad_norm": 17099.634765625,
"learning_rate": 0.0005683168316831683,
"loss": 1.0085,
"step": 6760
},
{
"epoch": 5.756504052700282,
"grad_norm": 19988.625,
"learning_rate": 0.0005673267326732673,
"loss": 1.0079,
"step": 6770
},
{
"epoch": 5.765004098236241,
"grad_norm": 17239.306640625,
"learning_rate": 0.0005663366336633663,
"loss": 1.0075,
"step": 6780
},
{
"epoch": 5.773504143772199,
"grad_norm": 16728.1171875,
"learning_rate": 0.0005653465346534653,
"loss": 1.0062,
"step": 6790
},
{
"epoch": 5.782004189308157,
"grad_norm": 15729.82421875,
"learning_rate": 0.0005643564356435643,
"loss": 1.0074,
"step": 6800
},
{
"epoch": 5.790504234844115,
"grad_norm": 19562.201171875,
"learning_rate": 0.0005633663366336634,
"loss": 1.0075,
"step": 6810
},
{
"epoch": 5.799004280380073,
"grad_norm": 20575.86328125,
"learning_rate": 0.0005623762376237624,
"loss": 1.0056,
"step": 6820
},
{
"epoch": 5.807504325916032,
"grad_norm": 19013.0625,
"learning_rate": 0.0005613861386138615,
"loss": 1.0058,
"step": 6830
},
{
"epoch": 5.81600437145199,
"grad_norm": 18319.556640625,
"learning_rate": 0.0005603960396039604,
"loss": 1.0055,
"step": 6840
},
{
"epoch": 5.81600437145199,
"eval_accuracy": 0.7846662939402871,
"eval_loss": 0.9171181321144104,
"eval_runtime": 1033.9438,
"eval_samples_per_second": 482.984,
"eval_steps_per_second": 0.822,
"step": 6840
},
{
"epoch": 5.824504416987948,
"grad_norm": 19040.82421875,
"learning_rate": 0.0005594059405940595,
"loss": 1.0077,
"step": 6850
},
{
"epoch": 5.833004462523906,
"grad_norm": 19041.265625,
"learning_rate": 0.0005584158415841585,
"loss": 1.0064,
"step": 6860
},
{
"epoch": 5.841504508059865,
"grad_norm": 15583.31640625,
"learning_rate": 0.0005574257425742575,
"loss": 1.0052,
"step": 6870
},
{
"epoch": 5.850004553595823,
"grad_norm": 15519.5576171875,
"learning_rate": 0.0005564356435643565,
"loss": 1.0066,
"step": 6880
},
{
"epoch": 5.858504599131781,
"grad_norm": 19509.908203125,
"learning_rate": 0.0005554455445544555,
"loss": 1.005,
"step": 6890
},
{
"epoch": 5.867004644667739,
"grad_norm": 16785.40625,
"learning_rate": 0.0005544554455445545,
"loss": 1.0045,
"step": 6900
},
{
"epoch": 5.875504690203697,
"grad_norm": 20995.373046875,
"learning_rate": 0.0005534653465346536,
"loss": 1.0028,
"step": 6910
},
{
"epoch": 5.884004735739656,
"grad_norm": 16980.791015625,
"learning_rate": 0.0005524752475247525,
"loss": 1.0058,
"step": 6920
},
{
"epoch": 5.892504781275614,
"grad_norm": 17058.552734375,
"learning_rate": 0.0005514851485148516,
"loss": 1.0032,
"step": 6930
},
{
"epoch": 5.892504781275614,
"eval_accuracy": 0.7852809138194954,
"eval_loss": 0.9138230085372925,
"eval_runtime": 1038.0576,
"eval_samples_per_second": 481.07,
"eval_steps_per_second": 0.819,
"step": 6930
},
{
"epoch": 5.9010048268115725,
"grad_norm": 15984.5615234375,
"learning_rate": 0.0005504950495049505,
"loss": 1.002,
"step": 6940
},
{
"epoch": 5.90950487234753,
"grad_norm": 18076.875,
"learning_rate": 0.0005495049504950496,
"loss": 1.0029,
"step": 6950
},
{
"epoch": 5.918004917883489,
"grad_norm": 17178.27734375,
"learning_rate": 0.0005485148514851486,
"loss": 1.0024,
"step": 6960
},
{
"epoch": 5.926504963419447,
"grad_norm": 17936.349609375,
"learning_rate": 0.0005475247524752476,
"loss": 0.9992,
"step": 6970
},
{
"epoch": 5.935005008955406,
"grad_norm": 16773.853515625,
"learning_rate": 0.0005465346534653466,
"loss": 1.002,
"step": 6980
},
{
"epoch": 5.9435050544913635,
"grad_norm": 18567.22265625,
"learning_rate": 0.0005455445544554456,
"loss": 1.0019,
"step": 6990
},
{
"epoch": 5.952005100027321,
"grad_norm": 20730.1484375,
"learning_rate": 0.0005445544554455446,
"loss": 1.0032,
"step": 7000
},
{
"epoch": 5.96050514556328,
"grad_norm": 18468.900390625,
"learning_rate": 0.0005435643564356437,
"loss": 1.0029,
"step": 7010
},
{
"epoch": 5.969005191099238,
"grad_norm": 21202.24609375,
"learning_rate": 0.0005425742574257426,
"loss": 1.0019,
"step": 7020
},
{
"epoch": 5.969005191099238,
"eval_accuracy": 0.7857404893569327,
"eval_loss": 0.9110856652259827,
"eval_runtime": 1036.1907,
"eval_samples_per_second": 481.936,
"eval_steps_per_second": 0.82,
"step": 7020
},
{
"epoch": 5.977505236635197,
"grad_norm": 17446.140625,
"learning_rate": 0.0005415841584158417,
"loss": 1.0013,
"step": 7030
},
{
"epoch": 5.9860052821711545,
"grad_norm": 20787.7109375,
"learning_rate": 0.0005405940594059406,
"loss": 1.0006,
"step": 7040
},
{
"epoch": 5.994505327707113,
"grad_norm": 15513.68359375,
"learning_rate": 0.0005396039603960396,
"loss": 1.001,
"step": 7050
},
{
"epoch": 6.003005373243071,
"grad_norm": 17996.607421875,
"learning_rate": 0.0005386138613861387,
"loss": 1.0003,
"step": 7060
},
{
"epoch": 6.011505418779029,
"grad_norm": 17095.095703125,
"learning_rate": 0.0005376237623762376,
"loss": 1.0007,
"step": 7070
},
{
"epoch": 6.020005464314988,
"grad_norm": 17025.11328125,
"learning_rate": 0.0005366336633663367,
"loss": 1.0003,
"step": 7080
},
{
"epoch": 6.0285055098509455,
"grad_norm": 15303.3544921875,
"learning_rate": 0.0005356435643564356,
"loss": 1.0,
"step": 7090
},
{
"epoch": 6.037005555386904,
"grad_norm": 18501.576171875,
"learning_rate": 0.0005346534653465347,
"loss": 0.9981,
"step": 7100
},
{
"epoch": 6.045505600922862,
"grad_norm": 19782.1171875,
"learning_rate": 0.0005336633663366337,
"loss": 0.9989,
"step": 7110
},
{
"epoch": 6.045505600922862,
"eval_accuracy": 0.7861505064886588,
"eval_loss": 0.9094851613044739,
"eval_runtime": 1034.5248,
"eval_samples_per_second": 482.712,
"eval_steps_per_second": 0.822,
"step": 7110
},
{
"epoch": 6.05400564645882,
"grad_norm": 19169.7890625,
"learning_rate": 0.0005326732673267327,
"loss": 0.9998,
"step": 7120
},
{
"epoch": 6.062505691994779,
"grad_norm": 19047.958984375,
"learning_rate": 0.0005316831683168317,
"loss": 0.9973,
"step": 7130
},
{
"epoch": 6.0710057375307365,
"grad_norm": 20607.556640625,
"learning_rate": 0.0005306930693069307,
"loss": 0.9962,
"step": 7140
},
{
"epoch": 6.079505783066695,
"grad_norm": 16297.8662109375,
"learning_rate": 0.0005297029702970297,
"loss": 0.9968,
"step": 7150
},
{
"epoch": 6.088005828602653,
"grad_norm": 16640.685546875,
"learning_rate": 0.0005287128712871288,
"loss": 0.9987,
"step": 7160
},
{
"epoch": 6.096505874138612,
"grad_norm": 16723.65625,
"learning_rate": 0.0005277227722772277,
"loss": 0.9981,
"step": 7170
},
{
"epoch": 6.10500591967457,
"grad_norm": 18703.06640625,
"learning_rate": 0.0005267326732673268,
"loss": 0.9959,
"step": 7180
},
{
"epoch": 6.113505965210528,
"grad_norm": 16973.158203125,
"learning_rate": 0.0005257425742574257,
"loss": 0.9988,
"step": 7190
},
{
"epoch": 6.122006010746486,
"grad_norm": 16989.619140625,
"learning_rate": 0.0005247524752475248,
"loss": 0.9955,
"step": 7200
},
{
"epoch": 6.122006010746486,
"eval_accuracy": 0.7866073008018492,
"eval_loss": 0.9084503650665283,
"eval_runtime": 1037.0895,
"eval_samples_per_second": 481.519,
"eval_steps_per_second": 0.82,
"step": 7200
},
{
"epoch": 6.130506056282444,
"grad_norm": 16469.03515625,
"learning_rate": 0.0005237623762376238,
"loss": 0.9962,
"step": 7210
},
{
"epoch": 6.139006101818403,
"grad_norm": 16646.052734375,
"learning_rate": 0.0005227722772277228,
"loss": 0.9978,
"step": 7220
},
{
"epoch": 6.147506147354361,
"grad_norm": 18823.22265625,
"learning_rate": 0.0005217821782178218,
"loss": 0.9967,
"step": 7230
},
{
"epoch": 6.156006192890319,
"grad_norm": 16628.587890625,
"learning_rate": 0.0005207920792079208,
"loss": 0.9968,
"step": 7240
},
{
"epoch": 6.164506238426277,
"grad_norm": 17544.67578125,
"learning_rate": 0.0005198019801980198,
"loss": 0.9958,
"step": 7250
},
{
"epoch": 6.173006283962236,
"grad_norm": 16480.125,
"learning_rate": 0.0005188118811881189,
"loss": 0.997,
"step": 7260
},
{
"epoch": 6.181506329498194,
"grad_norm": 18358.3984375,
"learning_rate": 0.0005178217821782178,
"loss": 0.9956,
"step": 7270
},
{
"epoch": 6.190006375034152,
"grad_norm": 19029.322265625,
"learning_rate": 0.0005168316831683169,
"loss": 0.9977,
"step": 7280
},
{
"epoch": 6.19850642057011,
"grad_norm": 16023.6572265625,
"learning_rate": 0.0005158415841584158,
"loss": 0.9944,
"step": 7290
},
{
"epoch": 6.19850642057011,
"eval_accuracy": 0.7871460102262176,
"eval_loss": 0.9043031930923462,
"eval_runtime": 1032.4584,
"eval_samples_per_second": 483.679,
"eval_steps_per_second": 0.823,
"step": 7290
},
{
"epoch": 6.207006466106068,
"grad_norm": 17711.34765625,
"learning_rate": 0.0005148514851485149,
"loss": 0.9931,
"step": 7300
},
{
"epoch": 6.215506511642027,
"grad_norm": 15834.974609375,
"learning_rate": 0.0005138613861386139,
"loss": 0.9935,
"step": 7310
},
{
"epoch": 6.224006557177985,
"grad_norm": 15997.443359375,
"learning_rate": 0.0005128712871287129,
"loss": 0.9943,
"step": 7320
},
{
"epoch": 6.2325066027139435,
"grad_norm": 19241.9921875,
"learning_rate": 0.0005118811881188119,
"loss": 0.9942,
"step": 7330
},
{
"epoch": 6.241006648249901,
"grad_norm": 14933.837890625,
"learning_rate": 0.0005108910891089109,
"loss": 0.9949,
"step": 7340
},
{
"epoch": 6.249506693785859,
"grad_norm": 16723.673828125,
"learning_rate": 0.0005099009900990099,
"loss": 0.9937,
"step": 7350
},
{
"epoch": 6.258006739321818,
"grad_norm": 19108.876953125,
"learning_rate": 0.000508910891089109,
"loss": 0.9915,
"step": 7360
},
{
"epoch": 6.266506784857776,
"grad_norm": 14807.8232421875,
"learning_rate": 0.0005079207920792079,
"loss": 0.9912,
"step": 7370
},
{
"epoch": 6.2750068303937345,
"grad_norm": 16829.048828125,
"learning_rate": 0.000506930693069307,
"loss": 0.9917,
"step": 7380
},
{
"epoch": 6.2750068303937345,
"eval_accuracy": 0.7875770735366631,
"eval_loss": 0.9019953608512878,
"eval_runtime": 1035.5292,
"eval_samples_per_second": 482.244,
"eval_steps_per_second": 0.821,
"step": 7380
},
{
"epoch": 6.283506875929692,
"grad_norm": 17801.87109375,
"learning_rate": 0.0005059405940594059,
"loss": 0.9928,
"step": 7390
},
{
"epoch": 6.292006921465651,
"grad_norm": 18446.126953125,
"learning_rate": 0.000504950495049505,
"loss": 0.9915,
"step": 7400
},
{
"epoch": 6.300506967001609,
"grad_norm": 16055.3349609375,
"learning_rate": 0.000503960396039604,
"loss": 0.9903,
"step": 7410
},
{
"epoch": 6.309007012537567,
"grad_norm": 16897.765625,
"learning_rate": 0.000502970297029703,
"loss": 0.9908,
"step": 7420
},
{
"epoch": 6.3175070580735255,
"grad_norm": 19176.193359375,
"learning_rate": 0.000501980198019802,
"loss": 0.9898,
"step": 7430
},
{
"epoch": 6.326007103609483,
"grad_norm": 14390.6298828125,
"learning_rate": 0.000500990099009901,
"loss": 0.9904,
"step": 7440
},
{
"epoch": 6.334507149145442,
"grad_norm": 21656.654296875,
"learning_rate": 0.0005,
"loss": 0.9916,
"step": 7450
},
{
"epoch": 6.3430071946814,
"grad_norm": 18097.005859375,
"learning_rate": 0.0004990099009900991,
"loss": 0.9902,
"step": 7460
},
{
"epoch": 6.351507240217359,
"grad_norm": 17372.796875,
"learning_rate": 0.000498019801980198,
"loss": 0.9896,
"step": 7470
},
{
"epoch": 6.351507240217359,
"eval_accuracy": 0.7878735288789341,
"eval_loss": 0.9011977314949036,
"eval_runtime": 1035.7365,
"eval_samples_per_second": 482.148,
"eval_steps_per_second": 0.821,
"step": 7470
},
{
"epoch": 6.3600072857533165,
"grad_norm": 16490.4375,
"learning_rate": 0.0004970297029702971,
"loss": 0.989,
"step": 7480
},
{
"epoch": 6.368507331289274,
"grad_norm": 18962.072265625,
"learning_rate": 0.000496039603960396,
"loss": 0.9887,
"step": 7490
},
{
"epoch": 6.377007376825233,
"grad_norm": 19760.28125,
"learning_rate": 0.0004950495049504951,
"loss": 0.9891,
"step": 7500
},
{
"epoch": 6.385507422361191,
"grad_norm": 16014.2646484375,
"learning_rate": 0.0004940594059405941,
"loss": 0.9921,
"step": 7510
},
{
"epoch": 6.39400746789715,
"grad_norm": 18045.798828125,
"learning_rate": 0.000493069306930693,
"loss": 0.9886,
"step": 7520
},
{
"epoch": 6.4025075134331075,
"grad_norm": 18082.34375,
"learning_rate": 0.0004920792079207921,
"loss": 0.987,
"step": 7530
},
{
"epoch": 6.411007558969066,
"grad_norm": 17991.447265625,
"learning_rate": 0.000491089108910891,
"loss": 0.99,
"step": 7540
},
{
"epoch": 6.419507604505024,
"grad_norm": 15685.39453125,
"learning_rate": 0.0004900990099009901,
"loss": 0.9884,
"step": 7550
},
{
"epoch": 6.428007650040982,
"grad_norm": 16992.146484375,
"learning_rate": 0.0004891089108910892,
"loss": 0.9878,
"step": 7560
},
{
"epoch": 6.428007650040982,
"eval_accuracy": 0.7882446558820121,
"eval_loss": 0.8991919755935669,
"eval_runtime": 1034.5241,
"eval_samples_per_second": 482.713,
"eval_steps_per_second": 0.822,
"step": 7560
},
{
"epoch": 6.436507695576941,
"grad_norm": 18808.25390625,
"learning_rate": 0.0004881188118811881,
"loss": 0.9872,
"step": 7570
},
{
"epoch": 6.4450077411128985,
"grad_norm": 18018.94921875,
"learning_rate": 0.00048712871287128715,
"loss": 0.9882,
"step": 7580
},
{
"epoch": 6.453507786648857,
"grad_norm": 16361.677734375,
"learning_rate": 0.00048613861386138615,
"loss": 0.9867,
"step": 7590
},
{
"epoch": 6.462007832184815,
"grad_norm": 16750.744140625,
"learning_rate": 0.00048514851485148515,
"loss": 0.9878,
"step": 7600
},
{
"epoch": 6.470507877720774,
"grad_norm": 19413.93359375,
"learning_rate": 0.00048415841584158414,
"loss": 0.9868,
"step": 7610
},
{
"epoch": 6.479007923256732,
"grad_norm": 17738.328125,
"learning_rate": 0.00048316831683168314,
"loss": 0.9869,
"step": 7620
},
{
"epoch": 6.4875079687926895,
"grad_norm": 16600.763671875,
"learning_rate": 0.0004821782178217822,
"loss": 0.9868,
"step": 7630
},
{
"epoch": 6.496008014328648,
"grad_norm": 19458.306640625,
"learning_rate": 0.0004811881188118812,
"loss": 0.9842,
"step": 7640
},
{
"epoch": 6.504508059864606,
"grad_norm": 15481.515625,
"learning_rate": 0.0004801980198019802,
"loss": 0.9886,
"step": 7650
},
{
"epoch": 6.504508059864606,
"eval_accuracy": 0.7887328723155168,
"eval_loss": 0.8970336318016052,
"eval_runtime": 1037.2132,
"eval_samples_per_second": 481.461,
"eval_steps_per_second": 0.82,
"step": 7650
},
{
"epoch": 6.513402750371877,
"grad_norm": 20331.080078125,
"learning_rate": 0.0004792079207920792,
"loss": 0.9857,
"step": 7660
},
{
"epoch": 6.5219027959078355,
"grad_norm": 17513.203125,
"learning_rate": 0.0004782178217821782,
"loss": 0.9846,
"step": 7670
},
{
"epoch": 6.530402841443793,
"grad_norm": 18388.54296875,
"learning_rate": 0.00047722772277227724,
"loss": 0.9853,
"step": 7680
},
{
"epoch": 6.538902886979752,
"grad_norm": 18303.26953125,
"learning_rate": 0.00047623762376237624,
"loss": 0.9853,
"step": 7690
},
{
"epoch": 6.54740293251571,
"grad_norm": 16617.505859375,
"learning_rate": 0.00047524752475247524,
"loss": 0.9859,
"step": 7700
},
{
"epoch": 6.555902978051668,
"grad_norm": 17269.935546875,
"learning_rate": 0.00047425742574257423,
"loss": 0.9855,
"step": 7710
},
{
"epoch": 6.5644030235876265,
"grad_norm": 16519.927734375,
"learning_rate": 0.00047326732673267323,
"loss": 0.9843,
"step": 7720
},
{
"epoch": 6.572903069123584,
"grad_norm": 15807.8505859375,
"learning_rate": 0.0004722772277227723,
"loss": 0.983,
"step": 7730
},
{
"epoch": 6.581403114659543,
"grad_norm": 15810.3408203125,
"learning_rate": 0.0004712871287128713,
"loss": 0.9861,
"step": 7740
},
{
"epoch": 6.581403114659543,
"eval_accuracy": 0.7891655969340131,
"eval_loss": 0.8951780200004578,
"eval_runtime": 1035.4986,
"eval_samples_per_second": 482.259,
"eval_steps_per_second": 0.821,
"step": 7740
},
{
"epoch": 6.589903160195501,
"grad_norm": 16766.341796875,
"learning_rate": 0.0004702970297029703,
"loss": 0.9851,
"step": 7750
},
{
"epoch": 6.59840320573146,
"grad_norm": 20359.9140625,
"learning_rate": 0.0004693069306930693,
"loss": 0.9844,
"step": 7760
},
{
"epoch": 6.6069032512674175,
"grad_norm": 16457.33984375,
"learning_rate": 0.00046831683168316833,
"loss": 0.9822,
"step": 7770
},
{
"epoch": 6.615403296803375,
"grad_norm": 18674.6015625,
"learning_rate": 0.0004673267326732674,
"loss": 0.983,
"step": 7780
},
{
"epoch": 6.623903342339334,
"grad_norm": 18476.642578125,
"learning_rate": 0.0004663366336633664,
"loss": 0.9826,
"step": 7790
},
{
"epoch": 6.632403387875292,
"grad_norm": 15693.6640625,
"learning_rate": 0.0004653465346534654,
"loss": 0.9806,
"step": 7800
},
{
"epoch": 6.640903433411251,
"grad_norm": 18678.359375,
"learning_rate": 0.0004643564356435644,
"loss": 0.9826,
"step": 7810
},
{
"epoch": 6.6494034789472085,
"grad_norm": 18860.5390625,
"learning_rate": 0.0004633663366336634,
"loss": 0.9843,
"step": 7820
},
{
"epoch": 6.657903524483167,
"grad_norm": 18948.015625,
"learning_rate": 0.00046237623762376243,
"loss": 0.9814,
"step": 7830
},
{
"epoch": 6.657903524483167,
"eval_accuracy": 0.789608308121163,
"eval_loss": 0.8939031362533569,
"eval_runtime": 1028.1352,
"eval_samples_per_second": 485.712,
"eval_steps_per_second": 0.827,
"step": 7830
},
{
"epoch": 6.666403570019125,
"grad_norm": 17197.728515625,
"learning_rate": 0.00046138613861386143,
"loss": 0.9816,
"step": 7840
},
{
"epoch": 6.674903615555083,
"grad_norm": 21051.08984375,
"learning_rate": 0.0004603960396039604,
"loss": 0.9812,
"step": 7850
},
{
"epoch": 6.683403661091042,
"grad_norm": 19255.662109375,
"learning_rate": 0.0004594059405940594,
"loss": 0.9836,
"step": 7860
},
{
"epoch": 6.6919037066269995,
"grad_norm": 16365.4599609375,
"learning_rate": 0.0004584158415841584,
"loss": 0.9824,
"step": 7870
},
{
"epoch": 6.700403752162958,
"grad_norm": 17249.7109375,
"learning_rate": 0.0004574257425742575,
"loss": 0.9793,
"step": 7880
},
{
"epoch": 6.708903797698916,
"grad_norm": 15945.6630859375,
"learning_rate": 0.00045643564356435647,
"loss": 0.9819,
"step": 7890
},
{
"epoch": 6.717403843234875,
"grad_norm": 16529.767578125,
"learning_rate": 0.00045544554455445547,
"loss": 0.9801,
"step": 7900
},
{
"epoch": 6.725903888770833,
"grad_norm": 16305.484375,
"learning_rate": 0.00045445544554455447,
"loss": 0.9823,
"step": 7910
},
{
"epoch": 6.7344039343067905,
"grad_norm": 17304.328125,
"learning_rate": 0.00045346534653465347,
"loss": 0.9782,
"step": 7920
},
{
"epoch": 6.7344039343067905,
"eval_accuracy": 0.7898104897932849,
"eval_loss": 0.8923720121383667,
"eval_runtime": 1032.3531,
"eval_samples_per_second": 483.728,
"eval_steps_per_second": 0.823,
"step": 7920
},
{
"epoch": 6.742903979842749,
"grad_norm": 16057.078125,
"learning_rate": 0.0004524752475247525,
"loss": 0.9804,
"step": 7930
},
{
"epoch": 6.751404025378707,
"grad_norm": 16539.8046875,
"learning_rate": 0.0004514851485148515,
"loss": 0.9816,
"step": 7940
},
{
"epoch": 6.759904070914666,
"grad_norm": 18756.751953125,
"learning_rate": 0.0004504950495049505,
"loss": 0.9803,
"step": 7950
},
{
"epoch": 6.768404116450624,
"grad_norm": 16140.8369140625,
"learning_rate": 0.0004495049504950495,
"loss": 0.9807,
"step": 7960
},
{
"epoch": 6.776904161986582,
"grad_norm": 16101.984375,
"learning_rate": 0.0004485148514851485,
"loss": 0.9795,
"step": 7970
},
{
"epoch": 6.78540420752254,
"grad_norm": 18084.240234375,
"learning_rate": 0.00044752475247524756,
"loss": 0.9798,
"step": 7980
},
{
"epoch": 6.793904253058498,
"grad_norm": 17187.412109375,
"learning_rate": 0.00044653465346534656,
"loss": 0.9801,
"step": 7990
},
{
"epoch": 6.802404298594457,
"grad_norm": 16965.23046875,
"learning_rate": 0.00044554455445544556,
"loss": 0.9784,
"step": 8000
},
{
"epoch": 6.810904344130415,
"grad_norm": 16516.47265625,
"learning_rate": 0.00044455445544554456,
"loss": 0.9786,
"step": 8010
},
{
"epoch": 6.810904344130415,
"eval_accuracy": 0.7902325926967286,
"eval_loss": 0.8902812004089355,
"eval_runtime": 1032.3456,
"eval_samples_per_second": 483.731,
"eval_steps_per_second": 0.823,
"step": 8010
},
{
"epoch": 6.819404389666373,
"grad_norm": 18656.958984375,
"learning_rate": 0.00044356435643564356,
"loss": 0.9794,
"step": 8020
},
{
"epoch": 6.827904435202331,
"grad_norm": 18381.744140625,
"learning_rate": 0.0004425742574257426,
"loss": 0.978,
"step": 8030
},
{
"epoch": 6.83640448073829,
"grad_norm": 16134.7392578125,
"learning_rate": 0.0004415841584158416,
"loss": 0.9783,
"step": 8040
},
{
"epoch": 6.844904526274248,
"grad_norm": 15535.76953125,
"learning_rate": 0.0004405940594059406,
"loss": 0.9771,
"step": 8050
},
{
"epoch": 6.853404571810206,
"grad_norm": 17276.91015625,
"learning_rate": 0.0004396039603960396,
"loss": 0.9778,
"step": 8060
},
{
"epoch": 6.861904617346164,
"grad_norm": 16477.451171875,
"learning_rate": 0.0004386138613861386,
"loss": 0.9762,
"step": 8070
},
{
"epoch": 6.870404662882122,
"grad_norm": 17235.421875,
"learning_rate": 0.00043762376237623765,
"loss": 0.9758,
"step": 8080
},
{
"epoch": 6.878904708418081,
"grad_norm": 23922.306640625,
"learning_rate": 0.00043663366336633665,
"loss": 0.9771,
"step": 8090
},
{
"epoch": 6.887404753954039,
"grad_norm": 20163.564453125,
"learning_rate": 0.00043564356435643565,
"loss": 0.9786,
"step": 8100
},
{
"epoch": 6.887404753954039,
"eval_accuracy": 0.7905720984224686,
"eval_loss": 0.8882827162742615,
"eval_runtime": 1035.2187,
"eval_samples_per_second": 482.389,
"eval_steps_per_second": 0.821,
"step": 8100
},
{
"epoch": 6.895904799489998,
"grad_norm": 22567.283203125,
"learning_rate": 0.00043465346534653465,
"loss": 0.9767,
"step": 8110
},
{
"epoch": 6.904404845025955,
"grad_norm": 18677.17578125,
"learning_rate": 0.00043366336633663365,
"loss": 0.9768,
"step": 8120
},
{
"epoch": 6.912904890561913,
"grad_norm": 18427.8046875,
"learning_rate": 0.0004326732673267327,
"loss": 0.9765,
"step": 8130
},
{
"epoch": 6.921404936097872,
"grad_norm": 16562.19140625,
"learning_rate": 0.0004316831683168317,
"loss": 0.9778,
"step": 8140
},
{
"epoch": 6.92990498163383,
"grad_norm": 18019.8046875,
"learning_rate": 0.0004306930693069307,
"loss": 0.9765,
"step": 8150
},
{
"epoch": 6.938405027169789,
"grad_norm": 15273.8798828125,
"learning_rate": 0.0004297029702970297,
"loss": 0.9756,
"step": 8160
},
{
"epoch": 6.946905072705746,
"grad_norm": 16736.18359375,
"learning_rate": 0.0004287128712871287,
"loss": 0.9739,
"step": 8170
},
{
"epoch": 6.955405118241705,
"grad_norm": 17612.673828125,
"learning_rate": 0.00042772277227722774,
"loss": 0.9772,
"step": 8180
},
{
"epoch": 6.963905163777663,
"grad_norm": 15665.3857421875,
"learning_rate": 0.00042673267326732674,
"loss": 0.9742,
"step": 8190
},
{
"epoch": 6.963905163777663,
"eval_accuracy": 0.7910029077910385,
"eval_loss": 0.8868052363395691,
"eval_runtime": 1033.6186,
"eval_samples_per_second": 483.136,
"eval_steps_per_second": 0.822,
"step": 8190
},
{
"epoch": 6.972405209313621,
"grad_norm": 16346.849609375,
"learning_rate": 0.00042574257425742574,
"loss": 0.9773,
"step": 8200
},
{
"epoch": 6.98090525484958,
"grad_norm": 15904.78515625,
"learning_rate": 0.00042475247524752474,
"loss": 0.9756,
"step": 8210
},
{
"epoch": 6.989405300385537,
"grad_norm": 16064.7041015625,
"learning_rate": 0.00042376237623762374,
"loss": 0.974,
"step": 8220
},
{
"epoch": 6.997905345921496,
"grad_norm": 17884.13671875,
"learning_rate": 0.0004227722772277228,
"loss": 0.9747,
"step": 8230
},
{
"epoch": 7.006405391457454,
"grad_norm": 14780.166015625,
"learning_rate": 0.0004217821782178218,
"loss": 0.9743,
"step": 8240
},
{
"epoch": 7.014905436993413,
"grad_norm": 17861.595703125,
"learning_rate": 0.0004207920792079208,
"loss": 0.9744,
"step": 8250
},
{
"epoch": 7.023405482529371,
"grad_norm": 16821.708984375,
"learning_rate": 0.0004198019801980198,
"loss": 0.9732,
"step": 8260
},
{
"epoch": 7.031905528065329,
"grad_norm": 18202.1796875,
"learning_rate": 0.0004188118811881188,
"loss": 0.9746,
"step": 8270
},
{
"epoch": 7.040405573601287,
"grad_norm": 15645.1865234375,
"learning_rate": 0.00041782178217821784,
"loss": 0.9735,
"step": 8280
},
{
"epoch": 7.040405573601287,
"eval_accuracy": 0.7912855873371129,
"eval_loss": 0.8865892887115479,
"eval_runtime": 1039.2279,
"eval_samples_per_second": 480.528,
"eval_steps_per_second": 0.818,
"step": 8280
},
{
"epoch": 7.048905619137245,
"grad_norm": 16949.755859375,
"learning_rate": 0.00041683168316831683,
"loss": 0.9736,
"step": 8290
},
{
"epoch": 7.057405664673204,
"grad_norm": 18461.337890625,
"learning_rate": 0.00041584158415841583,
"loss": 0.9728,
"step": 8300
},
{
"epoch": 7.065905710209162,
"grad_norm": 17668.466796875,
"learning_rate": 0.00041485148514851483,
"loss": 0.9721,
"step": 8310
},
{
"epoch": 7.07440575574512,
"grad_norm": 15126.234375,
"learning_rate": 0.00041386138613861383,
"loss": 0.9718,
"step": 8320
},
{
"epoch": 7.082905801281078,
"grad_norm": 16266.1083984375,
"learning_rate": 0.0004128712871287129,
"loss": 0.9733,
"step": 8330
},
{
"epoch": 7.091405846817037,
"grad_norm": 19422.59375,
"learning_rate": 0.0004118811881188119,
"loss": 0.9733,
"step": 8340
},
{
"epoch": 7.099905892352995,
"grad_norm": 17168.087890625,
"learning_rate": 0.0004108910891089109,
"loss": 0.9723,
"step": 8350
},
{
"epoch": 7.108405937888953,
"grad_norm": 16709.5625,
"learning_rate": 0.0004099009900990099,
"loss": 0.9712,
"step": 8360
},
{
"epoch": 7.116905983424911,
"grad_norm": 15946.3232421875,
"learning_rate": 0.0004089108910891089,
"loss": 0.9726,
"step": 8370
},
{
"epoch": 7.116905983424911,
"eval_accuracy": 0.7916597760806168,
"eval_loss": 0.8836008310317993,
"eval_runtime": 1033.0801,
"eval_samples_per_second": 483.388,
"eval_steps_per_second": 0.823,
"step": 8370
},
{
"epoch": 7.125406028960869,
"grad_norm": 13808.0322265625,
"learning_rate": 0.0004079207920792079,
"loss": 0.9693,
"step": 8380
},
{
"epoch": 7.133906074496828,
"grad_norm": 19899.408203125,
"learning_rate": 0.0004069306930693069,
"loss": 0.9695,
"step": 8390
},
{
"epoch": 7.142406120032786,
"grad_norm": 14987.80078125,
"learning_rate": 0.000405940594059406,
"loss": 0.9698,
"step": 8400
},
{
"epoch": 7.1509061655687445,
"grad_norm": 16658.63671875,
"learning_rate": 0.000404950495049505,
"loss": 0.9711,
"step": 8410
},
{
"epoch": 7.159406211104702,
"grad_norm": 16382.1494140625,
"learning_rate": 0.00040396039603960397,
"loss": 0.9717,
"step": 8420
},
{
"epoch": 7.16790625664066,
"grad_norm": 16640.423828125,
"learning_rate": 0.000402970297029703,
"loss": 0.9706,
"step": 8430
},
{
"epoch": 7.176406302176619,
"grad_norm": 17918.15625,
"learning_rate": 0.000401980198019802,
"loss": 0.9723,
"step": 8440
},
{
"epoch": 7.184906347712577,
"grad_norm": 16958.197265625,
"learning_rate": 0.000400990099009901,
"loss": 0.9704,
"step": 8450
},
{
"epoch": 7.1934063932485355,
"grad_norm": 15377.521484375,
"learning_rate": 0.0004,
"loss": 0.97,
"step": 8460
},
{
"epoch": 7.1934063932485355,
"eval_accuracy": 0.7920878335859806,
"eval_loss": 0.8810552954673767,
"eval_runtime": 1033.6339,
"eval_samples_per_second": 483.129,
"eval_steps_per_second": 0.822,
"step": 8460
},
{
"epoch": 7.201906438784493,
"grad_norm": 15135.03515625,
"learning_rate": 0.000399009900990099,
"loss": 0.9684,
"step": 8470
},
{
"epoch": 7.210406484320452,
"grad_norm": 16375.2431640625,
"learning_rate": 0.00039801980198019807,
"loss": 0.9691,
"step": 8480
},
{
"epoch": 7.21890652985641,
"grad_norm": 16684.763671875,
"learning_rate": 0.00039702970297029707,
"loss": 0.9678,
"step": 8490
},
{
"epoch": 7.227406575392368,
"grad_norm": 17551.935546875,
"learning_rate": 0.00039603960396039607,
"loss": 0.9688,
"step": 8500
},
{
"epoch": 7.2359066209283265,
"grad_norm": 17371.619140625,
"learning_rate": 0.00039504950495049506,
"loss": 0.9717,
"step": 8510
},
{
"epoch": 7.244406666464284,
"grad_norm": 20004.443359375,
"learning_rate": 0.00039405940594059406,
"loss": 0.9672,
"step": 8520
},
{
"epoch": 7.252906712000243,
"grad_norm": 18529.32421875,
"learning_rate": 0.0003930693069306931,
"loss": 0.9693,
"step": 8530
},
{
"epoch": 7.261406757536201,
"grad_norm": 16488.56640625,
"learning_rate": 0.0003920792079207921,
"loss": 0.9678,
"step": 8540
},
{
"epoch": 7.26990680307216,
"grad_norm": 22428.466796875,
"learning_rate": 0.0003910891089108911,
"loss": 0.9693,
"step": 8550
},
{
"epoch": 7.26990680307216,
"eval_accuracy": 0.792256339480986,
"eval_loss": 0.8810757398605347,
"eval_runtime": 1033.8252,
"eval_samples_per_second": 483.039,
"eval_steps_per_second": 0.822,
"step": 8550
},
{
"epoch": 7.2784068486081175,
"grad_norm": 17588.4453125,
"learning_rate": 0.0003900990099009901,
"loss": 0.9679,
"step": 8560
},
{
"epoch": 7.286906894144076,
"grad_norm": 18152.78515625,
"learning_rate": 0.0003891089108910891,
"loss": 0.9668,
"step": 8570
},
{
"epoch": 7.295406939680034,
"grad_norm": 14463.046875,
"learning_rate": 0.00038811881188118816,
"loss": 0.9679,
"step": 8580
},
{
"epoch": 7.303906985215992,
"grad_norm": 19075.265625,
"learning_rate": 0.00038712871287128716,
"loss": 0.9681,
"step": 8590
},
{
"epoch": 7.312407030751951,
"grad_norm": 21981.548828125,
"learning_rate": 0.00038613861386138616,
"loss": 0.9682,
"step": 8600
},
{
"epoch": 7.3209070762879085,
"grad_norm": 20439.08984375,
"learning_rate": 0.00038514851485148515,
"loss": 0.968,
"step": 8610
},
{
"epoch": 7.329407121823867,
"grad_norm": 17120.98046875,
"learning_rate": 0.00038415841584158415,
"loss": 0.9664,
"step": 8620
},
{
"epoch": 7.337907167359825,
"grad_norm": 18788.953125,
"learning_rate": 0.0003831683168316832,
"loss": 0.965,
"step": 8630
},
{
"epoch": 7.346407212895784,
"grad_norm": 14774.0830078125,
"learning_rate": 0.0003821782178217822,
"loss": 0.9666,
"step": 8640
},
{
"epoch": 7.346407212895784,
"eval_accuracy": 0.7926509573029413,
"eval_loss": 0.8783407807350159,
"eval_runtime": 1035.0192,
"eval_samples_per_second": 482.482,
"eval_steps_per_second": 0.821,
"step": 8640
},
{
"epoch": 7.354907258431742,
"grad_norm": 17107.38671875,
"learning_rate": 0.0003811881188118812,
"loss": 0.9666,
"step": 8650
},
{
"epoch": 7.3634073039676995,
"grad_norm": 15051.875,
"learning_rate": 0.0003801980198019802,
"loss": 0.9668,
"step": 8660
},
{
"epoch": 7.371907349503658,
"grad_norm": 16625.01171875,
"learning_rate": 0.0003792079207920792,
"loss": 0.9665,
"step": 8670
},
{
"epoch": 7.380407395039616,
"grad_norm": 15725.7158203125,
"learning_rate": 0.00037821782178217825,
"loss": 0.9665,
"step": 8680
},
{
"epoch": 7.388907440575575,
"grad_norm": 17703.29296875,
"learning_rate": 0.00037722772277227725,
"loss": 0.9653,
"step": 8690
},
{
"epoch": 7.397407486111533,
"grad_norm": 18342.701171875,
"learning_rate": 0.00037623762376237625,
"loss": 0.968,
"step": 8700
},
{
"epoch": 7.405907531647491,
"grad_norm": 16660.326171875,
"learning_rate": 0.00037524752475247524,
"loss": 0.9633,
"step": 8710
},
{
"epoch": 7.414407577183449,
"grad_norm": 14380.01171875,
"learning_rate": 0.00037425742574257424,
"loss": 0.9652,
"step": 8720
},
{
"epoch": 7.422907622719407,
"grad_norm": 15314.05859375,
"learning_rate": 0.0003732673267326733,
"loss": 0.9645,
"step": 8730
},
{
"epoch": 7.422907622719407,
"eval_accuracy": 0.7929085777897834,
"eval_loss": 0.8774629235267639,
"eval_runtime": 1036.1479,
"eval_samples_per_second": 481.956,
"eval_steps_per_second": 0.82,
"step": 8730
},
{
"epoch": 7.431407668255366,
"grad_norm": 15377.4306640625,
"learning_rate": 0.0003722772277227723,
"loss": 0.9661,
"step": 8740
},
{
"epoch": 7.439907713791324,
"grad_norm": 20652.443359375,
"learning_rate": 0.0003712871287128713,
"loss": 0.9638,
"step": 8750
},
{
"epoch": 7.448407759327282,
"grad_norm": 15570.021484375,
"learning_rate": 0.0003702970297029703,
"loss": 0.9647,
"step": 8760
},
{
"epoch": 7.45690780486324,
"grad_norm": 17774.888671875,
"learning_rate": 0.0003693069306930693,
"loss": 0.9651,
"step": 8770
},
{
"epoch": 7.465407850399199,
"grad_norm": 19272.056640625,
"learning_rate": 0.00036831683168316834,
"loss": 0.964,
"step": 8780
},
{
"epoch": 7.473907895935157,
"grad_norm": 17114.37109375,
"learning_rate": 0.00036732673267326734,
"loss": 0.963,
"step": 8790
},
{
"epoch": 7.482407941471115,
"grad_norm": 18956.96484375,
"learning_rate": 0.00036633663366336634,
"loss": 0.9648,
"step": 8800
},
{
"epoch": 7.490907987007073,
"grad_norm": 16027.392578125,
"learning_rate": 0.00036534653465346533,
"loss": 0.9646,
"step": 8810
},
{
"epoch": 7.499408032543031,
"grad_norm": 16447.48046875,
"learning_rate": 0.00036435643564356433,
"loss": 0.963,
"step": 8820
},
{
"epoch": 7.499408032543031,
"eval_accuracy": 0.7934318103664766,
"eval_loss": 0.8747227191925049,
"eval_runtime": 1031.384,
"eval_samples_per_second": 484.182,
"eval_steps_per_second": 0.824,
"step": 8820
},
{
"epoch": 7.50790807807899,
"grad_norm": 16898.626953125,
"learning_rate": 0.0003633663366336634,
"loss": 0.9651,
"step": 8830
},
{
"epoch": 7.516408123614948,
"grad_norm": 16217.8544921875,
"learning_rate": 0.0003623762376237624,
"loss": 0.9618,
"step": 8840
},
{
"epoch": 7.5249081691509065,
"grad_norm": 20356.615234375,
"learning_rate": 0.0003613861386138614,
"loss": 0.9601,
"step": 8850
},
{
"epoch": 7.533408214686864,
"grad_norm": 17453.779296875,
"learning_rate": 0.0003603960396039604,
"loss": 0.9627,
"step": 8860
},
{
"epoch": 7.541908260222822,
"grad_norm": 14681.1591796875,
"learning_rate": 0.0003594059405940594,
"loss": 0.9616,
"step": 8870
},
{
"epoch": 7.550408305758781,
"grad_norm": 16191.0478515625,
"learning_rate": 0.00035841584158415843,
"loss": 0.9623,
"step": 8880
},
{
"epoch": 7.558908351294739,
"grad_norm": 15307.25,
"learning_rate": 0.00035742574257425743,
"loss": 0.9604,
"step": 8890
},
{
"epoch": 7.5674083968306975,
"grad_norm": 18245.2265625,
"learning_rate": 0.0003564356435643564,
"loss": 0.9619,
"step": 8900
},
{
"epoch": 7.575908442366655,
"grad_norm": 16105.3232421875,
"learning_rate": 0.0003554455445544554,
"loss": 0.9643,
"step": 8910
},
{
"epoch": 7.575908442366655,
"eval_accuracy": 0.793696329436229,
"eval_loss": 0.8743059039115906,
"eval_runtime": 1035.7771,
"eval_samples_per_second": 482.129,
"eval_steps_per_second": 0.821,
"step": 8910
},
{
"epoch": 7.584408487902614,
"grad_norm": 16268.6376953125,
"learning_rate": 0.0003544554455445544,
"loss": 0.9632,
"step": 8920
},
{
"epoch": 7.592908533438572,
"grad_norm": 16207.1279296875,
"learning_rate": 0.0003534653465346535,
"loss": 0.9614,
"step": 8930
},
{
"epoch": 7.60140857897453,
"grad_norm": 15891.35546875,
"learning_rate": 0.0003524752475247525,
"loss": 0.9623,
"step": 8940
},
{
"epoch": 7.6099086245104886,
"grad_norm": 15914.4921875,
"learning_rate": 0.00035148514851485147,
"loss": 0.9609,
"step": 8950
},
{
"epoch": 7.618408670046446,
"grad_norm": 15182.478515625,
"learning_rate": 0.00035049504950495047,
"loss": 0.9601,
"step": 8960
},
{
"epoch": 7.626908715582405,
"grad_norm": 15736.513671875,
"learning_rate": 0.00034950495049504947,
"loss": 0.9613,
"step": 8970
},
{
"epoch": 7.635408761118363,
"grad_norm": 17880.693359375,
"learning_rate": 0.0003485148514851485,
"loss": 0.9606,
"step": 8980
},
{
"epoch": 7.643908806654322,
"grad_norm": 15555.0341796875,
"learning_rate": 0.0003475247524752475,
"loss": 0.9618,
"step": 8990
},
{
"epoch": 7.6524088521902796,
"grad_norm": 17536.287109375,
"learning_rate": 0.0003465346534653465,
"loss": 0.9606,
"step": 9000
},
{
"epoch": 7.6524088521902796,
"eval_accuracy": 0.7939732586317862,
"eval_loss": 0.8728435039520264,
"eval_runtime": 1035.2446,
"eval_samples_per_second": 482.377,
"eval_steps_per_second": 0.821,
"step": 9000
},
{
"epoch": 7.660908897726237,
"grad_norm": 16182.7158203125,
"learning_rate": 0.0003455445544554455,
"loss": 0.9605,
"step": 9010
},
{
"epoch": 7.669408943262196,
"grad_norm": 17641.724609375,
"learning_rate": 0.0003445544554455445,
"loss": 0.9611,
"step": 9020
},
{
"epoch": 7.677908988798154,
"grad_norm": 16540.5625,
"learning_rate": 0.0003435643564356436,
"loss": 0.9596,
"step": 9030
},
{
"epoch": 7.686409034334113,
"grad_norm": 15668.515625,
"learning_rate": 0.0003425742574257426,
"loss": 0.9592,
"step": 9040
},
{
"epoch": 7.694909079870071,
"grad_norm": 16943.318359375,
"learning_rate": 0.0003415841584158416,
"loss": 0.9619,
"step": 9050
},
{
"epoch": 7.703409125406029,
"grad_norm": 15736.9775390625,
"learning_rate": 0.0003405940594059406,
"loss": 0.959,
"step": 9060
},
{
"epoch": 7.711909170941987,
"grad_norm": 17073.3203125,
"learning_rate": 0.0003396039603960396,
"loss": 0.9609,
"step": 9070
},
{
"epoch": 7.720409216477945,
"grad_norm": 14390.5009765625,
"learning_rate": 0.00033861386138613867,
"loss": 0.9576,
"step": 9080
},
{
"epoch": 7.728909262013904,
"grad_norm": 17994.197265625,
"learning_rate": 0.00033762376237623766,
"loss": 0.9576,
"step": 9090
},
{
"epoch": 7.728909262013904,
"eval_accuracy": 0.7942283085944538,
"eval_loss": 0.8715931177139282,
"eval_runtime": 1034.9441,
"eval_samples_per_second": 482.517,
"eval_steps_per_second": 0.821,
"step": 9090
},
{
"epoch": 7.737409307549862,
"grad_norm": 14687.2646484375,
"learning_rate": 0.00033663366336633666,
"loss": 0.957,
"step": 9100
},
{
"epoch": 7.74590935308582,
"grad_norm": 16330.0478515625,
"learning_rate": 0.00033564356435643566,
"loss": 0.9595,
"step": 9110
},
{
"epoch": 7.754409398621778,
"grad_norm": 14873.998046875,
"learning_rate": 0.00033465346534653466,
"loss": 0.9596,
"step": 9120
},
{
"epoch": 7.762909444157737,
"grad_norm": 14278.1953125,
"learning_rate": 0.0003336633663366337,
"loss": 0.959,
"step": 9130
},
{
"epoch": 7.771409489693695,
"grad_norm": 17600.927734375,
"learning_rate": 0.0003326732673267327,
"loss": 0.9573,
"step": 9140
},
{
"epoch": 7.779909535229653,
"grad_norm": 15872.1806640625,
"learning_rate": 0.0003316831683168317,
"loss": 0.9573,
"step": 9150
},
{
"epoch": 7.788409580765611,
"grad_norm": 19368.9765625,
"learning_rate": 0.0003306930693069307,
"loss": 0.957,
"step": 9160
},
{
"epoch": 7.796909626301569,
"grad_norm": 19541.958984375,
"learning_rate": 0.0003297029702970297,
"loss": 0.9581,
"step": 9170
},
{
"epoch": 7.805409671837528,
"grad_norm": 16817.328125,
"learning_rate": 0.00032871287128712876,
"loss": 0.957,
"step": 9180
},
{
"epoch": 7.805409671837528,
"eval_accuracy": 0.7946937469682237,
"eval_loss": 0.8684272766113281,
"eval_runtime": 1033.8439,
"eval_samples_per_second": 483.03,
"eval_steps_per_second": 0.822,
"step": 9180
},
{
"epoch": 7.813909717373486,
"grad_norm": 18241.623046875,
"learning_rate": 0.00032772277227722775,
"loss": 0.9567,
"step": 9190
},
{
"epoch": 7.8224097629094445,
"grad_norm": 15321.79296875,
"learning_rate": 0.00032673267326732675,
"loss": 0.9574,
"step": 9200
},
{
"epoch": 7.830909808445402,
"grad_norm": 14790.1884765625,
"learning_rate": 0.00032574257425742575,
"loss": 0.9558,
"step": 9210
},
{
"epoch": 7.83940985398136,
"grad_norm": 13895.65234375,
"learning_rate": 0.00032475247524752475,
"loss": 0.9567,
"step": 9220
},
{
"epoch": 7.847909899517319,
"grad_norm": 17413.087890625,
"learning_rate": 0.0003237623762376238,
"loss": 0.9582,
"step": 9230
},
{
"epoch": 7.856409945053277,
"grad_norm": 15125.3447265625,
"learning_rate": 0.0003227722772277228,
"loss": 0.9555,
"step": 9240
},
{
"epoch": 7.8649099905892355,
"grad_norm": 15147.716796875,
"learning_rate": 0.0003217821782178218,
"loss": 0.9552,
"step": 9250
},
{
"epoch": 7.873410036125193,
"grad_norm": 16691.9765625,
"learning_rate": 0.0003207920792079208,
"loss": 0.9569,
"step": 9260
},
{
"epoch": 7.881910081661152,
"grad_norm": 15234.0234375,
"learning_rate": 0.0003198019801980198,
"loss": 0.957,
"step": 9270
},
{
"epoch": 7.881910081661152,
"eval_accuracy": 0.7946698774649099,
"eval_loss": 0.8689009547233582,
"eval_runtime": 1037.4044,
"eval_samples_per_second": 481.373,
"eval_steps_per_second": 0.819,
"step": 9270
},
{
"epoch": 7.89041012719711,
"grad_norm": 20217.794921875,
"learning_rate": 0.00031881188118811885,
"loss": 0.9558,
"step": 9280
},
{
"epoch": 7.898910172733068,
"grad_norm": 16890.482421875,
"learning_rate": 0.00031782178217821784,
"loss": 0.957,
"step": 9290
},
{
"epoch": 7.9074102182690265,
"grad_norm": 16310.05078125,
"learning_rate": 0.00031683168316831684,
"loss": 0.9557,
"step": 9300
},
{
"epoch": 7.915910263804984,
"grad_norm": 15330.0146484375,
"learning_rate": 0.00031584158415841584,
"loss": 0.9554,
"step": 9310
},
{
"epoch": 7.924410309340943,
"grad_norm": 14532.234375,
"learning_rate": 0.00031485148514851484,
"loss": 0.9542,
"step": 9320
},
{
"epoch": 7.932910354876901,
"grad_norm": 16035.0615234375,
"learning_rate": 0.0003138613861386139,
"loss": 0.9544,
"step": 9330
},
{
"epoch": 7.94141040041286,
"grad_norm": 16156.765625,
"learning_rate": 0.0003128712871287129,
"loss": 0.9549,
"step": 9340
},
{
"epoch": 7.9499104459488175,
"grad_norm": 16913.37109375,
"learning_rate": 0.0003118811881188119,
"loss": 0.9537,
"step": 9350
},
{
"epoch": 7.958410491484776,
"grad_norm": 16376.814453125,
"learning_rate": 0.0003108910891089109,
"loss": 0.9556,
"step": 9360
},
{
"epoch": 7.958410491484776,
"eval_accuracy": 0.7950512170789835,
"eval_loss": 0.8671656847000122,
"eval_runtime": 1035.1375,
"eval_samples_per_second": 482.427,
"eval_steps_per_second": 0.821,
"step": 9360
},
{
"epoch": 7.966910537020734,
"grad_norm": 14731.126953125,
"learning_rate": 0.0003099009900990099,
"loss": 0.9529,
"step": 9370
},
{
"epoch": 7.975410582556692,
"grad_norm": 16993.232421875,
"learning_rate": 0.00030891089108910894,
"loss": 0.9548,
"step": 9380
},
{
"epoch": 7.983910628092651,
"grad_norm": 16157.2880859375,
"learning_rate": 0.00030792079207920793,
"loss": 0.9543,
"step": 9390
},
{
"epoch": 7.9924106736286085,
"grad_norm": 17463.8984375,
"learning_rate": 0.00030693069306930693,
"loss": 0.9537,
"step": 9400
},
{
"epoch": 8.000910719164567,
"grad_norm": 18573.9765625,
"learning_rate": 0.00030594059405940593,
"loss": 0.9543,
"step": 9410
},
{
"epoch": 8.009410764700526,
"grad_norm": 18137.865234375,
"learning_rate": 0.00030495049504950493,
"loss": 0.9529,
"step": 9420
},
{
"epoch": 8.017910810236483,
"grad_norm": 17822.484375,
"learning_rate": 0.000303960396039604,
"loss": 0.9528,
"step": 9430
},
{
"epoch": 8.026410855772442,
"grad_norm": 16745.134765625,
"learning_rate": 0.000302970297029703,
"loss": 0.9529,
"step": 9440
},
{
"epoch": 8.0349109013084,
"grad_norm": 17063.685546875,
"learning_rate": 0.000301980198019802,
"loss": 0.9525,
"step": 9450
},
{
"epoch": 8.0349109013084,
"eval_accuracy": 0.7955066441462213,
"eval_loss": 0.8656001687049866,
"eval_runtime": 1034.9789,
"eval_samples_per_second": 482.501,
"eval_steps_per_second": 0.821,
"step": 9450
},
{
"epoch": 8.043410946844357,
"grad_norm": 14951.6552734375,
"learning_rate": 0.000300990099009901,
"loss": 0.9539,
"step": 9460
},
{
"epoch": 8.051910992380316,
"grad_norm": 18525.021484375,
"learning_rate": 0.0003,
"loss": 0.9529,
"step": 9470
},
{
"epoch": 8.060411037916275,
"grad_norm": 16683.45703125,
"learning_rate": 0.000299009900990099,
"loss": 0.9535,
"step": 9480
},
{
"epoch": 8.068911083452234,
"grad_norm": 17987.796875,
"learning_rate": 0.000298019801980198,
"loss": 0.9544,
"step": 9490
},
{
"epoch": 8.07741112898819,
"grad_norm": 14249.4453125,
"learning_rate": 0.000297029702970297,
"loss": 0.953,
"step": 9500
},
{
"epoch": 8.08591117452415,
"grad_norm": 15678.244140625,
"learning_rate": 0.000296039603960396,
"loss": 0.9517,
"step": 9510
},
{
"epoch": 8.094411220060108,
"grad_norm": 16945.619140625,
"learning_rate": 0.000295049504950495,
"loss": 0.9518,
"step": 9520
},
{
"epoch": 8.102911265596065,
"grad_norm": 16090.876953125,
"learning_rate": 0.00029405940594059407,
"loss": 0.9515,
"step": 9530
},
{
"epoch": 8.111411311132024,
"grad_norm": 15359.59375,
"learning_rate": 0.00029306930693069307,
"loss": 0.9509,
"step": 9540
},
{
"epoch": 8.111411311132024,
"eval_accuracy": 0.7956073067738585,
"eval_loss": 0.8639572858810425,
"eval_runtime": 1036.0688,
"eval_samples_per_second": 481.993,
"eval_steps_per_second": 0.82,
"step": 9540
},
{
"epoch": 8.119911356667982,
"grad_norm": 14432.5576171875,
"learning_rate": 0.00029207920792079207,
"loss": 0.9498,
"step": 9550
},
{
"epoch": 8.128411402203941,
"grad_norm": 15058.017578125,
"learning_rate": 0.00029108910891089107,
"loss": 0.9512,
"step": 9560
},
{
"epoch": 8.136911447739898,
"grad_norm": 14191.0791015625,
"learning_rate": 0.00029009900990099006,
"loss": 0.9517,
"step": 9570
},
{
"epoch": 8.145411493275857,
"grad_norm": 15950.939453125,
"learning_rate": 0.0002891089108910891,
"loss": 0.9506,
"step": 9580
},
{
"epoch": 8.153911538811816,
"grad_norm": 17082.619140625,
"learning_rate": 0.0002881188118811881,
"loss": 0.9513,
"step": 9590
},
{
"epoch": 8.162411584347772,
"grad_norm": 14874.2890625,
"learning_rate": 0.0002871287128712871,
"loss": 0.9501,
"step": 9600
},
{
"epoch": 8.170911629883731,
"grad_norm": 14615.6669921875,
"learning_rate": 0.0002861386138613861,
"loss": 0.951,
"step": 9610
},
{
"epoch": 8.17941167541969,
"grad_norm": 15949.583984375,
"learning_rate": 0.0002851485148514851,
"loss": 0.9511,
"step": 9620
},
{
"epoch": 8.187911720955649,
"grad_norm": 15501.064453125,
"learning_rate": 0.00028415841584158416,
"loss": 0.9503,
"step": 9630
},
{
"epoch": 8.187911720955649,
"eval_accuracy": 0.7959485569423415,
"eval_loss": 0.8628361821174622,
"eval_runtime": 1038.2269,
"eval_samples_per_second": 480.991,
"eval_steps_per_second": 0.819,
"step": 9630
},
{
"epoch": 8.196411766491606,
"grad_norm": 14815.34375,
"learning_rate": 0.00028316831683168316,
"loss": 0.9516,
"step": 9640
},
{
"epoch": 8.204911812027564,
"grad_norm": 19284.322265625,
"learning_rate": 0.00028217821782178216,
"loss": 0.9508,
"step": 9650
},
{
"epoch": 8.213411857563523,
"grad_norm": 15244.177734375,
"learning_rate": 0.0002811881188118812,
"loss": 0.9494,
"step": 9660
},
{
"epoch": 8.22191190309948,
"grad_norm": 14576.9482421875,
"learning_rate": 0.0002801980198019802,
"loss": 0.9485,
"step": 9670
},
{
"epoch": 8.230411948635439,
"grad_norm": 15386.0654296875,
"learning_rate": 0.00027920792079207926,
"loss": 0.9501,
"step": 9680
},
{
"epoch": 8.238911994171398,
"grad_norm": 15885.4248046875,
"learning_rate": 0.00027821782178217826,
"loss": 0.9493,
"step": 9690
},
{
"epoch": 8.247412039707356,
"grad_norm": 16190.021484375,
"learning_rate": 0.00027722772277227726,
"loss": 0.9498,
"step": 9700
},
{
"epoch": 8.255912085243313,
"grad_norm": 14406.8857421875,
"learning_rate": 0.00027623762376237626,
"loss": 0.9473,
"step": 9710
},
{
"epoch": 8.264412130779272,
"grad_norm": 14419.498046875,
"learning_rate": 0.00027524752475247525,
"loss": 0.9499,
"step": 9720
},
{
"epoch": 8.264412130779272,
"eval_accuracy": 0.7962577759360605,
"eval_loss": 0.8620018362998962,
"eval_runtime": 1037.7206,
"eval_samples_per_second": 481.226,
"eval_steps_per_second": 0.819,
"step": 9720
},
{
"epoch": 8.27291217631523,
"grad_norm": 16101.1533203125,
"learning_rate": 0.0002742574257425743,
"loss": 0.9502,
"step": 9730
},
{
"epoch": 8.281412221851188,
"grad_norm": 15671.9033203125,
"learning_rate": 0.0002732673267326733,
"loss": 0.9498,
"step": 9740
},
{
"epoch": 8.289912267387146,
"grad_norm": 15215.625,
"learning_rate": 0.0002722772277227723,
"loss": 0.9482,
"step": 9750
},
{
"epoch": 8.298412312923105,
"grad_norm": 16016.99609375,
"learning_rate": 0.0002712871287128713,
"loss": 0.9494,
"step": 9760
},
{
"epoch": 8.306912358459064,
"grad_norm": 17474.0859375,
"learning_rate": 0.0002702970297029703,
"loss": 0.9497,
"step": 9770
},
{
"epoch": 8.31541240399502,
"grad_norm": 13742.4052734375,
"learning_rate": 0.00026930693069306935,
"loss": 0.947,
"step": 9780
},
{
"epoch": 8.32391244953098,
"grad_norm": 15790.54296875,
"learning_rate": 0.00026831683168316835,
"loss": 0.9471,
"step": 9790
},
{
"epoch": 8.332412495066938,
"grad_norm": 16040.060546875,
"learning_rate": 0.00026732673267326735,
"loss": 0.9489,
"step": 9800
},
{
"epoch": 8.340912540602897,
"grad_norm": 14828.8095703125,
"learning_rate": 0.00026633663366336635,
"loss": 0.9482,
"step": 9810
},
{
"epoch": 8.340912540602897,
"eval_accuracy": 0.7962955020928217,
"eval_loss": 0.8608699440956116,
"eval_runtime": 1034.0211,
"eval_samples_per_second": 482.948,
"eval_steps_per_second": 0.822,
"step": 9810
},
{
"epoch": 8.349412586138854,
"grad_norm": 16481.876953125,
"learning_rate": 0.00026534653465346534,
"loss": 0.9478,
"step": 9820
},
{
"epoch": 8.357912631674813,
"grad_norm": 16586.9609375,
"learning_rate": 0.0002643564356435644,
"loss": 0.9491,
"step": 9830
},
{
"epoch": 8.366412677210771,
"grad_norm": 16209.767578125,
"learning_rate": 0.0002633663366336634,
"loss": 0.9484,
"step": 9840
},
{
"epoch": 8.374912722746728,
"grad_norm": 15675.52734375,
"learning_rate": 0.0002623762376237624,
"loss": 0.9483,
"step": 9850
},
{
"epoch": 8.383412768282687,
"grad_norm": 16780.919921875,
"learning_rate": 0.0002613861386138614,
"loss": 0.9466,
"step": 9860
},
{
"epoch": 8.391912813818646,
"grad_norm": 14785.990234375,
"learning_rate": 0.0002603960396039604,
"loss": 0.9459,
"step": 9870
},
{
"epoch": 8.400412859354603,
"grad_norm": 14482.4794921875,
"learning_rate": 0.00025940594059405944,
"loss": 0.9464,
"step": 9880
},
{
"epoch": 8.408912904890562,
"grad_norm": 17260.060546875,
"learning_rate": 0.00025841584158415844,
"loss": 0.947,
"step": 9890
},
{
"epoch": 8.41741295042652,
"grad_norm": 15805.9501953125,
"learning_rate": 0.00025742574257425744,
"loss": 0.9454,
"step": 9900
},
{
"epoch": 8.41741295042652,
"eval_accuracy": 0.7969292599781366,
"eval_loss": 0.8581969738006592,
"eval_runtime": 1037.2614,
"eval_samples_per_second": 481.439,
"eval_steps_per_second": 0.819,
"step": 9900
},
{
"epoch": 8.425912995962479,
"grad_norm": 19391.345703125,
"learning_rate": 0.00025643564356435644,
"loss": 0.9447,
"step": 9910
},
{
"epoch": 8.434413041498436,
"grad_norm": 15079.96875,
"learning_rate": 0.00025544554455445543,
"loss": 0.9472,
"step": 9920
},
{
"epoch": 8.442913087034395,
"grad_norm": 16290.3056640625,
"learning_rate": 0.0002544554455445545,
"loss": 0.9455,
"step": 9930
},
{
"epoch": 8.451413132570353,
"grad_norm": 14815.7783203125,
"learning_rate": 0.0002534653465346535,
"loss": 0.9453,
"step": 9940
},
{
"epoch": 8.459913178106312,
"grad_norm": 14300.2734375,
"learning_rate": 0.0002524752475247525,
"loss": 0.9448,
"step": 9950
},
{
"epoch": 8.46841322364227,
"grad_norm": 16484.212890625,
"learning_rate": 0.0002514851485148515,
"loss": 0.9471,
"step": 9960
},
{
"epoch": 8.476913269178228,
"grad_norm": 14792.8935546875,
"learning_rate": 0.0002504950495049505,
"loss": 0.9459,
"step": 9970
},
{
"epoch": 8.485413314714187,
"grad_norm": 15095.5869140625,
"learning_rate": 0.00024950495049504953,
"loss": 0.9439,
"step": 9980
},
{
"epoch": 8.493913360250144,
"grad_norm": 15246.3203125,
"learning_rate": 0.00024851485148514853,
"loss": 0.9459,
"step": 9990
},
{
"epoch": 8.493913360250144,
"eval_accuracy": 0.7969879777309228,
"eval_loss": 0.8577154278755188,
"eval_runtime": 1035.8486,
"eval_samples_per_second": 482.096,
"eval_steps_per_second": 0.821,
"step": 9990
},
{
"epoch": 8.502413405786102,
"grad_norm": 15397.66796875,
"learning_rate": 0.00024752475247524753,
"loss": 0.9443,
"step": 10000
},
{
"epoch": 8.510913451322061,
"grad_norm": 15283.72265625,
"learning_rate": 0.0002465346534653465,
"loss": 0.9446,
"step": 10010
},
{
"epoch": 8.519413496858018,
"grad_norm": 15016.5234375,
"learning_rate": 0.0002455445544554455,
"loss": 0.9446,
"step": 10020
},
{
"epoch": 8.527913542393977,
"grad_norm": 14347.6044921875,
"learning_rate": 0.0002445544554455446,
"loss": 0.945,
"step": 10030
},
{
"epoch": 8.536413587929935,
"grad_norm": 17782.130859375,
"learning_rate": 0.00024356435643564357,
"loss": 0.9445,
"step": 10040
},
{
"epoch": 8.544913633465894,
"grad_norm": 14884.66796875,
"learning_rate": 0.00024257425742574257,
"loss": 0.9429,
"step": 10050
},
{
"epoch": 8.553413679001851,
"grad_norm": 17597.05859375,
"learning_rate": 0.00024158415841584157,
"loss": 0.9443,
"step": 10060
},
{
"epoch": 8.56191372453781,
"grad_norm": 17000.009765625,
"learning_rate": 0.0002405940594059406,
"loss": 0.9449,
"step": 10070
},
{
"epoch": 8.570413770073769,
"grad_norm": 17303.015625,
"learning_rate": 0.0002396039603960396,
"loss": 0.9444,
"step": 10080
},
{
"epoch": 8.570413770073769,
"eval_accuracy": 0.7970984014477712,
"eval_loss": 0.8577408194541931,
"eval_runtime": 1033.5768,
"eval_samples_per_second": 483.155,
"eval_steps_per_second": 0.822,
"step": 10080
},
{
"epoch": 8.578913815609727,
"grad_norm": 14883.8759765625,
"learning_rate": 0.00023861386138613862,
"loss": 0.9429,
"step": 10090
},
{
"epoch": 8.587413861145684,
"grad_norm": 16271.083984375,
"learning_rate": 0.00023762376237623762,
"loss": 0.9441,
"step": 10100
},
{
"epoch": 8.595913906681643,
"grad_norm": 18421.546875,
"learning_rate": 0.00023663366336633662,
"loss": 0.9444,
"step": 10110
},
{
"epoch": 8.604413952217602,
"grad_norm": 19686.576171875,
"learning_rate": 0.00023564356435643564,
"loss": 0.9434,
"step": 10120
},
{
"epoch": 8.612913997753559,
"grad_norm": 14919.0126953125,
"learning_rate": 0.00023465346534653464,
"loss": 0.9427,
"step": 10130
},
{
"epoch": 8.621414043289517,
"grad_norm": 16443.49609375,
"learning_rate": 0.0002336633663366337,
"loss": 0.9412,
"step": 10140
},
{
"epoch": 8.629914088825476,
"grad_norm": 14300.4599609375,
"learning_rate": 0.0002326732673267327,
"loss": 0.9436,
"step": 10150
},
{
"epoch": 8.638414134361435,
"grad_norm": 17048.576171875,
"learning_rate": 0.0002316831683168317,
"loss": 0.9438,
"step": 10160
},
{
"epoch": 8.646914179897392,
"grad_norm": 14095.205078125,
"learning_rate": 0.00023069306930693071,
"loss": 0.9419,
"step": 10170
},
{
"epoch": 8.646914179897392,
"eval_accuracy": 0.797700358706027,
"eval_loss": 0.854520857334137,
"eval_runtime": 1038.4284,
"eval_samples_per_second": 480.898,
"eval_steps_per_second": 0.819,
"step": 10170
},
{
"epoch": 8.65541422543335,
"grad_norm": 16067.9052734375,
"learning_rate": 0.0002297029702970297,
"loss": 0.9409,
"step": 10180
},
{
"epoch": 8.66391427096931,
"grad_norm": 14993.923828125,
"learning_rate": 0.00022871287128712874,
"loss": 0.9421,
"step": 10190
},
{
"epoch": 8.672414316505266,
"grad_norm": 14386.7099609375,
"learning_rate": 0.00022772277227722774,
"loss": 0.9428,
"step": 10200
},
{
"epoch": 8.680914362041225,
"grad_norm": 15530.26953125,
"learning_rate": 0.00022673267326732673,
"loss": 0.9425,
"step": 10210
},
{
"epoch": 8.689414407577184,
"grad_norm": 15528.4541015625,
"learning_rate": 0.00022574257425742576,
"loss": 0.941,
"step": 10220
},
{
"epoch": 8.697914453113142,
"grad_norm": 14304.1318359375,
"learning_rate": 0.00022475247524752476,
"loss": 0.9402,
"step": 10230
},
{
"epoch": 8.7064144986491,
"grad_norm": 15193.4375,
"learning_rate": 0.00022376237623762378,
"loss": 0.9415,
"step": 10240
},
{
"epoch": 8.714914544185058,
"grad_norm": 14937.7109375,
"learning_rate": 0.00022277227722772278,
"loss": 0.9421,
"step": 10250
},
{
"epoch": 8.723414589721017,
"grad_norm": 15369.6669921875,
"learning_rate": 0.00022178217821782178,
"loss": 0.9415,
"step": 10260
},
{
"epoch": 8.723414589721017,
"eval_accuracy": 0.797684357783119,
"eval_loss": 0.8541524410247803,
"eval_runtime": 1037.1311,
"eval_samples_per_second": 481.499,
"eval_steps_per_second": 0.82,
"step": 10260
},
{
"epoch": 8.731914635256974,
"grad_norm": 17404.70703125,
"learning_rate": 0.0002207920792079208,
"loss": 0.9417,
"step": 10270
},
{
"epoch": 8.740414680792933,
"grad_norm": 16622.16796875,
"learning_rate": 0.0002198019801980198,
"loss": 0.9409,
"step": 10280
},
{
"epoch": 8.748914726328891,
"grad_norm": 13510.8125,
"learning_rate": 0.00021881188118811883,
"loss": 0.9409,
"step": 10290
},
{
"epoch": 8.75741477186485,
"grad_norm": 14584.3173828125,
"learning_rate": 0.00021782178217821783,
"loss": 0.9411,
"step": 10300
},
{
"epoch": 8.765914817400807,
"grad_norm": 13928.373046875,
"learning_rate": 0.00021683168316831682,
"loss": 0.9408,
"step": 10310
},
{
"epoch": 8.774414862936766,
"grad_norm": 13927.751953125,
"learning_rate": 0.00021584158415841585,
"loss": 0.9404,
"step": 10320
},
{
"epoch": 8.782914908472724,
"grad_norm": 15996.322265625,
"learning_rate": 0.00021485148514851485,
"loss": 0.9415,
"step": 10330
},
{
"epoch": 8.791414954008681,
"grad_norm": 14078.8876953125,
"learning_rate": 0.00021386138613861387,
"loss": 0.9392,
"step": 10340
},
{
"epoch": 8.79991499954464,
"grad_norm": 14735.3623046875,
"learning_rate": 0.00021287128712871287,
"loss": 0.9395,
"step": 10350
},
{
"epoch": 8.79991499954464,
"eval_accuracy": 0.7978062402609887,
"eval_loss": 0.8535209894180298,
"eval_runtime": 1036.7203,
"eval_samples_per_second": 481.69,
"eval_steps_per_second": 0.82,
"step": 10350
},
{
"epoch": 8.808415045080599,
"grad_norm": 14420.7861328125,
"learning_rate": 0.00021188118811881187,
"loss": 0.9412,
"step": 10360
},
{
"epoch": 8.816915090616558,
"grad_norm": 14367.3955078125,
"learning_rate": 0.0002108910891089109,
"loss": 0.9382,
"step": 10370
},
{
"epoch": 8.825415136152515,
"grad_norm": 14271.1162109375,
"learning_rate": 0.0002099009900990099,
"loss": 0.9406,
"step": 10380
},
{
"epoch": 8.833915181688473,
"grad_norm": 14588.7724609375,
"learning_rate": 0.00020891089108910892,
"loss": 0.9397,
"step": 10390
},
{
"epoch": 8.842415227224432,
"grad_norm": 13847.20703125,
"learning_rate": 0.00020792079207920792,
"loss": 0.9394,
"step": 10400
},
{
"epoch": 8.850915272760389,
"grad_norm": 14702.3779296875,
"learning_rate": 0.00020693069306930691,
"loss": 0.9386,
"step": 10410
},
{
"epoch": 8.859415318296348,
"grad_norm": 13664.92578125,
"learning_rate": 0.00020594059405940594,
"loss": 0.939,
"step": 10420
},
{
"epoch": 8.867915363832306,
"grad_norm": 14917.166015625,
"learning_rate": 0.00020495049504950494,
"loss": 0.9392,
"step": 10430
},
{
"epoch": 8.876415409368265,
"grad_norm": 15233.123046875,
"learning_rate": 0.00020396039603960396,
"loss": 0.9411,
"step": 10440
},
{
"epoch": 8.876415409368265,
"eval_accuracy": 0.7982283642493349,
"eval_loss": 0.8512039184570312,
"eval_runtime": 1037.2463,
"eval_samples_per_second": 481.446,
"eval_steps_per_second": 0.819,
"step": 10440
},
{
"epoch": 8.884915454904222,
"grad_norm": 14748.51953125,
"learning_rate": 0.000202970297029703,
"loss": 0.9379,
"step": 10450
},
{
"epoch": 8.893415500440181,
"grad_norm": 14725.625,
"learning_rate": 0.00020198019801980199,
"loss": 0.9394,
"step": 10460
},
{
"epoch": 8.90191554597614,
"grad_norm": 14719.560546875,
"learning_rate": 0.000200990099009901,
"loss": 0.9396,
"step": 10470
},
{
"epoch": 8.910415591512097,
"grad_norm": 15252.0546875,
"learning_rate": 0.0002,
"loss": 0.9386,
"step": 10480
},
{
"epoch": 8.918915637048055,
"grad_norm": 14756.244140625,
"learning_rate": 0.00019900990099009903,
"loss": 0.9378,
"step": 10490
},
{
"epoch": 8.927415682584014,
"grad_norm": 14705.61328125,
"learning_rate": 0.00019801980198019803,
"loss": 0.9377,
"step": 10500
},
{
"epoch": 8.935915728119973,
"grad_norm": 14071.0009765625,
"learning_rate": 0.00019702970297029703,
"loss": 0.9394,
"step": 10510
},
{
"epoch": 8.94441577365593,
"grad_norm": 13337.140625,
"learning_rate": 0.00019603960396039606,
"loss": 0.9384,
"step": 10520
},
{
"epoch": 8.952915819191889,
"grad_norm": 14109.1650390625,
"learning_rate": 0.00019504950495049505,
"loss": 0.9362,
"step": 10530
},
{
"epoch": 8.952915819191889,
"eval_accuracy": 0.7982979521046931,
"eval_loss": 0.8517733812332153,
"eval_runtime": 1038.3162,
"eval_samples_per_second": 480.95,
"eval_steps_per_second": 0.819,
"step": 10530
},
{
"epoch": 8.961415864727847,
"grad_norm": 17169.69140625,
"learning_rate": 0.00019405940594059408,
"loss": 0.9373,
"step": 10540
},
{
"epoch": 8.969915910263804,
"grad_norm": 14149.8935546875,
"learning_rate": 0.00019306930693069308,
"loss": 0.9365,
"step": 10550
},
{
"epoch": 8.978415955799763,
"grad_norm": 16067.255859375,
"learning_rate": 0.00019207920792079208,
"loss": 0.9371,
"step": 10560
},
{
"epoch": 8.986916001335722,
"grad_norm": 14375.869140625,
"learning_rate": 0.0001910891089108911,
"loss": 0.9375,
"step": 10570
},
{
"epoch": 8.99541604687168,
"grad_norm": 14484.7421875,
"learning_rate": 0.0001900990099009901,
"loss": 0.9372,
"step": 10580
},
{
"epoch": 9.003916092407637,
"grad_norm": 13759.2275390625,
"learning_rate": 0.00018910891089108913,
"loss": 0.9361,
"step": 10590
},
{
"epoch": 9.012416137943596,
"grad_norm": 14114.7119140625,
"learning_rate": 0.00018811881188118812,
"loss": 0.9369,
"step": 10600
},
{
"epoch": 9.020916183479555,
"grad_norm": 13548.9287109375,
"learning_rate": 0.00018712871287128712,
"loss": 0.9363,
"step": 10610
},
{
"epoch": 9.029416229015512,
"grad_norm": 13845.412109375,
"learning_rate": 0.00018613861386138615,
"loss": 0.9369,
"step": 10620
},
{
"epoch": 9.029416229015512,
"eval_accuracy": 0.7986381309887046,
"eval_loss": 0.8504059314727783,
"eval_runtime": 1037.8808,
"eval_samples_per_second": 481.152,
"eval_steps_per_second": 0.819,
"step": 10620
},
{
"epoch": 9.03791627455147,
"grad_norm": 14264.1455078125,
"learning_rate": 0.00018514851485148514,
"loss": 0.9371,
"step": 10630
},
{
"epoch": 9.04641632008743,
"grad_norm": 12720.0791015625,
"learning_rate": 0.00018415841584158417,
"loss": 0.9351,
"step": 10640
},
{
"epoch": 9.054916365623388,
"grad_norm": 14962.2216796875,
"learning_rate": 0.00018316831683168317,
"loss": 0.9347,
"step": 10650
},
{
"epoch": 9.063416411159345,
"grad_norm": 14343.6259765625,
"learning_rate": 0.00018217821782178217,
"loss": 0.9366,
"step": 10660
},
{
"epoch": 9.071916456695304,
"grad_norm": 15166.927734375,
"learning_rate": 0.0001811881188118812,
"loss": 0.9379,
"step": 10670
},
{
"epoch": 9.080416502231262,
"grad_norm": 13523.9873046875,
"learning_rate": 0.0001801980198019802,
"loss": 0.9369,
"step": 10680
},
{
"epoch": 9.08891654776722,
"grad_norm": 15057.4638671875,
"learning_rate": 0.00017920792079207922,
"loss": 0.9368,
"step": 10690
},
{
"epoch": 9.097416593303178,
"grad_norm": 13766.353515625,
"learning_rate": 0.0001782178217821782,
"loss": 0.935,
"step": 10700
},
{
"epoch": 9.105916638839137,
"grad_norm": 15783.3056640625,
"learning_rate": 0.0001772277227722772,
"loss": 0.9365,
"step": 10710
},
{
"epoch": 9.105916638839137,
"eval_accuracy": 0.7986280323340061,
"eval_loss": 0.8495949506759644,
"eval_runtime": 1036.1574,
"eval_samples_per_second": 481.952,
"eval_steps_per_second": 0.82,
"step": 10710
},
{
"epoch": 9.114416684375096,
"grad_norm": 14088.314453125,
"learning_rate": 0.00017623762376237624,
"loss": 0.9363,
"step": 10720
},
{
"epoch": 9.122916729911053,
"grad_norm": 14255.0244140625,
"learning_rate": 0.00017524752475247524,
"loss": 0.9351,
"step": 10730
},
{
"epoch": 9.131416775447011,
"grad_norm": 15012.841796875,
"learning_rate": 0.00017425742574257426,
"loss": 0.9341,
"step": 10740
},
{
"epoch": 9.13991682098297,
"grad_norm": 15262.3740234375,
"learning_rate": 0.00017326732673267326,
"loss": 0.9342,
"step": 10750
},
{
"epoch": 9.148416866518927,
"grad_norm": 14390.744140625,
"learning_rate": 0.00017227722772277226,
"loss": 0.9359,
"step": 10760
},
{
"epoch": 9.156916912054886,
"grad_norm": 14124.2373046875,
"learning_rate": 0.0001712871287128713,
"loss": 0.9353,
"step": 10770
},
{
"epoch": 9.165416957590844,
"grad_norm": 13361.82421875,
"learning_rate": 0.0001702970297029703,
"loss": 0.9352,
"step": 10780
},
{
"epoch": 9.173917003126803,
"grad_norm": 13731.3984375,
"learning_rate": 0.00016930693069306933,
"loss": 0.9345,
"step": 10790
},
{
"epoch": 9.18241704866276,
"grad_norm": 13207.9716796875,
"learning_rate": 0.00016831683168316833,
"loss": 0.9336,
"step": 10800
},
{
"epoch": 9.18241704866276,
"eval_accuracy": 0.7991505659503615,
"eval_loss": 0.8472453355789185,
"eval_runtime": 1035.1715,
"eval_samples_per_second": 482.411,
"eval_steps_per_second": 0.821,
"step": 10800
},
{
"epoch": 9.190917094198719,
"grad_norm": 14880.1533203125,
"learning_rate": 0.00016732673267326733,
"loss": 0.9333,
"step": 10810
},
{
"epoch": 9.199417139734678,
"grad_norm": 12751.9345703125,
"learning_rate": 0.00016633663366336635,
"loss": 0.935,
"step": 10820
},
{
"epoch": 9.207917185270635,
"grad_norm": 13587.4658203125,
"learning_rate": 0.00016534653465346535,
"loss": 0.9336,
"step": 10830
},
{
"epoch": 9.216417230806593,
"grad_norm": 13916.77734375,
"learning_rate": 0.00016435643564356438,
"loss": 0.9348,
"step": 10840
},
{
"epoch": 9.224917276342552,
"grad_norm": 14346.2119140625,
"learning_rate": 0.00016336633663366338,
"loss": 0.9328,
"step": 10850
},
{
"epoch": 9.23341732187851,
"grad_norm": 11983.17578125,
"learning_rate": 0.00016237623762376237,
"loss": 0.9341,
"step": 10860
},
{
"epoch": 9.241917367414468,
"grad_norm": 13424.4794921875,
"learning_rate": 0.0001613861386138614,
"loss": 0.9341,
"step": 10870
},
{
"epoch": 9.250417412950426,
"grad_norm": 13093.9267578125,
"learning_rate": 0.0001603960396039604,
"loss": 0.9336,
"step": 10880
},
{
"epoch": 9.258917458486385,
"grad_norm": 14450.0888671875,
"learning_rate": 0.00015940594059405942,
"loss": 0.9326,
"step": 10890
},
{
"epoch": 9.258917458486385,
"eval_accuracy": 0.7993350474335167,
"eval_loss": 0.846928596496582,
"eval_runtime": 1034.8058,
"eval_samples_per_second": 482.581,
"eval_steps_per_second": 0.821,
"step": 10890
},
{
"epoch": 9.267417504022344,
"grad_norm": 12639.7099609375,
"learning_rate": 0.00015841584158415842,
"loss": 0.9331,
"step": 10900
},
{
"epoch": 9.2759175495583,
"grad_norm": 14105.6357421875,
"learning_rate": 0.00015742574257425742,
"loss": 0.9323,
"step": 10910
},
{
"epoch": 9.28441759509426,
"grad_norm": 13878.3876953125,
"learning_rate": 0.00015643564356435644,
"loss": 0.9319,
"step": 10920
},
{
"epoch": 9.292917640630218,
"grad_norm": 14202.0380859375,
"learning_rate": 0.00015544554455445544,
"loss": 0.9309,
"step": 10930
},
{
"epoch": 9.301417686166175,
"grad_norm": 12203.2275390625,
"learning_rate": 0.00015445544554455447,
"loss": 0.9323,
"step": 10940
},
{
"epoch": 9.309917731702134,
"grad_norm": 12799.16015625,
"learning_rate": 0.00015346534653465347,
"loss": 0.933,
"step": 10950
},
{
"epoch": 9.318417777238093,
"grad_norm": 14222.4921875,
"learning_rate": 0.00015247524752475246,
"loss": 0.9335,
"step": 10960
},
{
"epoch": 9.32691782277405,
"grad_norm": 13150.4560546875,
"learning_rate": 0.0001514851485148515,
"loss": 0.9334,
"step": 10970
},
{
"epoch": 9.335417868310008,
"grad_norm": 13035.662109375,
"learning_rate": 0.0001504950495049505,
"loss": 0.9336,
"step": 10980
},
{
"epoch": 9.335417868310008,
"eval_accuracy": 0.7995125837200675,
"eval_loss": 0.8457638621330261,
"eval_runtime": 1036.0457,
"eval_samples_per_second": 482.004,
"eval_steps_per_second": 0.82,
"step": 10980
},
{
"epoch": 9.343917913845967,
"grad_norm": 13063.5947265625,
"learning_rate": 0.0001495049504950495,
"loss": 0.9358,
"step": 10990
},
{
"epoch": 9.352417959381926,
"grad_norm": 13189.0205078125,
"learning_rate": 0.0001485148514851485,
"loss": 0.9314,
"step": 11000
},
{
"epoch": 9.360918004917883,
"grad_norm": 14116.0537109375,
"learning_rate": 0.0001475247524752475,
"loss": 0.9322,
"step": 11010
},
{
"epoch": 9.369418050453842,
"grad_norm": 13098.5810546875,
"learning_rate": 0.00014653465346534653,
"loss": 0.9306,
"step": 11020
},
{
"epoch": 9.3779180959898,
"grad_norm": 12630.921875,
"learning_rate": 0.00014554455445544553,
"loss": 0.9294,
"step": 11030
},
{
"epoch": 9.386418141525759,
"grad_norm": 12936.2509765625,
"learning_rate": 0.00014455445544554456,
"loss": 0.9323,
"step": 11040
},
{
"epoch": 9.394918187061716,
"grad_norm": 13526.9853515625,
"learning_rate": 0.00014356435643564356,
"loss": 0.9311,
"step": 11050
},
{
"epoch": 9.403418232597675,
"grad_norm": 13259.3603515625,
"learning_rate": 0.00014257425742574255,
"loss": 0.9332,
"step": 11060
},
{
"epoch": 9.411918278133633,
"grad_norm": 14287.5361328125,
"learning_rate": 0.00014158415841584158,
"loss": 0.9321,
"step": 11070
},
{
"epoch": 9.411918278133633,
"eval_accuracy": 0.7996804125615878,
"eval_loss": 0.8447943329811096,
"eval_runtime": 1035.7879,
"eval_samples_per_second": 482.124,
"eval_steps_per_second": 0.821,
"step": 11070
},
{
"epoch": 9.42041832366959,
"grad_norm": 12443.166015625,
"learning_rate": 0.0001405940594059406,
"loss": 0.9303,
"step": 11080
},
{
"epoch": 9.42891836920555,
"grad_norm": 13510.2119140625,
"learning_rate": 0.00013960396039603963,
"loss": 0.9298,
"step": 11090
},
{
"epoch": 9.437418414741508,
"grad_norm": 11732.4970703125,
"learning_rate": 0.00013861386138613863,
"loss": 0.9314,
"step": 11100
},
{
"epoch": 9.445918460277467,
"grad_norm": 13648.421875,
"learning_rate": 0.00013762376237623763,
"loss": 0.9318,
"step": 11110
},
{
"epoch": 9.454418505813424,
"grad_norm": 12602.56640625,
"learning_rate": 0.00013663366336633665,
"loss": 0.9318,
"step": 11120
},
{
"epoch": 9.462918551349382,
"grad_norm": 13816.3720703125,
"learning_rate": 0.00013564356435643565,
"loss": 0.9319,
"step": 11130
},
{
"epoch": 9.471418596885341,
"grad_norm": 14125.748046875,
"learning_rate": 0.00013465346534653468,
"loss": 0.9297,
"step": 11140
},
{
"epoch": 9.479918642421298,
"grad_norm": 13495.037109375,
"learning_rate": 0.00013366336633663367,
"loss": 0.9314,
"step": 11150
},
{
"epoch": 9.488418687957257,
"grad_norm": 12772.404296875,
"learning_rate": 0.00013267326732673267,
"loss": 0.9283,
"step": 11160
},
{
"epoch": 9.488418687957257,
"eval_accuracy": 0.799732345307766,
"eval_loss": 0.8448570370674133,
"eval_runtime": 1038.7893,
"eval_samples_per_second": 480.731,
"eval_steps_per_second": 0.818,
"step": 11160
},
{
"epoch": 9.496918733493215,
"grad_norm": 13424.62890625,
"learning_rate": 0.0001316831683168317,
"loss": 0.9297,
"step": 11170
},
{
"epoch": 9.505418779029174,
"grad_norm": 14624.0283203125,
"learning_rate": 0.0001306930693069307,
"loss": 0.9295,
"step": 11180
},
{
"epoch": 9.513918824565131,
"grad_norm": 12779.9169921875,
"learning_rate": 0.00012970297029702972,
"loss": 0.9322,
"step": 11190
},
{
"epoch": 9.52241887010109,
"grad_norm": 13789.7421875,
"learning_rate": 0.00012871287128712872,
"loss": 0.9312,
"step": 11200
},
{
"epoch": 9.530918915637049,
"grad_norm": 12679.0703125,
"learning_rate": 0.00012772277227722772,
"loss": 0.9289,
"step": 11210
},
{
"epoch": 9.539418961173006,
"grad_norm": 14479.2919921875,
"learning_rate": 0.00012673267326732674,
"loss": 0.931,
"step": 11220
},
{
"epoch": 9.547919006708964,
"grad_norm": 12816.5322265625,
"learning_rate": 0.00012574257425742574,
"loss": 0.9304,
"step": 11230
},
{
"epoch": 9.556419052244923,
"grad_norm": 13485.4990234375,
"learning_rate": 0.00012475247524752477,
"loss": 0.928,
"step": 11240
},
{
"epoch": 9.564919097780882,
"grad_norm": 13575.083984375,
"learning_rate": 0.00012376237623762376,
"loss": 0.929,
"step": 11250
},
{
"epoch": 9.564919097780882,
"eval_accuracy": 0.7999083549113648,
"eval_loss": 0.8442253470420837,
"eval_runtime": 1035.3058,
"eval_samples_per_second": 482.348,
"eval_steps_per_second": 0.821,
"step": 11250
},
{
"epoch": 9.573419143316839,
"grad_norm": 13030.95703125,
"learning_rate": 0.00012277227722772276,
"loss": 0.9303,
"step": 11260
},
{
"epoch": 9.581919188852797,
"grad_norm": 12883.271484375,
"learning_rate": 0.00012178217821782179,
"loss": 0.9304,
"step": 11270
},
{
"epoch": 9.590419234388756,
"grad_norm": 12350.5322265625,
"learning_rate": 0.00012079207920792079,
"loss": 0.9298,
"step": 11280
},
{
"epoch": 9.598919279924713,
"grad_norm": 12253.1220703125,
"learning_rate": 0.0001198019801980198,
"loss": 0.9301,
"step": 11290
},
{
"epoch": 9.607419325460672,
"grad_norm": 13372.912109375,
"learning_rate": 0.00011881188118811881,
"loss": 0.9298,
"step": 11300
},
{
"epoch": 9.61591937099663,
"grad_norm": 12150.8623046875,
"learning_rate": 0.00011782178217821782,
"loss": 0.9276,
"step": 11310
},
{
"epoch": 9.62441941653259,
"grad_norm": 12663.35546875,
"learning_rate": 0.00011683168316831685,
"loss": 0.9305,
"step": 11320
},
{
"epoch": 9.632919462068546,
"grad_norm": 12773.25,
"learning_rate": 0.00011584158415841584,
"loss": 0.929,
"step": 11330
},
{
"epoch": 9.641419507604505,
"grad_norm": 12251.90234375,
"learning_rate": 0.00011485148514851486,
"loss": 0.9282,
"step": 11340
},
{
"epoch": 9.641419507604505,
"eval_accuracy": 0.8002882889018181,
"eval_loss": 0.842271089553833,
"eval_runtime": 1038.9797,
"eval_samples_per_second": 480.643,
"eval_steps_per_second": 0.818,
"step": 11340
},
{
"epoch": 9.649919553140464,
"grad_norm": 12380.44140625,
"learning_rate": 0.00011386138613861387,
"loss": 0.9275,
"step": 11350
},
{
"epoch": 9.65841959867642,
"grad_norm": 12090.271484375,
"learning_rate": 0.00011287128712871288,
"loss": 0.9275,
"step": 11360
},
{
"epoch": 9.66691964421238,
"grad_norm": 13393.486328125,
"learning_rate": 0.00011188118811881189,
"loss": 0.9293,
"step": 11370
},
{
"epoch": 9.675419689748338,
"grad_norm": 11574.048828125,
"learning_rate": 0.00011089108910891089,
"loss": 0.9269,
"step": 11380
},
{
"epoch": 9.683919735284297,
"grad_norm": 13117.8916015625,
"learning_rate": 0.0001099009900990099,
"loss": 0.928,
"step": 11390
},
{
"epoch": 9.692419780820254,
"grad_norm": 13132.0810546875,
"learning_rate": 0.00010891089108910891,
"loss": 0.9272,
"step": 11400
},
{
"epoch": 9.700919826356213,
"grad_norm": 13526.349609375,
"learning_rate": 0.00010792079207920792,
"loss": 0.9271,
"step": 11410
},
{
"epoch": 9.709419871892171,
"grad_norm": 13662.8740234375,
"learning_rate": 0.00010693069306930694,
"loss": 0.9261,
"step": 11420
},
{
"epoch": 9.717919917428128,
"grad_norm": 12285.236328125,
"learning_rate": 0.00010594059405940593,
"loss": 0.9271,
"step": 11430
},
{
"epoch": 9.717919917428128,
"eval_accuracy": 0.8004033551877789,
"eval_loss": 0.8413528800010681,
"eval_runtime": 1034.3691,
"eval_samples_per_second": 482.785,
"eval_steps_per_second": 0.822,
"step": 11430
},
{
"epoch": 9.726419962964087,
"grad_norm": 12984.609375,
"learning_rate": 0.00010495049504950495,
"loss": 0.9272,
"step": 11440
},
{
"epoch": 9.734920008500046,
"grad_norm": 13035.4326171875,
"learning_rate": 0.00010396039603960396,
"loss": 0.927,
"step": 11450
},
{
"epoch": 9.743420054036005,
"grad_norm": 12569.7998046875,
"learning_rate": 0.00010297029702970297,
"loss": 0.9276,
"step": 11460
},
{
"epoch": 9.751920099571961,
"grad_norm": 13077.8876953125,
"learning_rate": 0.00010198019801980198,
"loss": 0.9271,
"step": 11470
},
{
"epoch": 9.76042014510792,
"grad_norm": 12659.830078125,
"learning_rate": 0.00010099009900990099,
"loss": 0.9276,
"step": 11480
},
{
"epoch": 9.768920190643879,
"grad_norm": 13212.763671875,
"learning_rate": 0.0001,
"loss": 0.9268,
"step": 11490
},
{
"epoch": 9.777420236179836,
"grad_norm": 12120.390625,
"learning_rate": 9.900990099009902e-05,
"loss": 0.9267,
"step": 11500
},
{
"epoch": 9.785920281715795,
"grad_norm": 12504.2646484375,
"learning_rate": 9.801980198019803e-05,
"loss": 0.9281,
"step": 11510
},
{
"epoch": 9.794420327251753,
"grad_norm": 12107.5166015625,
"learning_rate": 9.702970297029704e-05,
"loss": 0.9264,
"step": 11520
},
{
"epoch": 9.794420327251753,
"eval_accuracy": 0.800677157520577,
"eval_loss": 0.840539276599884,
"eval_runtime": 1035.1246,
"eval_samples_per_second": 482.433,
"eval_steps_per_second": 0.821,
"step": 11520
},
{
"epoch": 9.802920372787712,
"grad_norm": 13302.0888671875,
"learning_rate": 9.603960396039604e-05,
"loss": 0.9249,
"step": 11530
},
{
"epoch": 9.811420418323669,
"grad_norm": 11464.931640625,
"learning_rate": 9.504950495049505e-05,
"loss": 0.928,
"step": 11540
},
{
"epoch": 9.819920463859628,
"grad_norm": 11620.611328125,
"learning_rate": 9.405940594059406e-05,
"loss": 0.9269,
"step": 11550
},
{
"epoch": 9.828420509395587,
"grad_norm": 12684.673828125,
"learning_rate": 9.306930693069307e-05,
"loss": 0.9264,
"step": 11560
},
{
"epoch": 9.836920554931543,
"grad_norm": 12113.302734375,
"learning_rate": 9.207920792079209e-05,
"loss": 0.9239,
"step": 11570
},
{
"epoch": 9.845420600467502,
"grad_norm": 12095.48828125,
"learning_rate": 9.108910891089108e-05,
"loss": 0.9262,
"step": 11580
},
{
"epoch": 9.853920646003461,
"grad_norm": 12388.09765625,
"learning_rate": 9.00990099009901e-05,
"loss": 0.9241,
"step": 11590
},
{
"epoch": 9.86242069153942,
"grad_norm": 12502.1171875,
"learning_rate": 8.91089108910891e-05,
"loss": 0.9256,
"step": 11600
},
{
"epoch": 9.870920737075377,
"grad_norm": 12518.240234375,
"learning_rate": 8.811881188118812e-05,
"loss": 0.9268,
"step": 11610
},
{
"epoch": 9.870920737075377,
"eval_accuracy": 0.800816669053509,
"eval_loss": 0.8393772840499878,
"eval_runtime": 1035.7736,
"eval_samples_per_second": 482.13,
"eval_steps_per_second": 0.821,
"step": 11610
},
{
"epoch": 9.879420782611335,
"grad_norm": 11850.7548828125,
"learning_rate": 8.712871287128713e-05,
"loss": 0.9257,
"step": 11620
},
{
"epoch": 9.887920828147294,
"grad_norm": 12279.2197265625,
"learning_rate": 8.613861386138613e-05,
"loss": 0.9251,
"step": 11630
},
{
"epoch": 9.896420873683251,
"grad_norm": 12372.4765625,
"learning_rate": 8.514851485148515e-05,
"loss": 0.9241,
"step": 11640
},
{
"epoch": 9.90492091921921,
"grad_norm": 11712.88671875,
"learning_rate": 8.415841584158417e-05,
"loss": 0.9233,
"step": 11650
},
{
"epoch": 9.913420964755169,
"grad_norm": 12502.9453125,
"learning_rate": 8.316831683168318e-05,
"loss": 0.9252,
"step": 11660
},
{
"epoch": 9.921921010291127,
"grad_norm": 13177.66796875,
"learning_rate": 8.217821782178219e-05,
"loss": 0.9237,
"step": 11670
},
{
"epoch": 9.930421055827084,
"grad_norm": 12558.2802734375,
"learning_rate": 8.118811881188119e-05,
"loss": 0.9251,
"step": 11680
},
{
"epoch": 9.938921101363043,
"grad_norm": 11745.330078125,
"learning_rate": 8.01980198019802e-05,
"loss": 0.9264,
"step": 11690
},
{
"epoch": 9.947421146899002,
"grad_norm": 11350.08203125,
"learning_rate": 7.920792079207921e-05,
"loss": 0.924,
"step": 11700
},
{
"epoch": 9.947421146899002,
"eval_accuracy": 0.8009528506484064,
"eval_loss": 0.8385128974914551,
"eval_runtime": 1035.7833,
"eval_samples_per_second": 482.126,
"eval_steps_per_second": 0.821,
"step": 11700
},
{
"epoch": 9.955921192434959,
"grad_norm": 11514.369140625,
"learning_rate": 7.821782178217822e-05,
"loss": 0.9236,
"step": 11710
},
{
"epoch": 9.964421237970917,
"grad_norm": 11468.5146484375,
"learning_rate": 7.722772277227723e-05,
"loss": 0.9226,
"step": 11720
},
{
"epoch": 9.972921283506876,
"grad_norm": 11347.8369140625,
"learning_rate": 7.623762376237623e-05,
"loss": 0.9244,
"step": 11730
},
{
"epoch": 9.981421329042835,
"grad_norm": 11171.818359375,
"learning_rate": 7.524752475247524e-05,
"loss": 0.9247,
"step": 11740
},
{
"epoch": 9.989921374578792,
"grad_norm": 11165.7451171875,
"learning_rate": 7.425742574257426e-05,
"loss": 0.926,
"step": 11750
},
{
"epoch": 9.99842142011475,
"grad_norm": 11604.375,
"learning_rate": 7.326732673267327e-05,
"loss": 0.9237,
"step": 11760
},
{
"epoch": 10.00692146565071,
"grad_norm": 11761.7099609375,
"learning_rate": 7.227722772277228e-05,
"loss": 0.9245,
"step": 11770
},
{
"epoch": 10.015421511186666,
"grad_norm": 11363.318359375,
"learning_rate": 7.128712871287128e-05,
"loss": 0.9242,
"step": 11780
},
{
"epoch": 10.023921556722625,
"grad_norm": 11492.59765625,
"learning_rate": 7.02970297029703e-05,
"loss": 0.9243,
"step": 11790
},
{
"epoch": 10.023921556722625,
"eval_accuracy": 0.8013337325505296,
"eval_loss": 0.8371462225914001,
"eval_runtime": 1036.3502,
"eval_samples_per_second": 481.862,
"eval_steps_per_second": 0.82,
"step": 11790
},
{
"epoch": 10.032421602258584,
"grad_norm": 11367.05859375,
"learning_rate": 6.930693069306931e-05,
"loss": 0.9241,
"step": 11800
},
{
"epoch": 10.040921647794542,
"grad_norm": 11473.013671875,
"learning_rate": 6.831683168316833e-05,
"loss": 0.9223,
"step": 11810
},
{
"epoch": 10.0494216933305,
"grad_norm": 11349.96484375,
"learning_rate": 6.732673267326734e-05,
"loss": 0.9225,
"step": 11820
},
{
"epoch": 10.057921738866458,
"grad_norm": 11236.7431640625,
"learning_rate": 6.633663366336634e-05,
"loss": 0.9236,
"step": 11830
},
{
"epoch": 10.066421784402417,
"grad_norm": 11882.5078125,
"learning_rate": 6.534653465346535e-05,
"loss": 0.9227,
"step": 11840
},
{
"epoch": 10.074921829938374,
"grad_norm": 11027.5966796875,
"learning_rate": 6.435643564356436e-05,
"loss": 0.9212,
"step": 11850
},
{
"epoch": 10.083421875474333,
"grad_norm": 11009.87890625,
"learning_rate": 6.336633663366337e-05,
"loss": 0.9235,
"step": 11860
},
{
"epoch": 10.091921921010291,
"grad_norm": 10943.912109375,
"learning_rate": 6.237623762376238e-05,
"loss": 0.9236,
"step": 11870
},
{
"epoch": 10.10042196654625,
"grad_norm": 11583.521484375,
"learning_rate": 6.138613861386138e-05,
"loss": 0.9251,
"step": 11880
},
{
"epoch": 10.10042196654625,
"eval_accuracy": 0.8014076660668283,
"eval_loss": 0.8366426825523376,
"eval_runtime": 1035.5282,
"eval_samples_per_second": 482.245,
"eval_steps_per_second": 0.821,
"step": 11880
},
{
"epoch": 10.108922012082207,
"grad_norm": 10507.802734375,
"learning_rate": 6.039603960396039e-05,
"loss": 0.9218,
"step": 11890
},
{
"epoch": 10.117422057618166,
"grad_norm": 12043.7763671875,
"learning_rate": 5.9405940594059404e-05,
"loss": 0.9213,
"step": 11900
},
{
"epoch": 10.125922103154124,
"grad_norm": 10901.1728515625,
"learning_rate": 5.841584158415842e-05,
"loss": 0.92,
"step": 11910
},
{
"epoch": 10.134422148690081,
"grad_norm": 11496.796875,
"learning_rate": 5.742574257425743e-05,
"loss": 0.922,
"step": 11920
},
{
"epoch": 10.14292219422604,
"grad_norm": 11207.1455078125,
"learning_rate": 5.643564356435644e-05,
"loss": 0.9207,
"step": 11930
},
{
"epoch": 10.151422239761999,
"grad_norm": 10568.88671875,
"learning_rate": 5.5445544554455445e-05,
"loss": 0.923,
"step": 11940
},
{
"epoch": 10.159922285297958,
"grad_norm": 11236.5009765625,
"learning_rate": 5.4455445544554456e-05,
"loss": 0.9226,
"step": 11950
},
{
"epoch": 10.168422330833915,
"grad_norm": 10793.302734375,
"learning_rate": 5.346534653465347e-05,
"loss": 0.921,
"step": 11960
},
{
"epoch": 10.176922376369873,
"grad_norm": 10562.205078125,
"learning_rate": 5.247524752475247e-05,
"loss": 0.9211,
"step": 11970
},
{
"epoch": 10.176922376369873,
"eval_accuracy": 0.8014045174418934,
"eval_loss": 0.836881697177887,
"eval_runtime": 1035.287,
"eval_samples_per_second": 482.357,
"eval_steps_per_second": 0.821,
"step": 11970
},
{
"epoch": 10.185422421905832,
"grad_norm": 10615.4091796875,
"learning_rate": 5.1485148514851485e-05,
"loss": 0.9216,
"step": 11980
},
{
"epoch": 10.19392246744179,
"grad_norm": 11039.455078125,
"learning_rate": 5.0495049504950497e-05,
"loss": 0.9193,
"step": 11990
},
{
"epoch": 10.202422512977748,
"grad_norm": 11875.810546875,
"learning_rate": 4.950495049504951e-05,
"loss": 0.9221,
"step": 12000
},
{
"epoch": 10.210922558513706,
"grad_norm": 10764.8984375,
"learning_rate": 4.851485148514852e-05,
"loss": 0.922,
"step": 12010
},
{
"epoch": 10.219422604049665,
"grad_norm": 10610.0224609375,
"learning_rate": 4.7524752475247525e-05,
"loss": 0.9225,
"step": 12020
},
{
"epoch": 10.227922649585622,
"grad_norm": 10679.6396484375,
"learning_rate": 4.653465346534654e-05,
"loss": 0.921,
"step": 12030
},
{
"epoch": 10.23642269512158,
"grad_norm": 10098.7451171875,
"learning_rate": 4.554455445544554e-05,
"loss": 0.9227,
"step": 12040
},
{
"epoch": 10.24492274065754,
"grad_norm": 11032.216796875,
"learning_rate": 4.455445544554455e-05,
"loss": 0.9223,
"step": 12050
},
{
"epoch": 10.253422786193497,
"grad_norm": 11303.1201171875,
"learning_rate": 4.3564356435643565e-05,
"loss": 0.9223,
"step": 12060
},
{
"epoch": 10.253422786193497,
"eval_accuracy": 0.8015698536327949,
"eval_loss": 0.8357640504837036,
"eval_runtime": 1032.412,
"eval_samples_per_second": 483.7,
"eval_steps_per_second": 0.823,
"step": 12060
},
{
"epoch": 10.261922831729455,
"grad_norm": 10582.2373046875,
"learning_rate": 4.257425742574258e-05,
"loss": 0.9192,
"step": 12070
},
{
"epoch": 10.270422877265414,
"grad_norm": 10299.9345703125,
"learning_rate": 4.158415841584159e-05,
"loss": 0.9209,
"step": 12080
},
{
"epoch": 10.278922922801373,
"grad_norm": 10322.740234375,
"learning_rate": 4.0594059405940594e-05,
"loss": 0.922,
"step": 12090
},
{
"epoch": 10.28742296833733,
"grad_norm": 10159.0244140625,
"learning_rate": 3.9603960396039605e-05,
"loss": 0.9189,
"step": 12100
},
{
"epoch": 10.295923013873288,
"grad_norm": 10132.728515625,
"learning_rate": 3.861386138613862e-05,
"loss": 0.9209,
"step": 12110
},
{
"epoch": 10.304423059409247,
"grad_norm": 10718.8837890625,
"learning_rate": 3.762376237623762e-05,
"loss": 0.9208,
"step": 12120
},
{
"epoch": 10.312923104945206,
"grad_norm": 10195.5869140625,
"learning_rate": 3.6633663366336634e-05,
"loss": 0.9225,
"step": 12130
},
{
"epoch": 10.321423150481163,
"grad_norm": 10577.4375,
"learning_rate": 3.564356435643564e-05,
"loss": 0.9204,
"step": 12140
},
{
"epoch": 10.329923196017122,
"grad_norm": 10045.123046875,
"learning_rate": 3.465346534653466e-05,
"loss": 0.9182,
"step": 12150
},
{
"epoch": 10.329923196017122,
"eval_accuracy": 0.8016586056285943,
"eval_loss": 0.835513710975647,
"eval_runtime": 1039.9403,
"eval_samples_per_second": 480.199,
"eval_steps_per_second": 0.817,
"step": 12150
},
{
"epoch": 10.33842324155308,
"grad_norm": 10554.9501953125,
"learning_rate": 3.366336633663367e-05,
"loss": 0.9219,
"step": 12160
},
{
"epoch": 10.346923287089037,
"grad_norm": 10395.685546875,
"learning_rate": 3.2673267326732674e-05,
"loss": 0.9189,
"step": 12170
},
{
"epoch": 10.355423332624996,
"grad_norm": 10223.9580078125,
"learning_rate": 3.1683168316831686e-05,
"loss": 0.9205,
"step": 12180
},
{
"epoch": 10.363923378160955,
"grad_norm": 10570.7958984375,
"learning_rate": 3.069306930693069e-05,
"loss": 0.9207,
"step": 12190
},
{
"epoch": 10.372423423696914,
"grad_norm": 9892.7470703125,
"learning_rate": 2.9702970297029702e-05,
"loss": 0.918,
"step": 12200
},
{
"epoch": 10.38092346923287,
"grad_norm": 9921.88671875,
"learning_rate": 2.8712871287128714e-05,
"loss": 0.9196,
"step": 12210
},
{
"epoch": 10.38942351476883,
"grad_norm": 9997.732421875,
"learning_rate": 2.7722772277227722e-05,
"loss": 0.9217,
"step": 12220
},
{
"epoch": 10.397923560304788,
"grad_norm": 10292.541015625,
"learning_rate": 2.6732673267326734e-05,
"loss": 0.9195,
"step": 12230
},
{
"epoch": 10.406423605840745,
"grad_norm": 10017.322265625,
"learning_rate": 2.5742574257425742e-05,
"loss": 0.9215,
"step": 12240
},
{
"epoch": 10.406423605840745,
"eval_accuracy": 0.8019068191825514,
"eval_loss": 0.83404940366745,
"eval_runtime": 1036.3785,
"eval_samples_per_second": 481.849,
"eval_steps_per_second": 0.82,
"step": 12240
},
{
"epoch": 10.414923651376704,
"grad_norm": 10193.8349609375,
"learning_rate": 2.4752475247524754e-05,
"loss": 0.9206,
"step": 12250
},
{
"epoch": 10.423423696912662,
"grad_norm": 10791.517578125,
"learning_rate": 2.3762376237623762e-05,
"loss": 0.9192,
"step": 12260
},
{
"epoch": 10.431923742448621,
"grad_norm": 9776.1796875,
"learning_rate": 2.277227722772277e-05,
"loss": 0.9183,
"step": 12270
},
{
"epoch": 10.440423787984578,
"grad_norm": 9862.8271484375,
"learning_rate": 2.1782178217821783e-05,
"loss": 0.9194,
"step": 12280
},
{
"epoch": 10.448923833520537,
"grad_norm": 10056.2607421875,
"learning_rate": 2.0792079207920794e-05,
"loss": 0.9207,
"step": 12290
},
{
"epoch": 10.457423879056496,
"grad_norm": 9770.7578125,
"learning_rate": 1.9801980198019803e-05,
"loss": 0.9198,
"step": 12300
},
{
"epoch": 10.465923924592452,
"grad_norm": 10069.5634765625,
"learning_rate": 1.881188118811881e-05,
"loss": 0.9201,
"step": 12310
},
{
"epoch": 10.474423970128411,
"grad_norm": 9465.1181640625,
"learning_rate": 1.782178217821782e-05,
"loss": 0.9197,
"step": 12320
},
{
"epoch": 10.48292401566437,
"grad_norm": 9807.8330078125,
"learning_rate": 1.6831683168316834e-05,
"loss": 0.9179,
"step": 12330
},
{
"epoch": 10.48292401566437,
"eval_accuracy": 0.8019292754647317,
"eval_loss": 0.8340857625007629,
"eval_runtime": 1037.0091,
"eval_samples_per_second": 481.556,
"eval_steps_per_second": 0.82,
"step": 12330
},
{
"epoch": 10.491424061200329,
"grad_norm": 9650.10546875,
"learning_rate": 1.5841584158415843e-05,
"loss": 0.9205,
"step": 12340
},
{
"epoch": 10.499924106736286,
"grad_norm": 9682.900390625,
"learning_rate": 1.4851485148514851e-05,
"loss": 0.9195,
"step": 12350
},
{
"epoch": 10.508424152272244,
"grad_norm": 9976.9375,
"learning_rate": 1.3861386138613861e-05,
"loss": 0.9191,
"step": 12360
},
{
"epoch": 10.516924197808203,
"grad_norm": 9428.4052734375,
"learning_rate": 1.2871287128712871e-05,
"loss": 0.9189,
"step": 12370
},
{
"epoch": 10.52542424334416,
"grad_norm": 10025.7158203125,
"learning_rate": 1.1881188118811881e-05,
"loss": 0.918,
"step": 12380
},
{
"epoch": 10.533924288880119,
"grad_norm": 9750.935546875,
"learning_rate": 1.0891089108910891e-05,
"loss": 0.9187,
"step": 12390
},
{
"epoch": 10.542424334416078,
"grad_norm": 9010.52734375,
"learning_rate": 9.900990099009901e-06,
"loss": 0.9189,
"step": 12400
},
{
"epoch": 10.550924379952036,
"grad_norm": 9163.4169921875,
"learning_rate": 8.91089108910891e-06,
"loss": 0.9191,
"step": 12410
},
{
"epoch": 10.559424425487993,
"grad_norm": 9015.5771484375,
"learning_rate": 7.920792079207921e-06,
"loss": 0.9195,
"step": 12420
},
{
"epoch": 10.559424425487993,
"eval_accuracy": 0.802042445850802,
"eval_loss": 0.833830714225769,
"eval_runtime": 1033.7642,
"eval_samples_per_second": 483.068,
"eval_steps_per_second": 0.822,
"step": 12420
}
],
"logging_steps": 10,
"max_steps": 12500,
"num_input_tokens_seen": 0,
"num_train_epochs": 11,
"save_steps": 90,
"total_flos": 7.8411669992781e+18,
"train_batch_size": 288,
"trial_name": null,
"trial_params": null
}