task-13-google-gemma-2b / trainer_state.json
ManyingZ's picture
Upload folder using huggingface_hub
5f8ba07 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 148.87126296522268,
"eval_steps": 500,
"global_step": 30500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.09762050030506407,
"grad_norm": 0.8231198191642761,
"learning_rate": 4e-05,
"loss": 2.624,
"step": 20
},
{
"epoch": 0.19524100061012814,
"grad_norm": 1.040634274482727,
"learning_rate": 8e-05,
"loss": 2.5599,
"step": 40
},
{
"epoch": 0.2928615009151922,
"grad_norm": 1.0133676528930664,
"learning_rate": 0.00012,
"loss": 2.1985,
"step": 60
},
{
"epoch": 0.3904820012202563,
"grad_norm": 1.0250593423843384,
"learning_rate": 0.00016,
"loss": 1.9481,
"step": 80
},
{
"epoch": 0.4881025015253203,
"grad_norm": 0.9968363046646118,
"learning_rate": 0.0002,
"loss": 1.7738,
"step": 100
},
{
"epoch": 0.5857230018303844,
"grad_norm": 1.4279309511184692,
"learning_rate": 0.0001998688524590164,
"loss": 1.7329,
"step": 120
},
{
"epoch": 0.6833435021354485,
"grad_norm": 1.3037158250808716,
"learning_rate": 0.0001997377049180328,
"loss": 1.6784,
"step": 140
},
{
"epoch": 0.7809640024405126,
"grad_norm": 1.404618263244629,
"learning_rate": 0.00019960655737704918,
"loss": 1.6264,
"step": 160
},
{
"epoch": 0.8785845027455765,
"grad_norm": 1.058852195739746,
"learning_rate": 0.0001994754098360656,
"loss": 1.6077,
"step": 180
},
{
"epoch": 0.9762050030506406,
"grad_norm": 1.8023818731307983,
"learning_rate": 0.00019934426229508198,
"loss": 1.5252,
"step": 200
},
{
"epoch": 1.0738255033557047,
"grad_norm": 1.2388545274734497,
"learning_rate": 0.00019921311475409837,
"loss": 1.5621,
"step": 220
},
{
"epoch": 1.1714460036607688,
"grad_norm": 1.5438932180404663,
"learning_rate": 0.00019908196721311476,
"loss": 1.4877,
"step": 240
},
{
"epoch": 1.2690665039658329,
"grad_norm": 1.590529203414917,
"learning_rate": 0.00019895081967213115,
"loss": 1.5568,
"step": 260
},
{
"epoch": 1.366687004270897,
"grad_norm": 1.2864124774932861,
"learning_rate": 0.00019881967213114757,
"loss": 1.5082,
"step": 280
},
{
"epoch": 1.4643075045759608,
"grad_norm": 1.4031420946121216,
"learning_rate": 0.00019868852459016393,
"loss": 1.467,
"step": 300
},
{
"epoch": 1.561928004881025,
"grad_norm": 1.5916552543640137,
"learning_rate": 0.00019855737704918035,
"loss": 1.4055,
"step": 320
},
{
"epoch": 1.659548505186089,
"grad_norm": 1.5989781618118286,
"learning_rate": 0.00019842622950819674,
"loss": 1.358,
"step": 340
},
{
"epoch": 1.757169005491153,
"grad_norm": 1.7474373579025269,
"learning_rate": 0.00019829508196721313,
"loss": 1.3905,
"step": 360
},
{
"epoch": 1.8547895057962172,
"grad_norm": 1.8275643587112427,
"learning_rate": 0.00019816393442622951,
"loss": 1.4449,
"step": 380
},
{
"epoch": 1.9524100061012812,
"grad_norm": 1.7919280529022217,
"learning_rate": 0.0001980327868852459,
"loss": 1.4057,
"step": 400
},
{
"epoch": 2.0500305064063453,
"grad_norm": 2.0150179862976074,
"learning_rate": 0.00019790163934426232,
"loss": 1.3659,
"step": 420
},
{
"epoch": 2.1476510067114094,
"grad_norm": 2.0520079135894775,
"learning_rate": 0.00019777049180327868,
"loss": 1.3735,
"step": 440
},
{
"epoch": 2.2452715070164735,
"grad_norm": 2.1597766876220703,
"learning_rate": 0.0001976393442622951,
"loss": 1.318,
"step": 460
},
{
"epoch": 2.3428920073215376,
"grad_norm": 1.8374860286712646,
"learning_rate": 0.0001975081967213115,
"loss": 1.3469,
"step": 480
},
{
"epoch": 2.4405125076266017,
"grad_norm": 2.2065134048461914,
"learning_rate": 0.00019737704918032788,
"loss": 1.3822,
"step": 500
},
{
"epoch": 2.5381330079316657,
"grad_norm": 2.5370113849639893,
"learning_rate": 0.00019724590163934427,
"loss": 1.2652,
"step": 520
},
{
"epoch": 2.63575350823673,
"grad_norm": 2.0484936237335205,
"learning_rate": 0.00019711475409836066,
"loss": 1.2971,
"step": 540
},
{
"epoch": 2.733374008541794,
"grad_norm": 1.699357509613037,
"learning_rate": 0.00019698360655737707,
"loss": 1.3195,
"step": 560
},
{
"epoch": 2.830994508846858,
"grad_norm": 2.4636785984039307,
"learning_rate": 0.00019685245901639344,
"loss": 1.2474,
"step": 580
},
{
"epoch": 2.9286150091519216,
"grad_norm": 3.6711232662200928,
"learning_rate": 0.00019672131147540985,
"loss": 1.3091,
"step": 600
},
{
"epoch": 3.026235509456986,
"grad_norm": 1.9547876119613647,
"learning_rate": 0.00019659016393442624,
"loss": 1.2966,
"step": 620
},
{
"epoch": 3.1238560097620502,
"grad_norm": 2.135209560394287,
"learning_rate": 0.00019645901639344263,
"loss": 1.2201,
"step": 640
},
{
"epoch": 3.221476510067114,
"grad_norm": 2.2378487586975098,
"learning_rate": 0.00019632786885245902,
"loss": 1.2226,
"step": 660
},
{
"epoch": 3.319097010372178,
"grad_norm": 2.5407207012176514,
"learning_rate": 0.0001961967213114754,
"loss": 1.224,
"step": 680
},
{
"epoch": 3.416717510677242,
"grad_norm": 2.5137102603912354,
"learning_rate": 0.00019606557377049183,
"loss": 1.271,
"step": 700
},
{
"epoch": 3.514338010982306,
"grad_norm": 2.5121419429779053,
"learning_rate": 0.00019593442622950822,
"loss": 1.2469,
"step": 720
},
{
"epoch": 3.61195851128737,
"grad_norm": 2.5250003337860107,
"learning_rate": 0.0001958032786885246,
"loss": 1.1799,
"step": 740
},
{
"epoch": 3.7095790115924343,
"grad_norm": 1.8390239477157593,
"learning_rate": 0.000195672131147541,
"loss": 1.2629,
"step": 760
},
{
"epoch": 3.8071995118974984,
"grad_norm": 2.065147638320923,
"learning_rate": 0.00019554098360655738,
"loss": 1.2158,
"step": 780
},
{
"epoch": 3.9048200122025625,
"grad_norm": 2.7138381004333496,
"learning_rate": 0.00019540983606557377,
"loss": 1.1778,
"step": 800
},
{
"epoch": 4.002440512507627,
"grad_norm": 3.124082565307617,
"learning_rate": 0.00019527868852459016,
"loss": 1.2258,
"step": 820
},
{
"epoch": 4.100061012812691,
"grad_norm": 2.8344528675079346,
"learning_rate": 0.00019514754098360658,
"loss": 1.1718,
"step": 840
},
{
"epoch": 4.197681513117755,
"grad_norm": 2.2852656841278076,
"learning_rate": 0.00019501639344262297,
"loss": 1.132,
"step": 860
},
{
"epoch": 4.295302013422819,
"grad_norm": 2.662973642349243,
"learning_rate": 0.00019488524590163936,
"loss": 1.158,
"step": 880
},
{
"epoch": 4.392922513727883,
"grad_norm": 2.1522879600524902,
"learning_rate": 0.00019475409836065575,
"loss": 1.1263,
"step": 900
},
{
"epoch": 4.490543014032947,
"grad_norm": 2.541273832321167,
"learning_rate": 0.00019462295081967214,
"loss": 1.1807,
"step": 920
},
{
"epoch": 4.588163514338011,
"grad_norm": 2.0520172119140625,
"learning_rate": 0.00019449180327868855,
"loss": 1.1164,
"step": 940
},
{
"epoch": 4.685784014643075,
"grad_norm": 2.2306621074676514,
"learning_rate": 0.00019436065573770491,
"loss": 1.117,
"step": 960
},
{
"epoch": 4.783404514948139,
"grad_norm": 3.0042741298675537,
"learning_rate": 0.00019422950819672133,
"loss": 1.2039,
"step": 980
},
{
"epoch": 4.881025015253203,
"grad_norm": 2.1427409648895264,
"learning_rate": 0.00019409836065573772,
"loss": 1.1225,
"step": 1000
},
{
"epoch": 4.978645515558267,
"grad_norm": 3.128009557723999,
"learning_rate": 0.0001939672131147541,
"loss": 1.1619,
"step": 1020
},
{
"epoch": 5.0762660158633315,
"grad_norm": 3.7390689849853516,
"learning_rate": 0.0001938360655737705,
"loss": 1.1612,
"step": 1040
},
{
"epoch": 5.173886516168396,
"grad_norm": 2.554919958114624,
"learning_rate": 0.0001937049180327869,
"loss": 1.1003,
"step": 1060
},
{
"epoch": 5.27150701647346,
"grad_norm": 3.6076526641845703,
"learning_rate": 0.0001935737704918033,
"loss": 1.105,
"step": 1080
},
{
"epoch": 5.369127516778524,
"grad_norm": 3.665255069732666,
"learning_rate": 0.00019344262295081967,
"loss": 1.039,
"step": 1100
},
{
"epoch": 5.466748017083588,
"grad_norm": 2.7409827709198,
"learning_rate": 0.00019331147540983608,
"loss": 1.0942,
"step": 1120
},
{
"epoch": 5.564368517388652,
"grad_norm": 2.087810754776001,
"learning_rate": 0.00019318032786885247,
"loss": 1.0797,
"step": 1140
},
{
"epoch": 5.661989017693716,
"grad_norm": 2.9384937286376953,
"learning_rate": 0.00019304918032786886,
"loss": 1.1056,
"step": 1160
},
{
"epoch": 5.75960951799878,
"grad_norm": 2.7793357372283936,
"learning_rate": 0.00019291803278688525,
"loss": 1.1048,
"step": 1180
},
{
"epoch": 5.857230018303844,
"grad_norm": 2.1878364086151123,
"learning_rate": 0.00019278688524590164,
"loss": 1.1033,
"step": 1200
},
{
"epoch": 5.954850518608908,
"grad_norm": 2.3006527423858643,
"learning_rate": 0.00019265573770491806,
"loss": 1.0708,
"step": 1220
},
{
"epoch": 6.052471018913972,
"grad_norm": 2.899059295654297,
"learning_rate": 0.00019252459016393442,
"loss": 1.0686,
"step": 1240
},
{
"epoch": 6.150091519219036,
"grad_norm": 2.8099234104156494,
"learning_rate": 0.00019239344262295084,
"loss": 0.9753,
"step": 1260
},
{
"epoch": 6.2477120195241005,
"grad_norm": 3.555896282196045,
"learning_rate": 0.00019226229508196723,
"loss": 1.0251,
"step": 1280
},
{
"epoch": 6.345332519829164,
"grad_norm": 2.860172748565674,
"learning_rate": 0.00019213114754098362,
"loss": 1.0532,
"step": 1300
},
{
"epoch": 6.442953020134228,
"grad_norm": 3.038743734359741,
"learning_rate": 0.000192,
"loss": 1.0339,
"step": 1320
},
{
"epoch": 6.540573520439292,
"grad_norm": 3.338714838027954,
"learning_rate": 0.0001918688524590164,
"loss": 1.0632,
"step": 1340
},
{
"epoch": 6.638194020744356,
"grad_norm": 2.932199478149414,
"learning_rate": 0.0001917377049180328,
"loss": 1.1157,
"step": 1360
},
{
"epoch": 6.73581452104942,
"grad_norm": 3.783275604248047,
"learning_rate": 0.00019160655737704917,
"loss": 1.0041,
"step": 1380
},
{
"epoch": 6.833435021354484,
"grad_norm": 3.5673303604125977,
"learning_rate": 0.0001914754098360656,
"loss": 1.013,
"step": 1400
},
{
"epoch": 6.931055521659548,
"grad_norm": 2.917269468307495,
"learning_rate": 0.00019134426229508198,
"loss": 1.1148,
"step": 1420
},
{
"epoch": 7.028676021964612,
"grad_norm": 3.269557237625122,
"learning_rate": 0.00019121311475409837,
"loss": 0.9701,
"step": 1440
},
{
"epoch": 7.126296522269676,
"grad_norm": 3.337118625640869,
"learning_rate": 0.00019108196721311476,
"loss": 0.9921,
"step": 1460
},
{
"epoch": 7.22391702257474,
"grad_norm": 3.016709089279175,
"learning_rate": 0.00019095081967213115,
"loss": 0.9514,
"step": 1480
},
{
"epoch": 7.3215375228798045,
"grad_norm": 2.8816025257110596,
"learning_rate": 0.00019081967213114756,
"loss": 0.9545,
"step": 1500
},
{
"epoch": 7.419158023184869,
"grad_norm": 2.9574875831604004,
"learning_rate": 0.00019068852459016395,
"loss": 1.0192,
"step": 1520
},
{
"epoch": 7.516778523489933,
"grad_norm": 2.7323741912841797,
"learning_rate": 0.00019055737704918034,
"loss": 0.9998,
"step": 1540
},
{
"epoch": 7.614399023794997,
"grad_norm": 3.1870975494384766,
"learning_rate": 0.00019042622950819673,
"loss": 0.983,
"step": 1560
},
{
"epoch": 7.712019524100061,
"grad_norm": 2.5255308151245117,
"learning_rate": 0.00019029508196721312,
"loss": 1.0236,
"step": 1580
},
{
"epoch": 7.809640024405125,
"grad_norm": 3.190338134765625,
"learning_rate": 0.0001901639344262295,
"loss": 1.0263,
"step": 1600
},
{
"epoch": 7.907260524710189,
"grad_norm": 2.7343392372131348,
"learning_rate": 0.0001900327868852459,
"loss": 0.981,
"step": 1620
},
{
"epoch": 8.004881025015253,
"grad_norm": 2.4128050804138184,
"learning_rate": 0.00018990163934426232,
"loss": 1.0471,
"step": 1640
},
{
"epoch": 8.102501525320317,
"grad_norm": 3.3373966217041016,
"learning_rate": 0.0001897704918032787,
"loss": 0.9562,
"step": 1660
},
{
"epoch": 8.200122025625381,
"grad_norm": 4.268326282501221,
"learning_rate": 0.0001896393442622951,
"loss": 0.9365,
"step": 1680
},
{
"epoch": 8.297742525930445,
"grad_norm": 4.0396728515625,
"learning_rate": 0.00018950819672131148,
"loss": 0.9469,
"step": 1700
},
{
"epoch": 8.39536302623551,
"grad_norm": 2.5835509300231934,
"learning_rate": 0.00018937704918032787,
"loss": 0.8998,
"step": 1720
},
{
"epoch": 8.492983526540574,
"grad_norm": 4.017479419708252,
"learning_rate": 0.0001892459016393443,
"loss": 0.9843,
"step": 1740
},
{
"epoch": 8.590604026845638,
"grad_norm": 2.7477169036865234,
"learning_rate": 0.00018911475409836065,
"loss": 0.9164,
"step": 1760
},
{
"epoch": 8.688224527150702,
"grad_norm": 3.388787031173706,
"learning_rate": 0.00018898360655737707,
"loss": 0.9626,
"step": 1780
},
{
"epoch": 8.785845027455766,
"grad_norm": 2.938905954360962,
"learning_rate": 0.00018885245901639346,
"loss": 0.9662,
"step": 1800
},
{
"epoch": 8.88346552776083,
"grad_norm": 2.854912519454956,
"learning_rate": 0.00018872131147540985,
"loss": 1.0023,
"step": 1820
},
{
"epoch": 8.981086028065894,
"grad_norm": 2.120589017868042,
"learning_rate": 0.00018859016393442624,
"loss": 0.9616,
"step": 1840
},
{
"epoch": 9.078706528370958,
"grad_norm": 2.8356106281280518,
"learning_rate": 0.00018845901639344263,
"loss": 0.9589,
"step": 1860
},
{
"epoch": 9.176327028676022,
"grad_norm": 2.9466371536254883,
"learning_rate": 0.00018832786885245904,
"loss": 0.9196,
"step": 1880
},
{
"epoch": 9.273947528981086,
"grad_norm": 3.5278165340423584,
"learning_rate": 0.0001881967213114754,
"loss": 0.8946,
"step": 1900
},
{
"epoch": 9.37156802928615,
"grad_norm": 3.5878446102142334,
"learning_rate": 0.00018806557377049182,
"loss": 0.8855,
"step": 1920
},
{
"epoch": 9.469188529591214,
"grad_norm": 2.8943283557891846,
"learning_rate": 0.0001879344262295082,
"loss": 0.9094,
"step": 1940
},
{
"epoch": 9.566809029896278,
"grad_norm": 2.825993061065674,
"learning_rate": 0.0001878032786885246,
"loss": 0.9591,
"step": 1960
},
{
"epoch": 9.664429530201343,
"grad_norm": 3.7361643314361572,
"learning_rate": 0.000187672131147541,
"loss": 0.9368,
"step": 1980
},
{
"epoch": 9.762050030506407,
"grad_norm": 3.8567728996276855,
"learning_rate": 0.00018754098360655738,
"loss": 0.8843,
"step": 2000
},
{
"epoch": 9.85967053081147,
"grad_norm": 3.694343090057373,
"learning_rate": 0.0001874098360655738,
"loss": 0.9235,
"step": 2020
},
{
"epoch": 9.957291031116535,
"grad_norm": 4.091517448425293,
"learning_rate": 0.00018727868852459016,
"loss": 0.9418,
"step": 2040
},
{
"epoch": 10.054911531421599,
"grad_norm": 3.1028010845184326,
"learning_rate": 0.00018714754098360657,
"loss": 0.8961,
"step": 2060
},
{
"epoch": 10.152532031726663,
"grad_norm": 2.9279282093048096,
"learning_rate": 0.00018701639344262296,
"loss": 0.8457,
"step": 2080
},
{
"epoch": 10.250152532031727,
"grad_norm": 2.750516176223755,
"learning_rate": 0.00018688524590163935,
"loss": 0.8941,
"step": 2100
},
{
"epoch": 10.347773032336791,
"grad_norm": 3.211580276489258,
"learning_rate": 0.00018675409836065574,
"loss": 0.8799,
"step": 2120
},
{
"epoch": 10.445393532641855,
"grad_norm": 3.7746126651763916,
"learning_rate": 0.00018662295081967213,
"loss": 0.8288,
"step": 2140
},
{
"epoch": 10.54301403294692,
"grad_norm": 2.8005576133728027,
"learning_rate": 0.00018649180327868855,
"loss": 0.9162,
"step": 2160
},
{
"epoch": 10.640634533251983,
"grad_norm": 2.9262828826904297,
"learning_rate": 0.00018636065573770494,
"loss": 0.83,
"step": 2180
},
{
"epoch": 10.738255033557047,
"grad_norm": 4.560007095336914,
"learning_rate": 0.00018622950819672133,
"loss": 0.956,
"step": 2200
},
{
"epoch": 10.835875533862112,
"grad_norm": 2.39521861076355,
"learning_rate": 0.00018609836065573772,
"loss": 0.9038,
"step": 2220
},
{
"epoch": 10.933496034167176,
"grad_norm": 2.711963653564453,
"learning_rate": 0.0001859672131147541,
"loss": 0.9322,
"step": 2240
},
{
"epoch": 11.03111653447224,
"grad_norm": 2.9960291385650635,
"learning_rate": 0.0001858360655737705,
"loss": 0.8956,
"step": 2260
},
{
"epoch": 11.128737034777304,
"grad_norm": 3.7117059230804443,
"learning_rate": 0.00018570491803278688,
"loss": 0.8253,
"step": 2280
},
{
"epoch": 11.226357535082368,
"grad_norm": 2.5004775524139404,
"learning_rate": 0.0001855737704918033,
"loss": 0.8075,
"step": 2300
},
{
"epoch": 11.323978035387432,
"grad_norm": 2.690229654312134,
"learning_rate": 0.0001854426229508197,
"loss": 0.8528,
"step": 2320
},
{
"epoch": 11.421598535692496,
"grad_norm": 2.5582730770111084,
"learning_rate": 0.00018531147540983608,
"loss": 0.9235,
"step": 2340
},
{
"epoch": 11.51921903599756,
"grad_norm": 3.388638973236084,
"learning_rate": 0.00018518032786885247,
"loss": 0.8562,
"step": 2360
},
{
"epoch": 11.616839536302624,
"grad_norm": 2.6668691635131836,
"learning_rate": 0.00018504918032786886,
"loss": 0.8447,
"step": 2380
},
{
"epoch": 11.714460036607688,
"grad_norm": 2.5166635513305664,
"learning_rate": 0.00018491803278688527,
"loss": 0.8748,
"step": 2400
},
{
"epoch": 11.812080536912752,
"grad_norm": 2.641568183898926,
"learning_rate": 0.00018478688524590164,
"loss": 0.8401,
"step": 2420
},
{
"epoch": 11.909701037217816,
"grad_norm": 4.179554462432861,
"learning_rate": 0.00018465573770491805,
"loss": 0.8231,
"step": 2440
},
{
"epoch": 12.00732153752288,
"grad_norm": 3.014526844024658,
"learning_rate": 0.00018452459016393444,
"loss": 0.8792,
"step": 2460
},
{
"epoch": 12.104942037827945,
"grad_norm": 4.275758743286133,
"learning_rate": 0.0001843934426229508,
"loss": 0.8161,
"step": 2480
},
{
"epoch": 12.202562538133009,
"grad_norm": 2.949131965637207,
"learning_rate": 0.00018426229508196722,
"loss": 0.8304,
"step": 2500
},
{
"epoch": 12.300183038438073,
"grad_norm": 3.018010377883911,
"learning_rate": 0.0001841311475409836,
"loss": 0.8373,
"step": 2520
},
{
"epoch": 12.397803538743137,
"grad_norm": 4.015832424163818,
"learning_rate": 0.00018400000000000003,
"loss": 0.779,
"step": 2540
},
{
"epoch": 12.495424039048201,
"grad_norm": 3.1773481369018555,
"learning_rate": 0.0001838688524590164,
"loss": 0.8428,
"step": 2560
},
{
"epoch": 12.593044539353265,
"grad_norm": 3.50673508644104,
"learning_rate": 0.0001837377049180328,
"loss": 0.8148,
"step": 2580
},
{
"epoch": 12.690665039658327,
"grad_norm": 2.848912239074707,
"learning_rate": 0.0001836065573770492,
"loss": 0.8427,
"step": 2600
},
{
"epoch": 12.788285539963393,
"grad_norm": 3.488842725753784,
"learning_rate": 0.00018347540983606558,
"loss": 0.8374,
"step": 2620
},
{
"epoch": 12.885906040268456,
"grad_norm": 3.831639528274536,
"learning_rate": 0.00018334426229508197,
"loss": 0.7835,
"step": 2640
},
{
"epoch": 12.98352654057352,
"grad_norm": 3.8596065044403076,
"learning_rate": 0.00018321311475409836,
"loss": 0.8719,
"step": 2660
},
{
"epoch": 13.081147040878584,
"grad_norm": 4.338437557220459,
"learning_rate": 0.00018308196721311478,
"loss": 0.7768,
"step": 2680
},
{
"epoch": 13.178767541183648,
"grad_norm": 4.147676944732666,
"learning_rate": 0.00018295081967213114,
"loss": 0.7686,
"step": 2700
},
{
"epoch": 13.276388041488712,
"grad_norm": 3.430076837539673,
"learning_rate": 0.00018281967213114756,
"loss": 0.7069,
"step": 2720
},
{
"epoch": 13.374008541793776,
"grad_norm": 3.6902527809143066,
"learning_rate": 0.00018268852459016395,
"loss": 0.8241,
"step": 2740
},
{
"epoch": 13.47162904209884,
"grad_norm": 3.362191915512085,
"learning_rate": 0.00018255737704918034,
"loss": 0.8096,
"step": 2760
},
{
"epoch": 13.569249542403904,
"grad_norm": 3.1833972930908203,
"learning_rate": 0.00018242622950819673,
"loss": 0.8258,
"step": 2780
},
{
"epoch": 13.666870042708968,
"grad_norm": 3.001095771789551,
"learning_rate": 0.00018229508196721312,
"loss": 0.7793,
"step": 2800
},
{
"epoch": 13.764490543014032,
"grad_norm": 4.468006134033203,
"learning_rate": 0.00018216393442622953,
"loss": 0.7837,
"step": 2820
},
{
"epoch": 13.862111043319096,
"grad_norm": 3.7999494075775146,
"learning_rate": 0.00018203278688524592,
"loss": 0.8158,
"step": 2840
},
{
"epoch": 13.95973154362416,
"grad_norm": 2.7308099269866943,
"learning_rate": 0.0001819016393442623,
"loss": 0.8657,
"step": 2860
},
{
"epoch": 14.057352043929225,
"grad_norm": 3.631223678588867,
"learning_rate": 0.0001817704918032787,
"loss": 0.8219,
"step": 2880
},
{
"epoch": 14.154972544234289,
"grad_norm": 3.7246315479278564,
"learning_rate": 0.0001816393442622951,
"loss": 0.738,
"step": 2900
},
{
"epoch": 14.252593044539353,
"grad_norm": 3.8941688537597656,
"learning_rate": 0.00018150819672131148,
"loss": 0.7882,
"step": 2920
},
{
"epoch": 14.350213544844417,
"grad_norm": 4.167781352996826,
"learning_rate": 0.00018137704918032787,
"loss": 0.7396,
"step": 2940
},
{
"epoch": 14.44783404514948,
"grad_norm": 3.553039312362671,
"learning_rate": 0.00018124590163934429,
"loss": 0.7614,
"step": 2960
},
{
"epoch": 14.545454545454545,
"grad_norm": 3.1469032764434814,
"learning_rate": 0.00018111475409836067,
"loss": 0.7651,
"step": 2980
},
{
"epoch": 14.643075045759609,
"grad_norm": 5.265510559082031,
"learning_rate": 0.00018098360655737704,
"loss": 0.8169,
"step": 3000
},
{
"epoch": 14.740695546064673,
"grad_norm": 2.910022020339966,
"learning_rate": 0.00018085245901639345,
"loss": 0.7746,
"step": 3020
},
{
"epoch": 14.838316046369737,
"grad_norm": 2.8065860271453857,
"learning_rate": 0.00018072131147540984,
"loss": 0.8028,
"step": 3040
},
{
"epoch": 14.935936546674801,
"grad_norm": 2.44608736038208,
"learning_rate": 0.00018059016393442626,
"loss": 0.7915,
"step": 3060
},
{
"epoch": 15.033557046979865,
"grad_norm": 2.230102300643921,
"learning_rate": 0.00018045901639344262,
"loss": 0.7283,
"step": 3080
},
{
"epoch": 15.13117754728493,
"grad_norm": 3.1589298248291016,
"learning_rate": 0.00018032786885245904,
"loss": 0.7117,
"step": 3100
},
{
"epoch": 15.228798047589994,
"grad_norm": 3.1552984714508057,
"learning_rate": 0.00018019672131147543,
"loss": 0.7047,
"step": 3120
},
{
"epoch": 15.326418547895058,
"grad_norm": 3.4300222396850586,
"learning_rate": 0.0001800655737704918,
"loss": 0.7328,
"step": 3140
},
{
"epoch": 15.424039048200122,
"grad_norm": 2.5847089290618896,
"learning_rate": 0.0001799344262295082,
"loss": 0.7774,
"step": 3160
},
{
"epoch": 15.521659548505186,
"grad_norm": 2.6597721576690674,
"learning_rate": 0.0001798032786885246,
"loss": 0.7427,
"step": 3180
},
{
"epoch": 15.61928004881025,
"grad_norm": 3.029202699661255,
"learning_rate": 0.000179672131147541,
"loss": 0.7037,
"step": 3200
},
{
"epoch": 15.716900549115314,
"grad_norm": 4.356511116027832,
"learning_rate": 0.00017954098360655737,
"loss": 0.7797,
"step": 3220
},
{
"epoch": 15.814521049420378,
"grad_norm": 3.456430435180664,
"learning_rate": 0.0001794098360655738,
"loss": 0.7839,
"step": 3240
},
{
"epoch": 15.912141549725442,
"grad_norm": 2.608675718307495,
"learning_rate": 0.00017927868852459018,
"loss": 0.7987,
"step": 3260
},
{
"epoch": 16.009762050030506,
"grad_norm": 3.393676519393921,
"learning_rate": 0.00017914754098360657,
"loss": 0.7627,
"step": 3280
},
{
"epoch": 16.107382550335572,
"grad_norm": 3.7134811878204346,
"learning_rate": 0.00017901639344262296,
"loss": 0.6441,
"step": 3300
},
{
"epoch": 16.205003050640634,
"grad_norm": 4.507261276245117,
"learning_rate": 0.00017888524590163935,
"loss": 0.6972,
"step": 3320
},
{
"epoch": 16.3026235509457,
"grad_norm": 3.150489091873169,
"learning_rate": 0.00017875409836065576,
"loss": 0.7079,
"step": 3340
},
{
"epoch": 16.400244051250763,
"grad_norm": 3.351140022277832,
"learning_rate": 0.00017862295081967213,
"loss": 0.7319,
"step": 3360
},
{
"epoch": 16.49786455155583,
"grad_norm": 3.533992290496826,
"learning_rate": 0.00017849180327868852,
"loss": 0.7095,
"step": 3380
},
{
"epoch": 16.59548505186089,
"grad_norm": 3.086233139038086,
"learning_rate": 0.00017836065573770493,
"loss": 0.7653,
"step": 3400
},
{
"epoch": 16.693105552165953,
"grad_norm": 2.891784191131592,
"learning_rate": 0.00017822950819672132,
"loss": 0.7096,
"step": 3420
},
{
"epoch": 16.79072605247102,
"grad_norm": 3.5396888256073,
"learning_rate": 0.0001780983606557377,
"loss": 0.7385,
"step": 3440
},
{
"epoch": 16.888346552776085,
"grad_norm": 3.5215485095977783,
"learning_rate": 0.0001779672131147541,
"loss": 0.791,
"step": 3460
},
{
"epoch": 16.985967053081147,
"grad_norm": 3.2734858989715576,
"learning_rate": 0.00017783606557377052,
"loss": 0.7904,
"step": 3480
},
{
"epoch": 17.08358755338621,
"grad_norm": 3.1748645305633545,
"learning_rate": 0.0001777049180327869,
"loss": 0.6706,
"step": 3500
},
{
"epoch": 17.181208053691275,
"grad_norm": 2.6187920570373535,
"learning_rate": 0.00017757377049180327,
"loss": 0.7038,
"step": 3520
},
{
"epoch": 17.278828553996338,
"grad_norm": 2.9449117183685303,
"learning_rate": 0.00017744262295081969,
"loss": 0.7214,
"step": 3540
},
{
"epoch": 17.376449054301403,
"grad_norm": 2.324122667312622,
"learning_rate": 0.00017731147540983607,
"loss": 0.7377,
"step": 3560
},
{
"epoch": 17.474069554606466,
"grad_norm": 3.7478408813476562,
"learning_rate": 0.00017718032786885246,
"loss": 0.6557,
"step": 3580
},
{
"epoch": 17.57169005491153,
"grad_norm": 2.6289916038513184,
"learning_rate": 0.00017704918032786885,
"loss": 0.7385,
"step": 3600
},
{
"epoch": 17.669310555216594,
"grad_norm": 3.6009535789489746,
"learning_rate": 0.00017691803278688527,
"loss": 0.7495,
"step": 3620
},
{
"epoch": 17.76693105552166,
"grad_norm": 4.0030035972595215,
"learning_rate": 0.00017678688524590166,
"loss": 0.6941,
"step": 3640
},
{
"epoch": 17.864551555826722,
"grad_norm": 2.4140219688415527,
"learning_rate": 0.00017665573770491802,
"loss": 0.6882,
"step": 3660
},
{
"epoch": 17.962172056131788,
"grad_norm": 3.9716620445251465,
"learning_rate": 0.00017652459016393444,
"loss": 0.7051,
"step": 3680
},
{
"epoch": 18.05979255643685,
"grad_norm": 3.0810763835906982,
"learning_rate": 0.00017639344262295083,
"loss": 0.7016,
"step": 3700
},
{
"epoch": 18.157413056741916,
"grad_norm": 3.1639134883880615,
"learning_rate": 0.00017626229508196724,
"loss": 0.699,
"step": 3720
},
{
"epoch": 18.25503355704698,
"grad_norm": 3.6788012981414795,
"learning_rate": 0.0001761311475409836,
"loss": 0.6673,
"step": 3740
},
{
"epoch": 18.352654057352044,
"grad_norm": 3.2544620037078857,
"learning_rate": 0.00017600000000000002,
"loss": 0.7313,
"step": 3760
},
{
"epoch": 18.450274557657107,
"grad_norm": 3.900455951690674,
"learning_rate": 0.0001758688524590164,
"loss": 0.7271,
"step": 3780
},
{
"epoch": 18.547895057962172,
"grad_norm": 3.3089077472686768,
"learning_rate": 0.00017573770491803277,
"loss": 0.6452,
"step": 3800
},
{
"epoch": 18.645515558267235,
"grad_norm": 3.921875,
"learning_rate": 0.0001756065573770492,
"loss": 0.6498,
"step": 3820
},
{
"epoch": 18.7431360585723,
"grad_norm": 3.5202882289886475,
"learning_rate": 0.00017547540983606558,
"loss": 0.685,
"step": 3840
},
{
"epoch": 18.840756558877363,
"grad_norm": 4.708593368530273,
"learning_rate": 0.000175344262295082,
"loss": 0.6824,
"step": 3860
},
{
"epoch": 18.93837705918243,
"grad_norm": 3.8522069454193115,
"learning_rate": 0.00017521311475409836,
"loss": 0.6827,
"step": 3880
},
{
"epoch": 19.03599755948749,
"grad_norm": 3.08516788482666,
"learning_rate": 0.00017508196721311475,
"loss": 0.6848,
"step": 3900
},
{
"epoch": 19.133618059792557,
"grad_norm": 6.1903815269470215,
"learning_rate": 0.00017495081967213116,
"loss": 0.5924,
"step": 3920
},
{
"epoch": 19.23123856009762,
"grad_norm": 4.409329891204834,
"learning_rate": 0.00017481967213114753,
"loss": 0.6484,
"step": 3940
},
{
"epoch": 19.328859060402685,
"grad_norm": 3.1954550743103027,
"learning_rate": 0.00017468852459016394,
"loss": 0.6759,
"step": 3960
},
{
"epoch": 19.426479560707747,
"grad_norm": 4.925632953643799,
"learning_rate": 0.00017455737704918033,
"loss": 0.6447,
"step": 3980
},
{
"epoch": 19.524100061012813,
"grad_norm": 3.4807968139648438,
"learning_rate": 0.00017442622950819675,
"loss": 0.6905,
"step": 4000
},
{
"epoch": 19.621720561317876,
"grad_norm": 3.031001567840576,
"learning_rate": 0.0001742950819672131,
"loss": 0.68,
"step": 4020
},
{
"epoch": 19.71934106162294,
"grad_norm": 2.873664140701294,
"learning_rate": 0.0001741639344262295,
"loss": 0.6986,
"step": 4040
},
{
"epoch": 19.816961561928004,
"grad_norm": 3.4913737773895264,
"learning_rate": 0.00017403278688524592,
"loss": 0.7319,
"step": 4060
},
{
"epoch": 19.91458206223307,
"grad_norm": 4.207777976989746,
"learning_rate": 0.0001739016393442623,
"loss": 0.6479,
"step": 4080
},
{
"epoch": 20.012202562538132,
"grad_norm": 2.822380781173706,
"learning_rate": 0.0001737704918032787,
"loss": 0.683,
"step": 4100
},
{
"epoch": 20.109823062843198,
"grad_norm": 3.4512064456939697,
"learning_rate": 0.00017363934426229509,
"loss": 0.6329,
"step": 4120
},
{
"epoch": 20.20744356314826,
"grad_norm": 4.552077293395996,
"learning_rate": 0.0001735081967213115,
"loss": 0.6622,
"step": 4140
},
{
"epoch": 20.305064063453326,
"grad_norm": 3.7285406589508057,
"learning_rate": 0.00017337704918032786,
"loss": 0.6287,
"step": 4160
},
{
"epoch": 20.40268456375839,
"grad_norm": 3.0254034996032715,
"learning_rate": 0.00017324590163934425,
"loss": 0.6696,
"step": 4180
},
{
"epoch": 20.500305064063454,
"grad_norm": 3.474151372909546,
"learning_rate": 0.00017311475409836067,
"loss": 0.623,
"step": 4200
},
{
"epoch": 20.597925564368516,
"grad_norm": 3.686966896057129,
"learning_rate": 0.00017298360655737706,
"loss": 0.7025,
"step": 4220
},
{
"epoch": 20.695546064673582,
"grad_norm": 4.044944763183594,
"learning_rate": 0.00017285245901639345,
"loss": 0.6287,
"step": 4240
},
{
"epoch": 20.793166564978645,
"grad_norm": 4.093704700469971,
"learning_rate": 0.00017272131147540984,
"loss": 0.6523,
"step": 4260
},
{
"epoch": 20.89078706528371,
"grad_norm": 3.9666953086853027,
"learning_rate": 0.00017259016393442625,
"loss": 0.6513,
"step": 4280
},
{
"epoch": 20.988407565588773,
"grad_norm": 3.888939142227173,
"learning_rate": 0.00017245901639344264,
"loss": 0.682,
"step": 4300
},
{
"epoch": 21.08602806589384,
"grad_norm": 3.227346420288086,
"learning_rate": 0.000172327868852459,
"loss": 0.6342,
"step": 4320
},
{
"epoch": 21.1836485661989,
"grad_norm": 2.9979476928710938,
"learning_rate": 0.00017219672131147542,
"loss": 0.6028,
"step": 4340
},
{
"epoch": 21.281269066503967,
"grad_norm": 3.6194779872894287,
"learning_rate": 0.0001720655737704918,
"loss": 0.6294,
"step": 4360
},
{
"epoch": 21.37888956680903,
"grad_norm": 4.482040882110596,
"learning_rate": 0.0001719344262295082,
"loss": 0.6283,
"step": 4380
},
{
"epoch": 21.476510067114095,
"grad_norm": 3.0028133392333984,
"learning_rate": 0.0001718032786885246,
"loss": 0.6292,
"step": 4400
},
{
"epoch": 21.574130567419157,
"grad_norm": 3.0038769245147705,
"learning_rate": 0.00017167213114754098,
"loss": 0.6349,
"step": 4420
},
{
"epoch": 21.671751067724223,
"grad_norm": 2.854794502258301,
"learning_rate": 0.0001715409836065574,
"loss": 0.632,
"step": 4440
},
{
"epoch": 21.769371568029285,
"grad_norm": 4.145231246948242,
"learning_rate": 0.00017140983606557376,
"loss": 0.6518,
"step": 4460
},
{
"epoch": 21.86699206833435,
"grad_norm": 2.7917041778564453,
"learning_rate": 0.00017127868852459018,
"loss": 0.6595,
"step": 4480
},
{
"epoch": 21.964612568639414,
"grad_norm": 2.858912706375122,
"learning_rate": 0.00017114754098360656,
"loss": 0.6507,
"step": 4500
},
{
"epoch": 22.06223306894448,
"grad_norm": 4.860498428344727,
"learning_rate": 0.00017101639344262298,
"loss": 0.6025,
"step": 4520
},
{
"epoch": 22.15985356924954,
"grad_norm": 3.04990291595459,
"learning_rate": 0.00017088524590163934,
"loss": 0.5384,
"step": 4540
},
{
"epoch": 22.257474069554608,
"grad_norm": 3.768587827682495,
"learning_rate": 0.00017075409836065573,
"loss": 0.5799,
"step": 4560
},
{
"epoch": 22.35509456985967,
"grad_norm": 3.870783805847168,
"learning_rate": 0.00017062295081967215,
"loss": 0.6453,
"step": 4580
},
{
"epoch": 22.452715070164736,
"grad_norm": 4.49683952331543,
"learning_rate": 0.0001704918032786885,
"loss": 0.6728,
"step": 4600
},
{
"epoch": 22.550335570469798,
"grad_norm": 3.2914700508117676,
"learning_rate": 0.00017036065573770493,
"loss": 0.6644,
"step": 4620
},
{
"epoch": 22.647956070774864,
"grad_norm": 2.4408113956451416,
"learning_rate": 0.00017022950819672132,
"loss": 0.6034,
"step": 4640
},
{
"epoch": 22.745576571079926,
"grad_norm": 4.22420072555542,
"learning_rate": 0.00017009836065573773,
"loss": 0.6464,
"step": 4660
},
{
"epoch": 22.843197071384992,
"grad_norm": 3.422456979751587,
"learning_rate": 0.0001699672131147541,
"loss": 0.6017,
"step": 4680
},
{
"epoch": 22.940817571690054,
"grad_norm": 2.6850616931915283,
"learning_rate": 0.00016983606557377049,
"loss": 0.6059,
"step": 4700
},
{
"epoch": 23.03843807199512,
"grad_norm": 4.270930290222168,
"learning_rate": 0.0001697049180327869,
"loss": 0.6273,
"step": 4720
},
{
"epoch": 23.136058572300183,
"grad_norm": 3.8276960849761963,
"learning_rate": 0.0001695737704918033,
"loss": 0.5779,
"step": 4740
},
{
"epoch": 23.23367907260525,
"grad_norm": 4.058398723602295,
"learning_rate": 0.00016944262295081968,
"loss": 0.593,
"step": 4760
},
{
"epoch": 23.33129957291031,
"grad_norm": 4.362659454345703,
"learning_rate": 0.00016931147540983607,
"loss": 0.5927,
"step": 4780
},
{
"epoch": 23.428920073215377,
"grad_norm": 3.7719762325286865,
"learning_rate": 0.00016918032786885249,
"loss": 0.6487,
"step": 4800
},
{
"epoch": 23.52654057352044,
"grad_norm": 2.9975759983062744,
"learning_rate": 0.00016904918032786885,
"loss": 0.614,
"step": 4820
},
{
"epoch": 23.624161073825505,
"grad_norm": 3.0442936420440674,
"learning_rate": 0.00016891803278688524,
"loss": 0.5966,
"step": 4840
},
{
"epoch": 23.721781574130567,
"grad_norm": 2.8749730587005615,
"learning_rate": 0.00016878688524590165,
"loss": 0.6166,
"step": 4860
},
{
"epoch": 23.819402074435633,
"grad_norm": 3.6400644779205322,
"learning_rate": 0.00016865573770491804,
"loss": 0.6329,
"step": 4880
},
{
"epoch": 23.917022574740695,
"grad_norm": 3.3705861568450928,
"learning_rate": 0.00016852459016393443,
"loss": 0.5753,
"step": 4900
},
{
"epoch": 24.01464307504576,
"grad_norm": 2.9685709476470947,
"learning_rate": 0.00016839344262295082,
"loss": 0.6302,
"step": 4920
},
{
"epoch": 24.112263575350823,
"grad_norm": 3.0799593925476074,
"learning_rate": 0.0001682622950819672,
"loss": 0.5936,
"step": 4940
},
{
"epoch": 24.20988407565589,
"grad_norm": 2.598597764968872,
"learning_rate": 0.00016813114754098363,
"loss": 0.5836,
"step": 4960
},
{
"epoch": 24.30750457596095,
"grad_norm": 3.719918966293335,
"learning_rate": 0.000168,
"loss": 0.5754,
"step": 4980
},
{
"epoch": 24.405125076266017,
"grad_norm": 3.592268466949463,
"learning_rate": 0.0001678688524590164,
"loss": 0.5683,
"step": 5000
},
{
"epoch": 24.50274557657108,
"grad_norm": 2.984259605407715,
"learning_rate": 0.0001677377049180328,
"loss": 0.5795,
"step": 5020
},
{
"epoch": 24.600366076876146,
"grad_norm": 2.833623170852661,
"learning_rate": 0.00016760655737704919,
"loss": 0.618,
"step": 5040
},
{
"epoch": 24.697986577181208,
"grad_norm": 5.3122172355651855,
"learning_rate": 0.00016747540983606558,
"loss": 0.5863,
"step": 5060
},
{
"epoch": 24.795607077486274,
"grad_norm": 2.9093785285949707,
"learning_rate": 0.00016734426229508196,
"loss": 0.5863,
"step": 5080
},
{
"epoch": 24.893227577791336,
"grad_norm": 2.8090052604675293,
"learning_rate": 0.00016721311475409838,
"loss": 0.6045,
"step": 5100
},
{
"epoch": 24.990848078096402,
"grad_norm": 2.9333415031433105,
"learning_rate": 0.00016708196721311474,
"loss": 0.612,
"step": 5120
},
{
"epoch": 25.088468578401464,
"grad_norm": 2.933706760406494,
"learning_rate": 0.00016695081967213116,
"loss": 0.5534,
"step": 5140
},
{
"epoch": 25.18608907870653,
"grad_norm": 4.508391380310059,
"learning_rate": 0.00016681967213114755,
"loss": 0.5384,
"step": 5160
},
{
"epoch": 25.283709579011592,
"grad_norm": 4.106027603149414,
"learning_rate": 0.00016668852459016397,
"loss": 0.5464,
"step": 5180
},
{
"epoch": 25.381330079316655,
"grad_norm": 4.005650997161865,
"learning_rate": 0.00016655737704918033,
"loss": 0.5805,
"step": 5200
},
{
"epoch": 25.47895057962172,
"grad_norm": 3.259774923324585,
"learning_rate": 0.00016642622950819672,
"loss": 0.5966,
"step": 5220
},
{
"epoch": 25.576571079926783,
"grad_norm": 3.695301055908203,
"learning_rate": 0.00016629508196721313,
"loss": 0.5344,
"step": 5240
},
{
"epoch": 25.67419158023185,
"grad_norm": 2.9240682125091553,
"learning_rate": 0.0001661639344262295,
"loss": 0.6109,
"step": 5260
},
{
"epoch": 25.77181208053691,
"grad_norm": 5.290099620819092,
"learning_rate": 0.0001660327868852459,
"loss": 0.5831,
"step": 5280
},
{
"epoch": 25.869432580841977,
"grad_norm": 3.4903934001922607,
"learning_rate": 0.0001659016393442623,
"loss": 0.6106,
"step": 5300
},
{
"epoch": 25.96705308114704,
"grad_norm": 4.101973533630371,
"learning_rate": 0.0001657704918032787,
"loss": 0.6319,
"step": 5320
},
{
"epoch": 26.064673581452105,
"grad_norm": 3.657115936279297,
"learning_rate": 0.00016563934426229508,
"loss": 0.5218,
"step": 5340
},
{
"epoch": 26.162294081757167,
"grad_norm": 3.4918181896209717,
"learning_rate": 0.00016550819672131147,
"loss": 0.5848,
"step": 5360
},
{
"epoch": 26.259914582062233,
"grad_norm": 3.117476224899292,
"learning_rate": 0.00016537704918032789,
"loss": 0.5579,
"step": 5380
},
{
"epoch": 26.357535082367296,
"grad_norm": 4.165419578552246,
"learning_rate": 0.00016524590163934428,
"loss": 0.535,
"step": 5400
},
{
"epoch": 26.45515558267236,
"grad_norm": 4.682600021362305,
"learning_rate": 0.00016511475409836067,
"loss": 0.6043,
"step": 5420
},
{
"epoch": 26.552776082977424,
"grad_norm": 3.7358362674713135,
"learning_rate": 0.00016498360655737705,
"loss": 0.5605,
"step": 5440
},
{
"epoch": 26.65039658328249,
"grad_norm": 3.1440377235412598,
"learning_rate": 0.00016485245901639344,
"loss": 0.5344,
"step": 5460
},
{
"epoch": 26.748017083587552,
"grad_norm": 4.453829288482666,
"learning_rate": 0.00016472131147540983,
"loss": 0.564,
"step": 5480
},
{
"epoch": 26.845637583892618,
"grad_norm": 3.959137439727783,
"learning_rate": 0.00016459016393442622,
"loss": 0.5986,
"step": 5500
},
{
"epoch": 26.94325808419768,
"grad_norm": 3.9074790477752686,
"learning_rate": 0.00016445901639344264,
"loss": 0.5789,
"step": 5520
},
{
"epoch": 27.040878584502746,
"grad_norm": 2.9816102981567383,
"learning_rate": 0.00016432786885245903,
"loss": 0.5651,
"step": 5540
},
{
"epoch": 27.13849908480781,
"grad_norm": 2.41890549659729,
"learning_rate": 0.00016419672131147542,
"loss": 0.5405,
"step": 5560
},
{
"epoch": 27.236119585112874,
"grad_norm": 3.1744871139526367,
"learning_rate": 0.0001640655737704918,
"loss": 0.5646,
"step": 5580
},
{
"epoch": 27.333740085417936,
"grad_norm": 3.6110246181488037,
"learning_rate": 0.0001639344262295082,
"loss": 0.5532,
"step": 5600
},
{
"epoch": 27.431360585723002,
"grad_norm": 4.252525806427002,
"learning_rate": 0.0001638032786885246,
"loss": 0.5243,
"step": 5620
},
{
"epoch": 27.528981086028065,
"grad_norm": 3.710491895675659,
"learning_rate": 0.00016367213114754098,
"loss": 0.57,
"step": 5640
},
{
"epoch": 27.62660158633313,
"grad_norm": 3.5450077056884766,
"learning_rate": 0.0001635409836065574,
"loss": 0.5401,
"step": 5660
},
{
"epoch": 27.724222086638193,
"grad_norm": 3.5488388538360596,
"learning_rate": 0.00016340983606557378,
"loss": 0.5345,
"step": 5680
},
{
"epoch": 27.82184258694326,
"grad_norm": 4.381565570831299,
"learning_rate": 0.00016327868852459017,
"loss": 0.5362,
"step": 5700
},
{
"epoch": 27.91946308724832,
"grad_norm": 4.077484130859375,
"learning_rate": 0.00016314754098360656,
"loss": 0.6039,
"step": 5720
},
{
"epoch": 28.017083587553387,
"grad_norm": 3.217458486557007,
"learning_rate": 0.00016301639344262295,
"loss": 0.6027,
"step": 5740
},
{
"epoch": 28.11470408785845,
"grad_norm": 3.741176128387451,
"learning_rate": 0.00016288524590163937,
"loss": 0.4955,
"step": 5760
},
{
"epoch": 28.212324588163515,
"grad_norm": 4.43143892288208,
"learning_rate": 0.00016275409836065573,
"loss": 0.5326,
"step": 5780
},
{
"epoch": 28.309945088468577,
"grad_norm": 3.040767192840576,
"learning_rate": 0.00016262295081967214,
"loss": 0.5207,
"step": 5800
},
{
"epoch": 28.407565588773643,
"grad_norm": 3.47582745552063,
"learning_rate": 0.00016249180327868853,
"loss": 0.5524,
"step": 5820
},
{
"epoch": 28.505186089078705,
"grad_norm": 4.133968830108643,
"learning_rate": 0.00016236065573770492,
"loss": 0.5553,
"step": 5840
},
{
"epoch": 28.60280658938377,
"grad_norm": 3.2506773471832275,
"learning_rate": 0.0001622295081967213,
"loss": 0.5287,
"step": 5860
},
{
"epoch": 28.700427089688834,
"grad_norm": 4.527762413024902,
"learning_rate": 0.0001620983606557377,
"loss": 0.507,
"step": 5880
},
{
"epoch": 28.7980475899939,
"grad_norm": 4.723377704620361,
"learning_rate": 0.00016196721311475412,
"loss": 0.6089,
"step": 5900
},
{
"epoch": 28.89566809029896,
"grad_norm": 3.6172468662261963,
"learning_rate": 0.00016183606557377048,
"loss": 0.542,
"step": 5920
},
{
"epoch": 28.993288590604028,
"grad_norm": 3.735133647918701,
"learning_rate": 0.0001617049180327869,
"loss": 0.5656,
"step": 5940
},
{
"epoch": 29.09090909090909,
"grad_norm": 5.235744476318359,
"learning_rate": 0.0001615737704918033,
"loss": 0.5355,
"step": 5960
},
{
"epoch": 29.188529591214156,
"grad_norm": 4.541784763336182,
"learning_rate": 0.00016144262295081968,
"loss": 0.5122,
"step": 5980
},
{
"epoch": 29.286150091519218,
"grad_norm": 3.0607311725616455,
"learning_rate": 0.00016131147540983607,
"loss": 0.5047,
"step": 6000
},
{
"epoch": 29.383770591824284,
"grad_norm": 4.826447486877441,
"learning_rate": 0.00016118032786885245,
"loss": 0.5077,
"step": 6020
},
{
"epoch": 29.481391092129346,
"grad_norm": 3.6418137550354004,
"learning_rate": 0.00016104918032786887,
"loss": 0.5401,
"step": 6040
},
{
"epoch": 29.579011592434412,
"grad_norm": 6.086111545562744,
"learning_rate": 0.00016091803278688526,
"loss": 0.5311,
"step": 6060
},
{
"epoch": 29.676632092739474,
"grad_norm": 2.81947922706604,
"learning_rate": 0.00016078688524590165,
"loss": 0.523,
"step": 6080
},
{
"epoch": 29.77425259304454,
"grad_norm": 4.2681355476379395,
"learning_rate": 0.00016065573770491804,
"loss": 0.5439,
"step": 6100
},
{
"epoch": 29.871873093349603,
"grad_norm": 3.6135213375091553,
"learning_rate": 0.00016052459016393443,
"loss": 0.5177,
"step": 6120
},
{
"epoch": 29.96949359365467,
"grad_norm": 3.3115713596343994,
"learning_rate": 0.00016039344262295082,
"loss": 0.581,
"step": 6140
},
{
"epoch": 30.06711409395973,
"grad_norm": 3.5007998943328857,
"learning_rate": 0.0001602622950819672,
"loss": 0.5092,
"step": 6160
},
{
"epoch": 30.164734594264797,
"grad_norm": 2.9520623683929443,
"learning_rate": 0.00016013114754098362,
"loss": 0.4636,
"step": 6180
},
{
"epoch": 30.26235509456986,
"grad_norm": 3.135349750518799,
"learning_rate": 0.00016,
"loss": 0.5391,
"step": 6200
},
{
"epoch": 30.359975594874925,
"grad_norm": 4.124795913696289,
"learning_rate": 0.0001598688524590164,
"loss": 0.5185,
"step": 6220
},
{
"epoch": 30.457596095179987,
"grad_norm": 3.6246888637542725,
"learning_rate": 0.0001597377049180328,
"loss": 0.532,
"step": 6240
},
{
"epoch": 30.555216595485053,
"grad_norm": 3.1558213233947754,
"learning_rate": 0.00015960655737704918,
"loss": 0.5261,
"step": 6260
},
{
"epoch": 30.652837095790115,
"grad_norm": 3.2610013484954834,
"learning_rate": 0.0001594754098360656,
"loss": 0.5104,
"step": 6280
},
{
"epoch": 30.75045759609518,
"grad_norm": 3.8648669719696045,
"learning_rate": 0.00015934426229508196,
"loss": 0.538,
"step": 6300
},
{
"epoch": 30.848078096400243,
"grad_norm": 3.0972094535827637,
"learning_rate": 0.00015921311475409838,
"loss": 0.5362,
"step": 6320
},
{
"epoch": 30.94569859670531,
"grad_norm": 4.238556385040283,
"learning_rate": 0.00015908196721311477,
"loss": 0.54,
"step": 6340
},
{
"epoch": 31.04331909701037,
"grad_norm": 4.820079326629639,
"learning_rate": 0.00015895081967213116,
"loss": 0.4893,
"step": 6360
},
{
"epoch": 31.140939597315437,
"grad_norm": 3.2895243167877197,
"learning_rate": 0.00015881967213114754,
"loss": 0.4502,
"step": 6380
},
{
"epoch": 31.2385600976205,
"grad_norm": 3.5625295639038086,
"learning_rate": 0.00015868852459016393,
"loss": 0.5033,
"step": 6400
},
{
"epoch": 31.336180597925566,
"grad_norm": 3.8144872188568115,
"learning_rate": 0.00015855737704918035,
"loss": 0.515,
"step": 6420
},
{
"epoch": 31.433801098230628,
"grad_norm": 4.3353271484375,
"learning_rate": 0.0001584262295081967,
"loss": 0.5356,
"step": 6440
},
{
"epoch": 31.531421598535694,
"grad_norm": 4.205932140350342,
"learning_rate": 0.00015829508196721313,
"loss": 0.5295,
"step": 6460
},
{
"epoch": 31.629042098840756,
"grad_norm": 4.171141147613525,
"learning_rate": 0.00015816393442622952,
"loss": 0.4502,
"step": 6480
},
{
"epoch": 31.726662599145822,
"grad_norm": 3.7091734409332275,
"learning_rate": 0.0001580327868852459,
"loss": 0.4885,
"step": 6500
},
{
"epoch": 31.824283099450884,
"grad_norm": 3.9756462574005127,
"learning_rate": 0.0001579016393442623,
"loss": 0.546,
"step": 6520
},
{
"epoch": 31.92190359975595,
"grad_norm": 4.506149768829346,
"learning_rate": 0.0001577704918032787,
"loss": 0.513,
"step": 6540
},
{
"epoch": 32.01952410006101,
"grad_norm": 3.1480486392974854,
"learning_rate": 0.0001576393442622951,
"loss": 0.5799,
"step": 6560
},
{
"epoch": 32.117144600366075,
"grad_norm": 2.476172685623169,
"learning_rate": 0.00015750819672131147,
"loss": 0.455,
"step": 6580
},
{
"epoch": 32.214765100671144,
"grad_norm": 3.0843114852905273,
"learning_rate": 0.00015737704918032788,
"loss": 0.4701,
"step": 6600
},
{
"epoch": 32.31238560097621,
"grad_norm": 3.4512338638305664,
"learning_rate": 0.00015724590163934427,
"loss": 0.4908,
"step": 6620
},
{
"epoch": 32.41000610128127,
"grad_norm": 3.0433006286621094,
"learning_rate": 0.00015711475409836066,
"loss": 0.4857,
"step": 6640
},
{
"epoch": 32.50762660158633,
"grad_norm": 3.1231653690338135,
"learning_rate": 0.00015698360655737705,
"loss": 0.5053,
"step": 6660
},
{
"epoch": 32.6052471018914,
"grad_norm": 3.121457576751709,
"learning_rate": 0.00015685245901639344,
"loss": 0.4962,
"step": 6680
},
{
"epoch": 32.70286760219646,
"grad_norm": 3.6397838592529297,
"learning_rate": 0.00015672131147540986,
"loss": 0.5083,
"step": 6700
},
{
"epoch": 32.800488102501525,
"grad_norm": 4.634298801422119,
"learning_rate": 0.00015659016393442622,
"loss": 0.5268,
"step": 6720
},
{
"epoch": 32.89810860280659,
"grad_norm": 3.6741957664489746,
"learning_rate": 0.00015645901639344263,
"loss": 0.5518,
"step": 6740
},
{
"epoch": 32.99572910311166,
"grad_norm": 3.7481977939605713,
"learning_rate": 0.00015632786885245902,
"loss": 0.5242,
"step": 6760
},
{
"epoch": 33.09334960341672,
"grad_norm": 3.2336630821228027,
"learning_rate": 0.0001561967213114754,
"loss": 0.4854,
"step": 6780
},
{
"epoch": 33.19097010372178,
"grad_norm": 3.153439521789551,
"learning_rate": 0.0001560655737704918,
"loss": 0.4512,
"step": 6800
},
{
"epoch": 33.288590604026844,
"grad_norm": 4.410555839538574,
"learning_rate": 0.0001559344262295082,
"loss": 0.4135,
"step": 6820
},
{
"epoch": 33.38621110433191,
"grad_norm": 4.749181747436523,
"learning_rate": 0.0001558032786885246,
"loss": 0.5023,
"step": 6840
},
{
"epoch": 33.483831604636975,
"grad_norm": 5.4415459632873535,
"learning_rate": 0.000155672131147541,
"loss": 0.4793,
"step": 6860
},
{
"epoch": 33.58145210494204,
"grad_norm": 4.493134498596191,
"learning_rate": 0.0001555409836065574,
"loss": 0.4827,
"step": 6880
},
{
"epoch": 33.6790726052471,
"grad_norm": 5.345827579498291,
"learning_rate": 0.00015540983606557378,
"loss": 0.4878,
"step": 6900
},
{
"epoch": 33.77669310555217,
"grad_norm": 5.483647346496582,
"learning_rate": 0.00015527868852459017,
"loss": 0.4976,
"step": 6920
},
{
"epoch": 33.87431360585723,
"grad_norm": 3.2896759510040283,
"learning_rate": 0.00015514754098360656,
"loss": 0.5163,
"step": 6940
},
{
"epoch": 33.971934106162294,
"grad_norm": 3.426711082458496,
"learning_rate": 0.00015501639344262294,
"loss": 0.6068,
"step": 6960
},
{
"epoch": 34.06955460646736,
"grad_norm": 5.182300567626953,
"learning_rate": 0.00015488524590163936,
"loss": 0.4459,
"step": 6980
},
{
"epoch": 34.16717510677242,
"grad_norm": 3.9542787075042725,
"learning_rate": 0.00015475409836065575,
"loss": 0.4688,
"step": 7000
},
{
"epoch": 34.26479560707749,
"grad_norm": 3.966188669204712,
"learning_rate": 0.00015462295081967214,
"loss": 0.4913,
"step": 7020
},
{
"epoch": 34.36241610738255,
"grad_norm": 3.175008535385132,
"learning_rate": 0.00015449180327868853,
"loss": 0.4564,
"step": 7040
},
{
"epoch": 34.46003660768761,
"grad_norm": 5.1779584884643555,
"learning_rate": 0.00015436065573770492,
"loss": 0.4807,
"step": 7060
},
{
"epoch": 34.557657107992675,
"grad_norm": 3.4940545558929443,
"learning_rate": 0.00015422950819672133,
"loss": 0.4948,
"step": 7080
},
{
"epoch": 34.655277608297745,
"grad_norm": 4.1444196701049805,
"learning_rate": 0.0001540983606557377,
"loss": 0.4756,
"step": 7100
},
{
"epoch": 34.75289810860281,
"grad_norm": 5.219560146331787,
"learning_rate": 0.00015396721311475411,
"loss": 0.5207,
"step": 7120
},
{
"epoch": 34.85051860890787,
"grad_norm": 3.7527666091918945,
"learning_rate": 0.0001538360655737705,
"loss": 0.484,
"step": 7140
},
{
"epoch": 34.94813910921293,
"grad_norm": 3.0565104484558105,
"learning_rate": 0.0001537049180327869,
"loss": 0.4795,
"step": 7160
},
{
"epoch": 35.045759609518,
"grad_norm": 2.8091022968292236,
"learning_rate": 0.00015357377049180328,
"loss": 0.4657,
"step": 7180
},
{
"epoch": 35.14338010982306,
"grad_norm": 2.8542001247406006,
"learning_rate": 0.00015344262295081967,
"loss": 0.4532,
"step": 7200
},
{
"epoch": 35.241000610128125,
"grad_norm": 3.4882893562316895,
"learning_rate": 0.0001533114754098361,
"loss": 0.4629,
"step": 7220
},
{
"epoch": 35.33862111043319,
"grad_norm": 3.3786075115203857,
"learning_rate": 0.00015318032786885245,
"loss": 0.4388,
"step": 7240
},
{
"epoch": 35.43624161073826,
"grad_norm": 4.602392673492432,
"learning_rate": 0.00015304918032786887,
"loss": 0.4788,
"step": 7260
},
{
"epoch": 35.53386211104332,
"grad_norm": 3.4427242279052734,
"learning_rate": 0.00015291803278688526,
"loss": 0.474,
"step": 7280
},
{
"epoch": 35.63148261134838,
"grad_norm": 3.069155216217041,
"learning_rate": 0.00015278688524590165,
"loss": 0.4385,
"step": 7300
},
{
"epoch": 35.729103111653444,
"grad_norm": 3.971026659011841,
"learning_rate": 0.00015265573770491803,
"loss": 0.5036,
"step": 7320
},
{
"epoch": 35.82672361195851,
"grad_norm": 3.3750109672546387,
"learning_rate": 0.00015252459016393442,
"loss": 0.4957,
"step": 7340
},
{
"epoch": 35.924344112263576,
"grad_norm": 3.382016181945801,
"learning_rate": 0.00015239344262295084,
"loss": 0.5176,
"step": 7360
},
{
"epoch": 36.02196461256864,
"grad_norm": 2.5936732292175293,
"learning_rate": 0.0001522622950819672,
"loss": 0.4807,
"step": 7380
},
{
"epoch": 36.1195851128737,
"grad_norm": 3.9948599338531494,
"learning_rate": 0.00015213114754098362,
"loss": 0.4285,
"step": 7400
},
{
"epoch": 36.21720561317877,
"grad_norm": 3.3358399868011475,
"learning_rate": 0.000152,
"loss": 0.5013,
"step": 7420
},
{
"epoch": 36.31482611348383,
"grad_norm": 3.7922635078430176,
"learning_rate": 0.0001518688524590164,
"loss": 0.4572,
"step": 7440
},
{
"epoch": 36.412446613788894,
"grad_norm": 2.7253546714782715,
"learning_rate": 0.0001517377049180328,
"loss": 0.4151,
"step": 7460
},
{
"epoch": 36.51006711409396,
"grad_norm": 5.344095706939697,
"learning_rate": 0.00015160655737704918,
"loss": 0.4718,
"step": 7480
},
{
"epoch": 36.607687614399026,
"grad_norm": 3.768033981323242,
"learning_rate": 0.0001514754098360656,
"loss": 0.4406,
"step": 7500
},
{
"epoch": 36.70530811470409,
"grad_norm": 3.884988784790039,
"learning_rate": 0.00015134426229508198,
"loss": 0.4605,
"step": 7520
},
{
"epoch": 36.80292861500915,
"grad_norm": 4.368161678314209,
"learning_rate": 0.00015121311475409837,
"loss": 0.4609,
"step": 7540
},
{
"epoch": 36.90054911531421,
"grad_norm": 5.337547779083252,
"learning_rate": 0.00015108196721311476,
"loss": 0.5078,
"step": 7560
},
{
"epoch": 36.99816961561928,
"grad_norm": 3.7230191230773926,
"learning_rate": 0.00015095081967213115,
"loss": 0.5003,
"step": 7580
},
{
"epoch": 37.095790115924345,
"grad_norm": 2.9239673614501953,
"learning_rate": 0.00015081967213114754,
"loss": 0.4312,
"step": 7600
},
{
"epoch": 37.19341061622941,
"grad_norm": 2.9180755615234375,
"learning_rate": 0.00015068852459016393,
"loss": 0.4876,
"step": 7620
},
{
"epoch": 37.29103111653447,
"grad_norm": 3.115180253982544,
"learning_rate": 0.00015055737704918035,
"loss": 0.4507,
"step": 7640
},
{
"epoch": 37.38865161683954,
"grad_norm": 4.005005359649658,
"learning_rate": 0.00015042622950819673,
"loss": 0.458,
"step": 7660
},
{
"epoch": 37.4862721171446,
"grad_norm": 3.3003063201904297,
"learning_rate": 0.00015029508196721312,
"loss": 0.4805,
"step": 7680
},
{
"epoch": 37.58389261744966,
"grad_norm": 2.783266544342041,
"learning_rate": 0.00015016393442622951,
"loss": 0.4742,
"step": 7700
},
{
"epoch": 37.681513117754726,
"grad_norm": 3.3052291870117188,
"learning_rate": 0.0001500327868852459,
"loss": 0.4415,
"step": 7720
},
{
"epoch": 37.779133618059795,
"grad_norm": 4.352939605712891,
"learning_rate": 0.00014990163934426232,
"loss": 0.4489,
"step": 7740
},
{
"epoch": 37.87675411836486,
"grad_norm": 2.9084181785583496,
"learning_rate": 0.00014977049180327868,
"loss": 0.439,
"step": 7760
},
{
"epoch": 37.97437461866992,
"grad_norm": 3.3146231174468994,
"learning_rate": 0.0001496393442622951,
"loss": 0.4568,
"step": 7780
},
{
"epoch": 38.07199511897498,
"grad_norm": 3.5614709854125977,
"learning_rate": 0.0001495081967213115,
"loss": 0.4648,
"step": 7800
},
{
"epoch": 38.16961561928005,
"grad_norm": 3.782663345336914,
"learning_rate": 0.00014937704918032788,
"loss": 0.4636,
"step": 7820
},
{
"epoch": 38.267236119585114,
"grad_norm": 3.2751007080078125,
"learning_rate": 0.00014924590163934427,
"loss": 0.4506,
"step": 7840
},
{
"epoch": 38.364856619890176,
"grad_norm": 4.198697090148926,
"learning_rate": 0.00014911475409836066,
"loss": 0.4271,
"step": 7860
},
{
"epoch": 38.46247712019524,
"grad_norm": 3.5415964126586914,
"learning_rate": 0.00014898360655737707,
"loss": 0.4112,
"step": 7880
},
{
"epoch": 38.56009762050031,
"grad_norm": 3.3707010746002197,
"learning_rate": 0.00014885245901639343,
"loss": 0.4788,
"step": 7900
},
{
"epoch": 38.65771812080537,
"grad_norm": 3.5206825733184814,
"learning_rate": 0.00014872131147540985,
"loss": 0.4456,
"step": 7920
},
{
"epoch": 38.75533862111043,
"grad_norm": 3.870487689971924,
"learning_rate": 0.00014859016393442624,
"loss": 0.4346,
"step": 7940
},
{
"epoch": 38.852959121415495,
"grad_norm": 4.031324863433838,
"learning_rate": 0.00014845901639344263,
"loss": 0.4272,
"step": 7960
},
{
"epoch": 38.950579621720564,
"grad_norm": 3.552272081375122,
"learning_rate": 0.00014832786885245902,
"loss": 0.455,
"step": 7980
},
{
"epoch": 39.04820012202563,
"grad_norm": 4.429355621337891,
"learning_rate": 0.0001481967213114754,
"loss": 0.474,
"step": 8000
},
{
"epoch": 39.14582062233069,
"grad_norm": 2.7380106449127197,
"learning_rate": 0.00014806557377049182,
"loss": 0.3892,
"step": 8020
},
{
"epoch": 39.24344112263575,
"grad_norm": 3.533512830734253,
"learning_rate": 0.0001479344262295082,
"loss": 0.4225,
"step": 8040
},
{
"epoch": 39.34106162294082,
"grad_norm": 4.216946125030518,
"learning_rate": 0.0001478032786885246,
"loss": 0.4288,
"step": 8060
},
{
"epoch": 39.43868212324588,
"grad_norm": 4.304781436920166,
"learning_rate": 0.000147672131147541,
"loss": 0.4678,
"step": 8080
},
{
"epoch": 39.536302623550945,
"grad_norm": 3.1035361289978027,
"learning_rate": 0.00014754098360655738,
"loss": 0.436,
"step": 8100
},
{
"epoch": 39.63392312385601,
"grad_norm": 4.312572956085205,
"learning_rate": 0.00014740983606557377,
"loss": 0.4418,
"step": 8120
},
{
"epoch": 39.73154362416108,
"grad_norm": 4.513503074645996,
"learning_rate": 0.00014727868852459016,
"loss": 0.4613,
"step": 8140
},
{
"epoch": 39.82916412446614,
"grad_norm": 3.7922120094299316,
"learning_rate": 0.00014714754098360658,
"loss": 0.4339,
"step": 8160
},
{
"epoch": 39.9267846247712,
"grad_norm": 4.1964616775512695,
"learning_rate": 0.00014701639344262297,
"loss": 0.4799,
"step": 8180
},
{
"epoch": 40.024405125076264,
"grad_norm": 3.0314066410064697,
"learning_rate": 0.00014688524590163936,
"loss": 0.4324,
"step": 8200
},
{
"epoch": 40.12202562538133,
"grad_norm": 3.27500319480896,
"learning_rate": 0.00014675409836065575,
"loss": 0.4353,
"step": 8220
},
{
"epoch": 40.219646125686396,
"grad_norm": 3.132371664047241,
"learning_rate": 0.00014662295081967214,
"loss": 0.4221,
"step": 8240
},
{
"epoch": 40.31726662599146,
"grad_norm": 4.715281009674072,
"learning_rate": 0.00014649180327868852,
"loss": 0.4344,
"step": 8260
},
{
"epoch": 40.41488712629652,
"grad_norm": 3.406825065612793,
"learning_rate": 0.00014636065573770491,
"loss": 0.4362,
"step": 8280
},
{
"epoch": 40.51250762660159,
"grad_norm": 3.668811082839966,
"learning_rate": 0.00014622950819672133,
"loss": 0.4217,
"step": 8300
},
{
"epoch": 40.61012812690665,
"grad_norm": 4.667581081390381,
"learning_rate": 0.00014609836065573772,
"loss": 0.4115,
"step": 8320
},
{
"epoch": 40.707748627211714,
"grad_norm": 4.834279537200928,
"learning_rate": 0.0001459672131147541,
"loss": 0.4292,
"step": 8340
},
{
"epoch": 40.80536912751678,
"grad_norm": 4.528176307678223,
"learning_rate": 0.0001458360655737705,
"loss": 0.4521,
"step": 8360
},
{
"epoch": 40.902989627821846,
"grad_norm": 4.21350622177124,
"learning_rate": 0.0001457049180327869,
"loss": 0.4593,
"step": 8380
},
{
"epoch": 41.00061012812691,
"grad_norm": 3.694138526916504,
"learning_rate": 0.0001455737704918033,
"loss": 0.434,
"step": 8400
},
{
"epoch": 41.09823062843197,
"grad_norm": 3.2935163974761963,
"learning_rate": 0.00014544262295081967,
"loss": 0.3825,
"step": 8420
},
{
"epoch": 41.19585112873703,
"grad_norm": 4.162698745727539,
"learning_rate": 0.00014531147540983608,
"loss": 0.3929,
"step": 8440
},
{
"epoch": 41.2934716290421,
"grad_norm": 3.650602340698242,
"learning_rate": 0.00014518032786885247,
"loss": 0.4134,
"step": 8460
},
{
"epoch": 41.391092129347165,
"grad_norm": 4.435211181640625,
"learning_rate": 0.00014504918032786886,
"loss": 0.4438,
"step": 8480
},
{
"epoch": 41.48871262965223,
"grad_norm": 3.3488500118255615,
"learning_rate": 0.00014491803278688525,
"loss": 0.3997,
"step": 8500
},
{
"epoch": 41.58633312995729,
"grad_norm": 3.115065574645996,
"learning_rate": 0.00014478688524590164,
"loss": 0.4443,
"step": 8520
},
{
"epoch": 41.68395363026236,
"grad_norm": 3.3622965812683105,
"learning_rate": 0.00014465573770491806,
"loss": 0.4335,
"step": 8540
},
{
"epoch": 41.78157413056742,
"grad_norm": 3.420353889465332,
"learning_rate": 0.00014452459016393442,
"loss": 0.4868,
"step": 8560
},
{
"epoch": 41.87919463087248,
"grad_norm": 4.102779388427734,
"learning_rate": 0.00014439344262295084,
"loss": 0.4201,
"step": 8580
},
{
"epoch": 41.976815131177545,
"grad_norm": 3.436736583709717,
"learning_rate": 0.00014426229508196722,
"loss": 0.4527,
"step": 8600
},
{
"epoch": 42.074435631482615,
"grad_norm": 3.5886449813842773,
"learning_rate": 0.00014413114754098361,
"loss": 0.4072,
"step": 8620
},
{
"epoch": 42.17205613178768,
"grad_norm": 3.397456645965576,
"learning_rate": 0.000144,
"loss": 0.3978,
"step": 8640
},
{
"epoch": 42.26967663209274,
"grad_norm": 3.1985559463500977,
"learning_rate": 0.0001438688524590164,
"loss": 0.4335,
"step": 8660
},
{
"epoch": 42.3672971323978,
"grad_norm": 3.780562400817871,
"learning_rate": 0.0001437377049180328,
"loss": 0.3899,
"step": 8680
},
{
"epoch": 42.464917632702864,
"grad_norm": 3.2960257530212402,
"learning_rate": 0.00014360655737704917,
"loss": 0.4167,
"step": 8700
},
{
"epoch": 42.56253813300793,
"grad_norm": 3.9879534244537354,
"learning_rate": 0.0001434754098360656,
"loss": 0.3799,
"step": 8720
},
{
"epoch": 42.660158633312996,
"grad_norm": 4.146468162536621,
"learning_rate": 0.00014334426229508198,
"loss": 0.4341,
"step": 8740
},
{
"epoch": 42.75777913361806,
"grad_norm": 3.994354724884033,
"learning_rate": 0.00014321311475409837,
"loss": 0.458,
"step": 8760
},
{
"epoch": 42.85539963392312,
"grad_norm": 3.043018341064453,
"learning_rate": 0.00014308196721311476,
"loss": 0.4277,
"step": 8780
},
{
"epoch": 42.95302013422819,
"grad_norm": 3.4882431030273438,
"learning_rate": 0.00014295081967213115,
"loss": 0.4491,
"step": 8800
},
{
"epoch": 43.05064063453325,
"grad_norm": 3.6433932781219482,
"learning_rate": 0.00014281967213114756,
"loss": 0.4009,
"step": 8820
},
{
"epoch": 43.148261134838314,
"grad_norm": 3.4401466846466064,
"learning_rate": 0.00014268852459016395,
"loss": 0.4191,
"step": 8840
},
{
"epoch": 43.24588163514338,
"grad_norm": 3.963707685470581,
"learning_rate": 0.00014255737704918034,
"loss": 0.3961,
"step": 8860
},
{
"epoch": 43.343502135448446,
"grad_norm": 3.0136444568634033,
"learning_rate": 0.00014242622950819673,
"loss": 0.4052,
"step": 8880
},
{
"epoch": 43.44112263575351,
"grad_norm": 3.807661533355713,
"learning_rate": 0.00014229508196721312,
"loss": 0.3929,
"step": 8900
},
{
"epoch": 43.53874313605857,
"grad_norm": 3.722825765609741,
"learning_rate": 0.0001421639344262295,
"loss": 0.4183,
"step": 8920
},
{
"epoch": 43.63636363636363,
"grad_norm": 3.470569610595703,
"learning_rate": 0.0001420327868852459,
"loss": 0.4169,
"step": 8940
},
{
"epoch": 43.7339841366687,
"grad_norm": 3.4966650009155273,
"learning_rate": 0.00014190163934426231,
"loss": 0.4121,
"step": 8960
},
{
"epoch": 43.831604636973765,
"grad_norm": 4.7074174880981445,
"learning_rate": 0.0001417704918032787,
"loss": 0.4634,
"step": 8980
},
{
"epoch": 43.92922513727883,
"grad_norm": 3.609379291534424,
"learning_rate": 0.0001416393442622951,
"loss": 0.4405,
"step": 9000
},
{
"epoch": 44.02684563758389,
"grad_norm": 3.337123394012451,
"learning_rate": 0.00014150819672131148,
"loss": 0.3883,
"step": 9020
},
{
"epoch": 44.12446613788896,
"grad_norm": 3.321878433227539,
"learning_rate": 0.00014137704918032787,
"loss": 0.4024,
"step": 9040
},
{
"epoch": 44.22208663819402,
"grad_norm": 3.427896738052368,
"learning_rate": 0.00014124590163934426,
"loss": 0.4032,
"step": 9060
},
{
"epoch": 44.31970713849908,
"grad_norm": 4.160934925079346,
"learning_rate": 0.00014111475409836065,
"loss": 0.3718,
"step": 9080
},
{
"epoch": 44.417327638804146,
"grad_norm": 3.591139316558838,
"learning_rate": 0.00014098360655737707,
"loss": 0.3829,
"step": 9100
},
{
"epoch": 44.514948139109215,
"grad_norm": 2.9896583557128906,
"learning_rate": 0.00014085245901639346,
"loss": 0.4231,
"step": 9120
},
{
"epoch": 44.61256863941428,
"grad_norm": 4.670585632324219,
"learning_rate": 0.00014072131147540985,
"loss": 0.4285,
"step": 9140
},
{
"epoch": 44.71018913971934,
"grad_norm": 4.087858200073242,
"learning_rate": 0.00014059016393442624,
"loss": 0.4061,
"step": 9160
},
{
"epoch": 44.8078096400244,
"grad_norm": 4.61581563949585,
"learning_rate": 0.00014045901639344262,
"loss": 0.4245,
"step": 9180
},
{
"epoch": 44.90543014032947,
"grad_norm": 4.715487957000732,
"learning_rate": 0.00014032786885245904,
"loss": 0.4223,
"step": 9200
},
{
"epoch": 45.003050640634534,
"grad_norm": 3.8034169673919678,
"learning_rate": 0.0001401967213114754,
"loss": 0.407,
"step": 9220
},
{
"epoch": 45.100671140939596,
"grad_norm": 2.813342332839966,
"learning_rate": 0.00014006557377049182,
"loss": 0.3936,
"step": 9240
},
{
"epoch": 45.19829164124466,
"grad_norm": 3.130986452102661,
"learning_rate": 0.0001399344262295082,
"loss": 0.3897,
"step": 9260
},
{
"epoch": 45.29591214154973,
"grad_norm": 3.6343700885772705,
"learning_rate": 0.0001398032786885246,
"loss": 0.3611,
"step": 9280
},
{
"epoch": 45.39353264185479,
"grad_norm": 3.8157641887664795,
"learning_rate": 0.000139672131147541,
"loss": 0.419,
"step": 9300
},
{
"epoch": 45.49115314215985,
"grad_norm": 3.9056179523468018,
"learning_rate": 0.00013954098360655738,
"loss": 0.3746,
"step": 9320
},
{
"epoch": 45.588773642464915,
"grad_norm": 3.952216625213623,
"learning_rate": 0.0001394098360655738,
"loss": 0.3876,
"step": 9340
},
{
"epoch": 45.686394142769984,
"grad_norm": 3.924835443496704,
"learning_rate": 0.00013927868852459016,
"loss": 0.4199,
"step": 9360
},
{
"epoch": 45.78401464307505,
"grad_norm": 4.389228343963623,
"learning_rate": 0.00013914754098360657,
"loss": 0.3949,
"step": 9380
},
{
"epoch": 45.88163514338011,
"grad_norm": 3.311279296875,
"learning_rate": 0.00013901639344262296,
"loss": 0.408,
"step": 9400
},
{
"epoch": 45.97925564368517,
"grad_norm": 3.3349342346191406,
"learning_rate": 0.00013888524590163935,
"loss": 0.4578,
"step": 9420
},
{
"epoch": 46.07687614399024,
"grad_norm": 2.8730790615081787,
"learning_rate": 0.00013875409836065574,
"loss": 0.3864,
"step": 9440
},
{
"epoch": 46.1744966442953,
"grad_norm": 5.034847259521484,
"learning_rate": 0.00013862295081967213,
"loss": 0.3707,
"step": 9460
},
{
"epoch": 46.272117144600365,
"grad_norm": 4.460078716278076,
"learning_rate": 0.00013849180327868855,
"loss": 0.3823,
"step": 9480
},
{
"epoch": 46.36973764490543,
"grad_norm": 3.5436556339263916,
"learning_rate": 0.0001383606557377049,
"loss": 0.4063,
"step": 9500
},
{
"epoch": 46.4673581452105,
"grad_norm": 3.1509130001068115,
"learning_rate": 0.00013822950819672133,
"loss": 0.3805,
"step": 9520
},
{
"epoch": 46.56497864551556,
"grad_norm": 3.3589935302734375,
"learning_rate": 0.00013809836065573771,
"loss": 0.4332,
"step": 9540
},
{
"epoch": 46.66259914582062,
"grad_norm": 3.7478573322296143,
"learning_rate": 0.0001379672131147541,
"loss": 0.3721,
"step": 9560
},
{
"epoch": 46.760219646125684,
"grad_norm": 3.930676221847534,
"learning_rate": 0.0001378360655737705,
"loss": 0.3847,
"step": 9580
},
{
"epoch": 46.85784014643075,
"grad_norm": 4.218331336975098,
"learning_rate": 0.00013770491803278688,
"loss": 0.4248,
"step": 9600
},
{
"epoch": 46.955460646735816,
"grad_norm": 3.993345022201538,
"learning_rate": 0.0001375737704918033,
"loss": 0.3976,
"step": 9620
},
{
"epoch": 47.05308114704088,
"grad_norm": 3.4169094562530518,
"learning_rate": 0.0001374426229508197,
"loss": 0.4126,
"step": 9640
},
{
"epoch": 47.15070164734594,
"grad_norm": 3.3602797985076904,
"learning_rate": 0.00013731147540983608,
"loss": 0.3675,
"step": 9660
},
{
"epoch": 47.24832214765101,
"grad_norm": 3.6925911903381348,
"learning_rate": 0.00013718032786885247,
"loss": 0.3732,
"step": 9680
},
{
"epoch": 47.34594264795607,
"grad_norm": 3.3294148445129395,
"learning_rate": 0.00013704918032786886,
"loss": 0.3576,
"step": 9700
},
{
"epoch": 47.443563148261134,
"grad_norm": 4.107402801513672,
"learning_rate": 0.00013691803278688525,
"loss": 0.376,
"step": 9720
},
{
"epoch": 47.5411836485662,
"grad_norm": 3.5636045932769775,
"learning_rate": 0.00013678688524590164,
"loss": 0.3918,
"step": 9740
},
{
"epoch": 47.638804148871266,
"grad_norm": 4.105456829071045,
"learning_rate": 0.00013665573770491805,
"loss": 0.3864,
"step": 9760
},
{
"epoch": 47.73642464917633,
"grad_norm": 3.6053054332733154,
"learning_rate": 0.00013652459016393444,
"loss": 0.4009,
"step": 9780
},
{
"epoch": 47.83404514948139,
"grad_norm": 4.397184371948242,
"learning_rate": 0.00013639344262295083,
"loss": 0.4094,
"step": 9800
},
{
"epoch": 47.93166564978645,
"grad_norm": 3.121466875076294,
"learning_rate": 0.00013626229508196722,
"loss": 0.4276,
"step": 9820
},
{
"epoch": 48.02928615009152,
"grad_norm": 3.9471912384033203,
"learning_rate": 0.0001361311475409836,
"loss": 0.3981,
"step": 9840
},
{
"epoch": 48.126906650396585,
"grad_norm": 2.690577983856201,
"learning_rate": 0.00013600000000000003,
"loss": 0.3281,
"step": 9860
},
{
"epoch": 48.22452715070165,
"grad_norm": 2.87288498878479,
"learning_rate": 0.0001358688524590164,
"loss": 0.4053,
"step": 9880
},
{
"epoch": 48.32214765100671,
"grad_norm": 3.6237525939941406,
"learning_rate": 0.0001357377049180328,
"loss": 0.3436,
"step": 9900
},
{
"epoch": 48.41976815131178,
"grad_norm": 3.1614110469818115,
"learning_rate": 0.0001356065573770492,
"loss": 0.3732,
"step": 9920
},
{
"epoch": 48.51738865161684,
"grad_norm": 4.337550640106201,
"learning_rate": 0.00013547540983606556,
"loss": 0.3558,
"step": 9940
},
{
"epoch": 48.6150091519219,
"grad_norm": 2.904919385910034,
"learning_rate": 0.00013534426229508197,
"loss": 0.4165,
"step": 9960
},
{
"epoch": 48.712629652226966,
"grad_norm": 3.3883869647979736,
"learning_rate": 0.00013521311475409836,
"loss": 0.37,
"step": 9980
},
{
"epoch": 48.810250152532035,
"grad_norm": 3.388711452484131,
"learning_rate": 0.00013508196721311478,
"loss": 0.4075,
"step": 10000
},
{
"epoch": 48.9078706528371,
"grad_norm": 3.972365617752075,
"learning_rate": 0.00013495081967213114,
"loss": 0.3944,
"step": 10020
},
{
"epoch": 49.00549115314216,
"grad_norm": 2.836026668548584,
"learning_rate": 0.00013481967213114756,
"loss": 0.4315,
"step": 10040
},
{
"epoch": 49.10311165344722,
"grad_norm": 3.727874755859375,
"learning_rate": 0.00013468852459016395,
"loss": 0.2791,
"step": 10060
},
{
"epoch": 49.20073215375229,
"grad_norm": 3.036069869995117,
"learning_rate": 0.00013455737704918034,
"loss": 0.358,
"step": 10080
},
{
"epoch": 49.298352654057354,
"grad_norm": 3.044581651687622,
"learning_rate": 0.00013442622950819673,
"loss": 0.423,
"step": 10100
},
{
"epoch": 49.395973154362416,
"grad_norm": 3.956517457962036,
"learning_rate": 0.00013429508196721311,
"loss": 0.3988,
"step": 10120
},
{
"epoch": 49.49359365466748,
"grad_norm": 4.359031677246094,
"learning_rate": 0.00013416393442622953,
"loss": 0.352,
"step": 10140
},
{
"epoch": 49.59121415497255,
"grad_norm": 4.619460582733154,
"learning_rate": 0.0001340327868852459,
"loss": 0.357,
"step": 10160
},
{
"epoch": 49.68883465527761,
"grad_norm": 4.051796913146973,
"learning_rate": 0.0001339016393442623,
"loss": 0.4048,
"step": 10180
},
{
"epoch": 49.78645515558267,
"grad_norm": 3.5017082691192627,
"learning_rate": 0.0001337704918032787,
"loss": 0.3937,
"step": 10200
},
{
"epoch": 49.884075655887735,
"grad_norm": 4.194153308868408,
"learning_rate": 0.0001336393442622951,
"loss": 0.3989,
"step": 10220
},
{
"epoch": 49.981696156192804,
"grad_norm": 3.5315494537353516,
"learning_rate": 0.00013350819672131148,
"loss": 0.4032,
"step": 10240
},
{
"epoch": 50.079316656497866,
"grad_norm": 3.8597731590270996,
"learning_rate": 0.00013337704918032787,
"loss": 0.3786,
"step": 10260
},
{
"epoch": 50.17693715680293,
"grad_norm": 3.1231532096862793,
"learning_rate": 0.00013324590163934428,
"loss": 0.3764,
"step": 10280
},
{
"epoch": 50.27455765710799,
"grad_norm": 3.272097110748291,
"learning_rate": 0.00013311475409836067,
"loss": 0.3943,
"step": 10300
},
{
"epoch": 50.37217815741306,
"grad_norm": 5.884640216827393,
"learning_rate": 0.00013298360655737706,
"loss": 0.312,
"step": 10320
},
{
"epoch": 50.46979865771812,
"grad_norm": 3.424651622772217,
"learning_rate": 0.00013285245901639345,
"loss": 0.3602,
"step": 10340
},
{
"epoch": 50.567419158023185,
"grad_norm": 3.703669786453247,
"learning_rate": 0.00013272131147540984,
"loss": 0.3992,
"step": 10360
},
{
"epoch": 50.66503965832825,
"grad_norm": 3.4607040882110596,
"learning_rate": 0.00013259016393442623,
"loss": 0.3773,
"step": 10380
},
{
"epoch": 50.76266015863331,
"grad_norm": 4.518139362335205,
"learning_rate": 0.00013245901639344262,
"loss": 0.3649,
"step": 10400
},
{
"epoch": 50.86028065893838,
"grad_norm": 3.7281651496887207,
"learning_rate": 0.00013232786885245904,
"loss": 0.3797,
"step": 10420
},
{
"epoch": 50.95790115924344,
"grad_norm": 3.4530558586120605,
"learning_rate": 0.00013219672131147543,
"loss": 0.3854,
"step": 10440
},
{
"epoch": 51.0555216595485,
"grad_norm": 3.220987558364868,
"learning_rate": 0.0001320655737704918,
"loss": 0.3606,
"step": 10460
},
{
"epoch": 51.153142159853566,
"grad_norm": 4.0888352394104,
"learning_rate": 0.0001319344262295082,
"loss": 0.3342,
"step": 10480
},
{
"epoch": 51.250762660158635,
"grad_norm": 3.6605489253997803,
"learning_rate": 0.0001318032786885246,
"loss": 0.3785,
"step": 10500
},
{
"epoch": 51.3483831604637,
"grad_norm": 3.1391944885253906,
"learning_rate": 0.000131672131147541,
"loss": 0.3741,
"step": 10520
},
{
"epoch": 51.44600366076876,
"grad_norm": 4.535563945770264,
"learning_rate": 0.00013154098360655737,
"loss": 0.3897,
"step": 10540
},
{
"epoch": 51.54362416107382,
"grad_norm": 2.85030198097229,
"learning_rate": 0.0001314098360655738,
"loss": 0.3335,
"step": 10560
},
{
"epoch": 51.64124466137889,
"grad_norm": 3.172010660171509,
"learning_rate": 0.00013127868852459018,
"loss": 0.3223,
"step": 10580
},
{
"epoch": 51.738865161683954,
"grad_norm": 3.834644079208374,
"learning_rate": 0.00013114754098360654,
"loss": 0.3897,
"step": 10600
},
{
"epoch": 51.836485661989016,
"grad_norm": 3.7358360290527344,
"learning_rate": 0.00013101639344262296,
"loss": 0.3853,
"step": 10620
},
{
"epoch": 51.93410616229408,
"grad_norm": 4.027935981750488,
"learning_rate": 0.00013088524590163935,
"loss": 0.3775,
"step": 10640
},
{
"epoch": 52.03172666259915,
"grad_norm": 2.9343104362487793,
"learning_rate": 0.00013075409836065576,
"loss": 0.3878,
"step": 10660
},
{
"epoch": 52.12934716290421,
"grad_norm": 3.9840517044067383,
"learning_rate": 0.00013062295081967213,
"loss": 0.3334,
"step": 10680
},
{
"epoch": 52.22696766320927,
"grad_norm": 2.7341692447662354,
"learning_rate": 0.00013049180327868854,
"loss": 0.3643,
"step": 10700
},
{
"epoch": 52.324588163514335,
"grad_norm": 4.158634662628174,
"learning_rate": 0.00013036065573770493,
"loss": 0.3497,
"step": 10720
},
{
"epoch": 52.422208663819404,
"grad_norm": 3.440031051635742,
"learning_rate": 0.00013022950819672132,
"loss": 0.3917,
"step": 10740
},
{
"epoch": 52.51982916412447,
"grad_norm": 3.250298500061035,
"learning_rate": 0.0001300983606557377,
"loss": 0.3472,
"step": 10760
},
{
"epoch": 52.61744966442953,
"grad_norm": 3.672611951828003,
"learning_rate": 0.0001299672131147541,
"loss": 0.3315,
"step": 10780
},
{
"epoch": 52.71507016473459,
"grad_norm": 2.9064407348632812,
"learning_rate": 0.00012983606557377052,
"loss": 0.4121,
"step": 10800
},
{
"epoch": 52.81269066503966,
"grad_norm": 3.6620004177093506,
"learning_rate": 0.00012970491803278688,
"loss": 0.3402,
"step": 10820
},
{
"epoch": 52.91031116534472,
"grad_norm": 4.237580299377441,
"learning_rate": 0.0001295737704918033,
"loss": 0.4054,
"step": 10840
},
{
"epoch": 53.007931665649785,
"grad_norm": 5.529778957366943,
"learning_rate": 0.00012944262295081968,
"loss": 0.359,
"step": 10860
},
{
"epoch": 53.10555216595485,
"grad_norm": 3.2727859020233154,
"learning_rate": 0.00012931147540983607,
"loss": 0.3431,
"step": 10880
},
{
"epoch": 53.20317266625992,
"grad_norm": 3.652461290359497,
"learning_rate": 0.00012918032786885246,
"loss": 0.3441,
"step": 10900
},
{
"epoch": 53.30079316656498,
"grad_norm": 3.462115526199341,
"learning_rate": 0.00012904918032786885,
"loss": 0.3754,
"step": 10920
},
{
"epoch": 53.39841366687004,
"grad_norm": 2.771514654159546,
"learning_rate": 0.00012891803278688527,
"loss": 0.3568,
"step": 10940
},
{
"epoch": 53.496034167175104,
"grad_norm": 3.806295394897461,
"learning_rate": 0.00012878688524590166,
"loss": 0.3357,
"step": 10960
},
{
"epoch": 53.59365466748017,
"grad_norm": 4.707630634307861,
"learning_rate": 0.00012865573770491802,
"loss": 0.3461,
"step": 10980
},
{
"epoch": 53.691275167785236,
"grad_norm": 4.196589946746826,
"learning_rate": 0.00012852459016393444,
"loss": 0.3578,
"step": 11000
},
{
"epoch": 53.7888956680903,
"grad_norm": 3.576612949371338,
"learning_rate": 0.00012839344262295083,
"loss": 0.3946,
"step": 11020
},
{
"epoch": 53.88651616839536,
"grad_norm": 3.4126014709472656,
"learning_rate": 0.00012826229508196722,
"loss": 0.3441,
"step": 11040
},
{
"epoch": 53.98413666870043,
"grad_norm": 3.8211019039154053,
"learning_rate": 0.0001281311475409836,
"loss": 0.3652,
"step": 11060
},
{
"epoch": 54.08175716900549,
"grad_norm": 3.447636127471924,
"learning_rate": 0.00012800000000000002,
"loss": 0.3315,
"step": 11080
},
{
"epoch": 54.179377669310554,
"grad_norm": 4.556658744812012,
"learning_rate": 0.0001278688524590164,
"loss": 0.37,
"step": 11100
},
{
"epoch": 54.27699816961562,
"grad_norm": 3.292505979537964,
"learning_rate": 0.00012773770491803277,
"loss": 0.3386,
"step": 11120
},
{
"epoch": 54.374618669920686,
"grad_norm": 2.7513251304626465,
"learning_rate": 0.0001276065573770492,
"loss": 0.3497,
"step": 11140
},
{
"epoch": 54.47223917022575,
"grad_norm": 3.0105299949645996,
"learning_rate": 0.00012747540983606558,
"loss": 0.3235,
"step": 11160
},
{
"epoch": 54.56985967053081,
"grad_norm": 3.2643628120422363,
"learning_rate": 0.000127344262295082,
"loss": 0.3673,
"step": 11180
},
{
"epoch": 54.66748017083587,
"grad_norm": 3.5350589752197266,
"learning_rate": 0.00012721311475409836,
"loss": 0.3733,
"step": 11200
},
{
"epoch": 54.76510067114094,
"grad_norm": 5.337443828582764,
"learning_rate": 0.00012708196721311477,
"loss": 0.3376,
"step": 11220
},
{
"epoch": 54.862721171446005,
"grad_norm": 3.618621826171875,
"learning_rate": 0.00012695081967213116,
"loss": 0.3372,
"step": 11240
},
{
"epoch": 54.96034167175107,
"grad_norm": 3.5537171363830566,
"learning_rate": 0.00012681967213114753,
"loss": 0.3639,
"step": 11260
},
{
"epoch": 55.05796217205613,
"grad_norm": 3.529486656188965,
"learning_rate": 0.00012668852459016394,
"loss": 0.3512,
"step": 11280
},
{
"epoch": 55.1555826723612,
"grad_norm": 4.277002811431885,
"learning_rate": 0.00012655737704918033,
"loss": 0.3841,
"step": 11300
},
{
"epoch": 55.25320317266626,
"grad_norm": 3.5343832969665527,
"learning_rate": 0.00012642622950819675,
"loss": 0.3757,
"step": 11320
},
{
"epoch": 55.35082367297132,
"grad_norm": 3.350728988647461,
"learning_rate": 0.0001262950819672131,
"loss": 0.3179,
"step": 11340
},
{
"epoch": 55.448444173276386,
"grad_norm": 4.037693500518799,
"learning_rate": 0.0001261639344262295,
"loss": 0.3685,
"step": 11360
},
{
"epoch": 55.546064673581455,
"grad_norm": 4.390751838684082,
"learning_rate": 0.00012603278688524592,
"loss": 0.3161,
"step": 11380
},
{
"epoch": 55.64368517388652,
"grad_norm": 4.671621322631836,
"learning_rate": 0.0001259016393442623,
"loss": 0.3268,
"step": 11400
},
{
"epoch": 55.74130567419158,
"grad_norm": 4.637950897216797,
"learning_rate": 0.0001257704918032787,
"loss": 0.357,
"step": 11420
},
{
"epoch": 55.83892617449664,
"grad_norm": 5.1185455322265625,
"learning_rate": 0.00012563934426229508,
"loss": 0.3628,
"step": 11440
},
{
"epoch": 55.93654667480171,
"grad_norm": 5.654429912567139,
"learning_rate": 0.0001255081967213115,
"loss": 0.3519,
"step": 11460
},
{
"epoch": 56.034167175106774,
"grad_norm": 3.041135549545288,
"learning_rate": 0.00012537704918032786,
"loss": 0.3529,
"step": 11480
},
{
"epoch": 56.131787675411836,
"grad_norm": 4.020240783691406,
"learning_rate": 0.00012524590163934425,
"loss": 0.3597,
"step": 11500
},
{
"epoch": 56.2294081757169,
"grad_norm": 3.730290412902832,
"learning_rate": 0.00012511475409836067,
"loss": 0.3137,
"step": 11520
},
{
"epoch": 56.32702867602197,
"grad_norm": 3.314856767654419,
"learning_rate": 0.00012498360655737706,
"loss": 0.3122,
"step": 11540
},
{
"epoch": 56.42464917632703,
"grad_norm": 4.214298248291016,
"learning_rate": 0.00012485245901639345,
"loss": 0.343,
"step": 11560
},
{
"epoch": 56.52226967663209,
"grad_norm": 4.5516791343688965,
"learning_rate": 0.00012472131147540984,
"loss": 0.3714,
"step": 11580
},
{
"epoch": 56.619890176937155,
"grad_norm": 5.098419666290283,
"learning_rate": 0.00012459016393442625,
"loss": 0.3489,
"step": 11600
},
{
"epoch": 56.717510677242224,
"grad_norm": 2.3853330612182617,
"learning_rate": 0.00012445901639344262,
"loss": 0.3389,
"step": 11620
},
{
"epoch": 56.815131177547286,
"grad_norm": 4.595917224884033,
"learning_rate": 0.000124327868852459,
"loss": 0.3392,
"step": 11640
},
{
"epoch": 56.91275167785235,
"grad_norm": 4.49993896484375,
"learning_rate": 0.00012419672131147542,
"loss": 0.3419,
"step": 11660
},
{
"epoch": 57.01037217815741,
"grad_norm": 3.2568199634552,
"learning_rate": 0.0001240655737704918,
"loss": 0.3614,
"step": 11680
},
{
"epoch": 57.10799267846248,
"grad_norm": 4.653031826019287,
"learning_rate": 0.0001239344262295082,
"loss": 0.3096,
"step": 11700
},
{
"epoch": 57.20561317876754,
"grad_norm": 2.5655269622802734,
"learning_rate": 0.0001238032786885246,
"loss": 0.3306,
"step": 11720
},
{
"epoch": 57.303233679072605,
"grad_norm": 3.4758191108703613,
"learning_rate": 0.000123672131147541,
"loss": 0.3513,
"step": 11740
},
{
"epoch": 57.40085417937767,
"grad_norm": 3.107544183731079,
"learning_rate": 0.0001235409836065574,
"loss": 0.3231,
"step": 11760
},
{
"epoch": 57.49847467968274,
"grad_norm": 3.983060121536255,
"learning_rate": 0.00012340983606557376,
"loss": 0.3366,
"step": 11780
},
{
"epoch": 57.5960951799878,
"grad_norm": 3.987680673599243,
"learning_rate": 0.00012327868852459017,
"loss": 0.3365,
"step": 11800
},
{
"epoch": 57.69371568029286,
"grad_norm": 3.7637107372283936,
"learning_rate": 0.00012314754098360656,
"loss": 0.3432,
"step": 11820
},
{
"epoch": 57.79133618059792,
"grad_norm": 3.193894386291504,
"learning_rate": 0.00012301639344262295,
"loss": 0.3668,
"step": 11840
},
{
"epoch": 57.88895668090299,
"grad_norm": 3.078507423400879,
"learning_rate": 0.00012288524590163934,
"loss": 0.3922,
"step": 11860
},
{
"epoch": 57.986577181208055,
"grad_norm": 3.561068058013916,
"learning_rate": 0.00012275409836065573,
"loss": 0.3285,
"step": 11880
},
{
"epoch": 58.08419768151312,
"grad_norm": 2.9965078830718994,
"learning_rate": 0.00012262295081967215,
"loss": 0.2991,
"step": 11900
},
{
"epoch": 58.18181818181818,
"grad_norm": 3.8305752277374268,
"learning_rate": 0.0001224918032786885,
"loss": 0.3728,
"step": 11920
},
{
"epoch": 58.27943868212325,
"grad_norm": 3.708678960800171,
"learning_rate": 0.00012236065573770493,
"loss": 0.3078,
"step": 11940
},
{
"epoch": 58.37705918242831,
"grad_norm": 3.8988072872161865,
"learning_rate": 0.00012222950819672132,
"loss": 0.3389,
"step": 11960
},
{
"epoch": 58.474679682733374,
"grad_norm": 3.550420045852661,
"learning_rate": 0.00012209836065573773,
"loss": 0.4069,
"step": 11980
},
{
"epoch": 58.572300183038436,
"grad_norm": 3.799747943878174,
"learning_rate": 0.0001219672131147541,
"loss": 0.3058,
"step": 12000
},
{
"epoch": 58.669920683343506,
"grad_norm": 4.734778881072998,
"learning_rate": 0.0001218360655737705,
"loss": 0.3199,
"step": 12020
},
{
"epoch": 58.76754118364857,
"grad_norm": 3.449169158935547,
"learning_rate": 0.0001217049180327869,
"loss": 0.3309,
"step": 12040
},
{
"epoch": 58.86516168395363,
"grad_norm": 3.164651393890381,
"learning_rate": 0.00012157377049180328,
"loss": 0.3338,
"step": 12060
},
{
"epoch": 58.96278218425869,
"grad_norm": 3.976644515991211,
"learning_rate": 0.00012144262295081968,
"loss": 0.3207,
"step": 12080
},
{
"epoch": 59.060402684563755,
"grad_norm": 4.6017632484436035,
"learning_rate": 0.00012131147540983607,
"loss": 0.317,
"step": 12100
},
{
"epoch": 59.158023184868824,
"grad_norm": 6.806497573852539,
"learning_rate": 0.00012118032786885247,
"loss": 0.2943,
"step": 12120
},
{
"epoch": 59.25564368517389,
"grad_norm": 3.545241355895996,
"learning_rate": 0.00012104918032786885,
"loss": 0.3093,
"step": 12140
},
{
"epoch": 59.35326418547895,
"grad_norm": 2.435683488845825,
"learning_rate": 0.00012091803278688525,
"loss": 0.3235,
"step": 12160
},
{
"epoch": 59.45088468578401,
"grad_norm": 3.507638931274414,
"learning_rate": 0.00012078688524590165,
"loss": 0.3405,
"step": 12180
},
{
"epoch": 59.54850518608908,
"grad_norm": 3.9438552856445312,
"learning_rate": 0.00012065573770491804,
"loss": 0.3311,
"step": 12200
},
{
"epoch": 59.64612568639414,
"grad_norm": 5.4473652839660645,
"learning_rate": 0.00012052459016393443,
"loss": 0.3424,
"step": 12220
},
{
"epoch": 59.743746186699205,
"grad_norm": 3.204779624938965,
"learning_rate": 0.00012039344262295082,
"loss": 0.3297,
"step": 12240
},
{
"epoch": 59.84136668700427,
"grad_norm": 3.356764793395996,
"learning_rate": 0.00012026229508196722,
"loss": 0.3698,
"step": 12260
},
{
"epoch": 59.93898718730934,
"grad_norm": 3.6063504219055176,
"learning_rate": 0.0001201311475409836,
"loss": 0.3479,
"step": 12280
},
{
"epoch": 60.0366076876144,
"grad_norm": 3.49246883392334,
"learning_rate": 0.00012,
"loss": 0.3394,
"step": 12300
},
{
"epoch": 60.13422818791946,
"grad_norm": 4.206517219543457,
"learning_rate": 0.0001198688524590164,
"loss": 0.2825,
"step": 12320
},
{
"epoch": 60.231848688224524,
"grad_norm": 3.532618999481201,
"learning_rate": 0.0001197377049180328,
"loss": 0.3162,
"step": 12340
},
{
"epoch": 60.32946918852959,
"grad_norm": 3.6355433464050293,
"learning_rate": 0.00011960655737704917,
"loss": 0.3138,
"step": 12360
},
{
"epoch": 60.427089688834656,
"grad_norm": 3.1386282444000244,
"learning_rate": 0.00011947540983606557,
"loss": 0.335,
"step": 12380
},
{
"epoch": 60.52471018913972,
"grad_norm": 2.946789026260376,
"learning_rate": 0.00011934426229508198,
"loss": 0.321,
"step": 12400
},
{
"epoch": 60.62233068944478,
"grad_norm": 4.056975841522217,
"learning_rate": 0.00011921311475409838,
"loss": 0.3647,
"step": 12420
},
{
"epoch": 60.71995118974985,
"grad_norm": 3.8449249267578125,
"learning_rate": 0.00011908196721311476,
"loss": 0.3189,
"step": 12440
},
{
"epoch": 60.81757169005491,
"grad_norm": 3.464470148086548,
"learning_rate": 0.00011895081967213116,
"loss": 0.3711,
"step": 12460
},
{
"epoch": 60.915192190359974,
"grad_norm": 3.6908376216888428,
"learning_rate": 0.00011881967213114755,
"loss": 0.3303,
"step": 12480
},
{
"epoch": 61.01281269066504,
"grad_norm": 4.723835468292236,
"learning_rate": 0.00011868852459016392,
"loss": 0.305,
"step": 12500
},
{
"epoch": 61.110433190970106,
"grad_norm": 3.6280064582824707,
"learning_rate": 0.00011855737704918033,
"loss": 0.3103,
"step": 12520
},
{
"epoch": 61.20805369127517,
"grad_norm": 3.640411138534546,
"learning_rate": 0.00011842622950819673,
"loss": 0.3017,
"step": 12540
},
{
"epoch": 61.30567419158023,
"grad_norm": 4.304122447967529,
"learning_rate": 0.00011829508196721313,
"loss": 0.3452,
"step": 12560
},
{
"epoch": 61.40329469188529,
"grad_norm": 3.4708502292633057,
"learning_rate": 0.00011816393442622951,
"loss": 0.3195,
"step": 12580
},
{
"epoch": 61.50091519219036,
"grad_norm": 3.257230758666992,
"learning_rate": 0.00011803278688524591,
"loss": 0.3115,
"step": 12600
},
{
"epoch": 61.598535692495425,
"grad_norm": 3.0247697830200195,
"learning_rate": 0.0001179016393442623,
"loss": 0.314,
"step": 12620
},
{
"epoch": 61.69615619280049,
"grad_norm": 2.5768277645111084,
"learning_rate": 0.0001177704918032787,
"loss": 0.3195,
"step": 12640
},
{
"epoch": 61.79377669310555,
"grad_norm": 4.08043909072876,
"learning_rate": 0.00011763934426229508,
"loss": 0.3312,
"step": 12660
},
{
"epoch": 61.89139719341062,
"grad_norm": 5.277688503265381,
"learning_rate": 0.00011750819672131148,
"loss": 0.3607,
"step": 12680
},
{
"epoch": 61.98901769371568,
"grad_norm": 4.2203288078308105,
"learning_rate": 0.00011737704918032789,
"loss": 0.3233,
"step": 12700
},
{
"epoch": 62.08663819402074,
"grad_norm": 3.8487565517425537,
"learning_rate": 0.00011724590163934426,
"loss": 0.2777,
"step": 12720
},
{
"epoch": 62.184258694325806,
"grad_norm": 3.509904623031616,
"learning_rate": 0.00011711475409836066,
"loss": 0.3418,
"step": 12740
},
{
"epoch": 62.281879194630875,
"grad_norm": 3.7442731857299805,
"learning_rate": 0.00011698360655737705,
"loss": 0.3115,
"step": 12760
},
{
"epoch": 62.37949969493594,
"grad_norm": 3.3927197456359863,
"learning_rate": 0.00011685245901639346,
"loss": 0.3304,
"step": 12780
},
{
"epoch": 62.477120195241,
"grad_norm": 4.529331684112549,
"learning_rate": 0.00011672131147540983,
"loss": 0.3246,
"step": 12800
},
{
"epoch": 62.57474069554606,
"grad_norm": 2.9640207290649414,
"learning_rate": 0.00011659016393442623,
"loss": 0.3228,
"step": 12820
},
{
"epoch": 62.67236119585113,
"grad_norm": 4.769125938415527,
"learning_rate": 0.00011645901639344264,
"loss": 0.3552,
"step": 12840
},
{
"epoch": 62.769981696156194,
"grad_norm": 4.817038536071777,
"learning_rate": 0.00011632786885245903,
"loss": 0.269,
"step": 12860
},
{
"epoch": 62.867602196461256,
"grad_norm": 4.221793174743652,
"learning_rate": 0.0001161967213114754,
"loss": 0.3192,
"step": 12880
},
{
"epoch": 62.96522269676632,
"grad_norm": 3.8652241230010986,
"learning_rate": 0.0001160655737704918,
"loss": 0.3415,
"step": 12900
},
{
"epoch": 63.06284319707139,
"grad_norm": 3.7543272972106934,
"learning_rate": 0.00011593442622950821,
"loss": 0.2707,
"step": 12920
},
{
"epoch": 63.16046369737645,
"grad_norm": 3.2344672679901123,
"learning_rate": 0.00011580327868852458,
"loss": 0.3133,
"step": 12940
},
{
"epoch": 63.25808419768151,
"grad_norm": 2.9289095401763916,
"learning_rate": 0.00011567213114754099,
"loss": 0.3111,
"step": 12960
},
{
"epoch": 63.355704697986575,
"grad_norm": 4.034417629241943,
"learning_rate": 0.00011554098360655739,
"loss": 0.3443,
"step": 12980
},
{
"epoch": 63.453325198291644,
"grad_norm": 3.571948528289795,
"learning_rate": 0.00011540983606557378,
"loss": 0.3061,
"step": 13000
},
{
"epoch": 63.550945698596706,
"grad_norm": 3.8213961124420166,
"learning_rate": 0.00011527868852459016,
"loss": 0.2943,
"step": 13020
},
{
"epoch": 63.64856619890177,
"grad_norm": 4.343628883361816,
"learning_rate": 0.00011514754098360656,
"loss": 0.3499,
"step": 13040
},
{
"epoch": 63.74618669920683,
"grad_norm": 3.373922348022461,
"learning_rate": 0.00011501639344262296,
"loss": 0.3193,
"step": 13060
},
{
"epoch": 63.8438071995119,
"grad_norm": 4.19200325012207,
"learning_rate": 0.00011488524590163936,
"loss": 0.3221,
"step": 13080
},
{
"epoch": 63.94142769981696,
"grad_norm": 3.28054141998291,
"learning_rate": 0.00011475409836065574,
"loss": 0.328,
"step": 13100
},
{
"epoch": 64.03904820012202,
"grad_norm": 3.887531042098999,
"learning_rate": 0.00011462295081967214,
"loss": 0.3161,
"step": 13120
},
{
"epoch": 64.1366687004271,
"grad_norm": 3.633239507675171,
"learning_rate": 0.00011449180327868853,
"loss": 0.3001,
"step": 13140
},
{
"epoch": 64.23428920073215,
"grad_norm": 3.5353002548217773,
"learning_rate": 0.00011436065573770491,
"loss": 0.3085,
"step": 13160
},
{
"epoch": 64.33190970103722,
"grad_norm": 3.3285086154937744,
"learning_rate": 0.00011422950819672131,
"loss": 0.2807,
"step": 13180
},
{
"epoch": 64.42953020134229,
"grad_norm": 4.186200141906738,
"learning_rate": 0.00011409836065573771,
"loss": 0.3072,
"step": 13200
},
{
"epoch": 64.52715070164734,
"grad_norm": 4.000847339630127,
"learning_rate": 0.00011396721311475412,
"loss": 0.2794,
"step": 13220
},
{
"epoch": 64.62477120195241,
"grad_norm": 3.643327474594116,
"learning_rate": 0.00011383606557377049,
"loss": 0.3292,
"step": 13240
},
{
"epoch": 64.72239170225747,
"grad_norm": 4.381191730499268,
"learning_rate": 0.00011370491803278688,
"loss": 0.3311,
"step": 13260
},
{
"epoch": 64.82001220256254,
"grad_norm": 3.9231462478637695,
"learning_rate": 0.00011357377049180329,
"loss": 0.3239,
"step": 13280
},
{
"epoch": 64.9176327028676,
"grad_norm": 3.68272066116333,
"learning_rate": 0.00011344262295081969,
"loss": 0.3573,
"step": 13300
},
{
"epoch": 65.01525320317266,
"grad_norm": 3.5405211448669434,
"learning_rate": 0.00011331147540983606,
"loss": 0.3026,
"step": 13320
},
{
"epoch": 65.11287370347773,
"grad_norm": 4.24603796005249,
"learning_rate": 0.00011318032786885247,
"loss": 0.2984,
"step": 13340
},
{
"epoch": 65.2104942037828,
"grad_norm": 3.9128637313842773,
"learning_rate": 0.00011304918032786887,
"loss": 0.3203,
"step": 13360
},
{
"epoch": 65.30811470408786,
"grad_norm": 3.794922113418579,
"learning_rate": 0.00011291803278688525,
"loss": 0.2696,
"step": 13380
},
{
"epoch": 65.40573520439293,
"grad_norm": 7.95392370223999,
"learning_rate": 0.00011278688524590164,
"loss": 0.3349,
"step": 13400
},
{
"epoch": 65.50335570469798,
"grad_norm": 4.138427257537842,
"learning_rate": 0.00011265573770491804,
"loss": 0.303,
"step": 13420
},
{
"epoch": 65.60097620500305,
"grad_norm": 3.470979690551758,
"learning_rate": 0.00011252459016393444,
"loss": 0.3019,
"step": 13440
},
{
"epoch": 65.69859670530812,
"grad_norm": 3.5186235904693604,
"learning_rate": 0.00011239344262295082,
"loss": 0.3254,
"step": 13460
},
{
"epoch": 65.79621720561317,
"grad_norm": 5.092533111572266,
"learning_rate": 0.00011226229508196722,
"loss": 0.3167,
"step": 13480
},
{
"epoch": 65.89383770591824,
"grad_norm": 4.472383499145508,
"learning_rate": 0.00011213114754098362,
"loss": 0.3102,
"step": 13500
},
{
"epoch": 65.99145820622331,
"grad_norm": 4.198476314544678,
"learning_rate": 0.00011200000000000001,
"loss": 0.3085,
"step": 13520
},
{
"epoch": 66.08907870652837,
"grad_norm": 4.235730171203613,
"learning_rate": 0.00011186885245901639,
"loss": 0.2754,
"step": 13540
},
{
"epoch": 66.18669920683344,
"grad_norm": 3.673928737640381,
"learning_rate": 0.00011173770491803279,
"loss": 0.2883,
"step": 13560
},
{
"epoch": 66.2843197071385,
"grad_norm": 3.8741283416748047,
"learning_rate": 0.0001116065573770492,
"loss": 0.3055,
"step": 13580
},
{
"epoch": 66.38194020744356,
"grad_norm": 3.2982194423675537,
"learning_rate": 0.00011147540983606557,
"loss": 0.3309,
"step": 13600
},
{
"epoch": 66.47956070774863,
"grad_norm": 3.1477532386779785,
"learning_rate": 0.00011134426229508197,
"loss": 0.299,
"step": 13620
},
{
"epoch": 66.57718120805369,
"grad_norm": 3.723947525024414,
"learning_rate": 0.00011121311475409838,
"loss": 0.3016,
"step": 13640
},
{
"epoch": 66.67480170835876,
"grad_norm": 4.094952583312988,
"learning_rate": 0.00011108196721311476,
"loss": 0.3045,
"step": 13660
},
{
"epoch": 66.77242220866383,
"grad_norm": 3.607595205307007,
"learning_rate": 0.00011095081967213114,
"loss": 0.2983,
"step": 13680
},
{
"epoch": 66.87004270896888,
"grad_norm": 4.301547527313232,
"learning_rate": 0.00011081967213114754,
"loss": 0.2783,
"step": 13700
},
{
"epoch": 66.96766320927395,
"grad_norm": 3.570128917694092,
"learning_rate": 0.00011068852459016395,
"loss": 0.3547,
"step": 13720
},
{
"epoch": 67.065283709579,
"grad_norm": 4.2378621101379395,
"learning_rate": 0.00011055737704918035,
"loss": 0.3097,
"step": 13740
},
{
"epoch": 67.16290420988408,
"grad_norm": 3.6283559799194336,
"learning_rate": 0.00011042622950819672,
"loss": 0.2773,
"step": 13760
},
{
"epoch": 67.26052471018915,
"grad_norm": 3.317201614379883,
"learning_rate": 0.00011029508196721311,
"loss": 0.3179,
"step": 13780
},
{
"epoch": 67.3581452104942,
"grad_norm": 3.120859146118164,
"learning_rate": 0.00011016393442622952,
"loss": 0.3239,
"step": 13800
},
{
"epoch": 67.45576571079927,
"grad_norm": 3.163276433944702,
"learning_rate": 0.00011003278688524589,
"loss": 0.2875,
"step": 13820
},
{
"epoch": 67.55338621110434,
"grad_norm": 3.597426652908325,
"learning_rate": 0.0001099016393442623,
"loss": 0.2739,
"step": 13840
},
{
"epoch": 67.6510067114094,
"grad_norm": 3.572763204574585,
"learning_rate": 0.0001097704918032787,
"loss": 0.3009,
"step": 13860
},
{
"epoch": 67.74862721171446,
"grad_norm": 3.660034656524658,
"learning_rate": 0.0001096393442622951,
"loss": 0.3089,
"step": 13880
},
{
"epoch": 67.84624771201952,
"grad_norm": 4.177897930145264,
"learning_rate": 0.00010950819672131148,
"loss": 0.3205,
"step": 13900
},
{
"epoch": 67.94386821232459,
"grad_norm": 5.586677074432373,
"learning_rate": 0.00010937704918032787,
"loss": 0.3263,
"step": 13920
},
{
"epoch": 68.04148871262966,
"grad_norm": 2.8578503131866455,
"learning_rate": 0.00010924590163934427,
"loss": 0.3139,
"step": 13940
},
{
"epoch": 68.13910921293471,
"grad_norm": 3.129765748977661,
"learning_rate": 0.00010911475409836067,
"loss": 0.2729,
"step": 13960
},
{
"epoch": 68.23672971323978,
"grad_norm": 3.238567590713501,
"learning_rate": 0.00010898360655737705,
"loss": 0.3008,
"step": 13980
},
{
"epoch": 68.33435021354484,
"grad_norm": 3.5786325931549072,
"learning_rate": 0.00010885245901639345,
"loss": 0.2818,
"step": 14000
},
{
"epoch": 68.4319707138499,
"grad_norm": 3.5092949867248535,
"learning_rate": 0.00010872131147540985,
"loss": 0.3034,
"step": 14020
},
{
"epoch": 68.52959121415498,
"grad_norm": 4.172147274017334,
"learning_rate": 0.00010859016393442623,
"loss": 0.2917,
"step": 14040
},
{
"epoch": 68.62721171446003,
"grad_norm": 3.5723907947540283,
"learning_rate": 0.00010845901639344262,
"loss": 0.3119,
"step": 14060
},
{
"epoch": 68.7248322147651,
"grad_norm": 3.2721340656280518,
"learning_rate": 0.00010832786885245902,
"loss": 0.2698,
"step": 14080
},
{
"epoch": 68.82245271507017,
"grad_norm": 4.181795597076416,
"learning_rate": 0.00010819672131147543,
"loss": 0.3181,
"step": 14100
},
{
"epoch": 68.92007321537523,
"grad_norm": 4.145654678344727,
"learning_rate": 0.0001080655737704918,
"loss": 0.2837,
"step": 14120
},
{
"epoch": 69.0176937156803,
"grad_norm": 4.098430633544922,
"learning_rate": 0.0001079344262295082,
"loss": 0.3281,
"step": 14140
},
{
"epoch": 69.11531421598535,
"grad_norm": 3.6833229064941406,
"learning_rate": 0.00010780327868852461,
"loss": 0.3059,
"step": 14160
},
{
"epoch": 69.21293471629042,
"grad_norm": 3.802940845489502,
"learning_rate": 0.00010767213114754098,
"loss": 0.2648,
"step": 14180
},
{
"epoch": 69.31055521659549,
"grad_norm": 2.9509127140045166,
"learning_rate": 0.00010754098360655737,
"loss": 0.2828,
"step": 14200
},
{
"epoch": 69.40817571690054,
"grad_norm": 3.6474950313568115,
"learning_rate": 0.00010740983606557378,
"loss": 0.2848,
"step": 14220
},
{
"epoch": 69.50579621720561,
"grad_norm": 3.762017250061035,
"learning_rate": 0.00010727868852459018,
"loss": 0.3305,
"step": 14240
},
{
"epoch": 69.60341671751068,
"grad_norm": 3.3296637535095215,
"learning_rate": 0.00010714754098360655,
"loss": 0.2885,
"step": 14260
},
{
"epoch": 69.70103721781574,
"grad_norm": 4.1584296226501465,
"learning_rate": 0.00010701639344262296,
"loss": 0.3162,
"step": 14280
},
{
"epoch": 69.79865771812081,
"grad_norm": 3.365910768508911,
"learning_rate": 0.00010688524590163935,
"loss": 0.2855,
"step": 14300
},
{
"epoch": 69.89627821842586,
"grad_norm": 3.9548261165618896,
"learning_rate": 0.00010675409836065575,
"loss": 0.3001,
"step": 14320
},
{
"epoch": 69.99389871873093,
"grad_norm": 4.556044578552246,
"learning_rate": 0.00010662295081967212,
"loss": 0.2855,
"step": 14340
},
{
"epoch": 70.091519219036,
"grad_norm": 4.781431198120117,
"learning_rate": 0.00010649180327868853,
"loss": 0.2741,
"step": 14360
},
{
"epoch": 70.18913971934106,
"grad_norm": 2.645362377166748,
"learning_rate": 0.00010636065573770493,
"loss": 0.2885,
"step": 14380
},
{
"epoch": 70.28676021964613,
"grad_norm": 4.382370471954346,
"learning_rate": 0.0001062295081967213,
"loss": 0.3038,
"step": 14400
},
{
"epoch": 70.3843807199512,
"grad_norm": 2.908954381942749,
"learning_rate": 0.00010609836065573771,
"loss": 0.2765,
"step": 14420
},
{
"epoch": 70.48200122025625,
"grad_norm": 3.298417091369629,
"learning_rate": 0.0001059672131147541,
"loss": 0.2986,
"step": 14440
},
{
"epoch": 70.57962172056132,
"grad_norm": 3.324769973754883,
"learning_rate": 0.0001058360655737705,
"loss": 0.2712,
"step": 14460
},
{
"epoch": 70.67724222086638,
"grad_norm": 4.584010124206543,
"learning_rate": 0.00010570491803278688,
"loss": 0.3049,
"step": 14480
},
{
"epoch": 70.77486272117144,
"grad_norm": 3.648829221725464,
"learning_rate": 0.00010557377049180328,
"loss": 0.3169,
"step": 14500
},
{
"epoch": 70.87248322147651,
"grad_norm": 3.729729175567627,
"learning_rate": 0.00010544262295081968,
"loss": 0.2818,
"step": 14520
},
{
"epoch": 70.97010372178157,
"grad_norm": 3.572127103805542,
"learning_rate": 0.00010531147540983609,
"loss": 0.303,
"step": 14540
},
{
"epoch": 71.06772422208664,
"grad_norm": 3.4848947525024414,
"learning_rate": 0.00010518032786885246,
"loss": 0.2796,
"step": 14560
},
{
"epoch": 71.16534472239171,
"grad_norm": 3.478776454925537,
"learning_rate": 0.00010504918032786885,
"loss": 0.2718,
"step": 14580
},
{
"epoch": 71.26296522269676,
"grad_norm": 3.6878888607025146,
"learning_rate": 0.00010491803278688525,
"loss": 0.2611,
"step": 14600
},
{
"epoch": 71.36058572300183,
"grad_norm": 3.0455853939056396,
"learning_rate": 0.00010478688524590163,
"loss": 0.3063,
"step": 14620
},
{
"epoch": 71.45820622330689,
"grad_norm": 4.242178916931152,
"learning_rate": 0.00010465573770491803,
"loss": 0.2811,
"step": 14640
},
{
"epoch": 71.55582672361196,
"grad_norm": 2.9485058784484863,
"learning_rate": 0.00010452459016393444,
"loss": 0.278,
"step": 14660
},
{
"epoch": 71.65344722391703,
"grad_norm": 4.489981651306152,
"learning_rate": 0.00010439344262295083,
"loss": 0.2833,
"step": 14680
},
{
"epoch": 71.75106772422208,
"grad_norm": 3.748349666595459,
"learning_rate": 0.00010426229508196721,
"loss": 0.3362,
"step": 14700
},
{
"epoch": 71.84868822452715,
"grad_norm": 3.5393285751342773,
"learning_rate": 0.0001041311475409836,
"loss": 0.2803,
"step": 14720
},
{
"epoch": 71.94630872483222,
"grad_norm": 3.8631341457366943,
"learning_rate": 0.00010400000000000001,
"loss": 0.3103,
"step": 14740
},
{
"epoch": 72.04392922513728,
"grad_norm": 3.3096232414245605,
"learning_rate": 0.00010386885245901641,
"loss": 0.3059,
"step": 14760
},
{
"epoch": 72.14154972544235,
"grad_norm": 3.3514437675476074,
"learning_rate": 0.00010373770491803279,
"loss": 0.2832,
"step": 14780
},
{
"epoch": 72.2391702257474,
"grad_norm": 3.548172950744629,
"learning_rate": 0.00010360655737704919,
"loss": 0.3087,
"step": 14800
},
{
"epoch": 72.33679072605247,
"grad_norm": 2.9847452640533447,
"learning_rate": 0.00010347540983606558,
"loss": 0.2565,
"step": 14820
},
{
"epoch": 72.43441122635754,
"grad_norm": 2.827442169189453,
"learning_rate": 0.00010334426229508197,
"loss": 0.2622,
"step": 14840
},
{
"epoch": 72.5320317266626,
"grad_norm": 5.385785102844238,
"learning_rate": 0.00010321311475409836,
"loss": 0.2684,
"step": 14860
},
{
"epoch": 72.62965222696766,
"grad_norm": 2.8758130073547363,
"learning_rate": 0.00010308196721311476,
"loss": 0.3032,
"step": 14880
},
{
"epoch": 72.72727272727273,
"grad_norm": 3.0120010375976562,
"learning_rate": 0.00010295081967213116,
"loss": 0.2822,
"step": 14900
},
{
"epoch": 72.82489322757779,
"grad_norm": 4.078627109527588,
"learning_rate": 0.00010281967213114754,
"loss": 0.2852,
"step": 14920
},
{
"epoch": 72.92251372788286,
"grad_norm": 3.6309173107147217,
"learning_rate": 0.00010268852459016394,
"loss": 0.2961,
"step": 14940
},
{
"epoch": 73.02013422818791,
"grad_norm": 5.154388427734375,
"learning_rate": 0.00010255737704918033,
"loss": 0.2926,
"step": 14960
},
{
"epoch": 73.11775472849298,
"grad_norm": 3.6091132164001465,
"learning_rate": 0.00010242622950819673,
"loss": 0.2855,
"step": 14980
},
{
"epoch": 73.21537522879805,
"grad_norm": 3.872344970703125,
"learning_rate": 0.00010229508196721311,
"loss": 0.2728,
"step": 15000
},
{
"epoch": 73.31299572910311,
"grad_norm": 3.2357425689697266,
"learning_rate": 0.00010216393442622951,
"loss": 0.3022,
"step": 15020
},
{
"epoch": 73.41061622940818,
"grad_norm": 3.10544490814209,
"learning_rate": 0.00010203278688524592,
"loss": 0.2792,
"step": 15040
},
{
"epoch": 73.50823672971325,
"grad_norm": 3.796975612640381,
"learning_rate": 0.00010190163934426229,
"loss": 0.2975,
"step": 15060
},
{
"epoch": 73.6058572300183,
"grad_norm": 4.492702484130859,
"learning_rate": 0.0001017704918032787,
"loss": 0.2628,
"step": 15080
},
{
"epoch": 73.70347773032337,
"grad_norm": 3.9376978874206543,
"learning_rate": 0.00010163934426229508,
"loss": 0.2428,
"step": 15100
},
{
"epoch": 73.80109823062843,
"grad_norm": 4.3148980140686035,
"learning_rate": 0.00010150819672131149,
"loss": 0.3151,
"step": 15120
},
{
"epoch": 73.8987187309335,
"grad_norm": 3.7548837661743164,
"learning_rate": 0.00010137704918032786,
"loss": 0.2676,
"step": 15140
},
{
"epoch": 73.99633923123857,
"grad_norm": 3.9053032398223877,
"learning_rate": 0.00010124590163934427,
"loss": 0.298,
"step": 15160
},
{
"epoch": 74.09395973154362,
"grad_norm": 4.0978007316589355,
"learning_rate": 0.00010111475409836067,
"loss": 0.2597,
"step": 15180
},
{
"epoch": 74.19158023184869,
"grad_norm": 2.8652732372283936,
"learning_rate": 0.00010098360655737706,
"loss": 0.2842,
"step": 15200
},
{
"epoch": 74.28920073215376,
"grad_norm": 3.180609941482544,
"learning_rate": 0.00010085245901639345,
"loss": 0.2737,
"step": 15220
},
{
"epoch": 74.38682123245881,
"grad_norm": 4.571623802185059,
"learning_rate": 0.00010072131147540984,
"loss": 0.3213,
"step": 15240
},
{
"epoch": 74.48444173276388,
"grad_norm": 3.073421001434326,
"learning_rate": 0.00010059016393442624,
"loss": 0.2922,
"step": 15260
},
{
"epoch": 74.58206223306894,
"grad_norm": 5.4209980964660645,
"learning_rate": 0.00010045901639344261,
"loss": 0.2575,
"step": 15280
},
{
"epoch": 74.67968273337401,
"grad_norm": 4.5742621421813965,
"learning_rate": 0.00010032786885245902,
"loss": 0.3044,
"step": 15300
},
{
"epoch": 74.77730323367908,
"grad_norm": 3.712390422821045,
"learning_rate": 0.00010019672131147542,
"loss": 0.2891,
"step": 15320
},
{
"epoch": 74.87492373398413,
"grad_norm": 3.149919033050537,
"learning_rate": 0.00010006557377049181,
"loss": 0.275,
"step": 15340
},
{
"epoch": 74.9725442342892,
"grad_norm": 4.481144428253174,
"learning_rate": 9.99344262295082e-05,
"loss": 0.2544,
"step": 15360
},
{
"epoch": 75.07016473459427,
"grad_norm": 3.6176226139068604,
"learning_rate": 9.980327868852459e-05,
"loss": 0.2656,
"step": 15380
},
{
"epoch": 75.16778523489933,
"grad_norm": 3.986781120300293,
"learning_rate": 9.967213114754099e-05,
"loss": 0.2712,
"step": 15400
},
{
"epoch": 75.2654057352044,
"grad_norm": 2.7806153297424316,
"learning_rate": 9.954098360655738e-05,
"loss": 0.2542,
"step": 15420
},
{
"epoch": 75.36302623550945,
"grad_norm": 4.493511199951172,
"learning_rate": 9.940983606557378e-05,
"loss": 0.2871,
"step": 15440
},
{
"epoch": 75.46064673581452,
"grad_norm": 4.610682964324951,
"learning_rate": 9.927868852459017e-05,
"loss": 0.2699,
"step": 15460
},
{
"epoch": 75.55826723611959,
"grad_norm": 3.7209839820861816,
"learning_rate": 9.914754098360656e-05,
"loss": 0.3086,
"step": 15480
},
{
"epoch": 75.65588773642465,
"grad_norm": 4.1248931884765625,
"learning_rate": 9.901639344262295e-05,
"loss": 0.2744,
"step": 15500
},
{
"epoch": 75.75350823672972,
"grad_norm": 3.7576661109924316,
"learning_rate": 9.888524590163934e-05,
"loss": 0.2791,
"step": 15520
},
{
"epoch": 75.85112873703477,
"grad_norm": 3.6697514057159424,
"learning_rate": 9.875409836065574e-05,
"loss": 0.2779,
"step": 15540
},
{
"epoch": 75.94874923733984,
"grad_norm": 4.156905651092529,
"learning_rate": 9.862295081967213e-05,
"loss": 0.2806,
"step": 15560
},
{
"epoch": 76.04636973764491,
"grad_norm": 3.244154214859009,
"learning_rate": 9.849180327868854e-05,
"loss": 0.2772,
"step": 15580
},
{
"epoch": 76.14399023794996,
"grad_norm": 4.677572250366211,
"learning_rate": 9.836065573770493e-05,
"loss": 0.2695,
"step": 15600
},
{
"epoch": 76.24161073825503,
"grad_norm": 3.442591667175293,
"learning_rate": 9.822950819672132e-05,
"loss": 0.309,
"step": 15620
},
{
"epoch": 76.3392312385601,
"grad_norm": 2.930734395980835,
"learning_rate": 9.80983606557377e-05,
"loss": 0.2543,
"step": 15640
},
{
"epoch": 76.43685173886516,
"grad_norm": 3.7919161319732666,
"learning_rate": 9.796721311475411e-05,
"loss": 0.2802,
"step": 15660
},
{
"epoch": 76.53447223917023,
"grad_norm": 3.598411798477173,
"learning_rate": 9.78360655737705e-05,
"loss": 0.2433,
"step": 15680
},
{
"epoch": 76.63209273947528,
"grad_norm": 3.6771163940429688,
"learning_rate": 9.770491803278689e-05,
"loss": 0.2529,
"step": 15700
},
{
"epoch": 76.72971323978035,
"grad_norm": 3.1646664142608643,
"learning_rate": 9.757377049180329e-05,
"loss": 0.2801,
"step": 15720
},
{
"epoch": 76.82733374008542,
"grad_norm": 2.325819969177246,
"learning_rate": 9.744262295081968e-05,
"loss": 0.2609,
"step": 15740
},
{
"epoch": 76.92495424039048,
"grad_norm": 4.598663330078125,
"learning_rate": 9.731147540983607e-05,
"loss": 0.2936,
"step": 15760
},
{
"epoch": 77.02257474069555,
"grad_norm": 2.9897961616516113,
"learning_rate": 9.718032786885246e-05,
"loss": 0.2891,
"step": 15780
},
{
"epoch": 77.12019524100062,
"grad_norm": 4.264781951904297,
"learning_rate": 9.704918032786886e-05,
"loss": 0.2677,
"step": 15800
},
{
"epoch": 77.21781574130567,
"grad_norm": 3.616536855697632,
"learning_rate": 9.691803278688525e-05,
"loss": 0.2282,
"step": 15820
},
{
"epoch": 77.31543624161074,
"grad_norm": 3.4226481914520264,
"learning_rate": 9.678688524590165e-05,
"loss": 0.2792,
"step": 15840
},
{
"epoch": 77.4130567419158,
"grad_norm": 3.012357473373413,
"learning_rate": 9.665573770491804e-05,
"loss": 0.2644,
"step": 15860
},
{
"epoch": 77.51067724222086,
"grad_norm": 3.582298517227173,
"learning_rate": 9.652459016393443e-05,
"loss": 0.2944,
"step": 15880
},
{
"epoch": 77.60829774252593,
"grad_norm": 4.138154983520508,
"learning_rate": 9.639344262295082e-05,
"loss": 0.3097,
"step": 15900
},
{
"epoch": 77.70591824283099,
"grad_norm": 4.919800281524658,
"learning_rate": 9.626229508196721e-05,
"loss": 0.2666,
"step": 15920
},
{
"epoch": 77.80353874313606,
"grad_norm": 3.3038241863250732,
"learning_rate": 9.613114754098361e-05,
"loss": 0.2909,
"step": 15940
},
{
"epoch": 77.90115924344113,
"grad_norm": 2.6817522048950195,
"learning_rate": 9.6e-05,
"loss": 0.2565,
"step": 15960
},
{
"epoch": 77.99877974374618,
"grad_norm": 4.992214679718018,
"learning_rate": 9.58688524590164e-05,
"loss": 0.2767,
"step": 15980
},
{
"epoch": 78.09640024405125,
"grad_norm": 3.589238166809082,
"learning_rate": 9.57377049180328e-05,
"loss": 0.2483,
"step": 16000
},
{
"epoch": 78.19402074435631,
"grad_norm": 4.856943607330322,
"learning_rate": 9.560655737704918e-05,
"loss": 0.2698,
"step": 16020
},
{
"epoch": 78.29164124466138,
"grad_norm": 3.963756561279297,
"learning_rate": 9.547540983606557e-05,
"loss": 0.2522,
"step": 16040
},
{
"epoch": 78.38926174496645,
"grad_norm": 3.843501567840576,
"learning_rate": 9.534426229508198e-05,
"loss": 0.2502,
"step": 16060
},
{
"epoch": 78.4868822452715,
"grad_norm": 3.1323421001434326,
"learning_rate": 9.521311475409837e-05,
"loss": 0.2693,
"step": 16080
},
{
"epoch": 78.58450274557657,
"grad_norm": 3.3395233154296875,
"learning_rate": 9.508196721311476e-05,
"loss": 0.2806,
"step": 16100
},
{
"epoch": 78.68212324588164,
"grad_norm": 3.2387075424194336,
"learning_rate": 9.495081967213116e-05,
"loss": 0.266,
"step": 16120
},
{
"epoch": 78.7797437461867,
"grad_norm": 3.9725232124328613,
"learning_rate": 9.481967213114755e-05,
"loss": 0.2753,
"step": 16140
},
{
"epoch": 78.87736424649177,
"grad_norm": 3.88059401512146,
"learning_rate": 9.468852459016394e-05,
"loss": 0.3028,
"step": 16160
},
{
"epoch": 78.97498474679682,
"grad_norm": 4.623359680175781,
"learning_rate": 9.455737704918033e-05,
"loss": 0.2816,
"step": 16180
},
{
"epoch": 79.07260524710189,
"grad_norm": 3.237917423248291,
"learning_rate": 9.442622950819673e-05,
"loss": 0.2545,
"step": 16200
},
{
"epoch": 79.17022574740696,
"grad_norm": 3.4442942142486572,
"learning_rate": 9.429508196721312e-05,
"loss": 0.2596,
"step": 16220
},
{
"epoch": 79.26784624771201,
"grad_norm": 4.398017406463623,
"learning_rate": 9.416393442622952e-05,
"loss": 0.2731,
"step": 16240
},
{
"epoch": 79.36546674801708,
"grad_norm": 3.892063856124878,
"learning_rate": 9.403278688524591e-05,
"loss": 0.266,
"step": 16260
},
{
"epoch": 79.46308724832215,
"grad_norm": 4.046624183654785,
"learning_rate": 9.39016393442623e-05,
"loss": 0.2368,
"step": 16280
},
{
"epoch": 79.56070774862721,
"grad_norm": 5.123999118804932,
"learning_rate": 9.377049180327869e-05,
"loss": 0.2409,
"step": 16300
},
{
"epoch": 79.65832824893228,
"grad_norm": 3.1601643562316895,
"learning_rate": 9.363934426229508e-05,
"loss": 0.2525,
"step": 16320
},
{
"epoch": 79.75594874923733,
"grad_norm": 4.788114070892334,
"learning_rate": 9.350819672131148e-05,
"loss": 0.2998,
"step": 16340
},
{
"epoch": 79.8535692495424,
"grad_norm": 3.9707436561584473,
"learning_rate": 9.337704918032787e-05,
"loss": 0.2724,
"step": 16360
},
{
"epoch": 79.95118974984747,
"grad_norm": 3.7388081550598145,
"learning_rate": 9.324590163934427e-05,
"loss": 0.2823,
"step": 16380
},
{
"epoch": 80.04881025015253,
"grad_norm": 2.7105419635772705,
"learning_rate": 9.311475409836066e-05,
"loss": 0.2602,
"step": 16400
},
{
"epoch": 80.1464307504576,
"grad_norm": 3.460477113723755,
"learning_rate": 9.298360655737705e-05,
"loss": 0.2428,
"step": 16420
},
{
"epoch": 80.24405125076267,
"grad_norm": 3.526585817337036,
"learning_rate": 9.285245901639344e-05,
"loss": 0.2982,
"step": 16440
},
{
"epoch": 80.34167175106772,
"grad_norm": 3.746425151824951,
"learning_rate": 9.272131147540985e-05,
"loss": 0.2365,
"step": 16460
},
{
"epoch": 80.43929225137279,
"grad_norm": 3.2804489135742188,
"learning_rate": 9.259016393442623e-05,
"loss": 0.2832,
"step": 16480
},
{
"epoch": 80.53691275167785,
"grad_norm": 3.136016607284546,
"learning_rate": 9.245901639344264e-05,
"loss": 0.2744,
"step": 16500
},
{
"epoch": 80.63453325198292,
"grad_norm": 4.361492156982422,
"learning_rate": 9.232786885245903e-05,
"loss": 0.2726,
"step": 16520
},
{
"epoch": 80.73215375228799,
"grad_norm": 3.3838353157043457,
"learning_rate": 9.21967213114754e-05,
"loss": 0.2651,
"step": 16540
},
{
"epoch": 80.82977425259304,
"grad_norm": 3.8092405796051025,
"learning_rate": 9.20655737704918e-05,
"loss": 0.2519,
"step": 16560
},
{
"epoch": 80.92739475289811,
"grad_norm": 4.455862522125244,
"learning_rate": 9.19344262295082e-05,
"loss": 0.2518,
"step": 16580
},
{
"epoch": 81.02501525320318,
"grad_norm": 3.073539972305298,
"learning_rate": 9.18032786885246e-05,
"loss": 0.272,
"step": 16600
},
{
"epoch": 81.12263575350823,
"grad_norm": 4.361839294433594,
"learning_rate": 9.167213114754099e-05,
"loss": 0.2468,
"step": 16620
},
{
"epoch": 81.2202562538133,
"grad_norm": 5.322413444519043,
"learning_rate": 9.154098360655739e-05,
"loss": 0.2339,
"step": 16640
},
{
"epoch": 81.31787675411836,
"grad_norm": 4.6976704597473145,
"learning_rate": 9.140983606557378e-05,
"loss": 0.27,
"step": 16660
},
{
"epoch": 81.41549725442343,
"grad_norm": 3.608119487762451,
"learning_rate": 9.127868852459017e-05,
"loss": 0.2615,
"step": 16680
},
{
"epoch": 81.5131177547285,
"grad_norm": 3.837738513946533,
"learning_rate": 9.114754098360656e-05,
"loss": 0.2534,
"step": 16700
},
{
"epoch": 81.61073825503355,
"grad_norm": 3.750638723373413,
"learning_rate": 9.101639344262296e-05,
"loss": 0.2533,
"step": 16720
},
{
"epoch": 81.70835875533862,
"grad_norm": 3.916907548904419,
"learning_rate": 9.088524590163935e-05,
"loss": 0.2847,
"step": 16740
},
{
"epoch": 81.80597925564369,
"grad_norm": 3.8545727729797363,
"learning_rate": 9.075409836065574e-05,
"loss": 0.2554,
"step": 16760
},
{
"epoch": 81.90359975594875,
"grad_norm": 2.8878276348114014,
"learning_rate": 9.062295081967214e-05,
"loss": 0.2793,
"step": 16780
},
{
"epoch": 82.00122025625382,
"grad_norm": 3.6857919692993164,
"learning_rate": 9.049180327868852e-05,
"loss": 0.2703,
"step": 16800
},
{
"epoch": 82.09884075655887,
"grad_norm": 3.670179605484009,
"learning_rate": 9.036065573770492e-05,
"loss": 0.2641,
"step": 16820
},
{
"epoch": 82.19646125686394,
"grad_norm": 3.4155802726745605,
"learning_rate": 9.022950819672131e-05,
"loss": 0.2528,
"step": 16840
},
{
"epoch": 82.29408175716901,
"grad_norm": 3.139631986618042,
"learning_rate": 9.009836065573771e-05,
"loss": 0.2485,
"step": 16860
},
{
"epoch": 82.39170225747407,
"grad_norm": 3.919480800628662,
"learning_rate": 8.99672131147541e-05,
"loss": 0.2586,
"step": 16880
},
{
"epoch": 82.48932275777914,
"grad_norm": 4.068678379058838,
"learning_rate": 8.98360655737705e-05,
"loss": 0.2836,
"step": 16900
},
{
"epoch": 82.5869432580842,
"grad_norm": 3.544682502746582,
"learning_rate": 8.97049180327869e-05,
"loss": 0.2515,
"step": 16920
},
{
"epoch": 82.68456375838926,
"grad_norm": 3.8421828746795654,
"learning_rate": 8.957377049180328e-05,
"loss": 0.2654,
"step": 16940
},
{
"epoch": 82.78218425869433,
"grad_norm": 3.0508570671081543,
"learning_rate": 8.944262295081967e-05,
"loss": 0.2742,
"step": 16960
},
{
"epoch": 82.87980475899938,
"grad_norm": 3.601579427719116,
"learning_rate": 8.931147540983606e-05,
"loss": 0.2542,
"step": 16980
},
{
"epoch": 82.97742525930445,
"grad_norm": 3.657724618911743,
"learning_rate": 8.918032786885247e-05,
"loss": 0.2253,
"step": 17000
},
{
"epoch": 83.07504575960952,
"grad_norm": 5.093318462371826,
"learning_rate": 8.904918032786886e-05,
"loss": 0.2429,
"step": 17020
},
{
"epoch": 83.17266625991458,
"grad_norm": 3.160510301589966,
"learning_rate": 8.891803278688526e-05,
"loss": 0.2416,
"step": 17040
},
{
"epoch": 83.27028676021965,
"grad_norm": 3.9046666622161865,
"learning_rate": 8.878688524590163e-05,
"loss": 0.2526,
"step": 17060
},
{
"epoch": 83.36790726052472,
"grad_norm": 4.196876525878906,
"learning_rate": 8.865573770491804e-05,
"loss": 0.2518,
"step": 17080
},
{
"epoch": 83.46552776082977,
"grad_norm": 3.5083820819854736,
"learning_rate": 8.852459016393443e-05,
"loss": 0.2392,
"step": 17100
},
{
"epoch": 83.56314826113484,
"grad_norm": 3.099393606185913,
"learning_rate": 8.839344262295083e-05,
"loss": 0.2783,
"step": 17120
},
{
"epoch": 83.6607687614399,
"grad_norm": 3.532540798187256,
"learning_rate": 8.826229508196722e-05,
"loss": 0.261,
"step": 17140
},
{
"epoch": 83.75838926174497,
"grad_norm": 3.2548046112060547,
"learning_rate": 8.813114754098362e-05,
"loss": 0.2622,
"step": 17160
},
{
"epoch": 83.85600976205004,
"grad_norm": 4.520061492919922,
"learning_rate": 8.800000000000001e-05,
"loss": 0.272,
"step": 17180
},
{
"epoch": 83.95363026235509,
"grad_norm": 3.1347196102142334,
"learning_rate": 8.786885245901639e-05,
"loss": 0.2504,
"step": 17200
},
{
"epoch": 84.05125076266016,
"grad_norm": 3.6954307556152344,
"learning_rate": 8.773770491803279e-05,
"loss": 0.2566,
"step": 17220
},
{
"epoch": 84.14887126296523,
"grad_norm": 4.145720481872559,
"learning_rate": 8.760655737704918e-05,
"loss": 0.2393,
"step": 17240
},
{
"epoch": 84.24649176327028,
"grad_norm": 3.575308084487915,
"learning_rate": 8.747540983606558e-05,
"loss": 0.2667,
"step": 17260
},
{
"epoch": 84.34411226357535,
"grad_norm": 4.094547271728516,
"learning_rate": 8.734426229508197e-05,
"loss": 0.244,
"step": 17280
},
{
"epoch": 84.44173276388041,
"grad_norm": 3.583008050918579,
"learning_rate": 8.721311475409837e-05,
"loss": 0.2262,
"step": 17300
},
{
"epoch": 84.53935326418548,
"grad_norm": 4.159909248352051,
"learning_rate": 8.708196721311475e-05,
"loss": 0.2625,
"step": 17320
},
{
"epoch": 84.63697376449055,
"grad_norm": 3.5979392528533936,
"learning_rate": 8.695081967213115e-05,
"loss": 0.2421,
"step": 17340
},
{
"epoch": 84.7345942647956,
"grad_norm": 2.0888795852661133,
"learning_rate": 8.681967213114754e-05,
"loss": 0.2555,
"step": 17360
},
{
"epoch": 84.83221476510067,
"grad_norm": 3.8699913024902344,
"learning_rate": 8.668852459016393e-05,
"loss": 0.2733,
"step": 17380
},
{
"epoch": 84.92983526540573,
"grad_norm": 3.1710896492004395,
"learning_rate": 8.655737704918033e-05,
"loss": 0.2824,
"step": 17400
},
{
"epoch": 85.0274557657108,
"grad_norm": 3.3995399475097656,
"learning_rate": 8.642622950819672e-05,
"loss": 0.2555,
"step": 17420
},
{
"epoch": 85.12507626601587,
"grad_norm": 4.362491130828857,
"learning_rate": 8.629508196721313e-05,
"loss": 0.2535,
"step": 17440
},
{
"epoch": 85.22269676632092,
"grad_norm": 3.141589641571045,
"learning_rate": 8.61639344262295e-05,
"loss": 0.2368,
"step": 17460
},
{
"epoch": 85.32031726662599,
"grad_norm": 2.7880699634552,
"learning_rate": 8.60327868852459e-05,
"loss": 0.2382,
"step": 17480
},
{
"epoch": 85.41793776693106,
"grad_norm": 3.9995205402374268,
"learning_rate": 8.59016393442623e-05,
"loss": 0.2484,
"step": 17500
},
{
"epoch": 85.51555826723612,
"grad_norm": 4.39112663269043,
"learning_rate": 8.57704918032787e-05,
"loss": 0.254,
"step": 17520
},
{
"epoch": 85.61317876754119,
"grad_norm": 3.7383909225463867,
"learning_rate": 8.563934426229509e-05,
"loss": 0.2685,
"step": 17540
},
{
"epoch": 85.71079926784624,
"grad_norm": 2.9596993923187256,
"learning_rate": 8.550819672131149e-05,
"loss": 0.2725,
"step": 17560
},
{
"epoch": 85.80841976815131,
"grad_norm": 3.4363205432891846,
"learning_rate": 8.537704918032787e-05,
"loss": 0.2518,
"step": 17580
},
{
"epoch": 85.90604026845638,
"grad_norm": 4.1404709815979,
"learning_rate": 8.524590163934426e-05,
"loss": 0.2409,
"step": 17600
},
{
"epoch": 86.00366076876143,
"grad_norm": 3.080606460571289,
"learning_rate": 8.511475409836066e-05,
"loss": 0.2646,
"step": 17620
},
{
"epoch": 86.1012812690665,
"grad_norm": 4.42800760269165,
"learning_rate": 8.498360655737705e-05,
"loss": 0.2274,
"step": 17640
},
{
"epoch": 86.19890176937157,
"grad_norm": 3.3812572956085205,
"learning_rate": 8.485245901639345e-05,
"loss": 0.2165,
"step": 17660
},
{
"epoch": 86.29652226967663,
"grad_norm": 3.873788833618164,
"learning_rate": 8.472131147540984e-05,
"loss": 0.2464,
"step": 17680
},
{
"epoch": 86.3941427699817,
"grad_norm": 4.275656700134277,
"learning_rate": 8.459016393442624e-05,
"loss": 0.2236,
"step": 17700
},
{
"epoch": 86.49176327028675,
"grad_norm": 3.0242347717285156,
"learning_rate": 8.445901639344262e-05,
"loss": 0.2518,
"step": 17720
},
{
"epoch": 86.58938377059182,
"grad_norm": 3.052992105484009,
"learning_rate": 8.432786885245902e-05,
"loss": 0.2659,
"step": 17740
},
{
"epoch": 86.68700427089689,
"grad_norm": 5.510742664337158,
"learning_rate": 8.419672131147541e-05,
"loss": 0.269,
"step": 17760
},
{
"epoch": 86.78462477120195,
"grad_norm": 3.8788983821868896,
"learning_rate": 8.406557377049181e-05,
"loss": 0.2548,
"step": 17780
},
{
"epoch": 86.88224527150702,
"grad_norm": 4.494417190551758,
"learning_rate": 8.39344262295082e-05,
"loss": 0.2558,
"step": 17800
},
{
"epoch": 86.97986577181209,
"grad_norm": 4.6547040939331055,
"learning_rate": 8.380327868852459e-05,
"loss": 0.2775,
"step": 17820
},
{
"epoch": 87.07748627211714,
"grad_norm": 3.511051654815674,
"learning_rate": 8.367213114754098e-05,
"loss": 0.2233,
"step": 17840
},
{
"epoch": 87.17510677242221,
"grad_norm": 4.116011142730713,
"learning_rate": 8.354098360655737e-05,
"loss": 0.2246,
"step": 17860
},
{
"epoch": 87.27272727272727,
"grad_norm": 3.525118589401245,
"learning_rate": 8.340983606557377e-05,
"loss": 0.2058,
"step": 17880
},
{
"epoch": 87.37034777303234,
"grad_norm": 3.5380094051361084,
"learning_rate": 8.327868852459016e-05,
"loss": 0.2467,
"step": 17900
},
{
"epoch": 87.4679682733374,
"grad_norm": 3.6875052452087402,
"learning_rate": 8.314754098360657e-05,
"loss": 0.2967,
"step": 17920
},
{
"epoch": 87.56558877364246,
"grad_norm": 3.565765142440796,
"learning_rate": 8.301639344262296e-05,
"loss": 0.2434,
"step": 17940
},
{
"epoch": 87.66320927394753,
"grad_norm": 3.3109848499298096,
"learning_rate": 8.288524590163935e-05,
"loss": 0.2538,
"step": 17960
},
{
"epoch": 87.7608297742526,
"grad_norm": 3.278052568435669,
"learning_rate": 8.275409836065573e-05,
"loss": 0.2739,
"step": 17980
},
{
"epoch": 87.85845027455765,
"grad_norm": 5.82271671295166,
"learning_rate": 8.262295081967214e-05,
"loss": 0.244,
"step": 18000
},
{
"epoch": 87.95607077486272,
"grad_norm": 4.266513824462891,
"learning_rate": 8.249180327868853e-05,
"loss": 0.2546,
"step": 18020
},
{
"epoch": 88.05369127516778,
"grad_norm": 3.985555648803711,
"learning_rate": 8.236065573770492e-05,
"loss": 0.2673,
"step": 18040
},
{
"epoch": 88.15131177547285,
"grad_norm": 3.4252424240112305,
"learning_rate": 8.222950819672132e-05,
"loss": 0.2308,
"step": 18060
},
{
"epoch": 88.24893227577792,
"grad_norm": 3.1381208896636963,
"learning_rate": 8.209836065573771e-05,
"loss": 0.24,
"step": 18080
},
{
"epoch": 88.34655277608297,
"grad_norm": 4.053617000579834,
"learning_rate": 8.19672131147541e-05,
"loss": 0.2646,
"step": 18100
},
{
"epoch": 88.44417327638804,
"grad_norm": 3.525423288345337,
"learning_rate": 8.183606557377049e-05,
"loss": 0.2535,
"step": 18120
},
{
"epoch": 88.54179377669311,
"grad_norm": 3.593766212463379,
"learning_rate": 8.170491803278689e-05,
"loss": 0.2426,
"step": 18140
},
{
"epoch": 88.63941427699817,
"grad_norm": 4.087725639343262,
"learning_rate": 8.157377049180328e-05,
"loss": 0.2196,
"step": 18160
},
{
"epoch": 88.73703477730324,
"grad_norm": 4.209127902984619,
"learning_rate": 8.144262295081968e-05,
"loss": 0.2593,
"step": 18180
},
{
"epoch": 88.83465527760829,
"grad_norm": 3.9146687984466553,
"learning_rate": 8.131147540983607e-05,
"loss": 0.2492,
"step": 18200
},
{
"epoch": 88.93227577791336,
"grad_norm": 3.4198620319366455,
"learning_rate": 8.118032786885246e-05,
"loss": 0.2569,
"step": 18220
},
{
"epoch": 89.02989627821843,
"grad_norm": 3.5384342670440674,
"learning_rate": 8.104918032786885e-05,
"loss": 0.2298,
"step": 18240
},
{
"epoch": 89.12751677852349,
"grad_norm": 3.252002239227295,
"learning_rate": 8.091803278688524e-05,
"loss": 0.2296,
"step": 18260
},
{
"epoch": 89.22513727882856,
"grad_norm": 3.5200119018554688,
"learning_rate": 8.078688524590164e-05,
"loss": 0.2476,
"step": 18280
},
{
"epoch": 89.32275777913362,
"grad_norm": 4.229335784912109,
"learning_rate": 8.065573770491803e-05,
"loss": 0.2427,
"step": 18300
},
{
"epoch": 89.42037827943868,
"grad_norm": 3.418846368789673,
"learning_rate": 8.052459016393444e-05,
"loss": 0.248,
"step": 18320
},
{
"epoch": 89.51799877974375,
"grad_norm": 4.674842357635498,
"learning_rate": 8.039344262295082e-05,
"loss": 0.2388,
"step": 18340
},
{
"epoch": 89.6156192800488,
"grad_norm": 2.8320224285125732,
"learning_rate": 8.026229508196721e-05,
"loss": 0.2737,
"step": 18360
},
{
"epoch": 89.71323978035387,
"grad_norm": 3.764437198638916,
"learning_rate": 8.01311475409836e-05,
"loss": 0.2217,
"step": 18380
},
{
"epoch": 89.81086028065894,
"grad_norm": 3.7620530128479004,
"learning_rate": 8e-05,
"loss": 0.2339,
"step": 18400
},
{
"epoch": 89.908480780964,
"grad_norm": 3.451263904571533,
"learning_rate": 7.98688524590164e-05,
"loss": 0.239,
"step": 18420
},
{
"epoch": 90.00610128126907,
"grad_norm": 3.327611207962036,
"learning_rate": 7.97377049180328e-05,
"loss": 0.2509,
"step": 18440
},
{
"epoch": 90.10372178157414,
"grad_norm": 3.0182056427001953,
"learning_rate": 7.960655737704919e-05,
"loss": 0.2173,
"step": 18460
},
{
"epoch": 90.20134228187919,
"grad_norm": 4.298709869384766,
"learning_rate": 7.947540983606558e-05,
"loss": 0.2371,
"step": 18480
},
{
"epoch": 90.29896278218426,
"grad_norm": 6.2249979972839355,
"learning_rate": 7.934426229508197e-05,
"loss": 0.2504,
"step": 18500
},
{
"epoch": 90.39658328248932,
"grad_norm": 3.3702869415283203,
"learning_rate": 7.921311475409836e-05,
"loss": 0.2264,
"step": 18520
},
{
"epoch": 90.49420378279439,
"grad_norm": 4.575166702270508,
"learning_rate": 7.908196721311476e-05,
"loss": 0.2296,
"step": 18540
},
{
"epoch": 90.59182428309946,
"grad_norm": 3.4049549102783203,
"learning_rate": 7.895081967213115e-05,
"loss": 0.2488,
"step": 18560
},
{
"epoch": 90.68944478340451,
"grad_norm": 3.5313448905944824,
"learning_rate": 7.881967213114755e-05,
"loss": 0.2414,
"step": 18580
},
{
"epoch": 90.78706528370958,
"grad_norm": 3.3291146755218506,
"learning_rate": 7.868852459016394e-05,
"loss": 0.225,
"step": 18600
},
{
"epoch": 90.88468578401465,
"grad_norm": 2.99589204788208,
"learning_rate": 7.855737704918033e-05,
"loss": 0.2548,
"step": 18620
},
{
"epoch": 90.9823062843197,
"grad_norm": 3.2335972785949707,
"learning_rate": 7.842622950819672e-05,
"loss": 0.269,
"step": 18640
},
{
"epoch": 91.07992678462477,
"grad_norm": 4.912237167358398,
"learning_rate": 7.829508196721311e-05,
"loss": 0.2392,
"step": 18660
},
{
"epoch": 91.17754728492983,
"grad_norm": 4.516569137573242,
"learning_rate": 7.816393442622951e-05,
"loss": 0.2213,
"step": 18680
},
{
"epoch": 91.2751677852349,
"grad_norm": 3.225470542907715,
"learning_rate": 7.80327868852459e-05,
"loss": 0.2414,
"step": 18700
},
{
"epoch": 91.37278828553997,
"grad_norm": 3.231811761856079,
"learning_rate": 7.79016393442623e-05,
"loss": 0.2384,
"step": 18720
},
{
"epoch": 91.47040878584502,
"grad_norm": 3.199504852294922,
"learning_rate": 7.77704918032787e-05,
"loss": 0.2336,
"step": 18740
},
{
"epoch": 91.5680292861501,
"grad_norm": 3.4679319858551025,
"learning_rate": 7.763934426229508e-05,
"loss": 0.2136,
"step": 18760
},
{
"epoch": 91.66564978645516,
"grad_norm": 4.474179267883301,
"learning_rate": 7.750819672131147e-05,
"loss": 0.2454,
"step": 18780
},
{
"epoch": 91.76327028676022,
"grad_norm": 4.39286994934082,
"learning_rate": 7.737704918032788e-05,
"loss": 0.2377,
"step": 18800
},
{
"epoch": 91.86089078706529,
"grad_norm": 2.924795627593994,
"learning_rate": 7.724590163934426e-05,
"loss": 0.2475,
"step": 18820
},
{
"epoch": 91.95851128737034,
"grad_norm": 3.4778449535369873,
"learning_rate": 7.711475409836067e-05,
"loss": 0.2712,
"step": 18840
},
{
"epoch": 92.05613178767541,
"grad_norm": 5.152170181274414,
"learning_rate": 7.698360655737706e-05,
"loss": 0.2336,
"step": 18860
},
{
"epoch": 92.15375228798048,
"grad_norm": 2.8238580226898193,
"learning_rate": 7.685245901639345e-05,
"loss": 0.2504,
"step": 18880
},
{
"epoch": 92.25137278828554,
"grad_norm": 3.870866537094116,
"learning_rate": 7.672131147540984e-05,
"loss": 0.25,
"step": 18900
},
{
"epoch": 92.3489932885906,
"grad_norm": 5.052824020385742,
"learning_rate": 7.659016393442622e-05,
"loss": 0.2517,
"step": 18920
},
{
"epoch": 92.44661378889568,
"grad_norm": 3.048907518386841,
"learning_rate": 7.645901639344263e-05,
"loss": 0.2347,
"step": 18940
},
{
"epoch": 92.54423428920073,
"grad_norm": 3.2473690509796143,
"learning_rate": 7.632786885245902e-05,
"loss": 0.2362,
"step": 18960
},
{
"epoch": 92.6418547895058,
"grad_norm": 4.635523319244385,
"learning_rate": 7.619672131147542e-05,
"loss": 0.2274,
"step": 18980
},
{
"epoch": 92.73947528981085,
"grad_norm": 2.6211893558502197,
"learning_rate": 7.606557377049181e-05,
"loss": 0.2258,
"step": 19000
},
{
"epoch": 92.83709579011592,
"grad_norm": 3.725900888442993,
"learning_rate": 7.59344262295082e-05,
"loss": 0.2407,
"step": 19020
},
{
"epoch": 92.934716290421,
"grad_norm": 4.358538627624512,
"learning_rate": 7.580327868852459e-05,
"loss": 0.2268,
"step": 19040
},
{
"epoch": 93.03233679072605,
"grad_norm": 3.608492851257324,
"learning_rate": 7.567213114754099e-05,
"loss": 0.2114,
"step": 19060
},
{
"epoch": 93.12995729103112,
"grad_norm": 3.455113410949707,
"learning_rate": 7.554098360655738e-05,
"loss": 0.2194,
"step": 19080
},
{
"epoch": 93.22757779133617,
"grad_norm": 3.443164825439453,
"learning_rate": 7.540983606557377e-05,
"loss": 0.2305,
"step": 19100
},
{
"epoch": 93.32519829164124,
"grad_norm": 3.579775810241699,
"learning_rate": 7.527868852459017e-05,
"loss": 0.2358,
"step": 19120
},
{
"epoch": 93.42281879194631,
"grad_norm": 2.8307318687438965,
"learning_rate": 7.514754098360656e-05,
"loss": 0.2289,
"step": 19140
},
{
"epoch": 93.52043929225137,
"grad_norm": 4.2809834480285645,
"learning_rate": 7.501639344262295e-05,
"loss": 0.2246,
"step": 19160
},
{
"epoch": 93.61805979255644,
"grad_norm": 4.093430042266846,
"learning_rate": 7.488524590163934e-05,
"loss": 0.2416,
"step": 19180
},
{
"epoch": 93.7156802928615,
"grad_norm": 3.562998056411743,
"learning_rate": 7.475409836065574e-05,
"loss": 0.2453,
"step": 19200
},
{
"epoch": 93.81330079316656,
"grad_norm": 3.611504554748535,
"learning_rate": 7.462295081967213e-05,
"loss": 0.2389,
"step": 19220
},
{
"epoch": 93.91092129347163,
"grad_norm": 3.638408899307251,
"learning_rate": 7.449180327868854e-05,
"loss": 0.2354,
"step": 19240
},
{
"epoch": 94.00854179377669,
"grad_norm": 2.646662473678589,
"learning_rate": 7.436065573770493e-05,
"loss": 0.2588,
"step": 19260
},
{
"epoch": 94.10616229408176,
"grad_norm": 2.8512191772460938,
"learning_rate": 7.422950819672131e-05,
"loss": 0.2219,
"step": 19280
},
{
"epoch": 94.20378279438683,
"grad_norm": 4.569155693054199,
"learning_rate": 7.40983606557377e-05,
"loss": 0.2151,
"step": 19300
},
{
"epoch": 94.30140329469188,
"grad_norm": 3.036120653152466,
"learning_rate": 7.39672131147541e-05,
"loss": 0.2133,
"step": 19320
},
{
"epoch": 94.39902379499695,
"grad_norm": 3.7911667823791504,
"learning_rate": 7.38360655737705e-05,
"loss": 0.206,
"step": 19340
},
{
"epoch": 94.49664429530202,
"grad_norm": 3.4659411907196045,
"learning_rate": 7.370491803278689e-05,
"loss": 0.2175,
"step": 19360
},
{
"epoch": 94.59426479560707,
"grad_norm": 3.7196578979492188,
"learning_rate": 7.357377049180329e-05,
"loss": 0.2508,
"step": 19380
},
{
"epoch": 94.69188529591214,
"grad_norm": 3.499638557434082,
"learning_rate": 7.344262295081968e-05,
"loss": 0.2595,
"step": 19400
},
{
"epoch": 94.7895057962172,
"grad_norm": 3.3197853565216064,
"learning_rate": 7.331147540983607e-05,
"loss": 0.2498,
"step": 19420
},
{
"epoch": 94.88712629652227,
"grad_norm": 3.819153308868408,
"learning_rate": 7.318032786885246e-05,
"loss": 0.24,
"step": 19440
},
{
"epoch": 94.98474679682734,
"grad_norm": 3.229252815246582,
"learning_rate": 7.304918032786886e-05,
"loss": 0.2402,
"step": 19460
},
{
"epoch": 95.0823672971324,
"grad_norm": 3.8045654296875,
"learning_rate": 7.291803278688525e-05,
"loss": 0.2235,
"step": 19480
},
{
"epoch": 95.17998779743746,
"grad_norm": 3.8064322471618652,
"learning_rate": 7.278688524590165e-05,
"loss": 0.2374,
"step": 19500
},
{
"epoch": 95.27760829774253,
"grad_norm": 2.9170637130737305,
"learning_rate": 7.265573770491804e-05,
"loss": 0.223,
"step": 19520
},
{
"epoch": 95.37522879804759,
"grad_norm": 3.6795055866241455,
"learning_rate": 7.252459016393443e-05,
"loss": 0.2256,
"step": 19540
},
{
"epoch": 95.47284929835266,
"grad_norm": 3.1766109466552734,
"learning_rate": 7.239344262295082e-05,
"loss": 0.2536,
"step": 19560
},
{
"epoch": 95.57046979865771,
"grad_norm": 3.862264633178711,
"learning_rate": 7.226229508196721e-05,
"loss": 0.2339,
"step": 19580
},
{
"epoch": 95.66809029896278,
"grad_norm": 4.323207855224609,
"learning_rate": 7.213114754098361e-05,
"loss": 0.2434,
"step": 19600
},
{
"epoch": 95.76571079926785,
"grad_norm": 2.973966598510742,
"learning_rate": 7.2e-05,
"loss": 0.219,
"step": 19620
},
{
"epoch": 95.8633312995729,
"grad_norm": 3.5967094898223877,
"learning_rate": 7.18688524590164e-05,
"loss": 0.243,
"step": 19640
},
{
"epoch": 95.96095179987798,
"grad_norm": 3.993166923522949,
"learning_rate": 7.17377049180328e-05,
"loss": 0.2263,
"step": 19660
},
{
"epoch": 96.05857230018304,
"grad_norm": 3.425034284591675,
"learning_rate": 7.160655737704918e-05,
"loss": 0.2158,
"step": 19680
},
{
"epoch": 96.1561928004881,
"grad_norm": 3.4514410495758057,
"learning_rate": 7.147540983606557e-05,
"loss": 0.1976,
"step": 19700
},
{
"epoch": 96.25381330079317,
"grad_norm": 5.355749607086182,
"learning_rate": 7.134426229508198e-05,
"loss": 0.221,
"step": 19720
},
{
"epoch": 96.35143380109822,
"grad_norm": 3.296389102935791,
"learning_rate": 7.121311475409837e-05,
"loss": 0.2606,
"step": 19740
},
{
"epoch": 96.4490543014033,
"grad_norm": 3.1039505004882812,
"learning_rate": 7.108196721311475e-05,
"loss": 0.2532,
"step": 19760
},
{
"epoch": 96.54667480170836,
"grad_norm": 3.3837499618530273,
"learning_rate": 7.095081967213116e-05,
"loss": 0.2351,
"step": 19780
},
{
"epoch": 96.64429530201342,
"grad_norm": 4.072084426879883,
"learning_rate": 7.081967213114755e-05,
"loss": 0.2214,
"step": 19800
},
{
"epoch": 96.74191580231849,
"grad_norm": 3.669661283493042,
"learning_rate": 7.068852459016394e-05,
"loss": 0.2168,
"step": 19820
},
{
"epoch": 96.83953630262356,
"grad_norm": 3.449476957321167,
"learning_rate": 7.055737704918033e-05,
"loss": 0.2256,
"step": 19840
},
{
"epoch": 96.93715680292861,
"grad_norm": 3.7562010288238525,
"learning_rate": 7.042622950819673e-05,
"loss": 0.233,
"step": 19860
},
{
"epoch": 97.03477730323368,
"grad_norm": 3.0482330322265625,
"learning_rate": 7.029508196721312e-05,
"loss": 0.2388,
"step": 19880
},
{
"epoch": 97.13239780353874,
"grad_norm": 2.8242592811584473,
"learning_rate": 7.016393442622952e-05,
"loss": 0.247,
"step": 19900
},
{
"epoch": 97.2300183038438,
"grad_norm": 3.9347472190856934,
"learning_rate": 7.003278688524591e-05,
"loss": 0.214,
"step": 19920
},
{
"epoch": 97.32763880414888,
"grad_norm": 3.7674787044525146,
"learning_rate": 6.99016393442623e-05,
"loss": 0.2321,
"step": 19940
},
{
"epoch": 97.42525930445393,
"grad_norm": 3.0752851963043213,
"learning_rate": 6.977049180327869e-05,
"loss": 0.228,
"step": 19960
},
{
"epoch": 97.522879804759,
"grad_norm": 4.3538899421691895,
"learning_rate": 6.963934426229508e-05,
"loss": 0.2196,
"step": 19980
},
{
"epoch": 97.62050030506407,
"grad_norm": 2.677072763442993,
"learning_rate": 6.950819672131148e-05,
"loss": 0.2519,
"step": 20000
},
{
"epoch": 97.71812080536913,
"grad_norm": 3.771855592727661,
"learning_rate": 6.937704918032787e-05,
"loss": 0.1967,
"step": 20020
},
{
"epoch": 97.8157413056742,
"grad_norm": 4.431488990783691,
"learning_rate": 6.924590163934427e-05,
"loss": 0.2112,
"step": 20040
},
{
"epoch": 97.91336180597925,
"grad_norm": 3.970080852508545,
"learning_rate": 6.911475409836066e-05,
"loss": 0.2295,
"step": 20060
},
{
"epoch": 98.01098230628432,
"grad_norm": 2.521176338195801,
"learning_rate": 6.898360655737705e-05,
"loss": 0.2495,
"step": 20080
},
{
"epoch": 98.10860280658939,
"grad_norm": 3.8167226314544678,
"learning_rate": 6.885245901639344e-05,
"loss": 0.2123,
"step": 20100
},
{
"epoch": 98.20622330689444,
"grad_norm": 3.245234727859497,
"learning_rate": 6.872131147540984e-05,
"loss": 0.229,
"step": 20120
},
{
"epoch": 98.30384380719951,
"grad_norm": 3.270099401473999,
"learning_rate": 6.859016393442623e-05,
"loss": 0.1986,
"step": 20140
},
{
"epoch": 98.40146430750458,
"grad_norm": 3.133777379989624,
"learning_rate": 6.845901639344262e-05,
"loss": 0.2407,
"step": 20160
},
{
"epoch": 98.49908480780964,
"grad_norm": 4.500607490539551,
"learning_rate": 6.832786885245903e-05,
"loss": 0.223,
"step": 20180
},
{
"epoch": 98.59670530811471,
"grad_norm": 3.4543895721435547,
"learning_rate": 6.819672131147542e-05,
"loss": 0.2373,
"step": 20200
},
{
"epoch": 98.69432580841976,
"grad_norm": 3.2081384658813477,
"learning_rate": 6.80655737704918e-05,
"loss": 0.2262,
"step": 20220
},
{
"epoch": 98.79194630872483,
"grad_norm": 4.9004926681518555,
"learning_rate": 6.79344262295082e-05,
"loss": 0.2335,
"step": 20240
},
{
"epoch": 98.8895668090299,
"grad_norm": 3.9303877353668213,
"learning_rate": 6.78032786885246e-05,
"loss": 0.2061,
"step": 20260
},
{
"epoch": 98.98718730933496,
"grad_norm": 3.722957134246826,
"learning_rate": 6.767213114754099e-05,
"loss": 0.2518,
"step": 20280
},
{
"epoch": 99.08480780964003,
"grad_norm": 4.127925872802734,
"learning_rate": 6.754098360655739e-05,
"loss": 0.2133,
"step": 20300
},
{
"epoch": 99.1824283099451,
"grad_norm": 4.3639726638793945,
"learning_rate": 6.740983606557378e-05,
"loss": 0.2119,
"step": 20320
},
{
"epoch": 99.28004881025015,
"grad_norm": 3.288351058959961,
"learning_rate": 6.727868852459017e-05,
"loss": 0.2307,
"step": 20340
},
{
"epoch": 99.37766931055522,
"grad_norm": 3.5033068656921387,
"learning_rate": 6.714754098360656e-05,
"loss": 0.2227,
"step": 20360
},
{
"epoch": 99.47528981086027,
"grad_norm": 3.1117262840270996,
"learning_rate": 6.701639344262295e-05,
"loss": 0.2257,
"step": 20380
},
{
"epoch": 99.57291031116534,
"grad_norm": 4.022121906280518,
"learning_rate": 6.688524590163935e-05,
"loss": 0.2336,
"step": 20400
},
{
"epoch": 99.67053081147041,
"grad_norm": 3.4611172676086426,
"learning_rate": 6.675409836065574e-05,
"loss": 0.2192,
"step": 20420
},
{
"epoch": 99.76815131177547,
"grad_norm": 3.3648264408111572,
"learning_rate": 6.662295081967214e-05,
"loss": 0.2482,
"step": 20440
},
{
"epoch": 99.86577181208054,
"grad_norm": 3.8535125255584717,
"learning_rate": 6.649180327868853e-05,
"loss": 0.2314,
"step": 20460
},
{
"epoch": 99.96339231238561,
"grad_norm": 3.8780901432037354,
"learning_rate": 6.636065573770492e-05,
"loss": 0.2203,
"step": 20480
},
{
"epoch": 100.06101281269066,
"grad_norm": 3.479278326034546,
"learning_rate": 6.622950819672131e-05,
"loss": 0.2003,
"step": 20500
},
{
"epoch": 100.15863331299573,
"grad_norm": 3.530697822570801,
"learning_rate": 6.609836065573771e-05,
"loss": 0.2084,
"step": 20520
},
{
"epoch": 100.25625381330079,
"grad_norm": 3.3677327632904053,
"learning_rate": 6.59672131147541e-05,
"loss": 0.2394,
"step": 20540
},
{
"epoch": 100.35387431360586,
"grad_norm": 2.9369733333587646,
"learning_rate": 6.58360655737705e-05,
"loss": 0.2553,
"step": 20560
},
{
"epoch": 100.45149481391093,
"grad_norm": 3.9915401935577393,
"learning_rate": 6.57049180327869e-05,
"loss": 0.2393,
"step": 20580
},
{
"epoch": 100.54911531421598,
"grad_norm": 3.6804075241088867,
"learning_rate": 6.557377049180327e-05,
"loss": 0.2266,
"step": 20600
},
{
"epoch": 100.64673581452105,
"grad_norm": 4.8830156326293945,
"learning_rate": 6.544262295081967e-05,
"loss": 0.19,
"step": 20620
},
{
"epoch": 100.74435631482612,
"grad_norm": 4.2519731521606445,
"learning_rate": 6.531147540983606e-05,
"loss": 0.2416,
"step": 20640
},
{
"epoch": 100.84197681513118,
"grad_norm": 3.413914203643799,
"learning_rate": 6.518032786885247e-05,
"loss": 0.2026,
"step": 20660
},
{
"epoch": 100.93959731543625,
"grad_norm": 3.642609119415283,
"learning_rate": 6.504918032786886e-05,
"loss": 0.2227,
"step": 20680
},
{
"epoch": 101.0372178157413,
"grad_norm": 4.770230770111084,
"learning_rate": 6.491803278688526e-05,
"loss": 0.2277,
"step": 20700
},
{
"epoch": 101.13483831604637,
"grad_norm": 4.115867614746094,
"learning_rate": 6.478688524590165e-05,
"loss": 0.205,
"step": 20720
},
{
"epoch": 101.23245881635144,
"grad_norm": 4.3617048263549805,
"learning_rate": 6.465573770491804e-05,
"loss": 0.2048,
"step": 20740
},
{
"epoch": 101.3300793166565,
"grad_norm": 3.818500518798828,
"learning_rate": 6.452459016393443e-05,
"loss": 0.2245,
"step": 20760
},
{
"epoch": 101.42769981696156,
"grad_norm": 3.3728935718536377,
"learning_rate": 6.439344262295083e-05,
"loss": 0.2342,
"step": 20780
},
{
"epoch": 101.52532031726662,
"grad_norm": 3.2103302478790283,
"learning_rate": 6.426229508196722e-05,
"loss": 0.2024,
"step": 20800
},
{
"epoch": 101.62294081757169,
"grad_norm": 3.5993459224700928,
"learning_rate": 6.413114754098361e-05,
"loss": 0.2272,
"step": 20820
},
{
"epoch": 101.72056131787676,
"grad_norm": 3.478533983230591,
"learning_rate": 6.400000000000001e-05,
"loss": 0.236,
"step": 20840
},
{
"epoch": 101.81818181818181,
"grad_norm": 4.245722770690918,
"learning_rate": 6.386885245901639e-05,
"loss": 0.2246,
"step": 20860
},
{
"epoch": 101.91580231848688,
"grad_norm": 3.5051097869873047,
"learning_rate": 6.373770491803279e-05,
"loss": 0.2194,
"step": 20880
},
{
"epoch": 102.01342281879195,
"grad_norm": 2.882301092147827,
"learning_rate": 6.360655737704918e-05,
"loss": 0.2196,
"step": 20900
},
{
"epoch": 102.111043319097,
"grad_norm": 3.080702066421509,
"learning_rate": 6.347540983606558e-05,
"loss": 0.2109,
"step": 20920
},
{
"epoch": 102.20866381940208,
"grad_norm": 2.9892094135284424,
"learning_rate": 6.334426229508197e-05,
"loss": 0.2086,
"step": 20940
},
{
"epoch": 102.30628431970713,
"grad_norm": 3.401524782180786,
"learning_rate": 6.321311475409837e-05,
"loss": 0.202,
"step": 20960
},
{
"epoch": 102.4039048200122,
"grad_norm": 3.0444400310516357,
"learning_rate": 6.308196721311475e-05,
"loss": 0.2071,
"step": 20980
},
{
"epoch": 102.50152532031727,
"grad_norm": 3.028918743133545,
"learning_rate": 6.295081967213115e-05,
"loss": 0.226,
"step": 21000
},
{
"epoch": 102.59914582062233,
"grad_norm": 3.2344133853912354,
"learning_rate": 6.281967213114754e-05,
"loss": 0.2417,
"step": 21020
},
{
"epoch": 102.6967663209274,
"grad_norm": 3.1719000339508057,
"learning_rate": 6.268852459016393e-05,
"loss": 0.2363,
"step": 21040
},
{
"epoch": 102.79438682123246,
"grad_norm": 3.474695920944214,
"learning_rate": 6.255737704918033e-05,
"loss": 0.2278,
"step": 21060
},
{
"epoch": 102.89200732153752,
"grad_norm": 4.550293445587158,
"learning_rate": 6.242622950819672e-05,
"loss": 0.2087,
"step": 21080
},
{
"epoch": 102.98962782184259,
"grad_norm": 3.7696001529693604,
"learning_rate": 6.229508196721313e-05,
"loss": 0.2305,
"step": 21100
},
{
"epoch": 103.08724832214764,
"grad_norm": 3.388397216796875,
"learning_rate": 6.21639344262295e-05,
"loss": 0.2133,
"step": 21120
},
{
"epoch": 103.18486882245271,
"grad_norm": 4.243392467498779,
"learning_rate": 6.20327868852459e-05,
"loss": 0.213,
"step": 21140
},
{
"epoch": 103.28248932275778,
"grad_norm": 2.646786689758301,
"learning_rate": 6.19016393442623e-05,
"loss": 0.2132,
"step": 21160
},
{
"epoch": 103.38010982306284,
"grad_norm": 3.157848596572876,
"learning_rate": 6.17704918032787e-05,
"loss": 0.1941,
"step": 21180
},
{
"epoch": 103.47773032336791,
"grad_norm": 3.383357286453247,
"learning_rate": 6.163934426229509e-05,
"loss": 0.2211,
"step": 21200
},
{
"epoch": 103.57535082367298,
"grad_norm": 2.940901279449463,
"learning_rate": 6.150819672131148e-05,
"loss": 0.1934,
"step": 21220
},
{
"epoch": 103.67297132397803,
"grad_norm": 4.2973103523254395,
"learning_rate": 6.137704918032787e-05,
"loss": 0.2516,
"step": 21240
},
{
"epoch": 103.7705918242831,
"grad_norm": 3.586219310760498,
"learning_rate": 6.124590163934426e-05,
"loss": 0.262,
"step": 21260
},
{
"epoch": 103.86821232458816,
"grad_norm": 3.7726240158081055,
"learning_rate": 6.111475409836066e-05,
"loss": 0.1876,
"step": 21280
},
{
"epoch": 103.96583282489323,
"grad_norm": 3.5038113594055176,
"learning_rate": 6.098360655737705e-05,
"loss": 0.2156,
"step": 21300
},
{
"epoch": 104.0634533251983,
"grad_norm": 3.6055893898010254,
"learning_rate": 6.085245901639345e-05,
"loss": 0.2386,
"step": 21320
},
{
"epoch": 104.16107382550335,
"grad_norm": 4.037646770477295,
"learning_rate": 6.072131147540984e-05,
"loss": 0.229,
"step": 21340
},
{
"epoch": 104.25869432580842,
"grad_norm": 3.209284543991089,
"learning_rate": 6.0590163934426236e-05,
"loss": 0.2204,
"step": 21360
},
{
"epoch": 104.35631482611349,
"grad_norm": 3.388456106185913,
"learning_rate": 6.0459016393442625e-05,
"loss": 0.2155,
"step": 21380
},
{
"epoch": 104.45393532641855,
"grad_norm": 2.9914731979370117,
"learning_rate": 6.032786885245902e-05,
"loss": 0.2205,
"step": 21400
},
{
"epoch": 104.55155582672361,
"grad_norm": 3.4753572940826416,
"learning_rate": 6.019672131147541e-05,
"loss": 0.226,
"step": 21420
},
{
"epoch": 104.64917632702867,
"grad_norm": 3.400149345397949,
"learning_rate": 6.00655737704918e-05,
"loss": 0.2117,
"step": 21440
},
{
"epoch": 104.74679682733374,
"grad_norm": 3.2313406467437744,
"learning_rate": 5.99344262295082e-05,
"loss": 0.216,
"step": 21460
},
{
"epoch": 104.84441732763881,
"grad_norm": 4.102112770080566,
"learning_rate": 5.9803278688524586e-05,
"loss": 0.1962,
"step": 21480
},
{
"epoch": 104.94203782794386,
"grad_norm": 3.6872830390930176,
"learning_rate": 5.967213114754099e-05,
"loss": 0.2212,
"step": 21500
},
{
"epoch": 105.03965832824893,
"grad_norm": 3.1488096714019775,
"learning_rate": 5.954098360655738e-05,
"loss": 0.1919,
"step": 21520
},
{
"epoch": 105.137278828554,
"grad_norm": 3.4388973712921143,
"learning_rate": 5.9409836065573774e-05,
"loss": 0.247,
"step": 21540
},
{
"epoch": 105.23489932885906,
"grad_norm": 3.776465892791748,
"learning_rate": 5.927868852459016e-05,
"loss": 0.1856,
"step": 21560
},
{
"epoch": 105.33251982916413,
"grad_norm": 3.552684783935547,
"learning_rate": 5.9147540983606566e-05,
"loss": 0.2042,
"step": 21580
},
{
"epoch": 105.43014032946918,
"grad_norm": 2.8810436725616455,
"learning_rate": 5.9016393442622956e-05,
"loss": 0.2353,
"step": 21600
},
{
"epoch": 105.52776082977425,
"grad_norm": 3.2408454418182373,
"learning_rate": 5.888524590163935e-05,
"loss": 0.2148,
"step": 21620
},
{
"epoch": 105.62538133007932,
"grad_norm": 3.5531325340270996,
"learning_rate": 5.875409836065574e-05,
"loss": 0.2106,
"step": 21640
},
{
"epoch": 105.72300183038438,
"grad_norm": 4.188174247741699,
"learning_rate": 5.862295081967213e-05,
"loss": 0.2211,
"step": 21660
},
{
"epoch": 105.82062233068945,
"grad_norm": 4.232003688812256,
"learning_rate": 5.849180327868853e-05,
"loss": 0.2139,
"step": 21680
},
{
"epoch": 105.91824283099452,
"grad_norm": 3.7038660049438477,
"learning_rate": 5.8360655737704916e-05,
"loss": 0.2048,
"step": 21700
},
{
"epoch": 106.01586333129957,
"grad_norm": 2.9898719787597656,
"learning_rate": 5.822950819672132e-05,
"loss": 0.2184,
"step": 21720
},
{
"epoch": 106.11348383160464,
"grad_norm": 3.482365846633911,
"learning_rate": 5.80983606557377e-05,
"loss": 0.2171,
"step": 21740
},
{
"epoch": 106.2111043319097,
"grad_norm": 3.3595712184906006,
"learning_rate": 5.7967213114754104e-05,
"loss": 0.2041,
"step": 21760
},
{
"epoch": 106.30872483221476,
"grad_norm": 3.7470903396606445,
"learning_rate": 5.7836065573770494e-05,
"loss": 0.1946,
"step": 21780
},
{
"epoch": 106.40634533251983,
"grad_norm": 3.418549060821533,
"learning_rate": 5.770491803278689e-05,
"loss": 0.2013,
"step": 21800
},
{
"epoch": 106.50396583282489,
"grad_norm": 3.115424394607544,
"learning_rate": 5.757377049180328e-05,
"loss": 0.2064,
"step": 21820
},
{
"epoch": 106.60158633312996,
"grad_norm": 3.2718710899353027,
"learning_rate": 5.744262295081968e-05,
"loss": 0.2266,
"step": 21840
},
{
"epoch": 106.69920683343503,
"grad_norm": 2.9478771686553955,
"learning_rate": 5.731147540983607e-05,
"loss": 0.2321,
"step": 21860
},
{
"epoch": 106.79682733374008,
"grad_norm": 3.479456663131714,
"learning_rate": 5.7180327868852454e-05,
"loss": 0.2317,
"step": 21880
},
{
"epoch": 106.89444783404515,
"grad_norm": 3.934882879257202,
"learning_rate": 5.704918032786886e-05,
"loss": 0.2074,
"step": 21900
},
{
"epoch": 106.99206833435021,
"grad_norm": 3.40720272064209,
"learning_rate": 5.6918032786885246e-05,
"loss": 0.2123,
"step": 21920
},
{
"epoch": 107.08968883465528,
"grad_norm": 2.825427770614624,
"learning_rate": 5.678688524590164e-05,
"loss": 0.1897,
"step": 21940
},
{
"epoch": 107.18730933496035,
"grad_norm": 3.6434717178344727,
"learning_rate": 5.665573770491803e-05,
"loss": 0.207,
"step": 21960
},
{
"epoch": 107.2849298352654,
"grad_norm": 4.119269847869873,
"learning_rate": 5.6524590163934435e-05,
"loss": 0.2086,
"step": 21980
},
{
"epoch": 107.38255033557047,
"grad_norm": 2.835963487625122,
"learning_rate": 5.639344262295082e-05,
"loss": 0.1942,
"step": 22000
},
{
"epoch": 107.48017083587554,
"grad_norm": 3.175858736038208,
"learning_rate": 5.626229508196722e-05,
"loss": 0.2089,
"step": 22020
},
{
"epoch": 107.5777913361806,
"grad_norm": 3.6265509128570557,
"learning_rate": 5.613114754098361e-05,
"loss": 0.2164,
"step": 22040
},
{
"epoch": 107.67541183648567,
"grad_norm": 2.90639591217041,
"learning_rate": 5.6000000000000006e-05,
"loss": 0.2121,
"step": 22060
},
{
"epoch": 107.77303233679072,
"grad_norm": 4.155264377593994,
"learning_rate": 5.5868852459016395e-05,
"loss": 0.219,
"step": 22080
},
{
"epoch": 107.87065283709579,
"grad_norm": 3.730433940887451,
"learning_rate": 5.5737704918032785e-05,
"loss": 0.2369,
"step": 22100
},
{
"epoch": 107.96827333740086,
"grad_norm": 2.8690969944000244,
"learning_rate": 5.560655737704919e-05,
"loss": 0.2189,
"step": 22120
},
{
"epoch": 108.06589383770591,
"grad_norm": 3.061427116394043,
"learning_rate": 5.547540983606557e-05,
"loss": 0.2203,
"step": 22140
},
{
"epoch": 108.16351433801098,
"grad_norm": 3.6155471801757812,
"learning_rate": 5.534426229508197e-05,
"loss": 0.2043,
"step": 22160
},
{
"epoch": 108.26113483831605,
"grad_norm": 3.1074283123016357,
"learning_rate": 5.521311475409836e-05,
"loss": 0.2095,
"step": 22180
},
{
"epoch": 108.35875533862111,
"grad_norm": 3.6168534755706787,
"learning_rate": 5.508196721311476e-05,
"loss": 0.2339,
"step": 22200
},
{
"epoch": 108.45637583892618,
"grad_norm": 2.9254798889160156,
"learning_rate": 5.495081967213115e-05,
"loss": 0.214,
"step": 22220
},
{
"epoch": 108.55399633923123,
"grad_norm": 3.2898197174072266,
"learning_rate": 5.481967213114755e-05,
"loss": 0.1981,
"step": 22240
},
{
"epoch": 108.6516168395363,
"grad_norm": 3.0724892616271973,
"learning_rate": 5.4688524590163933e-05,
"loss": 0.1997,
"step": 22260
},
{
"epoch": 108.74923733984137,
"grad_norm": 3.3344459533691406,
"learning_rate": 5.4557377049180336e-05,
"loss": 0.2243,
"step": 22280
},
{
"epoch": 108.84685784014643,
"grad_norm": 4.20386266708374,
"learning_rate": 5.4426229508196726e-05,
"loss": 0.1973,
"step": 22300
},
{
"epoch": 108.9444783404515,
"grad_norm": 3.0647547245025635,
"learning_rate": 5.4295081967213115e-05,
"loss": 0.1984,
"step": 22320
},
{
"epoch": 109.04209884075657,
"grad_norm": 3.2792584896087646,
"learning_rate": 5.416393442622951e-05,
"loss": 0.2209,
"step": 22340
},
{
"epoch": 109.13971934106162,
"grad_norm": 2.900493860244751,
"learning_rate": 5.40327868852459e-05,
"loss": 0.2123,
"step": 22360
},
{
"epoch": 109.23733984136669,
"grad_norm": 2.754514217376709,
"learning_rate": 5.3901639344262304e-05,
"loss": 0.2049,
"step": 22380
},
{
"epoch": 109.33496034167175,
"grad_norm": 2.955946445465088,
"learning_rate": 5.3770491803278686e-05,
"loss": 0.2164,
"step": 22400
},
{
"epoch": 109.43258084197682,
"grad_norm": 3.0447773933410645,
"learning_rate": 5.363934426229509e-05,
"loss": 0.1903,
"step": 22420
},
{
"epoch": 109.53020134228188,
"grad_norm": 2.8788065910339355,
"learning_rate": 5.350819672131148e-05,
"loss": 0.2018,
"step": 22440
},
{
"epoch": 109.62782184258694,
"grad_norm": 3.766073703765869,
"learning_rate": 5.3377049180327875e-05,
"loss": 0.2092,
"step": 22460
},
{
"epoch": 109.72544234289201,
"grad_norm": 4.014832973480225,
"learning_rate": 5.3245901639344264e-05,
"loss": 0.2138,
"step": 22480
},
{
"epoch": 109.82306284319706,
"grad_norm": 2.987813949584961,
"learning_rate": 5.311475409836065e-05,
"loss": 0.2253,
"step": 22500
},
{
"epoch": 109.92068334350213,
"grad_norm": 2.980419158935547,
"learning_rate": 5.298360655737705e-05,
"loss": 0.2159,
"step": 22520
},
{
"epoch": 110.0183038438072,
"grad_norm": 3.1005942821502686,
"learning_rate": 5.285245901639344e-05,
"loss": 0.1907,
"step": 22540
},
{
"epoch": 110.11592434411226,
"grad_norm": 3.2539725303649902,
"learning_rate": 5.272131147540984e-05,
"loss": 0.2023,
"step": 22560
},
{
"epoch": 110.21354484441733,
"grad_norm": 3.4975571632385254,
"learning_rate": 5.259016393442623e-05,
"loss": 0.2101,
"step": 22580
},
{
"epoch": 110.3111653447224,
"grad_norm": 4.542675018310547,
"learning_rate": 5.245901639344263e-05,
"loss": 0.2071,
"step": 22600
},
{
"epoch": 110.40878584502745,
"grad_norm": 2.5775978565216064,
"learning_rate": 5.2327868852459017e-05,
"loss": 0.2027,
"step": 22620
},
{
"epoch": 110.50640634533252,
"grad_norm": 2.964486598968506,
"learning_rate": 5.219672131147541e-05,
"loss": 0.2281,
"step": 22640
},
{
"epoch": 110.60402684563758,
"grad_norm": 3.883513927459717,
"learning_rate": 5.20655737704918e-05,
"loss": 0.2083,
"step": 22660
},
{
"epoch": 110.70164734594265,
"grad_norm": 3.696744441986084,
"learning_rate": 5.1934426229508205e-05,
"loss": 0.2044,
"step": 22680
},
{
"epoch": 110.79926784624772,
"grad_norm": 3.104335308074951,
"learning_rate": 5.1803278688524594e-05,
"loss": 0.216,
"step": 22700
},
{
"epoch": 110.89688834655277,
"grad_norm": 3.6240875720977783,
"learning_rate": 5.1672131147540984e-05,
"loss": 0.2086,
"step": 22720
},
{
"epoch": 110.99450884685784,
"grad_norm": 4.600063323974609,
"learning_rate": 5.154098360655738e-05,
"loss": 0.1983,
"step": 22740
},
{
"epoch": 111.09212934716291,
"grad_norm": 4.165120601654053,
"learning_rate": 5.140983606557377e-05,
"loss": 0.207,
"step": 22760
},
{
"epoch": 111.18974984746797,
"grad_norm": 3.1178033351898193,
"learning_rate": 5.1278688524590165e-05,
"loss": 0.1836,
"step": 22780
},
{
"epoch": 111.28737034777303,
"grad_norm": 3.8547523021698,
"learning_rate": 5.1147540983606555e-05,
"loss": 0.213,
"step": 22800
},
{
"epoch": 111.38499084807809,
"grad_norm": 3.626835346221924,
"learning_rate": 5.101639344262296e-05,
"loss": 0.1901,
"step": 22820
},
{
"epoch": 111.48261134838316,
"grad_norm": 3.89408540725708,
"learning_rate": 5.088524590163935e-05,
"loss": 0.2151,
"step": 22840
},
{
"epoch": 111.58023184868823,
"grad_norm": 3.3434460163116455,
"learning_rate": 5.075409836065574e-05,
"loss": 0.2165,
"step": 22860
},
{
"epoch": 111.67785234899328,
"grad_norm": 3.1831305027008057,
"learning_rate": 5.062295081967213e-05,
"loss": 0.1911,
"step": 22880
},
{
"epoch": 111.77547284929835,
"grad_norm": 3.6153972148895264,
"learning_rate": 5.049180327868853e-05,
"loss": 0.2076,
"step": 22900
},
{
"epoch": 111.87309334960342,
"grad_norm": 2.87998628616333,
"learning_rate": 5.036065573770492e-05,
"loss": 0.2121,
"step": 22920
},
{
"epoch": 111.97071384990848,
"grad_norm": 2.9351823329925537,
"learning_rate": 5.022950819672131e-05,
"loss": 0.2126,
"step": 22940
},
{
"epoch": 112.06833435021355,
"grad_norm": 3.098568916320801,
"learning_rate": 5.009836065573771e-05,
"loss": 0.2169,
"step": 22960
},
{
"epoch": 112.1659548505186,
"grad_norm": 4.321131229400635,
"learning_rate": 4.99672131147541e-05,
"loss": 0.2028,
"step": 22980
},
{
"epoch": 112.26357535082367,
"grad_norm": 2.1096384525299072,
"learning_rate": 4.9836065573770496e-05,
"loss": 0.2041,
"step": 23000
},
{
"epoch": 112.36119585112874,
"grad_norm": 3.537277936935425,
"learning_rate": 4.970491803278689e-05,
"loss": 0.1988,
"step": 23020
},
{
"epoch": 112.4588163514338,
"grad_norm": 4.58275842666626,
"learning_rate": 4.957377049180328e-05,
"loss": 0.1855,
"step": 23040
},
{
"epoch": 112.55643685173887,
"grad_norm": 3.290548086166382,
"learning_rate": 4.944262295081967e-05,
"loss": 0.2041,
"step": 23060
},
{
"epoch": 112.65405735204394,
"grad_norm": 4.11199951171875,
"learning_rate": 4.931147540983607e-05,
"loss": 0.2062,
"step": 23080
},
{
"epoch": 112.75167785234899,
"grad_norm": 4.122567176818848,
"learning_rate": 4.918032786885246e-05,
"loss": 0.2003,
"step": 23100
},
{
"epoch": 112.84929835265406,
"grad_norm": 2.775681495666504,
"learning_rate": 4.904918032786885e-05,
"loss": 0.2201,
"step": 23120
},
{
"epoch": 112.94691885295912,
"grad_norm": 2.8952252864837646,
"learning_rate": 4.891803278688525e-05,
"loss": 0.2068,
"step": 23140
},
{
"epoch": 113.04453935326418,
"grad_norm": 2.8353431224823,
"learning_rate": 4.8786885245901645e-05,
"loss": 0.2087,
"step": 23160
},
{
"epoch": 113.14215985356925,
"grad_norm": 4.125601768493652,
"learning_rate": 4.8655737704918034e-05,
"loss": 0.1973,
"step": 23180
},
{
"epoch": 113.23978035387431,
"grad_norm": 3.5090670585632324,
"learning_rate": 4.852459016393443e-05,
"loss": 0.1847,
"step": 23200
},
{
"epoch": 113.33740085417938,
"grad_norm": 2.688890218734741,
"learning_rate": 4.8393442622950826e-05,
"loss": 0.1842,
"step": 23220
},
{
"epoch": 113.43502135448445,
"grad_norm": 4.937670707702637,
"learning_rate": 4.8262295081967216e-05,
"loss": 0.214,
"step": 23240
},
{
"epoch": 113.5326418547895,
"grad_norm": 4.070579528808594,
"learning_rate": 4.8131147540983605e-05,
"loss": 0.2203,
"step": 23260
},
{
"epoch": 113.63026235509457,
"grad_norm": 3.052457809448242,
"learning_rate": 4.8e-05,
"loss": 0.2116,
"step": 23280
},
{
"epoch": 113.72788285539963,
"grad_norm": 3.2423150539398193,
"learning_rate": 4.78688524590164e-05,
"loss": 0.192,
"step": 23300
},
{
"epoch": 113.8255033557047,
"grad_norm": 2.9941837787628174,
"learning_rate": 4.773770491803279e-05,
"loss": 0.1916,
"step": 23320
},
{
"epoch": 113.92312385600977,
"grad_norm": 3.3031139373779297,
"learning_rate": 4.760655737704918e-05,
"loss": 0.2261,
"step": 23340
},
{
"epoch": 114.02074435631482,
"grad_norm": 3.531646490097046,
"learning_rate": 4.747540983606558e-05,
"loss": 0.2186,
"step": 23360
},
{
"epoch": 114.11836485661989,
"grad_norm": 2.753261089324951,
"learning_rate": 4.734426229508197e-05,
"loss": 0.1979,
"step": 23380
},
{
"epoch": 114.21598535692496,
"grad_norm": 3.2900078296661377,
"learning_rate": 4.7213114754098365e-05,
"loss": 0.1892,
"step": 23400
},
{
"epoch": 114.31360585723002,
"grad_norm": 3.6558027267456055,
"learning_rate": 4.708196721311476e-05,
"loss": 0.2202,
"step": 23420
},
{
"epoch": 114.41122635753509,
"grad_norm": 2.9026436805725098,
"learning_rate": 4.695081967213115e-05,
"loss": 0.2193,
"step": 23440
},
{
"epoch": 114.50884685784014,
"grad_norm": 3.0958666801452637,
"learning_rate": 4.681967213114754e-05,
"loss": 0.1861,
"step": 23460
},
{
"epoch": 114.60646735814521,
"grad_norm": 3.4162192344665527,
"learning_rate": 4.6688524590163936e-05,
"loss": 0.1813,
"step": 23480
},
{
"epoch": 114.70408785845028,
"grad_norm": 3.4320363998413086,
"learning_rate": 4.655737704918033e-05,
"loss": 0.1849,
"step": 23500
},
{
"epoch": 114.80170835875533,
"grad_norm": 3.931405544281006,
"learning_rate": 4.642622950819672e-05,
"loss": 0.2091,
"step": 23520
},
{
"epoch": 114.8993288590604,
"grad_norm": 2.463747024536133,
"learning_rate": 4.629508196721312e-05,
"loss": 0.2013,
"step": 23540
},
{
"epoch": 114.99694935936547,
"grad_norm": 3.4332265853881836,
"learning_rate": 4.616393442622951e-05,
"loss": 0.2224,
"step": 23560
},
{
"epoch": 115.09456985967053,
"grad_norm": 4.516382217407227,
"learning_rate": 4.60327868852459e-05,
"loss": 0.191,
"step": 23580
},
{
"epoch": 115.1921903599756,
"grad_norm": 3.7761762142181396,
"learning_rate": 4.59016393442623e-05,
"loss": 0.1972,
"step": 23600
},
{
"epoch": 115.28981086028065,
"grad_norm": 3.425079584121704,
"learning_rate": 4.5770491803278695e-05,
"loss": 0.2053,
"step": 23620
},
{
"epoch": 115.38743136058572,
"grad_norm": 3.098958969116211,
"learning_rate": 4.5639344262295084e-05,
"loss": 0.1994,
"step": 23640
},
{
"epoch": 115.48505186089079,
"grad_norm": 4.242363452911377,
"learning_rate": 4.550819672131148e-05,
"loss": 0.2062,
"step": 23660
},
{
"epoch": 115.58267236119585,
"grad_norm": 3.9641380310058594,
"learning_rate": 4.537704918032787e-05,
"loss": 0.1892,
"step": 23680
},
{
"epoch": 115.68029286150092,
"grad_norm": 3.817915678024292,
"learning_rate": 4.524590163934426e-05,
"loss": 0.2076,
"step": 23700
},
{
"epoch": 115.77791336180599,
"grad_norm": 4.9167680740356445,
"learning_rate": 4.5114754098360655e-05,
"loss": 0.1991,
"step": 23720
},
{
"epoch": 115.87553386211104,
"grad_norm": 4.05020809173584,
"learning_rate": 4.498360655737705e-05,
"loss": 0.2154,
"step": 23740
},
{
"epoch": 115.97315436241611,
"grad_norm": 5.213686466217041,
"learning_rate": 4.485245901639345e-05,
"loss": 0.1893,
"step": 23760
},
{
"epoch": 116.07077486272117,
"grad_norm": 2.9788458347320557,
"learning_rate": 4.472131147540984e-05,
"loss": 0.2015,
"step": 23780
},
{
"epoch": 116.16839536302624,
"grad_norm": 3.5112173557281494,
"learning_rate": 4.459016393442623e-05,
"loss": 0.2058,
"step": 23800
},
{
"epoch": 116.2660158633313,
"grad_norm": 3.3048148155212402,
"learning_rate": 4.445901639344263e-05,
"loss": 0.2107,
"step": 23820
},
{
"epoch": 116.36363636363636,
"grad_norm": 3.895615339279175,
"learning_rate": 4.432786885245902e-05,
"loss": 0.1744,
"step": 23840
},
{
"epoch": 116.46125686394143,
"grad_norm": 3.7658941745758057,
"learning_rate": 4.4196721311475415e-05,
"loss": 0.1935,
"step": 23860
},
{
"epoch": 116.5588773642465,
"grad_norm": 3.4818522930145264,
"learning_rate": 4.406557377049181e-05,
"loss": 0.2028,
"step": 23880
},
{
"epoch": 116.65649786455155,
"grad_norm": 2.971785068511963,
"learning_rate": 4.3934426229508194e-05,
"loss": 0.1993,
"step": 23900
},
{
"epoch": 116.75411836485662,
"grad_norm": 4.34529447555542,
"learning_rate": 4.380327868852459e-05,
"loss": 0.1887,
"step": 23920
},
{
"epoch": 116.85173886516168,
"grad_norm": 3.1781437397003174,
"learning_rate": 4.3672131147540986e-05,
"loss": 0.2241,
"step": 23940
},
{
"epoch": 116.94935936546675,
"grad_norm": 3.8372061252593994,
"learning_rate": 4.3540983606557375e-05,
"loss": 0.2103,
"step": 23960
},
{
"epoch": 117.04697986577182,
"grad_norm": 3.981776237487793,
"learning_rate": 4.340983606557377e-05,
"loss": 0.1859,
"step": 23980
},
{
"epoch": 117.14460036607687,
"grad_norm": 4.0308685302734375,
"learning_rate": 4.327868852459017e-05,
"loss": 0.1841,
"step": 24000
},
{
"epoch": 117.24222086638194,
"grad_norm": 3.2919681072235107,
"learning_rate": 4.3147540983606564e-05,
"loss": 0.1964,
"step": 24020
},
{
"epoch": 117.33984136668701,
"grad_norm": 4.155374050140381,
"learning_rate": 4.301639344262295e-05,
"loss": 0.1896,
"step": 24040
},
{
"epoch": 117.43746186699207,
"grad_norm": 3.5654046535491943,
"learning_rate": 4.288524590163935e-05,
"loss": 0.1913,
"step": 24060
},
{
"epoch": 117.53508236729714,
"grad_norm": 2.805769681930542,
"learning_rate": 4.2754098360655745e-05,
"loss": 0.1995,
"step": 24080
},
{
"epoch": 117.63270286760219,
"grad_norm": 2.990482807159424,
"learning_rate": 4.262295081967213e-05,
"loss": 0.2034,
"step": 24100
},
{
"epoch": 117.73032336790726,
"grad_norm": 3.236433506011963,
"learning_rate": 4.2491803278688524e-05,
"loss": 0.1948,
"step": 24120
},
{
"epoch": 117.82794386821233,
"grad_norm": 4.235846996307373,
"learning_rate": 4.236065573770492e-05,
"loss": 0.2014,
"step": 24140
},
{
"epoch": 117.92556436851739,
"grad_norm": 3.238041400909424,
"learning_rate": 4.222950819672131e-05,
"loss": 0.2106,
"step": 24160
},
{
"epoch": 118.02318486882245,
"grad_norm": 3.18979811668396,
"learning_rate": 4.2098360655737706e-05,
"loss": 0.1918,
"step": 24180
},
{
"epoch": 118.12080536912751,
"grad_norm": 2.699910879135132,
"learning_rate": 4.19672131147541e-05,
"loss": 0.2121,
"step": 24200
},
{
"epoch": 118.21842586943258,
"grad_norm": 3.6096503734588623,
"learning_rate": 4.183606557377049e-05,
"loss": 0.2165,
"step": 24220
},
{
"epoch": 118.31604636973765,
"grad_norm": 3.384746789932251,
"learning_rate": 4.170491803278689e-05,
"loss": 0.1884,
"step": 24240
},
{
"epoch": 118.4136668700427,
"grad_norm": 3.606003761291504,
"learning_rate": 4.1573770491803283e-05,
"loss": 0.1932,
"step": 24260
},
{
"epoch": 118.51128737034777,
"grad_norm": 2.9768433570861816,
"learning_rate": 4.144262295081967e-05,
"loss": 0.1802,
"step": 24280
},
{
"epoch": 118.60890787065284,
"grad_norm": 3.784608840942383,
"learning_rate": 4.131147540983607e-05,
"loss": 0.1919,
"step": 24300
},
{
"epoch": 118.7065283709579,
"grad_norm": 3.2883870601654053,
"learning_rate": 4.118032786885246e-05,
"loss": 0.2015,
"step": 24320
},
{
"epoch": 118.80414887126297,
"grad_norm": 4.078542232513428,
"learning_rate": 4.1049180327868854e-05,
"loss": 0.1994,
"step": 24340
},
{
"epoch": 118.90176937156802,
"grad_norm": 3.7962749004364014,
"learning_rate": 4.0918032786885244e-05,
"loss": 0.1958,
"step": 24360
},
{
"epoch": 118.99938987187309,
"grad_norm": 3.4484541416168213,
"learning_rate": 4.078688524590164e-05,
"loss": 0.1926,
"step": 24380
},
{
"epoch": 119.09701037217816,
"grad_norm": 2.9271483421325684,
"learning_rate": 4.0655737704918036e-05,
"loss": 0.1818,
"step": 24400
},
{
"epoch": 119.19463087248322,
"grad_norm": 5.304340839385986,
"learning_rate": 4.0524590163934425e-05,
"loss": 0.192,
"step": 24420
},
{
"epoch": 119.29225137278829,
"grad_norm": 4.88720178604126,
"learning_rate": 4.039344262295082e-05,
"loss": 0.1975,
"step": 24440
},
{
"epoch": 119.38987187309336,
"grad_norm": 3.7572522163391113,
"learning_rate": 4.026229508196722e-05,
"loss": 0.1856,
"step": 24460
},
{
"epoch": 119.48749237339841,
"grad_norm": 3.3659627437591553,
"learning_rate": 4.013114754098361e-05,
"loss": 0.1961,
"step": 24480
},
{
"epoch": 119.58511287370348,
"grad_norm": 3.1397533416748047,
"learning_rate": 4e-05,
"loss": 0.1991,
"step": 24500
},
{
"epoch": 119.68273337400854,
"grad_norm": 3.9070587158203125,
"learning_rate": 3.98688524590164e-05,
"loss": 0.1946,
"step": 24520
},
{
"epoch": 119.7803538743136,
"grad_norm": 3.3265631198883057,
"learning_rate": 3.973770491803279e-05,
"loss": 0.194,
"step": 24540
},
{
"epoch": 119.87797437461867,
"grad_norm": 3.5021114349365234,
"learning_rate": 3.960655737704918e-05,
"loss": 0.2084,
"step": 24560
},
{
"epoch": 119.97559487492373,
"grad_norm": 3.7133572101593018,
"learning_rate": 3.9475409836065574e-05,
"loss": 0.2102,
"step": 24580
},
{
"epoch": 120.0732153752288,
"grad_norm": 3.27691912651062,
"learning_rate": 3.934426229508197e-05,
"loss": 0.1864,
"step": 24600
},
{
"epoch": 120.17083587553387,
"grad_norm": 4.74030065536499,
"learning_rate": 3.921311475409836e-05,
"loss": 0.1967,
"step": 24620
},
{
"epoch": 120.26845637583892,
"grad_norm": 4.080429553985596,
"learning_rate": 3.9081967213114756e-05,
"loss": 0.1888,
"step": 24640
},
{
"epoch": 120.36607687614399,
"grad_norm": 3.5443646907806396,
"learning_rate": 3.895081967213115e-05,
"loss": 0.199,
"step": 24660
},
{
"epoch": 120.46369737644905,
"grad_norm": 4.0403923988342285,
"learning_rate": 3.881967213114754e-05,
"loss": 0.2002,
"step": 24680
},
{
"epoch": 120.56131787675412,
"grad_norm": 4.144064426422119,
"learning_rate": 3.868852459016394e-05,
"loss": 0.193,
"step": 24700
},
{
"epoch": 120.65893837705919,
"grad_norm": 5.478573799133301,
"learning_rate": 3.8557377049180334e-05,
"loss": 0.19,
"step": 24720
},
{
"epoch": 120.75655887736424,
"grad_norm": 3.799931287765503,
"learning_rate": 3.842622950819672e-05,
"loss": 0.1876,
"step": 24740
},
{
"epoch": 120.85417937766931,
"grad_norm": 3.1754050254821777,
"learning_rate": 3.829508196721311e-05,
"loss": 0.1881,
"step": 24760
},
{
"epoch": 120.95179987797438,
"grad_norm": 3.4417614936828613,
"learning_rate": 3.816393442622951e-05,
"loss": 0.2023,
"step": 24780
},
{
"epoch": 121.04942037827944,
"grad_norm": 2.689237594604492,
"learning_rate": 3.8032786885245905e-05,
"loss": 0.1918,
"step": 24800
},
{
"epoch": 121.1470408785845,
"grad_norm": 2.6631858348846436,
"learning_rate": 3.7901639344262294e-05,
"loss": 0.1681,
"step": 24820
},
{
"epoch": 121.24466137888956,
"grad_norm": 4.315155982971191,
"learning_rate": 3.777049180327869e-05,
"loss": 0.2037,
"step": 24840
},
{
"epoch": 121.34228187919463,
"grad_norm": 3.0183377265930176,
"learning_rate": 3.7639344262295086e-05,
"loss": 0.1901,
"step": 24860
},
{
"epoch": 121.4399023794997,
"grad_norm": 2.389106035232544,
"learning_rate": 3.7508196721311476e-05,
"loss": 0.2162,
"step": 24880
},
{
"epoch": 121.53752287980475,
"grad_norm": 3.5406622886657715,
"learning_rate": 3.737704918032787e-05,
"loss": 0.1887,
"step": 24900
},
{
"epoch": 121.63514338010982,
"grad_norm": 3.272151231765747,
"learning_rate": 3.724590163934427e-05,
"loss": 0.2156,
"step": 24920
},
{
"epoch": 121.7327638804149,
"grad_norm": 3.1407511234283447,
"learning_rate": 3.711475409836066e-05,
"loss": 0.1833,
"step": 24940
},
{
"epoch": 121.83038438071995,
"grad_norm": 3.3532192707061768,
"learning_rate": 3.698360655737705e-05,
"loss": 0.1858,
"step": 24960
},
{
"epoch": 121.92800488102502,
"grad_norm": 3.1784684658050537,
"learning_rate": 3.685245901639344e-05,
"loss": 0.2,
"step": 24980
},
{
"epoch": 122.02562538133007,
"grad_norm": 3.5837671756744385,
"learning_rate": 3.672131147540984e-05,
"loss": 0.2073,
"step": 25000
},
{
"epoch": 122.12324588163514,
"grad_norm": 3.2040927410125732,
"learning_rate": 3.659016393442623e-05,
"loss": 0.1712,
"step": 25020
},
{
"epoch": 122.22086638194021,
"grad_norm": 3.3119306564331055,
"learning_rate": 3.6459016393442625e-05,
"loss": 0.1879,
"step": 25040
},
{
"epoch": 122.31848688224527,
"grad_norm": 3.2154297828674316,
"learning_rate": 3.632786885245902e-05,
"loss": 0.1976,
"step": 25060
},
{
"epoch": 122.41610738255034,
"grad_norm": 2.9070613384246826,
"learning_rate": 3.619672131147541e-05,
"loss": 0.2077,
"step": 25080
},
{
"epoch": 122.5137278828554,
"grad_norm": 3.866682529449463,
"learning_rate": 3.6065573770491806e-05,
"loss": 0.1673,
"step": 25100
},
{
"epoch": 122.61134838316046,
"grad_norm": 4.236519813537598,
"learning_rate": 3.59344262295082e-05,
"loss": 0.2096,
"step": 25120
},
{
"epoch": 122.70896888346553,
"grad_norm": 3.386503219604492,
"learning_rate": 3.580327868852459e-05,
"loss": 0.1962,
"step": 25140
},
{
"epoch": 122.80658938377059,
"grad_norm": 2.7001540660858154,
"learning_rate": 3.567213114754099e-05,
"loss": 0.1731,
"step": 25160
},
{
"epoch": 122.90420988407566,
"grad_norm": 3.281818389892578,
"learning_rate": 3.554098360655738e-05,
"loss": 0.2225,
"step": 25180
},
{
"epoch": 123.00183038438072,
"grad_norm": 4.1742095947265625,
"learning_rate": 3.5409836065573773e-05,
"loss": 0.1947,
"step": 25200
},
{
"epoch": 123.09945088468578,
"grad_norm": 3.328521490097046,
"learning_rate": 3.527868852459016e-05,
"loss": 0.1607,
"step": 25220
},
{
"epoch": 123.19707138499085,
"grad_norm": 4.815507888793945,
"learning_rate": 3.514754098360656e-05,
"loss": 0.1977,
"step": 25240
},
{
"epoch": 123.29469188529592,
"grad_norm": 3.736438512802124,
"learning_rate": 3.5016393442622955e-05,
"loss": 0.2105,
"step": 25260
},
{
"epoch": 123.39231238560097,
"grad_norm": 3.4552454948425293,
"learning_rate": 3.4885245901639344e-05,
"loss": 0.2002,
"step": 25280
},
{
"epoch": 123.48993288590604,
"grad_norm": 2.3649439811706543,
"learning_rate": 3.475409836065574e-05,
"loss": 0.1823,
"step": 25300
},
{
"epoch": 123.5875533862111,
"grad_norm": 3.0043177604675293,
"learning_rate": 3.462295081967214e-05,
"loss": 0.1922,
"step": 25320
},
{
"epoch": 123.68517388651617,
"grad_norm": 4.2065253257751465,
"learning_rate": 3.4491803278688526e-05,
"loss": 0.1943,
"step": 25340
},
{
"epoch": 123.78279438682124,
"grad_norm": 3.961331605911255,
"learning_rate": 3.436065573770492e-05,
"loss": 0.1859,
"step": 25360
},
{
"epoch": 123.88041488712629,
"grad_norm": 4.434045791625977,
"learning_rate": 3.422950819672131e-05,
"loss": 0.1985,
"step": 25380
},
{
"epoch": 123.97803538743136,
"grad_norm": 2.87196946144104,
"learning_rate": 3.409836065573771e-05,
"loss": 0.2046,
"step": 25400
},
{
"epoch": 124.07565588773643,
"grad_norm": 2.9702465534210205,
"learning_rate": 3.39672131147541e-05,
"loss": 0.1665,
"step": 25420
},
{
"epoch": 124.17327638804149,
"grad_norm": 4.106092929840088,
"learning_rate": 3.383606557377049e-05,
"loss": 0.1966,
"step": 25440
},
{
"epoch": 124.27089688834656,
"grad_norm": 3.1091065406799316,
"learning_rate": 3.370491803278689e-05,
"loss": 0.2078,
"step": 25460
},
{
"epoch": 124.36851738865161,
"grad_norm": 3.4036500453948975,
"learning_rate": 3.357377049180328e-05,
"loss": 0.1814,
"step": 25480
},
{
"epoch": 124.46613788895668,
"grad_norm": 3.7814347743988037,
"learning_rate": 3.3442622950819675e-05,
"loss": 0.195,
"step": 25500
},
{
"epoch": 124.56375838926175,
"grad_norm": 4.475528240203857,
"learning_rate": 3.331147540983607e-05,
"loss": 0.1984,
"step": 25520
},
{
"epoch": 124.6613788895668,
"grad_norm": 3.35787296295166,
"learning_rate": 3.318032786885246e-05,
"loss": 0.1809,
"step": 25540
},
{
"epoch": 124.75899938987187,
"grad_norm": 3.594639301300049,
"learning_rate": 3.3049180327868857e-05,
"loss": 0.1844,
"step": 25560
},
{
"epoch": 124.85661989017694,
"grad_norm": 3.5354974269866943,
"learning_rate": 3.291803278688525e-05,
"loss": 0.1953,
"step": 25580
},
{
"epoch": 124.954240390482,
"grad_norm": 2.810798168182373,
"learning_rate": 3.2786885245901635e-05,
"loss": 0.1911,
"step": 25600
},
{
"epoch": 125.05186089078707,
"grad_norm": 3.3332059383392334,
"learning_rate": 3.265573770491803e-05,
"loss": 0.1593,
"step": 25620
},
{
"epoch": 125.14948139109212,
"grad_norm": 2.8333628177642822,
"learning_rate": 3.252459016393443e-05,
"loss": 0.1709,
"step": 25640
},
{
"epoch": 125.2471018913972,
"grad_norm": 2.9144675731658936,
"learning_rate": 3.2393442622950824e-05,
"loss": 0.2073,
"step": 25660
},
{
"epoch": 125.34472239170226,
"grad_norm": 3.185001850128174,
"learning_rate": 3.226229508196721e-05,
"loss": 0.1665,
"step": 25680
},
{
"epoch": 125.44234289200732,
"grad_norm": 2.8228659629821777,
"learning_rate": 3.213114754098361e-05,
"loss": 0.163,
"step": 25700
},
{
"epoch": 125.53996339231239,
"grad_norm": 2.8857967853546143,
"learning_rate": 3.2000000000000005e-05,
"loss": 0.2057,
"step": 25720
},
{
"epoch": 125.63758389261746,
"grad_norm": 2.9855597019195557,
"learning_rate": 3.1868852459016395e-05,
"loss": 0.1872,
"step": 25740
},
{
"epoch": 125.73520439292251,
"grad_norm": 3.5702884197235107,
"learning_rate": 3.173770491803279e-05,
"loss": 0.1992,
"step": 25760
},
{
"epoch": 125.83282489322758,
"grad_norm": 5.302943706512451,
"learning_rate": 3.160655737704919e-05,
"loss": 0.1995,
"step": 25780
},
{
"epoch": 125.93044539353264,
"grad_norm": 3.9966931343078613,
"learning_rate": 3.1475409836065576e-05,
"loss": 0.2046,
"step": 25800
},
{
"epoch": 126.0280658938377,
"grad_norm": 2.6985690593719482,
"learning_rate": 3.1344262295081966e-05,
"loss": 0.2084,
"step": 25820
},
{
"epoch": 126.12568639414278,
"grad_norm": 3.317439079284668,
"learning_rate": 3.121311475409836e-05,
"loss": 0.1971,
"step": 25840
},
{
"epoch": 126.22330689444783,
"grad_norm": 3.718867301940918,
"learning_rate": 3.108196721311475e-05,
"loss": 0.1759,
"step": 25860
},
{
"epoch": 126.3209273947529,
"grad_norm": 3.7418856620788574,
"learning_rate": 3.095081967213115e-05,
"loss": 0.2067,
"step": 25880
},
{
"epoch": 126.41854789505797,
"grad_norm": 3.9287869930267334,
"learning_rate": 3.0819672131147544e-05,
"loss": 0.2032,
"step": 25900
},
{
"epoch": 126.51616839536302,
"grad_norm": 1.9036474227905273,
"learning_rate": 3.068852459016393e-05,
"loss": 0.1905,
"step": 25920
},
{
"epoch": 126.6137888956681,
"grad_norm": 2.602092981338501,
"learning_rate": 3.055737704918033e-05,
"loss": 0.1851,
"step": 25940
},
{
"epoch": 126.71140939597315,
"grad_norm": 3.2280685901641846,
"learning_rate": 3.0426229508196725e-05,
"loss": 0.1938,
"step": 25960
},
{
"epoch": 126.80902989627822,
"grad_norm": 4.606971263885498,
"learning_rate": 3.0295081967213118e-05,
"loss": 0.1644,
"step": 25980
},
{
"epoch": 126.90665039658329,
"grad_norm": 3.755833864212036,
"learning_rate": 3.016393442622951e-05,
"loss": 0.1836,
"step": 26000
},
{
"epoch": 127.00427089688834,
"grad_norm": 3.572577953338623,
"learning_rate": 3.00327868852459e-05,
"loss": 0.1733,
"step": 26020
},
{
"epoch": 127.10189139719341,
"grad_norm": 3.1435470581054688,
"learning_rate": 2.9901639344262293e-05,
"loss": 0.1815,
"step": 26040
},
{
"epoch": 127.19951189749847,
"grad_norm": 3.3165206909179688,
"learning_rate": 2.977049180327869e-05,
"loss": 0.183,
"step": 26060
},
{
"epoch": 127.29713239780354,
"grad_norm": 3.269935131072998,
"learning_rate": 2.963934426229508e-05,
"loss": 0.1854,
"step": 26080
},
{
"epoch": 127.3947528981086,
"grad_norm": 3.6275577545166016,
"learning_rate": 2.9508196721311478e-05,
"loss": 0.1798,
"step": 26100
},
{
"epoch": 127.49237339841366,
"grad_norm": 3.3832483291625977,
"learning_rate": 2.937704918032787e-05,
"loss": 0.1722,
"step": 26120
},
{
"epoch": 127.58999389871873,
"grad_norm": 3.828364610671997,
"learning_rate": 2.9245901639344263e-05,
"loss": 0.1853,
"step": 26140
},
{
"epoch": 127.6876143990238,
"grad_norm": 3.7207860946655273,
"learning_rate": 2.911475409836066e-05,
"loss": 0.1948,
"step": 26160
},
{
"epoch": 127.78523489932886,
"grad_norm": 3.281031847000122,
"learning_rate": 2.8983606557377052e-05,
"loss": 0.2063,
"step": 26180
},
{
"epoch": 127.88285539963393,
"grad_norm": 3.1116421222686768,
"learning_rate": 2.8852459016393445e-05,
"loss": 0.191,
"step": 26200
},
{
"epoch": 127.98047589993898,
"grad_norm": 4.254022121429443,
"learning_rate": 2.872131147540984e-05,
"loss": 0.202,
"step": 26220
},
{
"epoch": 128.07809640024405,
"grad_norm": 3.1279819011688232,
"learning_rate": 2.8590163934426227e-05,
"loss": 0.1934,
"step": 26240
},
{
"epoch": 128.1757169005491,
"grad_norm": 3.3977596759796143,
"learning_rate": 2.8459016393442623e-05,
"loss": 0.1807,
"step": 26260
},
{
"epoch": 128.2733374008542,
"grad_norm": 5.261218070983887,
"learning_rate": 2.8327868852459016e-05,
"loss": 0.17,
"step": 26280
},
{
"epoch": 128.37095790115924,
"grad_norm": 4.153654098510742,
"learning_rate": 2.819672131147541e-05,
"loss": 0.1891,
"step": 26300
},
{
"epoch": 128.4685784014643,
"grad_norm": 3.449397563934326,
"learning_rate": 2.8065573770491805e-05,
"loss": 0.1846,
"step": 26320
},
{
"epoch": 128.56619890176938,
"grad_norm": 3.43442702293396,
"learning_rate": 2.7934426229508198e-05,
"loss": 0.1807,
"step": 26340
},
{
"epoch": 128.66381940207444,
"grad_norm": 2.9243948459625244,
"learning_rate": 2.7803278688524594e-05,
"loss": 0.1704,
"step": 26360
},
{
"epoch": 128.7614399023795,
"grad_norm": 3.9830613136291504,
"learning_rate": 2.7672131147540987e-05,
"loss": 0.1909,
"step": 26380
},
{
"epoch": 128.85906040268458,
"grad_norm": 3.0765368938446045,
"learning_rate": 2.754098360655738e-05,
"loss": 0.1954,
"step": 26400
},
{
"epoch": 128.95668090298963,
"grad_norm": 2.540853977203369,
"learning_rate": 2.7409836065573775e-05,
"loss": 0.2116,
"step": 26420
},
{
"epoch": 129.0543014032947,
"grad_norm": 3.55985689163208,
"learning_rate": 2.7278688524590168e-05,
"loss": 0.193,
"step": 26440
},
{
"epoch": 129.15192190359974,
"grad_norm": 3.974700450897217,
"learning_rate": 2.7147540983606558e-05,
"loss": 0.1673,
"step": 26460
},
{
"epoch": 129.24954240390483,
"grad_norm": 4.614022731781006,
"learning_rate": 2.701639344262295e-05,
"loss": 0.1968,
"step": 26480
},
{
"epoch": 129.34716290420988,
"grad_norm": 3.9277889728546143,
"learning_rate": 2.6885245901639343e-05,
"loss": 0.1643,
"step": 26500
},
{
"epoch": 129.44478340451494,
"grad_norm": 3.4351913928985596,
"learning_rate": 2.675409836065574e-05,
"loss": 0.1903,
"step": 26520
},
{
"epoch": 129.54240390482002,
"grad_norm": 3.3347392082214355,
"learning_rate": 2.6622950819672132e-05,
"loss": 0.1789,
"step": 26540
},
{
"epoch": 129.64002440512508,
"grad_norm": 3.970414161682129,
"learning_rate": 2.6491803278688525e-05,
"loss": 0.1994,
"step": 26560
},
{
"epoch": 129.73764490543013,
"grad_norm": 3.648883819580078,
"learning_rate": 2.636065573770492e-05,
"loss": 0.1877,
"step": 26580
},
{
"epoch": 129.8352654057352,
"grad_norm": 3.3394792079925537,
"learning_rate": 2.6229508196721314e-05,
"loss": 0.1838,
"step": 26600
},
{
"epoch": 129.93288590604027,
"grad_norm": 2.924798011779785,
"learning_rate": 2.6098360655737706e-05,
"loss": 0.1913,
"step": 26620
},
{
"epoch": 130.03050640634532,
"grad_norm": 3.448457717895508,
"learning_rate": 2.5967213114754103e-05,
"loss": 0.1993,
"step": 26640
},
{
"epoch": 130.1281269066504,
"grad_norm": 2.458868980407715,
"learning_rate": 2.5836065573770492e-05,
"loss": 0.1996,
"step": 26660
},
{
"epoch": 130.22574740695546,
"grad_norm": 4.39287805557251,
"learning_rate": 2.5704918032786885e-05,
"loss": 0.1894,
"step": 26680
},
{
"epoch": 130.32336790726052,
"grad_norm": 3.347745180130005,
"learning_rate": 2.5573770491803277e-05,
"loss": 0.1642,
"step": 26700
},
{
"epoch": 130.4209884075656,
"grad_norm": 3.0466248989105225,
"learning_rate": 2.5442622950819674e-05,
"loss": 0.1976,
"step": 26720
},
{
"epoch": 130.51860890787066,
"grad_norm": 2.580834150314331,
"learning_rate": 2.5311475409836066e-05,
"loss": 0.1764,
"step": 26740
},
{
"epoch": 130.6162294081757,
"grad_norm": 3.6775128841400146,
"learning_rate": 2.518032786885246e-05,
"loss": 0.1725,
"step": 26760
},
{
"epoch": 130.71384990848077,
"grad_norm": 3.829058885574341,
"learning_rate": 2.5049180327868855e-05,
"loss": 0.1776,
"step": 26780
},
{
"epoch": 130.81147040878585,
"grad_norm": 4.047943592071533,
"learning_rate": 2.4918032786885248e-05,
"loss": 0.1943,
"step": 26800
},
{
"epoch": 130.9090909090909,
"grad_norm": 3.2828705310821533,
"learning_rate": 2.478688524590164e-05,
"loss": 0.1864,
"step": 26820
},
{
"epoch": 131.00671140939596,
"grad_norm": 4.2563958168029785,
"learning_rate": 2.4655737704918033e-05,
"loss": 0.1838,
"step": 26840
},
{
"epoch": 131.10433190970105,
"grad_norm": 3.279503583908081,
"learning_rate": 2.4524590163934426e-05,
"loss": 0.2027,
"step": 26860
},
{
"epoch": 131.2019524100061,
"grad_norm": 3.8052897453308105,
"learning_rate": 2.4393442622950822e-05,
"loss": 0.1916,
"step": 26880
},
{
"epoch": 131.29957291031116,
"grad_norm": 3.12294602394104,
"learning_rate": 2.4262295081967215e-05,
"loss": 0.1542,
"step": 26900
},
{
"epoch": 131.39719341061624,
"grad_norm": 4.636548042297363,
"learning_rate": 2.4131147540983608e-05,
"loss": 0.1677,
"step": 26920
},
{
"epoch": 131.4948139109213,
"grad_norm": 2.8608415126800537,
"learning_rate": 2.4e-05,
"loss": 0.194,
"step": 26940
},
{
"epoch": 131.59243441122635,
"grad_norm": 3.7946908473968506,
"learning_rate": 2.3868852459016393e-05,
"loss": 0.1764,
"step": 26960
},
{
"epoch": 131.69005491153143,
"grad_norm": 3.1568832397460938,
"learning_rate": 2.373770491803279e-05,
"loss": 0.1865,
"step": 26980
},
{
"epoch": 131.7876754118365,
"grad_norm": 4.103198528289795,
"learning_rate": 2.3606557377049182e-05,
"loss": 0.1753,
"step": 27000
},
{
"epoch": 131.88529591214154,
"grad_norm": 3.655327796936035,
"learning_rate": 2.3475409836065575e-05,
"loss": 0.1965,
"step": 27020
},
{
"epoch": 131.98291641244663,
"grad_norm": 3.908200263977051,
"learning_rate": 2.3344262295081968e-05,
"loss": 0.1791,
"step": 27040
},
{
"epoch": 132.08053691275168,
"grad_norm": 3.2557129859924316,
"learning_rate": 2.321311475409836e-05,
"loss": 0.1715,
"step": 27060
},
{
"epoch": 132.17815741305674,
"grad_norm": 3.102268695831299,
"learning_rate": 2.3081967213114757e-05,
"loss": 0.191,
"step": 27080
},
{
"epoch": 132.2757779133618,
"grad_norm": 2.7028493881225586,
"learning_rate": 2.295081967213115e-05,
"loss": 0.1802,
"step": 27100
},
{
"epoch": 132.37339841366688,
"grad_norm": 2.5240046977996826,
"learning_rate": 2.2819672131147542e-05,
"loss": 0.1618,
"step": 27120
},
{
"epoch": 132.47101891397193,
"grad_norm": 3.0410265922546387,
"learning_rate": 2.2688524590163935e-05,
"loss": 0.1833,
"step": 27140
},
{
"epoch": 132.568639414277,
"grad_norm": 3.677824020385742,
"learning_rate": 2.2557377049180328e-05,
"loss": 0.1804,
"step": 27160
},
{
"epoch": 132.66625991458207,
"grad_norm": 2.826828718185425,
"learning_rate": 2.2426229508196724e-05,
"loss": 0.1915,
"step": 27180
},
{
"epoch": 132.76388041488713,
"grad_norm": 4.393260955810547,
"learning_rate": 2.2295081967213117e-05,
"loss": 0.1952,
"step": 27200
},
{
"epoch": 132.86150091519218,
"grad_norm": 3.1146349906921387,
"learning_rate": 2.216393442622951e-05,
"loss": 0.1973,
"step": 27220
},
{
"epoch": 132.95912141549726,
"grad_norm": 3.343693494796753,
"learning_rate": 2.2032786885245905e-05,
"loss": 0.1822,
"step": 27240
},
{
"epoch": 133.05674191580232,
"grad_norm": 3.0667552947998047,
"learning_rate": 2.1901639344262295e-05,
"loss": 0.1807,
"step": 27260
},
{
"epoch": 133.15436241610738,
"grad_norm": 3.497859001159668,
"learning_rate": 2.1770491803278688e-05,
"loss": 0.1525,
"step": 27280
},
{
"epoch": 133.25198291641246,
"grad_norm": 3.295478343963623,
"learning_rate": 2.1639344262295084e-05,
"loss": 0.1971,
"step": 27300
},
{
"epoch": 133.3496034167175,
"grad_norm": 3.3247010707855225,
"learning_rate": 2.1508196721311476e-05,
"loss": 0.1963,
"step": 27320
},
{
"epoch": 133.44722391702257,
"grad_norm": 4.269167900085449,
"learning_rate": 2.1377049180327873e-05,
"loss": 0.1967,
"step": 27340
},
{
"epoch": 133.54484441732765,
"grad_norm": 3.7575721740722656,
"learning_rate": 2.1245901639344262e-05,
"loss": 0.1788,
"step": 27360
},
{
"epoch": 133.6424649176327,
"grad_norm": 4.189979553222656,
"learning_rate": 2.1114754098360655e-05,
"loss": 0.1922,
"step": 27380
},
{
"epoch": 133.74008541793776,
"grad_norm": 3.48610782623291,
"learning_rate": 2.098360655737705e-05,
"loss": 0.1749,
"step": 27400
},
{
"epoch": 133.83770591824282,
"grad_norm": 3.320037364959717,
"learning_rate": 2.0852459016393444e-05,
"loss": 0.1886,
"step": 27420
},
{
"epoch": 133.9353264185479,
"grad_norm": 3.347099781036377,
"learning_rate": 2.0721311475409836e-05,
"loss": 0.1717,
"step": 27440
},
{
"epoch": 134.03294691885296,
"grad_norm": 2.9406418800354004,
"learning_rate": 2.059016393442623e-05,
"loss": 0.1702,
"step": 27460
},
{
"epoch": 134.130567419158,
"grad_norm": 2.3794620037078857,
"learning_rate": 2.0459016393442622e-05,
"loss": 0.1872,
"step": 27480
},
{
"epoch": 134.2281879194631,
"grad_norm": 3.1150660514831543,
"learning_rate": 2.0327868852459018e-05,
"loss": 0.2023,
"step": 27500
},
{
"epoch": 134.32580841976815,
"grad_norm": 3.679694414138794,
"learning_rate": 2.019672131147541e-05,
"loss": 0.1858,
"step": 27520
},
{
"epoch": 134.4234289200732,
"grad_norm": 2.665882110595703,
"learning_rate": 2.0065573770491804e-05,
"loss": 0.1782,
"step": 27540
},
{
"epoch": 134.5210494203783,
"grad_norm": 3.0053212642669678,
"learning_rate": 1.99344262295082e-05,
"loss": 0.1773,
"step": 27560
},
{
"epoch": 134.61866992068335,
"grad_norm": 2.689307451248169,
"learning_rate": 1.980327868852459e-05,
"loss": 0.1588,
"step": 27580
},
{
"epoch": 134.7162904209884,
"grad_norm": 4.318088054656982,
"learning_rate": 1.9672131147540985e-05,
"loss": 0.1436,
"step": 27600
},
{
"epoch": 134.81391092129348,
"grad_norm": 3.3378102779388428,
"learning_rate": 1.9540983606557378e-05,
"loss": 0.1919,
"step": 27620
},
{
"epoch": 134.91153142159854,
"grad_norm": 3.2355871200561523,
"learning_rate": 1.940983606557377e-05,
"loss": 0.1828,
"step": 27640
},
{
"epoch": 135.0091519219036,
"grad_norm": 3.1335229873657227,
"learning_rate": 1.9278688524590167e-05,
"loss": 0.1982,
"step": 27660
},
{
"epoch": 135.10677242220865,
"grad_norm": 4.066319465637207,
"learning_rate": 1.9147540983606556e-05,
"loss": 0.1526,
"step": 27680
},
{
"epoch": 135.20439292251373,
"grad_norm": 3.631089925765991,
"learning_rate": 1.9016393442622952e-05,
"loss": 0.1776,
"step": 27700
},
{
"epoch": 135.3020134228188,
"grad_norm": 3.5840327739715576,
"learning_rate": 1.8885245901639345e-05,
"loss": 0.1922,
"step": 27720
},
{
"epoch": 135.39963392312384,
"grad_norm": 2.926558256149292,
"learning_rate": 1.8754098360655738e-05,
"loss": 0.1847,
"step": 27740
},
{
"epoch": 135.49725442342893,
"grad_norm": 4.487957000732422,
"learning_rate": 1.8622950819672134e-05,
"loss": 0.1896,
"step": 27760
},
{
"epoch": 135.59487492373398,
"grad_norm": 3.209500789642334,
"learning_rate": 1.8491803278688523e-05,
"loss": 0.1827,
"step": 27780
},
{
"epoch": 135.69249542403904,
"grad_norm": 2.8735058307647705,
"learning_rate": 1.836065573770492e-05,
"loss": 0.1671,
"step": 27800
},
{
"epoch": 135.79011592434412,
"grad_norm": 3.221266508102417,
"learning_rate": 1.8229508196721312e-05,
"loss": 0.1861,
"step": 27820
},
{
"epoch": 135.88773642464918,
"grad_norm": 3.1269659996032715,
"learning_rate": 1.8098360655737705e-05,
"loss": 0.1857,
"step": 27840
},
{
"epoch": 135.98535692495423,
"grad_norm": 3.3478143215179443,
"learning_rate": 1.79672131147541e-05,
"loss": 0.1853,
"step": 27860
},
{
"epoch": 136.08297742525932,
"grad_norm": 2.8102643489837646,
"learning_rate": 1.7836065573770494e-05,
"loss": 0.1691,
"step": 27880
},
{
"epoch": 136.18059792556437,
"grad_norm": 2.9172909259796143,
"learning_rate": 1.7704918032786887e-05,
"loss": 0.1805,
"step": 27900
},
{
"epoch": 136.27821842586943,
"grad_norm": 3.6803557872772217,
"learning_rate": 1.757377049180328e-05,
"loss": 0.1808,
"step": 27920
},
{
"epoch": 136.3758389261745,
"grad_norm": 3.6122121810913086,
"learning_rate": 1.7442622950819672e-05,
"loss": 0.1825,
"step": 27940
},
{
"epoch": 136.47345942647956,
"grad_norm": 2.836901903152466,
"learning_rate": 1.731147540983607e-05,
"loss": 0.1762,
"step": 27960
},
{
"epoch": 136.57107992678462,
"grad_norm": 4.931893348693848,
"learning_rate": 1.718032786885246e-05,
"loss": 0.2097,
"step": 27980
},
{
"epoch": 136.66870042708968,
"grad_norm": 3.634223461151123,
"learning_rate": 1.7049180327868854e-05,
"loss": 0.1764,
"step": 28000
},
{
"epoch": 136.76632092739476,
"grad_norm": 4.417370319366455,
"learning_rate": 1.6918032786885247e-05,
"loss": 0.1695,
"step": 28020
},
{
"epoch": 136.8639414276998,
"grad_norm": 3.6770431995391846,
"learning_rate": 1.678688524590164e-05,
"loss": 0.1762,
"step": 28040
},
{
"epoch": 136.96156192800487,
"grad_norm": 3.134272813796997,
"learning_rate": 1.6655737704918036e-05,
"loss": 0.1603,
"step": 28060
},
{
"epoch": 137.05918242830995,
"grad_norm": 3.993882179260254,
"learning_rate": 1.6524590163934428e-05,
"loss": 0.1927,
"step": 28080
},
{
"epoch": 137.156802928615,
"grad_norm": 3.1614527702331543,
"learning_rate": 1.6393442622950818e-05,
"loss": 0.1739,
"step": 28100
},
{
"epoch": 137.25442342892006,
"grad_norm": 2.9293642044067383,
"learning_rate": 1.6262295081967214e-05,
"loss": 0.1916,
"step": 28120
},
{
"epoch": 137.35204392922515,
"grad_norm": 3.292917251586914,
"learning_rate": 1.6131147540983607e-05,
"loss": 0.1638,
"step": 28140
},
{
"epoch": 137.4496644295302,
"grad_norm": 3.048471212387085,
"learning_rate": 1.6000000000000003e-05,
"loss": 0.1877,
"step": 28160
},
{
"epoch": 137.54728492983526,
"grad_norm": 3.398252487182617,
"learning_rate": 1.5868852459016395e-05,
"loss": 0.1685,
"step": 28180
},
{
"epoch": 137.64490543014034,
"grad_norm": 3.906764268875122,
"learning_rate": 1.5737704918032788e-05,
"loss": 0.1772,
"step": 28200
},
{
"epoch": 137.7425259304454,
"grad_norm": 3.1852598190307617,
"learning_rate": 1.560655737704918e-05,
"loss": 0.168,
"step": 28220
},
{
"epoch": 137.84014643075045,
"grad_norm": 2.9285385608673096,
"learning_rate": 1.5475409836065574e-05,
"loss": 0.1733,
"step": 28240
},
{
"epoch": 137.93776693105553,
"grad_norm": 3.9846913814544678,
"learning_rate": 1.5344262295081966e-05,
"loss": 0.1906,
"step": 28260
},
{
"epoch": 138.0353874313606,
"grad_norm": 3.3251185417175293,
"learning_rate": 1.5213114754098363e-05,
"loss": 0.1872,
"step": 28280
},
{
"epoch": 138.13300793166565,
"grad_norm": 3.7904114723205566,
"learning_rate": 1.5081967213114755e-05,
"loss": 0.1811,
"step": 28300
},
{
"epoch": 138.2306284319707,
"grad_norm": 3.5363810062408447,
"learning_rate": 1.4950819672131146e-05,
"loss": 0.1803,
"step": 28320
},
{
"epoch": 138.32824893227578,
"grad_norm": 3.430577516555786,
"learning_rate": 1.481967213114754e-05,
"loss": 0.1619,
"step": 28340
},
{
"epoch": 138.42586943258084,
"grad_norm": 3.1190154552459717,
"learning_rate": 1.4688524590163935e-05,
"loss": 0.1909,
"step": 28360
},
{
"epoch": 138.5234899328859,
"grad_norm": 2.656212091445923,
"learning_rate": 1.455737704918033e-05,
"loss": 0.1682,
"step": 28380
},
{
"epoch": 138.62111043319098,
"grad_norm": 3.1637041568756104,
"learning_rate": 1.4426229508196722e-05,
"loss": 0.1729,
"step": 28400
},
{
"epoch": 138.71873093349603,
"grad_norm": 3.313136100769043,
"learning_rate": 1.4295081967213114e-05,
"loss": 0.1853,
"step": 28420
},
{
"epoch": 138.8163514338011,
"grad_norm": 3.63885498046875,
"learning_rate": 1.4163934426229508e-05,
"loss": 0.1783,
"step": 28440
},
{
"epoch": 138.91397193410617,
"grad_norm": 4.251205921173096,
"learning_rate": 1.4032786885245902e-05,
"loss": 0.1703,
"step": 28460
},
{
"epoch": 139.01159243441123,
"grad_norm": 3.76887583732605,
"learning_rate": 1.3901639344262297e-05,
"loss": 0.1871,
"step": 28480
},
{
"epoch": 139.10921293471628,
"grad_norm": 3.2986671924591064,
"learning_rate": 1.377049180327869e-05,
"loss": 0.1728,
"step": 28500
},
{
"epoch": 139.20683343502137,
"grad_norm": 2.5212690830230713,
"learning_rate": 1.3639344262295084e-05,
"loss": 0.175,
"step": 28520
},
{
"epoch": 139.30445393532642,
"grad_norm": 4.492109298706055,
"learning_rate": 1.3508196721311475e-05,
"loss": 0.1838,
"step": 28540
},
{
"epoch": 139.40207443563148,
"grad_norm": 3.0226120948791504,
"learning_rate": 1.337704918032787e-05,
"loss": 0.1753,
"step": 28560
},
{
"epoch": 139.49969493593656,
"grad_norm": 2.4843361377716064,
"learning_rate": 1.3245901639344262e-05,
"loss": 0.1915,
"step": 28580
},
{
"epoch": 139.59731543624162,
"grad_norm": 3.4304590225219727,
"learning_rate": 1.3114754098360657e-05,
"loss": 0.1703,
"step": 28600
},
{
"epoch": 139.69493593654667,
"grad_norm": 3.242751121520996,
"learning_rate": 1.2983606557377051e-05,
"loss": 0.176,
"step": 28620
},
{
"epoch": 139.79255643685173,
"grad_norm": 3.2713284492492676,
"learning_rate": 1.2852459016393442e-05,
"loss": 0.1838,
"step": 28640
},
{
"epoch": 139.8901769371568,
"grad_norm": 2.9921929836273193,
"learning_rate": 1.2721311475409837e-05,
"loss": 0.1764,
"step": 28660
},
{
"epoch": 139.98779743746186,
"grad_norm": 3.6933350563049316,
"learning_rate": 1.259016393442623e-05,
"loss": 0.1803,
"step": 28680
},
{
"epoch": 140.08541793776692,
"grad_norm": 2.947892665863037,
"learning_rate": 1.2459016393442624e-05,
"loss": 0.1732,
"step": 28700
},
{
"epoch": 140.183038438072,
"grad_norm": 3.5047738552093506,
"learning_rate": 1.2327868852459017e-05,
"loss": 0.1751,
"step": 28720
},
{
"epoch": 140.28065893837706,
"grad_norm": 2.9315600395202637,
"learning_rate": 1.2196721311475411e-05,
"loss": 0.1879,
"step": 28740
},
{
"epoch": 140.3782794386821,
"grad_norm": 3.6878082752227783,
"learning_rate": 1.2065573770491804e-05,
"loss": 0.1551,
"step": 28760
},
{
"epoch": 140.4758999389872,
"grad_norm": 3.2162342071533203,
"learning_rate": 1.1934426229508197e-05,
"loss": 0.183,
"step": 28780
},
{
"epoch": 140.57352043929225,
"grad_norm": 3.6583456993103027,
"learning_rate": 1.1803278688524591e-05,
"loss": 0.1784,
"step": 28800
},
{
"epoch": 140.6711409395973,
"grad_norm": 2.780412435531616,
"learning_rate": 1.1672131147540984e-05,
"loss": 0.1885,
"step": 28820
},
{
"epoch": 140.7687614399024,
"grad_norm": 2.3773672580718994,
"learning_rate": 1.1540983606557378e-05,
"loss": 0.1611,
"step": 28840
},
{
"epoch": 140.86638194020745,
"grad_norm": 3.6276142597198486,
"learning_rate": 1.1409836065573771e-05,
"loss": 0.1806,
"step": 28860
},
{
"epoch": 140.9640024405125,
"grad_norm": 3.3313121795654297,
"learning_rate": 1.1278688524590164e-05,
"loss": 0.1741,
"step": 28880
},
{
"epoch": 141.06162294081759,
"grad_norm": 3.1109941005706787,
"learning_rate": 1.1147540983606558e-05,
"loss": 0.1747,
"step": 28900
},
{
"epoch": 141.15924344112264,
"grad_norm": 2.3083291053771973,
"learning_rate": 1.1016393442622953e-05,
"loss": 0.16,
"step": 28920
},
{
"epoch": 141.2568639414277,
"grad_norm": 3.6427536010742188,
"learning_rate": 1.0885245901639344e-05,
"loss": 0.1792,
"step": 28940
},
{
"epoch": 141.35448444173275,
"grad_norm": 3.2836146354675293,
"learning_rate": 1.0754098360655738e-05,
"loss": 0.1921,
"step": 28960
},
{
"epoch": 141.45210494203783,
"grad_norm": 3.712411642074585,
"learning_rate": 1.0622950819672131e-05,
"loss": 0.1834,
"step": 28980
},
{
"epoch": 141.5497254423429,
"grad_norm": 2.6064016819000244,
"learning_rate": 1.0491803278688525e-05,
"loss": 0.1975,
"step": 29000
},
{
"epoch": 141.64734594264795,
"grad_norm": 3.3907470703125,
"learning_rate": 1.0360655737704918e-05,
"loss": 0.172,
"step": 29020
},
{
"epoch": 141.74496644295303,
"grad_norm": 3.0713050365448,
"learning_rate": 1.0229508196721311e-05,
"loss": 0.1547,
"step": 29040
},
{
"epoch": 141.84258694325808,
"grad_norm": 5.218588352203369,
"learning_rate": 1.0098360655737705e-05,
"loss": 0.1712,
"step": 29060
},
{
"epoch": 141.94020744356314,
"grad_norm": 3.205132007598877,
"learning_rate": 9.9672131147541e-06,
"loss": 0.1629,
"step": 29080
},
{
"epoch": 142.03782794386822,
"grad_norm": 3.580003499984741,
"learning_rate": 9.836065573770493e-06,
"loss": 0.1877,
"step": 29100
},
{
"epoch": 142.13544844417328,
"grad_norm": 3.393789529800415,
"learning_rate": 9.704918032786885e-06,
"loss": 0.1928,
"step": 29120
},
{
"epoch": 142.23306894447833,
"grad_norm": 2.592445135116577,
"learning_rate": 9.573770491803278e-06,
"loss": 0.1638,
"step": 29140
},
{
"epoch": 142.33068944478342,
"grad_norm": 4.577868938446045,
"learning_rate": 9.442622950819673e-06,
"loss": 0.1868,
"step": 29160
},
{
"epoch": 142.42830994508847,
"grad_norm": 3.1620144844055176,
"learning_rate": 9.311475409836067e-06,
"loss": 0.1635,
"step": 29180
},
{
"epoch": 142.52593044539353,
"grad_norm": 3.382749319076538,
"learning_rate": 9.18032786885246e-06,
"loss": 0.1469,
"step": 29200
},
{
"epoch": 142.6235509456986,
"grad_norm": 3.314983606338501,
"learning_rate": 9.049180327868853e-06,
"loss": 0.1719,
"step": 29220
},
{
"epoch": 142.72117144600367,
"grad_norm": 2.643578290939331,
"learning_rate": 8.918032786885247e-06,
"loss": 0.1587,
"step": 29240
},
{
"epoch": 142.81879194630872,
"grad_norm": 2.4660592079162598,
"learning_rate": 8.78688524590164e-06,
"loss": 0.2031,
"step": 29260
},
{
"epoch": 142.91641244661378,
"grad_norm": 3.2867209911346436,
"learning_rate": 8.655737704918034e-06,
"loss": 0.1679,
"step": 29280
},
{
"epoch": 143.01403294691886,
"grad_norm": 3.2089104652404785,
"learning_rate": 8.524590163934427e-06,
"loss": 0.171,
"step": 29300
},
{
"epoch": 143.11165344722392,
"grad_norm": 2.9183740615844727,
"learning_rate": 8.39344262295082e-06,
"loss": 0.1642,
"step": 29320
},
{
"epoch": 143.20927394752897,
"grad_norm": 2.9158482551574707,
"learning_rate": 8.262295081967214e-06,
"loss": 0.1743,
"step": 29340
},
{
"epoch": 143.30689444783405,
"grad_norm": 3.256065607070923,
"learning_rate": 8.131147540983607e-06,
"loss": 0.1816,
"step": 29360
},
{
"epoch": 143.4045149481391,
"grad_norm": 2.916098117828369,
"learning_rate": 8.000000000000001e-06,
"loss": 0.1575,
"step": 29380
},
{
"epoch": 143.50213544844416,
"grad_norm": 3.414485454559326,
"learning_rate": 7.868852459016394e-06,
"loss": 0.1584,
"step": 29400
},
{
"epoch": 143.59975594874925,
"grad_norm": 3.5028018951416016,
"learning_rate": 7.737704918032787e-06,
"loss": 0.1731,
"step": 29420
},
{
"epoch": 143.6973764490543,
"grad_norm": 2.7573652267456055,
"learning_rate": 7.606557377049181e-06,
"loss": 0.1848,
"step": 29440
},
{
"epoch": 143.79499694935936,
"grad_norm": 3.1906893253326416,
"learning_rate": 7.475409836065573e-06,
"loss": 0.1842,
"step": 29460
},
{
"epoch": 143.89261744966444,
"grad_norm": 2.384742021560669,
"learning_rate": 7.344262295081968e-06,
"loss": 0.1852,
"step": 29480
},
{
"epoch": 143.9902379499695,
"grad_norm": 3.7481918334960938,
"learning_rate": 7.213114754098361e-06,
"loss": 0.1794,
"step": 29500
},
{
"epoch": 144.08785845027455,
"grad_norm": 2.903989791870117,
"learning_rate": 7.081967213114754e-06,
"loss": 0.1745,
"step": 29520
},
{
"epoch": 144.1854789505796,
"grad_norm": 3.4449713230133057,
"learning_rate": 6.9508196721311484e-06,
"loss": 0.1874,
"step": 29540
},
{
"epoch": 144.2830994508847,
"grad_norm": 2.9290127754211426,
"learning_rate": 6.819672131147542e-06,
"loss": 0.1736,
"step": 29560
},
{
"epoch": 144.38071995118975,
"grad_norm": 2.805908203125,
"learning_rate": 6.688524590163935e-06,
"loss": 0.187,
"step": 29580
},
{
"epoch": 144.4783404514948,
"grad_norm": 2.9539241790771484,
"learning_rate": 6.557377049180328e-06,
"loss": 0.1856,
"step": 29600
},
{
"epoch": 144.57596095179989,
"grad_norm": 2.8198204040527344,
"learning_rate": 6.426229508196721e-06,
"loss": 0.1733,
"step": 29620
},
{
"epoch": 144.67358145210494,
"grad_norm": 3.3926849365234375,
"learning_rate": 6.295081967213115e-06,
"loss": 0.18,
"step": 29640
},
{
"epoch": 144.77120195241,
"grad_norm": 4.100579261779785,
"learning_rate": 6.163934426229508e-06,
"loss": 0.1568,
"step": 29660
},
{
"epoch": 144.86882245271508,
"grad_norm": 3.2875492572784424,
"learning_rate": 6.032786885245902e-06,
"loss": 0.1561,
"step": 29680
},
{
"epoch": 144.96644295302013,
"grad_norm": 2.626185417175293,
"learning_rate": 5.9016393442622956e-06,
"loss": 0.1714,
"step": 29700
},
{
"epoch": 145.0640634533252,
"grad_norm": 4.3447265625,
"learning_rate": 5.770491803278689e-06,
"loss": 0.1951,
"step": 29720
},
{
"epoch": 145.16168395363027,
"grad_norm": 4.056821346282959,
"learning_rate": 5.639344262295082e-06,
"loss": 0.1621,
"step": 29740
},
{
"epoch": 145.25930445393533,
"grad_norm": 3.4116666316986084,
"learning_rate": 5.508196721311476e-06,
"loss": 0.179,
"step": 29760
},
{
"epoch": 145.35692495424038,
"grad_norm": 2.810452699661255,
"learning_rate": 5.377049180327869e-06,
"loss": 0.1723,
"step": 29780
},
{
"epoch": 145.45454545454547,
"grad_norm": 3.708115816116333,
"learning_rate": 5.245901639344263e-06,
"loss": 0.1675,
"step": 29800
},
{
"epoch": 145.55216595485052,
"grad_norm": 4.000546455383301,
"learning_rate": 5.1147540983606555e-06,
"loss": 0.1716,
"step": 29820
},
{
"epoch": 145.64978645515558,
"grad_norm": 3.2421109676361084,
"learning_rate": 4.98360655737705e-06,
"loss": 0.1656,
"step": 29840
},
{
"epoch": 145.74740695546063,
"grad_norm": 3.13706111907959,
"learning_rate": 4.852459016393443e-06,
"loss": 0.1822,
"step": 29860
},
{
"epoch": 145.84502745576572,
"grad_norm": 3.364842176437378,
"learning_rate": 4.721311475409836e-06,
"loss": 0.1772,
"step": 29880
},
{
"epoch": 145.94264795607077,
"grad_norm": 3.2013063430786133,
"learning_rate": 4.59016393442623e-06,
"loss": 0.1717,
"step": 29900
},
{
"epoch": 146.04026845637583,
"grad_norm": 4.174123287200928,
"learning_rate": 4.4590163934426235e-06,
"loss": 0.1731,
"step": 29920
},
{
"epoch": 146.1378889566809,
"grad_norm": 2.8885281085968018,
"learning_rate": 4.327868852459017e-06,
"loss": 0.1421,
"step": 29940
},
{
"epoch": 146.23550945698597,
"grad_norm": 2.6078240871429443,
"learning_rate": 4.19672131147541e-06,
"loss": 0.1898,
"step": 29960
},
{
"epoch": 146.33312995729102,
"grad_norm": 4.43600606918335,
"learning_rate": 4.0655737704918034e-06,
"loss": 0.1787,
"step": 29980
},
{
"epoch": 146.4307504575961,
"grad_norm": 2.9905123710632324,
"learning_rate": 3.934426229508197e-06,
"loss": 0.1866,
"step": 30000
},
{
"epoch": 146.52837095790116,
"grad_norm": 2.8058741092681885,
"learning_rate": 3.8032786885245906e-06,
"loss": 0.1751,
"step": 30020
},
{
"epoch": 146.62599145820622,
"grad_norm": 2.752978563308716,
"learning_rate": 3.672131147540984e-06,
"loss": 0.1826,
"step": 30040
},
{
"epoch": 146.7236119585113,
"grad_norm": 3.0315961837768555,
"learning_rate": 3.540983606557377e-06,
"loss": 0.1623,
"step": 30060
},
{
"epoch": 146.82123245881635,
"grad_norm": 3.3782765865325928,
"learning_rate": 3.409836065573771e-06,
"loss": 0.1784,
"step": 30080
},
{
"epoch": 146.9188529591214,
"grad_norm": 2.961002826690674,
"learning_rate": 3.278688524590164e-06,
"loss": 0.1764,
"step": 30100
},
{
"epoch": 147.0164734594265,
"grad_norm": 3.775484561920166,
"learning_rate": 3.1475409836065574e-06,
"loss": 0.1595,
"step": 30120
},
{
"epoch": 147.11409395973155,
"grad_norm": 4.18531608581543,
"learning_rate": 3.016393442622951e-06,
"loss": 0.1599,
"step": 30140
},
{
"epoch": 147.2117144600366,
"grad_norm": 2.9723432064056396,
"learning_rate": 2.8852459016393446e-06,
"loss": 0.159,
"step": 30160
},
{
"epoch": 147.30933496034166,
"grad_norm": 3.2833070755004883,
"learning_rate": 2.754098360655738e-06,
"loss": 0.1657,
"step": 30180
},
{
"epoch": 147.40695546064674,
"grad_norm": 3.4174959659576416,
"learning_rate": 2.6229508196721314e-06,
"loss": 0.175,
"step": 30200
},
{
"epoch": 147.5045759609518,
"grad_norm": 3.3127195835113525,
"learning_rate": 2.491803278688525e-06,
"loss": 0.1815,
"step": 30220
},
{
"epoch": 147.60219646125685,
"grad_norm": 3.7137949466705322,
"learning_rate": 2.360655737704918e-06,
"loss": 0.198,
"step": 30240
},
{
"epoch": 147.69981696156194,
"grad_norm": 2.630924701690674,
"learning_rate": 2.2295081967213117e-06,
"loss": 0.1687,
"step": 30260
},
{
"epoch": 147.797437461867,
"grad_norm": 3.330245018005371,
"learning_rate": 2.098360655737705e-06,
"loss": 0.1581,
"step": 30280
},
{
"epoch": 147.89505796217205,
"grad_norm": 3.237410068511963,
"learning_rate": 1.9672131147540985e-06,
"loss": 0.2022,
"step": 30300
},
{
"epoch": 147.99267846247713,
"grad_norm": 2.633331537246704,
"learning_rate": 1.836065573770492e-06,
"loss": 0.1613,
"step": 30320
},
{
"epoch": 148.09029896278219,
"grad_norm": 2.527902603149414,
"learning_rate": 1.7049180327868855e-06,
"loss": 0.1749,
"step": 30340
},
{
"epoch": 148.18791946308724,
"grad_norm": 2.9230234622955322,
"learning_rate": 1.5737704918032787e-06,
"loss": 0.1464,
"step": 30360
},
{
"epoch": 148.28553996339232,
"grad_norm": 2.591038703918457,
"learning_rate": 1.4426229508196723e-06,
"loss": 0.1819,
"step": 30380
},
{
"epoch": 148.38316046369738,
"grad_norm": 3.6826913356781006,
"learning_rate": 1.3114754098360657e-06,
"loss": 0.1909,
"step": 30400
},
{
"epoch": 148.48078096400243,
"grad_norm": 3.1828205585479736,
"learning_rate": 1.180327868852459e-06,
"loss": 0.1727,
"step": 30420
},
{
"epoch": 148.57840146430752,
"grad_norm": 3.3356974124908447,
"learning_rate": 1.0491803278688525e-06,
"loss": 0.1624,
"step": 30440
},
{
"epoch": 148.67602196461257,
"grad_norm": 3.1692721843719482,
"learning_rate": 9.18032786885246e-07,
"loss": 0.1769,
"step": 30460
},
{
"epoch": 148.77364246491763,
"grad_norm": 2.968018054962158,
"learning_rate": 7.868852459016393e-07,
"loss": 0.1594,
"step": 30480
},
{
"epoch": 148.87126296522268,
"grad_norm": 3.693136692047119,
"learning_rate": 6.557377049180328e-07,
"loss": 0.1927,
"step": 30500
}
],
"logging_steps": 20,
"max_steps": 30600,
"num_input_tokens_seen": 0,
"num_train_epochs": 150,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3.924112697660375e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}