sonnet-llama-3.2-3b / trainer_state.json
legionlm's picture
Upload trainer_state.json
b1e612b verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.22003725936659077,
"eval_steps": 500,
"global_step": 1063,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.00020699648105982198,
"grad_norm": 0.3105248212814331,
"learning_rate": 1e-05,
"loss": 0.3097,
"step": 1
},
{
"epoch": 0.00041399296211964395,
"grad_norm": 0.35843613743782043,
"learning_rate": 2e-05,
"loss": 0.3333,
"step": 2
},
{
"epoch": 0.000620989443179466,
"grad_norm": 0.31169694662094116,
"learning_rate": 3e-05,
"loss": 0.3367,
"step": 3
},
{
"epoch": 0.0008279859242392879,
"grad_norm": 0.4025513529777527,
"learning_rate": 4e-05,
"loss": 0.3568,
"step": 4
},
{
"epoch": 0.0010349824052991099,
"grad_norm": 0.38164222240448,
"learning_rate": 5e-05,
"loss": 0.375,
"step": 5
},
{
"epoch": 0.001241978886358932,
"grad_norm": 0.41811275482177734,
"learning_rate": 6e-05,
"loss": 0.3421,
"step": 6
},
{
"epoch": 0.001448975367418754,
"grad_norm": 0.40563058853149414,
"learning_rate": 7e-05,
"loss": 0.3046,
"step": 7
},
{
"epoch": 0.0016559718484785758,
"grad_norm": 0.32561907172203064,
"learning_rate": 8e-05,
"loss": 0.3298,
"step": 8
},
{
"epoch": 0.0018629683295383979,
"grad_norm": 0.2336910218000412,
"learning_rate": 9e-05,
"loss": 0.2331,
"step": 9
},
{
"epoch": 0.0020699648105982197,
"grad_norm": 0.2214404195547104,
"learning_rate": 0.0001,
"loss": 0.2747,
"step": 10
},
{
"epoch": 0.002276961291658042,
"grad_norm": 0.1755189299583435,
"learning_rate": 0.00011000000000000002,
"loss": 0.215,
"step": 11
},
{
"epoch": 0.002483957772717864,
"grad_norm": 0.13464422523975372,
"learning_rate": 0.00012,
"loss": 0.1854,
"step": 12
},
{
"epoch": 0.0026909542537776857,
"grad_norm": 0.1568724364042282,
"learning_rate": 0.00013000000000000002,
"loss": 0.1954,
"step": 13
},
{
"epoch": 0.002897950734837508,
"grad_norm": 0.1855469048023224,
"learning_rate": 0.00014,
"loss": 0.1788,
"step": 14
},
{
"epoch": 0.00310494721589733,
"grad_norm": 0.16363286972045898,
"learning_rate": 0.00015000000000000001,
"loss": 0.1816,
"step": 15
},
{
"epoch": 0.0033119436969571516,
"grad_norm": 0.14675703644752502,
"learning_rate": 0.00016,
"loss": 0.144,
"step": 16
},
{
"epoch": 0.003518940178016974,
"grad_norm": 0.13585928082466125,
"learning_rate": 0.00017,
"loss": 0.1265,
"step": 17
},
{
"epoch": 0.0037259366590767957,
"grad_norm": 0.13742923736572266,
"learning_rate": 0.00018,
"loss": 0.1373,
"step": 18
},
{
"epoch": 0.003932933140136618,
"grad_norm": 0.09649409353733063,
"learning_rate": 0.00019,
"loss": 0.1161,
"step": 19
},
{
"epoch": 0.004139929621196439,
"grad_norm": 0.08537352085113525,
"learning_rate": 0.0002,
"loss": 0.1086,
"step": 20
},
{
"epoch": 0.004346926102256262,
"grad_norm": 0.07582477480173111,
"learning_rate": 0.00019995842860112245,
"loss": 0.1047,
"step": 21
},
{
"epoch": 0.004553922583316084,
"grad_norm": 0.07321921736001968,
"learning_rate": 0.00019991685720224486,
"loss": 0.1095,
"step": 22
},
{
"epoch": 0.004760919064375905,
"grad_norm": 0.07746334373950958,
"learning_rate": 0.0001998752858033673,
"loss": 0.08,
"step": 23
},
{
"epoch": 0.004967915545435728,
"grad_norm": 0.06693358719348907,
"learning_rate": 0.00019983371440448973,
"loss": 0.0932,
"step": 24
},
{
"epoch": 0.00517491202649555,
"grad_norm": 0.09097249805927277,
"learning_rate": 0.00019979214300561214,
"loss": 0.095,
"step": 25
},
{
"epoch": 0.005381908507555371,
"grad_norm": 0.06965727359056473,
"learning_rate": 0.00019975057160673458,
"loss": 0.0847,
"step": 26
},
{
"epoch": 0.005588904988615194,
"grad_norm": 0.060345325618982315,
"learning_rate": 0.00019970900020785701,
"loss": 0.0609,
"step": 27
},
{
"epoch": 0.005795901469675016,
"grad_norm": 0.06169256567955017,
"learning_rate": 0.00019966742880897945,
"loss": 0.0651,
"step": 28
},
{
"epoch": 0.006002897950734837,
"grad_norm": 0.07536791265010834,
"learning_rate": 0.00019962585741010186,
"loss": 0.0614,
"step": 29
},
{
"epoch": 0.00620989443179466,
"grad_norm": 0.06803280860185623,
"learning_rate": 0.0001995842860112243,
"loss": 0.0589,
"step": 30
},
{
"epoch": 0.006416890912854482,
"grad_norm": 0.09668745845556259,
"learning_rate": 0.0001995427146123467,
"loss": 0.0511,
"step": 31
},
{
"epoch": 0.006623887393914303,
"grad_norm": 0.060853298753499985,
"learning_rate": 0.00019950114321346915,
"loss": 0.0428,
"step": 32
},
{
"epoch": 0.0068308838749741255,
"grad_norm": 0.04822453856468201,
"learning_rate": 0.00019945957181459156,
"loss": 0.0442,
"step": 33
},
{
"epoch": 0.007037880356033948,
"grad_norm": 0.07689573615789413,
"learning_rate": 0.000199418000415714,
"loss": 0.0448,
"step": 34
},
{
"epoch": 0.007244876837093769,
"grad_norm": 0.07483747601509094,
"learning_rate": 0.00019937642901683643,
"loss": 0.0449,
"step": 35
},
{
"epoch": 0.0074518733181535915,
"grad_norm": 0.06677654385566711,
"learning_rate": 0.00019933485761795884,
"loss": 0.0368,
"step": 36
},
{
"epoch": 0.007658869799213414,
"grad_norm": 0.06132747232913971,
"learning_rate": 0.00019929328621908128,
"loss": 0.0365,
"step": 37
},
{
"epoch": 0.007865866280273236,
"grad_norm": 0.04215759411454201,
"learning_rate": 0.00019925171482020371,
"loss": 0.0307,
"step": 38
},
{
"epoch": 0.008072862761333057,
"grad_norm": 0.05545351654291153,
"learning_rate": 0.00019921014342132612,
"loss": 0.0308,
"step": 39
},
{
"epoch": 0.008279859242392879,
"grad_norm": 0.06014477089047432,
"learning_rate": 0.00019916857202244856,
"loss": 0.0335,
"step": 40
},
{
"epoch": 0.008486855723452702,
"grad_norm": 0.06840485334396362,
"learning_rate": 0.000199127000623571,
"loss": 0.0287,
"step": 41
},
{
"epoch": 0.008693852204512523,
"grad_norm": 0.0705009400844574,
"learning_rate": 0.0001990854292246934,
"loss": 0.0271,
"step": 42
},
{
"epoch": 0.008900848685572345,
"grad_norm": 0.05818328261375427,
"learning_rate": 0.00019904385782581585,
"loss": 0.0253,
"step": 43
},
{
"epoch": 0.009107845166632168,
"grad_norm": 0.03936947509646416,
"learning_rate": 0.00019900228642693828,
"loss": 0.0216,
"step": 44
},
{
"epoch": 0.00931484164769199,
"grad_norm": 0.044559165835380554,
"learning_rate": 0.00019896071502806072,
"loss": 0.0212,
"step": 45
},
{
"epoch": 0.00952183812875181,
"grad_norm": 0.04012398421764374,
"learning_rate": 0.00019891914362918313,
"loss": 0.014,
"step": 46
},
{
"epoch": 0.009728834609811634,
"grad_norm": 0.048441193997859955,
"learning_rate": 0.00019887757223030557,
"loss": 0.0144,
"step": 47
},
{
"epoch": 0.009935831090871455,
"grad_norm": 0.046538762748241425,
"learning_rate": 0.000198836000831428,
"loss": 0.0111,
"step": 48
},
{
"epoch": 0.010142827571931277,
"grad_norm": 0.04020772501826286,
"learning_rate": 0.0001987944294325504,
"loss": 0.0103,
"step": 49
},
{
"epoch": 0.0103498240529911,
"grad_norm": 0.030860010534524918,
"learning_rate": 0.00019875285803367285,
"loss": 0.0071,
"step": 50
},
{
"epoch": 0.010556820534050921,
"grad_norm": 0.0534852109849453,
"learning_rate": 0.0001987112866347953,
"loss": 0.0082,
"step": 51
},
{
"epoch": 0.010763817015110743,
"grad_norm": 0.061904191970825195,
"learning_rate": 0.0001986697152359177,
"loss": 0.0092,
"step": 52
},
{
"epoch": 0.010970813496170566,
"grad_norm": 0.06135628744959831,
"learning_rate": 0.0001986281438370401,
"loss": 0.0091,
"step": 53
},
{
"epoch": 0.011177809977230387,
"grad_norm": 0.060930103063583374,
"learning_rate": 0.00019858657243816254,
"loss": 0.0103,
"step": 54
},
{
"epoch": 0.011384806458290209,
"grad_norm": 0.038847871124744415,
"learning_rate": 0.00019854500103928498,
"loss": 0.0077,
"step": 55
},
{
"epoch": 0.011591802939350032,
"grad_norm": 0.01643364317715168,
"learning_rate": 0.0001985034296404074,
"loss": 0.0065,
"step": 56
},
{
"epoch": 0.011798799420409853,
"grad_norm": 0.03226076811552048,
"learning_rate": 0.00019846185824152983,
"loss": 0.0059,
"step": 57
},
{
"epoch": 0.012005795901469675,
"grad_norm": 0.045181699097156525,
"learning_rate": 0.00019842028684265227,
"loss": 0.0075,
"step": 58
},
{
"epoch": 0.012212792382529498,
"grad_norm": 0.0442410409450531,
"learning_rate": 0.00019837871544377468,
"loss": 0.007,
"step": 59
},
{
"epoch": 0.01241978886358932,
"grad_norm": 0.04646408557891846,
"learning_rate": 0.0001983371440448971,
"loss": 0.009,
"step": 60
},
{
"epoch": 0.01262678534464914,
"grad_norm": 0.037814535200595856,
"learning_rate": 0.00019829557264601955,
"loss": 0.0062,
"step": 61
},
{
"epoch": 0.012833781825708964,
"grad_norm": 0.029840698465704918,
"learning_rate": 0.000198254001247142,
"loss": 0.0075,
"step": 62
},
{
"epoch": 0.013040778306768785,
"grad_norm": 0.009363808669149876,
"learning_rate": 0.0001982124298482644,
"loss": 0.0054,
"step": 63
},
{
"epoch": 0.013247774787828607,
"grad_norm": 0.02504296600818634,
"learning_rate": 0.00019817085844938683,
"loss": 0.0051,
"step": 64
},
{
"epoch": 0.01345477126888843,
"grad_norm": 0.03861517831683159,
"learning_rate": 0.00019812928705050927,
"loss": 0.0061,
"step": 65
},
{
"epoch": 0.013661767749948251,
"grad_norm": 0.04222877696156502,
"learning_rate": 0.00019808771565163168,
"loss": 0.0074,
"step": 66
},
{
"epoch": 0.013868764231008072,
"grad_norm": 0.03613612800836563,
"learning_rate": 0.00019804614425275412,
"loss": 0.0063,
"step": 67
},
{
"epoch": 0.014075760712067896,
"grad_norm": 0.02817763015627861,
"learning_rate": 0.00019800457285387655,
"loss": 0.0074,
"step": 68
},
{
"epoch": 0.014282757193127717,
"grad_norm": 0.01215298566967249,
"learning_rate": 0.00019796300145499896,
"loss": 0.0057,
"step": 69
},
{
"epoch": 0.014489753674187538,
"grad_norm": 0.01781376637518406,
"learning_rate": 0.0001979214300561214,
"loss": 0.0067,
"step": 70
},
{
"epoch": 0.014696750155247362,
"grad_norm": 0.02770097553730011,
"learning_rate": 0.00019787985865724384,
"loss": 0.0057,
"step": 71
},
{
"epoch": 0.014903746636307183,
"grad_norm": 0.029749557375907898,
"learning_rate": 0.00019783828725836625,
"loss": 0.0063,
"step": 72
},
{
"epoch": 0.015110743117367004,
"grad_norm": 0.02843872457742691,
"learning_rate": 0.00019779671585948869,
"loss": 0.0076,
"step": 73
},
{
"epoch": 0.015317739598426828,
"grad_norm": 0.019167358055710793,
"learning_rate": 0.00019775514446061112,
"loss": 0.0059,
"step": 74
},
{
"epoch": 0.015524736079486649,
"grad_norm": 0.01691405475139618,
"learning_rate": 0.00019771357306173353,
"loss": 0.0072,
"step": 75
},
{
"epoch": 0.015731732560546472,
"grad_norm": 0.00585334375500679,
"learning_rate": 0.00019767200166285594,
"loss": 0.0047,
"step": 76
},
{
"epoch": 0.015938729041606294,
"grad_norm": 0.019127612933516502,
"learning_rate": 0.00019763043026397838,
"loss": 0.0047,
"step": 77
},
{
"epoch": 0.016145725522666115,
"grad_norm": 0.026558954268693924,
"learning_rate": 0.00019758885886510082,
"loss": 0.0064,
"step": 78
},
{
"epoch": 0.016352722003725936,
"grad_norm": 0.0275382362306118,
"learning_rate": 0.00019754728746622325,
"loss": 0.006,
"step": 79
},
{
"epoch": 0.016559718484785758,
"grad_norm": 0.016686394810676575,
"learning_rate": 0.00019750571606734566,
"loss": 0.0069,
"step": 80
},
{
"epoch": 0.01676671496584558,
"grad_norm": 0.011664893478155136,
"learning_rate": 0.0001974641446684681,
"loss": 0.0055,
"step": 81
},
{
"epoch": 0.016973711446905404,
"grad_norm": 0.010350242257118225,
"learning_rate": 0.00019742257326959054,
"loss": 0.0045,
"step": 82
},
{
"epoch": 0.017180707927965225,
"grad_norm": 0.018541481345891953,
"learning_rate": 0.00019738100187071295,
"loss": 0.0056,
"step": 83
},
{
"epoch": 0.017387704409025047,
"grad_norm": 0.015899088233709335,
"learning_rate": 0.00019733943047183539,
"loss": 0.0047,
"step": 84
},
{
"epoch": 0.017594700890084868,
"grad_norm": 0.01706838794052601,
"learning_rate": 0.00019729785907295782,
"loss": 0.0059,
"step": 85
},
{
"epoch": 0.01780169737114469,
"grad_norm": 0.01610150933265686,
"learning_rate": 0.00019725628767408023,
"loss": 0.0065,
"step": 86
},
{
"epoch": 0.01800869385220451,
"grad_norm": 0.011388594284653664,
"learning_rate": 0.00019721471627520267,
"loss": 0.0054,
"step": 87
},
{
"epoch": 0.018215690333264336,
"grad_norm": 0.010438695549964905,
"learning_rate": 0.0001971731448763251,
"loss": 0.005,
"step": 88
},
{
"epoch": 0.018422686814324157,
"grad_norm": 0.01277916319668293,
"learning_rate": 0.00019713157347744752,
"loss": 0.0045,
"step": 89
},
{
"epoch": 0.01862968329538398,
"grad_norm": 0.016964582726359367,
"learning_rate": 0.00019709000207856995,
"loss": 0.0064,
"step": 90
},
{
"epoch": 0.0188366797764438,
"grad_norm": 0.015511998906731606,
"learning_rate": 0.0001970484306796924,
"loss": 0.0046,
"step": 91
},
{
"epoch": 0.01904367625750362,
"grad_norm": 0.014610686339437962,
"learning_rate": 0.00019700685928081483,
"loss": 0.0068,
"step": 92
},
{
"epoch": 0.019250672738563443,
"grad_norm": 0.011182552203536034,
"learning_rate": 0.00019696528788193724,
"loss": 0.005,
"step": 93
},
{
"epoch": 0.019457669219623268,
"grad_norm": 0.00904427282512188,
"learning_rate": 0.00019692371648305967,
"loss": 0.0052,
"step": 94
},
{
"epoch": 0.01966466570068309,
"grad_norm": 0.013450189493596554,
"learning_rate": 0.0001968821450841821,
"loss": 0.0053,
"step": 95
},
{
"epoch": 0.01987166218174291,
"grad_norm": 0.011684760451316833,
"learning_rate": 0.00019684057368530452,
"loss": 0.0051,
"step": 96
},
{
"epoch": 0.020078658662802732,
"grad_norm": 0.009463542141020298,
"learning_rate": 0.00019679900228642693,
"loss": 0.0055,
"step": 97
},
{
"epoch": 0.020285655143862553,
"grad_norm": 0.007411513477563858,
"learning_rate": 0.00019675743088754937,
"loss": 0.005,
"step": 98
},
{
"epoch": 0.020492651624922375,
"grad_norm": 0.013031812384724617,
"learning_rate": 0.0001967158594886718,
"loss": 0.0063,
"step": 99
},
{
"epoch": 0.0206996481059822,
"grad_norm": 0.006821679417043924,
"learning_rate": 0.00019667428808979422,
"loss": 0.0058,
"step": 100
},
{
"epoch": 0.02090664458704202,
"grad_norm": 0.009770995937287807,
"learning_rate": 0.00019663271669091665,
"loss": 0.0045,
"step": 101
},
{
"epoch": 0.021113641068101843,
"grad_norm": 0.007675915956497192,
"learning_rate": 0.0001965911452920391,
"loss": 0.0049,
"step": 102
},
{
"epoch": 0.021320637549161664,
"grad_norm": 0.009076464921236038,
"learning_rate": 0.0001965495738931615,
"loss": 0.0048,
"step": 103
},
{
"epoch": 0.021527634030221485,
"grad_norm": 0.010678350925445557,
"learning_rate": 0.00019650800249428394,
"loss": 0.0056,
"step": 104
},
{
"epoch": 0.021734630511281307,
"grad_norm": 0.029563505202531815,
"learning_rate": 0.00019646643109540637,
"loss": 0.0049,
"step": 105
},
{
"epoch": 0.02194162699234113,
"grad_norm": 0.0072585404850542545,
"learning_rate": 0.00019642485969652878,
"loss": 0.005,
"step": 106
},
{
"epoch": 0.022148623473400953,
"grad_norm": 0.009124008938670158,
"learning_rate": 0.00019638328829765122,
"loss": 0.0045,
"step": 107
},
{
"epoch": 0.022355619954460774,
"grad_norm": 0.005743277724832296,
"learning_rate": 0.00019634171689877366,
"loss": 0.0051,
"step": 108
},
{
"epoch": 0.022562616435520596,
"grad_norm": 0.01300257071852684,
"learning_rate": 0.0001963001454998961,
"loss": 0.0049,
"step": 109
},
{
"epoch": 0.022769612916580417,
"grad_norm": 0.02877631224691868,
"learning_rate": 0.0001962585741010185,
"loss": 0.0053,
"step": 110
},
{
"epoch": 0.02297660939764024,
"grad_norm": 0.010237788781523705,
"learning_rate": 0.00019621700270214094,
"loss": 0.0046,
"step": 111
},
{
"epoch": 0.023183605878700064,
"grad_norm": 0.010189997963607311,
"learning_rate": 0.00019617543130326338,
"loss": 0.0056,
"step": 112
},
{
"epoch": 0.023390602359759885,
"grad_norm": 0.010226712562143803,
"learning_rate": 0.0001961338599043858,
"loss": 0.005,
"step": 113
},
{
"epoch": 0.023597598840819706,
"grad_norm": 0.011194237507879734,
"learning_rate": 0.00019609228850550823,
"loss": 0.0054,
"step": 114
},
{
"epoch": 0.023804595321879528,
"grad_norm": 0.0065891253761947155,
"learning_rate": 0.00019605071710663066,
"loss": 0.0057,
"step": 115
},
{
"epoch": 0.02401159180293935,
"grad_norm": 0.008131214417517185,
"learning_rate": 0.00019600914570775307,
"loss": 0.0056,
"step": 116
},
{
"epoch": 0.02421858828399917,
"grad_norm": 0.013916265219449997,
"learning_rate": 0.0001959675743088755,
"loss": 0.0051,
"step": 117
},
{
"epoch": 0.024425584765058996,
"grad_norm": 0.00658258656039834,
"learning_rate": 0.00019592600290999795,
"loss": 0.0048,
"step": 118
},
{
"epoch": 0.024632581246118817,
"grad_norm": 0.005407229065895081,
"learning_rate": 0.00019588443151112036,
"loss": 0.0053,
"step": 119
},
{
"epoch": 0.02483957772717864,
"grad_norm": 0.009581067599356174,
"learning_rate": 0.00019584286011224277,
"loss": 0.0048,
"step": 120
},
{
"epoch": 0.02504657420823846,
"grad_norm": 0.011583163402974606,
"learning_rate": 0.0001958012887133652,
"loss": 0.0049,
"step": 121
},
{
"epoch": 0.02525357068929828,
"grad_norm": 0.017572317272424698,
"learning_rate": 0.00019575971731448764,
"loss": 0.0049,
"step": 122
},
{
"epoch": 0.025460567170358103,
"grad_norm": 0.018644949421286583,
"learning_rate": 0.00019571814591561005,
"loss": 0.0068,
"step": 123
},
{
"epoch": 0.025667563651417927,
"grad_norm": 0.014408939518034458,
"learning_rate": 0.0001956765745167325,
"loss": 0.0056,
"step": 124
},
{
"epoch": 0.02587456013247775,
"grad_norm": 0.0033774918410927057,
"learning_rate": 0.00019563500311785493,
"loss": 0.0053,
"step": 125
},
{
"epoch": 0.02608155661353757,
"grad_norm": 0.005626993719488382,
"learning_rate": 0.00019559343171897736,
"loss": 0.0046,
"step": 126
},
{
"epoch": 0.02628855309459739,
"grad_norm": 0.018701711669564247,
"learning_rate": 0.00019555186032009977,
"loss": 0.0077,
"step": 127
},
{
"epoch": 0.026495549575657213,
"grad_norm": 0.016636714339256287,
"learning_rate": 0.0001955102889212222,
"loss": 0.005,
"step": 128
},
{
"epoch": 0.026702546056717034,
"grad_norm": 0.013526069931685925,
"learning_rate": 0.00019546871752234465,
"loss": 0.0047,
"step": 129
},
{
"epoch": 0.02690954253777686,
"grad_norm": 0.024728018790483475,
"learning_rate": 0.00019542714612346706,
"loss": 0.0045,
"step": 130
},
{
"epoch": 0.02711653901883668,
"grad_norm": 0.02217916212975979,
"learning_rate": 0.0001953855747245895,
"loss": 0.0045,
"step": 131
},
{
"epoch": 0.027323535499896502,
"grad_norm": 0.010518092662096024,
"learning_rate": 0.00019534400332571193,
"loss": 0.0065,
"step": 132
},
{
"epoch": 0.027530531980956324,
"grad_norm": 0.008342115208506584,
"learning_rate": 0.00019530243192683434,
"loss": 0.0063,
"step": 133
},
{
"epoch": 0.027737528462016145,
"grad_norm": 0.008312125690281391,
"learning_rate": 0.00019526086052795678,
"loss": 0.0047,
"step": 134
},
{
"epoch": 0.027944524943075966,
"grad_norm": 0.006928388494998217,
"learning_rate": 0.00019521928912907921,
"loss": 0.005,
"step": 135
},
{
"epoch": 0.02815152142413579,
"grad_norm": 0.0073064775206148624,
"learning_rate": 0.00019517771773020162,
"loss": 0.0061,
"step": 136
},
{
"epoch": 0.028358517905195613,
"grad_norm": 0.007849021814763546,
"learning_rate": 0.00019513614633132406,
"loss": 0.0052,
"step": 137
},
{
"epoch": 0.028565514386255434,
"grad_norm": 0.006330843083560467,
"learning_rate": 0.0001950945749324465,
"loss": 0.006,
"step": 138
},
{
"epoch": 0.028772510867315255,
"grad_norm": 0.002727820537984371,
"learning_rate": 0.00019505300353356894,
"loss": 0.0047,
"step": 139
},
{
"epoch": 0.028979507348375077,
"grad_norm": 0.006755925714969635,
"learning_rate": 0.00019501143213469135,
"loss": 0.0043,
"step": 140
},
{
"epoch": 0.0291865038294349,
"grad_norm": 0.007393544539809227,
"learning_rate": 0.00019496986073581376,
"loss": 0.0047,
"step": 141
},
{
"epoch": 0.029393500310494723,
"grad_norm": 0.007699685171246529,
"learning_rate": 0.0001949282893369362,
"loss": 0.0053,
"step": 142
},
{
"epoch": 0.029600496791554545,
"grad_norm": 0.003382055787369609,
"learning_rate": 0.0001948867179380586,
"loss": 0.0042,
"step": 143
},
{
"epoch": 0.029807493272614366,
"grad_norm": 0.01596757024526596,
"learning_rate": 0.00019484514653918104,
"loss": 0.0078,
"step": 144
},
{
"epoch": 0.030014489753674187,
"grad_norm": 0.00668082432821393,
"learning_rate": 0.00019480357514030348,
"loss": 0.0052,
"step": 145
},
{
"epoch": 0.03022148623473401,
"grad_norm": 0.007241010665893555,
"learning_rate": 0.00019476200374142591,
"loss": 0.0054,
"step": 146
},
{
"epoch": 0.03042848271579383,
"grad_norm": 0.009339329786598682,
"learning_rate": 0.00019472043234254832,
"loss": 0.0066,
"step": 147
},
{
"epoch": 0.030635479196853655,
"grad_norm": 0.007252382580190897,
"learning_rate": 0.00019467886094367076,
"loss": 0.0046,
"step": 148
},
{
"epoch": 0.030842475677913476,
"grad_norm": 0.004665658809244633,
"learning_rate": 0.0001946372895447932,
"loss": 0.0044,
"step": 149
},
{
"epoch": 0.031049472158973298,
"grad_norm": 0.006148173939436674,
"learning_rate": 0.0001945957181459156,
"loss": 0.0049,
"step": 150
},
{
"epoch": 0.03125646864003312,
"grad_norm": 0.004451967775821686,
"learning_rate": 0.00019455414674703805,
"loss": 0.0061,
"step": 151
},
{
"epoch": 0.031463465121092944,
"grad_norm": 0.004053746350109577,
"learning_rate": 0.00019451257534816048,
"loss": 0.0047,
"step": 152
},
{
"epoch": 0.03167046160215276,
"grad_norm": 0.009219355881214142,
"learning_rate": 0.0001944710039492829,
"loss": 0.0048,
"step": 153
},
{
"epoch": 0.03187745808321259,
"grad_norm": 0.020168175920844078,
"learning_rate": 0.00019442943255040533,
"loss": 0.0063,
"step": 154
},
{
"epoch": 0.032084454564272405,
"grad_norm": 0.0030542612075805664,
"learning_rate": 0.00019438786115152777,
"loss": 0.0048,
"step": 155
},
{
"epoch": 0.03229145104533223,
"grad_norm": 0.00561768002808094,
"learning_rate": 0.0001943462897526502,
"loss": 0.0047,
"step": 156
},
{
"epoch": 0.032498447526392055,
"grad_norm": 0.004265373572707176,
"learning_rate": 0.0001943047183537726,
"loss": 0.0062,
"step": 157
},
{
"epoch": 0.03270544400745187,
"grad_norm": 0.005523406434804201,
"learning_rate": 0.00019426314695489505,
"loss": 0.0044,
"step": 158
},
{
"epoch": 0.0329124404885117,
"grad_norm": 0.009324849583208561,
"learning_rate": 0.0001942215755560175,
"loss": 0.0055,
"step": 159
},
{
"epoch": 0.033119436969571515,
"grad_norm": 0.012468270026147366,
"learning_rate": 0.0001941800041571399,
"loss": 0.0045,
"step": 160
},
{
"epoch": 0.03332643345063134,
"grad_norm": 0.0018690524157136679,
"learning_rate": 0.00019413843275826233,
"loss": 0.0047,
"step": 161
},
{
"epoch": 0.03353342993169116,
"grad_norm": 0.009851769544184208,
"learning_rate": 0.00019409686135938477,
"loss": 0.0052,
"step": 162
},
{
"epoch": 0.03374042641275098,
"grad_norm": 0.0014802832156419754,
"learning_rate": 0.00019405528996050718,
"loss": 0.0045,
"step": 163
},
{
"epoch": 0.03394742289381081,
"grad_norm": 0.00177583540789783,
"learning_rate": 0.0001940137185616296,
"loss": 0.0046,
"step": 164
},
{
"epoch": 0.034154419374870626,
"grad_norm": 0.0017039773520082235,
"learning_rate": 0.00019397214716275203,
"loss": 0.0044,
"step": 165
},
{
"epoch": 0.03436141585593045,
"grad_norm": 0.010193880647420883,
"learning_rate": 0.00019393057576387447,
"loss": 0.0054,
"step": 166
},
{
"epoch": 0.03456841233699027,
"grad_norm": 0.011056206189095974,
"learning_rate": 0.00019388900436499688,
"loss": 0.0051,
"step": 167
},
{
"epoch": 0.034775408818050094,
"grad_norm": 0.004655253142118454,
"learning_rate": 0.0001938474329661193,
"loss": 0.0046,
"step": 168
},
{
"epoch": 0.03498240529910992,
"grad_norm": 0.0035964485723525286,
"learning_rate": 0.00019380586156724175,
"loss": 0.0045,
"step": 169
},
{
"epoch": 0.035189401780169736,
"grad_norm": 0.0026267431676387787,
"learning_rate": 0.00019376429016836416,
"loss": 0.0047,
"step": 170
},
{
"epoch": 0.03539639826122956,
"grad_norm": 0.008452721871435642,
"learning_rate": 0.0001937227187694866,
"loss": 0.0059,
"step": 171
},
{
"epoch": 0.03560339474228938,
"grad_norm": 0.006845233030617237,
"learning_rate": 0.00019368114737060903,
"loss": 0.0051,
"step": 172
},
{
"epoch": 0.035810391223349204,
"grad_norm": 0.005468891002237797,
"learning_rate": 0.00019363957597173144,
"loss": 0.0059,
"step": 173
},
{
"epoch": 0.03601738770440902,
"grad_norm": 0.007444227579981089,
"learning_rate": 0.00019359800457285388,
"loss": 0.0044,
"step": 174
},
{
"epoch": 0.03622438418546885,
"grad_norm": 0.011544904671609402,
"learning_rate": 0.00019355643317397632,
"loss": 0.0053,
"step": 175
},
{
"epoch": 0.03643138066652867,
"grad_norm": 0.04777868092060089,
"learning_rate": 0.00019351486177509875,
"loss": 0.005,
"step": 176
},
{
"epoch": 0.03663837714758849,
"grad_norm": 0.004177747759968042,
"learning_rate": 0.00019347329037622116,
"loss": 0.0048,
"step": 177
},
{
"epoch": 0.036845373628648315,
"grad_norm": 0.0020709133241325617,
"learning_rate": 0.0001934317189773436,
"loss": 0.0045,
"step": 178
},
{
"epoch": 0.03705237010970813,
"grad_norm": 0.012106567621231079,
"learning_rate": 0.00019339014757846604,
"loss": 0.0043,
"step": 179
},
{
"epoch": 0.03725936659076796,
"grad_norm": 0.005318734794855118,
"learning_rate": 0.00019334857617958845,
"loss": 0.005,
"step": 180
},
{
"epoch": 0.03746636307182778,
"grad_norm": 0.04815113916993141,
"learning_rate": 0.00019330700478071089,
"loss": 0.0056,
"step": 181
},
{
"epoch": 0.0376733595528876,
"grad_norm": 0.029821882024407387,
"learning_rate": 0.00019326543338183332,
"loss": 0.0053,
"step": 182
},
{
"epoch": 0.037880356033947425,
"grad_norm": 0.010615061037242413,
"learning_rate": 0.00019322386198295573,
"loss": 0.0049,
"step": 183
},
{
"epoch": 0.03808735251500724,
"grad_norm": 0.00788772851228714,
"learning_rate": 0.00019318229058407817,
"loss": 0.0043,
"step": 184
},
{
"epoch": 0.03829434899606707,
"grad_norm": 0.0045401486568152905,
"learning_rate": 0.00019314071918520058,
"loss": 0.0056,
"step": 185
},
{
"epoch": 0.038501345477126886,
"grad_norm": 0.004325198009610176,
"learning_rate": 0.00019309914778632302,
"loss": 0.0059,
"step": 186
},
{
"epoch": 0.03870834195818671,
"grad_norm": 0.012164851650595665,
"learning_rate": 0.00019305757638744543,
"loss": 0.0046,
"step": 187
},
{
"epoch": 0.038915338439246536,
"grad_norm": 0.02496037259697914,
"learning_rate": 0.00019301600498856786,
"loss": 0.0068,
"step": 188
},
{
"epoch": 0.039122334920306354,
"grad_norm": 0.011729522608220577,
"learning_rate": 0.0001929744335896903,
"loss": 0.0047,
"step": 189
},
{
"epoch": 0.03932933140136618,
"grad_norm": 0.003992550540715456,
"learning_rate": 0.0001929328621908127,
"loss": 0.0047,
"step": 190
},
{
"epoch": 0.039536327882425996,
"grad_norm": 0.012220533564686775,
"learning_rate": 0.00019289129079193515,
"loss": 0.0059,
"step": 191
},
{
"epoch": 0.03974332436348582,
"grad_norm": 0.0037318835966289043,
"learning_rate": 0.00019284971939305759,
"loss": 0.0049,
"step": 192
},
{
"epoch": 0.039950320844545646,
"grad_norm": 0.008259239606559277,
"learning_rate": 0.00019280814799418002,
"loss": 0.0047,
"step": 193
},
{
"epoch": 0.040157317325605464,
"grad_norm": 0.0029760177712887526,
"learning_rate": 0.00019276657659530243,
"loss": 0.0046,
"step": 194
},
{
"epoch": 0.04036431380666529,
"grad_norm": 0.007835526019334793,
"learning_rate": 0.00019272500519642487,
"loss": 0.0045,
"step": 195
},
{
"epoch": 0.04057131028772511,
"grad_norm": 0.019788436591625214,
"learning_rate": 0.0001926834337975473,
"loss": 0.0056,
"step": 196
},
{
"epoch": 0.04077830676878493,
"grad_norm": 0.0054263221099972725,
"learning_rate": 0.00019264186239866972,
"loss": 0.0047,
"step": 197
},
{
"epoch": 0.04098530324984475,
"grad_norm": 0.006738686002790928,
"learning_rate": 0.00019260029099979215,
"loss": 0.0053,
"step": 198
},
{
"epoch": 0.041192299730904575,
"grad_norm": 0.0019211384933441877,
"learning_rate": 0.0001925587196009146,
"loss": 0.0046,
"step": 199
},
{
"epoch": 0.0413992962119644,
"grad_norm": 0.0064520747400820255,
"learning_rate": 0.000192517148202037,
"loss": 0.0049,
"step": 200
},
{
"epoch": 0.04160629269302422,
"grad_norm": 0.0034363584127277136,
"learning_rate": 0.00019247557680315944,
"loss": 0.0048,
"step": 201
},
{
"epoch": 0.04181328917408404,
"grad_norm": 0.021868525072932243,
"learning_rate": 0.00019243400540428187,
"loss": 0.0055,
"step": 202
},
{
"epoch": 0.04202028565514386,
"grad_norm": 0.011735360138118267,
"learning_rate": 0.0001923924340054043,
"loss": 0.0059,
"step": 203
},
{
"epoch": 0.042227282136203685,
"grad_norm": 0.008518829010426998,
"learning_rate": 0.00019235086260652672,
"loss": 0.0062,
"step": 204
},
{
"epoch": 0.0424342786172635,
"grad_norm": 0.007598051335662603,
"learning_rate": 0.00019230929120764916,
"loss": 0.005,
"step": 205
},
{
"epoch": 0.04264127509832333,
"grad_norm": 0.004902805667370558,
"learning_rate": 0.0001922677198087716,
"loss": 0.0047,
"step": 206
},
{
"epoch": 0.04284827157938315,
"grad_norm": 0.01092604547739029,
"learning_rate": 0.00019222614840989398,
"loss": 0.0062,
"step": 207
},
{
"epoch": 0.04305526806044297,
"grad_norm": 0.017829621210694313,
"learning_rate": 0.00019218457701101642,
"loss": 0.0069,
"step": 208
},
{
"epoch": 0.043262264541502796,
"grad_norm": 0.0056928307749331,
"learning_rate": 0.00019214300561213885,
"loss": 0.0052,
"step": 209
},
{
"epoch": 0.043469261022562614,
"grad_norm": 0.0032917701173573732,
"learning_rate": 0.0001921014342132613,
"loss": 0.0049,
"step": 210
},
{
"epoch": 0.04367625750362244,
"grad_norm": 0.00267777475528419,
"learning_rate": 0.0001920598628143837,
"loss": 0.0047,
"step": 211
},
{
"epoch": 0.04388325398468226,
"grad_norm": 0.005063153337687254,
"learning_rate": 0.00019201829141550614,
"loss": 0.0065,
"step": 212
},
{
"epoch": 0.04409025046574208,
"grad_norm": 0.007821328938007355,
"learning_rate": 0.00019197672001662857,
"loss": 0.0073,
"step": 213
},
{
"epoch": 0.044297246946801906,
"grad_norm": 0.0022040351759642363,
"learning_rate": 0.00019193514861775098,
"loss": 0.0042,
"step": 214
},
{
"epoch": 0.044504243427861724,
"grad_norm": 0.010549996979534626,
"learning_rate": 0.00019189357721887342,
"loss": 0.0048,
"step": 215
},
{
"epoch": 0.04471123990892155,
"grad_norm": 0.003572909627109766,
"learning_rate": 0.00019185200581999586,
"loss": 0.0048,
"step": 216
},
{
"epoch": 0.04491823638998137,
"grad_norm": 0.0047572036273777485,
"learning_rate": 0.00019181043442111827,
"loss": 0.0046,
"step": 217
},
{
"epoch": 0.04512523287104119,
"grad_norm": 0.003330536652356386,
"learning_rate": 0.0001917688630222407,
"loss": 0.0043,
"step": 218
},
{
"epoch": 0.04533222935210102,
"grad_norm": 0.003489327384158969,
"learning_rate": 0.00019172729162336314,
"loss": 0.0044,
"step": 219
},
{
"epoch": 0.045539225833160835,
"grad_norm": 0.006631118711084127,
"learning_rate": 0.00019168572022448555,
"loss": 0.0046,
"step": 220
},
{
"epoch": 0.04574622231422066,
"grad_norm": 0.0016155489720404148,
"learning_rate": 0.000191644148825608,
"loss": 0.0045,
"step": 221
},
{
"epoch": 0.04595321879528048,
"grad_norm": 0.00484581058844924,
"learning_rate": 0.00019160257742673043,
"loss": 0.0049,
"step": 222
},
{
"epoch": 0.0461602152763403,
"grad_norm": 0.016010191291570663,
"learning_rate": 0.00019156100602785286,
"loss": 0.0052,
"step": 223
},
{
"epoch": 0.04636721175740013,
"grad_norm": 0.0038696257397532463,
"learning_rate": 0.00019151943462897527,
"loss": 0.0046,
"step": 224
},
{
"epoch": 0.046574208238459945,
"grad_norm": 0.003132582874968648,
"learning_rate": 0.0001914778632300977,
"loss": 0.0055,
"step": 225
},
{
"epoch": 0.04678120471951977,
"grad_norm": 0.003856977680698037,
"learning_rate": 0.00019143629183122015,
"loss": 0.0048,
"step": 226
},
{
"epoch": 0.04698820120057959,
"grad_norm": 0.0084042027592659,
"learning_rate": 0.00019139472043234256,
"loss": 0.005,
"step": 227
},
{
"epoch": 0.04719519768163941,
"grad_norm": 0.004033537581562996,
"learning_rate": 0.000191353149033465,
"loss": 0.0053,
"step": 228
},
{
"epoch": 0.04740219416269923,
"grad_norm": 0.002745938254520297,
"learning_rate": 0.0001913115776345874,
"loss": 0.005,
"step": 229
},
{
"epoch": 0.047609190643759056,
"grad_norm": 0.0033783107064664364,
"learning_rate": 0.00019127000623570984,
"loss": 0.0045,
"step": 230
},
{
"epoch": 0.04781618712481888,
"grad_norm": 0.011202207766473293,
"learning_rate": 0.00019122843483683225,
"loss": 0.0054,
"step": 231
},
{
"epoch": 0.0480231836058787,
"grad_norm": 0.006426738575100899,
"learning_rate": 0.0001911868634379547,
"loss": 0.0049,
"step": 232
},
{
"epoch": 0.04823018008693852,
"grad_norm": 0.0026592197827994823,
"learning_rate": 0.00019114529203907713,
"loss": 0.0045,
"step": 233
},
{
"epoch": 0.04843717656799834,
"grad_norm": 0.0044938791543245316,
"learning_rate": 0.00019110372064019954,
"loss": 0.0048,
"step": 234
},
{
"epoch": 0.048644173049058166,
"grad_norm": 0.004537639208137989,
"learning_rate": 0.00019106214924132197,
"loss": 0.0043,
"step": 235
},
{
"epoch": 0.04885116953011799,
"grad_norm": 0.007899290882050991,
"learning_rate": 0.0001910205778424444,
"loss": 0.0047,
"step": 236
},
{
"epoch": 0.04905816601117781,
"grad_norm": 0.003346335142850876,
"learning_rate": 0.00019097900644356682,
"loss": 0.0046,
"step": 237
},
{
"epoch": 0.049265162492237634,
"grad_norm": 0.012666060589253902,
"learning_rate": 0.00019093743504468926,
"loss": 0.0051,
"step": 238
},
{
"epoch": 0.04947215897329745,
"grad_norm": 0.005454343743622303,
"learning_rate": 0.0001908958636458117,
"loss": 0.0052,
"step": 239
},
{
"epoch": 0.04967915545435728,
"grad_norm": 0.0051568858325481415,
"learning_rate": 0.00019085429224693413,
"loss": 0.0054,
"step": 240
},
{
"epoch": 0.049886151935417095,
"grad_norm": 0.005354621913284063,
"learning_rate": 0.00019081272084805654,
"loss": 0.0045,
"step": 241
},
{
"epoch": 0.05009314841647692,
"grad_norm": 0.00728578818961978,
"learning_rate": 0.00019077114944917898,
"loss": 0.0056,
"step": 242
},
{
"epoch": 0.050300144897536744,
"grad_norm": 0.012164080515503883,
"learning_rate": 0.00019072957805030141,
"loss": 0.0055,
"step": 243
},
{
"epoch": 0.05050714137859656,
"grad_norm": 0.006451157853007317,
"learning_rate": 0.00019068800665142382,
"loss": 0.0045,
"step": 244
},
{
"epoch": 0.05071413785965639,
"grad_norm": 0.04119205102324486,
"learning_rate": 0.00019064643525254626,
"loss": 0.0048,
"step": 245
},
{
"epoch": 0.050921134340716205,
"grad_norm": 0.006683278828859329,
"learning_rate": 0.0001906048638536687,
"loss": 0.0054,
"step": 246
},
{
"epoch": 0.05112813082177603,
"grad_norm": 0.004711155779659748,
"learning_rate": 0.0001905632924547911,
"loss": 0.0052,
"step": 247
},
{
"epoch": 0.051335127302835855,
"grad_norm": 0.022338053211569786,
"learning_rate": 0.00019052172105591355,
"loss": 0.0051,
"step": 248
},
{
"epoch": 0.05154212378389567,
"grad_norm": 0.0033723730593919754,
"learning_rate": 0.00019048014965703598,
"loss": 0.0041,
"step": 249
},
{
"epoch": 0.0517491202649555,
"grad_norm": 0.013185818679630756,
"learning_rate": 0.00019043857825815842,
"loss": 0.0071,
"step": 250
},
{
"epoch": 0.051956116746015316,
"grad_norm": 0.01329563558101654,
"learning_rate": 0.0001903970068592808,
"loss": 0.0047,
"step": 251
},
{
"epoch": 0.05216311322707514,
"grad_norm": 0.0038430816493928432,
"learning_rate": 0.00019035543546040324,
"loss": 0.0049,
"step": 252
},
{
"epoch": 0.05237010970813496,
"grad_norm": 0.006868777330964804,
"learning_rate": 0.00019031386406152568,
"loss": 0.0045,
"step": 253
},
{
"epoch": 0.05257710618919478,
"grad_norm": 0.002608188660815358,
"learning_rate": 0.0001902722926626481,
"loss": 0.005,
"step": 254
},
{
"epoch": 0.05278410267025461,
"grad_norm": 0.005354885943233967,
"learning_rate": 0.00019023072126377052,
"loss": 0.0058,
"step": 255
},
{
"epoch": 0.052991099151314426,
"grad_norm": 0.006729793269187212,
"learning_rate": 0.00019018914986489296,
"loss": 0.0056,
"step": 256
},
{
"epoch": 0.05319809563237425,
"grad_norm": 0.0042056431993842125,
"learning_rate": 0.0001901475784660154,
"loss": 0.0042,
"step": 257
},
{
"epoch": 0.05340509211343407,
"grad_norm": 0.006845582276582718,
"learning_rate": 0.0001901060070671378,
"loss": 0.0048,
"step": 258
},
{
"epoch": 0.053612088594493894,
"grad_norm": 0.004010177683085203,
"learning_rate": 0.00019006443566826025,
"loss": 0.005,
"step": 259
},
{
"epoch": 0.05381908507555372,
"grad_norm": 0.0019074059091508389,
"learning_rate": 0.00019002286426938268,
"loss": 0.005,
"step": 260
},
{
"epoch": 0.05402608155661354,
"grad_norm": 0.002619614126160741,
"learning_rate": 0.0001899812928705051,
"loss": 0.0043,
"step": 261
},
{
"epoch": 0.05423307803767336,
"grad_norm": 0.002715028589591384,
"learning_rate": 0.00018993972147162753,
"loss": 0.0046,
"step": 262
},
{
"epoch": 0.05444007451873318,
"grad_norm": 0.004556506406515837,
"learning_rate": 0.00018989815007274997,
"loss": 0.0046,
"step": 263
},
{
"epoch": 0.054647070999793004,
"grad_norm": 0.010189310647547245,
"learning_rate": 0.00018985657867387238,
"loss": 0.0047,
"step": 264
},
{
"epoch": 0.05485406748085282,
"grad_norm": 0.002120416844263673,
"learning_rate": 0.0001898150072749948,
"loss": 0.0051,
"step": 265
},
{
"epoch": 0.05506106396191265,
"grad_norm": 0.008936331607401371,
"learning_rate": 0.00018977343587611725,
"loss": 0.005,
"step": 266
},
{
"epoch": 0.05526806044297247,
"grad_norm": 0.0026866530533879995,
"learning_rate": 0.00018973186447723966,
"loss": 0.0047,
"step": 267
},
{
"epoch": 0.05547505692403229,
"grad_norm": 0.009859035722911358,
"learning_rate": 0.0001896902930783621,
"loss": 0.0052,
"step": 268
},
{
"epoch": 0.055682053405092115,
"grad_norm": 0.0024370807223021984,
"learning_rate": 0.00018964872167948453,
"loss": 0.0053,
"step": 269
},
{
"epoch": 0.05588904988615193,
"grad_norm": 0.006978074554353952,
"learning_rate": 0.00018960715028060697,
"loss": 0.0062,
"step": 270
},
{
"epoch": 0.05609604636721176,
"grad_norm": 0.0037223445251584053,
"learning_rate": 0.00018956557888172938,
"loss": 0.0047,
"step": 271
},
{
"epoch": 0.05630304284827158,
"grad_norm": 0.018946697935461998,
"learning_rate": 0.00018952400748285182,
"loss": 0.0054,
"step": 272
},
{
"epoch": 0.0565100393293314,
"grad_norm": 0.009405361488461494,
"learning_rate": 0.00018948243608397423,
"loss": 0.0043,
"step": 273
},
{
"epoch": 0.056717035810391225,
"grad_norm": 0.0008082574931904674,
"learning_rate": 0.00018944086468509667,
"loss": 0.0045,
"step": 274
},
{
"epoch": 0.05692403229145104,
"grad_norm": 0.004950170870870352,
"learning_rate": 0.00018939929328621908,
"loss": 0.0044,
"step": 275
},
{
"epoch": 0.05713102877251087,
"grad_norm": 0.010343370027840137,
"learning_rate": 0.0001893577218873415,
"loss": 0.0046,
"step": 276
},
{
"epoch": 0.057338025253570686,
"grad_norm": 0.005096866749227047,
"learning_rate": 0.00018931615048846395,
"loss": 0.005,
"step": 277
},
{
"epoch": 0.05754502173463051,
"grad_norm": 0.0033187547232955694,
"learning_rate": 0.00018927457908958636,
"loss": 0.0041,
"step": 278
},
{
"epoch": 0.057752018215690336,
"grad_norm": 0.0031861995812505484,
"learning_rate": 0.0001892330076907088,
"loss": 0.0059,
"step": 279
},
{
"epoch": 0.057959014696750154,
"grad_norm": 0.0053593809716403484,
"learning_rate": 0.00018919143629183123,
"loss": 0.0047,
"step": 280
},
{
"epoch": 0.05816601117780998,
"grad_norm": 0.005950809922069311,
"learning_rate": 0.00018914986489295364,
"loss": 0.0046,
"step": 281
},
{
"epoch": 0.0583730076588698,
"grad_norm": 0.006816082634031773,
"learning_rate": 0.00018910829349407608,
"loss": 0.0048,
"step": 282
},
{
"epoch": 0.05858000413992962,
"grad_norm": 0.005741049535572529,
"learning_rate": 0.00018906672209519852,
"loss": 0.0048,
"step": 283
},
{
"epoch": 0.058787000620989446,
"grad_norm": 0.004367714747786522,
"learning_rate": 0.00018902515069632093,
"loss": 0.0044,
"step": 284
},
{
"epoch": 0.058993997102049264,
"grad_norm": 0.0051370360888540745,
"learning_rate": 0.00018898357929744336,
"loss": 0.0049,
"step": 285
},
{
"epoch": 0.05920099358310909,
"grad_norm": 0.00557302962988615,
"learning_rate": 0.0001889420078985658,
"loss": 0.005,
"step": 286
},
{
"epoch": 0.05940799006416891,
"grad_norm": 0.004271439276635647,
"learning_rate": 0.00018890043649968824,
"loss": 0.0042,
"step": 287
},
{
"epoch": 0.05961498654522873,
"grad_norm": 0.00942255649715662,
"learning_rate": 0.00018885886510081065,
"loss": 0.005,
"step": 288
},
{
"epoch": 0.05982198302628855,
"grad_norm": 0.001850366359576583,
"learning_rate": 0.00018881729370193309,
"loss": 0.0049,
"step": 289
},
{
"epoch": 0.060028979507348375,
"grad_norm": 0.005905089899897575,
"learning_rate": 0.00018877572230305552,
"loss": 0.0049,
"step": 290
},
{
"epoch": 0.0602359759884082,
"grad_norm": 0.006148039363324642,
"learning_rate": 0.00018873415090417793,
"loss": 0.0058,
"step": 291
},
{
"epoch": 0.06044297246946802,
"grad_norm": 0.004467264749109745,
"learning_rate": 0.00018869257950530037,
"loss": 0.0046,
"step": 292
},
{
"epoch": 0.06064996895052784,
"grad_norm": 0.0026016500778496265,
"learning_rate": 0.0001886510081064228,
"loss": 0.0046,
"step": 293
},
{
"epoch": 0.06085696543158766,
"grad_norm": 0.00536306481808424,
"learning_rate": 0.00018860943670754522,
"loss": 0.0046,
"step": 294
},
{
"epoch": 0.061063961912647485,
"grad_norm": 0.00043015365372411907,
"learning_rate": 0.00018856786530866763,
"loss": 0.0051,
"step": 295
},
{
"epoch": 0.06127095839370731,
"grad_norm": 0.002035475103184581,
"learning_rate": 0.00018852629390979006,
"loss": 0.0044,
"step": 296
},
{
"epoch": 0.06147795487476713,
"grad_norm": 0.004757678601890802,
"learning_rate": 0.0001884847225109125,
"loss": 0.0055,
"step": 297
},
{
"epoch": 0.06168495135582695,
"grad_norm": 0.004153924528509378,
"learning_rate": 0.0001884431511120349,
"loss": 0.0047,
"step": 298
},
{
"epoch": 0.06189194783688677,
"grad_norm": 0.004693943541496992,
"learning_rate": 0.00018840157971315735,
"loss": 0.006,
"step": 299
},
{
"epoch": 0.062098944317946596,
"grad_norm": 0.005683131981641054,
"learning_rate": 0.00018836000831427979,
"loss": 0.0054,
"step": 300
},
{
"epoch": 0.062305940799006414,
"grad_norm": 0.01867171749472618,
"learning_rate": 0.0001883184369154022,
"loss": 0.0047,
"step": 301
},
{
"epoch": 0.06251293728006624,
"grad_norm": 0.010631178505718708,
"learning_rate": 0.00018827686551652463,
"loss": 0.0052,
"step": 302
},
{
"epoch": 0.06271993376112606,
"grad_norm": 0.0025544106028974056,
"learning_rate": 0.00018823529411764707,
"loss": 0.0044,
"step": 303
},
{
"epoch": 0.06292693024218589,
"grad_norm": 0.005773225799202919,
"learning_rate": 0.0001881937227187695,
"loss": 0.0047,
"step": 304
},
{
"epoch": 0.0631339267232457,
"grad_norm": 0.008623667992651463,
"learning_rate": 0.00018815215131989192,
"loss": 0.0041,
"step": 305
},
{
"epoch": 0.06334092320430552,
"grad_norm": 0.0023567613679915667,
"learning_rate": 0.00018811057992101435,
"loss": 0.0044,
"step": 306
},
{
"epoch": 0.06354791968536534,
"grad_norm": 0.005568039603531361,
"learning_rate": 0.0001880690085221368,
"loss": 0.0048,
"step": 307
},
{
"epoch": 0.06375491616642517,
"grad_norm": 0.006429716479033232,
"learning_rate": 0.0001880274371232592,
"loss": 0.0049,
"step": 308
},
{
"epoch": 0.06396191264748499,
"grad_norm": 0.0012258924543857574,
"learning_rate": 0.00018798586572438164,
"loss": 0.0046,
"step": 309
},
{
"epoch": 0.06416890912854481,
"grad_norm": 0.013900945894420147,
"learning_rate": 0.00018794429432550407,
"loss": 0.0054,
"step": 310
},
{
"epoch": 0.06437590560960464,
"grad_norm": 0.0164532121270895,
"learning_rate": 0.00018790272292662648,
"loss": 0.0052,
"step": 311
},
{
"epoch": 0.06458290209066446,
"grad_norm": 0.0039049319457262754,
"learning_rate": 0.00018786115152774892,
"loss": 0.0047,
"step": 312
},
{
"epoch": 0.06478989857172428,
"grad_norm": 0.0025435080751776695,
"learning_rate": 0.00018781958012887136,
"loss": 0.0045,
"step": 313
},
{
"epoch": 0.06499689505278411,
"grad_norm": 0.0066621373407542706,
"learning_rate": 0.00018777800872999377,
"loss": 0.0057,
"step": 314
},
{
"epoch": 0.06520389153384393,
"grad_norm": 0.0011938404059037566,
"learning_rate": 0.0001877364373311162,
"loss": 0.0045,
"step": 315
},
{
"epoch": 0.06541088801490375,
"grad_norm": 0.005898007657378912,
"learning_rate": 0.00018769486593223864,
"loss": 0.0047,
"step": 316
},
{
"epoch": 0.06561788449596356,
"grad_norm": 0.004224811680614948,
"learning_rate": 0.00018765329453336105,
"loss": 0.0059,
"step": 317
},
{
"epoch": 0.0658248809770234,
"grad_norm": 0.02106441557407379,
"learning_rate": 0.00018761172313448346,
"loss": 0.0051,
"step": 318
},
{
"epoch": 0.06603187745808321,
"grad_norm": 0.0010845439974218607,
"learning_rate": 0.0001875701517356059,
"loss": 0.0042,
"step": 319
},
{
"epoch": 0.06623887393914303,
"grad_norm": 0.007267239037901163,
"learning_rate": 0.00018752858033672834,
"loss": 0.0048,
"step": 320
},
{
"epoch": 0.06644587042020286,
"grad_norm": 0.0066713071428239346,
"learning_rate": 0.00018748700893785077,
"loss": 0.0049,
"step": 321
},
{
"epoch": 0.06665286690126268,
"grad_norm": 0.007623916491866112,
"learning_rate": 0.00018744543753897318,
"loss": 0.0052,
"step": 322
},
{
"epoch": 0.0668598633823225,
"grad_norm": 0.003484464716166258,
"learning_rate": 0.00018740386614009562,
"loss": 0.0045,
"step": 323
},
{
"epoch": 0.06706685986338232,
"grad_norm": 0.013743946328759193,
"learning_rate": 0.00018736229474121806,
"loss": 0.0049,
"step": 324
},
{
"epoch": 0.06727385634444215,
"grad_norm": 0.0030819710809737444,
"learning_rate": 0.00018732072334234047,
"loss": 0.0052,
"step": 325
},
{
"epoch": 0.06748085282550197,
"grad_norm": 0.014786194078624249,
"learning_rate": 0.0001872791519434629,
"loss": 0.0058,
"step": 326
},
{
"epoch": 0.06768784930656178,
"grad_norm": 0.007248689886182547,
"learning_rate": 0.00018723758054458534,
"loss": 0.0046,
"step": 327
},
{
"epoch": 0.06789484578762162,
"grad_norm": 0.010181601159274578,
"learning_rate": 0.00018719600914570775,
"loss": 0.0053,
"step": 328
},
{
"epoch": 0.06810184226868143,
"grad_norm": 0.004160667769610882,
"learning_rate": 0.0001871544377468302,
"loss": 0.0056,
"step": 329
},
{
"epoch": 0.06830883874974125,
"grad_norm": 0.0021253142040222883,
"learning_rate": 0.00018711286634795263,
"loss": 0.0045,
"step": 330
},
{
"epoch": 0.06851583523080107,
"grad_norm": 0.0058175004087388515,
"learning_rate": 0.00018707129494907504,
"loss": 0.0058,
"step": 331
},
{
"epoch": 0.0687228317118609,
"grad_norm": 0.005257429089397192,
"learning_rate": 0.00018702972355019747,
"loss": 0.0049,
"step": 332
},
{
"epoch": 0.06892982819292072,
"grad_norm": 0.003563474863767624,
"learning_rate": 0.0001869881521513199,
"loss": 0.0043,
"step": 333
},
{
"epoch": 0.06913682467398054,
"grad_norm": 0.002636804012581706,
"learning_rate": 0.00018694658075244235,
"loss": 0.0046,
"step": 334
},
{
"epoch": 0.06934382115504037,
"grad_norm": 0.009255892597138882,
"learning_rate": 0.00018690500935356476,
"loss": 0.0053,
"step": 335
},
{
"epoch": 0.06955081763610019,
"grad_norm": 0.0023986424785107374,
"learning_rate": 0.0001868634379546872,
"loss": 0.0045,
"step": 336
},
{
"epoch": 0.06975781411716,
"grad_norm": 0.006371774710714817,
"learning_rate": 0.00018682186655580963,
"loss": 0.0046,
"step": 337
},
{
"epoch": 0.06996481059821984,
"grad_norm": 0.009518152102828026,
"learning_rate": 0.00018678029515693204,
"loss": 0.0049,
"step": 338
},
{
"epoch": 0.07017180707927965,
"grad_norm": 0.003742037108168006,
"learning_rate": 0.00018673872375805445,
"loss": 0.0064,
"step": 339
},
{
"epoch": 0.07037880356033947,
"grad_norm": 0.009771923534572124,
"learning_rate": 0.0001866971523591769,
"loss": 0.0045,
"step": 340
},
{
"epoch": 0.07058580004139929,
"grad_norm": 0.01101437397301197,
"learning_rate": 0.00018665558096029933,
"loss": 0.0055,
"step": 341
},
{
"epoch": 0.07079279652245912,
"grad_norm": 0.008826150558888912,
"learning_rate": 0.00018661400956142174,
"loss": 0.005,
"step": 342
},
{
"epoch": 0.07099979300351894,
"grad_norm": 0.006243105512112379,
"learning_rate": 0.00018657243816254417,
"loss": 0.0048,
"step": 343
},
{
"epoch": 0.07120678948457876,
"grad_norm": 0.0014233957044780254,
"learning_rate": 0.0001865308667636666,
"loss": 0.0045,
"step": 344
},
{
"epoch": 0.07141378596563859,
"grad_norm": 0.002639338606968522,
"learning_rate": 0.00018648929536478902,
"loss": 0.0053,
"step": 345
},
{
"epoch": 0.07162078244669841,
"grad_norm": 0.003536937525495887,
"learning_rate": 0.00018644772396591146,
"loss": 0.0048,
"step": 346
},
{
"epoch": 0.07182777892775823,
"grad_norm": 0.0018274744506925344,
"learning_rate": 0.0001864061525670339,
"loss": 0.0043,
"step": 347
},
{
"epoch": 0.07203477540881804,
"grad_norm": 0.004306804854422808,
"learning_rate": 0.0001863645811681563,
"loss": 0.0058,
"step": 348
},
{
"epoch": 0.07224177188987788,
"grad_norm": 0.003877132898196578,
"learning_rate": 0.00018632300976927874,
"loss": 0.0058,
"step": 349
},
{
"epoch": 0.0724487683709377,
"grad_norm": 0.0018924670293927193,
"learning_rate": 0.00018628143837040118,
"loss": 0.0044,
"step": 350
},
{
"epoch": 0.07265576485199751,
"grad_norm": 0.005626944359391928,
"learning_rate": 0.00018623986697152361,
"loss": 0.0046,
"step": 351
},
{
"epoch": 0.07286276133305734,
"grad_norm": 0.006948824506253004,
"learning_rate": 0.00018619829557264602,
"loss": 0.0057,
"step": 352
},
{
"epoch": 0.07306975781411716,
"grad_norm": 0.002097270218655467,
"learning_rate": 0.00018615672417376846,
"loss": 0.0047,
"step": 353
},
{
"epoch": 0.07327675429517698,
"grad_norm": 0.0013399182353168726,
"learning_rate": 0.0001861151527748909,
"loss": 0.0047,
"step": 354
},
{
"epoch": 0.0734837507762368,
"grad_norm": 0.010953530669212341,
"learning_rate": 0.0001860735813760133,
"loss": 0.0061,
"step": 355
},
{
"epoch": 0.07369074725729663,
"grad_norm": 0.0024627153761684895,
"learning_rate": 0.00018603200997713575,
"loss": 0.0061,
"step": 356
},
{
"epoch": 0.07389774373835645,
"grad_norm": 0.002271963283419609,
"learning_rate": 0.00018599043857825818,
"loss": 0.0043,
"step": 357
},
{
"epoch": 0.07410474021941627,
"grad_norm": 0.0036786317359656096,
"learning_rate": 0.0001859488671793806,
"loss": 0.0044,
"step": 358
},
{
"epoch": 0.0743117367004761,
"grad_norm": 0.006179209798574448,
"learning_rate": 0.00018590729578050303,
"loss": 0.0049,
"step": 359
},
{
"epoch": 0.07451873318153591,
"grad_norm": 0.0028029060922563076,
"learning_rate": 0.00018586572438162547,
"loss": 0.0051,
"step": 360
},
{
"epoch": 0.07472572966259573,
"grad_norm": 0.003495444543659687,
"learning_rate": 0.00018582415298274788,
"loss": 0.0055,
"step": 361
},
{
"epoch": 0.07493272614365556,
"grad_norm": 0.005056194495409727,
"learning_rate": 0.0001857825815838703,
"loss": 0.0052,
"step": 362
},
{
"epoch": 0.07513972262471538,
"grad_norm": 0.005216763820499182,
"learning_rate": 0.00018574101018499272,
"loss": 0.0064,
"step": 363
},
{
"epoch": 0.0753467191057752,
"grad_norm": 0.006161578465253115,
"learning_rate": 0.00018569943878611516,
"loss": 0.0051,
"step": 364
},
{
"epoch": 0.07555371558683502,
"grad_norm": 0.0022837778087705374,
"learning_rate": 0.00018565786738723757,
"loss": 0.0046,
"step": 365
},
{
"epoch": 0.07576071206789485,
"grad_norm": 0.0017736790468916297,
"learning_rate": 0.00018561629598836,
"loss": 0.0043,
"step": 366
},
{
"epoch": 0.07596770854895467,
"grad_norm": 0.004538076464086771,
"learning_rate": 0.00018557472458948245,
"loss": 0.0048,
"step": 367
},
{
"epoch": 0.07617470503001449,
"grad_norm": 0.0018419534899294376,
"learning_rate": 0.00018553315319060488,
"loss": 0.0043,
"step": 368
},
{
"epoch": 0.07638170151107432,
"grad_norm": 0.004529993515461683,
"learning_rate": 0.0001854915817917273,
"loss": 0.0044,
"step": 369
},
{
"epoch": 0.07658869799213414,
"grad_norm": 0.0072746858932077885,
"learning_rate": 0.00018545001039284973,
"loss": 0.0045,
"step": 370
},
{
"epoch": 0.07679569447319395,
"grad_norm": 0.013997476547956467,
"learning_rate": 0.00018540843899397217,
"loss": 0.0048,
"step": 371
},
{
"epoch": 0.07700269095425377,
"grad_norm": 0.0026135060470551252,
"learning_rate": 0.00018536686759509458,
"loss": 0.0043,
"step": 372
},
{
"epoch": 0.0772096874353136,
"grad_norm": 0.008647504262626171,
"learning_rate": 0.000185325296196217,
"loss": 0.0043,
"step": 373
},
{
"epoch": 0.07741668391637342,
"grad_norm": 0.0028080667834728956,
"learning_rate": 0.00018528372479733945,
"loss": 0.0045,
"step": 374
},
{
"epoch": 0.07762368039743324,
"grad_norm": 0.00593935651704669,
"learning_rate": 0.00018524215339846186,
"loss": 0.0043,
"step": 375
},
{
"epoch": 0.07783067687849307,
"grad_norm": 0.005520394071936607,
"learning_rate": 0.0001852005819995843,
"loss": 0.0044,
"step": 376
},
{
"epoch": 0.07803767335955289,
"grad_norm": 0.004040780942887068,
"learning_rate": 0.00018515901060070673,
"loss": 0.0044,
"step": 377
},
{
"epoch": 0.07824466984061271,
"grad_norm": 0.006418270990252495,
"learning_rate": 0.00018511743920182914,
"loss": 0.0044,
"step": 378
},
{
"epoch": 0.07845166632167253,
"grad_norm": 0.01350860670208931,
"learning_rate": 0.00018507586780295158,
"loss": 0.0044,
"step": 379
},
{
"epoch": 0.07865866280273236,
"grad_norm": 0.011060641147196293,
"learning_rate": 0.00018503429640407402,
"loss": 0.004,
"step": 380
},
{
"epoch": 0.07886565928379217,
"grad_norm": 0.004036550410091877,
"learning_rate": 0.00018499272500519646,
"loss": 0.0038,
"step": 381
},
{
"epoch": 0.07907265576485199,
"grad_norm": 0.003959359601140022,
"learning_rate": 0.00018495115360631887,
"loss": 0.0038,
"step": 382
},
{
"epoch": 0.07927965224591182,
"grad_norm": 0.005686459131538868,
"learning_rate": 0.00018490958220744128,
"loss": 0.0047,
"step": 383
},
{
"epoch": 0.07948664872697164,
"grad_norm": 0.007630357053130865,
"learning_rate": 0.0001848680108085637,
"loss": 0.0042,
"step": 384
},
{
"epoch": 0.07969364520803146,
"grad_norm": 0.005852843634784222,
"learning_rate": 0.00018482643940968615,
"loss": 0.0036,
"step": 385
},
{
"epoch": 0.07990064168909129,
"grad_norm": 0.005719276610761881,
"learning_rate": 0.00018478486801080856,
"loss": 0.0033,
"step": 386
},
{
"epoch": 0.08010763817015111,
"grad_norm": 0.007002281956374645,
"learning_rate": 0.000184743296611931,
"loss": 0.0033,
"step": 387
},
{
"epoch": 0.08031463465121093,
"grad_norm": 0.007220590952783823,
"learning_rate": 0.00018470172521305343,
"loss": 0.0046,
"step": 388
},
{
"epoch": 0.08052163113227075,
"grad_norm": 0.005552125629037619,
"learning_rate": 0.00018466015381417584,
"loss": 0.0028,
"step": 389
},
{
"epoch": 0.08072862761333058,
"grad_norm": 0.011994168162345886,
"learning_rate": 0.00018461858241529828,
"loss": 0.0028,
"step": 390
},
{
"epoch": 0.0809356240943904,
"grad_norm": 0.011316240765154362,
"learning_rate": 0.00018457701101642072,
"loss": 0.0026,
"step": 391
},
{
"epoch": 0.08114262057545021,
"grad_norm": 0.006591492332518101,
"learning_rate": 0.00018453543961754313,
"loss": 0.0022,
"step": 392
},
{
"epoch": 0.08134961705651005,
"grad_norm": 0.007715560495853424,
"learning_rate": 0.00018449386821866556,
"loss": 0.0032,
"step": 393
},
{
"epoch": 0.08155661353756986,
"grad_norm": 0.009517872706055641,
"learning_rate": 0.000184452296819788,
"loss": 0.0016,
"step": 394
},
{
"epoch": 0.08176361001862968,
"grad_norm": 0.79290372133255,
"learning_rate": 0.0001844107254209104,
"loss": 0.0117,
"step": 395
},
{
"epoch": 0.0819706064996895,
"grad_norm": 0.10478183627128601,
"learning_rate": 0.00018436915402203285,
"loss": 0.0041,
"step": 396
},
{
"epoch": 0.08217760298074933,
"grad_norm": 0.11330251395702362,
"learning_rate": 0.00018432758262315529,
"loss": 0.0349,
"step": 397
},
{
"epoch": 0.08238459946180915,
"grad_norm": 0.022089608013629913,
"learning_rate": 0.00018428601122427772,
"loss": 0.0028,
"step": 398
},
{
"epoch": 0.08259159594286897,
"grad_norm": 0.045795392245054245,
"learning_rate": 0.00018424443982540013,
"loss": 0.0027,
"step": 399
},
{
"epoch": 0.0827985924239288,
"grad_norm": 0.052710726857185364,
"learning_rate": 0.00018420286842652257,
"loss": 0.0033,
"step": 400
},
{
"epoch": 0.08300558890498862,
"grad_norm": 0.0732388123869896,
"learning_rate": 0.000184161297027645,
"loss": 0.0031,
"step": 401
},
{
"epoch": 0.08321258538604843,
"grad_norm": 0.03560757264494896,
"learning_rate": 0.00018411972562876742,
"loss": 0.0025,
"step": 402
},
{
"epoch": 0.08341958186710825,
"grad_norm": 0.039032500237226486,
"learning_rate": 0.00018407815422988985,
"loss": 0.0027,
"step": 403
},
{
"epoch": 0.08362657834816808,
"grad_norm": 0.018673432990908623,
"learning_rate": 0.0001840365828310123,
"loss": 0.0018,
"step": 404
},
{
"epoch": 0.0838335748292279,
"grad_norm": 0.06421470642089844,
"learning_rate": 0.0001839950114321347,
"loss": 0.0293,
"step": 405
},
{
"epoch": 0.08404057131028772,
"grad_norm": 0.05356355383992195,
"learning_rate": 0.0001839534400332571,
"loss": 0.0234,
"step": 406
},
{
"epoch": 0.08424756779134755,
"grad_norm": 0.047022175043821335,
"learning_rate": 0.00018391186863437955,
"loss": 0.0198,
"step": 407
},
{
"epoch": 0.08445456427240737,
"grad_norm": 0.017866840586066246,
"learning_rate": 0.00018387029723550199,
"loss": 0.0017,
"step": 408
},
{
"epoch": 0.08466156075346719,
"grad_norm": 0.010899499990046024,
"learning_rate": 0.0001838287258366244,
"loss": 0.0013,
"step": 409
},
{
"epoch": 0.084868557234527,
"grad_norm": 0.0167918112128973,
"learning_rate": 0.00018378715443774683,
"loss": 0.0024,
"step": 410
},
{
"epoch": 0.08507555371558684,
"grad_norm": 0.02067534811794758,
"learning_rate": 0.00018374558303886927,
"loss": 0.0015,
"step": 411
},
{
"epoch": 0.08528255019664666,
"grad_norm": 0.01670040749013424,
"learning_rate": 0.00018370401163999168,
"loss": 0.0015,
"step": 412
},
{
"epoch": 0.08548954667770647,
"grad_norm": 0.008924894034862518,
"learning_rate": 0.00018366244024111412,
"loss": 0.0029,
"step": 413
},
{
"epoch": 0.0856965431587663,
"grad_norm": 0.013602840714156628,
"learning_rate": 0.00018362086884223655,
"loss": 0.0009,
"step": 414
},
{
"epoch": 0.08590353963982612,
"grad_norm": 0.013082594610750675,
"learning_rate": 0.000183579297443359,
"loss": 0.002,
"step": 415
},
{
"epoch": 0.08611053612088594,
"grad_norm": 0.012215960770845413,
"learning_rate": 0.0001835377260444814,
"loss": 0.0017,
"step": 416
},
{
"epoch": 0.08631753260194577,
"grad_norm": 0.16738812625408173,
"learning_rate": 0.00018349615464560384,
"loss": 0.0173,
"step": 417
},
{
"epoch": 0.08652452908300559,
"grad_norm": 0.006629611365497112,
"learning_rate": 0.00018345458324672627,
"loss": 0.0006,
"step": 418
},
{
"epoch": 0.08673152556406541,
"grad_norm": 0.006643650587648153,
"learning_rate": 0.00018341301184784868,
"loss": 0.0015,
"step": 419
},
{
"epoch": 0.08693852204512523,
"grad_norm": 0.0039656031876802444,
"learning_rate": 0.00018337144044897112,
"loss": 0.0003,
"step": 420
},
{
"epoch": 0.08714551852618506,
"grad_norm": 0.00564931146800518,
"learning_rate": 0.00018332986905009356,
"loss": 0.0004,
"step": 421
},
{
"epoch": 0.08735251500724488,
"grad_norm": 0.014363352209329605,
"learning_rate": 0.00018328829765121597,
"loss": 0.0006,
"step": 422
},
{
"epoch": 0.0875595114883047,
"grad_norm": 0.006862149108201265,
"learning_rate": 0.0001832467262523384,
"loss": 0.0004,
"step": 423
},
{
"epoch": 0.08776650796936453,
"grad_norm": 0.005224172957241535,
"learning_rate": 0.00018320515485346084,
"loss": 0.0003,
"step": 424
},
{
"epoch": 0.08797350445042434,
"grad_norm": 0.012813829816877842,
"learning_rate": 0.00018316358345458325,
"loss": 0.0005,
"step": 425
},
{
"epoch": 0.08818050093148416,
"grad_norm": 0.0045601376332342625,
"learning_rate": 0.0001831220120557057,
"loss": 0.0003,
"step": 426
},
{
"epoch": 0.08838749741254398,
"grad_norm": 0.002229505218565464,
"learning_rate": 0.0001830804406568281,
"loss": 0.0002,
"step": 427
},
{
"epoch": 0.08859449389360381,
"grad_norm": 0.005202361848205328,
"learning_rate": 0.00018303886925795054,
"loss": 0.0003,
"step": 428
},
{
"epoch": 0.08880149037466363,
"grad_norm": 0.010837195441126823,
"learning_rate": 0.00018299729785907295,
"loss": 0.0016,
"step": 429
},
{
"epoch": 0.08900848685572345,
"grad_norm": 0.006401981692761183,
"learning_rate": 0.00018295572646019538,
"loss": 0.0002,
"step": 430
},
{
"epoch": 0.08921548333678328,
"grad_norm": 0.0025153057649731636,
"learning_rate": 0.00018291415506131782,
"loss": 0.0001,
"step": 431
},
{
"epoch": 0.0894224798178431,
"grad_norm": 0.009693821892142296,
"learning_rate": 0.00018287258366244023,
"loss": 0.0004,
"step": 432
},
{
"epoch": 0.08962947629890292,
"grad_norm": 0.0013723783195018768,
"learning_rate": 0.00018283101226356267,
"loss": 0.0001,
"step": 433
},
{
"epoch": 0.08983647277996273,
"grad_norm": 0.008555575273931026,
"learning_rate": 0.0001827894408646851,
"loss": 0.0018,
"step": 434
},
{
"epoch": 0.09004346926102257,
"grad_norm": 0.0028277519159018993,
"learning_rate": 0.00018274786946580754,
"loss": 0.0002,
"step": 435
},
{
"epoch": 0.09025046574208238,
"grad_norm": 0.014325006864964962,
"learning_rate": 0.00018270629806692995,
"loss": 0.0009,
"step": 436
},
{
"epoch": 0.0904574622231422,
"grad_norm": 0.008406553417444229,
"learning_rate": 0.0001826647266680524,
"loss": 0.0012,
"step": 437
},
{
"epoch": 0.09066445870420203,
"grad_norm": 0.0018985685892403126,
"learning_rate": 0.00018262315526917483,
"loss": 0.0002,
"step": 438
},
{
"epoch": 0.09087145518526185,
"grad_norm": 0.008550492115318775,
"learning_rate": 0.00018258158387029724,
"loss": 0.0024,
"step": 439
},
{
"epoch": 0.09107845166632167,
"grad_norm": 0.0008987212786450982,
"learning_rate": 0.00018254001247141967,
"loss": 0.0001,
"step": 440
},
{
"epoch": 0.0912854481473815,
"grad_norm": 0.002059886697679758,
"learning_rate": 0.0001824984410725421,
"loss": 0.0001,
"step": 441
},
{
"epoch": 0.09149244462844132,
"grad_norm": 0.003429644973948598,
"learning_rate": 0.00018245686967366452,
"loss": 0.0001,
"step": 442
},
{
"epoch": 0.09169944110950114,
"grad_norm": 0.0026945085264742374,
"learning_rate": 0.00018241529827478696,
"loss": 0.0011,
"step": 443
},
{
"epoch": 0.09190643759056095,
"grad_norm": 0.00299448031000793,
"learning_rate": 0.0001823737268759094,
"loss": 0.0001,
"step": 444
},
{
"epoch": 0.09211343407162079,
"grad_norm": 0.004376763943582773,
"learning_rate": 0.00018233215547703183,
"loss": 0.0011,
"step": 445
},
{
"epoch": 0.0923204305526806,
"grad_norm": 0.005078963004052639,
"learning_rate": 0.00018229058407815424,
"loss": 0.0002,
"step": 446
},
{
"epoch": 0.09252742703374042,
"grad_norm": 0.009306194260716438,
"learning_rate": 0.00018224901267927668,
"loss": 0.0003,
"step": 447
},
{
"epoch": 0.09273442351480025,
"grad_norm": 0.0030481938738375902,
"learning_rate": 0.00018220744128039912,
"loss": 0.0013,
"step": 448
},
{
"epoch": 0.09294141999586007,
"grad_norm": 0.013268841430544853,
"learning_rate": 0.0001821658698815215,
"loss": 0.0017,
"step": 449
},
{
"epoch": 0.09314841647691989,
"grad_norm": 0.006063752807676792,
"learning_rate": 0.00018212429848264394,
"loss": 0.0002,
"step": 450
},
{
"epoch": 0.09335541295797971,
"grad_norm": 0.00182344822678715,
"learning_rate": 0.00018208272708376637,
"loss": 0.0001,
"step": 451
},
{
"epoch": 0.09356240943903954,
"grad_norm": 0.009372780099511147,
"learning_rate": 0.0001820411556848888,
"loss": 0.0005,
"step": 452
},
{
"epoch": 0.09376940592009936,
"grad_norm": 0.003136920742690563,
"learning_rate": 0.00018199958428601122,
"loss": 0.0001,
"step": 453
},
{
"epoch": 0.09397640240115918,
"grad_norm": 0.030062230303883553,
"learning_rate": 0.00018195801288713366,
"loss": 0.0006,
"step": 454
},
{
"epoch": 0.09418339888221901,
"grad_norm": 0.004309754353016615,
"learning_rate": 0.0001819164414882561,
"loss": 0.0007,
"step": 455
},
{
"epoch": 0.09439039536327883,
"grad_norm": 0.010606180876493454,
"learning_rate": 0.0001818748700893785,
"loss": 0.0002,
"step": 456
},
{
"epoch": 0.09459739184433864,
"grad_norm": 0.0054748812690377235,
"learning_rate": 0.00018183329869050094,
"loss": 0.0005,
"step": 457
},
{
"epoch": 0.09480438832539846,
"grad_norm": 0.001673020888119936,
"learning_rate": 0.00018179172729162338,
"loss": 0.0001,
"step": 458
},
{
"epoch": 0.0950113848064583,
"grad_norm": 0.004401384387165308,
"learning_rate": 0.0001817501558927458,
"loss": 0.0001,
"step": 459
},
{
"epoch": 0.09521838128751811,
"grad_norm": 0.005755012389272451,
"learning_rate": 0.00018170858449386822,
"loss": 0.0012,
"step": 460
},
{
"epoch": 0.09542537776857793,
"grad_norm": 0.004951901733875275,
"learning_rate": 0.00018166701309499066,
"loss": 0.0012,
"step": 461
},
{
"epoch": 0.09563237424963776,
"grad_norm": 0.0014370133867487311,
"learning_rate": 0.0001816254416961131,
"loss": 0.0001,
"step": 462
},
{
"epoch": 0.09583937073069758,
"grad_norm": 0.012078122235834599,
"learning_rate": 0.0001815838702972355,
"loss": 0.0013,
"step": 463
},
{
"epoch": 0.0960463672117574,
"grad_norm": 0.002765175886452198,
"learning_rate": 0.00018154229889835795,
"loss": 0.0009,
"step": 464
},
{
"epoch": 0.09625336369281723,
"grad_norm": 0.0017701378092169762,
"learning_rate": 0.00018150072749948038,
"loss": 0.0001,
"step": 465
},
{
"epoch": 0.09646036017387705,
"grad_norm": 0.003232579445466399,
"learning_rate": 0.0001814591561006028,
"loss": 0.0001,
"step": 466
},
{
"epoch": 0.09666735665493686,
"grad_norm": 0.004619232844561338,
"learning_rate": 0.00018141758470172523,
"loss": 0.0012,
"step": 467
},
{
"epoch": 0.09687435313599668,
"grad_norm": 0.008126890286803246,
"learning_rate": 0.00018137601330284767,
"loss": 0.0011,
"step": 468
},
{
"epoch": 0.09708134961705651,
"grad_norm": 0.004719397984445095,
"learning_rate": 0.00018133444190397008,
"loss": 0.0009,
"step": 469
},
{
"epoch": 0.09728834609811633,
"grad_norm": 0.005153202451765537,
"learning_rate": 0.00018129287050509251,
"loss": 0.0012,
"step": 470
},
{
"epoch": 0.09749534257917615,
"grad_norm": 0.01385215763002634,
"learning_rate": 0.00018125129910621492,
"loss": 0.0022,
"step": 471
},
{
"epoch": 0.09770233906023598,
"grad_norm": 0.004983994178473949,
"learning_rate": 0.00018120972770733736,
"loss": 0.0004,
"step": 472
},
{
"epoch": 0.0979093355412958,
"grad_norm": 0.007088206708431244,
"learning_rate": 0.00018116815630845977,
"loss": 0.0005,
"step": 473
},
{
"epoch": 0.09811633202235562,
"grad_norm": 0.004754175432026386,
"learning_rate": 0.0001811265849095822,
"loss": 0.0006,
"step": 474
},
{
"epoch": 0.09832332850341544,
"grad_norm": 0.004105637315660715,
"learning_rate": 0.00018108501351070465,
"loss": 0.0001,
"step": 475
},
{
"epoch": 0.09853032498447527,
"grad_norm": 0.002687152475118637,
"learning_rate": 0.00018104344211182706,
"loss": 0.0001,
"step": 476
},
{
"epoch": 0.09873732146553509,
"grad_norm": 0.0023124783765524626,
"learning_rate": 0.0001810018707129495,
"loss": 0.0001,
"step": 477
},
{
"epoch": 0.0989443179465949,
"grad_norm": 0.007885076105594635,
"learning_rate": 0.00018096029931407193,
"loss": 0.0006,
"step": 478
},
{
"epoch": 0.09915131442765474,
"grad_norm": 0.0014087413437664509,
"learning_rate": 0.00018091872791519434,
"loss": 0.0007,
"step": 479
},
{
"epoch": 0.09935831090871455,
"grad_norm": 0.0055119190365076065,
"learning_rate": 0.00018087715651631678,
"loss": 0.0002,
"step": 480
},
{
"epoch": 0.09956530738977437,
"grad_norm": 0.0003096537839155644,
"learning_rate": 0.0001808355851174392,
"loss": 0.0,
"step": 481
},
{
"epoch": 0.09977230387083419,
"grad_norm": 0.001899409806355834,
"learning_rate": 0.00018079401371856165,
"loss": 0.0002,
"step": 482
},
{
"epoch": 0.09997930035189402,
"grad_norm": 0.005635194014757872,
"learning_rate": 0.00018075244231968406,
"loss": 0.0001,
"step": 483
},
{
"epoch": 0.10018629683295384,
"grad_norm": 0.005856087896972895,
"learning_rate": 0.0001807108709208065,
"loss": 0.001,
"step": 484
},
{
"epoch": 0.10039329331401366,
"grad_norm": 0.005273948423564434,
"learning_rate": 0.00018066929952192893,
"loss": 0.0013,
"step": 485
},
{
"epoch": 0.10060028979507349,
"grad_norm": 0.001652201754041016,
"learning_rate": 0.00018062772812305134,
"loss": 0.0001,
"step": 486
},
{
"epoch": 0.1008072862761333,
"grad_norm": 0.006849886849522591,
"learning_rate": 0.00018058615672417378,
"loss": 0.0001,
"step": 487
},
{
"epoch": 0.10101428275719312,
"grad_norm": 0.009754250757396221,
"learning_rate": 0.00018054458532529622,
"loss": 0.0023,
"step": 488
},
{
"epoch": 0.10122127923825296,
"grad_norm": 0.0038455536123365164,
"learning_rate": 0.00018050301392641863,
"loss": 0.0016,
"step": 489
},
{
"epoch": 0.10142827571931277,
"grad_norm": 0.0006483698962256312,
"learning_rate": 0.00018046144252754107,
"loss": 0.0,
"step": 490
},
{
"epoch": 0.10163527220037259,
"grad_norm": 0.008700639940798283,
"learning_rate": 0.0001804198711286635,
"loss": 0.0007,
"step": 491
},
{
"epoch": 0.10184226868143241,
"grad_norm": 0.004151639994233847,
"learning_rate": 0.00018037829972978594,
"loss": 0.0006,
"step": 492
},
{
"epoch": 0.10204926516249224,
"grad_norm": 0.003242357401177287,
"learning_rate": 0.00018033672833090832,
"loss": 0.0001,
"step": 493
},
{
"epoch": 0.10225626164355206,
"grad_norm": 0.0038309101946651936,
"learning_rate": 0.00018029515693203076,
"loss": 0.0002,
"step": 494
},
{
"epoch": 0.10246325812461188,
"grad_norm": 0.0032492976170033216,
"learning_rate": 0.0001802535855331532,
"loss": 0.0001,
"step": 495
},
{
"epoch": 0.10267025460567171,
"grad_norm": 0.005621058400720358,
"learning_rate": 0.0001802120141342756,
"loss": 0.0002,
"step": 496
},
{
"epoch": 0.10287725108673153,
"grad_norm": 0.007613383699208498,
"learning_rate": 0.00018017044273539804,
"loss": 0.0013,
"step": 497
},
{
"epoch": 0.10308424756779135,
"grad_norm": 0.004469983279705048,
"learning_rate": 0.00018012887133652048,
"loss": 0.0001,
"step": 498
},
{
"epoch": 0.10329124404885116,
"grad_norm": 0.010518516413867474,
"learning_rate": 0.00018008729993764292,
"loss": 0.0002,
"step": 499
},
{
"epoch": 0.103498240529911,
"grad_norm": 0.0019439860479906201,
"learning_rate": 0.00018004572853876533,
"loss": 0.0001,
"step": 500
},
{
"epoch": 0.10370523701097081,
"grad_norm": 0.007315011695027351,
"learning_rate": 0.00018000415713988776,
"loss": 0.0002,
"step": 501
},
{
"epoch": 0.10391223349203063,
"grad_norm": 0.0027510204818099737,
"learning_rate": 0.0001799625857410102,
"loss": 0.0002,
"step": 502
},
{
"epoch": 0.10411922997309046,
"grad_norm": 0.007963057607412338,
"learning_rate": 0.0001799210143421326,
"loss": 0.0003,
"step": 503
},
{
"epoch": 0.10432622645415028,
"grad_norm": 0.004816776607185602,
"learning_rate": 0.00017987944294325505,
"loss": 0.001,
"step": 504
},
{
"epoch": 0.1045332229352101,
"grad_norm": 0.0010107432026416063,
"learning_rate": 0.00017983787154437749,
"loss": 0.0,
"step": 505
},
{
"epoch": 0.10474021941626992,
"grad_norm": 0.002849761163815856,
"learning_rate": 0.0001797963001454999,
"loss": 0.0001,
"step": 506
},
{
"epoch": 0.10494721589732975,
"grad_norm": 0.008465790189802647,
"learning_rate": 0.00017975472874662233,
"loss": 0.0017,
"step": 507
},
{
"epoch": 0.10515421237838957,
"grad_norm": 0.001012888620607555,
"learning_rate": 0.00017971315734774477,
"loss": 0.0,
"step": 508
},
{
"epoch": 0.10536120885944938,
"grad_norm": 0.005154603160917759,
"learning_rate": 0.0001796715859488672,
"loss": 0.0001,
"step": 509
},
{
"epoch": 0.10556820534050922,
"grad_norm": 0.012283824384212494,
"learning_rate": 0.00017963001454998962,
"loss": 0.0004,
"step": 510
},
{
"epoch": 0.10577520182156903,
"grad_norm": 0.001696570310741663,
"learning_rate": 0.00017958844315111205,
"loss": 0.0001,
"step": 511
},
{
"epoch": 0.10598219830262885,
"grad_norm": 0.0002515624219086021,
"learning_rate": 0.0001795468717522345,
"loss": 0.0,
"step": 512
},
{
"epoch": 0.10618919478368868,
"grad_norm": 0.0006335057551041245,
"learning_rate": 0.0001795053003533569,
"loss": 0.0,
"step": 513
},
{
"epoch": 0.1063961912647485,
"grad_norm": 0.0007280270801857114,
"learning_rate": 0.00017946372895447934,
"loss": 0.0,
"step": 514
},
{
"epoch": 0.10660318774580832,
"grad_norm": 0.0040188622660934925,
"learning_rate": 0.00017942215755560175,
"loss": 0.0002,
"step": 515
},
{
"epoch": 0.10681018422686814,
"grad_norm": 0.0054796175099909306,
"learning_rate": 0.00017938058615672419,
"loss": 0.0009,
"step": 516
},
{
"epoch": 0.10701718070792797,
"grad_norm": 0.0034792469814419746,
"learning_rate": 0.0001793390147578466,
"loss": 0.0001,
"step": 517
},
{
"epoch": 0.10722417718898779,
"grad_norm": 0.00432013813406229,
"learning_rate": 0.00017929744335896903,
"loss": 0.0002,
"step": 518
},
{
"epoch": 0.1074311736700476,
"grad_norm": 0.004862105939537287,
"learning_rate": 0.00017925587196009147,
"loss": 0.001,
"step": 519
},
{
"epoch": 0.10763817015110744,
"grad_norm": 0.002249139128252864,
"learning_rate": 0.00017921430056121388,
"loss": 0.0,
"step": 520
},
{
"epoch": 0.10784516663216726,
"grad_norm": 0.006818681955337524,
"learning_rate": 0.00017917272916233632,
"loss": 0.0004,
"step": 521
},
{
"epoch": 0.10805216311322707,
"grad_norm": 0.008624670095741749,
"learning_rate": 0.00017913115776345875,
"loss": 0.0004,
"step": 522
},
{
"epoch": 0.10825915959428689,
"grad_norm": 0.0007210278417915106,
"learning_rate": 0.00017908958636458116,
"loss": 0.0,
"step": 523
},
{
"epoch": 0.10846615607534672,
"grad_norm": 0.010820691473782063,
"learning_rate": 0.0001790480149657036,
"loss": 0.0003,
"step": 524
},
{
"epoch": 0.10867315255640654,
"grad_norm": 0.0023018312640488148,
"learning_rate": 0.00017900644356682604,
"loss": 0.0001,
"step": 525
},
{
"epoch": 0.10888014903746636,
"grad_norm": 0.00019024198991246521,
"learning_rate": 0.00017896487216794845,
"loss": 0.0,
"step": 526
},
{
"epoch": 0.10908714551852619,
"grad_norm": 0.005011410918086767,
"learning_rate": 0.00017892330076907088,
"loss": 0.0003,
"step": 527
},
{
"epoch": 0.10929414199958601,
"grad_norm": 0.007016469724476337,
"learning_rate": 0.00017888172937019332,
"loss": 0.0009,
"step": 528
},
{
"epoch": 0.10950113848064583,
"grad_norm": 0.008118787780404091,
"learning_rate": 0.00017884015797131576,
"loss": 0.0003,
"step": 529
},
{
"epoch": 0.10970813496170564,
"grad_norm": 0.005854643415659666,
"learning_rate": 0.00017879858657243817,
"loss": 0.0011,
"step": 530
},
{
"epoch": 0.10991513144276548,
"grad_norm": 0.0038967933505773544,
"learning_rate": 0.0001787570151735606,
"loss": 0.001,
"step": 531
},
{
"epoch": 0.1101221279238253,
"grad_norm": 0.003802061313763261,
"learning_rate": 0.00017871544377468304,
"loss": 0.0018,
"step": 532
},
{
"epoch": 0.11032912440488511,
"grad_norm": 0.004740913398563862,
"learning_rate": 0.00017867387237580545,
"loss": 0.0001,
"step": 533
},
{
"epoch": 0.11053612088594494,
"grad_norm": 0.008046228438615799,
"learning_rate": 0.0001786323009769279,
"loss": 0.0005,
"step": 534
},
{
"epoch": 0.11074311736700476,
"grad_norm": 0.0014560514828190207,
"learning_rate": 0.00017859072957805033,
"loss": 0.0,
"step": 535
},
{
"epoch": 0.11095011384806458,
"grad_norm": 0.009222283028066158,
"learning_rate": 0.00017854915817917274,
"loss": 0.0002,
"step": 536
},
{
"epoch": 0.11115711032912441,
"grad_norm": 0.005452610552310944,
"learning_rate": 0.00017850758678029515,
"loss": 0.0013,
"step": 537
},
{
"epoch": 0.11136410681018423,
"grad_norm": 0.0007964425021782517,
"learning_rate": 0.00017846601538141758,
"loss": 0.0,
"step": 538
},
{
"epoch": 0.11157110329124405,
"grad_norm": 0.007167865987867117,
"learning_rate": 0.00017842444398254002,
"loss": 0.0011,
"step": 539
},
{
"epoch": 0.11177809977230387,
"grad_norm": 0.01039041206240654,
"learning_rate": 0.00017838287258366243,
"loss": 0.0007,
"step": 540
},
{
"epoch": 0.1119850962533637,
"grad_norm": 0.0001594142959220335,
"learning_rate": 0.00017834130118478487,
"loss": 0.0,
"step": 541
},
{
"epoch": 0.11219209273442352,
"grad_norm": 0.0023100003600120544,
"learning_rate": 0.0001782997297859073,
"loss": 0.0007,
"step": 542
},
{
"epoch": 0.11239908921548333,
"grad_norm": 0.003656044602394104,
"learning_rate": 0.00017825815838702972,
"loss": 0.0014,
"step": 543
},
{
"epoch": 0.11260608569654317,
"grad_norm": 0.0014201959129422903,
"learning_rate": 0.00017821658698815215,
"loss": 0.0001,
"step": 544
},
{
"epoch": 0.11281308217760298,
"grad_norm": 0.015617001801729202,
"learning_rate": 0.0001781750155892746,
"loss": 0.0002,
"step": 545
},
{
"epoch": 0.1130200786586628,
"grad_norm": 0.008762934245169163,
"learning_rate": 0.00017813344419039703,
"loss": 0.0003,
"step": 546
},
{
"epoch": 0.11322707513972262,
"grad_norm": 0.002274678787216544,
"learning_rate": 0.00017809187279151944,
"loss": 0.0001,
"step": 547
},
{
"epoch": 0.11343407162078245,
"grad_norm": 0.003864066442474723,
"learning_rate": 0.00017805030139264187,
"loss": 0.0005,
"step": 548
},
{
"epoch": 0.11364106810184227,
"grad_norm": 0.006032771430909634,
"learning_rate": 0.0001780087299937643,
"loss": 0.0003,
"step": 549
},
{
"epoch": 0.11384806458290209,
"grad_norm": 0.0011935516959056258,
"learning_rate": 0.00017796715859488672,
"loss": 0.0,
"step": 550
},
{
"epoch": 0.11405506106396192,
"grad_norm": 0.013128140941262245,
"learning_rate": 0.00017792558719600916,
"loss": 0.0006,
"step": 551
},
{
"epoch": 0.11426205754502174,
"grad_norm": 0.000645163469016552,
"learning_rate": 0.0001778840157971316,
"loss": 0.0,
"step": 552
},
{
"epoch": 0.11446905402608155,
"grad_norm": 0.011457535438239574,
"learning_rate": 0.000177842444398254,
"loss": 0.0003,
"step": 553
},
{
"epoch": 0.11467605050714137,
"grad_norm": 0.003057427005842328,
"learning_rate": 0.00017780087299937644,
"loss": 0.0012,
"step": 554
},
{
"epoch": 0.1148830469882012,
"grad_norm": 0.0017485780408605933,
"learning_rate": 0.00017775930160049888,
"loss": 0.0006,
"step": 555
},
{
"epoch": 0.11509004346926102,
"grad_norm": 0.0005046813748776913,
"learning_rate": 0.0001777177302016213,
"loss": 0.0,
"step": 556
},
{
"epoch": 0.11529703995032084,
"grad_norm": 0.007802332751452923,
"learning_rate": 0.00017767615880274373,
"loss": 0.0009,
"step": 557
},
{
"epoch": 0.11550403643138067,
"grad_norm": 0.0074394443072378635,
"learning_rate": 0.00017763458740386616,
"loss": 0.0021,
"step": 558
},
{
"epoch": 0.11571103291244049,
"grad_norm": 0.007590603083372116,
"learning_rate": 0.00017759301600498857,
"loss": 0.0004,
"step": 559
},
{
"epoch": 0.11591802939350031,
"grad_norm": 0.005805825348943472,
"learning_rate": 0.00017755144460611098,
"loss": 0.0019,
"step": 560
},
{
"epoch": 0.11612502587456014,
"grad_norm": 0.008781611919403076,
"learning_rate": 0.00017750987320723342,
"loss": 0.0017,
"step": 561
},
{
"epoch": 0.11633202235561996,
"grad_norm": 0.0010139975929632783,
"learning_rate": 0.00017746830180835586,
"loss": 0.0,
"step": 562
},
{
"epoch": 0.11653901883667978,
"grad_norm": 0.003940982278436422,
"learning_rate": 0.0001774267304094783,
"loss": 0.0001,
"step": 563
},
{
"epoch": 0.1167460153177396,
"grad_norm": 0.013145407661795616,
"learning_rate": 0.0001773851590106007,
"loss": 0.0012,
"step": 564
},
{
"epoch": 0.11695301179879942,
"grad_norm": 0.0011159079149365425,
"learning_rate": 0.00017734358761172314,
"loss": 0.0001,
"step": 565
},
{
"epoch": 0.11716000827985924,
"grad_norm": 0.01270979829132557,
"learning_rate": 0.00017730201621284558,
"loss": 0.0017,
"step": 566
},
{
"epoch": 0.11736700476091906,
"grad_norm": 0.004431411158293486,
"learning_rate": 0.000177260444813968,
"loss": 0.0001,
"step": 567
},
{
"epoch": 0.11757400124197889,
"grad_norm": 0.0041249035857617855,
"learning_rate": 0.00017721887341509042,
"loss": 0.0003,
"step": 568
},
{
"epoch": 0.11778099772303871,
"grad_norm": 0.0006257002823986113,
"learning_rate": 0.00017717730201621286,
"loss": 0.0,
"step": 569
},
{
"epoch": 0.11798799420409853,
"grad_norm": 0.004501596093177795,
"learning_rate": 0.00017713573061733527,
"loss": 0.0001,
"step": 570
},
{
"epoch": 0.11819499068515835,
"grad_norm": 0.00742512010037899,
"learning_rate": 0.0001770941592184577,
"loss": 0.0002,
"step": 571
},
{
"epoch": 0.11840198716621818,
"grad_norm": 0.010421551764011383,
"learning_rate": 0.00017705258781958015,
"loss": 0.0008,
"step": 572
},
{
"epoch": 0.118608983647278,
"grad_norm": 0.0010451058624312282,
"learning_rate": 0.00017701101642070256,
"loss": 0.0001,
"step": 573
},
{
"epoch": 0.11881598012833781,
"grad_norm": 0.0014272347325459123,
"learning_rate": 0.000176969445021825,
"loss": 0.0001,
"step": 574
},
{
"epoch": 0.11902297660939765,
"grad_norm": 0.007021667901426554,
"learning_rate": 0.00017692787362294743,
"loss": 0.0002,
"step": 575
},
{
"epoch": 0.11922997309045746,
"grad_norm": 0.0050498368218541145,
"learning_rate": 0.00017688630222406987,
"loss": 0.001,
"step": 576
},
{
"epoch": 0.11943696957151728,
"grad_norm": 0.0004365240456536412,
"learning_rate": 0.00017684473082519228,
"loss": 0.0,
"step": 577
},
{
"epoch": 0.1196439660525771,
"grad_norm": 0.007502545602619648,
"learning_rate": 0.00017680315942631471,
"loss": 0.0002,
"step": 578
},
{
"epoch": 0.11985096253363693,
"grad_norm": 0.00824455451220274,
"learning_rate": 0.00017676158802743715,
"loss": 0.0004,
"step": 579
},
{
"epoch": 0.12005795901469675,
"grad_norm": 0.003414528677240014,
"learning_rate": 0.00017672001662855956,
"loss": 0.0011,
"step": 580
},
{
"epoch": 0.12026495549575657,
"grad_norm": 0.004608092829585075,
"learning_rate": 0.00017667844522968197,
"loss": 0.0014,
"step": 581
},
{
"epoch": 0.1204719519768164,
"grad_norm": 0.006573988124728203,
"learning_rate": 0.0001766368738308044,
"loss": 0.0002,
"step": 582
},
{
"epoch": 0.12067894845787622,
"grad_norm": 0.006878604646772146,
"learning_rate": 0.00017659530243192685,
"loss": 0.0005,
"step": 583
},
{
"epoch": 0.12088594493893604,
"grad_norm": 0.0013765916228294373,
"learning_rate": 0.00017655373103304926,
"loss": 0.0001,
"step": 584
},
{
"epoch": 0.12109294141999585,
"grad_norm": 0.009517376311123371,
"learning_rate": 0.0001765121596341717,
"loss": 0.0005,
"step": 585
},
{
"epoch": 0.12129993790105568,
"grad_norm": 0.02729586698114872,
"learning_rate": 0.00017647058823529413,
"loss": 0.0013,
"step": 586
},
{
"epoch": 0.1215069343821155,
"grad_norm": 0.005033944733440876,
"learning_rate": 0.00017642901683641654,
"loss": 0.0011,
"step": 587
},
{
"epoch": 0.12171393086317532,
"grad_norm": 0.001488934038206935,
"learning_rate": 0.00017638744543753898,
"loss": 0.0001,
"step": 588
},
{
"epoch": 0.12192092734423515,
"grad_norm": 0.004233523737639189,
"learning_rate": 0.0001763458740386614,
"loss": 0.0013,
"step": 589
},
{
"epoch": 0.12212792382529497,
"grad_norm": 0.001819688593968749,
"learning_rate": 0.00017630430263978382,
"loss": 0.0002,
"step": 590
},
{
"epoch": 0.12233492030635479,
"grad_norm": 0.0051133958622813225,
"learning_rate": 0.00017626273124090626,
"loss": 0.0003,
"step": 591
},
{
"epoch": 0.12254191678741462,
"grad_norm": 0.007632863707840443,
"learning_rate": 0.0001762211598420287,
"loss": 0.0005,
"step": 592
},
{
"epoch": 0.12274891326847444,
"grad_norm": 0.0009289845474995673,
"learning_rate": 0.00017617958844315113,
"loss": 0.0,
"step": 593
},
{
"epoch": 0.12295590974953426,
"grad_norm": 0.00543027650564909,
"learning_rate": 0.00017613801704427354,
"loss": 0.0001,
"step": 594
},
{
"epoch": 0.12316290623059407,
"grad_norm": 0.002607417991384864,
"learning_rate": 0.00017609644564539598,
"loss": 0.0002,
"step": 595
},
{
"epoch": 0.1233699027116539,
"grad_norm": 0.025557972490787506,
"learning_rate": 0.00017605487424651842,
"loss": 0.0002,
"step": 596
},
{
"epoch": 0.12357689919271372,
"grad_norm": 0.0016189507441595197,
"learning_rate": 0.00017601330284764083,
"loss": 0.0001,
"step": 597
},
{
"epoch": 0.12378389567377354,
"grad_norm": 0.004612909164279699,
"learning_rate": 0.00017597173144876327,
"loss": 0.0002,
"step": 598
},
{
"epoch": 0.12399089215483337,
"grad_norm": 0.00019464526849333197,
"learning_rate": 0.0001759301600498857,
"loss": 0.0,
"step": 599
},
{
"epoch": 0.12419788863589319,
"grad_norm": 0.0013309603091329336,
"learning_rate": 0.0001758885886510081,
"loss": 0.0007,
"step": 600
},
{
"epoch": 0.12440488511695301,
"grad_norm": 0.002917417325079441,
"learning_rate": 0.00017584701725213055,
"loss": 0.0001,
"step": 601
},
{
"epoch": 0.12461188159801283,
"grad_norm": 0.004730269778519869,
"learning_rate": 0.000175805445853253,
"loss": 0.0001,
"step": 602
},
{
"epoch": 0.12481887807907266,
"grad_norm": 0.0036635478027164936,
"learning_rate": 0.0001757638744543754,
"loss": 0.0001,
"step": 603
},
{
"epoch": 0.12502587456013248,
"grad_norm": 0.002084661042317748,
"learning_rate": 0.0001757223030554978,
"loss": 0.0001,
"step": 604
},
{
"epoch": 0.1252328710411923,
"grad_norm": 0.006881284527480602,
"learning_rate": 0.00017568073165662024,
"loss": 0.0002,
"step": 605
},
{
"epoch": 0.1254398675222521,
"grad_norm": 0.0007496286416426301,
"learning_rate": 0.00017563916025774268,
"loss": 0.0,
"step": 606
},
{
"epoch": 0.12564686400331193,
"grad_norm": 0.0013991744490340352,
"learning_rate": 0.0001755975888588651,
"loss": 0.0,
"step": 607
},
{
"epoch": 0.12585386048437178,
"grad_norm": 0.00578208127990365,
"learning_rate": 0.00017555601745998753,
"loss": 0.0016,
"step": 608
},
{
"epoch": 0.1260608569654316,
"grad_norm": 0.0005476415390148759,
"learning_rate": 0.00017551444606110996,
"loss": 0.0,
"step": 609
},
{
"epoch": 0.1262678534464914,
"grad_norm": 0.003824407234787941,
"learning_rate": 0.0001754728746622324,
"loss": 0.0001,
"step": 610
},
{
"epoch": 0.12647484992755123,
"grad_norm": 0.0068860347382724285,
"learning_rate": 0.0001754313032633548,
"loss": 0.0003,
"step": 611
},
{
"epoch": 0.12668184640861105,
"grad_norm": 0.001763600972481072,
"learning_rate": 0.00017538973186447725,
"loss": 0.0002,
"step": 612
},
{
"epoch": 0.12688884288967087,
"grad_norm": 0.0029042328242212534,
"learning_rate": 0.00017534816046559969,
"loss": 0.0019,
"step": 613
},
{
"epoch": 0.12709583937073068,
"grad_norm": 0.026835285127162933,
"learning_rate": 0.0001753065890667221,
"loss": 0.0006,
"step": 614
},
{
"epoch": 0.12730283585179053,
"grad_norm": 0.0025784820318222046,
"learning_rate": 0.00017526501766784453,
"loss": 0.0002,
"step": 615
},
{
"epoch": 0.12750983233285035,
"grad_norm": 0.000811999780125916,
"learning_rate": 0.00017522344626896697,
"loss": 0.0,
"step": 616
},
{
"epoch": 0.12771682881391017,
"grad_norm": 0.0023158304393291473,
"learning_rate": 0.00017518187487008938,
"loss": 0.0001,
"step": 617
},
{
"epoch": 0.12792382529496998,
"grad_norm": 0.00527742225676775,
"learning_rate": 0.00017514030347121182,
"loss": 0.0011,
"step": 618
},
{
"epoch": 0.1281308217760298,
"grad_norm": 0.004715193063020706,
"learning_rate": 0.00017509873207233425,
"loss": 0.0026,
"step": 619
},
{
"epoch": 0.12833781825708962,
"grad_norm": 0.001638007932342589,
"learning_rate": 0.00017505716067345666,
"loss": 0.0001,
"step": 620
},
{
"epoch": 0.12854481473814944,
"grad_norm": 0.0012813376961275935,
"learning_rate": 0.0001750155892745791,
"loss": 0.0001,
"step": 621
},
{
"epoch": 0.12875181121920928,
"grad_norm": 0.006484480109065771,
"learning_rate": 0.00017497401787570154,
"loss": 0.0004,
"step": 622
},
{
"epoch": 0.1289588077002691,
"grad_norm": 0.00035095165367238224,
"learning_rate": 0.00017493244647682398,
"loss": 0.0,
"step": 623
},
{
"epoch": 0.12916580418132892,
"grad_norm": 0.004927014000713825,
"learning_rate": 0.00017489087507794639,
"loss": 0.0012,
"step": 624
},
{
"epoch": 0.12937280066238874,
"grad_norm": 0.00287305167876184,
"learning_rate": 0.0001748493036790688,
"loss": 0.0001,
"step": 625
},
{
"epoch": 0.12957979714344856,
"grad_norm": 0.003079169662669301,
"learning_rate": 0.00017480773228019123,
"loss": 0.0001,
"step": 626
},
{
"epoch": 0.12978679362450837,
"grad_norm": 0.0018820518162101507,
"learning_rate": 0.00017476616088131367,
"loss": 0.0001,
"step": 627
},
{
"epoch": 0.12999379010556822,
"grad_norm": 0.004426770843565464,
"learning_rate": 0.00017472458948243608,
"loss": 0.0001,
"step": 628
},
{
"epoch": 0.13020078658662804,
"grad_norm": 0.008074757643043995,
"learning_rate": 0.00017468301808355852,
"loss": 0.0002,
"step": 629
},
{
"epoch": 0.13040778306768785,
"grad_norm": 0.004479815252125263,
"learning_rate": 0.00017464144668468095,
"loss": 0.0002,
"step": 630
},
{
"epoch": 0.13061477954874767,
"grad_norm": 0.0016544251702725887,
"learning_rate": 0.00017459987528580336,
"loss": 0.0,
"step": 631
},
{
"epoch": 0.1308217760298075,
"grad_norm": 0.0007902836659923196,
"learning_rate": 0.0001745583038869258,
"loss": 0.0,
"step": 632
},
{
"epoch": 0.1310287725108673,
"grad_norm": 0.011000900529325008,
"learning_rate": 0.00017451673248804824,
"loss": 0.0005,
"step": 633
},
{
"epoch": 0.13123576899192713,
"grad_norm": 0.00046783004654571414,
"learning_rate": 0.00017447516108917065,
"loss": 0.0,
"step": 634
},
{
"epoch": 0.13144276547298697,
"grad_norm": 0.003358067711815238,
"learning_rate": 0.00017443358969029308,
"loss": 0.0018,
"step": 635
},
{
"epoch": 0.1316497619540468,
"grad_norm": 0.0025496368762105703,
"learning_rate": 0.00017439201829141552,
"loss": 0.0015,
"step": 636
},
{
"epoch": 0.1318567584351066,
"grad_norm": 0.0016015061410143971,
"learning_rate": 0.00017435044689253793,
"loss": 0.0002,
"step": 637
},
{
"epoch": 0.13206375491616643,
"grad_norm": 0.0038993649650365114,
"learning_rate": 0.00017430887549366037,
"loss": 0.0001,
"step": 638
},
{
"epoch": 0.13227075139722624,
"grad_norm": 0.0033800469245761633,
"learning_rate": 0.0001742673040947828,
"loss": 0.0001,
"step": 639
},
{
"epoch": 0.13247774787828606,
"grad_norm": 0.0008187236380763352,
"learning_rate": 0.00017422573269590524,
"loss": 0.0,
"step": 640
},
{
"epoch": 0.13268474435934588,
"grad_norm": 0.005097914487123489,
"learning_rate": 0.00017418416129702765,
"loss": 0.0004,
"step": 641
},
{
"epoch": 0.13289174084040573,
"grad_norm": 0.0009978336747735739,
"learning_rate": 0.0001741425898981501,
"loss": 0.0,
"step": 642
},
{
"epoch": 0.13309873732146554,
"grad_norm": 0.004832749720662832,
"learning_rate": 0.00017410101849927253,
"loss": 0.0012,
"step": 643
},
{
"epoch": 0.13330573380252536,
"grad_norm": 0.0038694944232702255,
"learning_rate": 0.00017405944710039494,
"loss": 0.0001,
"step": 644
},
{
"epoch": 0.13351273028358518,
"grad_norm": 0.001419690903276205,
"learning_rate": 0.00017401787570151737,
"loss": 0.0,
"step": 645
},
{
"epoch": 0.133719726764645,
"grad_norm": 0.006202602293342352,
"learning_rate": 0.0001739763043026398,
"loss": 0.0002,
"step": 646
},
{
"epoch": 0.13392672324570482,
"grad_norm": 0.0008485604776069522,
"learning_rate": 0.00017393473290376222,
"loss": 0.0,
"step": 647
},
{
"epoch": 0.13413371972676463,
"grad_norm": 0.0050230189226567745,
"learning_rate": 0.00017389316150488463,
"loss": 0.0022,
"step": 648
},
{
"epoch": 0.13434071620782448,
"grad_norm": 0.002081549260765314,
"learning_rate": 0.00017385159010600707,
"loss": 0.0001,
"step": 649
},
{
"epoch": 0.1345477126888843,
"grad_norm": 0.001964141381904483,
"learning_rate": 0.0001738100187071295,
"loss": 0.0002,
"step": 650
},
{
"epoch": 0.13475470916994411,
"grad_norm": 0.0006888345233164728,
"learning_rate": 0.00017376844730825192,
"loss": 0.0,
"step": 651
},
{
"epoch": 0.13496170565100393,
"grad_norm": 0.002313450677320361,
"learning_rate": 0.00017372687590937435,
"loss": 0.0,
"step": 652
},
{
"epoch": 0.13516870213206375,
"grad_norm": 0.007078672293573618,
"learning_rate": 0.0001736853045104968,
"loss": 0.0006,
"step": 653
},
{
"epoch": 0.13537569861312357,
"grad_norm": 0.005166813265532255,
"learning_rate": 0.0001736437331116192,
"loss": 0.0011,
"step": 654
},
{
"epoch": 0.1355826950941834,
"grad_norm": 0.007185124326497316,
"learning_rate": 0.00017360216171274164,
"loss": 0.0001,
"step": 655
},
{
"epoch": 0.13578969157524323,
"grad_norm": 0.005528238136321306,
"learning_rate": 0.00017356059031386407,
"loss": 0.0001,
"step": 656
},
{
"epoch": 0.13599668805630305,
"grad_norm": 0.0077844299376010895,
"learning_rate": 0.0001735190189149865,
"loss": 0.0011,
"step": 657
},
{
"epoch": 0.13620368453736287,
"grad_norm": 0.00246329209767282,
"learning_rate": 0.00017347744751610892,
"loss": 0.0008,
"step": 658
},
{
"epoch": 0.13641068101842269,
"grad_norm": 0.005287639796733856,
"learning_rate": 0.00017343587611723136,
"loss": 0.0002,
"step": 659
},
{
"epoch": 0.1366176774994825,
"grad_norm": 0.006681959610432386,
"learning_rate": 0.0001733943047183538,
"loss": 0.0013,
"step": 660
},
{
"epoch": 0.13682467398054232,
"grad_norm": 0.0063599334098398685,
"learning_rate": 0.0001733527333194762,
"loss": 0.0003,
"step": 661
},
{
"epoch": 0.13703167046160214,
"grad_norm": 0.007015643175691366,
"learning_rate": 0.00017331116192059864,
"loss": 0.0003,
"step": 662
},
{
"epoch": 0.13723866694266199,
"grad_norm": 0.0003168722032569349,
"learning_rate": 0.00017326959052172108,
"loss": 0.0,
"step": 663
},
{
"epoch": 0.1374456634237218,
"grad_norm": 0.006562775932252407,
"learning_rate": 0.0001732280191228435,
"loss": 0.0008,
"step": 664
},
{
"epoch": 0.13765265990478162,
"grad_norm": 0.003267984837293625,
"learning_rate": 0.00017318644772396593,
"loss": 0.0001,
"step": 665
},
{
"epoch": 0.13785965638584144,
"grad_norm": 0.007215241901576519,
"learning_rate": 0.00017314487632508836,
"loss": 0.0021,
"step": 666
},
{
"epoch": 0.13806665286690126,
"grad_norm": 0.001962031237781048,
"learning_rate": 0.00017310330492621077,
"loss": 0.0,
"step": 667
},
{
"epoch": 0.13827364934796108,
"grad_norm": 0.007086516357958317,
"learning_rate": 0.0001730617335273332,
"loss": 0.0021,
"step": 668
},
{
"epoch": 0.1384806458290209,
"grad_norm": 0.0063016172498464584,
"learning_rate": 0.00017302016212845562,
"loss": 0.0001,
"step": 669
},
{
"epoch": 0.13868764231008074,
"grad_norm": 0.007975582964718342,
"learning_rate": 0.00017297859072957806,
"loss": 0.0002,
"step": 670
},
{
"epoch": 0.13889463879114056,
"grad_norm": 0.0030251971911638975,
"learning_rate": 0.00017293701933070047,
"loss": 0.0002,
"step": 671
},
{
"epoch": 0.13910163527220037,
"grad_norm": 0.00741973053663969,
"learning_rate": 0.0001728954479318229,
"loss": 0.0003,
"step": 672
},
{
"epoch": 0.1393086317532602,
"grad_norm": 0.002640543272718787,
"learning_rate": 0.00017285387653294534,
"loss": 0.0015,
"step": 673
},
{
"epoch": 0.13951562823432,
"grad_norm": 0.0004313603858463466,
"learning_rate": 0.00017281230513406778,
"loss": 0.0,
"step": 674
},
{
"epoch": 0.13972262471537983,
"grad_norm": 0.0020118863321840763,
"learning_rate": 0.0001727707337351902,
"loss": 0.0005,
"step": 675
},
{
"epoch": 0.13992962119643967,
"grad_norm": 0.003337120870128274,
"learning_rate": 0.00017272916233631262,
"loss": 0.0007,
"step": 676
},
{
"epoch": 0.1401366176774995,
"grad_norm": 0.014386707916855812,
"learning_rate": 0.00017268759093743506,
"loss": 0.0004,
"step": 677
},
{
"epoch": 0.1403436141585593,
"grad_norm": 0.006729326210916042,
"learning_rate": 0.00017264601953855747,
"loss": 0.001,
"step": 678
},
{
"epoch": 0.14055061063961913,
"grad_norm": 0.001671936479397118,
"learning_rate": 0.0001726044481396799,
"loss": 0.0002,
"step": 679
},
{
"epoch": 0.14075760712067895,
"grad_norm": 0.007516622077673674,
"learning_rate": 0.00017256287674080235,
"loss": 0.001,
"step": 680
},
{
"epoch": 0.14096460360173876,
"grad_norm": 0.0027280249632894993,
"learning_rate": 0.00017252130534192476,
"loss": 0.0006,
"step": 681
},
{
"epoch": 0.14117160008279858,
"grad_norm": 0.010556796565651894,
"learning_rate": 0.0001724797339430472,
"loss": 0.0003,
"step": 682
},
{
"epoch": 0.14137859656385843,
"grad_norm": 0.0027946115005761385,
"learning_rate": 0.00017243816254416963,
"loss": 0.0001,
"step": 683
},
{
"epoch": 0.14158559304491825,
"grad_norm": 0.00467882351949811,
"learning_rate": 0.00017239659114529204,
"loss": 0.0001,
"step": 684
},
{
"epoch": 0.14179258952597806,
"grad_norm": 0.004167881328612566,
"learning_rate": 0.00017235501974641448,
"loss": 0.0012,
"step": 685
},
{
"epoch": 0.14199958600703788,
"grad_norm": 0.0034762704744935036,
"learning_rate": 0.00017231344834753691,
"loss": 0.0007,
"step": 686
},
{
"epoch": 0.1422065824880977,
"grad_norm": 0.0005650786333717406,
"learning_rate": 0.00017227187694865935,
"loss": 0.0,
"step": 687
},
{
"epoch": 0.14241357896915752,
"grad_norm": 0.0043237158097326756,
"learning_rate": 0.00017223030554978176,
"loss": 0.0001,
"step": 688
},
{
"epoch": 0.14262057545021734,
"grad_norm": 0.0071853832341730595,
"learning_rate": 0.0001721887341509042,
"loss": 0.0004,
"step": 689
},
{
"epoch": 0.14282757193127718,
"grad_norm": 0.01868472993373871,
"learning_rate": 0.00017214716275202664,
"loss": 0.0009,
"step": 690
},
{
"epoch": 0.143034568412337,
"grad_norm": 0.001339295064099133,
"learning_rate": 0.00017210559135314902,
"loss": 0.0009,
"step": 691
},
{
"epoch": 0.14324156489339682,
"grad_norm": 0.00664726085960865,
"learning_rate": 0.00017206401995427146,
"loss": 0.0002,
"step": 692
},
{
"epoch": 0.14344856137445663,
"grad_norm": 0.006592089310288429,
"learning_rate": 0.0001720224485553939,
"loss": 0.0001,
"step": 693
},
{
"epoch": 0.14365555785551645,
"grad_norm": 0.0005503061693161726,
"learning_rate": 0.00017198087715651633,
"loss": 0.0,
"step": 694
},
{
"epoch": 0.14386255433657627,
"grad_norm": 0.003913522697985172,
"learning_rate": 0.00017193930575763874,
"loss": 0.0002,
"step": 695
},
{
"epoch": 0.1440695508176361,
"grad_norm": 0.004871245473623276,
"learning_rate": 0.00017189773435876118,
"loss": 0.0002,
"step": 696
},
{
"epoch": 0.14427654729869593,
"grad_norm": 0.007188999559730291,
"learning_rate": 0.0001718561629598836,
"loss": 0.0002,
"step": 697
},
{
"epoch": 0.14448354377975575,
"grad_norm": 0.003864140482619405,
"learning_rate": 0.00017181459156100602,
"loss": 0.0013,
"step": 698
},
{
"epoch": 0.14469054026081557,
"grad_norm": 0.005774588789790869,
"learning_rate": 0.00017177302016212846,
"loss": 0.0004,
"step": 699
},
{
"epoch": 0.1448975367418754,
"grad_norm": 0.002636535558849573,
"learning_rate": 0.0001717314487632509,
"loss": 0.0007,
"step": 700
},
{
"epoch": 0.1451045332229352,
"grad_norm": 0.03907289355993271,
"learning_rate": 0.0001716898773643733,
"loss": 0.0019,
"step": 701
},
{
"epoch": 0.14531152970399502,
"grad_norm": 0.005653630942106247,
"learning_rate": 0.00017164830596549574,
"loss": 0.0003,
"step": 702
},
{
"epoch": 0.14551852618505484,
"grad_norm": 0.003644258715212345,
"learning_rate": 0.00017160673456661818,
"loss": 0.0001,
"step": 703
},
{
"epoch": 0.1457255226661147,
"grad_norm": 0.0028953952714800835,
"learning_rate": 0.00017156516316774062,
"loss": 0.0001,
"step": 704
},
{
"epoch": 0.1459325191471745,
"grad_norm": 0.005685892421752214,
"learning_rate": 0.00017152359176886303,
"loss": 0.0002,
"step": 705
},
{
"epoch": 0.14613951562823432,
"grad_norm": 0.00946901086717844,
"learning_rate": 0.00017148202036998547,
"loss": 0.0008,
"step": 706
},
{
"epoch": 0.14634651210929414,
"grad_norm": 0.004027761984616518,
"learning_rate": 0.0001714404489711079,
"loss": 0.0001,
"step": 707
},
{
"epoch": 0.14655350859035396,
"grad_norm": 0.0014218458672985435,
"learning_rate": 0.0001713988775722303,
"loss": 0.0,
"step": 708
},
{
"epoch": 0.14676050507141378,
"grad_norm": 0.0058472915552556515,
"learning_rate": 0.00017135730617335275,
"loss": 0.0002,
"step": 709
},
{
"epoch": 0.1469675015524736,
"grad_norm": 0.004684192128479481,
"learning_rate": 0.0001713157347744752,
"loss": 0.0002,
"step": 710
},
{
"epoch": 0.14717449803353344,
"grad_norm": 0.002729298546910286,
"learning_rate": 0.0001712741633755976,
"loss": 0.0001,
"step": 711
},
{
"epoch": 0.14738149451459326,
"grad_norm": 0.003782545682042837,
"learning_rate": 0.00017123259197672,
"loss": 0.0022,
"step": 712
},
{
"epoch": 0.14758849099565308,
"grad_norm": 0.004307260736823082,
"learning_rate": 0.00017119102057784244,
"loss": 0.0009,
"step": 713
},
{
"epoch": 0.1477954874767129,
"grad_norm": 0.01339892577379942,
"learning_rate": 0.00017114944917896488,
"loss": 0.001,
"step": 714
},
{
"epoch": 0.1480024839577727,
"grad_norm": 0.0017793363658711314,
"learning_rate": 0.0001711078777800873,
"loss": 0.0,
"step": 715
},
{
"epoch": 0.14820948043883253,
"grad_norm": 0.0005680687027052045,
"learning_rate": 0.00017106630638120973,
"loss": 0.0,
"step": 716
},
{
"epoch": 0.14841647691989235,
"grad_norm": 0.0010823605116456747,
"learning_rate": 0.00017102473498233216,
"loss": 0.0,
"step": 717
},
{
"epoch": 0.1486234734009522,
"grad_norm": 0.006135303992778063,
"learning_rate": 0.00017098316358345457,
"loss": 0.0002,
"step": 718
},
{
"epoch": 0.148830469882012,
"grad_norm": 0.003215776290744543,
"learning_rate": 0.000170941592184577,
"loss": 0.0008,
"step": 719
},
{
"epoch": 0.14903746636307183,
"grad_norm": 0.00970076397061348,
"learning_rate": 0.00017090002078569945,
"loss": 0.0008,
"step": 720
},
{
"epoch": 0.14924446284413165,
"grad_norm": 0.0037311650812625885,
"learning_rate": 0.00017085844938682189,
"loss": 0.0007,
"step": 721
},
{
"epoch": 0.14945145932519147,
"grad_norm": 0.0035531132016330957,
"learning_rate": 0.0001708168779879443,
"loss": 0.0007,
"step": 722
},
{
"epoch": 0.14965845580625128,
"grad_norm": 0.0013675455702468753,
"learning_rate": 0.00017077530658906673,
"loss": 0.0002,
"step": 723
},
{
"epoch": 0.14986545228731113,
"grad_norm": 0.003667420009151101,
"learning_rate": 0.00017073373519018917,
"loss": 0.0001,
"step": 724
},
{
"epoch": 0.15007244876837095,
"grad_norm": 0.0006531656836159527,
"learning_rate": 0.00017069216379131158,
"loss": 0.0,
"step": 725
},
{
"epoch": 0.15027944524943077,
"grad_norm": 0.0029405278619378805,
"learning_rate": 0.00017065059239243402,
"loss": 0.0001,
"step": 726
},
{
"epoch": 0.15048644173049058,
"grad_norm": 0.0020144616719335318,
"learning_rate": 0.00017060902099355645,
"loss": 0.0001,
"step": 727
},
{
"epoch": 0.1506934382115504,
"grad_norm": 0.003123146714642644,
"learning_rate": 0.00017056744959467886,
"loss": 0.0007,
"step": 728
},
{
"epoch": 0.15090043469261022,
"grad_norm": 0.005841000005602837,
"learning_rate": 0.0001705258781958013,
"loss": 0.0001,
"step": 729
},
{
"epoch": 0.15110743117367004,
"grad_norm": 0.001898916088975966,
"learning_rate": 0.00017048430679692374,
"loss": 0.0,
"step": 730
},
{
"epoch": 0.15131442765472988,
"grad_norm": 0.0005505726439878345,
"learning_rate": 0.00017044273539804615,
"loss": 0.0,
"step": 731
},
{
"epoch": 0.1515214241357897,
"grad_norm": 0.022630905732512474,
"learning_rate": 0.00017040116399916859,
"loss": 0.0003,
"step": 732
},
{
"epoch": 0.15172842061684952,
"grad_norm": 0.0018513459945097566,
"learning_rate": 0.00017035959260029102,
"loss": 0.0,
"step": 733
},
{
"epoch": 0.15193541709790934,
"grad_norm": 0.006640856619924307,
"learning_rate": 0.00017031802120141343,
"loss": 0.0015,
"step": 734
},
{
"epoch": 0.15214241357896915,
"grad_norm": 0.010431594215333462,
"learning_rate": 0.00017027644980253584,
"loss": 0.0007,
"step": 735
},
{
"epoch": 0.15234941006002897,
"grad_norm": 0.0009595350711606443,
"learning_rate": 0.00017023487840365828,
"loss": 0.0001,
"step": 736
},
{
"epoch": 0.1525564065410888,
"grad_norm": 0.0019930503331124783,
"learning_rate": 0.00017019330700478072,
"loss": 0.0001,
"step": 737
},
{
"epoch": 0.15276340302214864,
"grad_norm": 0.0020235483534634113,
"learning_rate": 0.00017015173560590313,
"loss": 0.0012,
"step": 738
},
{
"epoch": 0.15297039950320845,
"grad_norm": 0.000323964050039649,
"learning_rate": 0.00017011016420702556,
"loss": 0.0,
"step": 739
},
{
"epoch": 0.15317739598426827,
"grad_norm": 0.004805979318916798,
"learning_rate": 0.000170068592808148,
"loss": 0.0016,
"step": 740
},
{
"epoch": 0.1533843924653281,
"grad_norm": 0.0007103011594153941,
"learning_rate": 0.00017002702140927044,
"loss": 0.0,
"step": 741
},
{
"epoch": 0.1535913889463879,
"grad_norm": 0.00901501253247261,
"learning_rate": 0.00016998545001039285,
"loss": 0.0001,
"step": 742
},
{
"epoch": 0.15379838542744773,
"grad_norm": 0.01626206934452057,
"learning_rate": 0.00016994387861151528,
"loss": 0.0001,
"step": 743
},
{
"epoch": 0.15400538190850754,
"grad_norm": 0.006600509863346815,
"learning_rate": 0.00016990230721263772,
"loss": 0.0022,
"step": 744
},
{
"epoch": 0.1542123783895674,
"grad_norm": 0.0031586415134370327,
"learning_rate": 0.00016986073581376013,
"loss": 0.0003,
"step": 745
},
{
"epoch": 0.1544193748706272,
"grad_norm": 0.00408458337187767,
"learning_rate": 0.00016981916441488257,
"loss": 0.0014,
"step": 746
},
{
"epoch": 0.15462637135168703,
"grad_norm": 0.006417447701096535,
"learning_rate": 0.000169777593016005,
"loss": 0.0013,
"step": 747
},
{
"epoch": 0.15483336783274684,
"grad_norm": 0.002676580101251602,
"learning_rate": 0.00016973602161712742,
"loss": 0.0009,
"step": 748
},
{
"epoch": 0.15504036431380666,
"grad_norm": 0.003124868730083108,
"learning_rate": 0.00016969445021824985,
"loss": 0.0001,
"step": 749
},
{
"epoch": 0.15524736079486648,
"grad_norm": 0.005617608781903982,
"learning_rate": 0.0001696528788193723,
"loss": 0.0009,
"step": 750
},
{
"epoch": 0.1554543572759263,
"grad_norm": 0.0029069185256958008,
"learning_rate": 0.00016961130742049473,
"loss": 0.0008,
"step": 751
},
{
"epoch": 0.15566135375698614,
"grad_norm": 0.008944474160671234,
"learning_rate": 0.00016956973602161714,
"loss": 0.0017,
"step": 752
},
{
"epoch": 0.15586835023804596,
"grad_norm": 0.004935794975608587,
"learning_rate": 0.00016952816462273957,
"loss": 0.0012,
"step": 753
},
{
"epoch": 0.15607534671910578,
"grad_norm": 0.0005579759599640965,
"learning_rate": 0.000169486593223862,
"loss": 0.0,
"step": 754
},
{
"epoch": 0.1562823432001656,
"grad_norm": 0.00902874581515789,
"learning_rate": 0.00016944502182498442,
"loss": 0.0004,
"step": 755
},
{
"epoch": 0.15648933968122541,
"grad_norm": 0.00498725613579154,
"learning_rate": 0.00016940345042610683,
"loss": 0.0001,
"step": 756
},
{
"epoch": 0.15669633616228523,
"grad_norm": 0.0004982489626854658,
"learning_rate": 0.00016936187902722927,
"loss": 0.0,
"step": 757
},
{
"epoch": 0.15690333264334505,
"grad_norm": 0.0011680921306833625,
"learning_rate": 0.0001693203076283517,
"loss": 0.0,
"step": 758
},
{
"epoch": 0.1571103291244049,
"grad_norm": 0.0013553223107010126,
"learning_rate": 0.00016927873622947412,
"loss": 0.0,
"step": 759
},
{
"epoch": 0.15731732560546471,
"grad_norm": 0.00549361202865839,
"learning_rate": 0.00016923716483059655,
"loss": 0.0016,
"step": 760
},
{
"epoch": 0.15752432208652453,
"grad_norm": 0.004852932877838612,
"learning_rate": 0.000169195593431719,
"loss": 0.0002,
"step": 761
},
{
"epoch": 0.15773131856758435,
"grad_norm": 0.0046032629907131195,
"learning_rate": 0.0001691540220328414,
"loss": 0.002,
"step": 762
},
{
"epoch": 0.15793831504864417,
"grad_norm": 0.009385612793266773,
"learning_rate": 0.00016911245063396384,
"loss": 0.0003,
"step": 763
},
{
"epoch": 0.15814531152970399,
"grad_norm": 0.0024257150944322348,
"learning_rate": 0.00016907087923508627,
"loss": 0.0009,
"step": 764
},
{
"epoch": 0.1583523080107638,
"grad_norm": 0.002726235194131732,
"learning_rate": 0.00016902930783620868,
"loss": 0.0012,
"step": 765
},
{
"epoch": 0.15855930449182365,
"grad_norm": 0.006497920490801334,
"learning_rate": 0.00016898773643733112,
"loss": 0.0005,
"step": 766
},
{
"epoch": 0.15876630097288347,
"grad_norm": 0.012873928062617779,
"learning_rate": 0.00016894616503845356,
"loss": 0.0021,
"step": 767
},
{
"epoch": 0.15897329745394329,
"grad_norm": 0.009931253269314766,
"learning_rate": 0.000168904593639576,
"loss": 0.0004,
"step": 768
},
{
"epoch": 0.1591802939350031,
"grad_norm": 0.0012783849379047751,
"learning_rate": 0.0001688630222406984,
"loss": 0.0,
"step": 769
},
{
"epoch": 0.15938729041606292,
"grad_norm": 0.0025215751957148314,
"learning_rate": 0.00016882145084182084,
"loss": 0.0001,
"step": 770
},
{
"epoch": 0.15959428689712274,
"grad_norm": 0.00574857834726572,
"learning_rate": 0.00016877987944294328,
"loss": 0.0002,
"step": 771
},
{
"epoch": 0.15980128337818258,
"grad_norm": 0.0008691879920661449,
"learning_rate": 0.0001687383080440657,
"loss": 0.0,
"step": 772
},
{
"epoch": 0.1600082798592424,
"grad_norm": 0.008225478231906891,
"learning_rate": 0.00016869673664518813,
"loss": 0.0002,
"step": 773
},
{
"epoch": 0.16021527634030222,
"grad_norm": 0.003890304360538721,
"learning_rate": 0.00016865516524631056,
"loss": 0.0014,
"step": 774
},
{
"epoch": 0.16042227282136204,
"grad_norm": 0.0011641031596809626,
"learning_rate": 0.00016861359384743297,
"loss": 0.0001,
"step": 775
},
{
"epoch": 0.16062926930242186,
"grad_norm": 0.008769871667027473,
"learning_rate": 0.0001685720224485554,
"loss": 0.0011,
"step": 776
},
{
"epoch": 0.16083626578348167,
"grad_norm": 0.005050954408943653,
"learning_rate": 0.00016853045104967785,
"loss": 0.0002,
"step": 777
},
{
"epoch": 0.1610432622645415,
"grad_norm": 0.002180990530177951,
"learning_rate": 0.00016848887965080026,
"loss": 0.0008,
"step": 778
},
{
"epoch": 0.16125025874560134,
"grad_norm": 0.0015876460820436478,
"learning_rate": 0.00016844730825192267,
"loss": 0.0,
"step": 779
},
{
"epoch": 0.16145725522666116,
"grad_norm": 0.004357179626822472,
"learning_rate": 0.0001684057368530451,
"loss": 0.0003,
"step": 780
},
{
"epoch": 0.16166425170772097,
"grad_norm": 0.0034056720323860645,
"learning_rate": 0.00016836416545416754,
"loss": 0.0012,
"step": 781
},
{
"epoch": 0.1618712481887808,
"grad_norm": 0.005545208230614662,
"learning_rate": 0.00016832259405528995,
"loss": 0.0001,
"step": 782
},
{
"epoch": 0.1620782446698406,
"grad_norm": 0.0002129770437022671,
"learning_rate": 0.0001682810226564124,
"loss": 0.0,
"step": 783
},
{
"epoch": 0.16228524115090043,
"grad_norm": 0.0036753590684384108,
"learning_rate": 0.00016823945125753482,
"loss": 0.0001,
"step": 784
},
{
"epoch": 0.16249223763196025,
"grad_norm": 0.0018491502851247787,
"learning_rate": 0.00016819787985865723,
"loss": 0.0008,
"step": 785
},
{
"epoch": 0.1626992341130201,
"grad_norm": 0.0006519712042063475,
"learning_rate": 0.00016815630845977967,
"loss": 0.0,
"step": 786
},
{
"epoch": 0.1629062305940799,
"grad_norm": 0.011139947921037674,
"learning_rate": 0.0001681147370609021,
"loss": 0.0011,
"step": 787
},
{
"epoch": 0.16311322707513973,
"grad_norm": 0.0020866713020950556,
"learning_rate": 0.00016807316566202455,
"loss": 0.0002,
"step": 788
},
{
"epoch": 0.16332022355619955,
"grad_norm": 0.0034007905051112175,
"learning_rate": 0.00016803159426314696,
"loss": 0.0008,
"step": 789
},
{
"epoch": 0.16352722003725936,
"grad_norm": 0.0017938032979145646,
"learning_rate": 0.0001679900228642694,
"loss": 0.0009,
"step": 790
},
{
"epoch": 0.16373421651831918,
"grad_norm": 0.005385685246437788,
"learning_rate": 0.00016794845146539183,
"loss": 0.001,
"step": 791
},
{
"epoch": 0.163941212999379,
"grad_norm": 0.010079730302095413,
"learning_rate": 0.00016790688006651424,
"loss": 0.0004,
"step": 792
},
{
"epoch": 0.16414820948043884,
"grad_norm": 0.005826961249113083,
"learning_rate": 0.00016786530866763668,
"loss": 0.0001,
"step": 793
},
{
"epoch": 0.16435520596149866,
"grad_norm": 0.002885566558688879,
"learning_rate": 0.00016782373726875911,
"loss": 0.0001,
"step": 794
},
{
"epoch": 0.16456220244255848,
"grad_norm": 0.004031067714095116,
"learning_rate": 0.00016778216586988152,
"loss": 0.0002,
"step": 795
},
{
"epoch": 0.1647691989236183,
"grad_norm": 0.0019721402786672115,
"learning_rate": 0.00016774059447100396,
"loss": 0.0001,
"step": 796
},
{
"epoch": 0.16497619540467812,
"grad_norm": 0.002213244093582034,
"learning_rate": 0.0001676990230721264,
"loss": 0.0006,
"step": 797
},
{
"epoch": 0.16518319188573793,
"grad_norm": 0.001942839939147234,
"learning_rate": 0.00016765745167324884,
"loss": 0.0,
"step": 798
},
{
"epoch": 0.16539018836679775,
"grad_norm": 0.003173516597598791,
"learning_rate": 0.00016761588027437125,
"loss": 0.0001,
"step": 799
},
{
"epoch": 0.1655971848478576,
"grad_norm": 0.004877821542322636,
"learning_rate": 0.00016757430887549366,
"loss": 0.0001,
"step": 800
},
{
"epoch": 0.16580418132891742,
"grad_norm": 0.006676991004496813,
"learning_rate": 0.0001675327374766161,
"loss": 0.0002,
"step": 801
},
{
"epoch": 0.16601117780997723,
"grad_norm": 0.0022598986979573965,
"learning_rate": 0.0001674911660777385,
"loss": 0.0003,
"step": 802
},
{
"epoch": 0.16621817429103705,
"grad_norm": 0.00012318776862230152,
"learning_rate": 0.00016744959467886094,
"loss": 0.0,
"step": 803
},
{
"epoch": 0.16642517077209687,
"grad_norm": 0.006045771297067404,
"learning_rate": 0.00016740802327998338,
"loss": 0.0003,
"step": 804
},
{
"epoch": 0.1666321672531567,
"grad_norm": 0.004370058421045542,
"learning_rate": 0.0001673664518811058,
"loss": 0.0001,
"step": 805
},
{
"epoch": 0.1668391637342165,
"grad_norm": 0.005490643437951803,
"learning_rate": 0.00016732488048222822,
"loss": 0.0022,
"step": 806
},
{
"epoch": 0.16704616021527635,
"grad_norm": 0.007493430282920599,
"learning_rate": 0.00016728330908335066,
"loss": 0.001,
"step": 807
},
{
"epoch": 0.16725315669633617,
"grad_norm": 0.0006159085314720869,
"learning_rate": 0.0001672417376844731,
"loss": 0.0,
"step": 808
},
{
"epoch": 0.167460153177396,
"grad_norm": 0.002211883431300521,
"learning_rate": 0.0001672001662855955,
"loss": 0.0,
"step": 809
},
{
"epoch": 0.1676671496584558,
"grad_norm": 0.0028680090326815844,
"learning_rate": 0.00016715859488671794,
"loss": 0.0011,
"step": 810
},
{
"epoch": 0.16787414613951562,
"grad_norm": 0.004992680158466101,
"learning_rate": 0.00016711702348784038,
"loss": 0.0002,
"step": 811
},
{
"epoch": 0.16808114262057544,
"grad_norm": 0.0024819490499794483,
"learning_rate": 0.0001670754520889628,
"loss": 0.0001,
"step": 812
},
{
"epoch": 0.16828813910163526,
"grad_norm": 0.001662694732658565,
"learning_rate": 0.00016703388069008523,
"loss": 0.0001,
"step": 813
},
{
"epoch": 0.1684951355826951,
"grad_norm": 0.0027136337012052536,
"learning_rate": 0.00016699230929120767,
"loss": 0.0015,
"step": 814
},
{
"epoch": 0.16870213206375492,
"grad_norm": 0.0055983890779316425,
"learning_rate": 0.00016695073789233008,
"loss": 0.0006,
"step": 815
},
{
"epoch": 0.16890912854481474,
"grad_norm": 0.0005543065490201116,
"learning_rate": 0.0001669091664934525,
"loss": 0.0,
"step": 816
},
{
"epoch": 0.16911612502587456,
"grad_norm": 0.006743449252098799,
"learning_rate": 0.00016686759509457495,
"loss": 0.0006,
"step": 817
},
{
"epoch": 0.16932312150693438,
"grad_norm": 0.005361751653254032,
"learning_rate": 0.0001668260236956974,
"loss": 0.0002,
"step": 818
},
{
"epoch": 0.1695301179879942,
"grad_norm": 0.015542850829660892,
"learning_rate": 0.0001667844522968198,
"loss": 0.0001,
"step": 819
},
{
"epoch": 0.169737114469054,
"grad_norm": 0.006788911763578653,
"learning_rate": 0.00016674288089794223,
"loss": 0.0004,
"step": 820
},
{
"epoch": 0.16994411095011386,
"grad_norm": 0.006434622220695019,
"learning_rate": 0.00016670130949906467,
"loss": 0.0004,
"step": 821
},
{
"epoch": 0.17015110743117368,
"grad_norm": 0.0024506154004484415,
"learning_rate": 0.00016665973810018708,
"loss": 0.0007,
"step": 822
},
{
"epoch": 0.1703581039122335,
"grad_norm": 0.000382046215236187,
"learning_rate": 0.0001666181667013095,
"loss": 0.0,
"step": 823
},
{
"epoch": 0.1705651003932933,
"grad_norm": 0.00432636309415102,
"learning_rate": 0.00016657659530243193,
"loss": 0.0009,
"step": 824
},
{
"epoch": 0.17077209687435313,
"grad_norm": 0.005686972755938768,
"learning_rate": 0.00016653502390355436,
"loss": 0.0002,
"step": 825
},
{
"epoch": 0.17097909335541295,
"grad_norm": 0.005743528716266155,
"learning_rate": 0.00016649345250467677,
"loss": 0.0003,
"step": 826
},
{
"epoch": 0.1711860898364728,
"grad_norm": 0.002116352552548051,
"learning_rate": 0.0001664518811057992,
"loss": 0.0003,
"step": 827
},
{
"epoch": 0.1713930863175326,
"grad_norm": 0.002352718496695161,
"learning_rate": 0.00016641030970692165,
"loss": 0.0,
"step": 828
},
{
"epoch": 0.17160008279859243,
"grad_norm": 0.0044693113304674625,
"learning_rate": 0.00016636873830804406,
"loss": 0.0004,
"step": 829
},
{
"epoch": 0.17180707927965225,
"grad_norm": 0.0005167116178199649,
"learning_rate": 0.0001663271669091665,
"loss": 0.0,
"step": 830
},
{
"epoch": 0.17201407576071207,
"grad_norm": 0.005162122659385204,
"learning_rate": 0.00016628559551028893,
"loss": 0.0002,
"step": 831
},
{
"epoch": 0.17222107224177188,
"grad_norm": 0.00015954635455273092,
"learning_rate": 0.00016624402411141134,
"loss": 0.0,
"step": 832
},
{
"epoch": 0.1724280687228317,
"grad_norm": 0.0030487151816487312,
"learning_rate": 0.00016620245271253378,
"loss": 0.0001,
"step": 833
},
{
"epoch": 0.17263506520389155,
"grad_norm": 0.002151534892618656,
"learning_rate": 0.00016616088131365622,
"loss": 0.0002,
"step": 834
},
{
"epoch": 0.17284206168495136,
"grad_norm": 0.0044494629837572575,
"learning_rate": 0.00016611930991477865,
"loss": 0.0005,
"step": 835
},
{
"epoch": 0.17304905816601118,
"grad_norm": 0.00033838755916804075,
"learning_rate": 0.00016607773851590106,
"loss": 0.0,
"step": 836
},
{
"epoch": 0.173256054647071,
"grad_norm": 0.0005302856443449855,
"learning_rate": 0.0001660361671170235,
"loss": 0.0,
"step": 837
},
{
"epoch": 0.17346305112813082,
"grad_norm": 0.0013208640739321709,
"learning_rate": 0.00016599459571814594,
"loss": 0.0,
"step": 838
},
{
"epoch": 0.17367004760919064,
"grad_norm": 0.001052051316946745,
"learning_rate": 0.00016595302431926835,
"loss": 0.0,
"step": 839
},
{
"epoch": 0.17387704409025045,
"grad_norm": 0.018445929512381554,
"learning_rate": 0.00016591145292039079,
"loss": 0.0004,
"step": 840
},
{
"epoch": 0.1740840405713103,
"grad_norm": 0.0025256802327930927,
"learning_rate": 0.00016586988152151322,
"loss": 0.0001,
"step": 841
},
{
"epoch": 0.17429103705237012,
"grad_norm": 0.0014724883949384093,
"learning_rate": 0.00016582831012263563,
"loss": 0.0,
"step": 842
},
{
"epoch": 0.17449803353342994,
"grad_norm": 0.003576815826818347,
"learning_rate": 0.00016578673872375807,
"loss": 0.0001,
"step": 843
},
{
"epoch": 0.17470503001448975,
"grad_norm": 0.0006163385114632547,
"learning_rate": 0.00016574516732488048,
"loss": 0.0,
"step": 844
},
{
"epoch": 0.17491202649554957,
"grad_norm": 0.0011656074784696102,
"learning_rate": 0.00016570359592600292,
"loss": 0.0001,
"step": 845
},
{
"epoch": 0.1751190229766094,
"grad_norm": 0.0018338944064453244,
"learning_rate": 0.00016566202452712533,
"loss": 0.0001,
"step": 846
},
{
"epoch": 0.1753260194576692,
"grad_norm": 0.005035779904574156,
"learning_rate": 0.00016562045312824776,
"loss": 0.0014,
"step": 847
},
{
"epoch": 0.17553301593872905,
"grad_norm": 0.006770180072635412,
"learning_rate": 0.0001655788817293702,
"loss": 0.0012,
"step": 848
},
{
"epoch": 0.17574001241978887,
"grad_norm": 0.0003650276339612901,
"learning_rate": 0.0001655373103304926,
"loss": 0.0,
"step": 849
},
{
"epoch": 0.1759470089008487,
"grad_norm": 0.00023851868172641844,
"learning_rate": 0.00016549573893161505,
"loss": 0.0,
"step": 850
},
{
"epoch": 0.1761540053819085,
"grad_norm": 0.014695384539663792,
"learning_rate": 0.00016545416753273748,
"loss": 0.0011,
"step": 851
},
{
"epoch": 0.17636100186296833,
"grad_norm": 0.00036404369166120887,
"learning_rate": 0.00016541259613385992,
"loss": 0.0,
"step": 852
},
{
"epoch": 0.17656799834402814,
"grad_norm": 0.002682497026398778,
"learning_rate": 0.00016537102473498233,
"loss": 0.0012,
"step": 853
},
{
"epoch": 0.17677499482508796,
"grad_norm": 0.007028127089142799,
"learning_rate": 0.00016532945333610477,
"loss": 0.0002,
"step": 854
},
{
"epoch": 0.1769819913061478,
"grad_norm": 0.0012324461713433266,
"learning_rate": 0.0001652878819372272,
"loss": 0.0,
"step": 855
},
{
"epoch": 0.17718898778720762,
"grad_norm": 0.00292210397310555,
"learning_rate": 0.00016524631053834962,
"loss": 0.0,
"step": 856
},
{
"epoch": 0.17739598426826744,
"grad_norm": 0.0014698312152177095,
"learning_rate": 0.00016520473913947205,
"loss": 0.0002,
"step": 857
},
{
"epoch": 0.17760298074932726,
"grad_norm": 0.0022247894667088985,
"learning_rate": 0.0001651631677405945,
"loss": 0.0,
"step": 858
},
{
"epoch": 0.17780997723038708,
"grad_norm": 0.0006738615338690579,
"learning_rate": 0.0001651215963417169,
"loss": 0.0,
"step": 859
},
{
"epoch": 0.1780169737114469,
"grad_norm": 0.004056425765156746,
"learning_rate": 0.00016508002494283934,
"loss": 0.001,
"step": 860
},
{
"epoch": 0.17822397019250671,
"grad_norm": 0.006607827264815569,
"learning_rate": 0.00016503845354396177,
"loss": 0.0003,
"step": 861
},
{
"epoch": 0.17843096667356656,
"grad_norm": 0.007498994003981352,
"learning_rate": 0.00016499688214508418,
"loss": 0.0009,
"step": 862
},
{
"epoch": 0.17863796315462638,
"grad_norm": 0.002715140348300338,
"learning_rate": 0.00016495531074620662,
"loss": 0.0001,
"step": 863
},
{
"epoch": 0.1788449596356862,
"grad_norm": 0.006559406872838736,
"learning_rate": 0.00016491373934732906,
"loss": 0.0012,
"step": 864
},
{
"epoch": 0.179051956116746,
"grad_norm": 0.005900564603507519,
"learning_rate": 0.0001648721679484515,
"loss": 0.0001,
"step": 865
},
{
"epoch": 0.17925895259780583,
"grad_norm": 0.0014680642634630203,
"learning_rate": 0.00016483059654957388,
"loss": 0.0004,
"step": 866
},
{
"epoch": 0.17946594907886565,
"grad_norm": 0.0017862527165561914,
"learning_rate": 0.00016478902515069632,
"loss": 0.0001,
"step": 867
},
{
"epoch": 0.17967294555992547,
"grad_norm": 0.0010660128900781274,
"learning_rate": 0.00016474745375181875,
"loss": 0.0001,
"step": 868
},
{
"epoch": 0.1798799420409853,
"grad_norm": 0.010508016683161259,
"learning_rate": 0.0001647058823529412,
"loss": 0.0004,
"step": 869
},
{
"epoch": 0.18008693852204513,
"grad_norm": 0.0026938568335026503,
"learning_rate": 0.0001646643109540636,
"loss": 0.0001,
"step": 870
},
{
"epoch": 0.18029393500310495,
"grad_norm": 0.0015470877988263965,
"learning_rate": 0.00016462273955518604,
"loss": 0.0,
"step": 871
},
{
"epoch": 0.18050093148416477,
"grad_norm": 0.003435211256146431,
"learning_rate": 0.00016458116815630847,
"loss": 0.0001,
"step": 872
},
{
"epoch": 0.18070792796522459,
"grad_norm": 0.010342281311750412,
"learning_rate": 0.00016453959675743088,
"loss": 0.0007,
"step": 873
},
{
"epoch": 0.1809149244462844,
"grad_norm": 0.0007751841330900788,
"learning_rate": 0.00016449802535855332,
"loss": 0.0,
"step": 874
},
{
"epoch": 0.18112192092734425,
"grad_norm": 0.0003991715202573687,
"learning_rate": 0.00016445645395967576,
"loss": 0.0,
"step": 875
},
{
"epoch": 0.18132891740840407,
"grad_norm": 0.004742010496556759,
"learning_rate": 0.00016441488256079817,
"loss": 0.0001,
"step": 876
},
{
"epoch": 0.18153591388946388,
"grad_norm": 0.0010139705846086144,
"learning_rate": 0.0001643733111619206,
"loss": 0.0,
"step": 877
},
{
"epoch": 0.1817429103705237,
"grad_norm": 0.00697368336841464,
"learning_rate": 0.00016433173976304304,
"loss": 0.0003,
"step": 878
},
{
"epoch": 0.18194990685158352,
"grad_norm": 0.0056029148399829865,
"learning_rate": 0.00016429016836416545,
"loss": 0.0003,
"step": 879
},
{
"epoch": 0.18215690333264334,
"grad_norm": 0.0031287583988159895,
"learning_rate": 0.0001642485969652879,
"loss": 0.0001,
"step": 880
},
{
"epoch": 0.18236389981370316,
"grad_norm": 0.0005836491473019123,
"learning_rate": 0.00016420702556641033,
"loss": 0.0,
"step": 881
},
{
"epoch": 0.182570896294763,
"grad_norm": 0.006221551448106766,
"learning_rate": 0.00016416545416753276,
"loss": 0.0014,
"step": 882
},
{
"epoch": 0.18277789277582282,
"grad_norm": 0.00045936627429910004,
"learning_rate": 0.00016412388276865517,
"loss": 0.0,
"step": 883
},
{
"epoch": 0.18298488925688264,
"grad_norm": 0.0006924067274667323,
"learning_rate": 0.0001640823113697776,
"loss": 0.0,
"step": 884
},
{
"epoch": 0.18319188573794246,
"grad_norm": 0.010869395919144154,
"learning_rate": 0.00016404073997090005,
"loss": 0.0013,
"step": 885
},
{
"epoch": 0.18339888221900227,
"grad_norm": 0.00480787456035614,
"learning_rate": 0.00016399916857202246,
"loss": 0.0009,
"step": 886
},
{
"epoch": 0.1836058787000621,
"grad_norm": 0.0004685772000811994,
"learning_rate": 0.0001639575971731449,
"loss": 0.0,
"step": 887
},
{
"epoch": 0.1838128751811219,
"grad_norm": 0.00303410436026752,
"learning_rate": 0.0001639160257742673,
"loss": 0.0007,
"step": 888
},
{
"epoch": 0.18401987166218176,
"grad_norm": 0.001141547691076994,
"learning_rate": 0.00016387445437538974,
"loss": 0.0001,
"step": 889
},
{
"epoch": 0.18422686814324157,
"grad_norm": 0.007433968596160412,
"learning_rate": 0.00016383288297651215,
"loss": 0.0001,
"step": 890
},
{
"epoch": 0.1844338646243014,
"grad_norm": 0.003386344527825713,
"learning_rate": 0.0001637913115776346,
"loss": 0.0003,
"step": 891
},
{
"epoch": 0.1846408611053612,
"grad_norm": 0.0011372484732419252,
"learning_rate": 0.00016374974017875702,
"loss": 0.0,
"step": 892
},
{
"epoch": 0.18484785758642103,
"grad_norm": 0.0039020997937768698,
"learning_rate": 0.00016370816877987943,
"loss": 0.0001,
"step": 893
},
{
"epoch": 0.18505485406748085,
"grad_norm": 0.003088288474828005,
"learning_rate": 0.00016366659738100187,
"loss": 0.0009,
"step": 894
},
{
"epoch": 0.18526185054854066,
"grad_norm": 0.001126794726587832,
"learning_rate": 0.0001636250259821243,
"loss": 0.0001,
"step": 895
},
{
"epoch": 0.1854688470296005,
"grad_norm": 0.007449139375239611,
"learning_rate": 0.00016358345458324672,
"loss": 0.0004,
"step": 896
},
{
"epoch": 0.18567584351066033,
"grad_norm": 0.005704225040972233,
"learning_rate": 0.00016354188318436916,
"loss": 0.0016,
"step": 897
},
{
"epoch": 0.18588283999172014,
"grad_norm": 0.004983640741556883,
"learning_rate": 0.0001635003117854916,
"loss": 0.0016,
"step": 898
},
{
"epoch": 0.18608983647277996,
"grad_norm": 0.00034120268537662923,
"learning_rate": 0.00016345874038661403,
"loss": 0.0,
"step": 899
},
{
"epoch": 0.18629683295383978,
"grad_norm": 0.007043101824820042,
"learning_rate": 0.00016341716898773644,
"loss": 0.001,
"step": 900
},
{
"epoch": 0.1865038294348996,
"grad_norm": 0.004050271585583687,
"learning_rate": 0.00016337559758885888,
"loss": 0.0001,
"step": 901
},
{
"epoch": 0.18671082591595942,
"grad_norm": 0.001882696757093072,
"learning_rate": 0.00016333402618998131,
"loss": 0.0005,
"step": 902
},
{
"epoch": 0.18691782239701926,
"grad_norm": 0.002479350659996271,
"learning_rate": 0.00016329245479110372,
"loss": 0.0003,
"step": 903
},
{
"epoch": 0.18712481887807908,
"grad_norm": 0.00246567465364933,
"learning_rate": 0.00016325088339222616,
"loss": 0.0001,
"step": 904
},
{
"epoch": 0.1873318153591389,
"grad_norm": 0.0021426973398774862,
"learning_rate": 0.0001632093119933486,
"loss": 0.0006,
"step": 905
},
{
"epoch": 0.18753881184019872,
"grad_norm": 0.004363594576716423,
"learning_rate": 0.000163167740594471,
"loss": 0.0002,
"step": 906
},
{
"epoch": 0.18774580832125853,
"grad_norm": 0.004984852857887745,
"learning_rate": 0.00016312616919559345,
"loss": 0.0004,
"step": 907
},
{
"epoch": 0.18795280480231835,
"grad_norm": 0.004489907994866371,
"learning_rate": 0.00016308459779671588,
"loss": 0.0001,
"step": 908
},
{
"epoch": 0.18815980128337817,
"grad_norm": 0.0013233786448836327,
"learning_rate": 0.0001630430263978383,
"loss": 0.0,
"step": 909
},
{
"epoch": 0.18836679776443802,
"grad_norm": 0.00894436426460743,
"learning_rate": 0.0001630014549989607,
"loss": 0.0016,
"step": 910
},
{
"epoch": 0.18857379424549783,
"grad_norm": 0.001729694427922368,
"learning_rate": 0.00016295988360008314,
"loss": 0.001,
"step": 911
},
{
"epoch": 0.18878079072655765,
"grad_norm": 0.005005873739719391,
"learning_rate": 0.00016291831220120558,
"loss": 0.0005,
"step": 912
},
{
"epoch": 0.18898778720761747,
"grad_norm": 0.0007573101902380586,
"learning_rate": 0.00016287674080232799,
"loss": 0.0,
"step": 913
},
{
"epoch": 0.1891947836886773,
"grad_norm": 0.005315006244927645,
"learning_rate": 0.00016283516940345042,
"loss": 0.0001,
"step": 914
},
{
"epoch": 0.1894017801697371,
"grad_norm": 0.001140634878538549,
"learning_rate": 0.00016279359800457286,
"loss": 0.0001,
"step": 915
},
{
"epoch": 0.18960877665079692,
"grad_norm": 0.003881396260112524,
"learning_rate": 0.0001627520266056953,
"loss": 0.0001,
"step": 916
},
{
"epoch": 0.18981577313185677,
"grad_norm": 0.001353550935164094,
"learning_rate": 0.0001627104552068177,
"loss": 0.0,
"step": 917
},
{
"epoch": 0.1900227696129166,
"grad_norm": 0.001415180740877986,
"learning_rate": 0.00016266888380794014,
"loss": 0.0,
"step": 918
},
{
"epoch": 0.1902297660939764,
"grad_norm": 0.006398684810847044,
"learning_rate": 0.00016262731240906258,
"loss": 0.0006,
"step": 919
},
{
"epoch": 0.19043676257503622,
"grad_norm": 0.005204816348850727,
"learning_rate": 0.000162585741010185,
"loss": 0.0016,
"step": 920
},
{
"epoch": 0.19064375905609604,
"grad_norm": 0.0015194268198683858,
"learning_rate": 0.00016254416961130743,
"loss": 0.0007,
"step": 921
},
{
"epoch": 0.19085075553715586,
"grad_norm": 0.002916971454396844,
"learning_rate": 0.00016250259821242987,
"loss": 0.0015,
"step": 922
},
{
"epoch": 0.1910577520182157,
"grad_norm": 0.00017840563668869436,
"learning_rate": 0.00016246102681355228,
"loss": 0.0,
"step": 923
},
{
"epoch": 0.19126474849927552,
"grad_norm": 0.0017515165964141488,
"learning_rate": 0.0001624194554146747,
"loss": 0.0005,
"step": 924
},
{
"epoch": 0.19147174498033534,
"grad_norm": 0.0011207156348973513,
"learning_rate": 0.00016237788401579715,
"loss": 0.0,
"step": 925
},
{
"epoch": 0.19167874146139516,
"grad_norm": 0.00647772429510951,
"learning_rate": 0.00016233631261691956,
"loss": 0.0004,
"step": 926
},
{
"epoch": 0.19188573794245498,
"grad_norm": 0.0009239514474757016,
"learning_rate": 0.000162294741218042,
"loss": 0.0,
"step": 927
},
{
"epoch": 0.1920927344235148,
"grad_norm": 0.0009718858054839075,
"learning_rate": 0.00016225316981916443,
"loss": 0.0,
"step": 928
},
{
"epoch": 0.1922997309045746,
"grad_norm": 0.0009835285600274801,
"learning_rate": 0.00016221159842028687,
"loss": 0.0001,
"step": 929
},
{
"epoch": 0.19250672738563446,
"grad_norm": 0.003986849449574947,
"learning_rate": 0.00016217002702140928,
"loss": 0.0005,
"step": 930
},
{
"epoch": 0.19271372386669428,
"grad_norm": 0.0055690668523311615,
"learning_rate": 0.00016212845562253172,
"loss": 0.0002,
"step": 931
},
{
"epoch": 0.1929207203477541,
"grad_norm": 0.006283191032707691,
"learning_rate": 0.00016208688422365413,
"loss": 0.0003,
"step": 932
},
{
"epoch": 0.1931277168288139,
"grad_norm": 0.00035167241003364325,
"learning_rate": 0.00016204531282477656,
"loss": 0.0,
"step": 933
},
{
"epoch": 0.19333471330987373,
"grad_norm": 0.001550202607177198,
"learning_rate": 0.00016200374142589897,
"loss": 0.0001,
"step": 934
},
{
"epoch": 0.19354170979093355,
"grad_norm": 0.0009650330757722259,
"learning_rate": 0.0001619621700270214,
"loss": 0.0,
"step": 935
},
{
"epoch": 0.19374870627199337,
"grad_norm": 0.006459403783082962,
"learning_rate": 0.00016192059862814385,
"loss": 0.0009,
"step": 936
},
{
"epoch": 0.1939557027530532,
"grad_norm": 0.0006884423783048987,
"learning_rate": 0.00016187902722926626,
"loss": 0.0,
"step": 937
},
{
"epoch": 0.19416269923411303,
"grad_norm": 0.0034009867813438177,
"learning_rate": 0.0001618374558303887,
"loss": 0.0001,
"step": 938
},
{
"epoch": 0.19436969571517285,
"grad_norm": 0.001749175600707531,
"learning_rate": 0.00016179588443151113,
"loss": 0.0012,
"step": 939
},
{
"epoch": 0.19457669219623266,
"grad_norm": 0.005409194156527519,
"learning_rate": 0.00016175431303263354,
"loss": 0.0002,
"step": 940
},
{
"epoch": 0.19478368867729248,
"grad_norm": 0.0033904362935572863,
"learning_rate": 0.00016171274163375598,
"loss": 0.0001,
"step": 941
},
{
"epoch": 0.1949906851583523,
"grad_norm": 0.005800081882625818,
"learning_rate": 0.00016167117023487842,
"loss": 0.0014,
"step": 942
},
{
"epoch": 0.19519768163941212,
"grad_norm": 0.001085714902728796,
"learning_rate": 0.00016162959883600083,
"loss": 0.0001,
"step": 943
},
{
"epoch": 0.19540467812047196,
"grad_norm": 0.0017082407139241695,
"learning_rate": 0.00016158802743712326,
"loss": 0.0001,
"step": 944
},
{
"epoch": 0.19561167460153178,
"grad_norm": 0.0016056247986853123,
"learning_rate": 0.0001615464560382457,
"loss": 0.0009,
"step": 945
},
{
"epoch": 0.1958186710825916,
"grad_norm": 0.0005931173800490797,
"learning_rate": 0.00016150488463936814,
"loss": 0.0,
"step": 946
},
{
"epoch": 0.19602566756365142,
"grad_norm": 0.0002708766842260957,
"learning_rate": 0.00016146331324049055,
"loss": 0.0,
"step": 947
},
{
"epoch": 0.19623266404471124,
"grad_norm": 0.003350366372615099,
"learning_rate": 0.00016142174184161299,
"loss": 0.0004,
"step": 948
},
{
"epoch": 0.19643966052577105,
"grad_norm": 0.00030215582228265703,
"learning_rate": 0.00016138017044273542,
"loss": 0.0,
"step": 949
},
{
"epoch": 0.19664665700683087,
"grad_norm": 0.0013855715515092015,
"learning_rate": 0.00016133859904385783,
"loss": 0.0009,
"step": 950
},
{
"epoch": 0.19685365348789072,
"grad_norm": 0.0005864354898221791,
"learning_rate": 0.00016129702764498027,
"loss": 0.0,
"step": 951
},
{
"epoch": 0.19706064996895054,
"grad_norm": 0.0006372429197654128,
"learning_rate": 0.0001612554562461027,
"loss": 0.0,
"step": 952
},
{
"epoch": 0.19726764645001035,
"grad_norm": 0.0005041586118750274,
"learning_rate": 0.00016121388484722512,
"loss": 0.0,
"step": 953
},
{
"epoch": 0.19747464293107017,
"grad_norm": 0.0023472902830690145,
"learning_rate": 0.00016117231344834753,
"loss": 0.001,
"step": 954
},
{
"epoch": 0.19768163941213,
"grad_norm": 0.00015194782463368028,
"learning_rate": 0.00016113074204946996,
"loss": 0.0,
"step": 955
},
{
"epoch": 0.1978886358931898,
"grad_norm": 0.001190232578665018,
"learning_rate": 0.0001610891706505924,
"loss": 0.0002,
"step": 956
},
{
"epoch": 0.19809563237424963,
"grad_norm": 0.0018357646185904741,
"learning_rate": 0.0001610475992517148,
"loss": 0.0002,
"step": 957
},
{
"epoch": 0.19830262885530947,
"grad_norm": 0.007886867970228195,
"learning_rate": 0.00016100602785283725,
"loss": 0.0015,
"step": 958
},
{
"epoch": 0.1985096253363693,
"grad_norm": 0.00014407855633180588,
"learning_rate": 0.00016096445645395968,
"loss": 0.0,
"step": 959
},
{
"epoch": 0.1987166218174291,
"grad_norm": 0.0008407345740124583,
"learning_rate": 0.0001609228850550821,
"loss": 0.0001,
"step": 960
},
{
"epoch": 0.19892361829848892,
"grad_norm": 0.0005690194084309042,
"learning_rate": 0.00016088131365620453,
"loss": 0.0,
"step": 961
},
{
"epoch": 0.19913061477954874,
"grad_norm": 0.0001164446584880352,
"learning_rate": 0.00016083974225732697,
"loss": 0.0,
"step": 962
},
{
"epoch": 0.19933761126060856,
"grad_norm": 0.004800689872354269,
"learning_rate": 0.0001607981708584494,
"loss": 0.0009,
"step": 963
},
{
"epoch": 0.19954460774166838,
"grad_norm": 0.0038551113102585077,
"learning_rate": 0.00016075659945957182,
"loss": 0.0002,
"step": 964
},
{
"epoch": 0.19975160422272822,
"grad_norm": 0.00023845378018449992,
"learning_rate": 0.00016071502806069425,
"loss": 0.0,
"step": 965
},
{
"epoch": 0.19995860070378804,
"grad_norm": 0.0006543719209730625,
"learning_rate": 0.0001606734566618167,
"loss": 0.0,
"step": 966
},
{
"epoch": 0.20016559718484786,
"grad_norm": 0.0024344087578356266,
"learning_rate": 0.0001606318852629391,
"loss": 0.0011,
"step": 967
},
{
"epoch": 0.20037259366590768,
"grad_norm": 0.0006420607678592205,
"learning_rate": 0.00016059031386406154,
"loss": 0.0,
"step": 968
},
{
"epoch": 0.2005795901469675,
"grad_norm": 0.0016330952057614923,
"learning_rate": 0.00016054874246518397,
"loss": 0.0,
"step": 969
},
{
"epoch": 0.2007865866280273,
"grad_norm": 0.0013299377169460058,
"learning_rate": 0.00016050717106630638,
"loss": 0.0001,
"step": 970
},
{
"epoch": 0.20099358310908713,
"grad_norm": 0.0023206102196127176,
"learning_rate": 0.00016046559966742882,
"loss": 0.0006,
"step": 971
},
{
"epoch": 0.20120057959014698,
"grad_norm": 0.0030964380130171776,
"learning_rate": 0.00016042402826855126,
"loss": 0.0004,
"step": 972
},
{
"epoch": 0.2014075760712068,
"grad_norm": 0.0015272133750841022,
"learning_rate": 0.00016038245686967367,
"loss": 0.0,
"step": 973
},
{
"epoch": 0.2016145725522666,
"grad_norm": 0.0036174836568534374,
"learning_rate": 0.0001603408854707961,
"loss": 0.0002,
"step": 974
},
{
"epoch": 0.20182156903332643,
"grad_norm": 0.0014752513961866498,
"learning_rate": 0.00016029931407191854,
"loss": 0.0,
"step": 975
},
{
"epoch": 0.20202856551438625,
"grad_norm": 0.0008594008395448327,
"learning_rate": 0.00016025774267304095,
"loss": 0.0001,
"step": 976
},
{
"epoch": 0.20223556199544607,
"grad_norm": 0.007280942518264055,
"learning_rate": 0.00016021617127416336,
"loss": 0.001,
"step": 977
},
{
"epoch": 0.2024425584765059,
"grad_norm": 0.0013399991439655423,
"learning_rate": 0.0001601745998752858,
"loss": 0.0,
"step": 978
},
{
"epoch": 0.20264955495756573,
"grad_norm": 0.0015200217021629214,
"learning_rate": 0.00016013302847640824,
"loss": 0.0003,
"step": 979
},
{
"epoch": 0.20285655143862555,
"grad_norm": 0.008712020702660084,
"learning_rate": 0.00016009145707753067,
"loss": 0.0002,
"step": 980
},
{
"epoch": 0.20306354791968537,
"grad_norm": 0.004218498710542917,
"learning_rate": 0.00016004988567865308,
"loss": 0.0006,
"step": 981
},
{
"epoch": 0.20327054440074518,
"grad_norm": 0.0019195530330762267,
"learning_rate": 0.00016000831427977552,
"loss": 0.0008,
"step": 982
},
{
"epoch": 0.203477540881805,
"grad_norm": 0.0017306975787505507,
"learning_rate": 0.00015996674288089796,
"loss": 0.001,
"step": 983
},
{
"epoch": 0.20368453736286482,
"grad_norm": 0.006045056506991386,
"learning_rate": 0.00015992517148202037,
"loss": 0.0002,
"step": 984
},
{
"epoch": 0.20389153384392467,
"grad_norm": 0.002741064177826047,
"learning_rate": 0.0001598836000831428,
"loss": 0.0001,
"step": 985
},
{
"epoch": 0.20409853032498448,
"grad_norm": 0.0026846020482480526,
"learning_rate": 0.00015984202868426524,
"loss": 0.0003,
"step": 986
},
{
"epoch": 0.2043055268060443,
"grad_norm": 0.009860471822321415,
"learning_rate": 0.00015980045728538765,
"loss": 0.0002,
"step": 987
},
{
"epoch": 0.20451252328710412,
"grad_norm": 0.0001563982223160565,
"learning_rate": 0.0001597588858865101,
"loss": 0.0,
"step": 988
},
{
"epoch": 0.20471951976816394,
"grad_norm": 0.0035680129658430815,
"learning_rate": 0.00015971731448763253,
"loss": 0.0009,
"step": 989
},
{
"epoch": 0.20492651624922376,
"grad_norm": 0.00037079930189065635,
"learning_rate": 0.00015967574308875494,
"loss": 0.0,
"step": 990
},
{
"epoch": 0.20513351273028357,
"grad_norm": 0.006476435344666243,
"learning_rate": 0.00015963417168987737,
"loss": 0.0004,
"step": 991
},
{
"epoch": 0.20534050921134342,
"grad_norm": 0.0008909539901651442,
"learning_rate": 0.0001595926002909998,
"loss": 0.0001,
"step": 992
},
{
"epoch": 0.20554750569240324,
"grad_norm": 0.005418546497821808,
"learning_rate": 0.00015955102889212225,
"loss": 0.0009,
"step": 993
},
{
"epoch": 0.20575450217346306,
"grad_norm": 0.004925790708512068,
"learning_rate": 0.00015950945749324466,
"loss": 0.0009,
"step": 994
},
{
"epoch": 0.20596149865452287,
"grad_norm": 0.003130319295451045,
"learning_rate": 0.0001594678860943671,
"loss": 0.001,
"step": 995
},
{
"epoch": 0.2061684951355827,
"grad_norm": 0.0062978435307741165,
"learning_rate": 0.00015942631469548953,
"loss": 0.0002,
"step": 996
},
{
"epoch": 0.2063754916166425,
"grad_norm": 0.006842117290943861,
"learning_rate": 0.00015938474329661194,
"loss": 0.0003,
"step": 997
},
{
"epoch": 0.20658248809770233,
"grad_norm": 0.0006231727893464267,
"learning_rate": 0.00015934317189773435,
"loss": 0.0,
"step": 998
},
{
"epoch": 0.20678948457876217,
"grad_norm": 0.0008572920341975987,
"learning_rate": 0.0001593016004988568,
"loss": 0.0,
"step": 999
},
{
"epoch": 0.206996481059822,
"grad_norm": 0.0018585945945233107,
"learning_rate": 0.00015926002909997922,
"loss": 0.0001,
"step": 1000
},
{
"epoch": 0.2072034775408818,
"grad_norm": 0.002304868074133992,
"learning_rate": 0.00015921845770110163,
"loss": 0.0009,
"step": 1001
},
{
"epoch": 0.20741047402194163,
"grad_norm": 0.004541350062936544,
"learning_rate": 0.00015917688630222407,
"loss": 0.0,
"step": 1002
},
{
"epoch": 0.20761747050300144,
"grad_norm": 0.003366716904565692,
"learning_rate": 0.0001591353149033465,
"loss": 0.0011,
"step": 1003
},
{
"epoch": 0.20782446698406126,
"grad_norm": 0.002917301142588258,
"learning_rate": 0.00015909374350446892,
"loss": 0.0017,
"step": 1004
},
{
"epoch": 0.20803146346512108,
"grad_norm": 0.0004168320447206497,
"learning_rate": 0.00015905217210559136,
"loss": 0.0,
"step": 1005
},
{
"epoch": 0.20823845994618093,
"grad_norm": 0.001863375655375421,
"learning_rate": 0.0001590106007067138,
"loss": 0.001,
"step": 1006
},
{
"epoch": 0.20844545642724074,
"grad_norm": 0.001271730288863182,
"learning_rate": 0.0001589690293078362,
"loss": 0.0001,
"step": 1007
},
{
"epoch": 0.20865245290830056,
"grad_norm": 0.002366506028920412,
"learning_rate": 0.00015892745790895864,
"loss": 0.0009,
"step": 1008
},
{
"epoch": 0.20885944938936038,
"grad_norm": 0.0031757184769958258,
"learning_rate": 0.00015888588651008108,
"loss": 0.0001,
"step": 1009
},
{
"epoch": 0.2090664458704202,
"grad_norm": 0.0019097458571195602,
"learning_rate": 0.00015884431511120351,
"loss": 0.0,
"step": 1010
},
{
"epoch": 0.20927344235148002,
"grad_norm": 0.0008379930513910949,
"learning_rate": 0.00015880274371232592,
"loss": 0.0,
"step": 1011
},
{
"epoch": 0.20948043883253983,
"grad_norm": 0.005206478293985128,
"learning_rate": 0.00015876117231344836,
"loss": 0.002,
"step": 1012
},
{
"epoch": 0.20968743531359968,
"grad_norm": 0.0008720169425942004,
"learning_rate": 0.0001587196009145708,
"loss": 0.0,
"step": 1013
},
{
"epoch": 0.2098944317946595,
"grad_norm": 0.0041591702029109,
"learning_rate": 0.0001586780295156932,
"loss": 0.001,
"step": 1014
},
{
"epoch": 0.21010142827571932,
"grad_norm": 0.0004002148343715817,
"learning_rate": 0.00015863645811681565,
"loss": 0.0,
"step": 1015
},
{
"epoch": 0.21030842475677913,
"grad_norm": 0.00017360522178933024,
"learning_rate": 0.00015859488671793808,
"loss": 0.0,
"step": 1016
},
{
"epoch": 0.21051542123783895,
"grad_norm": 0.004276643507182598,
"learning_rate": 0.0001585533153190605,
"loss": 0.0,
"step": 1017
},
{
"epoch": 0.21072241771889877,
"grad_norm": 0.0027010890189558268,
"learning_rate": 0.00015851174392018293,
"loss": 0.0014,
"step": 1018
},
{
"epoch": 0.2109294141999586,
"grad_norm": 0.0048659988678991795,
"learning_rate": 0.00015847017252130537,
"loss": 0.0005,
"step": 1019
},
{
"epoch": 0.21113641068101843,
"grad_norm": 0.0003106594958808273,
"learning_rate": 0.00015842860112242778,
"loss": 0.0,
"step": 1020
},
{
"epoch": 0.21134340716207825,
"grad_norm": 0.0032943710684776306,
"learning_rate": 0.00015838702972355019,
"loss": 0.0012,
"step": 1021
},
{
"epoch": 0.21155040364313807,
"grad_norm": 0.0022477346938103437,
"learning_rate": 0.00015834545832467262,
"loss": 0.0002,
"step": 1022
},
{
"epoch": 0.2117574001241979,
"grad_norm": 0.0007089116843417287,
"learning_rate": 0.00015830388692579506,
"loss": 0.0,
"step": 1023
},
{
"epoch": 0.2119643966052577,
"grad_norm": 0.003983316943049431,
"learning_rate": 0.00015826231552691747,
"loss": 0.0002,
"step": 1024
},
{
"epoch": 0.21217139308631752,
"grad_norm": 0.0038651269860565662,
"learning_rate": 0.0001582207441280399,
"loss": 0.0017,
"step": 1025
},
{
"epoch": 0.21237838956737737,
"grad_norm": 0.006369102746248245,
"learning_rate": 0.00015817917272916234,
"loss": 0.0002,
"step": 1026
},
{
"epoch": 0.21258538604843719,
"grad_norm": 0.002382291480898857,
"learning_rate": 0.00015813760133028478,
"loss": 0.0004,
"step": 1027
},
{
"epoch": 0.212792382529497,
"grad_norm": 0.0016512110596522689,
"learning_rate": 0.0001580960299314072,
"loss": 0.0002,
"step": 1028
},
{
"epoch": 0.21299937901055682,
"grad_norm": 0.008482804521918297,
"learning_rate": 0.00015805445853252963,
"loss": 0.0002,
"step": 1029
},
{
"epoch": 0.21320637549161664,
"grad_norm": 0.005470529198646545,
"learning_rate": 0.00015801288713365207,
"loss": 0.0007,
"step": 1030
},
{
"epoch": 0.21341337197267646,
"grad_norm": 0.004424599930644035,
"learning_rate": 0.00015797131573477448,
"loss": 0.0007,
"step": 1031
},
{
"epoch": 0.21362036845373628,
"grad_norm": 0.0011165774194523692,
"learning_rate": 0.0001579297443358969,
"loss": 0.0,
"step": 1032
},
{
"epoch": 0.21382736493479612,
"grad_norm": 0.0008872200851328671,
"learning_rate": 0.00015788817293701935,
"loss": 0.0,
"step": 1033
},
{
"epoch": 0.21403436141585594,
"grad_norm": 0.010052971541881561,
"learning_rate": 0.00015784660153814176,
"loss": 0.0011,
"step": 1034
},
{
"epoch": 0.21424135789691576,
"grad_norm": 0.002929918933659792,
"learning_rate": 0.0001578050301392642,
"loss": 0.0015,
"step": 1035
},
{
"epoch": 0.21444835437797558,
"grad_norm": 0.008970585651695728,
"learning_rate": 0.00015776345874038663,
"loss": 0.0003,
"step": 1036
},
{
"epoch": 0.2146553508590354,
"grad_norm": 0.06671982258558273,
"learning_rate": 0.00015772188734150904,
"loss": 0.0008,
"step": 1037
},
{
"epoch": 0.2148623473400952,
"grad_norm": 0.0003781789855565876,
"learning_rate": 0.00015768031594263148,
"loss": 0.0,
"step": 1038
},
{
"epoch": 0.21506934382115503,
"grad_norm": 0.0004683547012973577,
"learning_rate": 0.00015763874454375392,
"loss": 0.0,
"step": 1039
},
{
"epoch": 0.21527634030221487,
"grad_norm": 0.005777016282081604,
"learning_rate": 0.00015759717314487635,
"loss": 0.0015,
"step": 1040
},
{
"epoch": 0.2154833367832747,
"grad_norm": 0.0068507413379848,
"learning_rate": 0.00015755560174599876,
"loss": 0.0002,
"step": 1041
},
{
"epoch": 0.2156903332643345,
"grad_norm": 0.0021482266020029783,
"learning_rate": 0.00015751403034712117,
"loss": 0.0004,
"step": 1042
},
{
"epoch": 0.21589732974539433,
"grad_norm": 0.003305132733657956,
"learning_rate": 0.0001574724589482436,
"loss": 0.0009,
"step": 1043
},
{
"epoch": 0.21610432622645415,
"grad_norm": 0.005324844736605883,
"learning_rate": 0.00015743088754936602,
"loss": 0.0012,
"step": 1044
},
{
"epoch": 0.21631132270751396,
"grad_norm": 0.004400115925818682,
"learning_rate": 0.00015738931615048846,
"loss": 0.0002,
"step": 1045
},
{
"epoch": 0.21651831918857378,
"grad_norm": 0.0030595625285059214,
"learning_rate": 0.0001573477447516109,
"loss": 0.0005,
"step": 1046
},
{
"epoch": 0.21672531566963363,
"grad_norm": 0.004036907572299242,
"learning_rate": 0.00015730617335273333,
"loss": 0.0,
"step": 1047
},
{
"epoch": 0.21693231215069345,
"grad_norm": 0.005875944159924984,
"learning_rate": 0.00015726460195385574,
"loss": 0.0004,
"step": 1048
},
{
"epoch": 0.21713930863175326,
"grad_norm": 0.003494358854368329,
"learning_rate": 0.00015722303055497818,
"loss": 0.0015,
"step": 1049
},
{
"epoch": 0.21734630511281308,
"grad_norm": 0.00041328632505610585,
"learning_rate": 0.00015718145915610062,
"loss": 0.0,
"step": 1050
},
{
"epoch": 0.2175533015938729,
"grad_norm": 0.0010599165689200163,
"learning_rate": 0.00015713988775722303,
"loss": 0.0,
"step": 1051
},
{
"epoch": 0.21776029807493272,
"grad_norm": 0.00022103896480984986,
"learning_rate": 0.00015709831635834546,
"loss": 0.0,
"step": 1052
},
{
"epoch": 0.21796729455599254,
"grad_norm": 0.00018703911337070167,
"learning_rate": 0.0001570567449594679,
"loss": 0.0,
"step": 1053
},
{
"epoch": 0.21817429103705238,
"grad_norm": 0.0001905701938085258,
"learning_rate": 0.0001570151735605903,
"loss": 0.0,
"step": 1054
},
{
"epoch": 0.2183812875181122,
"grad_norm": 0.01590561680495739,
"learning_rate": 0.00015697360216171275,
"loss": 0.0004,
"step": 1055
},
{
"epoch": 0.21858828399917202,
"grad_norm": 0.007658824324607849,
"learning_rate": 0.00015693203076283519,
"loss": 0.0001,
"step": 1056
},
{
"epoch": 0.21879528048023184,
"grad_norm": 0.0036896623205393553,
"learning_rate": 0.00015689045936395762,
"loss": 0.0011,
"step": 1057
},
{
"epoch": 0.21900227696129165,
"grad_norm": 0.006060061044991016,
"learning_rate": 0.00015684888796508003,
"loss": 0.0001,
"step": 1058
},
{
"epoch": 0.21920927344235147,
"grad_norm": 0.010098116472363472,
"learning_rate": 0.00015680731656620247,
"loss": 0.0012,
"step": 1059
},
{
"epoch": 0.2194162699234113,
"grad_norm": 0.0016395826824009418,
"learning_rate": 0.0001567657451673249,
"loss": 0.0007,
"step": 1060
},
{
"epoch": 0.21962326640447113,
"grad_norm": 0.004565931856632233,
"learning_rate": 0.00015672417376844732,
"loss": 0.0011,
"step": 1061
},
{
"epoch": 0.21983026288553095,
"grad_norm": 0.004525905009359121,
"learning_rate": 0.00015668260236956975,
"loss": 0.0002,
"step": 1062
},
{
"epoch": 0.22003725936659077,
"grad_norm": 0.0008571099024266005,
"learning_rate": 0.0001566410309706922,
"loss": 0.0,
"step": 1063
}
],
"logging_steps": 1,
"max_steps": 4831,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 5.0739590310395904e+17,
"train_batch_size": 12,
"trial_name": null,
"trial_params": null
}