trapoom555's picture
Upload checkpoints
7b65800
raw
history blame
49.3 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.3047851264858275,
"eval_steps": 500,
"global_step": 3000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.001015950421619425,
"grad_norm": 17.625,
"learning_rate": 5e-06,
"loss": 3.4264,
"step": 10
},
{
"epoch": 0.00203190084323885,
"grad_norm": 12.5625,
"learning_rate": 1e-05,
"loss": 3.432,
"step": 20
},
{
"epoch": 0.003047851264858275,
"grad_norm": 14.0625,
"learning_rate": 1.5e-05,
"loss": 3.23,
"step": 30
},
{
"epoch": 0.0040638016864777,
"grad_norm": 12.4375,
"learning_rate": 2e-05,
"loss": 2.9762,
"step": 40
},
{
"epoch": 0.005079752108097125,
"grad_norm": 10.0625,
"learning_rate": 2.5e-05,
"loss": 2.6173,
"step": 50
},
{
"epoch": 0.00609570252971655,
"grad_norm": 10.1875,
"learning_rate": 3e-05,
"loss": 2.2004,
"step": 60
},
{
"epoch": 0.007111652951335975,
"grad_norm": 7.03125,
"learning_rate": 3.5e-05,
"loss": 1.4176,
"step": 70
},
{
"epoch": 0.0081276033729554,
"grad_norm": 4.375,
"learning_rate": 4e-05,
"loss": 1.0122,
"step": 80
},
{
"epoch": 0.009143553794574825,
"grad_norm": 6.5625,
"learning_rate": 4.5e-05,
"loss": 0.9116,
"step": 90
},
{
"epoch": 0.01015950421619425,
"grad_norm": 5.28125,
"learning_rate": 5e-05,
"loss": 0.6832,
"step": 100
},
{
"epoch": 0.011175454637813675,
"grad_norm": 5.5,
"learning_rate": 4.9999870035728426e-05,
"loss": 0.7355,
"step": 110
},
{
"epoch": 0.0121914050594331,
"grad_norm": 5.1875,
"learning_rate": 4.9999480144264944e-05,
"loss": 0.6673,
"step": 120
},
{
"epoch": 0.013207355481052525,
"grad_norm": 4.5,
"learning_rate": 4.9998830329663314e-05,
"loss": 0.6792,
"step": 130
},
{
"epoch": 0.01422330590267195,
"grad_norm": 3.9375,
"learning_rate": 4.9997920598679756e-05,
"loss": 0.6207,
"step": 140
},
{
"epoch": 0.015239256324291375,
"grad_norm": 3.15625,
"learning_rate": 4.999675096077286e-05,
"loss": 0.483,
"step": 150
},
{
"epoch": 0.0162552067459108,
"grad_norm": 5.28125,
"learning_rate": 4.999532142810354e-05,
"loss": 0.5319,
"step": 160
},
{
"epoch": 0.017271157167530225,
"grad_norm": 4.59375,
"learning_rate": 4.999363201553483e-05,
"loss": 0.6052,
"step": 170
},
{
"epoch": 0.01828710758914965,
"grad_norm": 5.03125,
"learning_rate": 4.9991682740631794e-05,
"loss": 0.4258,
"step": 180
},
{
"epoch": 0.019303058010769075,
"grad_norm": 3.859375,
"learning_rate": 4.998947362366133e-05,
"loss": 0.4309,
"step": 190
},
{
"epoch": 0.0203190084323885,
"grad_norm": 3.328125,
"learning_rate": 4.998700468759193e-05,
"loss": 0.3957,
"step": 200
},
{
"epoch": 0.021334958854007924,
"grad_norm": 4.9375,
"learning_rate": 4.9984275958093475e-05,
"loss": 0.4777,
"step": 210
},
{
"epoch": 0.02235090927562735,
"grad_norm": 4.78125,
"learning_rate": 4.998128746353695e-05,
"loss": 0.3549,
"step": 220
},
{
"epoch": 0.023366859697246774,
"grad_norm": 4.0625,
"learning_rate": 4.997803923499417e-05,
"loss": 0.4447,
"step": 230
},
{
"epoch": 0.0243828101188662,
"grad_norm": 6.375,
"learning_rate": 4.99745313062374e-05,
"loss": 0.3808,
"step": 240
},
{
"epoch": 0.025398760540485624,
"grad_norm": 3.59375,
"learning_rate": 4.99707637137391e-05,
"loss": 0.3827,
"step": 250
},
{
"epoch": 0.02641471096210505,
"grad_norm": 3.015625,
"learning_rate": 4.996673649667145e-05,
"loss": 0.3694,
"step": 260
},
{
"epoch": 0.027430661383724474,
"grad_norm": 2.296875,
"learning_rate": 4.9962449696906e-05,
"loss": 0.3586,
"step": 270
},
{
"epoch": 0.0284466118053439,
"grad_norm": 4.125,
"learning_rate": 4.9957903359013214e-05,
"loss": 0.3832,
"step": 280
},
{
"epoch": 0.029462562226963324,
"grad_norm": 3.296875,
"learning_rate": 4.995309753026201e-05,
"loss": 0.328,
"step": 290
},
{
"epoch": 0.03047851264858275,
"grad_norm": 4.5,
"learning_rate": 4.994803226061927e-05,
"loss": 0.3667,
"step": 300
},
{
"epoch": 0.03149446307020217,
"grad_norm": 4.3125,
"learning_rate": 4.994270760274933e-05,
"loss": 0.3811,
"step": 310
},
{
"epoch": 0.0325104134918216,
"grad_norm": 3.421875,
"learning_rate": 4.99371236120134e-05,
"loss": 0.3065,
"step": 320
},
{
"epoch": 0.03352636391344102,
"grad_norm": 4.6875,
"learning_rate": 4.993128034646902e-05,
"loss": 0.4177,
"step": 330
},
{
"epoch": 0.03454231433506045,
"grad_norm": 3.046875,
"learning_rate": 4.992517786686947e-05,
"loss": 0.33,
"step": 340
},
{
"epoch": 0.03555826475667987,
"grad_norm": 1.8828125,
"learning_rate": 4.9918816236663077e-05,
"loss": 0.3287,
"step": 350
},
{
"epoch": 0.0365742151782993,
"grad_norm": 3.8125,
"learning_rate": 4.991219552199262e-05,
"loss": 0.2934,
"step": 360
},
{
"epoch": 0.03759016559991872,
"grad_norm": 4.28125,
"learning_rate": 4.99053157916946e-05,
"loss": 0.3176,
"step": 370
},
{
"epoch": 0.03860611602153815,
"grad_norm": 2.609375,
"learning_rate": 4.989817711729856e-05,
"loss": 0.3318,
"step": 380
},
{
"epoch": 0.03962206644315757,
"grad_norm": 2.375,
"learning_rate": 4.98907795730263e-05,
"loss": 0.3234,
"step": 390
},
{
"epoch": 0.040638016864777,
"grad_norm": 4.46875,
"learning_rate": 4.988312323579114e-05,
"loss": 0.267,
"step": 400
},
{
"epoch": 0.04165396728639642,
"grad_norm": 3.75,
"learning_rate": 4.98752081851971e-05,
"loss": 0.3081,
"step": 410
},
{
"epoch": 0.04266991770801585,
"grad_norm": 2.203125,
"learning_rate": 4.986703450353809e-05,
"loss": 0.2917,
"step": 420
},
{
"epoch": 0.04368586812963527,
"grad_norm": 1.6015625,
"learning_rate": 4.985860227579703e-05,
"loss": 0.2805,
"step": 430
},
{
"epoch": 0.0447018185512547,
"grad_norm": 3.140625,
"learning_rate": 4.984991158964499e-05,
"loss": 0.3534,
"step": 440
},
{
"epoch": 0.04571776897287412,
"grad_norm": 3.296875,
"learning_rate": 4.9840962535440265e-05,
"loss": 0.335,
"step": 450
},
{
"epoch": 0.04673371939449355,
"grad_norm": 3.25,
"learning_rate": 4.983175520622744e-05,
"loss": 0.2544,
"step": 460
},
{
"epoch": 0.04774966981611297,
"grad_norm": 2.25,
"learning_rate": 4.982228969773642e-05,
"loss": 0.3449,
"step": 470
},
{
"epoch": 0.0487656202377324,
"grad_norm": 4.9375,
"learning_rate": 4.9812566108381435e-05,
"loss": 0.2964,
"step": 480
},
{
"epoch": 0.04978157065935182,
"grad_norm": 1.5703125,
"learning_rate": 4.9802584539260035e-05,
"loss": 0.2799,
"step": 490
},
{
"epoch": 0.05079752108097125,
"grad_norm": 2.828125,
"learning_rate": 4.979234509415199e-05,
"loss": 0.3231,
"step": 500
},
{
"epoch": 0.05181347150259067,
"grad_norm": 2.9375,
"learning_rate": 4.978184787951828e-05,
"loss": 0.2943,
"step": 510
},
{
"epoch": 0.0528294219242101,
"grad_norm": 2.34375,
"learning_rate": 4.977109300449992e-05,
"loss": 0.2705,
"step": 520
},
{
"epoch": 0.05384537234582952,
"grad_norm": 3.140625,
"learning_rate": 4.9760080580916876e-05,
"loss": 0.2998,
"step": 530
},
{
"epoch": 0.05486132276744895,
"grad_norm": 3.5625,
"learning_rate": 4.974881072326688e-05,
"loss": 0.2595,
"step": 540
},
{
"epoch": 0.05587727318906837,
"grad_norm": 4.25,
"learning_rate": 4.9737283548724236e-05,
"loss": 0.2803,
"step": 550
},
{
"epoch": 0.0568932236106878,
"grad_norm": 4.0625,
"learning_rate": 4.97254991771386e-05,
"loss": 0.3511,
"step": 560
},
{
"epoch": 0.05790917403230722,
"grad_norm": 2.515625,
"learning_rate": 4.971345773103377e-05,
"loss": 0.312,
"step": 570
},
{
"epoch": 0.05892512445392665,
"grad_norm": 3.21875,
"learning_rate": 4.9701159335606365e-05,
"loss": 0.2482,
"step": 580
},
{
"epoch": 0.05994107487554607,
"grad_norm": 5.5,
"learning_rate": 4.968860411872454e-05,
"loss": 0.2537,
"step": 590
},
{
"epoch": 0.0609570252971655,
"grad_norm": 3.546875,
"learning_rate": 4.967579221092666e-05,
"loss": 0.3125,
"step": 600
},
{
"epoch": 0.06197297571878492,
"grad_norm": 2.984375,
"learning_rate": 4.966272374541996e-05,
"loss": 0.2354,
"step": 610
},
{
"epoch": 0.06298892614040434,
"grad_norm": 3.6875,
"learning_rate": 4.964939885807912e-05,
"loss": 0.3213,
"step": 620
},
{
"epoch": 0.06400487656202378,
"grad_norm": 2.140625,
"learning_rate": 4.9635817687444876e-05,
"loss": 0.3003,
"step": 630
},
{
"epoch": 0.0650208269836432,
"grad_norm": 3.484375,
"learning_rate": 4.962198037472259e-05,
"loss": 0.2996,
"step": 640
},
{
"epoch": 0.06603677740526262,
"grad_norm": 3.21875,
"learning_rate": 4.9607887063780776e-05,
"loss": 0.2257,
"step": 650
},
{
"epoch": 0.06705272782688204,
"grad_norm": 5.375,
"learning_rate": 4.9593537901149564e-05,
"loss": 0.223,
"step": 660
},
{
"epoch": 0.06806867824850148,
"grad_norm": 4.1875,
"learning_rate": 4.957893303601924e-05,
"loss": 0.3407,
"step": 670
},
{
"epoch": 0.0690846286701209,
"grad_norm": 3.328125,
"learning_rate": 4.956407262023866e-05,
"loss": 0.2589,
"step": 680
},
{
"epoch": 0.07010057909174032,
"grad_norm": 2.953125,
"learning_rate": 4.954895680831367e-05,
"loss": 0.2949,
"step": 690
},
{
"epoch": 0.07111652951335974,
"grad_norm": 4.0625,
"learning_rate": 4.9533585757405506e-05,
"loss": 0.2995,
"step": 700
},
{
"epoch": 0.07213247993497918,
"grad_norm": 4.625,
"learning_rate": 4.951795962732917e-05,
"loss": 0.2894,
"step": 710
},
{
"epoch": 0.0731484303565986,
"grad_norm": 3.0,
"learning_rate": 4.9502078580551755e-05,
"loss": 0.3082,
"step": 720
},
{
"epoch": 0.07416438077821802,
"grad_norm": 3.65625,
"learning_rate": 4.9485942782190734e-05,
"loss": 0.2308,
"step": 730
},
{
"epoch": 0.07518033119983744,
"grad_norm": 4.78125,
"learning_rate": 4.9469552400012306e-05,
"loss": 0.2272,
"step": 740
},
{
"epoch": 0.07619628162145688,
"grad_norm": 4.25,
"learning_rate": 4.94529076044296e-05,
"loss": 0.2701,
"step": 750
},
{
"epoch": 0.0772122320430763,
"grad_norm": 3.140625,
"learning_rate": 4.94360085685009e-05,
"loss": 0.2686,
"step": 760
},
{
"epoch": 0.07822818246469572,
"grad_norm": 0.765625,
"learning_rate": 4.9418855467927894e-05,
"loss": 0.2051,
"step": 770
},
{
"epoch": 0.07924413288631514,
"grad_norm": 1.796875,
"learning_rate": 4.940144848105379e-05,
"loss": 0.2267,
"step": 780
},
{
"epoch": 0.08026008330793458,
"grad_norm": 4.5625,
"learning_rate": 4.93837877888615e-05,
"loss": 0.2597,
"step": 790
},
{
"epoch": 0.081276033729554,
"grad_norm": 3.03125,
"learning_rate": 4.9365873574971745e-05,
"loss": 0.3701,
"step": 800
},
{
"epoch": 0.08229198415117342,
"grad_norm": 4.5625,
"learning_rate": 4.9347706025641136e-05,
"loss": 0.2559,
"step": 810
},
{
"epoch": 0.08330793457279284,
"grad_norm": 3.90625,
"learning_rate": 4.9329285329760275e-05,
"loss": 0.2799,
"step": 820
},
{
"epoch": 0.08432388499441228,
"grad_norm": 3.140625,
"learning_rate": 4.9310611678851735e-05,
"loss": 0.2866,
"step": 830
},
{
"epoch": 0.0853398354160317,
"grad_norm": 2.46875,
"learning_rate": 4.929168526706811e-05,
"loss": 0.3105,
"step": 840
},
{
"epoch": 0.08635578583765112,
"grad_norm": 13.625,
"learning_rate": 4.927250629119e-05,
"loss": 0.2454,
"step": 850
},
{
"epoch": 0.08737173625927054,
"grad_norm": 3.921875,
"learning_rate": 4.9253074950623925e-05,
"loss": 0.2424,
"step": 860
},
{
"epoch": 0.08838768668088998,
"grad_norm": 2.90625,
"learning_rate": 4.9233391447400286e-05,
"loss": 0.2481,
"step": 870
},
{
"epoch": 0.0894036371025094,
"grad_norm": 2.96875,
"learning_rate": 4.921345598617125e-05,
"loss": 0.2231,
"step": 880
},
{
"epoch": 0.09041958752412882,
"grad_norm": 5.375,
"learning_rate": 4.9193268774208654e-05,
"loss": 0.3447,
"step": 890
},
{
"epoch": 0.09143553794574824,
"grad_norm": 2.0,
"learning_rate": 4.9172830021401785e-05,
"loss": 0.229,
"step": 900
},
{
"epoch": 0.09245148836736768,
"grad_norm": 3.1875,
"learning_rate": 4.9152139940255245e-05,
"loss": 0.2122,
"step": 910
},
{
"epoch": 0.0934674387889871,
"grad_norm": 3.40625,
"learning_rate": 4.913119874588677e-05,
"loss": 0.2386,
"step": 920
},
{
"epoch": 0.09448338921060652,
"grad_norm": 1.4609375,
"learning_rate": 4.911000665602489e-05,
"loss": 0.1944,
"step": 930
},
{
"epoch": 0.09549933963222594,
"grad_norm": 5.0625,
"learning_rate": 4.9088563891006786e-05,
"loss": 0.2038,
"step": 940
},
{
"epoch": 0.09651529005384538,
"grad_norm": 4.53125,
"learning_rate": 4.906687067377592e-05,
"loss": 0.3122,
"step": 950
},
{
"epoch": 0.0975312404754648,
"grad_norm": 2.84375,
"learning_rate": 4.904492722987976e-05,
"loss": 0.3157,
"step": 960
},
{
"epoch": 0.09854719089708422,
"grad_norm": 2.171875,
"learning_rate": 4.902273378746738e-05,
"loss": 0.3077,
"step": 970
},
{
"epoch": 0.09956314131870364,
"grad_norm": 2.84375,
"learning_rate": 4.9000290577287165e-05,
"loss": 0.2756,
"step": 980
},
{
"epoch": 0.10057909174032308,
"grad_norm": 0.99609375,
"learning_rate": 4.897759783268434e-05,
"loss": 0.2915,
"step": 990
},
{
"epoch": 0.1015950421619425,
"grad_norm": 3.53125,
"learning_rate": 4.895465578959859e-05,
"loss": 0.2052,
"step": 1000
},
{
"epoch": 0.10261099258356192,
"grad_norm": 4.0,
"learning_rate": 4.893146468656159e-05,
"loss": 0.2499,
"step": 1010
},
{
"epoch": 0.10362694300518134,
"grad_norm": 1.65625,
"learning_rate": 4.890802476469452e-05,
"loss": 0.278,
"step": 1020
},
{
"epoch": 0.10464289342680078,
"grad_norm": 3.625,
"learning_rate": 4.888433626770558e-05,
"loss": 0.2143,
"step": 1030
},
{
"epoch": 0.1056588438484202,
"grad_norm": 5.0625,
"learning_rate": 4.886039944188741e-05,
"loss": 0.2878,
"step": 1040
},
{
"epoch": 0.10667479427003962,
"grad_norm": 4.5,
"learning_rate": 4.883621453611461e-05,
"loss": 0.2744,
"step": 1050
},
{
"epoch": 0.10769074469165904,
"grad_norm": 4.5625,
"learning_rate": 4.881178180184106e-05,
"loss": 0.2734,
"step": 1060
},
{
"epoch": 0.10870669511327848,
"grad_norm": 3.125,
"learning_rate": 4.878710149309735e-05,
"loss": 0.3574,
"step": 1070
},
{
"epoch": 0.1097226455348979,
"grad_norm": 3.0625,
"learning_rate": 4.876217386648816e-05,
"loss": 0.2625,
"step": 1080
},
{
"epoch": 0.11073859595651732,
"grad_norm": 4.0625,
"learning_rate": 4.873699918118955e-05,
"loss": 0.2437,
"step": 1090
},
{
"epoch": 0.11175454637813674,
"grad_norm": 1.59375,
"learning_rate": 4.87115776989463e-05,
"loss": 0.2051,
"step": 1100
},
{
"epoch": 0.11277049679975618,
"grad_norm": 4.375,
"learning_rate": 4.8685909684069153e-05,
"loss": 0.1727,
"step": 1110
},
{
"epoch": 0.1137864472213756,
"grad_norm": 2.28125,
"learning_rate": 4.865999540343211e-05,
"loss": 0.2256,
"step": 1120
},
{
"epoch": 0.11480239764299502,
"grad_norm": 2.265625,
"learning_rate": 4.86338351264696e-05,
"loss": 0.3529,
"step": 1130
},
{
"epoch": 0.11581834806461444,
"grad_norm": 2.34375,
"learning_rate": 4.8607429125173754e-05,
"loss": 0.2113,
"step": 1140
},
{
"epoch": 0.11683429848623388,
"grad_norm": 0.7578125,
"learning_rate": 4.858077767409149e-05,
"loss": 0.2759,
"step": 1150
},
{
"epoch": 0.1178502489078533,
"grad_norm": 3.640625,
"learning_rate": 4.855388105032174e-05,
"loss": 0.2482,
"step": 1160
},
{
"epoch": 0.11886619932947272,
"grad_norm": 3.5,
"learning_rate": 4.852673953351249e-05,
"loss": 0.1865,
"step": 1170
},
{
"epoch": 0.11988214975109214,
"grad_norm": 3.75,
"learning_rate": 4.849935340585796e-05,
"loss": 0.2659,
"step": 1180
},
{
"epoch": 0.12089810017271158,
"grad_norm": 3.375,
"learning_rate": 4.8471722952095586e-05,
"loss": 0.1506,
"step": 1190
},
{
"epoch": 0.121914050594331,
"grad_norm": 3.34375,
"learning_rate": 4.844384845950312e-05,
"loss": 0.307,
"step": 1200
},
{
"epoch": 0.12293000101595042,
"grad_norm": 1.578125,
"learning_rate": 4.841573021789561e-05,
"loss": 0.1952,
"step": 1210
},
{
"epoch": 0.12394595143756984,
"grad_norm": 1.2890625,
"learning_rate": 4.838736851962239e-05,
"loss": 0.1779,
"step": 1220
},
{
"epoch": 0.12496190185918928,
"grad_norm": 1.265625,
"learning_rate": 4.835876365956408e-05,
"loss": 0.1235,
"step": 1230
},
{
"epoch": 0.12597785228080868,
"grad_norm": 1.9609375,
"learning_rate": 4.8329915935129436e-05,
"loss": 0.1876,
"step": 1240
},
{
"epoch": 0.12699380270242813,
"grad_norm": 1.6328125,
"learning_rate": 4.830082564625235e-05,
"loss": 0.2188,
"step": 1250
},
{
"epoch": 0.12800975312404755,
"grad_norm": 3.96875,
"learning_rate": 4.8271493095388684e-05,
"loss": 0.2622,
"step": 1260
},
{
"epoch": 0.12902570354566698,
"grad_norm": 3.765625,
"learning_rate": 4.824191858751312e-05,
"loss": 0.2724,
"step": 1270
},
{
"epoch": 0.1300416539672864,
"grad_norm": 5.59375,
"learning_rate": 4.821210243011601e-05,
"loss": 0.2413,
"step": 1280
},
{
"epoch": 0.13105760438890582,
"grad_norm": 3.34375,
"learning_rate": 4.818204493320016e-05,
"loss": 0.2618,
"step": 1290
},
{
"epoch": 0.13207355481052524,
"grad_norm": 2.78125,
"learning_rate": 4.8151746409277634e-05,
"loss": 0.2295,
"step": 1300
},
{
"epoch": 0.13308950523214466,
"grad_norm": 3.1875,
"learning_rate": 4.8121207173366484e-05,
"loss": 0.2733,
"step": 1310
},
{
"epoch": 0.13410545565376408,
"grad_norm": 2.28125,
"learning_rate": 4.809042754298746e-05,
"loss": 0.2311,
"step": 1320
},
{
"epoch": 0.13512140607538353,
"grad_norm": 2.171875,
"learning_rate": 4.805940783816075e-05,
"loss": 0.2059,
"step": 1330
},
{
"epoch": 0.13613735649700295,
"grad_norm": 2.796875,
"learning_rate": 4.8028148381402625e-05,
"loss": 0.2102,
"step": 1340
},
{
"epoch": 0.13715330691862238,
"grad_norm": 2.96875,
"learning_rate": 4.7996649497722084e-05,
"loss": 0.2708,
"step": 1350
},
{
"epoch": 0.1381692573402418,
"grad_norm": 2.4375,
"learning_rate": 4.7964911514617485e-05,
"loss": 0.2429,
"step": 1360
},
{
"epoch": 0.13918520776186122,
"grad_norm": 5.8125,
"learning_rate": 4.793293476207312e-05,
"loss": 0.2725,
"step": 1370
},
{
"epoch": 0.14020115818348064,
"grad_norm": 2.40625,
"learning_rate": 4.790071957255585e-05,
"loss": 0.2098,
"step": 1380
},
{
"epoch": 0.14121710860510006,
"grad_norm": 4.25,
"learning_rate": 4.786826628101154e-05,
"loss": 0.2101,
"step": 1390
},
{
"epoch": 0.14223305902671948,
"grad_norm": 2.578125,
"learning_rate": 4.783557522486167e-05,
"loss": 0.2624,
"step": 1400
},
{
"epoch": 0.14324900944833893,
"grad_norm": 3.125,
"learning_rate": 4.780264674399978e-05,
"loss": 0.2518,
"step": 1410
},
{
"epoch": 0.14426495986995835,
"grad_norm": 3.671875,
"learning_rate": 4.7769481180787966e-05,
"loss": 0.3112,
"step": 1420
},
{
"epoch": 0.14528091029157778,
"grad_norm": 3.984375,
"learning_rate": 4.773607888005327e-05,
"loss": 0.2747,
"step": 1430
},
{
"epoch": 0.1462968607131972,
"grad_norm": 3.234375,
"learning_rate": 4.770244018908416e-05,
"loss": 0.1572,
"step": 1440
},
{
"epoch": 0.14731281113481662,
"grad_norm": 4.09375,
"learning_rate": 4.766856545762687e-05,
"loss": 0.2148,
"step": 1450
},
{
"epoch": 0.14832876155643604,
"grad_norm": 1.6875,
"learning_rate": 4.763445503788178e-05,
"loss": 0.2531,
"step": 1460
},
{
"epoch": 0.14934471197805546,
"grad_norm": 2.375,
"learning_rate": 4.760010928449976e-05,
"loss": 0.199,
"step": 1470
},
{
"epoch": 0.15036066239967488,
"grad_norm": 4.6875,
"learning_rate": 4.7565528554578485e-05,
"loss": 0.2366,
"step": 1480
},
{
"epoch": 0.15137661282129433,
"grad_norm": 5.4375,
"learning_rate": 4.75307132076587e-05,
"loss": 0.1862,
"step": 1490
},
{
"epoch": 0.15239256324291375,
"grad_norm": 2.484375,
"learning_rate": 4.749566360572049e-05,
"loss": 0.2143,
"step": 1500
},
{
"epoch": 0.15340851366453317,
"grad_norm": 2.1875,
"learning_rate": 4.746038011317955e-05,
"loss": 0.1877,
"step": 1510
},
{
"epoch": 0.1544244640861526,
"grad_norm": 2.84375,
"learning_rate": 4.742486309688333e-05,
"loss": 0.2831,
"step": 1520
},
{
"epoch": 0.15544041450777202,
"grad_norm": 2.015625,
"learning_rate": 4.738911292610732e-05,
"loss": 0.1708,
"step": 1530
},
{
"epoch": 0.15645636492939144,
"grad_norm": 3.953125,
"learning_rate": 4.735312997255107e-05,
"loss": 0.192,
"step": 1540
},
{
"epoch": 0.15747231535101086,
"grad_norm": 2.09375,
"learning_rate": 4.7316914610334475e-05,
"loss": 0.2586,
"step": 1550
},
{
"epoch": 0.15848826577263028,
"grad_norm": 3.6875,
"learning_rate": 4.728046721599378e-05,
"loss": 0.2141,
"step": 1560
},
{
"epoch": 0.15950421619424973,
"grad_norm": 2.9375,
"learning_rate": 4.724378816847771e-05,
"loss": 0.193,
"step": 1570
},
{
"epoch": 0.16052016661586915,
"grad_norm": 1.5625,
"learning_rate": 4.720687784914352e-05,
"loss": 0.191,
"step": 1580
},
{
"epoch": 0.16153611703748857,
"grad_norm": 3.75,
"learning_rate": 4.716973664175304e-05,
"loss": 0.2172,
"step": 1590
},
{
"epoch": 0.162552067459108,
"grad_norm": 3.125,
"learning_rate": 4.7132364932468645e-05,
"loss": 0.2134,
"step": 1600
},
{
"epoch": 0.16356801788072742,
"grad_norm": 4.09375,
"learning_rate": 4.709476310984932e-05,
"loss": 0.2055,
"step": 1610
},
{
"epoch": 0.16458396830234684,
"grad_norm": 3.875,
"learning_rate": 4.705693156484652e-05,
"loss": 0.2136,
"step": 1620
},
{
"epoch": 0.16559991872396626,
"grad_norm": 1.1796875,
"learning_rate": 4.7018870690800196e-05,
"loss": 0.1471,
"step": 1630
},
{
"epoch": 0.16661586914558568,
"grad_norm": 2.5,
"learning_rate": 4.698058088343465e-05,
"loss": 0.2308,
"step": 1640
},
{
"epoch": 0.16763181956720513,
"grad_norm": 1.390625,
"learning_rate": 4.6942062540854425e-05,
"loss": 0.2456,
"step": 1650
},
{
"epoch": 0.16864776998882455,
"grad_norm": 3.125,
"learning_rate": 4.69033160635402e-05,
"loss": 0.2654,
"step": 1660
},
{
"epoch": 0.16966372041044397,
"grad_norm": 3.984375,
"learning_rate": 4.6864341854344587e-05,
"loss": 0.2226,
"step": 1670
},
{
"epoch": 0.1706796708320634,
"grad_norm": 2.328125,
"learning_rate": 4.682514031848795e-05,
"loss": 0.2438,
"step": 1680
},
{
"epoch": 0.17169562125368282,
"grad_norm": 3.078125,
"learning_rate": 4.678571186355423e-05,
"loss": 0.1889,
"step": 1690
},
{
"epoch": 0.17271157167530224,
"grad_norm": 3.328125,
"learning_rate": 4.6746056899486644e-05,
"loss": 0.2117,
"step": 1700
},
{
"epoch": 0.17372752209692166,
"grad_norm": 2.78125,
"learning_rate": 4.67061758385835e-05,
"loss": 0.1953,
"step": 1710
},
{
"epoch": 0.17474347251854108,
"grad_norm": 3.09375,
"learning_rate": 4.6666069095493816e-05,
"loss": 0.1844,
"step": 1720
},
{
"epoch": 0.17575942294016053,
"grad_norm": 3.234375,
"learning_rate": 4.662573708721309e-05,
"loss": 0.2774,
"step": 1730
},
{
"epoch": 0.17677537336177995,
"grad_norm": 4.03125,
"learning_rate": 4.658518023307894e-05,
"loss": 0.2527,
"step": 1740
},
{
"epoch": 0.17779132378339937,
"grad_norm": 3.21875,
"learning_rate": 4.654439895476671e-05,
"loss": 0.2164,
"step": 1750
},
{
"epoch": 0.1788072742050188,
"grad_norm": 2.390625,
"learning_rate": 4.6503393676285146e-05,
"loss": 0.2424,
"step": 1760
},
{
"epoch": 0.17982322462663822,
"grad_norm": 1.8359375,
"learning_rate": 4.646216482397192e-05,
"loss": 0.2428,
"step": 1770
},
{
"epoch": 0.18083917504825764,
"grad_norm": 2.796875,
"learning_rate": 4.6420712826489275e-05,
"loss": 0.2155,
"step": 1780
},
{
"epoch": 0.18185512546987706,
"grad_norm": 0.69921875,
"learning_rate": 4.6379038114819485e-05,
"loss": 0.1544,
"step": 1790
},
{
"epoch": 0.18287107589149648,
"grad_norm": 3.40625,
"learning_rate": 4.6337141122260444e-05,
"loss": 0.2029,
"step": 1800
},
{
"epoch": 0.18388702631311593,
"grad_norm": 2.359375,
"learning_rate": 4.629502228442112e-05,
"loss": 0.1489,
"step": 1810
},
{
"epoch": 0.18490297673473535,
"grad_norm": 1.4453125,
"learning_rate": 4.6252682039217045e-05,
"loss": 0.2101,
"step": 1820
},
{
"epoch": 0.18591892715635477,
"grad_norm": 2.71875,
"learning_rate": 4.621012082686573e-05,
"loss": 0.2076,
"step": 1830
},
{
"epoch": 0.1869348775779742,
"grad_norm": 3.0625,
"learning_rate": 4.616733908988216e-05,
"loss": 0.2719,
"step": 1840
},
{
"epoch": 0.18795082799959362,
"grad_norm": 1.953125,
"learning_rate": 4.612433727307409e-05,
"loss": 0.2105,
"step": 1850
},
{
"epoch": 0.18896677842121304,
"grad_norm": 3.46875,
"learning_rate": 4.608111582353751e-05,
"loss": 0.1877,
"step": 1860
},
{
"epoch": 0.18998272884283246,
"grad_norm": 2.546875,
"learning_rate": 4.603767519065197e-05,
"loss": 0.2238,
"step": 1870
},
{
"epoch": 0.19099867926445188,
"grad_norm": 1.5703125,
"learning_rate": 4.599401582607589e-05,
"loss": 0.243,
"step": 1880
},
{
"epoch": 0.19201462968607133,
"grad_norm": 2.4375,
"learning_rate": 4.595013818374185e-05,
"loss": 0.1867,
"step": 1890
},
{
"epoch": 0.19303058010769075,
"grad_norm": 2.203125,
"learning_rate": 4.5906042719851925e-05,
"loss": 0.1994,
"step": 1900
},
{
"epoch": 0.19404653052931017,
"grad_norm": 3.984375,
"learning_rate": 4.586172989287291e-05,
"loss": 0.1899,
"step": 1910
},
{
"epoch": 0.1950624809509296,
"grad_norm": 2.6875,
"learning_rate": 4.5817200163531534e-05,
"loss": 0.2528,
"step": 1920
},
{
"epoch": 0.19607843137254902,
"grad_norm": 2.71875,
"learning_rate": 4.577245399480972e-05,
"loss": 0.2336,
"step": 1930
},
{
"epoch": 0.19709438179416844,
"grad_norm": 2.640625,
"learning_rate": 4.5727491851939715e-05,
"loss": 0.2204,
"step": 1940
},
{
"epoch": 0.19811033221578786,
"grad_norm": 1.78125,
"learning_rate": 4.568231420239929e-05,
"loss": 0.1656,
"step": 1950
},
{
"epoch": 0.19912628263740728,
"grad_norm": 3.15625,
"learning_rate": 4.563692151590687e-05,
"loss": 0.2105,
"step": 1960
},
{
"epoch": 0.20014223305902673,
"grad_norm": 1.3671875,
"learning_rate": 4.5591314264416666e-05,
"loss": 0.1464,
"step": 1970
},
{
"epoch": 0.20115818348064615,
"grad_norm": 4.25,
"learning_rate": 4.554549292211371e-05,
"loss": 0.2103,
"step": 1980
},
{
"epoch": 0.20217413390226557,
"grad_norm": 2.984375,
"learning_rate": 4.549945796540901e-05,
"loss": 0.144,
"step": 1990
},
{
"epoch": 0.203190084323885,
"grad_norm": 1.859375,
"learning_rate": 4.545320987293453e-05,
"loss": 0.1963,
"step": 2000
},
{
"epoch": 0.20420603474550442,
"grad_norm": 1.078125,
"learning_rate": 4.540674912553824e-05,
"loss": 0.2115,
"step": 2010
},
{
"epoch": 0.20522198516712384,
"grad_norm": 4.25,
"learning_rate": 4.536007620627911e-05,
"loss": 0.1682,
"step": 2020
},
{
"epoch": 0.20623793558874326,
"grad_norm": 2.71875,
"learning_rate": 4.531319160042212e-05,
"loss": 0.1992,
"step": 2030
},
{
"epoch": 0.20725388601036268,
"grad_norm": 1.2890625,
"learning_rate": 4.5266095795433126e-05,
"loss": 0.1134,
"step": 2040
},
{
"epoch": 0.20826983643198213,
"grad_norm": 3.296875,
"learning_rate": 4.5218789280973925e-05,
"loss": 0.1474,
"step": 2050
},
{
"epoch": 0.20928578685360155,
"grad_norm": 1.9375,
"learning_rate": 4.5171272548897024e-05,
"loss": 0.1955,
"step": 2060
},
{
"epoch": 0.21030173727522097,
"grad_norm": 2.734375,
"learning_rate": 4.512354609324063e-05,
"loss": 0.2042,
"step": 2070
},
{
"epoch": 0.2113176876968404,
"grad_norm": 2.921875,
"learning_rate": 4.507561041022347e-05,
"loss": 0.2174,
"step": 2080
},
{
"epoch": 0.21233363811845982,
"grad_norm": 2.40625,
"learning_rate": 4.502746599823963e-05,
"loss": 0.2634,
"step": 2090
},
{
"epoch": 0.21334958854007924,
"grad_norm": 1.71875,
"learning_rate": 4.497911335785339e-05,
"loss": 0.1884,
"step": 2100
},
{
"epoch": 0.21436553896169866,
"grad_norm": 0.79296875,
"learning_rate": 4.4930552991794e-05,
"loss": 0.1872,
"step": 2110
},
{
"epoch": 0.21538148938331808,
"grad_norm": 3.171875,
"learning_rate": 4.4881785404950474e-05,
"loss": 0.2233,
"step": 2120
},
{
"epoch": 0.21639743980493753,
"grad_norm": 2.59375,
"learning_rate": 4.483281110436631e-05,
"loss": 0.2374,
"step": 2130
},
{
"epoch": 0.21741339022655695,
"grad_norm": 3.328125,
"learning_rate": 4.478363059923426e-05,
"loss": 0.2545,
"step": 2140
},
{
"epoch": 0.21842934064817637,
"grad_norm": 2.3125,
"learning_rate": 4.4734244400891014e-05,
"loss": 0.2063,
"step": 2150
},
{
"epoch": 0.2194452910697958,
"grad_norm": 3.40625,
"learning_rate": 4.4684653022811865e-05,
"loss": 0.1219,
"step": 2160
},
{
"epoch": 0.22046124149141522,
"grad_norm": 4.1875,
"learning_rate": 4.463485698060541e-05,
"loss": 0.2805,
"step": 2170
},
{
"epoch": 0.22147719191303464,
"grad_norm": 2.3125,
"learning_rate": 4.458485679200814e-05,
"loss": 0.1998,
"step": 2180
},
{
"epoch": 0.22249314233465406,
"grad_norm": 3.578125,
"learning_rate": 4.453465297687912e-05,
"loss": 0.2489,
"step": 2190
},
{
"epoch": 0.22350909275627348,
"grad_norm": 2.59375,
"learning_rate": 4.448424605719452e-05,
"loss": 0.2731,
"step": 2200
},
{
"epoch": 0.22452504317789293,
"grad_norm": 3.28125,
"learning_rate": 4.443363655704224e-05,
"loss": 0.2425,
"step": 2210
},
{
"epoch": 0.22554099359951235,
"grad_norm": 2.78125,
"learning_rate": 4.438282500261641e-05,
"loss": 0.2938,
"step": 2220
},
{
"epoch": 0.22655694402113177,
"grad_norm": 1.1953125,
"learning_rate": 4.433181192221197e-05,
"loss": 0.1728,
"step": 2230
},
{
"epoch": 0.2275728944427512,
"grad_norm": 1.34375,
"learning_rate": 4.4280597846219155e-05,
"loss": 0.216,
"step": 2240
},
{
"epoch": 0.22858884486437062,
"grad_norm": 1.8515625,
"learning_rate": 4.422918330711796e-05,
"loss": 0.1612,
"step": 2250
},
{
"epoch": 0.22960479528599004,
"grad_norm": 1.90625,
"learning_rate": 4.417756883947263e-05,
"loss": 0.107,
"step": 2260
},
{
"epoch": 0.23062074570760946,
"grad_norm": 3.375,
"learning_rate": 4.412575497992611e-05,
"loss": 0.1756,
"step": 2270
},
{
"epoch": 0.23163669612922888,
"grad_norm": 4.375,
"learning_rate": 4.407374226719445e-05,
"loss": 0.234,
"step": 2280
},
{
"epoch": 0.23265264655084833,
"grad_norm": 3.25,
"learning_rate": 4.402153124206119e-05,
"loss": 0.2144,
"step": 2290
},
{
"epoch": 0.23366859697246775,
"grad_norm": 1.703125,
"learning_rate": 4.396912244737173e-05,
"loss": 0.1696,
"step": 2300
},
{
"epoch": 0.23468454739408717,
"grad_norm": 2.84375,
"learning_rate": 4.391651642802778e-05,
"loss": 0.2506,
"step": 2310
},
{
"epoch": 0.2357004978157066,
"grad_norm": 4.5,
"learning_rate": 4.386371373098155e-05,
"loss": 0.1686,
"step": 2320
},
{
"epoch": 0.23671644823732602,
"grad_norm": 2.515625,
"learning_rate": 4.381071490523018e-05,
"loss": 0.2403,
"step": 2330
},
{
"epoch": 0.23773239865894544,
"grad_norm": 4.4375,
"learning_rate": 4.3757520501809955e-05,
"loss": 0.1611,
"step": 2340
},
{
"epoch": 0.23874834908056486,
"grad_norm": 1.609375,
"learning_rate": 4.370413107379065e-05,
"loss": 0.1698,
"step": 2350
},
{
"epoch": 0.23976429950218428,
"grad_norm": 4.96875,
"learning_rate": 4.36505471762697e-05,
"loss": 0.1928,
"step": 2360
},
{
"epoch": 0.24078024992380373,
"grad_norm": 0.8984375,
"learning_rate": 4.3596769366366474e-05,
"loss": 0.2035,
"step": 2370
},
{
"epoch": 0.24179620034542315,
"grad_norm": 5.75,
"learning_rate": 4.354279820321649e-05,
"loss": 0.16,
"step": 2380
},
{
"epoch": 0.24281215076704257,
"grad_norm": 1.9453125,
"learning_rate": 4.34886342479656e-05,
"loss": 0.1851,
"step": 2390
},
{
"epoch": 0.243828101188662,
"grad_norm": 1.015625,
"learning_rate": 4.34342780637641e-05,
"loss": 0.1726,
"step": 2400
},
{
"epoch": 0.24484405161028142,
"grad_norm": 4.59375,
"learning_rate": 4.337973021576095e-05,
"loss": 0.2847,
"step": 2410
},
{
"epoch": 0.24586000203190084,
"grad_norm": 1.03125,
"learning_rate": 4.3324991271097846e-05,
"loss": 0.2528,
"step": 2420
},
{
"epoch": 0.24687595245352026,
"grad_norm": 2.1875,
"learning_rate": 4.3270061798903374e-05,
"loss": 0.1573,
"step": 2430
},
{
"epoch": 0.24789190287513968,
"grad_norm": 0.98046875,
"learning_rate": 4.321494237028701e-05,
"loss": 0.1703,
"step": 2440
},
{
"epoch": 0.24890785329675913,
"grad_norm": 3.8125,
"learning_rate": 4.31596335583333e-05,
"loss": 0.2613,
"step": 2450
},
{
"epoch": 0.24992380371837855,
"grad_norm": 4.0625,
"learning_rate": 4.310413593809579e-05,
"loss": 0.22,
"step": 2460
},
{
"epoch": 0.250939754139998,
"grad_norm": 3.15625,
"learning_rate": 4.304845008659108e-05,
"loss": 0.1263,
"step": 2470
},
{
"epoch": 0.25195570456161737,
"grad_norm": 3.046875,
"learning_rate": 4.2992576582792895e-05,
"loss": 0.1639,
"step": 2480
},
{
"epoch": 0.2529716549832368,
"grad_norm": 9.8125,
"learning_rate": 4.293651600762595e-05,
"loss": 0.2681,
"step": 2490
},
{
"epoch": 0.25398760540485626,
"grad_norm": 3.734375,
"learning_rate": 4.288026894395999e-05,
"loss": 0.2292,
"step": 2500
},
{
"epoch": 0.25500355582647566,
"grad_norm": 0.455078125,
"learning_rate": 4.2823835976603723e-05,
"loss": 0.2324,
"step": 2510
},
{
"epoch": 0.2560195062480951,
"grad_norm": 5.625,
"learning_rate": 4.276721769229869e-05,
"loss": 0.1834,
"step": 2520
},
{
"epoch": 0.2570354566697145,
"grad_norm": 1.3671875,
"learning_rate": 4.271041467971323e-05,
"loss": 0.1826,
"step": 2530
},
{
"epoch": 0.25805140709133395,
"grad_norm": 5.0625,
"learning_rate": 4.265342752943632e-05,
"loss": 0.2463,
"step": 2540
},
{
"epoch": 0.25906735751295334,
"grad_norm": 2.859375,
"learning_rate": 4.2596256833971425e-05,
"loss": 0.2598,
"step": 2550
},
{
"epoch": 0.2600833079345728,
"grad_norm": 1.8515625,
"learning_rate": 4.2538903187730374e-05,
"loss": 0.1148,
"step": 2560
},
{
"epoch": 0.26109925835619224,
"grad_norm": 2.71875,
"learning_rate": 4.248136718702716e-05,
"loss": 0.2123,
"step": 2570
},
{
"epoch": 0.26211520877781164,
"grad_norm": 4.5625,
"learning_rate": 4.242364943007172e-05,
"loss": 0.2369,
"step": 2580
},
{
"epoch": 0.2631311591994311,
"grad_norm": 2.296875,
"learning_rate": 4.236575051696377e-05,
"loss": 0.261,
"step": 2590
},
{
"epoch": 0.2641471096210505,
"grad_norm": 2.75,
"learning_rate": 4.2307671049686514e-05,
"loss": 0.1564,
"step": 2600
},
{
"epoch": 0.26516306004266993,
"grad_norm": 3.5,
"learning_rate": 4.2249411632100396e-05,
"loss": 0.1563,
"step": 2610
},
{
"epoch": 0.2661790104642893,
"grad_norm": 2.84375,
"learning_rate": 4.219097286993684e-05,
"loss": 0.1697,
"step": 2620
},
{
"epoch": 0.26719496088590877,
"grad_norm": 2.125,
"learning_rate": 4.2132355370791946e-05,
"loss": 0.1844,
"step": 2630
},
{
"epoch": 0.26821091130752817,
"grad_norm": 4.03125,
"learning_rate": 4.2073559744120156e-05,
"loss": 0.2144,
"step": 2640
},
{
"epoch": 0.2692268617291476,
"grad_norm": 2.375,
"learning_rate": 4.201458660122793e-05,
"loss": 0.2013,
"step": 2650
},
{
"epoch": 0.27024281215076706,
"grad_norm": 3.625,
"learning_rate": 4.1955436555267393e-05,
"loss": 0.2166,
"step": 2660
},
{
"epoch": 0.27125876257238646,
"grad_norm": 0.328125,
"learning_rate": 4.189611022122997e-05,
"loss": 0.1934,
"step": 2670
},
{
"epoch": 0.2722747129940059,
"grad_norm": 2.75,
"learning_rate": 4.1836608215939944e-05,
"loss": 0.2157,
"step": 2680
},
{
"epoch": 0.2732906634156253,
"grad_norm": 3.5,
"learning_rate": 4.17769311580481e-05,
"loss": 0.18,
"step": 2690
},
{
"epoch": 0.27430661383724475,
"grad_norm": 2.109375,
"learning_rate": 4.171707966802528e-05,
"loss": 0.2178,
"step": 2700
},
{
"epoch": 0.27532256425886414,
"grad_norm": 4.65625,
"learning_rate": 4.16570543681559e-05,
"loss": 0.1896,
"step": 2710
},
{
"epoch": 0.2763385146804836,
"grad_norm": 4.8125,
"learning_rate": 4.159685588253151e-05,
"loss": 0.1322,
"step": 2720
},
{
"epoch": 0.27735446510210304,
"grad_norm": 3.9375,
"learning_rate": 4.153648483704429e-05,
"loss": 0.184,
"step": 2730
},
{
"epoch": 0.27837041552372244,
"grad_norm": 4.53125,
"learning_rate": 4.147594185938057e-05,
"loss": 0.2451,
"step": 2740
},
{
"epoch": 0.2793863659453419,
"grad_norm": 1.0390625,
"learning_rate": 4.141522757901426e-05,
"loss": 0.2367,
"step": 2750
},
{
"epoch": 0.2804023163669613,
"grad_norm": 3.375,
"learning_rate": 4.1354342627200345e-05,
"loss": 0.179,
"step": 2760
},
{
"epoch": 0.28141826678858073,
"grad_norm": 2.953125,
"learning_rate": 4.1293287636968286e-05,
"loss": 0.1396,
"step": 2770
},
{
"epoch": 0.2824342172102001,
"grad_norm": 2.546875,
"learning_rate": 4.1232063243115485e-05,
"loss": 0.1963,
"step": 2780
},
{
"epoch": 0.28345016763181957,
"grad_norm": 5.09375,
"learning_rate": 4.117067008220063e-05,
"loss": 0.2457,
"step": 2790
},
{
"epoch": 0.28446611805343897,
"grad_norm": 2.046875,
"learning_rate": 4.110910879253712e-05,
"loss": 0.2262,
"step": 2800
},
{
"epoch": 0.2854820684750584,
"grad_norm": 2.1875,
"learning_rate": 4.104738001418641e-05,
"loss": 0.2499,
"step": 2810
},
{
"epoch": 0.28649801889667786,
"grad_norm": 2.59375,
"learning_rate": 4.098548438895135e-05,
"loss": 0.1667,
"step": 2820
},
{
"epoch": 0.28751396931829726,
"grad_norm": 2.875,
"learning_rate": 4.092342256036954e-05,
"loss": 0.2288,
"step": 2830
},
{
"epoch": 0.2885299197399167,
"grad_norm": 3.015625,
"learning_rate": 4.086119517370659e-05,
"loss": 0.2038,
"step": 2840
},
{
"epoch": 0.2895458701615361,
"grad_norm": 3.53125,
"learning_rate": 4.0798802875949485e-05,
"loss": 0.181,
"step": 2850
},
{
"epoch": 0.29056182058315555,
"grad_norm": 2.296875,
"learning_rate": 4.073624631579975e-05,
"loss": 0.1886,
"step": 2860
},
{
"epoch": 0.29157777100477494,
"grad_norm": 3.609375,
"learning_rate": 4.067352614366685e-05,
"loss": 0.2053,
"step": 2870
},
{
"epoch": 0.2925937214263944,
"grad_norm": 2.328125,
"learning_rate": 4.061064301166128e-05,
"loss": 0.1409,
"step": 2880
},
{
"epoch": 0.29360967184801384,
"grad_norm": 4.9375,
"learning_rate": 4.054759757358787e-05,
"loss": 0.184,
"step": 2890
},
{
"epoch": 0.29462562226963324,
"grad_norm": 4.6875,
"learning_rate": 4.048439048493898e-05,
"loss": 0.2306,
"step": 2900
},
{
"epoch": 0.2956415726912527,
"grad_norm": 4.09375,
"learning_rate": 4.0421022402887676e-05,
"loss": 0.1914,
"step": 2910
},
{
"epoch": 0.2966575231128721,
"grad_norm": 2.3125,
"learning_rate": 4.035749398628088e-05,
"loss": 0.1653,
"step": 2920
},
{
"epoch": 0.29767347353449153,
"grad_norm": 2.515625,
"learning_rate": 4.029380589563256e-05,
"loss": 0.1941,
"step": 2930
},
{
"epoch": 0.2986894239561109,
"grad_norm": 1.78125,
"learning_rate": 4.02299587931168e-05,
"loss": 0.1117,
"step": 2940
},
{
"epoch": 0.29970537437773037,
"grad_norm": 0.8359375,
"learning_rate": 4.0165953342560974e-05,
"loss": 0.1605,
"step": 2950
},
{
"epoch": 0.30072132479934977,
"grad_norm": 3.046875,
"learning_rate": 4.010179020943884e-05,
"loss": 0.1726,
"step": 2960
},
{
"epoch": 0.3017372752209692,
"grad_norm": 3.453125,
"learning_rate": 4.003747006086357e-05,
"loss": 0.2208,
"step": 2970
},
{
"epoch": 0.30275322564258866,
"grad_norm": 2.515625,
"learning_rate": 3.9972993565580866e-05,
"loss": 0.1325,
"step": 2980
},
{
"epoch": 0.30376917606420806,
"grad_norm": 3.046875,
"learning_rate": 3.9908361393962e-05,
"loss": 0.2014,
"step": 2990
},
{
"epoch": 0.3047851264858275,
"grad_norm": 2.28125,
"learning_rate": 3.984357421799681e-05,
"loss": 0.165,
"step": 3000
}
],
"logging_steps": 10,
"max_steps": 9843,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"total_flos": 0.0,
"train_batch_size": 7,
"trial_name": null,
"trial_params": null
}