{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 3782, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00026441036488630354, "grad_norm": 3.609375, "learning_rate": 5.277044854881267e-07, "loss": 1.8784, "step": 1 }, { "epoch": 0.0013220518244315177, "grad_norm": 3.609375, "learning_rate": 2.6385224274406333e-06, "loss": 1.8748, "step": 5 }, { "epoch": 0.0026441036488630354, "grad_norm": 3.578125, "learning_rate": 5.2770448548812665e-06, "loss": 1.8755, "step": 10 }, { "epoch": 0.003966155473294554, "grad_norm": 3.5, "learning_rate": 7.915567282321901e-06, "loss": 1.8741, "step": 15 }, { "epoch": 0.005288207297726071, "grad_norm": 3.359375, "learning_rate": 1.0554089709762533e-05, "loss": 1.8644, "step": 20 }, { "epoch": 0.006610259122157589, "grad_norm": 2.9375, "learning_rate": 1.3192612137203167e-05, "loss": 1.8583, "step": 25 }, { "epoch": 0.007932310946589107, "grad_norm": 2.5625, "learning_rate": 1.5831134564643802e-05, "loss": 1.8416, "step": 30 }, { "epoch": 0.009254362771020624, "grad_norm": 1.8671875, "learning_rate": 1.8469656992084432e-05, "loss": 1.8236, "step": 35 }, { "epoch": 0.010576414595452142, "grad_norm": 1.4921875, "learning_rate": 2.1108179419525066e-05, "loss": 1.8066, "step": 40 }, { "epoch": 0.011898466419883659, "grad_norm": 1.2265625, "learning_rate": 2.37467018469657e-05, "loss": 1.7951, "step": 45 }, { "epoch": 0.013220518244315178, "grad_norm": 1.1171875, "learning_rate": 2.6385224274406334e-05, "loss": 1.7783, "step": 50 }, { "epoch": 0.014542570068746695, "grad_norm": 0.92578125, "learning_rate": 2.9023746701846964e-05, "loss": 1.7577, "step": 55 }, { "epoch": 0.015864621893178214, "grad_norm": 0.8125, "learning_rate": 3.1662269129287604e-05, "loss": 1.7413, "step": 60 }, { "epoch": 0.01718667371760973, "grad_norm": 0.6953125, "learning_rate": 3.430079155672823e-05, "loss": 1.7217, "step": 65 }, { "epoch": 0.01850872554204125, "grad_norm": 0.69921875, "learning_rate": 3.6939313984168865e-05, "loss": 1.7064, "step": 70 }, { "epoch": 0.019830777366472766, "grad_norm": 0.57421875, "learning_rate": 3.95778364116095e-05, "loss": 1.6914, "step": 75 }, { "epoch": 0.021152829190904283, "grad_norm": 0.57421875, "learning_rate": 4.221635883905013e-05, "loss": 1.679, "step": 80 }, { "epoch": 0.0224748810153358, "grad_norm": 0.55859375, "learning_rate": 4.4854881266490766e-05, "loss": 1.6632, "step": 85 }, { "epoch": 0.023796932839767318, "grad_norm": 0.58203125, "learning_rate": 4.74934036939314e-05, "loss": 1.6472, "step": 90 }, { "epoch": 0.025118984664198835, "grad_norm": 0.546875, "learning_rate": 5.0131926121372033e-05, "loss": 1.6343, "step": 95 }, { "epoch": 0.026441036488630356, "grad_norm": 0.380859375, "learning_rate": 5.277044854881267e-05, "loss": 1.6313, "step": 100 }, { "epoch": 0.027763088313061873, "grad_norm": 0.484375, "learning_rate": 5.540897097625331e-05, "loss": 1.6094, "step": 105 }, { "epoch": 0.02908514013749339, "grad_norm": 0.5234375, "learning_rate": 5.804749340369393e-05, "loss": 1.5984, "step": 110 }, { "epoch": 0.030407191961924908, "grad_norm": 0.51171875, "learning_rate": 6.068601583113457e-05, "loss": 1.5843, "step": 115 }, { "epoch": 0.03172924378635643, "grad_norm": 0.322265625, "learning_rate": 6.332453825857521e-05, "loss": 1.5689, "step": 120 }, { "epoch": 0.03305129561078794, "grad_norm": 0.44921875, "learning_rate": 6.596306068601583e-05, "loss": 1.5575, "step": 125 }, { "epoch": 0.03437334743521946, "grad_norm": 0.3203125, "learning_rate": 6.860158311345646e-05, "loss": 1.5451, "step": 130 }, { "epoch": 0.03569539925965098, "grad_norm": 0.4140625, "learning_rate": 7.124010554089711e-05, "loss": 1.5333, "step": 135 }, { "epoch": 0.0370174510840825, "grad_norm": 0.48828125, "learning_rate": 7.387862796833773e-05, "loss": 1.5237, "step": 140 }, { "epoch": 0.03833950290851401, "grad_norm": 0.5, "learning_rate": 7.651715039577836e-05, "loss": 1.5105, "step": 145 }, { "epoch": 0.03966155473294553, "grad_norm": 0.3984375, "learning_rate": 7.9155672823219e-05, "loss": 1.4946, "step": 150 }, { "epoch": 0.040983606557377046, "grad_norm": 0.39453125, "learning_rate": 8.179419525065963e-05, "loss": 1.4843, "step": 155 }, { "epoch": 0.04230565838180857, "grad_norm": 0.3515625, "learning_rate": 8.443271767810026e-05, "loss": 1.4732, "step": 160 }, { "epoch": 0.04362771020624009, "grad_norm": 0.421875, "learning_rate": 8.70712401055409e-05, "loss": 1.4611, "step": 165 }, { "epoch": 0.0449497620306716, "grad_norm": 0.31640625, "learning_rate": 8.970976253298153e-05, "loss": 1.4485, "step": 170 }, { "epoch": 0.04627181385510312, "grad_norm": 0.341796875, "learning_rate": 9.234828496042217e-05, "loss": 1.4406, "step": 175 }, { "epoch": 0.047593865679534636, "grad_norm": 0.359375, "learning_rate": 9.49868073878628e-05, "loss": 1.4236, "step": 180 }, { "epoch": 0.04891591750396616, "grad_norm": 0.32421875, "learning_rate": 9.762532981530343e-05, "loss": 1.4107, "step": 185 }, { "epoch": 0.05023796932839767, "grad_norm": 0.435546875, "learning_rate": 0.00010026385224274407, "loss": 1.4052, "step": 190 }, { "epoch": 0.05156002115282919, "grad_norm": 0.458984375, "learning_rate": 0.00010290237467018471, "loss": 1.3859, "step": 195 }, { "epoch": 0.05288207297726071, "grad_norm": 0.51953125, "learning_rate": 0.00010554089709762533, "loss": 1.3742, "step": 200 }, { "epoch": 0.054204124801692226, "grad_norm": 0.478515625, "learning_rate": 0.00010817941952506597, "loss": 1.365, "step": 205 }, { "epoch": 0.05552617662612375, "grad_norm": 0.57421875, "learning_rate": 0.00011081794195250662, "loss": 1.3482, "step": 210 }, { "epoch": 0.05684822845055526, "grad_norm": 0.412109375, "learning_rate": 0.00011345646437994724, "loss": 1.3375, "step": 215 }, { "epoch": 0.05817028027498678, "grad_norm": 0.384765625, "learning_rate": 0.00011609498680738786, "loss": 1.333, "step": 220 }, { "epoch": 0.059492332099418295, "grad_norm": 0.63671875, "learning_rate": 0.00011873350923482852, "loss": 1.3238, "step": 225 }, { "epoch": 0.060814383923849816, "grad_norm": 0.55859375, "learning_rate": 0.00012137203166226914, "loss": 1.315, "step": 230 }, { "epoch": 0.06213643574828133, "grad_norm": 0.671875, "learning_rate": 0.00012401055408970977, "loss": 1.3028, "step": 235 }, { "epoch": 0.06345848757271286, "grad_norm": 0.470703125, "learning_rate": 0.00012664907651715042, "loss": 1.2869, "step": 240 }, { "epoch": 0.06478053939714437, "grad_norm": 0.435546875, "learning_rate": 0.00012928759894459104, "loss": 1.281, "step": 245 }, { "epoch": 0.06610259122157588, "grad_norm": 0.671875, "learning_rate": 0.00013192612137203166, "loss": 1.2747, "step": 250 }, { "epoch": 0.0674246430460074, "grad_norm": 0.625, "learning_rate": 0.0001345646437994723, "loss": 1.2697, "step": 255 }, { "epoch": 0.06874669487043893, "grad_norm": 1.0859375, "learning_rate": 0.00013720316622691292, "loss": 1.2549, "step": 260 }, { "epoch": 0.07006874669487044, "grad_norm": 0.75, "learning_rate": 0.00013984168865435357, "loss": 1.246, "step": 265 }, { "epoch": 0.07139079851930195, "grad_norm": 0.451171875, "learning_rate": 0.00014248021108179422, "loss": 1.2412, "step": 270 }, { "epoch": 0.07271285034373347, "grad_norm": 0.515625, "learning_rate": 0.00014511873350923484, "loss": 1.2343, "step": 275 }, { "epoch": 0.074034902168165, "grad_norm": 0.5390625, "learning_rate": 0.00014775725593667546, "loss": 1.2205, "step": 280 }, { "epoch": 0.07535695399259651, "grad_norm": 0.8984375, "learning_rate": 0.00015039577836411608, "loss": 1.2203, "step": 285 }, { "epoch": 0.07667900581702802, "grad_norm": 0.6875, "learning_rate": 0.00015303430079155673, "loss": 1.2122, "step": 290 }, { "epoch": 0.07800105764145955, "grad_norm": 1.1015625, "learning_rate": 0.00015567282321899737, "loss": 1.2035, "step": 295 }, { "epoch": 0.07932310946589106, "grad_norm": 0.3515625, "learning_rate": 0.000158311345646438, "loss": 1.1933, "step": 300 }, { "epoch": 0.08064516129032258, "grad_norm": 0.4453125, "learning_rate": 0.00016094986807387864, "loss": 1.1926, "step": 305 }, { "epoch": 0.08196721311475409, "grad_norm": 0.404296875, "learning_rate": 0.00016358839050131926, "loss": 1.1833, "step": 310 }, { "epoch": 0.08328926493918562, "grad_norm": 0.5703125, "learning_rate": 0.00016622691292875988, "loss": 1.1753, "step": 315 }, { "epoch": 0.08461131676361713, "grad_norm": 0.6484375, "learning_rate": 0.00016886543535620053, "loss": 1.1717, "step": 320 }, { "epoch": 0.08593336858804865, "grad_norm": 0.66015625, "learning_rate": 0.00017150395778364118, "loss": 1.1684, "step": 325 }, { "epoch": 0.08725542041248018, "grad_norm": 0.58984375, "learning_rate": 0.0001741424802110818, "loss": 1.1608, "step": 330 }, { "epoch": 0.08857747223691169, "grad_norm": 0.5625, "learning_rate": 0.00017678100263852244, "loss": 1.1563, "step": 335 }, { "epoch": 0.0898995240613432, "grad_norm": 0.47265625, "learning_rate": 0.00017941952506596306, "loss": 1.1501, "step": 340 }, { "epoch": 0.09122157588577472, "grad_norm": 0.58203125, "learning_rate": 0.00018205804749340368, "loss": 1.1451, "step": 345 }, { "epoch": 0.09254362771020624, "grad_norm": 0.59765625, "learning_rate": 0.00018469656992084433, "loss": 1.1338, "step": 350 }, { "epoch": 0.09386567953463776, "grad_norm": 0.484375, "learning_rate": 0.00018733509234828498, "loss": 1.1326, "step": 355 }, { "epoch": 0.09518773135906927, "grad_norm": 0.55859375, "learning_rate": 0.0001899736147757256, "loss": 1.1286, "step": 360 }, { "epoch": 0.0965097831835008, "grad_norm": 0.62109375, "learning_rate": 0.00019261213720316625, "loss": 1.1219, "step": 365 }, { "epoch": 0.09783183500793231, "grad_norm": 0.69140625, "learning_rate": 0.00019525065963060687, "loss": 1.1238, "step": 370 }, { "epoch": 0.09915388683236383, "grad_norm": 0.57421875, "learning_rate": 0.00019788918205804749, "loss": 1.1165, "step": 375 }, { "epoch": 0.10047593865679534, "grad_norm": 0.490234375, "learning_rate": 0.00019999995738663592, "loss": 1.1111, "step": 380 }, { "epoch": 0.10179799048122687, "grad_norm": 0.6484375, "learning_rate": 0.00019999846592270624, "loss": 1.0969, "step": 385 }, { "epoch": 0.10312004230565838, "grad_norm": 0.8046875, "learning_rate": 0.00019999484382688995, "loss": 1.1008, "step": 390 }, { "epoch": 0.1044420941300899, "grad_norm": 0.65625, "learning_rate": 0.00019998909117636182, "loss": 1.0969, "step": 395 }, { "epoch": 0.10576414595452142, "grad_norm": 0.56640625, "learning_rate": 0.00019998120809369154, "loss": 1.0888, "step": 400 }, { "epoch": 0.10708619777895294, "grad_norm": 1.1875, "learning_rate": 0.0001999711947468411, "loss": 1.0926, "step": 405 }, { "epoch": 0.10840824960338445, "grad_norm": 1.3515625, "learning_rate": 0.00019995905134916143, "loss": 1.08, "step": 410 }, { "epoch": 0.10973030142781597, "grad_norm": 0.58984375, "learning_rate": 0.0001999447781593875, "loss": 1.0778, "step": 415 }, { "epoch": 0.1110523532522475, "grad_norm": 0.453125, "learning_rate": 0.00019992837548163316, "loss": 1.0771, "step": 420 }, { "epoch": 0.11237440507667901, "grad_norm": 0.70703125, "learning_rate": 0.00019990984366538442, "loss": 1.0722, "step": 425 }, { "epoch": 0.11369645690111052, "grad_norm": 0.49609375, "learning_rate": 0.00019988918310549222, "loss": 1.0628, "step": 430 }, { "epoch": 0.11501850872554203, "grad_norm": 0.45703125, "learning_rate": 0.0001998663942421637, "loss": 1.0611, "step": 435 }, { "epoch": 0.11634056054997356, "grad_norm": 0.54296875, "learning_rate": 0.00019984147756095308, "loss": 1.0627, "step": 440 }, { "epoch": 0.11766261237440508, "grad_norm": 0.78515625, "learning_rate": 0.0001998144335927513, "loss": 1.0562, "step": 445 }, { "epoch": 0.11898466419883659, "grad_norm": 0.439453125, "learning_rate": 0.00019978526291377464, "loss": 1.0543, "step": 450 }, { "epoch": 0.12030671602326812, "grad_norm": 0.392578125, "learning_rate": 0.0001997539661455524, "loss": 1.0532, "step": 455 }, { "epoch": 0.12162876784769963, "grad_norm": 0.439453125, "learning_rate": 0.00019972054395491368, "loss": 1.0477, "step": 460 }, { "epoch": 0.12295081967213115, "grad_norm": 0.59765625, "learning_rate": 0.00019968499705397331, "loss": 1.044, "step": 465 }, { "epoch": 0.12427287149656266, "grad_norm": 0.451171875, "learning_rate": 0.00019964732620011651, "loss": 1.0451, "step": 470 }, { "epoch": 0.12559492332099417, "grad_norm": 0.365234375, "learning_rate": 0.00019960753219598281, "loss": 1.0385, "step": 475 }, { "epoch": 0.12691697514542571, "grad_norm": 0.4375, "learning_rate": 0.00019956561588944897, "loss": 1.0376, "step": 480 }, { "epoch": 0.12823902696985723, "grad_norm": 0.39453125, "learning_rate": 0.00019952157817361088, "loss": 1.0396, "step": 485 }, { "epoch": 0.12956107879428874, "grad_norm": 0.484375, "learning_rate": 0.0001994754199867645, "loss": 1.0342, "step": 490 }, { "epoch": 0.13088313061872026, "grad_norm": 0.640625, "learning_rate": 0.00019942714231238604, "loss": 1.036, "step": 495 }, { "epoch": 0.13220518244315177, "grad_norm": 0.451171875, "learning_rate": 0.00019937674617911077, "loss": 1.0231, "step": 500 }, { "epoch": 0.13352723426758328, "grad_norm": 0.400390625, "learning_rate": 0.00019932423266071122, "loss": 1.0286, "step": 505 }, { "epoch": 0.1348492860920148, "grad_norm": 0.322265625, "learning_rate": 0.00019926960287607436, "loss": 1.0238, "step": 510 }, { "epoch": 0.1361713379164463, "grad_norm": 0.39453125, "learning_rate": 0.0001992128579891777, "loss": 1.0216, "step": 515 }, { "epoch": 0.13749338974087785, "grad_norm": 0.287109375, "learning_rate": 0.00019915399920906432, "loss": 1.0237, "step": 520 }, { "epoch": 0.13881544156530937, "grad_norm": 0.443359375, "learning_rate": 0.0001990930277898175, "loss": 1.0137, "step": 525 }, { "epoch": 0.14013749338974088, "grad_norm": 0.5, "learning_rate": 0.00019902994503053363, "loss": 1.0206, "step": 530 }, { "epoch": 0.1414595452141724, "grad_norm": 0.470703125, "learning_rate": 0.0001989647522752947, "loss": 1.0189, "step": 535 }, { "epoch": 0.1427815970386039, "grad_norm": 0.486328125, "learning_rate": 0.0001988974509131397, "loss": 1.0109, "step": 540 }, { "epoch": 0.14410364886303542, "grad_norm": 0.337890625, "learning_rate": 0.00019882804237803488, "loss": 1.0069, "step": 545 }, { "epoch": 0.14542570068746694, "grad_norm": 0.224609375, "learning_rate": 0.00019875652814884326, "loss": 1.0097, "step": 550 }, { "epoch": 0.14674775251189848, "grad_norm": 0.48828125, "learning_rate": 0.00019868290974929328, "loss": 1.0033, "step": 555 }, { "epoch": 0.14806980433633, "grad_norm": 0.8671875, "learning_rate": 0.00019860718874794602, "loss": 1.0034, "step": 560 }, { "epoch": 0.1493918561607615, "grad_norm": 0.69921875, "learning_rate": 0.00019852936675816209, "loss": 1.0019, "step": 565 }, { "epoch": 0.15071390798519302, "grad_norm": 0.62890625, "learning_rate": 0.000198449445438067, "loss": 1.0004, "step": 570 }, { "epoch": 0.15203595980962453, "grad_norm": 0.5078125, "learning_rate": 0.00019836742649051602, "loss": 0.998, "step": 575 }, { "epoch": 0.15335801163405605, "grad_norm": 0.69140625, "learning_rate": 0.00019828331166305785, "loss": 0.9944, "step": 580 }, { "epoch": 0.15468006345848756, "grad_norm": 0.427734375, "learning_rate": 0.00019819710274789727, "loss": 0.9951, "step": 585 }, { "epoch": 0.1560021152829191, "grad_norm": 0.5, "learning_rate": 0.0001981088015818571, "loss": 0.9962, "step": 590 }, { "epoch": 0.15732416710735062, "grad_norm": 0.7265625, "learning_rate": 0.00019801841004633906, "loss": 0.9884, "step": 595 }, { "epoch": 0.15864621893178213, "grad_norm": 0.5546875, "learning_rate": 0.00019792593006728347, "loss": 0.9918, "step": 600 }, { "epoch": 0.15996827075621364, "grad_norm": 0.796875, "learning_rate": 0.00019783136361512858, "loss": 0.993, "step": 605 }, { "epoch": 0.16129032258064516, "grad_norm": 0.6875, "learning_rate": 0.00019773471270476822, "loss": 0.9902, "step": 610 }, { "epoch": 0.16261237440507667, "grad_norm": 0.5390625, "learning_rate": 0.0001976359793955091, "loss": 0.9843, "step": 615 }, { "epoch": 0.16393442622950818, "grad_norm": 0.306640625, "learning_rate": 0.0001975351657910269, "loss": 0.988, "step": 620 }, { "epoch": 0.16525647805393973, "grad_norm": 0.54296875, "learning_rate": 0.00019743227403932134, "loss": 0.9874, "step": 625 }, { "epoch": 0.16657852987837124, "grad_norm": 0.54296875, "learning_rate": 0.00019732730633267056, "loss": 0.9782, "step": 630 }, { "epoch": 0.16790058170280275, "grad_norm": 0.6171875, "learning_rate": 0.00019722026490758424, "loss": 0.9813, "step": 635 }, { "epoch": 0.16922263352723427, "grad_norm": 0.412109375, "learning_rate": 0.00019711115204475616, "loss": 0.9793, "step": 640 }, { "epoch": 0.17054468535166578, "grad_norm": 0.36328125, "learning_rate": 0.0001969999700690154, "loss": 0.9749, "step": 645 }, { "epoch": 0.1718667371760973, "grad_norm": 0.255859375, "learning_rate": 0.00019688672134927693, "loss": 0.9768, "step": 650 }, { "epoch": 0.1731887890005288, "grad_norm": 0.38671875, "learning_rate": 0.00019677140829849112, "loss": 0.973, "step": 655 }, { "epoch": 0.17451084082496035, "grad_norm": 0.4453125, "learning_rate": 0.00019665403337359227, "loss": 0.972, "step": 660 }, { "epoch": 0.17583289264939186, "grad_norm": 0.291015625, "learning_rate": 0.00019653459907544634, "loss": 0.9759, "step": 665 }, { "epoch": 0.17715494447382338, "grad_norm": 0.3671875, "learning_rate": 0.0001964131079487976, "loss": 0.9691, "step": 670 }, { "epoch": 0.1784769962982549, "grad_norm": 0.330078125, "learning_rate": 0.0001962895625822144, "loss": 0.9681, "step": 675 }, { "epoch": 0.1797990481226864, "grad_norm": 0.48828125, "learning_rate": 0.0001961639656080342, "loss": 0.9695, "step": 680 }, { "epoch": 0.18112109994711792, "grad_norm": 0.435546875, "learning_rate": 0.00019603631970230713, "loss": 0.9695, "step": 685 }, { "epoch": 0.18244315177154943, "grad_norm": 0.369140625, "learning_rate": 0.00019590662758473934, "loss": 0.9685, "step": 690 }, { "epoch": 0.18376520359598097, "grad_norm": 0.37109375, "learning_rate": 0.0001957748920186348, "loss": 0.9681, "step": 695 }, { "epoch": 0.1850872554204125, "grad_norm": 0.3984375, "learning_rate": 0.00019564111581083657, "loss": 0.9632, "step": 700 }, { "epoch": 0.186409307244844, "grad_norm": 0.330078125, "learning_rate": 0.00019550530181166692, "loss": 0.9683, "step": 705 }, { "epoch": 0.18773135906927552, "grad_norm": 0.4453125, "learning_rate": 0.0001953674529148666, "loss": 0.9628, "step": 710 }, { "epoch": 0.18905341089370703, "grad_norm": 0.259765625, "learning_rate": 0.00019522757205753323, "loss": 0.9625, "step": 715 }, { "epoch": 0.19037546271813854, "grad_norm": 0.294921875, "learning_rate": 0.00019508566222005866, "loss": 0.9583, "step": 720 }, { "epoch": 0.19169751454257006, "grad_norm": 0.330078125, "learning_rate": 0.00019494172642606553, "loss": 0.9634, "step": 725 }, { "epoch": 0.1930195663670016, "grad_norm": 0.4140625, "learning_rate": 0.00019479576774234284, "loss": 0.9602, "step": 730 }, { "epoch": 0.1943416181914331, "grad_norm": 0.345703125, "learning_rate": 0.00019464778927878048, "loss": 0.967, "step": 735 }, { "epoch": 0.19566367001586463, "grad_norm": 0.44140625, "learning_rate": 0.00019449779418830322, "loss": 0.9549, "step": 740 }, { "epoch": 0.19698572184029614, "grad_norm": 0.279296875, "learning_rate": 0.0001943457856668033, "loss": 0.9652, "step": 745 }, { "epoch": 0.19830777366472765, "grad_norm": 0.400390625, "learning_rate": 0.00019419176695307245, "loss": 0.9541, "step": 750 }, { "epoch": 0.19962982548915917, "grad_norm": 0.419921875, "learning_rate": 0.00019403574132873276, "loss": 0.9506, "step": 755 }, { "epoch": 0.20095187731359068, "grad_norm": 0.421875, "learning_rate": 0.00019387771211816705, "loss": 0.9572, "step": 760 }, { "epoch": 0.20227392913802222, "grad_norm": 0.2265625, "learning_rate": 0.00019371768268844762, "loss": 0.9525, "step": 765 }, { "epoch": 0.20359598096245374, "grad_norm": 0.3203125, "learning_rate": 0.0001935556564492649, "loss": 0.9542, "step": 770 }, { "epoch": 0.20491803278688525, "grad_norm": 0.388671875, "learning_rate": 0.0001933916368528545, "loss": 0.9532, "step": 775 }, { "epoch": 0.20624008461131677, "grad_norm": 0.318359375, "learning_rate": 0.00019322562739392394, "loss": 0.9511, "step": 780 }, { "epoch": 0.20756213643574828, "grad_norm": 0.4296875, "learning_rate": 0.00019305763160957788, "loss": 0.9536, "step": 785 }, { "epoch": 0.2088841882601798, "grad_norm": 0.322265625, "learning_rate": 0.00019288765307924299, "loss": 0.9511, "step": 790 }, { "epoch": 0.2102062400846113, "grad_norm": 0.3125, "learning_rate": 0.00019271569542459165, "loss": 0.9497, "step": 795 }, { "epoch": 0.21152829190904285, "grad_norm": 0.259765625, "learning_rate": 0.00019254176230946462, "loss": 0.9484, "step": 800 }, { "epoch": 0.21285034373347436, "grad_norm": 0.466796875, "learning_rate": 0.00019236585743979322, "loss": 0.9442, "step": 805 }, { "epoch": 0.21417239555790588, "grad_norm": 0.3671875, "learning_rate": 0.00019218798456352012, "loss": 0.9426, "step": 810 }, { "epoch": 0.2154944473823374, "grad_norm": 0.40234375, "learning_rate": 0.00019200814747051976, "loss": 0.9455, "step": 815 }, { "epoch": 0.2168164992067689, "grad_norm": 0.349609375, "learning_rate": 0.00019182634999251728, "loss": 0.9419, "step": 820 }, { "epoch": 0.21813855103120042, "grad_norm": 0.29296875, "learning_rate": 0.00019164259600300723, "loss": 0.947, "step": 825 }, { "epoch": 0.21946060285563193, "grad_norm": 0.3359375, "learning_rate": 0.00019145688941717075, "loss": 0.9478, "step": 830 }, { "epoch": 0.22078265468006344, "grad_norm": 0.396484375, "learning_rate": 0.0001912692341917922, "loss": 0.9433, "step": 835 }, { "epoch": 0.222104706504495, "grad_norm": 0.310546875, "learning_rate": 0.00019107963432517505, "loss": 0.9461, "step": 840 }, { "epoch": 0.2234267583289265, "grad_norm": 0.416015625, "learning_rate": 0.00019088809385705646, "loss": 0.9421, "step": 845 }, { "epoch": 0.22474881015335801, "grad_norm": 0.439453125, "learning_rate": 0.00019069461686852134, "loss": 0.9436, "step": 850 }, { "epoch": 0.22607086197778953, "grad_norm": 0.26171875, "learning_rate": 0.0001904992074819153, "loss": 0.9381, "step": 855 }, { "epoch": 0.22739291380222104, "grad_norm": 0.2470703125, "learning_rate": 0.00019030186986075703, "loss": 0.9408, "step": 860 }, { "epoch": 0.22871496562665256, "grad_norm": 0.3515625, "learning_rate": 0.0001901026082096492, "loss": 0.9371, "step": 865 }, { "epoch": 0.23003701745108407, "grad_norm": 0.404296875, "learning_rate": 0.00018990142677418923, "loss": 0.9387, "step": 870 }, { "epoch": 0.2313590692755156, "grad_norm": 0.455078125, "learning_rate": 0.00018969832984087873, "loss": 0.9372, "step": 875 }, { "epoch": 0.23268112109994712, "grad_norm": 0.265625, "learning_rate": 0.00018949332173703206, "loss": 0.936, "step": 880 }, { "epoch": 0.23400317292437864, "grad_norm": 0.298828125, "learning_rate": 0.0001892864068306843, "loss": 0.9405, "step": 885 }, { "epoch": 0.23532522474881015, "grad_norm": 0.373046875, "learning_rate": 0.00018907758953049805, "loss": 0.9333, "step": 890 }, { "epoch": 0.23664727657324167, "grad_norm": 0.3984375, "learning_rate": 0.00018886687428566954, "loss": 0.9356, "step": 895 }, { "epoch": 0.23796932839767318, "grad_norm": 0.50390625, "learning_rate": 0.00018865426558583383, "loss": 0.9344, "step": 900 }, { "epoch": 0.2392913802221047, "grad_norm": 0.291015625, "learning_rate": 0.00018843976796096917, "loss": 0.9355, "step": 905 }, { "epoch": 0.24061343204653624, "grad_norm": 0.353515625, "learning_rate": 0.00018822338598130047, "loss": 0.9336, "step": 910 }, { "epoch": 0.24193548387096775, "grad_norm": 0.2490234375, "learning_rate": 0.0001880051242572019, "loss": 0.9304, "step": 915 }, { "epoch": 0.24325753569539926, "grad_norm": 0.5703125, "learning_rate": 0.00018778498743909873, "loss": 0.931, "step": 920 }, { "epoch": 0.24457958751983078, "grad_norm": 0.259765625, "learning_rate": 0.00018756298021736808, "loss": 0.934, "step": 925 }, { "epoch": 0.2459016393442623, "grad_norm": 0.2890625, "learning_rate": 0.00018733910732223925, "loss": 0.9273, "step": 930 }, { "epoch": 0.2472236911686938, "grad_norm": 0.279296875, "learning_rate": 0.00018711337352369264, "loss": 0.9315, "step": 935 }, { "epoch": 0.24854574299312532, "grad_norm": 0.341796875, "learning_rate": 0.0001868857836313584, "loss": 0.9316, "step": 940 }, { "epoch": 0.24986779481755686, "grad_norm": 0.2578125, "learning_rate": 0.00018665634249441366, "loss": 0.9319, "step": 945 }, { "epoch": 0.25118984664198835, "grad_norm": 0.337890625, "learning_rate": 0.0001864250550014795, "loss": 0.9297, "step": 950 }, { "epoch": 0.25251189846641986, "grad_norm": 0.43359375, "learning_rate": 0.00018619192608051655, "loss": 0.9237, "step": 955 }, { "epoch": 0.25383395029085143, "grad_norm": 0.255859375, "learning_rate": 0.00018595696069872013, "loss": 0.9293, "step": 960 }, { "epoch": 0.25515600211528294, "grad_norm": 0.279296875, "learning_rate": 0.00018572016386241442, "loss": 0.9308, "step": 965 }, { "epoch": 0.25647805393971446, "grad_norm": 0.546875, "learning_rate": 0.00018548154061694572, "loss": 0.9207, "step": 970 }, { "epoch": 0.25780010576414597, "grad_norm": 0.369140625, "learning_rate": 0.00018524109604657496, "loss": 0.9285, "step": 975 }, { "epoch": 0.2591221575885775, "grad_norm": 0.3359375, "learning_rate": 0.00018499883527436947, "loss": 0.9268, "step": 980 }, { "epoch": 0.260444209413009, "grad_norm": 0.271484375, "learning_rate": 0.0001847547634620936, "loss": 0.9284, "step": 985 }, { "epoch": 0.2617662612374405, "grad_norm": 0.4296875, "learning_rate": 0.00018450888581009908, "loss": 0.9264, "step": 990 }, { "epoch": 0.263088313061872, "grad_norm": 0.359375, "learning_rate": 0.00018426120755721386, "loss": 0.926, "step": 995 }, { "epoch": 0.26441036488630354, "grad_norm": 0.345703125, "learning_rate": 0.0001840117339806308, "loss": 0.9218, "step": 1000 }, { "epoch": 0.26573241671073505, "grad_norm": 0.326171875, "learning_rate": 0.00018376047039579495, "loss": 0.9245, "step": 1005 }, { "epoch": 0.26705446853516657, "grad_norm": 0.2578125, "learning_rate": 0.00018350742215629055, "loss": 0.9221, "step": 1010 }, { "epoch": 0.2683765203595981, "grad_norm": 0.42578125, "learning_rate": 0.00018325259465372677, "loss": 0.926, "step": 1015 }, { "epoch": 0.2696985721840296, "grad_norm": 0.4140625, "learning_rate": 0.0001829959933176229, "loss": 0.9235, "step": 1020 }, { "epoch": 0.2710206240084611, "grad_norm": 0.4375, "learning_rate": 0.00018273762361529274, "loss": 0.9237, "step": 1025 }, { "epoch": 0.2723426758328926, "grad_norm": 0.2314453125, "learning_rate": 0.000182477491051728, "loss": 0.9203, "step": 1030 }, { "epoch": 0.2736647276573242, "grad_norm": 0.396484375, "learning_rate": 0.00018221560116948103, "loss": 0.9206, "step": 1035 }, { "epoch": 0.2749867794817557, "grad_norm": 0.345703125, "learning_rate": 0.00018195195954854676, "loss": 0.924, "step": 1040 }, { "epoch": 0.2763088313061872, "grad_norm": 0.53125, "learning_rate": 0.00018168657180624384, "loss": 0.918, "step": 1045 }, { "epoch": 0.27763088313061873, "grad_norm": 0.337890625, "learning_rate": 0.0001814194435970949, "loss": 0.9298, "step": 1050 }, { "epoch": 0.27895293495505025, "grad_norm": 0.7421875, "learning_rate": 0.00018115058061270598, "loss": 0.9195, "step": 1055 }, { "epoch": 0.28027498677948176, "grad_norm": 0.57421875, "learning_rate": 0.0001808799885816455, "loss": 0.9196, "step": 1060 }, { "epoch": 0.2815970386039133, "grad_norm": 0.39453125, "learning_rate": 0.00018060767326932194, "loss": 0.9154, "step": 1065 }, { "epoch": 0.2829190904283448, "grad_norm": 0.38671875, "learning_rate": 0.00018033364047786128, "loss": 0.9186, "step": 1070 }, { "epoch": 0.2842411422527763, "grad_norm": 0.400390625, "learning_rate": 0.00018005789604598303, "loss": 0.9173, "step": 1075 }, { "epoch": 0.2855631940772078, "grad_norm": 0.384765625, "learning_rate": 0.0001797804458488762, "loss": 0.9199, "step": 1080 }, { "epoch": 0.28688524590163933, "grad_norm": 0.71875, "learning_rate": 0.00017950129579807374, "loss": 0.9202, "step": 1085 }, { "epoch": 0.28820729772607084, "grad_norm": 0.55859375, "learning_rate": 0.00017922045184132698, "loss": 0.922, "step": 1090 }, { "epoch": 0.28952934955050236, "grad_norm": 0.640625, "learning_rate": 0.00017893791996247856, "loss": 0.9157, "step": 1095 }, { "epoch": 0.29085140137493387, "grad_norm": 0.263671875, "learning_rate": 0.00017865370618133511, "loss": 0.9145, "step": 1100 }, { "epoch": 0.29217345319936544, "grad_norm": 0.306640625, "learning_rate": 0.00017836781655353905, "loss": 0.9114, "step": 1105 }, { "epoch": 0.29349550502379695, "grad_norm": 0.314453125, "learning_rate": 0.00017808025717043938, "loss": 0.9128, "step": 1110 }, { "epoch": 0.29481755684822847, "grad_norm": 0.2392578125, "learning_rate": 0.00017779103415896193, "loss": 0.9182, "step": 1115 }, { "epoch": 0.29613960867266, "grad_norm": 0.421875, "learning_rate": 0.000177500153681479, "loss": 0.9127, "step": 1120 }, { "epoch": 0.2974616604970915, "grad_norm": 0.59765625, "learning_rate": 0.00017720762193567787, "loss": 0.9122, "step": 1125 }, { "epoch": 0.298783712321523, "grad_norm": 0.255859375, "learning_rate": 0.00017691344515442879, "loss": 0.9117, "step": 1130 }, { "epoch": 0.3001057641459545, "grad_norm": 0.2333984375, "learning_rate": 0.00017661762960565223, "loss": 0.9097, "step": 1135 }, { "epoch": 0.30142781597038604, "grad_norm": 0.376953125, "learning_rate": 0.0001763201815921853, "loss": 0.9085, "step": 1140 }, { "epoch": 0.30274986779481755, "grad_norm": 0.2412109375, "learning_rate": 0.0001760211074516474, "loss": 0.915, "step": 1145 }, { "epoch": 0.30407191961924906, "grad_norm": 0.291015625, "learning_rate": 0.00017572041355630536, "loss": 0.9145, "step": 1150 }, { "epoch": 0.3053939714436806, "grad_norm": 0.2294921875, "learning_rate": 0.00017541810631293742, "loss": 0.9144, "step": 1155 }, { "epoch": 0.3067160232681121, "grad_norm": 0.25, "learning_rate": 0.00017511419216269695, "loss": 0.9129, "step": 1160 }, { "epoch": 0.3080380750925436, "grad_norm": 0.2490234375, "learning_rate": 0.00017480867758097506, "loss": 0.9126, "step": 1165 }, { "epoch": 0.3093601269169751, "grad_norm": 0.2431640625, "learning_rate": 0.00017450156907726272, "loss": 0.9139, "step": 1170 }, { "epoch": 0.3106821787414067, "grad_norm": 0.2421875, "learning_rate": 0.00017419287319501197, "loss": 0.9064, "step": 1175 }, { "epoch": 0.3120042305658382, "grad_norm": 0.2578125, "learning_rate": 0.00017388259651149673, "loss": 0.9165, "step": 1180 }, { "epoch": 0.3133262823902697, "grad_norm": 0.263671875, "learning_rate": 0.00017357074563767225, "loss": 0.9102, "step": 1185 }, { "epoch": 0.31464833421470123, "grad_norm": 0.2412109375, "learning_rate": 0.00017325732721803466, "loss": 0.9102, "step": 1190 }, { "epoch": 0.31597038603913274, "grad_norm": 0.302734375, "learning_rate": 0.0001729423479304792, "loss": 0.9092, "step": 1195 }, { "epoch": 0.31729243786356426, "grad_norm": 0.30078125, "learning_rate": 0.00017262581448615794, "loss": 0.9076, "step": 1200 }, { "epoch": 0.31861448968799577, "grad_norm": 0.265625, "learning_rate": 0.00017230773362933687, "loss": 0.9073, "step": 1205 }, { "epoch": 0.3199365415124273, "grad_norm": 0.294921875, "learning_rate": 0.0001719881121372521, "loss": 0.9074, "step": 1210 }, { "epoch": 0.3212585933368588, "grad_norm": 0.2197265625, "learning_rate": 0.00017166695681996555, "loss": 0.912, "step": 1215 }, { "epoch": 0.3225806451612903, "grad_norm": 0.361328125, "learning_rate": 0.00017134427452021974, "loss": 0.9061, "step": 1220 }, { "epoch": 0.3239026969857218, "grad_norm": 0.2890625, "learning_rate": 0.00017102007211329214, "loss": 0.9084, "step": 1225 }, { "epoch": 0.32522474881015334, "grad_norm": 0.251953125, "learning_rate": 0.0001706943565068486, "loss": 0.9091, "step": 1230 }, { "epoch": 0.32654680063458486, "grad_norm": 0.2080078125, "learning_rate": 0.00017036713464079612, "loss": 0.9092, "step": 1235 }, { "epoch": 0.32786885245901637, "grad_norm": 0.326171875, "learning_rate": 0.0001700384134871351, "loss": 0.9069, "step": 1240 }, { "epoch": 0.32919090428344794, "grad_norm": 0.310546875, "learning_rate": 0.00016970820004981067, "loss": 0.9082, "step": 1245 }, { "epoch": 0.33051295610787945, "grad_norm": 0.33203125, "learning_rate": 0.00016937650136456358, "loss": 0.9086, "step": 1250 }, { "epoch": 0.33183500793231097, "grad_norm": 0.1875, "learning_rate": 0.0001690433244987802, "loss": 0.9041, "step": 1255 }, { "epoch": 0.3331570597567425, "grad_norm": 0.2578125, "learning_rate": 0.000168708676551342, "loss": 0.9098, "step": 1260 }, { "epoch": 0.334479111581174, "grad_norm": 0.30859375, "learning_rate": 0.00016837256465247418, "loss": 0.9024, "step": 1265 }, { "epoch": 0.3358011634056055, "grad_norm": 0.2412109375, "learning_rate": 0.00016803499596359392, "loss": 0.9061, "step": 1270 }, { "epoch": 0.337123215230037, "grad_norm": 0.232421875, "learning_rate": 0.0001676959776771577, "loss": 0.9094, "step": 1275 }, { "epoch": 0.33844526705446853, "grad_norm": 0.234375, "learning_rate": 0.00016735551701650803, "loss": 0.9035, "step": 1280 }, { "epoch": 0.33976731887890005, "grad_norm": 0.294921875, "learning_rate": 0.00016701362123571959, "loss": 0.9071, "step": 1285 }, { "epoch": 0.34108937070333156, "grad_norm": 0.421875, "learning_rate": 0.00016667029761944466, "loss": 0.9064, "step": 1290 }, { "epoch": 0.3424114225277631, "grad_norm": 0.2373046875, "learning_rate": 0.00016632555348275788, "loss": 0.9043, "step": 1295 }, { "epoch": 0.3437334743521946, "grad_norm": 0.32421875, "learning_rate": 0.00016597939617100046, "loss": 0.9039, "step": 1300 }, { "epoch": 0.3450555261766261, "grad_norm": 0.482421875, "learning_rate": 0.00016563183305962363, "loss": 0.9042, "step": 1305 }, { "epoch": 0.3463775780010576, "grad_norm": 0.310546875, "learning_rate": 0.0001652828715540314, "loss": 0.9026, "step": 1310 }, { "epoch": 0.34769962982548913, "grad_norm": 0.29296875, "learning_rate": 0.00016493251908942302, "loss": 0.9043, "step": 1315 }, { "epoch": 0.3490216816499207, "grad_norm": 0.392578125, "learning_rate": 0.0001645807831306343, "loss": 0.9035, "step": 1320 }, { "epoch": 0.3503437334743522, "grad_norm": 0.384765625, "learning_rate": 0.00016422767117197867, "loss": 0.9016, "step": 1325 }, { "epoch": 0.35166578529878373, "grad_norm": 0.3125, "learning_rate": 0.0001638731907370876, "loss": 0.8989, "step": 1330 }, { "epoch": 0.35298783712321524, "grad_norm": 0.271484375, "learning_rate": 0.00016351734937875007, "loss": 0.8979, "step": 1335 }, { "epoch": 0.35430988894764676, "grad_norm": 0.3125, "learning_rate": 0.00016316015467875188, "loss": 0.904, "step": 1340 }, { "epoch": 0.35563194077207827, "grad_norm": 0.326171875, "learning_rate": 0.00016280161424771396, "loss": 0.8991, "step": 1345 }, { "epoch": 0.3569539925965098, "grad_norm": 0.27734375, "learning_rate": 0.0001624417357249302, "loss": 0.901, "step": 1350 }, { "epoch": 0.3582760444209413, "grad_norm": 0.271484375, "learning_rate": 0.00016208052677820484, "loss": 0.904, "step": 1355 }, { "epoch": 0.3595980962453728, "grad_norm": 0.2470703125, "learning_rate": 0.00016171799510368895, "loss": 0.8977, "step": 1360 }, { "epoch": 0.3609201480698043, "grad_norm": 0.296875, "learning_rate": 0.00016135414842571643, "loss": 0.9002, "step": 1365 }, { "epoch": 0.36224219989423584, "grad_norm": 0.271484375, "learning_rate": 0.0001609889944966396, "loss": 0.8996, "step": 1370 }, { "epoch": 0.36356425171866735, "grad_norm": 0.27734375, "learning_rate": 0.0001606225410966638, "loss": 0.9047, "step": 1375 }, { "epoch": 0.36488630354309887, "grad_norm": 0.33984375, "learning_rate": 0.0001602547960336819, "loss": 0.9004, "step": 1380 }, { "epoch": 0.3662083553675304, "grad_norm": 0.3125, "learning_rate": 0.00015988576714310755, "loss": 0.902, "step": 1385 }, { "epoch": 0.36753040719196195, "grad_norm": 0.25, "learning_rate": 0.00015951546228770868, "loss": 0.8999, "step": 1390 }, { "epoch": 0.36885245901639346, "grad_norm": 0.248046875, "learning_rate": 0.0001591438893574396, "loss": 0.9047, "step": 1395 }, { "epoch": 0.370174510840825, "grad_norm": 0.2451171875, "learning_rate": 0.0001587710562692731, "loss": 0.8982, "step": 1400 }, { "epoch": 0.3714965626652565, "grad_norm": 0.26171875, "learning_rate": 0.00015839697096703176, "loss": 0.9009, "step": 1405 }, { "epoch": 0.372818614489688, "grad_norm": 0.359375, "learning_rate": 0.00015802164142121854, "loss": 0.9023, "step": 1410 }, { "epoch": 0.3741406663141195, "grad_norm": 0.392578125, "learning_rate": 0.0001576450756288471, "loss": 0.9018, "step": 1415 }, { "epoch": 0.37546271813855103, "grad_norm": 0.341796875, "learning_rate": 0.0001572672816132714, "loss": 0.9006, "step": 1420 }, { "epoch": 0.37678476996298255, "grad_norm": 0.212890625, "learning_rate": 0.00015688826742401465, "loss": 0.8996, "step": 1425 }, { "epoch": 0.37810682178741406, "grad_norm": 0.2314453125, "learning_rate": 0.00015650804113659793, "loss": 0.8957, "step": 1430 }, { "epoch": 0.3794288736118456, "grad_norm": 0.369140625, "learning_rate": 0.00015612661085236807, "loss": 0.901, "step": 1435 }, { "epoch": 0.3807509254362771, "grad_norm": 0.2265625, "learning_rate": 0.00015574398469832493, "loss": 0.8984, "step": 1440 }, { "epoch": 0.3820729772607086, "grad_norm": 0.287109375, "learning_rate": 0.00015536017082694846, "loss": 0.8983, "step": 1445 }, { "epoch": 0.3833950290851401, "grad_norm": 0.2392578125, "learning_rate": 0.00015497517741602486, "loss": 0.8992, "step": 1450 }, { "epoch": 0.38471708090957163, "grad_norm": 0.2158203125, "learning_rate": 0.00015458901266847232, "loss": 0.8988, "step": 1455 }, { "epoch": 0.3860391327340032, "grad_norm": 0.369140625, "learning_rate": 0.0001542016848121663, "loss": 0.8982, "step": 1460 }, { "epoch": 0.3873611845584347, "grad_norm": 0.314453125, "learning_rate": 0.00015381320209976426, "loss": 0.9009, "step": 1465 }, { "epoch": 0.3886832363828662, "grad_norm": 0.3203125, "learning_rate": 0.0001534235728085297, "loss": 0.9006, "step": 1470 }, { "epoch": 0.39000528820729774, "grad_norm": 0.2216796875, "learning_rate": 0.00015303280524015585, "loss": 0.9001, "step": 1475 }, { "epoch": 0.39132734003172925, "grad_norm": 0.228515625, "learning_rate": 0.0001526409077205889, "loss": 0.894, "step": 1480 }, { "epoch": 0.39264939185616077, "grad_norm": 0.244140625, "learning_rate": 0.00015224788859985043, "loss": 0.8955, "step": 1485 }, { "epoch": 0.3939714436805923, "grad_norm": 0.322265625, "learning_rate": 0.00015185375625185964, "loss": 0.902, "step": 1490 }, { "epoch": 0.3952934955050238, "grad_norm": 0.322265625, "learning_rate": 0.00015145851907425484, "loss": 0.8995, "step": 1495 }, { "epoch": 0.3966155473294553, "grad_norm": 0.1962890625, "learning_rate": 0.0001510621854882145, "loss": 0.8997, "step": 1500 }, { "epoch": 0.3979375991538868, "grad_norm": 0.2158203125, "learning_rate": 0.00015066476393827799, "loss": 0.8915, "step": 1505 }, { "epoch": 0.39925965097831834, "grad_norm": 0.2421875, "learning_rate": 0.00015026626289216542, "loss": 0.8916, "step": 1510 }, { "epoch": 0.40058170280274985, "grad_norm": 0.3125, "learning_rate": 0.0001498666908405975, "loss": 0.8961, "step": 1515 }, { "epoch": 0.40190375462718136, "grad_norm": 0.3359375, "learning_rate": 0.00014946605629711425, "loss": 0.8991, "step": 1520 }, { "epoch": 0.4032258064516129, "grad_norm": 0.306640625, "learning_rate": 0.00014906436779789402, "loss": 0.8932, "step": 1525 }, { "epoch": 0.40454785827604445, "grad_norm": 0.267578125, "learning_rate": 0.00014866163390157136, "loss": 0.8952, "step": 1530 }, { "epoch": 0.40586991010047596, "grad_norm": 0.51953125, "learning_rate": 0.00014825786318905469, "loss": 0.893, "step": 1535 }, { "epoch": 0.4071919619249075, "grad_norm": 0.43359375, "learning_rate": 0.00014785306426334354, "loss": 0.8961, "step": 1540 }, { "epoch": 0.408514013749339, "grad_norm": 0.244140625, "learning_rate": 0.0001474472457493452, "loss": 0.8916, "step": 1545 }, { "epoch": 0.4098360655737705, "grad_norm": 0.263671875, "learning_rate": 0.00014704041629369106, "loss": 0.9021, "step": 1550 }, { "epoch": 0.411158117398202, "grad_norm": 0.265625, "learning_rate": 0.00014663258456455212, "loss": 0.8992, "step": 1555 }, { "epoch": 0.41248016922263353, "grad_norm": 0.251953125, "learning_rate": 0.00014622375925145455, "loss": 0.8926, "step": 1560 }, { "epoch": 0.41380222104706504, "grad_norm": 0.25, "learning_rate": 0.0001458139490650945, "loss": 0.894, "step": 1565 }, { "epoch": 0.41512427287149656, "grad_norm": 0.33984375, "learning_rate": 0.0001454031627371524, "loss": 0.8948, "step": 1570 }, { "epoch": 0.41644632469592807, "grad_norm": 0.26953125, "learning_rate": 0.000144991409020107, "loss": 0.8933, "step": 1575 }, { "epoch": 0.4177683765203596, "grad_norm": 0.287109375, "learning_rate": 0.00014457869668704895, "loss": 0.9002, "step": 1580 }, { "epoch": 0.4190904283447911, "grad_norm": 0.201171875, "learning_rate": 0.0001441650345314936, "loss": 0.8961, "step": 1585 }, { "epoch": 0.4204124801692226, "grad_norm": 0.271484375, "learning_rate": 0.00014375043136719402, "loss": 0.8939, "step": 1590 }, { "epoch": 0.4217345319936541, "grad_norm": 0.2275390625, "learning_rate": 0.00014333489602795293, "loss": 0.8915, "step": 1595 }, { "epoch": 0.4230565838180857, "grad_norm": 0.265625, "learning_rate": 0.0001429184373674346, "loss": 0.8952, "step": 1600 }, { "epoch": 0.4243786356425172, "grad_norm": 0.2470703125, "learning_rate": 0.0001425010642589762, "loss": 0.8931, "step": 1605 }, { "epoch": 0.4257006874669487, "grad_norm": 0.283203125, "learning_rate": 0.0001420827855953986, "loss": 0.891, "step": 1610 }, { "epoch": 0.42702273929138024, "grad_norm": 0.267578125, "learning_rate": 0.00014166361028881716, "loss": 0.891, "step": 1615 }, { "epoch": 0.42834479111581175, "grad_norm": 0.2333984375, "learning_rate": 0.00014124354727045163, "loss": 0.8917, "step": 1620 }, { "epoch": 0.42966684294024327, "grad_norm": 0.2177734375, "learning_rate": 0.00014082260549043592, "loss": 0.8897, "step": 1625 }, { "epoch": 0.4309888947646748, "grad_norm": 0.34375, "learning_rate": 0.0001404007939176274, "loss": 0.895, "step": 1630 }, { "epoch": 0.4323109465891063, "grad_norm": 0.2060546875, "learning_rate": 0.00013997812153941584, "loss": 0.8965, "step": 1635 }, { "epoch": 0.4336329984135378, "grad_norm": 0.2314453125, "learning_rate": 0.00013955459736153183, "loss": 0.8954, "step": 1640 }, { "epoch": 0.4349550502379693, "grad_norm": 0.234375, "learning_rate": 0.00013913023040785502, "loss": 0.8982, "step": 1645 }, { "epoch": 0.43627710206240083, "grad_norm": 0.30859375, "learning_rate": 0.00013870502972022173, "loss": 0.8906, "step": 1650 }, { "epoch": 0.43759915388683235, "grad_norm": 0.29296875, "learning_rate": 0.00013827900435823236, "loss": 0.8945, "step": 1655 }, { "epoch": 0.43892120571126386, "grad_norm": 0.271484375, "learning_rate": 0.00013785216339905838, "loss": 0.8904, "step": 1660 }, { "epoch": 0.4402432575356954, "grad_norm": 0.41796875, "learning_rate": 0.0001374245159372489, "loss": 0.8924, "step": 1665 }, { "epoch": 0.4415653093601269, "grad_norm": 0.35546875, "learning_rate": 0.00013699607108453684, "loss": 0.8855, "step": 1670 }, { "epoch": 0.44288736118455846, "grad_norm": 0.26171875, "learning_rate": 0.00013656683796964495, "loss": 0.8914, "step": 1675 }, { "epoch": 0.44420941300899, "grad_norm": 0.2451171875, "learning_rate": 0.00013613682573809113, "loss": 0.8931, "step": 1680 }, { "epoch": 0.4455314648334215, "grad_norm": 0.359375, "learning_rate": 0.0001357060435519936, "loss": 0.8926, "step": 1685 }, { "epoch": 0.446853516657853, "grad_norm": 0.25390625, "learning_rate": 0.00013527450058987585, "loss": 0.8915, "step": 1690 }, { "epoch": 0.4481755684822845, "grad_norm": 0.197265625, "learning_rate": 0.0001348422060464709, "loss": 0.8878, "step": 1695 }, { "epoch": 0.44949762030671603, "grad_norm": 0.23046875, "learning_rate": 0.00013440916913252536, "loss": 0.89, "step": 1700 }, { "epoch": 0.45081967213114754, "grad_norm": 0.1767578125, "learning_rate": 0.0001339753990746034, "loss": 0.8881, "step": 1705 }, { "epoch": 0.45214172395557906, "grad_norm": 0.287109375, "learning_rate": 0.00013354090511488994, "loss": 0.8911, "step": 1710 }, { "epoch": 0.45346377578001057, "grad_norm": 0.283203125, "learning_rate": 0.00013310569651099388, "loss": 0.8904, "step": 1715 }, { "epoch": 0.4547858276044421, "grad_norm": 0.2314453125, "learning_rate": 0.00013266978253575074, "loss": 0.8902, "step": 1720 }, { "epoch": 0.4561078794288736, "grad_norm": 0.29296875, "learning_rate": 0.00013223317247702517, "loss": 0.8923, "step": 1725 }, { "epoch": 0.4574299312533051, "grad_norm": 0.244140625, "learning_rate": 0.000131795875637513, "loss": 0.8918, "step": 1730 }, { "epoch": 0.4587519830777366, "grad_norm": 0.34375, "learning_rate": 0.00013135790133454305, "loss": 0.8906, "step": 1735 }, { "epoch": 0.46007403490216814, "grad_norm": 0.31640625, "learning_rate": 0.00013091925889987865, "loss": 0.8894, "step": 1740 }, { "epoch": 0.4613960867265997, "grad_norm": 0.23046875, "learning_rate": 0.00013047995767951883, "loss": 0.8945, "step": 1745 }, { "epoch": 0.4627181385510312, "grad_norm": 0.306640625, "learning_rate": 0.00013004000703349891, "loss": 0.8918, "step": 1750 }, { "epoch": 0.46404019037546274, "grad_norm": 0.29296875, "learning_rate": 0.00012959941633569153, "loss": 0.8951, "step": 1755 }, { "epoch": 0.46536224219989425, "grad_norm": 0.1943359375, "learning_rate": 0.00012915819497360658, "loss": 0.8887, "step": 1760 }, { "epoch": 0.46668429402432576, "grad_norm": 0.251953125, "learning_rate": 0.00012871635234819136, "loss": 0.8897, "step": 1765 }, { "epoch": 0.4680063458487573, "grad_norm": 0.25390625, "learning_rate": 0.00012827389787363006, "loss": 0.892, "step": 1770 }, { "epoch": 0.4693283976731888, "grad_norm": 0.2119140625, "learning_rate": 0.00012783084097714347, "loss": 0.8879, "step": 1775 }, { "epoch": 0.4706504494976203, "grad_norm": 0.240234375, "learning_rate": 0.00012738719109878794, "loss": 0.891, "step": 1780 }, { "epoch": 0.4719725013220518, "grad_norm": 0.265625, "learning_rate": 0.00012694295769125424, "loss": 0.8906, "step": 1785 }, { "epoch": 0.47329455314648333, "grad_norm": 0.271484375, "learning_rate": 0.0001264981502196662, "loss": 0.8911, "step": 1790 }, { "epoch": 0.47461660497091485, "grad_norm": 0.26171875, "learning_rate": 0.000126052778161379, "loss": 0.895, "step": 1795 }, { "epoch": 0.47593865679534636, "grad_norm": 0.1806640625, "learning_rate": 0.00012560685100577738, "loss": 0.8892, "step": 1800 }, { "epoch": 0.4772607086197779, "grad_norm": 0.275390625, "learning_rate": 0.0001251603782540732, "loss": 0.8869, "step": 1805 }, { "epoch": 0.4785827604442094, "grad_norm": 0.197265625, "learning_rate": 0.0001247133694191032, "loss": 0.8881, "step": 1810 }, { "epoch": 0.47990481226864096, "grad_norm": 0.2216796875, "learning_rate": 0.0001242658340251263, "loss": 0.8821, "step": 1815 }, { "epoch": 0.48122686409307247, "grad_norm": 0.1923828125, "learning_rate": 0.00012381778160762058, "loss": 0.8936, "step": 1820 }, { "epoch": 0.482548915917504, "grad_norm": 0.208984375, "learning_rate": 0.00012336922171308018, "loss": 0.8916, "step": 1825 }, { "epoch": 0.4838709677419355, "grad_norm": 0.23828125, "learning_rate": 0.00012292016389881184, "loss": 0.8879, "step": 1830 }, { "epoch": 0.485193019566367, "grad_norm": 0.240234375, "learning_rate": 0.00012247061773273135, "loss": 0.8883, "step": 1835 }, { "epoch": 0.4865150713907985, "grad_norm": 0.3046875, "learning_rate": 0.00012202059279315954, "loss": 0.8888, "step": 1840 }, { "epoch": 0.48783712321523004, "grad_norm": 0.365234375, "learning_rate": 0.00012157009866861842, "loss": 0.8884, "step": 1845 }, { "epoch": 0.48915917503966155, "grad_norm": 0.2314453125, "learning_rate": 0.00012111914495762669, "loss": 0.8899, "step": 1850 }, { "epoch": 0.49048122686409307, "grad_norm": 0.427734375, "learning_rate": 0.00012066774126849529, "loss": 0.8879, "step": 1855 }, { "epoch": 0.4918032786885246, "grad_norm": 0.263671875, "learning_rate": 0.00012021589721912277, "loss": 0.8925, "step": 1860 }, { "epoch": 0.4931253305129561, "grad_norm": 0.275390625, "learning_rate": 0.00011976362243679014, "loss": 0.8885, "step": 1865 }, { "epoch": 0.4944473823373876, "grad_norm": 0.22265625, "learning_rate": 0.00011931092655795597, "loss": 0.8869, "step": 1870 }, { "epoch": 0.4957694341618191, "grad_norm": 0.44140625, "learning_rate": 0.00011885781922805101, "loss": 0.8892, "step": 1875 }, { "epoch": 0.49709148598625064, "grad_norm": 0.228515625, "learning_rate": 0.0001184043101012726, "loss": 0.8913, "step": 1880 }, { "epoch": 0.4984135378106822, "grad_norm": 0.232421875, "learning_rate": 0.000117950408840379, "loss": 0.8877, "step": 1885 }, { "epoch": 0.4997355896351137, "grad_norm": 0.1923828125, "learning_rate": 0.00011749612511648362, "loss": 0.8914, "step": 1890 }, { "epoch": 0.5010576414595452, "grad_norm": 0.30078125, "learning_rate": 0.0001170414686088488, "loss": 0.8906, "step": 1895 }, { "epoch": 0.5023796932839767, "grad_norm": 0.2451171875, "learning_rate": 0.00011658644900467965, "loss": 0.8923, "step": 1900 }, { "epoch": 0.5037017451084083, "grad_norm": 0.314453125, "learning_rate": 0.00011613107599891775, "loss": 0.8924, "step": 1905 }, { "epoch": 0.5050237969328397, "grad_norm": 0.259765625, "learning_rate": 0.00011567535929403435, "loss": 0.8869, "step": 1910 }, { "epoch": 0.5063458487572713, "grad_norm": 0.240234375, "learning_rate": 0.00011521930859982382, "loss": 0.8896, "step": 1915 }, { "epoch": 0.5076679005817029, "grad_norm": 0.2109375, "learning_rate": 0.00011476293363319675, "loss": 0.8869, "step": 1920 }, { "epoch": 0.5089899524061343, "grad_norm": 0.2197265625, "learning_rate": 0.0001143062441179729, "loss": 0.8869, "step": 1925 }, { "epoch": 0.5103120042305659, "grad_norm": 0.22265625, "learning_rate": 0.00011384924978467397, "loss": 0.8854, "step": 1930 }, { "epoch": 0.5116340560549973, "grad_norm": 0.24609375, "learning_rate": 0.00011339196037031627, "loss": 0.8905, "step": 1935 }, { "epoch": 0.5129561078794289, "grad_norm": 0.23828125, "learning_rate": 0.00011293438561820343, "loss": 0.8885, "step": 1940 }, { "epoch": 0.5142781597038604, "grad_norm": 0.1923828125, "learning_rate": 0.00011247653527771847, "loss": 0.8869, "step": 1945 }, { "epoch": 0.5156002115282919, "grad_norm": 0.2138671875, "learning_rate": 0.00011201841910411652, "loss": 0.8873, "step": 1950 }, { "epoch": 0.5169222633527234, "grad_norm": 0.2177734375, "learning_rate": 0.00011156004685831648, "loss": 0.893, "step": 1955 }, { "epoch": 0.518244315177155, "grad_norm": 0.298828125, "learning_rate": 0.0001111014283066935, "loss": 0.8841, "step": 1960 }, { "epoch": 0.5195663670015864, "grad_norm": 0.291015625, "learning_rate": 0.0001106425732208705, "loss": 0.8796, "step": 1965 }, { "epoch": 0.520888418826018, "grad_norm": 0.2890625, "learning_rate": 0.0001101834913775103, "loss": 0.8917, "step": 1970 }, { "epoch": 0.5222104706504495, "grad_norm": 0.35546875, "learning_rate": 0.00010972419255810704, "loss": 0.8918, "step": 1975 }, { "epoch": 0.523532522474881, "grad_norm": 0.259765625, "learning_rate": 0.000109264686548778, "loss": 0.8866, "step": 1980 }, { "epoch": 0.5248545742993125, "grad_norm": 0.203125, "learning_rate": 0.00010880498314005488, "loss": 0.8911, "step": 1985 }, { "epoch": 0.526176626123744, "grad_norm": 0.265625, "learning_rate": 0.00010834509212667537, "loss": 0.8856, "step": 1990 }, { "epoch": 0.5274986779481756, "grad_norm": 0.2890625, "learning_rate": 0.00010788502330737438, "loss": 0.8899, "step": 1995 }, { "epoch": 0.5288207297726071, "grad_norm": 0.248046875, "learning_rate": 0.00010742478648467522, "loss": 0.8883, "step": 2000 }, { "epoch": 0.5301427815970386, "grad_norm": 0.2138671875, "learning_rate": 0.00010696439146468085, "loss": 0.8899, "step": 2005 }, { "epoch": 0.5314648334214701, "grad_norm": 0.21875, "learning_rate": 0.00010650384805686482, "loss": 0.8894, "step": 2010 }, { "epoch": 0.5327868852459017, "grad_norm": 0.220703125, "learning_rate": 0.0001060431660738624, "loss": 0.8882, "step": 2015 }, { "epoch": 0.5341089370703331, "grad_norm": 0.236328125, "learning_rate": 0.00010558235533126142, "loss": 0.8855, "step": 2020 }, { "epoch": 0.5354309888947647, "grad_norm": 0.236328125, "learning_rate": 0.00010512142564739302, "loss": 0.8841, "step": 2025 }, { "epoch": 0.5367530407191962, "grad_norm": 0.1826171875, "learning_rate": 0.00010466038684312278, "loss": 0.8912, "step": 2030 }, { "epoch": 0.5380750925436277, "grad_norm": 0.2314453125, "learning_rate": 0.0001041992487416411, "loss": 0.8883, "step": 2035 }, { "epoch": 0.5393971443680592, "grad_norm": 0.24609375, "learning_rate": 0.00010373802116825418, "loss": 0.8863, "step": 2040 }, { "epoch": 0.5407191961924908, "grad_norm": 0.275390625, "learning_rate": 0.00010327671395017451, "loss": 0.8831, "step": 2045 }, { "epoch": 0.5420412480169222, "grad_norm": 0.283203125, "learning_rate": 0.00010281533691631155, "loss": 0.8831, "step": 2050 }, { "epoch": 0.5433632998413538, "grad_norm": 0.2333984375, "learning_rate": 0.00010235389989706232, "loss": 0.8888, "step": 2055 }, { "epoch": 0.5446853516657852, "grad_norm": 0.2431640625, "learning_rate": 0.0001018924127241019, "loss": 0.893, "step": 2060 }, { "epoch": 0.5460074034902168, "grad_norm": 0.2109375, "learning_rate": 0.000101430885230174, "loss": 0.8872, "step": 2065 }, { "epoch": 0.5473294553146484, "grad_norm": 0.294921875, "learning_rate": 0.00010096932724888146, "loss": 0.888, "step": 2070 }, { "epoch": 0.5486515071390798, "grad_norm": 0.2119140625, "learning_rate": 0.00010050774861447662, "loss": 0.8897, "step": 2075 }, { "epoch": 0.5499735589635114, "grad_norm": 0.1806640625, "learning_rate": 0.00010004615916165191, "loss": 0.8865, "step": 2080 }, { "epoch": 0.5512956107879429, "grad_norm": 0.21875, "learning_rate": 9.958456872533031e-05, "loss": 0.8856, "step": 2085 }, { "epoch": 0.5526176626123744, "grad_norm": 0.181640625, "learning_rate": 9.91229871404557e-05, "loss": 0.8828, "step": 2090 }, { "epoch": 0.5539397144368059, "grad_norm": 0.158203125, "learning_rate": 9.866142424178341e-05, "loss": 0.8896, "step": 2095 }, { "epoch": 0.5552617662612375, "grad_norm": 0.2060546875, "learning_rate": 9.819988986367053e-05, "loss": 0.8845, "step": 2100 }, { "epoch": 0.5565838180856689, "grad_norm": 0.208984375, "learning_rate": 9.773839383986664e-05, "loss": 0.8877, "step": 2105 }, { "epoch": 0.5579058699101005, "grad_norm": 0.2119140625, "learning_rate": 9.727694600330395e-05, "loss": 0.8835, "step": 2110 }, { "epoch": 0.559227921734532, "grad_norm": 0.2021484375, "learning_rate": 9.681555618588808e-05, "loss": 0.8855, "step": 2115 }, { "epoch": 0.5605499735589635, "grad_norm": 0.2021484375, "learning_rate": 9.635423421828835e-05, "loss": 0.8838, "step": 2120 }, { "epoch": 0.561872025383395, "grad_norm": 0.23046875, "learning_rate": 9.589298992972857e-05, "loss": 0.8901, "step": 2125 }, { "epoch": 0.5631940772078265, "grad_norm": 0.306640625, "learning_rate": 9.543183314777732e-05, "loss": 0.888, "step": 2130 }, { "epoch": 0.5645161290322581, "grad_norm": 0.244140625, "learning_rate": 9.497077369813885e-05, "loss": 0.8865, "step": 2135 }, { "epoch": 0.5658381808566896, "grad_norm": 0.29296875, "learning_rate": 9.450982140444344e-05, "loss": 0.8916, "step": 2140 }, { "epoch": 0.5671602326811211, "grad_norm": 0.21875, "learning_rate": 9.404898608803836e-05, "loss": 0.8869, "step": 2145 }, { "epoch": 0.5684822845055526, "grad_norm": 0.2412109375, "learning_rate": 9.358827756777837e-05, "loss": 0.8844, "step": 2150 }, { "epoch": 0.5698043363299842, "grad_norm": 0.23046875, "learning_rate": 9.312770565981677e-05, "loss": 0.8861, "step": 2155 }, { "epoch": 0.5711263881544156, "grad_norm": 0.2412109375, "learning_rate": 9.266728017739601e-05, "loss": 0.888, "step": 2160 }, { "epoch": 0.5724484399788472, "grad_norm": 0.1953125, "learning_rate": 9.220701093063868e-05, "loss": 0.8857, "step": 2165 }, { "epoch": 0.5737704918032787, "grad_norm": 0.2421875, "learning_rate": 9.174690772633865e-05, "loss": 0.8896, "step": 2170 }, { "epoch": 0.5750925436277102, "grad_norm": 0.326171875, "learning_rate": 9.128698036775179e-05, "loss": 0.884, "step": 2175 }, { "epoch": 0.5764145954521417, "grad_norm": 0.208984375, "learning_rate": 9.082723865438747e-05, "loss": 0.8903, "step": 2180 }, { "epoch": 0.5777366472765733, "grad_norm": 0.2001953125, "learning_rate": 9.036769238179947e-05, "loss": 0.8927, "step": 2185 }, { "epoch": 0.5790586991010047, "grad_norm": 0.21484375, "learning_rate": 8.990835134137744e-05, "loss": 0.8848, "step": 2190 }, { "epoch": 0.5803807509254363, "grad_norm": 0.166015625, "learning_rate": 8.944922532013811e-05, "loss": 0.884, "step": 2195 }, { "epoch": 0.5817028027498677, "grad_norm": 0.2578125, "learning_rate": 8.899032410051708e-05, "loss": 0.8832, "step": 2200 }, { "epoch": 0.5830248545742993, "grad_norm": 0.2119140625, "learning_rate": 8.853165746015997e-05, "loss": 0.8851, "step": 2205 }, { "epoch": 0.5843469063987309, "grad_norm": 0.21484375, "learning_rate": 8.807323517171444e-05, "loss": 0.8899, "step": 2210 }, { "epoch": 0.5856689582231623, "grad_norm": 0.181640625, "learning_rate": 8.761506700262172e-05, "loss": 0.8875, "step": 2215 }, { "epoch": 0.5869910100475939, "grad_norm": 0.19921875, "learning_rate": 8.715716271490877e-05, "loss": 0.8899, "step": 2220 }, { "epoch": 0.5883130618720254, "grad_norm": 0.185546875, "learning_rate": 8.669953206497994e-05, "loss": 0.8819, "step": 2225 }, { "epoch": 0.5896351136964569, "grad_norm": 0.25390625, "learning_rate": 8.624218480340942e-05, "loss": 0.8879, "step": 2230 }, { "epoch": 0.5909571655208884, "grad_norm": 0.296875, "learning_rate": 8.578513067473322e-05, "loss": 0.8877, "step": 2235 }, { "epoch": 0.59227921734532, "grad_norm": 0.1875, "learning_rate": 8.532837941724181e-05, "loss": 0.8849, "step": 2240 }, { "epoch": 0.5936012691697514, "grad_norm": 0.2109375, "learning_rate": 8.487194076277234e-05, "loss": 0.8826, "step": 2245 }, { "epoch": 0.594923320994183, "grad_norm": 0.259765625, "learning_rate": 8.441582443650161e-05, "loss": 0.8847, "step": 2250 }, { "epoch": 0.5962453728186144, "grad_norm": 0.1904296875, "learning_rate": 8.396004015673853e-05, "loss": 0.8825, "step": 2255 }, { "epoch": 0.597567424643046, "grad_norm": 0.2158203125, "learning_rate": 8.35045976347173e-05, "loss": 0.888, "step": 2260 }, { "epoch": 0.5988894764674775, "grad_norm": 0.1865234375, "learning_rate": 8.304950657439033e-05, "loss": 0.885, "step": 2265 }, { "epoch": 0.600211528291909, "grad_norm": 0.2353515625, "learning_rate": 8.259477667222172e-05, "loss": 0.8857, "step": 2270 }, { "epoch": 0.6015335801163406, "grad_norm": 0.296875, "learning_rate": 8.214041761698031e-05, "loss": 0.8876, "step": 2275 }, { "epoch": 0.6028556319407721, "grad_norm": 0.232421875, "learning_rate": 8.168643908953359e-05, "loss": 0.8846, "step": 2280 }, { "epoch": 0.6041776837652036, "grad_norm": 0.255859375, "learning_rate": 8.12328507626411e-05, "loss": 0.8863, "step": 2285 }, { "epoch": 0.6054997355896351, "grad_norm": 0.3359375, "learning_rate": 8.077966230074872e-05, "loss": 0.8889, "step": 2290 }, { "epoch": 0.6068217874140667, "grad_norm": 0.2197265625, "learning_rate": 8.032688335978247e-05, "loss": 0.8844, "step": 2295 }, { "epoch": 0.6081438392384981, "grad_norm": 0.2470703125, "learning_rate": 7.987452358694278e-05, "loss": 0.8894, "step": 2300 }, { "epoch": 0.6094658910629297, "grad_norm": 0.2060546875, "learning_rate": 7.942259262049911e-05, "loss": 0.8825, "step": 2305 }, { "epoch": 0.6107879428873612, "grad_norm": 0.236328125, "learning_rate": 7.897110008958441e-05, "loss": 0.8851, "step": 2310 }, { "epoch": 0.6121099947117927, "grad_norm": 0.2021484375, "learning_rate": 7.852005561399014e-05, "loss": 0.8903, "step": 2315 }, { "epoch": 0.6134320465362242, "grad_norm": 0.27734375, "learning_rate": 7.806946880396109e-05, "loss": 0.8873, "step": 2320 }, { "epoch": 0.6147540983606558, "grad_norm": 0.27734375, "learning_rate": 7.761934925999086e-05, "loss": 0.8862, "step": 2325 }, { "epoch": 0.6160761501850872, "grad_norm": 0.23046875, "learning_rate": 7.716970657261702e-05, "loss": 0.8846, "step": 2330 }, { "epoch": 0.6173982020095188, "grad_norm": 0.2255859375, "learning_rate": 7.672055032221709e-05, "loss": 0.8897, "step": 2335 }, { "epoch": 0.6187202538339502, "grad_norm": 0.2197265625, "learning_rate": 7.627189007880413e-05, "loss": 0.8873, "step": 2340 }, { "epoch": 0.6200423056583818, "grad_norm": 0.2021484375, "learning_rate": 7.582373540182304e-05, "loss": 0.8904, "step": 2345 }, { "epoch": 0.6213643574828134, "grad_norm": 0.208984375, "learning_rate": 7.537609583994667e-05, "loss": 0.8834, "step": 2350 }, { "epoch": 0.6226864093072448, "grad_norm": 0.2021484375, "learning_rate": 7.492898093087269e-05, "loss": 0.8825, "step": 2355 }, { "epoch": 0.6240084611316764, "grad_norm": 0.2021484375, "learning_rate": 7.448240020111993e-05, "loss": 0.8837, "step": 2360 }, { "epoch": 0.6253305129561079, "grad_norm": 0.1669921875, "learning_rate": 7.403636316582594e-05, "loss": 0.8885, "step": 2365 }, { "epoch": 0.6266525647805394, "grad_norm": 0.2158203125, "learning_rate": 7.359087932854371e-05, "loss": 0.8944, "step": 2370 }, { "epoch": 0.6279746166049709, "grad_norm": 0.369140625, "learning_rate": 7.314595818103959e-05, "loss": 0.889, "step": 2375 }, { "epoch": 0.6292966684294025, "grad_norm": 0.232421875, "learning_rate": 7.270160920309077e-05, "loss": 0.8869, "step": 2380 }, { "epoch": 0.6306187202538339, "grad_norm": 0.208984375, "learning_rate": 7.225784186228359e-05, "loss": 0.8829, "step": 2385 }, { "epoch": 0.6319407720782655, "grad_norm": 0.25390625, "learning_rate": 7.181466561381152e-05, "loss": 0.8889, "step": 2390 }, { "epoch": 0.633262823902697, "grad_norm": 0.287109375, "learning_rate": 7.13720899002739e-05, "loss": 0.8823, "step": 2395 }, { "epoch": 0.6345848757271285, "grad_norm": 0.21484375, "learning_rate": 7.093012415147462e-05, "loss": 0.884, "step": 2400 }, { "epoch": 0.63590692755156, "grad_norm": 0.1591796875, "learning_rate": 7.048877778422138e-05, "loss": 0.8848, "step": 2405 }, { "epoch": 0.6372289793759915, "grad_norm": 0.259765625, "learning_rate": 7.00480602021248e-05, "loss": 0.8821, "step": 2410 }, { "epoch": 0.638551031200423, "grad_norm": 0.291015625, "learning_rate": 6.960798079539835e-05, "loss": 0.8879, "step": 2415 }, { "epoch": 0.6398730830248546, "grad_norm": 0.2080078125, "learning_rate": 6.916854894065796e-05, "loss": 0.886, "step": 2420 }, { "epoch": 0.6411951348492861, "grad_norm": 0.251953125, "learning_rate": 6.87297740007225e-05, "loss": 0.8895, "step": 2425 }, { "epoch": 0.6425171866737176, "grad_norm": 0.189453125, "learning_rate": 6.829166532441421e-05, "loss": 0.8839, "step": 2430 }, { "epoch": 0.6438392384981492, "grad_norm": 0.203125, "learning_rate": 6.785423224635941e-05, "loss": 0.8834, "step": 2435 }, { "epoch": 0.6451612903225806, "grad_norm": 0.2021484375, "learning_rate": 6.741748408678975e-05, "loss": 0.8849, "step": 2440 }, { "epoch": 0.6464833421470122, "grad_norm": 0.1787109375, "learning_rate": 6.698143015134349e-05, "loss": 0.8804, "step": 2445 }, { "epoch": 0.6478053939714437, "grad_norm": 0.197265625, "learning_rate": 6.65460797308674e-05, "loss": 0.8845, "step": 2450 }, { "epoch": 0.6491274457958752, "grad_norm": 0.1611328125, "learning_rate": 6.611144210121861e-05, "loss": 0.8897, "step": 2455 }, { "epoch": 0.6504494976203067, "grad_norm": 0.1630859375, "learning_rate": 6.567752652306717e-05, "loss": 0.8913, "step": 2460 }, { "epoch": 0.6517715494447383, "grad_norm": 0.2470703125, "learning_rate": 6.52443422416985e-05, "loss": 0.8884, "step": 2465 }, { "epoch": 0.6530936012691697, "grad_norm": 0.17578125, "learning_rate": 6.48118984868167e-05, "loss": 0.8894, "step": 2470 }, { "epoch": 0.6544156530936013, "grad_norm": 0.2265625, "learning_rate": 6.438020447234759e-05, "loss": 0.8837, "step": 2475 }, { "epoch": 0.6557377049180327, "grad_norm": 0.1748046875, "learning_rate": 6.394926939624267e-05, "loss": 0.8864, "step": 2480 }, { "epoch": 0.6570597567424643, "grad_norm": 0.1796875, "learning_rate": 6.351910244028285e-05, "loss": 0.8859, "step": 2485 }, { "epoch": 0.6583818085668959, "grad_norm": 0.2314453125, "learning_rate": 6.308971276988313e-05, "loss": 0.885, "step": 2490 }, { "epoch": 0.6597038603913273, "grad_norm": 0.263671875, "learning_rate": 6.266110953389704e-05, "loss": 0.8835, "step": 2495 }, { "epoch": 0.6610259122157589, "grad_norm": 0.1806640625, "learning_rate": 6.223330186442194e-05, "loss": 0.8834, "step": 2500 }, { "epoch": 0.6623479640401904, "grad_norm": 0.201171875, "learning_rate": 6.180629887660425e-05, "loss": 0.8839, "step": 2505 }, { "epoch": 0.6636700158646219, "grad_norm": 0.1904296875, "learning_rate": 6.138010966844538e-05, "loss": 0.8868, "step": 2510 }, { "epoch": 0.6649920676890534, "grad_norm": 0.1962890625, "learning_rate": 6.095474332060776e-05, "loss": 0.8857, "step": 2515 }, { "epoch": 0.666314119513485, "grad_norm": 0.205078125, "learning_rate": 6.0530208896221527e-05, "loss": 0.8823, "step": 2520 }, { "epoch": 0.6676361713379164, "grad_norm": 0.28125, "learning_rate": 6.010651544069119e-05, "loss": 0.8835, "step": 2525 }, { "epoch": 0.668958223162348, "grad_norm": 0.259765625, "learning_rate": 5.968367198150316e-05, "loss": 0.8826, "step": 2530 }, { "epoch": 0.6702802749867794, "grad_norm": 0.1650390625, "learning_rate": 5.92616875280332e-05, "loss": 0.8825, "step": 2535 }, { "epoch": 0.671602326811211, "grad_norm": 0.2294921875, "learning_rate": 5.8840571071354565e-05, "loss": 0.8874, "step": 2540 }, { "epoch": 0.6729243786356425, "grad_norm": 0.197265625, "learning_rate": 5.842033158404636e-05, "loss": 0.8914, "step": 2545 }, { "epoch": 0.674246430460074, "grad_norm": 0.181640625, "learning_rate": 5.800097802000257e-05, "loss": 0.8844, "step": 2550 }, { "epoch": 0.6755684822845055, "grad_norm": 0.240234375, "learning_rate": 5.758251931424089e-05, "loss": 0.8809, "step": 2555 }, { "epoch": 0.6768905341089371, "grad_norm": 0.2421875, "learning_rate": 5.716496438271277e-05, "loss": 0.8828, "step": 2560 }, { "epoch": 0.6782125859333686, "grad_norm": 0.185546875, "learning_rate": 5.6748322122113264e-05, "loss": 0.8832, "step": 2565 }, { "epoch": 0.6795346377578001, "grad_norm": 0.23046875, "learning_rate": 5.633260140969132e-05, "loss": 0.885, "step": 2570 }, { "epoch": 0.6808566895822317, "grad_norm": 0.162109375, "learning_rate": 5.591781110306096e-05, "loss": 0.8837, "step": 2575 }, { "epoch": 0.6821787414066631, "grad_norm": 0.173828125, "learning_rate": 5.55039600400122e-05, "loss": 0.8817, "step": 2580 }, { "epoch": 0.6835007932310947, "grad_norm": 0.20703125, "learning_rate": 5.509105703832313e-05, "loss": 0.8858, "step": 2585 }, { "epoch": 0.6848228450555262, "grad_norm": 0.2451171875, "learning_rate": 5.467911089557169e-05, "loss": 0.8863, "step": 2590 }, { "epoch": 0.6861448968799577, "grad_norm": 0.208984375, "learning_rate": 5.426813038894839e-05, "loss": 0.8844, "step": 2595 }, { "epoch": 0.6874669487043892, "grad_norm": 0.203125, "learning_rate": 5.3858124275069244e-05, "loss": 0.8836, "step": 2600 }, { "epoch": 0.6887890005288207, "grad_norm": 0.25390625, "learning_rate": 5.344910128978934e-05, "loss": 0.8837, "step": 2605 }, { "epoch": 0.6901110523532522, "grad_norm": 0.1982421875, "learning_rate": 5.304107014801645e-05, "loss": 0.8845, "step": 2610 }, { "epoch": 0.6914331041776838, "grad_norm": 0.21875, "learning_rate": 5.263403954352564e-05, "loss": 0.8878, "step": 2615 }, { "epoch": 0.6927551560021152, "grad_norm": 0.1767578125, "learning_rate": 5.222801814877369e-05, "loss": 0.8879, "step": 2620 }, { "epoch": 0.6940772078265468, "grad_norm": 0.2001953125, "learning_rate": 5.182301461471474e-05, "loss": 0.8905, "step": 2625 }, { "epoch": 0.6953992596509783, "grad_norm": 0.201171875, "learning_rate": 5.141903757061555e-05, "loss": 0.8889, "step": 2630 }, { "epoch": 0.6967213114754098, "grad_norm": 0.2001953125, "learning_rate": 5.1016095623871865e-05, "loss": 0.8805, "step": 2635 }, { "epoch": 0.6980433632998414, "grad_norm": 0.2001953125, "learning_rate": 5.0614197359824955e-05, "loss": 0.8848, "step": 2640 }, { "epoch": 0.6993654151242729, "grad_norm": 0.2294921875, "learning_rate": 5.021335134157885e-05, "loss": 0.883, "step": 2645 }, { "epoch": 0.7006874669487044, "grad_norm": 0.216796875, "learning_rate": 4.981356610981756e-05, "loss": 0.8817, "step": 2650 }, { "epoch": 0.7020095187731359, "grad_norm": 0.1591796875, "learning_rate": 4.9414850182623465e-05, "loss": 0.8831, "step": 2655 }, { "epoch": 0.7033315705975675, "grad_norm": 0.1982421875, "learning_rate": 4.901721205529548e-05, "loss": 0.8845, "step": 2660 }, { "epoch": 0.7046536224219989, "grad_norm": 0.189453125, "learning_rate": 4.8620660200168425e-05, "loss": 0.8821, "step": 2665 }, { "epoch": 0.7059756742464305, "grad_norm": 0.19140625, "learning_rate": 4.822520306643207e-05, "loss": 0.8835, "step": 2670 }, { "epoch": 0.7072977260708619, "grad_norm": 0.255859375, "learning_rate": 4.783084907995156e-05, "loss": 0.8831, "step": 2675 }, { "epoch": 0.7086197778952935, "grad_norm": 0.158203125, "learning_rate": 4.743760664308753e-05, "loss": 0.8841, "step": 2680 }, { "epoch": 0.709941829719725, "grad_norm": 0.298828125, "learning_rate": 4.704548413451728e-05, "loss": 0.8843, "step": 2685 }, { "epoch": 0.7112638815441565, "grad_norm": 0.220703125, "learning_rate": 4.665448990905613e-05, "loss": 0.8843, "step": 2690 }, { "epoch": 0.712585933368588, "grad_norm": 0.265625, "learning_rate": 4.626463229747958e-05, "loss": 0.8832, "step": 2695 }, { "epoch": 0.7139079851930196, "grad_norm": 0.181640625, "learning_rate": 4.587591960634565e-05, "loss": 0.8852, "step": 2700 }, { "epoch": 0.7152300370174511, "grad_norm": 0.2421875, "learning_rate": 4.54883601178179e-05, "loss": 0.8844, "step": 2705 }, { "epoch": 0.7165520888418826, "grad_norm": 0.1943359375, "learning_rate": 4.510196208948909e-05, "loss": 0.8873, "step": 2710 }, { "epoch": 0.7178741406663142, "grad_norm": 0.17578125, "learning_rate": 4.4716733754205064e-05, "loss": 0.8872, "step": 2715 }, { "epoch": 0.7191961924907456, "grad_norm": 0.197265625, "learning_rate": 4.433268331988956e-05, "loss": 0.8891, "step": 2720 }, { "epoch": 0.7205182443151772, "grad_norm": 0.2392578125, "learning_rate": 4.394981896936908e-05, "loss": 0.8861, "step": 2725 }, { "epoch": 0.7218402961396087, "grad_norm": 0.17578125, "learning_rate": 4.356814886019871e-05, "loss": 0.8884, "step": 2730 }, { "epoch": 0.7231623479640402, "grad_norm": 0.162109375, "learning_rate": 4.3187681124488196e-05, "loss": 0.8865, "step": 2735 }, { "epoch": 0.7244843997884717, "grad_norm": 0.19921875, "learning_rate": 4.280842386872889e-05, "loss": 0.8817, "step": 2740 }, { "epoch": 0.7258064516129032, "grad_norm": 0.1865234375, "learning_rate": 4.243038517362068e-05, "loss": 0.8832, "step": 2745 }, { "epoch": 0.7271285034373347, "grad_norm": 0.2021484375, "learning_rate": 4.205357309390021e-05, "loss": 0.8856, "step": 2750 }, { "epoch": 0.7284505552617663, "grad_norm": 0.14453125, "learning_rate": 4.167799565816889e-05, "loss": 0.8862, "step": 2755 }, { "epoch": 0.7297726070861977, "grad_norm": 0.2021484375, "learning_rate": 4.130366086872215e-05, "loss": 0.8826, "step": 2760 }, { "epoch": 0.7310946589106293, "grad_norm": 0.22265625, "learning_rate": 4.093057670137865e-05, "loss": 0.8871, "step": 2765 }, { "epoch": 0.7324167107350608, "grad_norm": 0.2099609375, "learning_rate": 4.055875110531068e-05, "loss": 0.8862, "step": 2770 }, { "epoch": 0.7337387625594923, "grad_norm": 0.20703125, "learning_rate": 4.0188192002874404e-05, "loss": 0.8849, "step": 2775 }, { "epoch": 0.7350608143839239, "grad_norm": 0.2431640625, "learning_rate": 3.9818907289441375e-05, "loss": 0.8831, "step": 2780 }, { "epoch": 0.7363828662083554, "grad_norm": 0.205078125, "learning_rate": 3.9450904833230094e-05, "loss": 0.8879, "step": 2785 }, { "epoch": 0.7377049180327869, "grad_norm": 0.162109375, "learning_rate": 3.908419247513862e-05, "loss": 0.884, "step": 2790 }, { "epoch": 0.7390269698572184, "grad_norm": 0.173828125, "learning_rate": 3.871877802857714e-05, "loss": 0.8931, "step": 2795 }, { "epoch": 0.74034902168165, "grad_norm": 0.2578125, "learning_rate": 3.8354669279301914e-05, "loss": 0.8881, "step": 2800 }, { "epoch": 0.7416710735060814, "grad_norm": 0.294921875, "learning_rate": 3.7991873985248996e-05, "loss": 0.8866, "step": 2805 }, { "epoch": 0.742993125330513, "grad_norm": 0.2294921875, "learning_rate": 3.763039987636924e-05, "loss": 0.8826, "step": 2810 }, { "epoch": 0.7443151771549444, "grad_norm": 0.193359375, "learning_rate": 3.7270254654463356e-05, "loss": 0.8841, "step": 2815 }, { "epoch": 0.745637228979376, "grad_norm": 0.2041015625, "learning_rate": 3.6911445993018076e-05, "loss": 0.8903, "step": 2820 }, { "epoch": 0.7469592808038075, "grad_norm": 0.2177734375, "learning_rate": 3.6553981537042406e-05, "loss": 0.8823, "step": 2825 }, { "epoch": 0.748281332628239, "grad_norm": 0.259765625, "learning_rate": 3.619786890290482e-05, "loss": 0.8846, "step": 2830 }, { "epoch": 0.7496033844526705, "grad_norm": 0.205078125, "learning_rate": 3.5843115678171125e-05, "loss": 0.881, "step": 2835 }, { "epoch": 0.7509254362771021, "grad_norm": 0.255859375, "learning_rate": 3.5489729421442516e-05, "loss": 0.8845, "step": 2840 }, { "epoch": 0.7522474881015336, "grad_norm": 0.2412109375, "learning_rate": 3.513771766219487e-05, "loss": 0.8857, "step": 2845 }, { "epoch": 0.7535695399259651, "grad_norm": 0.220703125, "learning_rate": 3.4787087900617955e-05, "loss": 0.8882, "step": 2850 }, { "epoch": 0.7548915917503967, "grad_norm": 0.2421875, "learning_rate": 3.443784760745593e-05, "loss": 0.8853, "step": 2855 }, { "epoch": 0.7562136435748281, "grad_norm": 0.216796875, "learning_rate": 3.409000422384797e-05, "loss": 0.8892, "step": 2860 }, { "epoch": 0.7575356953992597, "grad_norm": 0.1845703125, "learning_rate": 3.3743565161169835e-05, "loss": 0.8842, "step": 2865 }, { "epoch": 0.7588577472236911, "grad_norm": 0.1953125, "learning_rate": 3.339853780087584e-05, "loss": 0.8863, "step": 2870 }, { "epoch": 0.7601797990481227, "grad_norm": 0.1962890625, "learning_rate": 3.3054929494341805e-05, "loss": 0.884, "step": 2875 }, { "epoch": 0.7615018508725542, "grad_norm": 0.2119140625, "learning_rate": 3.2712747562708115e-05, "loss": 0.886, "step": 2880 }, { "epoch": 0.7628239026969857, "grad_norm": 0.216796875, "learning_rate": 3.237199929672401e-05, "loss": 0.8833, "step": 2885 }, { "epoch": 0.7641459545214172, "grad_norm": 0.2431640625, "learning_rate": 3.203269195659202e-05, "loss": 0.8824, "step": 2890 }, { "epoch": 0.7654680063458488, "grad_norm": 0.2001953125, "learning_rate": 3.169483277181351e-05, "loss": 0.8864, "step": 2895 }, { "epoch": 0.7667900581702802, "grad_norm": 0.2001953125, "learning_rate": 3.135842894103436e-05, "loss": 0.8867, "step": 2900 }, { "epoch": 0.7681121099947118, "grad_norm": 0.1806640625, "learning_rate": 3.102348763189188e-05, "loss": 0.891, "step": 2905 }, { "epoch": 0.7694341618191433, "grad_norm": 0.2001953125, "learning_rate": 3.069001598086184e-05, "loss": 0.8902, "step": 2910 }, { "epoch": 0.7707562136435748, "grad_norm": 0.193359375, "learning_rate": 3.0358021093106594e-05, "loss": 0.885, "step": 2915 }, { "epoch": 0.7720782654680064, "grad_norm": 0.23046875, "learning_rate": 3.0027510042323537e-05, "loss": 0.8866, "step": 2920 }, { "epoch": 0.7734003172924379, "grad_norm": 0.2275390625, "learning_rate": 2.9698489870594616e-05, "loss": 0.8831, "step": 2925 }, { "epoch": 0.7747223691168694, "grad_norm": 0.244140625, "learning_rate": 2.937096758823601e-05, "loss": 0.8869, "step": 2930 }, { "epoch": 0.7760444209413009, "grad_norm": 0.1748046875, "learning_rate": 2.9044950173649e-05, "loss": 0.8816, "step": 2935 }, { "epoch": 0.7773664727657325, "grad_norm": 0.16796875, "learning_rate": 2.872044457317109e-05, "loss": 0.8875, "step": 2940 }, { "epoch": 0.7786885245901639, "grad_norm": 0.171875, "learning_rate": 2.839745770092821e-05, "loss": 0.8828, "step": 2945 }, { "epoch": 0.7800105764145955, "grad_norm": 0.1884765625, "learning_rate": 2.8075996438687148e-05, "loss": 0.8818, "step": 2950 }, { "epoch": 0.7813326282390269, "grad_norm": 0.212890625, "learning_rate": 2.7756067635709204e-05, "loss": 0.8911, "step": 2955 }, { "epoch": 0.7826546800634585, "grad_norm": 0.1904296875, "learning_rate": 2.7437678108604003e-05, "loss": 0.8859, "step": 2960 }, { "epoch": 0.78397673188789, "grad_norm": 0.1796875, "learning_rate": 2.712083464118441e-05, "loss": 0.8892, "step": 2965 }, { "epoch": 0.7852987837123215, "grad_norm": 0.205078125, "learning_rate": 2.68055439843219e-05, "loss": 0.8826, "step": 2970 }, { "epoch": 0.786620835536753, "grad_norm": 0.1962890625, "learning_rate": 2.6491812855802834e-05, "loss": 0.8877, "step": 2975 }, { "epoch": 0.7879428873611846, "grad_norm": 0.189453125, "learning_rate": 2.6179647940185238e-05, "loss": 0.8867, "step": 2980 }, { "epoch": 0.789264939185616, "grad_norm": 0.232421875, "learning_rate": 2.5869055888656336e-05, "loss": 0.8858, "step": 2985 }, { "epoch": 0.7905869910100476, "grad_norm": 0.234375, "learning_rate": 2.556004331889098e-05, "loss": 0.8859, "step": 2990 }, { "epoch": 0.7919090428344792, "grad_norm": 0.197265625, "learning_rate": 2.525261681491047e-05, "loss": 0.8849, "step": 2995 }, { "epoch": 0.7932310946589106, "grad_norm": 0.248046875, "learning_rate": 2.494678292694247e-05, "loss": 0.8823, "step": 3000 }, { "epoch": 0.7945531464833422, "grad_norm": 0.162109375, "learning_rate": 2.4642548171281244e-05, "loss": 0.8886, "step": 3005 }, { "epoch": 0.7958751983077736, "grad_norm": 0.2041015625, "learning_rate": 2.433991903014896e-05, "loss": 0.8889, "step": 3010 }, { "epoch": 0.7971972501322052, "grad_norm": 0.2255859375, "learning_rate": 2.403890195155748e-05, "loss": 0.8858, "step": 3015 }, { "epoch": 0.7985193019566367, "grad_norm": 0.1728515625, "learning_rate": 2.373950334917111e-05, "loss": 0.8818, "step": 3020 }, { "epoch": 0.7998413537810682, "grad_norm": 0.263671875, "learning_rate": 2.3441729602169748e-05, "loss": 0.8859, "step": 3025 }, { "epoch": 0.8011634056054997, "grad_norm": 0.2197265625, "learning_rate": 2.3145587055113183e-05, "loss": 0.8876, "step": 3030 }, { "epoch": 0.8024854574299313, "grad_norm": 0.25, "learning_rate": 2.2851082017805703e-05, "loss": 0.8843, "step": 3035 }, { "epoch": 0.8038075092543627, "grad_norm": 0.2412109375, "learning_rate": 2.255822076516184e-05, "loss": 0.8846, "step": 3040 }, { "epoch": 0.8051295610787943, "grad_norm": 0.2392578125, "learning_rate": 2.2267009537072536e-05, "loss": 0.8877, "step": 3045 }, { "epoch": 0.8064516129032258, "grad_norm": 0.263671875, "learning_rate": 2.197745453827226e-05, "loss": 0.8836, "step": 3050 }, { "epoch": 0.8077736647276573, "grad_norm": 0.171875, "learning_rate": 2.168956193820676e-05, "loss": 0.885, "step": 3055 }, { "epoch": 0.8090957165520889, "grad_norm": 0.1669921875, "learning_rate": 2.140333787090172e-05, "loss": 0.8896, "step": 3060 }, { "epoch": 0.8104177683765204, "grad_norm": 0.197265625, "learning_rate": 2.1118788434831894e-05, "loss": 0.8819, "step": 3065 }, { "epoch": 0.8117398202009519, "grad_norm": 0.19921875, "learning_rate": 2.083591969279136e-05, "loss": 0.8838, "step": 3070 }, { "epoch": 0.8130618720253834, "grad_norm": 0.1884765625, "learning_rate": 2.0554737671764134e-05, "loss": 0.8825, "step": 3075 }, { "epoch": 0.814383923849815, "grad_norm": 0.181640625, "learning_rate": 2.0275248362795995e-05, "loss": 0.8857, "step": 3080 }, { "epoch": 0.8157059756742464, "grad_norm": 0.30078125, "learning_rate": 1.999745772086655e-05, "loss": 0.8852, "step": 3085 }, { "epoch": 0.817028027498678, "grad_norm": 0.1904296875, "learning_rate": 1.9721371664762666e-05, "loss": 0.8864, "step": 3090 }, { "epoch": 0.8183500793231094, "grad_norm": 0.2451171875, "learning_rate": 1.9446996076952094e-05, "loss": 0.88, "step": 3095 }, { "epoch": 0.819672131147541, "grad_norm": 0.20703125, "learning_rate": 1.917433680345829e-05, "loss": 0.8872, "step": 3100 }, { "epoch": 0.8209941829719725, "grad_norm": 0.2119140625, "learning_rate": 1.8903399653735766e-05, "loss": 0.888, "step": 3105 }, { "epoch": 0.822316234796404, "grad_norm": 0.2021484375, "learning_rate": 1.8634190400546415e-05, "loss": 0.8882, "step": 3110 }, { "epoch": 0.8236382866208355, "grad_norm": 0.1845703125, "learning_rate": 1.8366714779836434e-05, "loss": 0.8887, "step": 3115 }, { "epoch": 0.8249603384452671, "grad_norm": 0.1728515625, "learning_rate": 1.8100978490614085e-05, "loss": 0.8856, "step": 3120 }, { "epoch": 0.8262823902696985, "grad_norm": 0.1796875, "learning_rate": 1.783698719482836e-05, "loss": 0.886, "step": 3125 }, { "epoch": 0.8276044420941301, "grad_norm": 0.181640625, "learning_rate": 1.7574746517248254e-05, "loss": 0.8807, "step": 3130 }, { "epoch": 0.8289264939185617, "grad_norm": 0.2431640625, "learning_rate": 1.7314262045343e-05, "loss": 0.8861, "step": 3135 }, { "epoch": 0.8302485457429931, "grad_norm": 0.2021484375, "learning_rate": 1.7055539329162963e-05, "loss": 0.8825, "step": 3140 }, { "epoch": 0.8315705975674247, "grad_norm": 0.193359375, "learning_rate": 1.6798583881221374e-05, "loss": 0.8803, "step": 3145 }, { "epoch": 0.8328926493918561, "grad_norm": 0.23046875, "learning_rate": 1.6543401176376916e-05, "loss": 0.8877, "step": 3150 }, { "epoch": 0.8342147012162877, "grad_norm": 0.2041015625, "learning_rate": 1.628999665171712e-05, "loss": 0.8832, "step": 3155 }, { "epoch": 0.8355367530407192, "grad_norm": 0.1796875, "learning_rate": 1.603837570644238e-05, "loss": 0.8854, "step": 3160 }, { "epoch": 0.8368588048651507, "grad_norm": 0.2578125, "learning_rate": 1.578854370175107e-05, "loss": 0.883, "step": 3165 }, { "epoch": 0.8381808566895822, "grad_norm": 0.189453125, "learning_rate": 1.554050596072516e-05, "loss": 0.8855, "step": 3170 }, { "epoch": 0.8395029085140138, "grad_norm": 0.1611328125, "learning_rate": 1.5294267768216975e-05, "loss": 0.8873, "step": 3175 }, { "epoch": 0.8408249603384452, "grad_norm": 0.212890625, "learning_rate": 1.5049834370736405e-05, "loss": 0.8869, "step": 3180 }, { "epoch": 0.8421470121628768, "grad_norm": 0.2001953125, "learning_rate": 1.4807210976339291e-05, "loss": 0.8853, "step": 3185 }, { "epoch": 0.8434690639873083, "grad_norm": 0.212890625, "learning_rate": 1.4566402754516328e-05, "loss": 0.8862, "step": 3190 }, { "epoch": 0.8447911158117398, "grad_norm": 0.2255859375, "learning_rate": 1.4327414836082976e-05, "loss": 0.884, "step": 3195 }, { "epoch": 0.8461131676361714, "grad_norm": 0.18359375, "learning_rate": 1.409025231307013e-05, "loss": 0.8904, "step": 3200 }, { "epoch": 0.8474352194606029, "grad_norm": 0.20703125, "learning_rate": 1.3854920238615688e-05, "loss": 0.8857, "step": 3205 }, { "epoch": 0.8487572712850344, "grad_norm": 0.189453125, "learning_rate": 1.3621423626856756e-05, "loss": 0.8863, "step": 3210 }, { "epoch": 0.8500793231094659, "grad_norm": 0.25390625, "learning_rate": 1.3389767452822943e-05, "loss": 0.8851, "step": 3215 }, { "epoch": 0.8514013749338974, "grad_norm": 0.2177734375, "learning_rate": 1.3159956652330251e-05, "loss": 0.886, "step": 3220 }, { "epoch": 0.8527234267583289, "grad_norm": 0.21484375, "learning_rate": 1.2931996121876033e-05, "loss": 0.8822, "step": 3225 }, { "epoch": 0.8540454785827605, "grad_norm": 0.1640625, "learning_rate": 1.2705890718534508e-05, "loss": 0.8833, "step": 3230 }, { "epoch": 0.8553675304071919, "grad_norm": 0.1904296875, "learning_rate": 1.2481645259853436e-05, "loss": 0.8854, "step": 3235 }, { "epoch": 0.8566895822316235, "grad_norm": 0.220703125, "learning_rate": 1.2259264523751335e-05, "loss": 0.8882, "step": 3240 }, { "epoch": 0.858011634056055, "grad_norm": 0.18359375, "learning_rate": 1.2038753248415768e-05, "loss": 0.8862, "step": 3245 }, { "epoch": 0.8593336858804865, "grad_norm": 0.1572265625, "learning_rate": 1.1820116132202374e-05, "loss": 0.8882, "step": 3250 }, { "epoch": 0.860655737704918, "grad_norm": 0.2431640625, "learning_rate": 1.16033578335347e-05, "loss": 0.888, "step": 3255 }, { "epoch": 0.8619777895293496, "grad_norm": 0.212890625, "learning_rate": 1.138848297080507e-05, "loss": 0.887, "step": 3260 }, { "epoch": 0.863299841353781, "grad_norm": 0.1904296875, "learning_rate": 1.1175496122276008e-05, "loss": 0.8841, "step": 3265 }, { "epoch": 0.8646218931782126, "grad_norm": 0.1728515625, "learning_rate": 1.0964401825982895e-05, "loss": 0.8828, "step": 3270 }, { "epoch": 0.8659439450026442, "grad_norm": 0.2294921875, "learning_rate": 1.075520457963708e-05, "loss": 0.8828, "step": 3275 }, { "epoch": 0.8672659968270756, "grad_norm": 0.2412109375, "learning_rate": 1.0547908840530197e-05, "loss": 0.8892, "step": 3280 }, { "epoch": 0.8685880486515072, "grad_norm": 0.1513671875, "learning_rate": 1.034251902543908e-05, "loss": 0.8842, "step": 3285 }, { "epoch": 0.8699101004759386, "grad_norm": 0.1669921875, "learning_rate": 1.01390395105318e-05, "loss": 0.8858, "step": 3290 }, { "epoch": 0.8712321523003702, "grad_norm": 0.1904296875, "learning_rate": 9.937474631274257e-06, "loss": 0.8896, "step": 3295 }, { "epoch": 0.8725542041248017, "grad_norm": 0.220703125, "learning_rate": 9.737828682337969e-06, "loss": 0.8844, "step": 3300 }, { "epoch": 0.8738762559492332, "grad_norm": 0.1806640625, "learning_rate": 9.540105917508391e-06, "loss": 0.885, "step": 3305 }, { "epoch": 0.8751983077736647, "grad_norm": 0.1826171875, "learning_rate": 9.34431054959447e-06, "loss": 0.8878, "step": 3310 }, { "epoch": 0.8765203595980963, "grad_norm": 0.1748046875, "learning_rate": 9.150446750338714e-06, "loss": 0.8867, "step": 3315 }, { "epoch": 0.8778424114225277, "grad_norm": 0.255859375, "learning_rate": 8.958518650328429e-06, "loss": 0.8846, "step": 3320 }, { "epoch": 0.8791644632469593, "grad_norm": 0.169921875, "learning_rate": 8.768530338907655e-06, "loss": 0.8856, "step": 3325 }, { "epoch": 0.8804865150713908, "grad_norm": 0.2099609375, "learning_rate": 8.580485864090015e-06, "loss": 0.8829, "step": 3330 }, { "epoch": 0.8818085668958223, "grad_norm": 0.1748046875, "learning_rate": 8.394389232472499e-06, "loss": 0.8895, "step": 3335 }, { "epoch": 0.8831306187202538, "grad_norm": 0.2470703125, "learning_rate": 8.210244409150158e-06, "loss": 0.8858, "step": 3340 }, { "epoch": 0.8844526705446853, "grad_norm": 0.189453125, "learning_rate": 8.028055317631455e-06, "loss": 0.8871, "step": 3345 }, { "epoch": 0.8857747223691169, "grad_norm": 0.2236328125, "learning_rate": 7.847825839754852e-06, "loss": 0.8837, "step": 3350 }, { "epoch": 0.8870967741935484, "grad_norm": 0.216796875, "learning_rate": 7.669559815605965e-06, "loss": 0.8839, "step": 3355 }, { "epoch": 0.88841882601798, "grad_norm": 0.1533203125, "learning_rate": 7.493261043435828e-06, "loss": 0.8802, "step": 3360 }, { "epoch": 0.8897408778424114, "grad_norm": 0.169921875, "learning_rate": 7.318933279579909e-06, "loss": 0.8831, "step": 3365 }, { "epoch": 0.891062929666843, "grad_norm": 0.189453125, "learning_rate": 7.146580238378131e-06, "loss": 0.8866, "step": 3370 }, { "epoch": 0.8923849814912744, "grad_norm": 0.1689453125, "learning_rate": 6.976205592095675e-06, "loss": 0.886, "step": 3375 }, { "epoch": 0.893707033315706, "grad_norm": 0.2119140625, "learning_rate": 6.8078129708447516e-06, "loss": 0.8897, "step": 3380 }, { "epoch": 0.8950290851401375, "grad_norm": 0.18359375, "learning_rate": 6.641405962507307e-06, "loss": 0.886, "step": 3385 }, { "epoch": 0.896351136964569, "grad_norm": 0.18359375, "learning_rate": 6.4769881126584996e-06, "loss": 0.8848, "step": 3390 }, { "epoch": 0.8976731887890005, "grad_norm": 0.1962890625, "learning_rate": 6.314562924491229e-06, "loss": 0.8863, "step": 3395 }, { "epoch": 0.8989952406134321, "grad_norm": 0.208984375, "learning_rate": 6.154133858741407e-06, "loss": 0.8861, "step": 3400 }, { "epoch": 0.9003172924378635, "grad_norm": 0.1962890625, "learning_rate": 5.995704333614327e-06, "loss": 0.8867, "step": 3405 }, { "epoch": 0.9016393442622951, "grad_norm": 0.28125, "learning_rate": 5.839277724711733e-06, "loss": 0.8874, "step": 3410 }, { "epoch": 0.9029613960867267, "grad_norm": 0.2373046875, "learning_rate": 5.6848573649599834e-06, "loss": 0.8782, "step": 3415 }, { "epoch": 0.9042834479111581, "grad_norm": 0.185546875, "learning_rate": 5.53244654453896e-06, "loss": 0.8851, "step": 3420 }, { "epoch": 0.9056054997355897, "grad_norm": 0.1953125, "learning_rate": 5.382048510812032e-06, "loss": 0.8814, "step": 3425 }, { "epoch": 0.9069275515600211, "grad_norm": 0.2294921875, "learning_rate": 5.233666468256804e-06, "loss": 0.888, "step": 3430 }, { "epoch": 0.9082496033844527, "grad_norm": 0.2041015625, "learning_rate": 5.0873035783969066e-06, "loss": 0.8862, "step": 3435 }, { "epoch": 0.9095716552088842, "grad_norm": 0.16796875, "learning_rate": 4.942962959734565e-06, "loss": 0.8835, "step": 3440 }, { "epoch": 0.9108937070333157, "grad_norm": 0.1904296875, "learning_rate": 4.800647687684223e-06, "loss": 0.8893, "step": 3445 }, { "epoch": 0.9122157588577472, "grad_norm": 0.2197265625, "learning_rate": 4.660360794506946e-06, "loss": 0.8826, "step": 3450 }, { "epoch": 0.9135378106821788, "grad_norm": 0.166015625, "learning_rate": 4.52210526924588e-06, "loss": 0.8845, "step": 3455 }, { "epoch": 0.9148598625066102, "grad_norm": 0.1943359375, "learning_rate": 4.385884057662515e-06, "loss": 0.8862, "step": 3460 }, { "epoch": 0.9161819143310418, "grad_norm": 0.2216796875, "learning_rate": 4.251700062173947e-06, "loss": 0.886, "step": 3465 }, { "epoch": 0.9175039661554732, "grad_norm": 0.220703125, "learning_rate": 4.119556141791048e-06, "loss": 0.888, "step": 3470 }, { "epoch": 0.9188260179799048, "grad_norm": 0.1826171875, "learning_rate": 3.989455112057483e-06, "loss": 0.8857, "step": 3475 }, { "epoch": 0.9201480698043363, "grad_norm": 0.2451171875, "learning_rate": 3.861399744989813e-06, "loss": 0.8871, "step": 3480 }, { "epoch": 0.9214701216287678, "grad_norm": 0.203125, "learning_rate": 3.7353927690183867e-06, "loss": 0.8852, "step": 3485 }, { "epoch": 0.9227921734531994, "grad_norm": 0.1884765625, "learning_rate": 3.6114368689291856e-06, "loss": 0.8834, "step": 3490 }, { "epoch": 0.9241142252776309, "grad_norm": 0.267578125, "learning_rate": 3.4895346858066724e-06, "loss": 0.8847, "step": 3495 }, { "epoch": 0.9254362771020624, "grad_norm": 0.2138671875, "learning_rate": 3.3696888169774677e-06, "loss": 0.8825, "step": 3500 }, { "epoch": 0.9267583289264939, "grad_norm": 0.1806640625, "learning_rate": 3.2519018159550496e-06, "loss": 0.8884, "step": 3505 }, { "epoch": 0.9280803807509255, "grad_norm": 0.3671875, "learning_rate": 3.1361761923853095e-06, "loss": 0.8816, "step": 3510 }, { "epoch": 0.9294024325753569, "grad_norm": 0.1904296875, "learning_rate": 3.0225144119931158e-06, "loss": 0.8828, "step": 3515 }, { "epoch": 0.9307244843997885, "grad_norm": 0.2734375, "learning_rate": 2.910918896529757e-06, "loss": 0.8851, "step": 3520 }, { "epoch": 0.93204653622422, "grad_norm": 0.236328125, "learning_rate": 2.8013920237213277e-06, "loss": 0.8851, "step": 3525 }, { "epoch": 0.9333685880486515, "grad_norm": 0.2451171875, "learning_rate": 2.693936127218133e-06, "loss": 0.8877, "step": 3530 }, { "epoch": 0.934690639873083, "grad_norm": 0.2294921875, "learning_rate": 2.5885534965448768e-06, "loss": 0.8848, "step": 3535 }, { "epoch": 0.9360126916975146, "grad_norm": 0.1513671875, "learning_rate": 2.4852463770519417e-06, "loss": 0.8823, "step": 3540 }, { "epoch": 0.937334743521946, "grad_norm": 0.2197265625, "learning_rate": 2.3840169698675087e-06, "loss": 0.8865, "step": 3545 }, { "epoch": 0.9386567953463776, "grad_norm": 0.2431640625, "learning_rate": 2.2848674318507125e-06, "loss": 0.8829, "step": 3550 }, { "epoch": 0.9399788471708092, "grad_norm": 0.2216796875, "learning_rate": 2.1877998755456266e-06, "loss": 0.8851, "step": 3555 }, { "epoch": 0.9413008989952406, "grad_norm": 0.1884765625, "learning_rate": 2.0928163691362856e-06, "loss": 0.8876, "step": 3560 }, { "epoch": 0.9426229508196722, "grad_norm": 0.169921875, "learning_rate": 1.9999189364025894e-06, "loss": 0.8861, "step": 3565 }, { "epoch": 0.9439450026441036, "grad_norm": 0.2275390625, "learning_rate": 1.909109556677269e-06, "loss": 0.8874, "step": 3570 }, { "epoch": 0.9452670544685352, "grad_norm": 0.1982421875, "learning_rate": 1.8203901648035648e-06, "loss": 0.8865, "step": 3575 }, { "epoch": 0.9465891062929667, "grad_norm": 0.181640625, "learning_rate": 1.7337626510941619e-06, "loss": 0.8876, "step": 3580 }, { "epoch": 0.9479111581173982, "grad_norm": 0.3046875, "learning_rate": 1.6492288612907748e-06, "loss": 0.8889, "step": 3585 }, { "epoch": 0.9492332099418297, "grad_norm": 0.208984375, "learning_rate": 1.5667905965249363e-06, "loss": 0.8842, "step": 3590 }, { "epoch": 0.9505552617662613, "grad_norm": 0.1630859375, "learning_rate": 1.486449613279539e-06, "loss": 0.8857, "step": 3595 }, { "epoch": 0.9518773135906927, "grad_norm": 0.208984375, "learning_rate": 1.408207623351454e-06, "loss": 0.8847, "step": 3600 }, { "epoch": 0.9531993654151243, "grad_norm": 0.2353515625, "learning_rate": 1.3320662938150485e-06, "loss": 0.8828, "step": 3605 }, { "epoch": 0.9545214172395557, "grad_norm": 0.1669921875, "learning_rate": 1.2580272469866483e-06, "loss": 0.8787, "step": 3610 }, { "epoch": 0.9558434690639873, "grad_norm": 0.2080078125, "learning_rate": 1.186092060389976e-06, "loss": 0.886, "step": 3615 }, { "epoch": 0.9571655208884188, "grad_norm": 0.146484375, "learning_rate": 1.1162622667226008e-06, "loss": 0.8837, "step": 3620 }, { "epoch": 0.9584875727128503, "grad_norm": 0.1806640625, "learning_rate": 1.0485393538231747e-06, "loss": 0.8846, "step": 3625 }, { "epoch": 0.9598096245372819, "grad_norm": 0.234375, "learning_rate": 9.829247646398144e-07, "loss": 0.8867, "step": 3630 }, { "epoch": 0.9611316763617134, "grad_norm": 0.1923828125, "learning_rate": 9.194198971993362e-07, "loss": 0.8869, "step": 3635 }, { "epoch": 0.9624537281861449, "grad_norm": 0.265625, "learning_rate": 8.580261045774474e-07, "loss": 0.8824, "step": 3640 }, { "epoch": 0.9637757800105764, "grad_norm": 0.16796875, "learning_rate": 7.98744694869924e-07, "loss": 0.8875, "step": 3645 }, { "epoch": 0.965097831835008, "grad_norm": 0.291015625, "learning_rate": 7.41576931164778e-07, "loss": 0.8856, "step": 3650 }, { "epoch": 0.9664198836594394, "grad_norm": 0.2236328125, "learning_rate": 6.86524031515301e-07, "loss": 0.8861, "step": 3655 }, { "epoch": 0.967741935483871, "grad_norm": 0.1806640625, "learning_rate": 6.335871689140959e-07, "loss": 0.8868, "step": 3660 }, { "epoch": 0.9690639873083025, "grad_norm": 0.1650390625, "learning_rate": 5.827674712681752e-07, "loss": 0.8873, "step": 3665 }, { "epoch": 0.970386039132734, "grad_norm": 0.2138671875, "learning_rate": 5.340660213748017e-07, "loss": 0.8833, "step": 3670 }, { "epoch": 0.9717080909571655, "grad_norm": 0.2041015625, "learning_rate": 4.874838568985296e-07, "loss": 0.8803, "step": 3675 }, { "epoch": 0.973030142781597, "grad_norm": 0.1962890625, "learning_rate": 4.430219703490335e-07, "loss": 0.8846, "step": 3680 }, { "epoch": 0.9743521946060285, "grad_norm": 0.28515625, "learning_rate": 4.0068130905996924e-07, "loss": 0.8873, "step": 3685 }, { "epoch": 0.9756742464304601, "grad_norm": 0.2392578125, "learning_rate": 3.604627751688239e-07, "loss": 0.8817, "step": 3690 }, { "epoch": 0.9769962982548915, "grad_norm": 0.1845703125, "learning_rate": 3.2236722559764176e-07, "loss": 0.8877, "step": 3695 }, { "epoch": 0.9783183500793231, "grad_norm": 0.19140625, "learning_rate": 2.86395472034795e-07, "loss": 0.8879, "step": 3700 }, { "epoch": 0.9796404019037547, "grad_norm": 0.2578125, "learning_rate": 2.525482809177082e-07, "loss": 0.8908, "step": 3705 }, { "epoch": 0.9809624537281861, "grad_norm": 0.173828125, "learning_rate": 2.2082637341647173e-07, "loss": 0.881, "step": 3710 }, { "epoch": 0.9822845055526177, "grad_norm": 0.2080078125, "learning_rate": 1.9123042541854265e-07, "loss": 0.8863, "step": 3715 }, { "epoch": 0.9836065573770492, "grad_norm": 0.16796875, "learning_rate": 1.6376106751430087e-07, "loss": 0.8909, "step": 3720 }, { "epoch": 0.9849286092014807, "grad_norm": 0.21484375, "learning_rate": 1.3841888498361544e-07, "loss": 0.8815, "step": 3725 }, { "epoch": 0.9862506610259122, "grad_norm": 0.23828125, "learning_rate": 1.1520441778339885e-07, "loss": 0.8851, "step": 3730 }, { "epoch": 0.9875727128503438, "grad_norm": 0.259765625, "learning_rate": 9.411816053608302e-08, "loss": 0.8855, "step": 3735 }, { "epoch": 0.9888947646747752, "grad_norm": 0.197265625, "learning_rate": 7.51605625190721e-08, "loss": 0.8903, "step": 3740 }, { "epoch": 0.9902168164992068, "grad_norm": 0.1845703125, "learning_rate": 5.833202765519463e-08, "loss": 0.8867, "step": 3745 }, { "epoch": 0.9915388683236382, "grad_norm": 0.1943359375, "learning_rate": 4.3632914504077026e-08, "loss": 0.8839, "step": 3750 }, { "epoch": 0.9928609201480698, "grad_norm": 0.193359375, "learning_rate": 3.106353625451641e-08, "loss": 0.8824, "step": 3755 }, { "epoch": 0.9941829719725013, "grad_norm": 0.1708984375, "learning_rate": 2.062416071780815e-08, "loss": 0.8881, "step": 3760 }, { "epoch": 0.9955050237969328, "grad_norm": 0.1962890625, "learning_rate": 1.2315010322028198e-08, "loss": 0.8869, "step": 3765 }, { "epoch": 0.9968270756213644, "grad_norm": 0.19921875, "learning_rate": 6.136262107292456e-09, "loss": 0.8892, "step": 3770 }, { "epoch": 0.9981491274457959, "grad_norm": 0.23828125, "learning_rate": 2.0880477220042163e-09, "loss": 0.8853, "step": 3775 }, { "epoch": 0.9994711792702274, "grad_norm": 0.234375, "learning_rate": 1.7045342003418895e-10, "loss": 0.8854, "step": 3780 }, { "epoch": 1.0, "step": 3782, "total_flos": 1.011923420184576e+18, "train_loss": 0.9601045426144794, "train_runtime": 898.0661, "train_samples_per_second": 2156.094, "train_steps_per_second": 4.211 } ], "logging_steps": 5, "max_steps": 3782, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.011923420184576e+18, "train_batch_size": 16, "trial_name": null, "trial_params": null }