{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.9938347718865597, "eval_steps": 500, "global_step": 606, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.004932182490752158, "grad_norm": 1.1257907152175903, "learning_rate": 5.0000000000000004e-08, "loss": 1.0986, "step": 1 }, { "epoch": 0.009864364981504316, "grad_norm": 1.5788811445236206, "learning_rate": 1.0000000000000001e-07, "loss": 1.1152, "step": 2 }, { "epoch": 0.014796547472256474, "grad_norm": 1.7341364622116089, "learning_rate": 1.5000000000000002e-07, "loss": 1.1392, "step": 3 }, { "epoch": 0.01972872996300863, "grad_norm": 1.3568042516708374, "learning_rate": 2.0000000000000002e-07, "loss": 1.1046, "step": 4 }, { "epoch": 0.02466091245376079, "grad_norm": 1.1873835325241089, "learning_rate": 2.5000000000000004e-07, "loss": 1.1123, "step": 5 }, { "epoch": 0.029593094944512947, "grad_norm": 1.0728861093521118, "learning_rate": 3.0000000000000004e-07, "loss": 1.1024, "step": 6 }, { "epoch": 0.0345252774352651, "grad_norm": 1.991363525390625, "learning_rate": 3.5000000000000004e-07, "loss": 1.1045, "step": 7 }, { "epoch": 0.03945745992601726, "grad_norm": 1.3626810312271118, "learning_rate": 4.0000000000000003e-07, "loss": 1.1414, "step": 8 }, { "epoch": 0.04438964241676942, "grad_norm": 1.2338333129882812, "learning_rate": 4.5000000000000003e-07, "loss": 1.1312, "step": 9 }, { "epoch": 0.04932182490752158, "grad_norm": 1.2466716766357422, "learning_rate": 5.000000000000001e-07, "loss": 1.0895, "step": 10 }, { "epoch": 0.05425400739827373, "grad_norm": 1.360613465309143, "learning_rate": 5.5e-07, "loss": 1.1285, "step": 11 }, { "epoch": 0.059186189889025895, "grad_norm": 2.145904541015625, "learning_rate": 6.000000000000001e-07, "loss": 1.079, "step": 12 }, { "epoch": 0.06411837237977805, "grad_norm": 1.5354256629943848, "learning_rate": 6.5e-07, "loss": 1.0897, "step": 13 }, { "epoch": 0.0690505548705302, "grad_norm": 1.2002555131912231, "learning_rate": 7.000000000000001e-07, "loss": 1.0474, "step": 14 }, { "epoch": 0.07398273736128237, "grad_norm": 1.213218331336975, "learning_rate": 7.5e-07, "loss": 1.0336, "step": 15 }, { "epoch": 0.07891491985203453, "grad_norm": 1.0160884857177734, "learning_rate": 8.000000000000001e-07, "loss": 1.0396, "step": 16 }, { "epoch": 0.08384710234278668, "grad_norm": 1.3291572332382202, "learning_rate": 8.500000000000001e-07, "loss": 1.0111, "step": 17 }, { "epoch": 0.08877928483353884, "grad_norm": 9.560614585876465, "learning_rate": 9.000000000000001e-07, "loss": 1.0343, "step": 18 }, { "epoch": 0.093711467324291, "grad_norm": 0.8536161184310913, "learning_rate": 9.500000000000001e-07, "loss": 1.0288, "step": 19 }, { "epoch": 0.09864364981504316, "grad_norm": 0.8779953122138977, "learning_rate": 1.0000000000000002e-06, "loss": 0.9296, "step": 20 }, { "epoch": 0.10357583230579531, "grad_norm": 0.9102424383163452, "learning_rate": 1.0500000000000001e-06, "loss": 0.9348, "step": 21 }, { "epoch": 0.10850801479654747, "grad_norm": 0.5936668515205383, "learning_rate": 1.1e-06, "loss": 0.9103, "step": 22 }, { "epoch": 0.11344019728729964, "grad_norm": 0.38109758496284485, "learning_rate": 1.1500000000000002e-06, "loss": 0.9412, "step": 23 }, { "epoch": 0.11837237977805179, "grad_norm": 0.6523936986923218, "learning_rate": 1.2000000000000002e-06, "loss": 0.8968, "step": 24 }, { "epoch": 0.12330456226880394, "grad_norm": 0.4067033529281616, "learning_rate": 1.25e-06, "loss": 0.9031, "step": 25 }, { "epoch": 0.1282367447595561, "grad_norm": 0.4689585566520691, "learning_rate": 1.3e-06, "loss": 0.8724, "step": 26 }, { "epoch": 0.13316892725030827, "grad_norm": 0.5714285373687744, "learning_rate": 1.3500000000000002e-06, "loss": 0.9195, "step": 27 }, { "epoch": 0.1381011097410604, "grad_norm": 0.43251562118530273, "learning_rate": 1.4000000000000001e-06, "loss": 0.8937, "step": 28 }, { "epoch": 0.14303329223181258, "grad_norm": 0.3626736104488373, "learning_rate": 1.45e-06, "loss": 0.8685, "step": 29 }, { "epoch": 0.14796547472256474, "grad_norm": 0.3419496715068817, "learning_rate": 1.5e-06, "loss": 0.8833, "step": 30 }, { "epoch": 0.15289765721331688, "grad_norm": 0.4578935503959656, "learning_rate": 1.5500000000000002e-06, "loss": 0.8287, "step": 31 }, { "epoch": 0.15782983970406905, "grad_norm": 0.4360993206501007, "learning_rate": 1.6000000000000001e-06, "loss": 0.8452, "step": 32 }, { "epoch": 0.16276202219482122, "grad_norm": 0.5611799359321594, "learning_rate": 1.6500000000000003e-06, "loss": 0.872, "step": 33 }, { "epoch": 0.16769420468557336, "grad_norm": 0.2829279601573944, "learning_rate": 1.7000000000000002e-06, "loss": 0.87, "step": 34 }, { "epoch": 0.17262638717632553, "grad_norm": 0.38662174344062805, "learning_rate": 1.75e-06, "loss": 0.816, "step": 35 }, { "epoch": 0.17755856966707767, "grad_norm": 0.4239096939563751, "learning_rate": 1.8000000000000001e-06, "loss": 0.8599, "step": 36 }, { "epoch": 0.18249075215782984, "grad_norm": 0.653641939163208, "learning_rate": 1.85e-06, "loss": 0.8456, "step": 37 }, { "epoch": 0.187422934648582, "grad_norm": 0.21455056965351105, "learning_rate": 1.9000000000000002e-06, "loss": 0.795, "step": 38 }, { "epoch": 0.19235511713933415, "grad_norm": 0.26148131489753723, "learning_rate": 1.9500000000000004e-06, "loss": 0.8405, "step": 39 }, { "epoch": 0.19728729963008632, "grad_norm": 0.31602978706359863, "learning_rate": 2.0000000000000003e-06, "loss": 0.8569, "step": 40 }, { "epoch": 0.20221948212083848, "grad_norm": 0.20896084606647491, "learning_rate": 2.05e-06, "loss": 0.801, "step": 41 }, { "epoch": 0.20715166461159062, "grad_norm": 0.21203891932964325, "learning_rate": 2.1000000000000002e-06, "loss": 0.8202, "step": 42 }, { "epoch": 0.2120838471023428, "grad_norm": 0.23695039749145508, "learning_rate": 2.15e-06, "loss": 0.8088, "step": 43 }, { "epoch": 0.21701602959309493, "grad_norm": 0.1736220270395279, "learning_rate": 2.2e-06, "loss": 0.8031, "step": 44 }, { "epoch": 0.2219482120838471, "grad_norm": 0.23390308022499084, "learning_rate": 2.25e-06, "loss": 0.7914, "step": 45 }, { "epoch": 0.22688039457459927, "grad_norm": 0.23420561850070953, "learning_rate": 2.3000000000000004e-06, "loss": 0.8092, "step": 46 }, { "epoch": 0.2318125770653514, "grad_norm": 0.19289737939834595, "learning_rate": 2.35e-06, "loss": 0.7767, "step": 47 }, { "epoch": 0.23674475955610358, "grad_norm": 0.23044224083423615, "learning_rate": 2.4000000000000003e-06, "loss": 0.8185, "step": 48 }, { "epoch": 0.24167694204685575, "grad_norm": 0.2052794247865677, "learning_rate": 2.4500000000000003e-06, "loss": 0.781, "step": 49 }, { "epoch": 0.2466091245376079, "grad_norm": 0.19429263472557068, "learning_rate": 2.5e-06, "loss": 0.8125, "step": 50 }, { "epoch": 0.25154130702836003, "grad_norm": 0.18123206496238708, "learning_rate": 2.55e-06, "loss": 0.7775, "step": 51 }, { "epoch": 0.2564734895191122, "grad_norm": 1.8012878894805908, "learning_rate": 2.6e-06, "loss": 0.7692, "step": 52 }, { "epoch": 0.26140567200986436, "grad_norm": 0.2761130928993225, "learning_rate": 2.6500000000000005e-06, "loss": 0.7996, "step": 53 }, { "epoch": 0.26633785450061653, "grad_norm": 0.1579173356294632, "learning_rate": 2.7000000000000004e-06, "loss": 0.7936, "step": 54 }, { "epoch": 0.2712700369913687, "grad_norm": 0.16990424692630768, "learning_rate": 2.7500000000000004e-06, "loss": 0.8067, "step": 55 }, { "epoch": 0.2762022194821208, "grad_norm": 1.0547306537628174, "learning_rate": 2.8000000000000003e-06, "loss": 0.7894, "step": 56 }, { "epoch": 0.281134401972873, "grad_norm": 0.18402548134326935, "learning_rate": 2.85e-06, "loss": 0.7564, "step": 57 }, { "epoch": 0.28606658446362515, "grad_norm": 0.19761355221271515, "learning_rate": 2.9e-06, "loss": 0.7754, "step": 58 }, { "epoch": 0.2909987669543773, "grad_norm": 0.26926949620246887, "learning_rate": 2.95e-06, "loss": 0.7698, "step": 59 }, { "epoch": 0.2959309494451295, "grad_norm": 0.19200852513313293, "learning_rate": 3e-06, "loss": 0.757, "step": 60 }, { "epoch": 0.3008631319358816, "grad_norm": 0.14297647774219513, "learning_rate": 3.05e-06, "loss": 0.7372, "step": 61 }, { "epoch": 0.30579531442663377, "grad_norm": 0.1689106523990631, "learning_rate": 3.1000000000000004e-06, "loss": 0.7511, "step": 62 }, { "epoch": 0.31072749691738594, "grad_norm": 0.3330126106739044, "learning_rate": 3.1500000000000003e-06, "loss": 0.747, "step": 63 }, { "epoch": 0.3156596794081381, "grad_norm": 0.2532431483268738, "learning_rate": 3.2000000000000003e-06, "loss": 0.798, "step": 64 }, { "epoch": 0.3205918618988903, "grad_norm": 0.14446307718753815, "learning_rate": 3.2500000000000002e-06, "loss": 0.7631, "step": 65 }, { "epoch": 0.32552404438964244, "grad_norm": 0.24439038336277008, "learning_rate": 3.3000000000000006e-06, "loss": 0.7405, "step": 66 }, { "epoch": 0.33045622688039455, "grad_norm": 0.15191350877285004, "learning_rate": 3.3500000000000005e-06, "loss": 0.7559, "step": 67 }, { "epoch": 0.3353884093711467, "grad_norm": 0.16212493181228638, "learning_rate": 3.4000000000000005e-06, "loss": 0.7452, "step": 68 }, { "epoch": 0.3403205918618989, "grad_norm": 0.1210586428642273, "learning_rate": 3.45e-06, "loss": 0.7817, "step": 69 }, { "epoch": 0.34525277435265106, "grad_norm": 0.35743728280067444, "learning_rate": 3.5e-06, "loss": 0.7811, "step": 70 }, { "epoch": 0.35018495684340323, "grad_norm": 0.12270136177539825, "learning_rate": 3.5500000000000003e-06, "loss": 0.7679, "step": 71 }, { "epoch": 0.35511713933415534, "grad_norm": 0.14078450202941895, "learning_rate": 3.6000000000000003e-06, "loss": 0.7479, "step": 72 }, { "epoch": 0.3600493218249075, "grad_norm": 0.11881807446479797, "learning_rate": 3.65e-06, "loss": 0.7475, "step": 73 }, { "epoch": 0.3649815043156597, "grad_norm": 0.11201618611812592, "learning_rate": 3.7e-06, "loss": 0.7396, "step": 74 }, { "epoch": 0.36991368680641185, "grad_norm": 0.14292244613170624, "learning_rate": 3.7500000000000005e-06, "loss": 0.7117, "step": 75 }, { "epoch": 0.374845869297164, "grad_norm": 0.18425297737121582, "learning_rate": 3.8000000000000005e-06, "loss": 0.718, "step": 76 }, { "epoch": 0.3797780517879161, "grad_norm": 0.16399289667606354, "learning_rate": 3.85e-06, "loss": 0.7314, "step": 77 }, { "epoch": 0.3847102342786683, "grad_norm": 0.1419249176979065, "learning_rate": 3.900000000000001e-06, "loss": 0.7996, "step": 78 }, { "epoch": 0.38964241676942046, "grad_norm": 0.154827281832695, "learning_rate": 3.95e-06, "loss": 0.7308, "step": 79 }, { "epoch": 0.39457459926017263, "grad_norm": 0.17852246761322021, "learning_rate": 4.000000000000001e-06, "loss": 0.7161, "step": 80 }, { "epoch": 0.3995067817509248, "grad_norm": 0.12098474055528641, "learning_rate": 4.05e-06, "loss": 0.7581, "step": 81 }, { "epoch": 0.40443896424167697, "grad_norm": 0.15087738633155823, "learning_rate": 4.1e-06, "loss": 0.728, "step": 82 }, { "epoch": 0.4093711467324291, "grad_norm": 0.23446552455425262, "learning_rate": 4.15e-06, "loss": 0.7089, "step": 83 }, { "epoch": 0.41430332922318125, "grad_norm": 0.12758222222328186, "learning_rate": 4.2000000000000004e-06, "loss": 0.7513, "step": 84 }, { "epoch": 0.4192355117139334, "grad_norm": 0.18490324914455414, "learning_rate": 4.25e-06, "loss": 0.7232, "step": 85 }, { "epoch": 0.4241676942046856, "grad_norm": 3.9814538955688477, "learning_rate": 4.3e-06, "loss": 0.7071, "step": 86 }, { "epoch": 0.42909987669543775, "grad_norm": 0.13754811882972717, "learning_rate": 4.350000000000001e-06, "loss": 0.7218, "step": 87 }, { "epoch": 0.43403205918618987, "grad_norm": 0.15279600024223328, "learning_rate": 4.4e-06, "loss": 0.7531, "step": 88 }, { "epoch": 0.43896424167694204, "grad_norm": 0.13091601431369781, "learning_rate": 4.450000000000001e-06, "loss": 0.6954, "step": 89 }, { "epoch": 0.4438964241676942, "grad_norm": 0.14220909774303436, "learning_rate": 4.5e-06, "loss": 0.7227, "step": 90 }, { "epoch": 0.44882860665844637, "grad_norm": 0.11386663466691971, "learning_rate": 4.5500000000000005e-06, "loss": 0.7196, "step": 91 }, { "epoch": 0.45376078914919854, "grad_norm": 0.16582414507865906, "learning_rate": 4.600000000000001e-06, "loss": 0.7155, "step": 92 }, { "epoch": 0.45869297163995065, "grad_norm": 0.2046297937631607, "learning_rate": 4.65e-06, "loss": 0.7201, "step": 93 }, { "epoch": 0.4636251541307028, "grad_norm": 0.11329102516174316, "learning_rate": 4.7e-06, "loss": 0.7233, "step": 94 }, { "epoch": 0.468557336621455, "grad_norm": 0.1256910264492035, "learning_rate": 4.75e-06, "loss": 0.7294, "step": 95 }, { "epoch": 0.47348951911220716, "grad_norm": 0.16878068447113037, "learning_rate": 4.800000000000001e-06, "loss": 0.7292, "step": 96 }, { "epoch": 0.4784217016029593, "grad_norm": 0.1140855923295021, "learning_rate": 4.85e-06, "loss": 0.7215, "step": 97 }, { "epoch": 0.4833538840937115, "grad_norm": 0.13400927186012268, "learning_rate": 4.9000000000000005e-06, "loss": 0.7446, "step": 98 }, { "epoch": 0.4882860665844636, "grad_norm": 0.13537898659706116, "learning_rate": 4.95e-06, "loss": 0.7083, "step": 99 }, { "epoch": 0.4932182490752158, "grad_norm": 0.23342657089233398, "learning_rate": 5e-06, "loss": 0.7172, "step": 100 }, { "epoch": 0.49815043156596794, "grad_norm": 2.2964117527008057, "learning_rate": 4.999951815503011e-06, "loss": 0.7564, "step": 101 }, { "epoch": 0.5030826140567201, "grad_norm": 0.20878185331821442, "learning_rate": 4.999807263869441e-06, "loss": 0.7255, "step": 102 }, { "epoch": 0.5080147965474723, "grad_norm": 0.26244527101516724, "learning_rate": 4.999566350671405e-06, "loss": 0.707, "step": 103 }, { "epoch": 0.5129469790382244, "grad_norm": 0.12257402390241623, "learning_rate": 4.999229085195532e-06, "loss": 0.7147, "step": 104 }, { "epoch": 0.5178791615289766, "grad_norm": 0.18218959867954254, "learning_rate": 4.998795480442595e-06, "loss": 0.7113, "step": 105 }, { "epoch": 0.5228113440197287, "grad_norm": 0.1487857848405838, "learning_rate": 4.998265553127013e-06, "loss": 0.7532, "step": 106 }, { "epoch": 0.5277435265104808, "grad_norm": 0.18883858621120453, "learning_rate": 4.997639323676214e-06, "loss": 0.6822, "step": 107 }, { "epoch": 0.5326757090012331, "grad_norm": 0.13591569662094116, "learning_rate": 4.996916816229838e-06, "loss": 0.7569, "step": 108 }, { "epoch": 0.5376078914919852, "grad_norm": 0.9631307721138, "learning_rate": 4.99609805863881e-06, "loss": 0.745, "step": 109 }, { "epoch": 0.5425400739827374, "grad_norm": 0.13628794252872467, "learning_rate": 4.995183082464269e-06, "loss": 0.732, "step": 110 }, { "epoch": 0.5474722564734895, "grad_norm": 2.4701926708221436, "learning_rate": 4.994171922976349e-06, "loss": 0.756, "step": 111 }, { "epoch": 0.5524044389642416, "grad_norm": 1.1472599506378174, "learning_rate": 4.993064619152818e-06, "loss": 0.7427, "step": 112 }, { "epoch": 0.5573366214549939, "grad_norm": 0.11852557212114334, "learning_rate": 4.991861213677578e-06, "loss": 0.7124, "step": 113 }, { "epoch": 0.562268803945746, "grad_norm": 0.09930302202701569, "learning_rate": 4.99056175293902e-06, "loss": 0.7065, "step": 114 }, { "epoch": 0.5672009864364982, "grad_norm": 0.12781856954097748, "learning_rate": 4.989166287028234e-06, "loss": 0.7298, "step": 115 }, { "epoch": 0.5721331689272503, "grad_norm": 0.39002031087875366, "learning_rate": 4.987674869737078e-06, "loss": 0.6996, "step": 116 }, { "epoch": 0.5770653514180024, "grad_norm": 0.1187821626663208, "learning_rate": 4.986087558556104e-06, "loss": 0.7333, "step": 117 }, { "epoch": 0.5819975339087546, "grad_norm": 0.11041084676980972, "learning_rate": 4.984404414672346e-06, "loss": 0.7101, "step": 118 }, { "epoch": 0.5869297163995068, "grad_norm": 0.1762731671333313, "learning_rate": 4.9826255029669575e-06, "loss": 0.7063, "step": 119 }, { "epoch": 0.591861898890259, "grad_norm": 0.12481699883937836, "learning_rate": 4.980750892012711e-06, "loss": 0.725, "step": 120 }, { "epoch": 0.5967940813810111, "grad_norm": 0.7090242505073547, "learning_rate": 4.978780654071355e-06, "loss": 0.7105, "step": 121 }, { "epoch": 0.6017262638717632, "grad_norm": 0.11730094999074936, "learning_rate": 4.976714865090827e-06, "loss": 0.7091, "step": 122 }, { "epoch": 0.6066584463625154, "grad_norm": 0.12327645719051361, "learning_rate": 4.974553604702332e-06, "loss": 0.739, "step": 123 }, { "epoch": 0.6115906288532675, "grad_norm": 0.22779831290245056, "learning_rate": 4.972296956217265e-06, "loss": 0.6924, "step": 124 }, { "epoch": 0.6165228113440198, "grad_norm": 0.3117700517177582, "learning_rate": 4.969945006624003e-06, "loss": 0.7166, "step": 125 }, { "epoch": 0.6214549938347719, "grad_norm": 0.1463162750005722, "learning_rate": 4.967497846584552e-06, "loss": 0.7236, "step": 126 }, { "epoch": 0.6263871763255241, "grad_norm": 0.11583642661571503, "learning_rate": 4.9649555704310545e-06, "loss": 0.6957, "step": 127 }, { "epoch": 0.6313193588162762, "grad_norm": 0.1344325989484787, "learning_rate": 4.962318276162148e-06, "loss": 0.7031, "step": 128 }, { "epoch": 0.6362515413070283, "grad_norm": 0.13892586529254913, "learning_rate": 4.959586065439189e-06, "loss": 0.6925, "step": 129 }, { "epoch": 0.6411837237977805, "grad_norm": 0.19953861832618713, "learning_rate": 4.956759043582339e-06, "loss": 0.72, "step": 130 }, { "epoch": 0.6461159062885327, "grad_norm": 0.2914772629737854, "learning_rate": 4.953837319566498e-06, "loss": 0.7175, "step": 131 }, { "epoch": 0.6510480887792849, "grad_norm": 0.109347403049469, "learning_rate": 4.950821006017107e-06, "loss": 0.6923, "step": 132 }, { "epoch": 0.655980271270037, "grad_norm": 0.2301577925682068, "learning_rate": 4.947710219205808e-06, "loss": 0.7093, "step": 133 }, { "epoch": 0.6609124537607891, "grad_norm": 0.1368824988603592, "learning_rate": 4.9445050790459585e-06, "loss": 0.6746, "step": 134 }, { "epoch": 0.6658446362515413, "grad_norm": 0.13323882222175598, "learning_rate": 4.9412057090880115e-06, "loss": 0.6919, "step": 135 }, { "epoch": 0.6707768187422934, "grad_norm": 0.14521068334579468, "learning_rate": 4.937812236514754e-06, "loss": 0.7352, "step": 136 }, { "epoch": 0.6757090012330457, "grad_norm": 0.11869832128286362, "learning_rate": 4.9343247921364e-06, "loss": 0.6759, "step": 137 }, { "epoch": 0.6806411837237978, "grad_norm": 0.20703694224357605, "learning_rate": 4.930743510385551e-06, "loss": 0.6984, "step": 138 }, { "epoch": 0.6855733662145499, "grad_norm": 0.13110846281051636, "learning_rate": 4.927068529312017e-06, "loss": 0.7143, "step": 139 }, { "epoch": 0.6905055487053021, "grad_norm": 0.13985510170459747, "learning_rate": 4.923299990577488e-06, "loss": 0.6899, "step": 140 }, { "epoch": 0.6954377311960542, "grad_norm": 0.1306031197309494, "learning_rate": 4.919438039450079e-06, "loss": 0.6801, "step": 141 }, { "epoch": 0.7003699136868065, "grad_norm": 0.10599593818187714, "learning_rate": 4.915482824798728e-06, "loss": 0.7191, "step": 142 }, { "epoch": 0.7053020961775586, "grad_norm": 0.1460665911436081, "learning_rate": 4.911434499087457e-06, "loss": 0.7134, "step": 143 }, { "epoch": 0.7102342786683107, "grad_norm": 0.1587439626455307, "learning_rate": 4.907293218369499e-06, "loss": 0.6987, "step": 144 }, { "epoch": 0.7151664611590629, "grad_norm": 0.22658978402614594, "learning_rate": 4.903059142281273e-06, "loss": 0.7049, "step": 145 }, { "epoch": 0.720098643649815, "grad_norm": 0.6049544215202332, "learning_rate": 4.8987324340362445e-06, "loss": 0.709, "step": 146 }, { "epoch": 0.7250308261405672, "grad_norm": 0.1274397075176239, "learning_rate": 4.894313260418617e-06, "loss": 0.6762, "step": 147 }, { "epoch": 0.7299630086313194, "grad_norm": 0.1254238486289978, "learning_rate": 4.889801791776921e-06, "loss": 0.6598, "step": 148 }, { "epoch": 0.7348951911220715, "grad_norm": 0.12932169437408447, "learning_rate": 4.885198202017431e-06, "loss": 0.6905, "step": 149 }, { "epoch": 0.7398273736128237, "grad_norm": 0.1857711523771286, "learning_rate": 4.880502668597475e-06, "loss": 0.7186, "step": 150 }, { "epoch": 0.7447595561035758, "grad_norm": 0.11666952073574066, "learning_rate": 4.875715372518585e-06, "loss": 0.6971, "step": 151 }, { "epoch": 0.749691738594328, "grad_norm": 0.11182554066181183, "learning_rate": 4.870836498319523e-06, "loss": 0.6876, "step": 152 }, { "epoch": 0.7546239210850801, "grad_norm": 0.11282117664813995, "learning_rate": 4.865866234069169e-06, "loss": 0.7039, "step": 153 }, { "epoch": 0.7595561035758323, "grad_norm": 0.11031971871852875, "learning_rate": 4.86080477135927e-06, "loss": 0.6703, "step": 154 }, { "epoch": 0.7644882860665845, "grad_norm": 0.1600004881620407, "learning_rate": 4.855652305297052e-06, "loss": 0.6776, "step": 155 }, { "epoch": 0.7694204685573366, "grad_norm": 0.13014693558216095, "learning_rate": 4.8504090344977036e-06, "loss": 0.6926, "step": 156 }, { "epoch": 0.7743526510480888, "grad_norm": 0.17829261720180511, "learning_rate": 4.84507516107672e-06, "loss": 0.7036, "step": 157 }, { "epoch": 0.7792848335388409, "grad_norm": 0.13715577125549316, "learning_rate": 4.839650890642104e-06, "loss": 0.6788, "step": 158 }, { "epoch": 0.7842170160295932, "grad_norm": 0.16660194098949432, "learning_rate": 4.834136432286452e-06, "loss": 0.7052, "step": 159 }, { "epoch": 0.7891491985203453, "grad_norm": 0.10915841907262802, "learning_rate": 4.828531998578885e-06, "loss": 0.6484, "step": 160 }, { "epoch": 0.7940813810110974, "grad_norm": 0.11565055698156357, "learning_rate": 4.822837805556858e-06, "loss": 0.6853, "step": 161 }, { "epoch": 0.7990135635018496, "grad_norm": 0.11839542537927628, "learning_rate": 4.817054072717833e-06, "loss": 0.6647, "step": 162 }, { "epoch": 0.8039457459926017, "grad_norm": 0.11023162305355072, "learning_rate": 4.811181023010815e-06, "loss": 0.7049, "step": 163 }, { "epoch": 0.8088779284833539, "grad_norm": 0.21431177854537964, "learning_rate": 4.805218882827761e-06, "loss": 0.6942, "step": 164 }, { "epoch": 0.813810110974106, "grad_norm": 0.11522458493709564, "learning_rate": 4.799167881994852e-06, "loss": 0.6929, "step": 165 }, { "epoch": 0.8187422934648582, "grad_norm": 0.12389165163040161, "learning_rate": 4.793028253763633e-06, "loss": 0.6994, "step": 166 }, { "epoch": 0.8236744759556104, "grad_norm": 0.11536920070648193, "learning_rate": 4.786800234802022e-06, "loss": 0.6959, "step": 167 }, { "epoch": 0.8286066584463625, "grad_norm": 0.1272544115781784, "learning_rate": 4.780484065185188e-06, "loss": 0.6626, "step": 168 }, { "epoch": 0.8335388409371147, "grad_norm": 0.16449759900569916, "learning_rate": 4.7740799883862966e-06, "loss": 0.6925, "step": 169 }, { "epoch": 0.8384710234278668, "grad_norm": 0.2522273361682892, "learning_rate": 4.767588251267121e-06, "loss": 0.7052, "step": 170 }, { "epoch": 0.843403205918619, "grad_norm": 0.09476125985383987, "learning_rate": 4.761009104068533e-06, "loss": 0.6722, "step": 171 }, { "epoch": 0.8483353884093712, "grad_norm": 0.23170353472232819, "learning_rate": 4.754342800400852e-06, "loss": 0.6794, "step": 172 }, { "epoch": 0.8532675709001233, "grad_norm": 0.14337614178657532, "learning_rate": 4.747589597234068e-06, "loss": 0.6564, "step": 173 }, { "epoch": 0.8581997533908755, "grad_norm": 0.17019444704055786, "learning_rate": 4.740749754887939e-06, "loss": 0.6688, "step": 174 }, { "epoch": 0.8631319358816276, "grad_norm": 0.1297682523727417, "learning_rate": 4.7338235370219556e-06, "loss": 0.7185, "step": 175 }, { "epoch": 0.8680641183723797, "grad_norm": 0.17522914707660675, "learning_rate": 4.726811210625176e-06, "loss": 0.7061, "step": 176 }, { "epoch": 0.872996300863132, "grad_norm": 0.3105694651603699, "learning_rate": 4.7197130460059385e-06, "loss": 0.688, "step": 177 }, { "epoch": 0.8779284833538841, "grad_norm": 0.10563701391220093, "learning_rate": 4.712529316781435e-06, "loss": 0.6695, "step": 178 }, { "epoch": 0.8828606658446363, "grad_norm": 0.16776345670223236, "learning_rate": 4.705260299867169e-06, "loss": 0.6855, "step": 179 }, { "epoch": 0.8877928483353884, "grad_norm": 0.10242436826229095, "learning_rate": 4.697906275466279e-06, "loss": 0.6768, "step": 180 }, { "epoch": 0.8927250308261405, "grad_norm": 0.1279619336128235, "learning_rate": 4.69046752705874e-06, "loss": 0.6854, "step": 181 }, { "epoch": 0.8976572133168927, "grad_norm": 0.190452441573143, "learning_rate": 4.682944341390431e-06, "loss": 0.7199, "step": 182 }, { "epoch": 0.9025893958076449, "grad_norm": 0.2059255987405777, "learning_rate": 4.675337008462085e-06, "loss": 0.6694, "step": 183 }, { "epoch": 0.9075215782983971, "grad_norm": 0.10989531874656677, "learning_rate": 4.667645821518111e-06, "loss": 0.6707, "step": 184 }, { "epoch": 0.9124537607891492, "grad_norm": 0.11381805688142776, "learning_rate": 4.659871077035289e-06, "loss": 0.6617, "step": 185 }, { "epoch": 0.9173859432799013, "grad_norm": 0.18315713107585907, "learning_rate": 4.65201307471134e-06, "loss": 0.6721, "step": 186 }, { "epoch": 0.9223181257706535, "grad_norm": 0.13762731850147247, "learning_rate": 4.644072117453377e-06, "loss": 0.6691, "step": 187 }, { "epoch": 0.9272503082614056, "grad_norm": 0.13186247646808624, "learning_rate": 4.636048511366222e-06, "loss": 0.6907, "step": 188 }, { "epoch": 0.9321824907521579, "grad_norm": 0.23819328844547272, "learning_rate": 4.6279425657406154e-06, "loss": 0.7147, "step": 189 }, { "epoch": 0.93711467324291, "grad_norm": 0.3640376329421997, "learning_rate": 4.619754593041287e-06, "loss": 0.6981, "step": 190 }, { "epoch": 0.9420468557336621, "grad_norm": 0.19634462893009186, "learning_rate": 4.6114849088949146e-06, "loss": 0.6704, "step": 191 }, { "epoch": 0.9469790382244143, "grad_norm": 0.11775479465723038, "learning_rate": 4.603133832077953e-06, "loss": 0.6909, "step": 192 }, { "epoch": 0.9519112207151664, "grad_norm": 0.10143059492111206, "learning_rate": 4.594701684504352e-06, "loss": 0.6625, "step": 193 }, { "epoch": 0.9568434032059187, "grad_norm": 0.11726385354995728, "learning_rate": 4.586188791213143e-06, "loss": 0.6692, "step": 194 }, { "epoch": 0.9617755856966708, "grad_norm": 0.11105194687843323, "learning_rate": 4.577595480355911e-06, "loss": 0.6963, "step": 195 }, { "epoch": 0.966707768187423, "grad_norm": 0.10140436887741089, "learning_rate": 4.568922083184144e-06, "loss": 0.6969, "step": 196 }, { "epoch": 0.9716399506781751, "grad_norm": 0.2534874677658081, "learning_rate": 4.560168934036467e-06, "loss": 0.6852, "step": 197 }, { "epoch": 0.9765721331689272, "grad_norm": 0.10877111554145813, "learning_rate": 4.55133637032575e-06, "loss": 0.6998, "step": 198 }, { "epoch": 0.9815043156596794, "grad_norm": 0.14067795872688293, "learning_rate": 4.542424732526105e-06, "loss": 0.6685, "step": 199 }, { "epoch": 0.9864364981504316, "grad_norm": 0.7147830128669739, "learning_rate": 4.533434364159761e-06, "loss": 0.6982, "step": 200 }, { "epoch": 0.9913686806411838, "grad_norm": 0.15026246011257172, "learning_rate": 4.524365611783818e-06, "loss": 0.6915, "step": 201 }, { "epoch": 0.9963008631319359, "grad_norm": 0.12991268932819366, "learning_rate": 4.515218824976895e-06, "loss": 0.6539, "step": 202 }, { "epoch": 1.0036991368680641, "grad_norm": 0.11360019445419312, "learning_rate": 4.505994356325648e-06, "loss": 0.6479, "step": 203 }, { "epoch": 1.0086313193588163, "grad_norm": 0.1208319365978241, "learning_rate": 4.496692561411182e-06, "loss": 0.6704, "step": 204 }, { "epoch": 1.0135635018495683, "grad_norm": 0.15796445310115814, "learning_rate": 4.487313798795347e-06, "loss": 0.6585, "step": 205 }, { "epoch": 1.0184956843403206, "grad_norm": 0.2212684154510498, "learning_rate": 4.477858430006906e-06, "loss": 0.6748, "step": 206 }, { "epoch": 1.0234278668310728, "grad_norm": 0.14605554938316345, "learning_rate": 4.468326819527613e-06, "loss": 0.6671, "step": 207 }, { "epoch": 1.028360049321825, "grad_norm": 0.11415625363588333, "learning_rate": 4.458719334778153e-06, "loss": 0.681, "step": 208 }, { "epoch": 1.033292231812577, "grad_norm": 0.113845095038414, "learning_rate": 4.449036346103982e-06, "loss": 0.6803, "step": 209 }, { "epoch": 1.0382244143033292, "grad_norm": 0.10725957155227661, "learning_rate": 4.43927822676105e-06, "loss": 0.648, "step": 210 }, { "epoch": 1.0431565967940815, "grad_norm": 0.12335722148418427, "learning_rate": 4.429445352901415e-06, "loss": 0.695, "step": 211 }, { "epoch": 1.0480887792848335, "grad_norm": 0.10324176400899887, "learning_rate": 4.419538103558742e-06, "loss": 0.6529, "step": 212 }, { "epoch": 1.0530209617755857, "grad_norm": 0.10406983643770218, "learning_rate": 4.409556860633692e-06, "loss": 0.6599, "step": 213 }, { "epoch": 1.057953144266338, "grad_norm": 0.11518678814172745, "learning_rate": 4.3995020088792e-06, "loss": 0.6601, "step": 214 }, { "epoch": 1.06288532675709, "grad_norm": 0.14593787491321564, "learning_rate": 4.3893739358856465e-06, "loss": 0.6714, "step": 215 }, { "epoch": 1.0678175092478421, "grad_norm": 0.17964190244674683, "learning_rate": 4.379173032065912e-06, "loss": 0.6441, "step": 216 }, { "epoch": 1.0727496917385944, "grad_norm": 0.12014143913984299, "learning_rate": 4.368899690640333e-06, "loss": 0.6526, "step": 217 }, { "epoch": 1.0776818742293466, "grad_norm": 0.2840045690536499, "learning_rate": 4.3585543076215405e-06, "loss": 0.6688, "step": 218 }, { "epoch": 1.0826140567200986, "grad_norm": 0.23248234391212463, "learning_rate": 4.3481372817991976e-06, "loss": 0.6798, "step": 219 }, { "epoch": 1.0875462392108508, "grad_norm": 0.11116907000541687, "learning_rate": 4.3376490147246205e-06, "loss": 0.6676, "step": 220 }, { "epoch": 1.092478421701603, "grad_norm": 0.10024593770503998, "learning_rate": 4.32708991069531e-06, "loss": 0.6718, "step": 221 }, { "epoch": 1.097410604192355, "grad_norm": 0.11960247159004211, "learning_rate": 4.31646037673936e-06, "loss": 0.665, "step": 222 }, { "epoch": 1.1023427866831073, "grad_norm": 0.2154683619737625, "learning_rate": 4.305760822599766e-06, "loss": 0.6782, "step": 223 }, { "epoch": 1.1072749691738595, "grad_norm": 0.13609148561954498, "learning_rate": 4.294991660718636e-06, "loss": 0.6715, "step": 224 }, { "epoch": 1.1122071516646117, "grad_norm": 0.12187358736991882, "learning_rate": 4.284153306221289e-06, "loss": 0.6785, "step": 225 }, { "epoch": 1.1171393341553637, "grad_norm": 0.10929637402296066, "learning_rate": 4.273246176900252e-06, "loss": 0.6757, "step": 226 }, { "epoch": 1.122071516646116, "grad_norm": 0.12787918746471405, "learning_rate": 4.262270693199159e-06, "loss": 0.6895, "step": 227 }, { "epoch": 1.1270036991368682, "grad_norm": 0.11202108860015869, "learning_rate": 4.2512272781965355e-06, "loss": 0.6823, "step": 228 }, { "epoch": 1.1319358816276202, "grad_norm": 0.5738335251808167, "learning_rate": 4.240116357589502e-06, "loss": 0.6724, "step": 229 }, { "epoch": 1.1368680641183724, "grad_norm": 0.13368858397006989, "learning_rate": 4.228938359677354e-06, "loss": 0.6583, "step": 230 }, { "epoch": 1.1418002466091246, "grad_norm": 0.10556632280349731, "learning_rate": 4.217693715345057e-06, "loss": 0.6444, "step": 231 }, { "epoch": 1.1467324290998766, "grad_norm": 0.28452828526496887, "learning_rate": 4.206382858046636e-06, "loss": 0.6638, "step": 232 }, { "epoch": 1.1516646115906288, "grad_norm": 0.19593405723571777, "learning_rate": 4.195006223788466e-06, "loss": 0.6537, "step": 233 }, { "epoch": 1.156596794081381, "grad_norm": 0.11627080291509628, "learning_rate": 4.183564251112466e-06, "loss": 0.6699, "step": 234 }, { "epoch": 1.161528976572133, "grad_norm": 0.15152066946029663, "learning_rate": 4.172057381079196e-06, "loss": 0.6685, "step": 235 }, { "epoch": 1.1664611590628853, "grad_norm": 0.20805396139621735, "learning_rate": 4.160486057250849e-06, "loss": 0.6605, "step": 236 }, { "epoch": 1.1713933415536375, "grad_norm": 0.16027909517288208, "learning_rate": 4.148850725674162e-06, "loss": 0.6959, "step": 237 }, { "epoch": 1.1763255240443897, "grad_norm": 0.12061706185340881, "learning_rate": 4.137151834863213e-06, "loss": 0.6706, "step": 238 }, { "epoch": 1.1812577065351417, "grad_norm": 0.09794213622808456, "learning_rate": 4.125389835782138e-06, "loss": 0.6846, "step": 239 }, { "epoch": 1.186189889025894, "grad_norm": 0.13422715663909912, "learning_rate": 4.113565181827745e-06, "loss": 0.6611, "step": 240 }, { "epoch": 1.1911220715166462, "grad_norm": 0.11418966203927994, "learning_rate": 4.101678328812035e-06, "loss": 0.6687, "step": 241 }, { "epoch": 1.1960542540073984, "grad_norm": 0.2175075113773346, "learning_rate": 4.0897297349446345e-06, "loss": 0.6933, "step": 242 }, { "epoch": 1.2009864364981504, "grad_norm": 0.1245436891913414, "learning_rate": 4.077719860815132e-06, "loss": 0.6538, "step": 243 }, { "epoch": 1.2059186189889026, "grad_norm": 0.12700359523296356, "learning_rate": 4.065649169375324e-06, "loss": 0.6555, "step": 244 }, { "epoch": 1.2108508014796548, "grad_norm": 0.12477461248636246, "learning_rate": 4.053518125921365e-06, "loss": 0.6664, "step": 245 }, { "epoch": 1.2157829839704068, "grad_norm": 0.16408082842826843, "learning_rate": 4.041327198075838e-06, "loss": 0.6482, "step": 246 }, { "epoch": 1.220715166461159, "grad_norm": 0.11284561455249786, "learning_rate": 4.029076855769722e-06, "loss": 0.6557, "step": 247 }, { "epoch": 1.2256473489519113, "grad_norm": 0.11041481792926788, "learning_rate": 4.016767571224285e-06, "loss": 0.6445, "step": 248 }, { "epoch": 1.2305795314426633, "grad_norm": 0.1686171293258667, "learning_rate": 4.0043998189328705e-06, "loss": 0.641, "step": 249 }, { "epoch": 1.2355117139334155, "grad_norm": 0.10102049261331558, "learning_rate": 3.991974075642621e-06, "loss": 0.6522, "step": 250 }, { "epoch": 1.2404438964241677, "grad_norm": 0.22298377752304077, "learning_rate": 3.9794908203360865e-06, "loss": 0.6942, "step": 251 }, { "epoch": 1.2453760789149197, "grad_norm": 0.11277928948402405, "learning_rate": 3.966950534212769e-06, "loss": 0.647, "step": 252 }, { "epoch": 1.250308261405672, "grad_norm": 0.12690778076648712, "learning_rate": 3.954353700670573e-06, "loss": 0.6682, "step": 253 }, { "epoch": 1.2552404438964242, "grad_norm": 0.1330857127904892, "learning_rate": 3.941700805287169e-06, "loss": 0.6902, "step": 254 }, { "epoch": 1.2601726263871762, "grad_norm": 0.10677850991487503, "learning_rate": 3.9289923358012735e-06, "loss": 0.6523, "step": 255 }, { "epoch": 1.2651048088779284, "grad_norm": 0.09997207671403885, "learning_rate": 3.9162287820938575e-06, "loss": 0.6396, "step": 256 }, { "epoch": 1.2700369913686806, "grad_norm": 0.10573034733533859, "learning_rate": 3.903410636169252e-06, "loss": 0.6631, "step": 257 }, { "epoch": 1.2749691738594329, "grad_norm": 0.11500944197177887, "learning_rate": 3.890538392136188e-06, "loss": 0.6485, "step": 258 }, { "epoch": 1.279901356350185, "grad_norm": 0.11942487210035324, "learning_rate": 3.877612546188749e-06, "loss": 0.6783, "step": 259 }, { "epoch": 1.284833538840937, "grad_norm": 0.22677327692508698, "learning_rate": 3.864633596587242e-06, "loss": 0.6761, "step": 260 }, { "epoch": 1.2897657213316893, "grad_norm": 0.10834953933954239, "learning_rate": 3.8516020436389945e-06, "loss": 0.6667, "step": 261 }, { "epoch": 1.2946979038224415, "grad_norm": 0.10524627566337585, "learning_rate": 3.838518389679065e-06, "loss": 0.6722, "step": 262 }, { "epoch": 1.2996300863131935, "grad_norm": 0.20889033377170563, "learning_rate": 3.825383139050881e-06, "loss": 0.6471, "step": 263 }, { "epoch": 1.3045622688039458, "grad_norm": 0.09844981133937836, "learning_rate": 3.812196798086799e-06, "loss": 0.6847, "step": 264 }, { "epoch": 1.309494451294698, "grad_norm": 0.10606922209262848, "learning_rate": 3.798959875088584e-06, "loss": 0.666, "step": 265 }, { "epoch": 1.31442663378545, "grad_norm": 0.1338571161031723, "learning_rate": 3.7856728803078168e-06, "loss": 0.6438, "step": 266 }, { "epoch": 1.3193588162762022, "grad_norm": 0.09691350907087326, "learning_rate": 3.7723363259262253e-06, "loss": 0.6567, "step": 267 }, { "epoch": 1.3242909987669544, "grad_norm": 0.12978605926036835, "learning_rate": 3.75895072603594e-06, "loss": 0.654, "step": 268 }, { "epoch": 1.3292231812577064, "grad_norm": 0.11295681446790695, "learning_rate": 3.7455165966196817e-06, "loss": 0.6981, "step": 269 }, { "epoch": 1.3341553637484587, "grad_norm": 0.11601907014846802, "learning_rate": 3.732034455530863e-06, "loss": 0.6891, "step": 270 }, { "epoch": 1.339087546239211, "grad_norm": 0.10687986761331558, "learning_rate": 3.718504822473634e-06, "loss": 0.6606, "step": 271 }, { "epoch": 1.344019728729963, "grad_norm": 0.11528074741363525, "learning_rate": 3.704928218982845e-06, "loss": 0.6481, "step": 272 }, { "epoch": 1.3489519112207151, "grad_norm": 0.18344081938266754, "learning_rate": 3.6913051684039435e-06, "loss": 0.6645, "step": 273 }, { "epoch": 1.3538840937114673, "grad_norm": 0.2689914405345917, "learning_rate": 3.6776361958728025e-06, "loss": 0.6648, "step": 274 }, { "epoch": 1.3588162762022196, "grad_norm": 0.3526904582977295, "learning_rate": 3.663921828295474e-06, "loss": 0.6679, "step": 275 }, { "epoch": 1.3637484586929716, "grad_norm": 0.12253173440694809, "learning_rate": 3.650162594327881e-06, "loss": 0.649, "step": 276 }, { "epoch": 1.3686806411837238, "grad_norm": 0.11910586804151535, "learning_rate": 3.6363590243554362e-06, "loss": 0.6372, "step": 277 }, { "epoch": 1.373612823674476, "grad_norm": 0.2642151117324829, "learning_rate": 3.6225116504726014e-06, "loss": 0.6465, "step": 278 }, { "epoch": 1.3785450061652282, "grad_norm": 0.12019569426774979, "learning_rate": 3.6086210064623735e-06, "loss": 0.6815, "step": 279 }, { "epoch": 1.3834771886559802, "grad_norm": 0.12397979199886322, "learning_rate": 3.5946876277757066e-06, "loss": 0.6836, "step": 280 }, { "epoch": 1.3884093711467325, "grad_norm": 0.12133345752954483, "learning_rate": 3.580712051510876e-06, "loss": 0.6514, "step": 281 }, { "epoch": 1.3933415536374847, "grad_norm": 0.12849684059619904, "learning_rate": 3.5666948163927716e-06, "loss": 0.6651, "step": 282 }, { "epoch": 1.3982737361282367, "grad_norm": 0.10554145276546478, "learning_rate": 3.5526364627521322e-06, "loss": 0.6784, "step": 283 }, { "epoch": 1.403205918618989, "grad_norm": 0.12648458778858185, "learning_rate": 3.5385375325047167e-06, "loss": 0.6738, "step": 284 }, { "epoch": 1.4081381011097411, "grad_norm": 0.10506008565425873, "learning_rate": 3.5243985691304146e-06, "loss": 0.6478, "step": 285 }, { "epoch": 1.4130702836004931, "grad_norm": 0.1196879968047142, "learning_rate": 3.5102201176522966e-06, "loss": 0.6516, "step": 286 }, { "epoch": 1.4180024660912454, "grad_norm": 0.10076049715280533, "learning_rate": 3.4960027246156043e-06, "loss": 0.6363, "step": 287 }, { "epoch": 1.4229346485819976, "grad_norm": 0.11161559075117111, "learning_rate": 3.4817469380666834e-06, "loss": 0.6501, "step": 288 }, { "epoch": 1.4278668310727496, "grad_norm": 0.10083605349063873, "learning_rate": 3.467453307531858e-06, "loss": 0.6715, "step": 289 }, { "epoch": 1.4327990135635018, "grad_norm": 0.16506797075271606, "learning_rate": 3.453122383996245e-06, "loss": 0.6404, "step": 290 }, { "epoch": 1.437731196054254, "grad_norm": 0.12411224842071533, "learning_rate": 3.4387547198825187e-06, "loss": 0.668, "step": 291 }, { "epoch": 1.442663378545006, "grad_norm": 0.1273510903120041, "learning_rate": 3.4243508690296133e-06, "loss": 0.6571, "step": 292 }, { "epoch": 1.4475955610357583, "grad_norm": 0.10576938837766647, "learning_rate": 3.409911386671375e-06, "loss": 0.656, "step": 293 }, { "epoch": 1.4525277435265105, "grad_norm": 0.10712239891290665, "learning_rate": 3.3954368294151603e-06, "loss": 0.6474, "step": 294 }, { "epoch": 1.4574599260172627, "grad_norm": 0.1035386323928833, "learning_rate": 3.380927755220376e-06, "loss": 0.6512, "step": 295 }, { "epoch": 1.462392108508015, "grad_norm": 0.1390061378479004, "learning_rate": 3.366384723376977e-06, "loss": 0.6538, "step": 296 }, { "epoch": 1.467324290998767, "grad_norm": 0.12152580916881561, "learning_rate": 3.351808294483902e-06, "loss": 0.6414, "step": 297 }, { "epoch": 1.4722564734895192, "grad_norm": 0.11257496476173401, "learning_rate": 3.3371990304274654e-06, "loss": 0.6793, "step": 298 }, { "epoch": 1.4771886559802714, "grad_norm": 0.10283617675304413, "learning_rate": 3.3225574943597005e-06, "loss": 0.6444, "step": 299 }, { "epoch": 1.4821208384710234, "grad_norm": 0.12117356061935425, "learning_rate": 3.3078842506766484e-06, "loss": 0.631, "step": 300 }, { "epoch": 1.4870530209617756, "grad_norm": 0.14655297994613647, "learning_rate": 3.2931798649966e-06, "loss": 0.6614, "step": 301 }, { "epoch": 1.4919852034525278, "grad_norm": 0.1442922204732895, "learning_rate": 3.2784449041382973e-06, "loss": 0.6602, "step": 302 }, { "epoch": 1.4969173859432798, "grad_norm": 0.11084005236625671, "learning_rate": 3.263679936099083e-06, "loss": 0.6616, "step": 303 }, { "epoch": 1.501849568434032, "grad_norm": 0.1568291336297989, "learning_rate": 3.248885530033004e-06, "loss": 0.6745, "step": 304 }, { "epoch": 1.5067817509247843, "grad_norm": 0.11827261000871658, "learning_rate": 3.2340622562288717e-06, "loss": 0.666, "step": 305 }, { "epoch": 1.5117139334155363, "grad_norm": 0.13850663602352142, "learning_rate": 3.2192106860882782e-06, "loss": 0.6425, "step": 306 }, { "epoch": 1.5166461159062885, "grad_norm": 0.11242467910051346, "learning_rate": 3.2043313921035747e-06, "loss": 0.6366, "step": 307 }, { "epoch": 1.5215782983970407, "grad_norm": 0.11277095973491669, "learning_rate": 3.189424947835797e-06, "loss": 0.6423, "step": 308 }, { "epoch": 1.5265104808877927, "grad_norm": 0.1134696900844574, "learning_rate": 3.174491927892561e-06, "loss": 0.6637, "step": 309 }, { "epoch": 1.531442663378545, "grad_norm": 0.10201112180948257, "learning_rate": 3.1595329079059102e-06, "loss": 0.6458, "step": 310 }, { "epoch": 1.5363748458692972, "grad_norm": 0.22865180671215057, "learning_rate": 3.144548464510127e-06, "loss": 0.6693, "step": 311 }, { "epoch": 1.5413070283600492, "grad_norm": 0.12015289813280106, "learning_rate": 3.129539175319505e-06, "loss": 0.6463, "step": 312 }, { "epoch": 1.5462392108508016, "grad_norm": 0.14742949604988098, "learning_rate": 3.114505618906086e-06, "loss": 0.6398, "step": 313 }, { "epoch": 1.5511713933415536, "grad_norm": 0.12067458778619766, "learning_rate": 3.0994483747773508e-06, "loss": 0.6618, "step": 314 }, { "epoch": 1.5561035758323056, "grad_norm": 0.5674145817756653, "learning_rate": 3.0843680233538885e-06, "loss": 0.6552, "step": 315 }, { "epoch": 1.561035758323058, "grad_norm": 0.1083153635263443, "learning_rate": 3.0692651459470164e-06, "loss": 0.6283, "step": 316 }, { "epoch": 1.56596794081381, "grad_norm": 0.177827849984169, "learning_rate": 3.0541403247363756e-06, "loss": 0.6423, "step": 317 }, { "epoch": 1.5709001233045623, "grad_norm": 0.10957730561494827, "learning_rate": 3.0389941427474873e-06, "loss": 0.6569, "step": 318 }, { "epoch": 1.5758323057953145, "grad_norm": 0.14152103662490845, "learning_rate": 3.0238271838292815e-06, "loss": 0.6607, "step": 319 }, { "epoch": 1.5807644882860665, "grad_norm": 0.2304203361272812, "learning_rate": 3.0086400326315853e-06, "loss": 0.6798, "step": 320 }, { "epoch": 1.5856966707768188, "grad_norm": 0.11011958122253418, "learning_rate": 2.9934332745825924e-06, "loss": 0.6308, "step": 321 }, { "epoch": 1.590628853267571, "grad_norm": 0.12152457237243652, "learning_rate": 2.9782074958662923e-06, "loss": 0.6423, "step": 322 }, { "epoch": 1.595561035758323, "grad_norm": 0.10855505615472794, "learning_rate": 2.962963283399877e-06, "loss": 0.6759, "step": 323 }, { "epoch": 1.6004932182490752, "grad_norm": 0.12242516875267029, "learning_rate": 2.947701224811113e-06, "loss": 0.6489, "step": 324 }, { "epoch": 1.6054254007398274, "grad_norm": 0.11491915583610535, "learning_rate": 2.932421908415694e-06, "loss": 0.6476, "step": 325 }, { "epoch": 1.6103575832305794, "grad_norm": 0.10774843394756317, "learning_rate": 2.9171259231945598e-06, "loss": 0.6535, "step": 326 }, { "epoch": 1.6152897657213316, "grad_norm": 0.10615874826908112, "learning_rate": 2.901813858771193e-06, "loss": 0.6477, "step": 327 }, { "epoch": 1.6202219482120839, "grad_norm": 0.09981298446655273, "learning_rate": 2.8864863053888927e-06, "loss": 0.6525, "step": 328 }, { "epoch": 1.6251541307028359, "grad_norm": 0.1059933453798294, "learning_rate": 2.871143853888017e-06, "loss": 0.6609, "step": 329 }, { "epoch": 1.6300863131935883, "grad_norm": 0.28401389718055725, "learning_rate": 2.8557870956832135e-06, "loss": 0.6738, "step": 330 }, { "epoch": 1.6350184956843403, "grad_norm": 0.1126745268702507, "learning_rate": 2.840416622740617e-06, "loss": 0.6443, "step": 331 }, { "epoch": 1.6399506781750923, "grad_norm": 0.10980167984962463, "learning_rate": 2.8250330275550337e-06, "loss": 0.6575, "step": 332 }, { "epoch": 1.6448828606658448, "grad_norm": 0.20389322936534882, "learning_rate": 2.8096369031271e-06, "loss": 0.6855, "step": 333 }, { "epoch": 1.6498150431565968, "grad_norm": 0.1709917038679123, "learning_rate": 2.7942288429404256e-06, "loss": 0.6752, "step": 334 }, { "epoch": 1.654747225647349, "grad_norm": 0.12198632955551147, "learning_rate": 2.778809440938714e-06, "loss": 0.6398, "step": 335 }, { "epoch": 1.6596794081381012, "grad_norm": 0.11568919569253922, "learning_rate": 2.763379291502868e-06, "loss": 0.6489, "step": 336 }, { "epoch": 1.6646115906288532, "grad_norm": 0.15780635178089142, "learning_rate": 2.7479389894280793e-06, "loss": 0.6489, "step": 337 }, { "epoch": 1.6695437731196054, "grad_norm": 0.1101425290107727, "learning_rate": 2.7324891299008987e-06, "loss": 0.6726, "step": 338 }, { "epoch": 1.6744759556103577, "grad_norm": 0.10121767222881317, "learning_rate": 2.7170303084762958e-06, "loss": 0.6726, "step": 339 }, { "epoch": 1.6794081381011097, "grad_norm": 0.1087949350476265, "learning_rate": 2.701563121054695e-06, "loss": 0.6175, "step": 340 }, { "epoch": 1.684340320591862, "grad_norm": 0.11207877844572067, "learning_rate": 2.6860881638590146e-06, "loss": 0.6565, "step": 341 }, { "epoch": 1.6892725030826141, "grad_norm": 0.11743203550577164, "learning_rate": 2.670606033411678e-06, "loss": 0.677, "step": 342 }, { "epoch": 1.6942046855733661, "grad_norm": 0.10707394033670425, "learning_rate": 2.6551173265116182e-06, "loss": 0.6601, "step": 343 }, { "epoch": 1.6991368680641183, "grad_norm": 0.10522118955850601, "learning_rate": 2.6396226402112768e-06, "loss": 0.6494, "step": 344 }, { "epoch": 1.7040690505548706, "grad_norm": 0.19015160202980042, "learning_rate": 2.624122571793586e-06, "loss": 0.6494, "step": 345 }, { "epoch": 1.7090012330456226, "grad_norm": 0.11446730047464371, "learning_rate": 2.6086177187489453e-06, "loss": 0.6183, "step": 346 }, { "epoch": 1.7139334155363748, "grad_norm": 0.17297939956188202, "learning_rate": 2.593108678752191e-06, "loss": 0.6572, "step": 347 }, { "epoch": 1.718865598027127, "grad_norm": 0.12407731264829636, "learning_rate": 2.5775960496395565e-06, "loss": 0.6496, "step": 348 }, { "epoch": 1.723797780517879, "grad_norm": 0.36749541759490967, "learning_rate": 2.562080429385626e-06, "loss": 0.6611, "step": 349 }, { "epoch": 1.7287299630086315, "grad_norm": 0.10345156490802765, "learning_rate": 2.5465624160802847e-06, "loss": 0.6716, "step": 350 }, { "epoch": 1.7336621454993835, "grad_norm": 0.11017563939094543, "learning_rate": 2.531042607905665e-06, "loss": 0.6499, "step": 351 }, { "epoch": 1.7385943279901355, "grad_norm": 0.09504040330648422, "learning_rate": 2.5155216031130882e-06, "loss": 0.6657, "step": 352 }, { "epoch": 1.743526510480888, "grad_norm": 0.11456336081027985, "learning_rate": 2.5e-06, "loss": 0.6433, "step": 353 }, { "epoch": 1.74845869297164, "grad_norm": 0.3017407953739166, "learning_rate": 2.4844783968869126e-06, "loss": 0.6678, "step": 354 }, { "epoch": 1.7533908754623921, "grad_norm": 0.17654098570346832, "learning_rate": 2.4689573920943358e-06, "loss": 0.6407, "step": 355 }, { "epoch": 1.7583230579531444, "grad_norm": 0.19192291796207428, "learning_rate": 2.4534375839197166e-06, "loss": 0.6538, "step": 356 }, { "epoch": 1.7632552404438964, "grad_norm": 0.101350799202919, "learning_rate": 2.4379195706143755e-06, "loss": 0.6563, "step": 357 }, { "epoch": 1.7681874229346486, "grad_norm": 0.09221342951059341, "learning_rate": 2.422403950360444e-06, "loss": 0.6465, "step": 358 }, { "epoch": 1.7731196054254008, "grad_norm": 0.1029122844338417, "learning_rate": 2.406891321247809e-06, "loss": 0.6524, "step": 359 }, { "epoch": 1.7780517879161528, "grad_norm": 0.10780887305736542, "learning_rate": 2.391382281251055e-06, "loss": 0.6577, "step": 360 }, { "epoch": 1.782983970406905, "grad_norm": 0.11179229617118835, "learning_rate": 2.375877428206415e-06, "loss": 0.6281, "step": 361 }, { "epoch": 1.7879161528976573, "grad_norm": 0.1713687628507614, "learning_rate": 2.360377359788724e-06, "loss": 0.6866, "step": 362 }, { "epoch": 1.7928483353884093, "grad_norm": 0.20977120101451874, "learning_rate": 2.3448826734883826e-06, "loss": 0.6524, "step": 363 }, { "epoch": 1.7977805178791615, "grad_norm": 0.10442214459180832, "learning_rate": 2.3293939665883233e-06, "loss": 0.6356, "step": 364 }, { "epoch": 1.8027127003699137, "grad_norm": 0.11700151115655899, "learning_rate": 2.313911836140986e-06, "loss": 0.635, "step": 365 }, { "epoch": 1.8076448828606657, "grad_norm": 0.20740574598312378, "learning_rate": 2.298436878945306e-06, "loss": 0.6407, "step": 366 }, { "epoch": 1.8125770653514182, "grad_norm": 0.10487078875303268, "learning_rate": 2.2829696915237055e-06, "loss": 0.6606, "step": 367 }, { "epoch": 1.8175092478421702, "grad_norm": 0.19342948496341705, "learning_rate": 2.267510870099101e-06, "loss": 0.6652, "step": 368 }, { "epoch": 1.8224414303329222, "grad_norm": 0.12832309305667877, "learning_rate": 2.252061010571921e-06, "loss": 0.6699, "step": 369 }, { "epoch": 1.8273736128236746, "grad_norm": 0.24777992069721222, "learning_rate": 2.2366207084971326e-06, "loss": 0.652, "step": 370 }, { "epoch": 1.8323057953144266, "grad_norm": 0.2240799218416214, "learning_rate": 2.2211905590612864e-06, "loss": 0.6559, "step": 371 }, { "epoch": 1.8372379778051788, "grad_norm": 0.21929948031902313, "learning_rate": 2.205771157059575e-06, "loss": 0.6524, "step": 372 }, { "epoch": 1.842170160295931, "grad_norm": 0.09952554106712341, "learning_rate": 2.1903630968729003e-06, "loss": 0.663, "step": 373 }, { "epoch": 1.847102342786683, "grad_norm": 0.13348402082920074, "learning_rate": 2.174966972444967e-06, "loss": 0.6594, "step": 374 }, { "epoch": 1.8520345252774353, "grad_norm": 0.13318030536174774, "learning_rate": 2.159583377259384e-06, "loss": 0.6271, "step": 375 }, { "epoch": 1.8569667077681875, "grad_norm": 0.10292787104845047, "learning_rate": 2.1442129043167877e-06, "loss": 0.6537, "step": 376 }, { "epoch": 1.8618988902589395, "grad_norm": 0.1853366643190384, "learning_rate": 2.1288561461119837e-06, "loss": 0.6434, "step": 377 }, { "epoch": 1.8668310727496917, "grad_norm": 0.3069738745689392, "learning_rate": 2.1135136946111077e-06, "loss": 0.6294, "step": 378 }, { "epoch": 1.871763255240444, "grad_norm": 0.1591643989086151, "learning_rate": 2.098186141228807e-06, "loss": 0.6596, "step": 379 }, { "epoch": 1.876695437731196, "grad_norm": 0.1068023145198822, "learning_rate": 2.0828740768054406e-06, "loss": 0.6559, "step": 380 }, { "epoch": 1.8816276202219482, "grad_norm": 0.13300898671150208, "learning_rate": 2.0675780915843068e-06, "loss": 0.63, "step": 381 }, { "epoch": 1.8865598027127004, "grad_norm": 0.13550280034542084, "learning_rate": 2.052298775188888e-06, "loss": 0.646, "step": 382 }, { "epoch": 1.8914919852034524, "grad_norm": 0.19939783215522766, "learning_rate": 2.0370367166001245e-06, "loss": 0.6511, "step": 383 }, { "epoch": 1.8964241676942046, "grad_norm": 0.13335835933685303, "learning_rate": 2.021792504133709e-06, "loss": 0.6584, "step": 384 }, { "epoch": 1.9013563501849569, "grad_norm": 0.0967630073428154, "learning_rate": 2.006566725417409e-06, "loss": 0.656, "step": 385 }, { "epoch": 1.9062885326757089, "grad_norm": 0.10085921734571457, "learning_rate": 1.991359967368416e-06, "loss": 0.6515, "step": 386 }, { "epoch": 1.9112207151664613, "grad_norm": 0.1494772732257843, "learning_rate": 1.97617281617072e-06, "loss": 0.6625, "step": 387 }, { "epoch": 1.9161528976572133, "grad_norm": 0.14001788198947906, "learning_rate": 1.9610058572525127e-06, "loss": 0.6265, "step": 388 }, { "epoch": 1.9210850801479655, "grad_norm": 0.10699108988046646, "learning_rate": 1.945859675263625e-06, "loss": 0.6446, "step": 389 }, { "epoch": 1.9260172626387178, "grad_norm": 0.10522957146167755, "learning_rate": 1.9307348540529845e-06, "loss": 0.6525, "step": 390 }, { "epoch": 1.9309494451294698, "grad_norm": 0.0986817479133606, "learning_rate": 1.9156319766461124e-06, "loss": 0.6325, "step": 391 }, { "epoch": 1.935881627620222, "grad_norm": 0.11651718616485596, "learning_rate": 1.90055162522265e-06, "loss": 0.6303, "step": 392 }, { "epoch": 1.9408138101109742, "grad_norm": 0.10605504363775253, "learning_rate": 1.8854943810939152e-06, "loss": 0.6559, "step": 393 }, { "epoch": 1.9457459926017262, "grad_norm": 0.09837733954191208, "learning_rate": 1.8704608246804956e-06, "loss": 0.6708, "step": 394 }, { "epoch": 1.9506781750924784, "grad_norm": 0.08927737921476364, "learning_rate": 1.8554515354898744e-06, "loss": 0.654, "step": 395 }, { "epoch": 1.9556103575832307, "grad_norm": 0.10312498360872269, "learning_rate": 1.840467092094091e-06, "loss": 0.6178, "step": 396 }, { "epoch": 1.9605425400739827, "grad_norm": 0.12431120872497559, "learning_rate": 1.8255080721074391e-06, "loss": 0.6804, "step": 397 }, { "epoch": 1.9654747225647349, "grad_norm": 0.17175759375095367, "learning_rate": 1.8105750521642035e-06, "loss": 0.6509, "step": 398 }, { "epoch": 1.970406905055487, "grad_norm": 0.6188605427742004, "learning_rate": 1.7956686078964257e-06, "loss": 0.6311, "step": 399 }, { "epoch": 1.975339087546239, "grad_norm": 0.15260924398899078, "learning_rate": 1.7807893139117222e-06, "loss": 0.6562, "step": 400 }, { "epoch": 1.9802712700369913, "grad_norm": 0.10048322379589081, "learning_rate": 1.7659377437711294e-06, "loss": 0.6562, "step": 401 }, { "epoch": 1.9852034525277436, "grad_norm": 0.14348655939102173, "learning_rate": 1.7511144699669967e-06, "loss": 0.6415, "step": 402 }, { "epoch": 1.9901356350184956, "grad_norm": 0.2633483111858368, "learning_rate": 1.7363200639009176e-06, "loss": 0.6289, "step": 403 }, { "epoch": 1.995067817509248, "grad_norm": 0.10529658943414688, "learning_rate": 1.7215550958617036e-06, "loss": 0.6494, "step": 404 }, { "epoch": 2.0024660912453762, "grad_norm": 0.12062520533800125, "learning_rate": 1.7068201350034017e-06, "loss": 0.6449, "step": 405 }, { "epoch": 2.0073982737361282, "grad_norm": 0.1379857212305069, "learning_rate": 1.692115749323353e-06, "loss": 0.6557, "step": 406 }, { "epoch": 2.0123304562268802, "grad_norm": 0.10167136788368225, "learning_rate": 1.6774425056402993e-06, "loss": 0.6362, "step": 407 }, { "epoch": 2.0172626387176327, "grad_norm": 0.1222464069724083, "learning_rate": 1.6628009695725348e-06, "loss": 0.6479, "step": 408 }, { "epoch": 2.0221948212083847, "grad_norm": 0.19944354891777039, "learning_rate": 1.648191705516099e-06, "loss": 0.6606, "step": 409 }, { "epoch": 2.0271270036991367, "grad_norm": 0.09848273545503616, "learning_rate": 1.6336152766230235e-06, "loss": 0.6408, "step": 410 }, { "epoch": 2.032059186189889, "grad_norm": 0.09147990494966507, "learning_rate": 1.6190722447796242e-06, "loss": 0.6169, "step": 411 }, { "epoch": 2.036991368680641, "grad_norm": 0.09940177202224731, "learning_rate": 1.6045631705848405e-06, "loss": 0.6397, "step": 412 }, { "epoch": 2.041923551171393, "grad_norm": 0.10459251701831818, "learning_rate": 1.5900886133286254e-06, "loss": 0.6209, "step": 413 }, { "epoch": 2.0468557336621456, "grad_norm": 0.10069490969181061, "learning_rate": 1.5756491309703875e-06, "loss": 0.6516, "step": 414 }, { "epoch": 2.0517879161528976, "grad_norm": 0.10702253878116608, "learning_rate": 1.561245280117482e-06, "loss": 0.6491, "step": 415 }, { "epoch": 2.05672009864365, "grad_norm": 0.10752403736114502, "learning_rate": 1.5468776160037558e-06, "loss": 0.6554, "step": 416 }, { "epoch": 2.061652281134402, "grad_norm": 0.1126432865858078, "learning_rate": 1.5325466924681425e-06, "loss": 0.642, "step": 417 }, { "epoch": 2.066584463625154, "grad_norm": 0.13967493176460266, "learning_rate": 1.5182530619333168e-06, "loss": 0.6271, "step": 418 }, { "epoch": 2.0715166461159065, "grad_norm": 0.107243612408638, "learning_rate": 1.5039972753843966e-06, "loss": 0.6296, "step": 419 }, { "epoch": 2.0764488286066585, "grad_norm": 0.12694260478019714, "learning_rate": 1.4897798823477045e-06, "loss": 0.6584, "step": 420 }, { "epoch": 2.0813810110974105, "grad_norm": 0.11211492121219635, "learning_rate": 1.4756014308695865e-06, "loss": 0.6394, "step": 421 }, { "epoch": 2.086313193588163, "grad_norm": 0.10561738908290863, "learning_rate": 1.4614624674952843e-06, "loss": 0.6333, "step": 422 }, { "epoch": 2.091245376078915, "grad_norm": 0.11454541981220245, "learning_rate": 1.4473635372478692e-06, "loss": 0.6185, "step": 423 }, { "epoch": 2.096177558569667, "grad_norm": 0.11180110275745392, "learning_rate": 1.4333051836072298e-06, "loss": 0.6266, "step": 424 }, { "epoch": 2.1011097410604194, "grad_norm": 0.10139421373605728, "learning_rate": 1.4192879484891253e-06, "loss": 0.6519, "step": 425 }, { "epoch": 2.1060419235511714, "grad_norm": 0.10878422111272812, "learning_rate": 1.405312372224294e-06, "loss": 0.6253, "step": 426 }, { "epoch": 2.1109741060419234, "grad_norm": 0.09218919277191162, "learning_rate": 1.3913789935376271e-06, "loss": 0.6373, "step": 427 }, { "epoch": 2.115906288532676, "grad_norm": 0.10704641044139862, "learning_rate": 1.3774883495273986e-06, "loss": 0.6498, "step": 428 }, { "epoch": 2.120838471023428, "grad_norm": 0.09850191324949265, "learning_rate": 1.363640975644564e-06, "loss": 0.6595, "step": 429 }, { "epoch": 2.12577065351418, "grad_norm": 0.13874687254428864, "learning_rate": 1.3498374056721198e-06, "loss": 0.6555, "step": 430 }, { "epoch": 2.1307028360049323, "grad_norm": 0.11779513210058212, "learning_rate": 1.3360781717045266e-06, "loss": 0.6217, "step": 431 }, { "epoch": 2.1356350184956843, "grad_norm": 0.22133827209472656, "learning_rate": 1.322363804127198e-06, "loss": 0.6624, "step": 432 }, { "epoch": 2.1405672009864363, "grad_norm": 0.1059156209230423, "learning_rate": 1.3086948315960567e-06, "loss": 0.6412, "step": 433 }, { "epoch": 2.1454993834771887, "grad_norm": 0.1276949644088745, "learning_rate": 1.295071781017156e-06, "loss": 0.6325, "step": 434 }, { "epoch": 2.1504315659679407, "grad_norm": 0.09928814321756363, "learning_rate": 1.2814951775263671e-06, "loss": 0.6165, "step": 435 }, { "epoch": 2.155363748458693, "grad_norm": 0.1885470151901245, "learning_rate": 1.267965544469137e-06, "loss": 0.6294, "step": 436 }, { "epoch": 2.160295930949445, "grad_norm": 0.11126084625720978, "learning_rate": 1.2544834033803183e-06, "loss": 0.6392, "step": 437 }, { "epoch": 2.165228113440197, "grad_norm": 0.09588748216629028, "learning_rate": 1.2410492739640592e-06, "loss": 0.6242, "step": 438 }, { "epoch": 2.1701602959309496, "grad_norm": 0.3128255009651184, "learning_rate": 1.227663674073775e-06, "loss": 0.6191, "step": 439 }, { "epoch": 2.1750924784217016, "grad_norm": 0.11006899923086166, "learning_rate": 1.2143271196921832e-06, "loss": 0.6418, "step": 440 }, { "epoch": 2.1800246609124536, "grad_norm": 0.12083282321691513, "learning_rate": 1.2010401249114166e-06, "loss": 0.6192, "step": 441 }, { "epoch": 2.184956843403206, "grad_norm": 0.13546785712242126, "learning_rate": 1.1878032019132016e-06, "loss": 0.6688, "step": 442 }, { "epoch": 2.189889025893958, "grad_norm": 0.1263391375541687, "learning_rate": 1.1746168609491198e-06, "loss": 0.639, "step": 443 }, { "epoch": 2.19482120838471, "grad_norm": 0.357994943857193, "learning_rate": 1.1614816103209363e-06, "loss": 0.6582, "step": 444 }, { "epoch": 2.1997533908754625, "grad_norm": 0.09652648866176605, "learning_rate": 1.148397956361007e-06, "loss": 0.6323, "step": 445 }, { "epoch": 2.2046855733662145, "grad_norm": 0.12993961572647095, "learning_rate": 1.1353664034127585e-06, "loss": 0.6461, "step": 446 }, { "epoch": 2.2096177558569665, "grad_norm": 0.10628235340118408, "learning_rate": 1.122387453811252e-06, "loss": 0.6315, "step": 447 }, { "epoch": 2.214549938347719, "grad_norm": 2.75348162651062, "learning_rate": 1.1094616078638123e-06, "loss": 0.6616, "step": 448 }, { "epoch": 2.219482120838471, "grad_norm": 0.13206490874290466, "learning_rate": 1.0965893638307484e-06, "loss": 0.6227, "step": 449 }, { "epoch": 2.2244143033292234, "grad_norm": 0.10024918615818024, "learning_rate": 1.083771217906143e-06, "loss": 0.6171, "step": 450 }, { "epoch": 2.2293464858199754, "grad_norm": 0.1251702457666397, "learning_rate": 1.071007664198727e-06, "loss": 0.6592, "step": 451 }, { "epoch": 2.2342786683107274, "grad_norm": 0.12185829877853394, "learning_rate": 1.0582991947128324e-06, "loss": 0.6235, "step": 452 }, { "epoch": 2.23921085080148, "grad_norm": 0.09852628409862518, "learning_rate": 1.0456462993294273e-06, "loss": 0.6207, "step": 453 }, { "epoch": 2.244143033292232, "grad_norm": 0.11288397014141083, "learning_rate": 1.0330494657872312e-06, "loss": 0.6497, "step": 454 }, { "epoch": 2.249075215782984, "grad_norm": 0.1829065978527069, "learning_rate": 1.0205091796639143e-06, "loss": 0.6385, "step": 455 }, { "epoch": 2.2540073982737363, "grad_norm": 0.13576146960258484, "learning_rate": 1.008025924357379e-06, "loss": 0.6214, "step": 456 }, { "epoch": 2.2589395807644883, "grad_norm": 0.1026289239525795, "learning_rate": 9.95600181067129e-07, "loss": 0.6422, "step": 457 }, { "epoch": 2.2638717632552403, "grad_norm": 0.11270825564861298, "learning_rate": 9.832324287757158e-07, "loss": 0.6441, "step": 458 }, { "epoch": 2.2688039457459928, "grad_norm": 0.15193237364292145, "learning_rate": 9.709231442302777e-07, "loss": 0.6211, "step": 459 }, { "epoch": 2.2737361282367448, "grad_norm": 0.09436095505952835, "learning_rate": 9.586728019241622e-07, "loss": 0.6341, "step": 460 }, { "epoch": 2.2786683107274968, "grad_norm": 0.503972053527832, "learning_rate": 9.464818740786357e-07, "loss": 0.6092, "step": 461 }, { "epoch": 2.283600493218249, "grad_norm": 0.1313806027173996, "learning_rate": 9.343508306246771e-07, "loss": 0.6338, "step": 462 }, { "epoch": 2.288532675709001, "grad_norm": 0.21144555509090424, "learning_rate": 9.222801391848688e-07, "loss": 0.6227, "step": 463 }, { "epoch": 2.293464858199753, "grad_norm": 0.09302227199077606, "learning_rate": 9.102702650553672e-07, "loss": 0.6538, "step": 464 }, { "epoch": 2.2983970406905057, "grad_norm": 0.15463946759700775, "learning_rate": 8.983216711879663e-07, "loss": 0.6429, "step": 465 }, { "epoch": 2.3033292231812577, "grad_norm": 0.09232427924871445, "learning_rate": 8.86434818172256e-07, "loss": 0.6414, "step": 466 }, { "epoch": 2.3082614056720097, "grad_norm": 0.4079054296016693, "learning_rate": 8.746101642178623e-07, "loss": 0.6505, "step": 467 }, { "epoch": 2.313193588162762, "grad_norm": 0.11529503017663956, "learning_rate": 8.628481651367876e-07, "loss": 0.6379, "step": 468 }, { "epoch": 2.318125770653514, "grad_norm": 0.2939020097255707, "learning_rate": 8.51149274325839e-07, "loss": 0.6404, "step": 469 }, { "epoch": 2.323057953144266, "grad_norm": 0.09394059330224991, "learning_rate": 8.395139427491517e-07, "loss": 0.6456, "step": 470 }, { "epoch": 2.3279901356350186, "grad_norm": 0.12832114100456238, "learning_rate": 8.279426189208057e-07, "loss": 0.6372, "step": 471 }, { "epoch": 2.3329223181257706, "grad_norm": 0.1322140395641327, "learning_rate": 8.164357488875349e-07, "loss": 0.6199, "step": 472 }, { "epoch": 2.337854500616523, "grad_norm": 0.1532570719718933, "learning_rate": 8.049937762115354e-07, "loss": 0.6197, "step": 473 }, { "epoch": 2.342786683107275, "grad_norm": 0.11840520799160004, "learning_rate": 7.936171419533653e-07, "loss": 0.6225, "step": 474 }, { "epoch": 2.347718865598027, "grad_norm": 0.11926340311765671, "learning_rate": 7.823062846549432e-07, "loss": 0.6369, "step": 475 }, { "epoch": 2.3526510480887795, "grad_norm": 0.10458600521087646, "learning_rate": 7.71061640322646e-07, "loss": 0.6192, "step": 476 }, { "epoch": 2.3575832305795315, "grad_norm": 0.09384766221046448, "learning_rate": 7.59883642410498e-07, "loss": 0.6319, "step": 477 }, { "epoch": 2.3625154130702835, "grad_norm": 0.13794849812984467, "learning_rate": 7.487727218034646e-07, "loss": 0.6627, "step": 478 }, { "epoch": 2.367447595561036, "grad_norm": 0.09522448480129242, "learning_rate": 7.377293068008421e-07, "loss": 0.612, "step": 479 }, { "epoch": 2.372379778051788, "grad_norm": 0.34490853548049927, "learning_rate": 7.267538230997487e-07, "loss": 0.6447, "step": 480 }, { "epoch": 2.37731196054254, "grad_norm": 0.3206603229045868, "learning_rate": 7.15846693778712e-07, "loss": 0.6633, "step": 481 }, { "epoch": 2.3822441430332923, "grad_norm": 0.10969394445419312, "learning_rate": 7.050083392813651e-07, "loss": 0.63, "step": 482 }, { "epoch": 2.3871763255240444, "grad_norm": 0.09678292274475098, "learning_rate": 6.942391774002352e-07, "loss": 0.6102, "step": 483 }, { "epoch": 2.392108508014797, "grad_norm": 0.12844984233379364, "learning_rate": 6.835396232606414e-07, "loss": 0.6065, "step": 484 }, { "epoch": 2.397040690505549, "grad_norm": 0.10880632698535919, "learning_rate": 6.729100893046897e-07, "loss": 0.6388, "step": 485 }, { "epoch": 2.401972872996301, "grad_norm": 0.15737678110599518, "learning_rate": 6.623509852753798e-07, "loss": 0.6288, "step": 486 }, { "epoch": 2.4069050554870532, "grad_norm": 0.26250696182250977, "learning_rate": 6.518627182008034e-07, "loss": 0.6355, "step": 487 }, { "epoch": 2.4118372379778052, "grad_norm": 0.1346411406993866, "learning_rate": 6.414456923784593e-07, "loss": 0.6637, "step": 488 }, { "epoch": 2.4167694204685573, "grad_norm": 0.10590661317110062, "learning_rate": 6.311003093596674e-07, "loss": 0.6191, "step": 489 }, { "epoch": 2.4217016029593097, "grad_norm": 0.10193908959627151, "learning_rate": 6.208269679340886e-07, "loss": 0.6224, "step": 490 }, { "epoch": 2.4266337854500617, "grad_norm": 0.12438759952783585, "learning_rate": 6.106260641143547e-07, "loss": 0.6291, "step": 491 }, { "epoch": 2.4315659679408137, "grad_norm": 0.11157315969467163, "learning_rate": 6.004979911208006e-07, "loss": 0.6265, "step": 492 }, { "epoch": 2.436498150431566, "grad_norm": 0.10205821692943573, "learning_rate": 5.904431393663088e-07, "loss": 0.6219, "step": 493 }, { "epoch": 2.441430332922318, "grad_norm": 0.0957137867808342, "learning_rate": 5.804618964412587e-07, "loss": 0.6376, "step": 494 }, { "epoch": 2.44636251541307, "grad_norm": 0.14211417734622955, "learning_rate": 5.705546470985851e-07, "loss": 0.6255, "step": 495 }, { "epoch": 2.4512946979038226, "grad_norm": 0.12571153044700623, "learning_rate": 5.607217732389503e-07, "loss": 0.6222, "step": 496 }, { "epoch": 2.4562268803945746, "grad_norm": 0.18337282538414001, "learning_rate": 5.509636538960183e-07, "loss": 0.661, "step": 497 }, { "epoch": 2.4611590628853266, "grad_norm": 0.10181237757205963, "learning_rate": 5.412806652218469e-07, "loss": 0.6097, "step": 498 }, { "epoch": 2.466091245376079, "grad_norm": 0.10895208269357681, "learning_rate": 5.316731804723877e-07, "loss": 0.6544, "step": 499 }, { "epoch": 2.471023427866831, "grad_norm": 0.11107916384935379, "learning_rate": 5.221415699930952e-07, "loss": 0.6326, "step": 500 }, { "epoch": 2.475955610357583, "grad_norm": 0.09523724764585495, "learning_rate": 5.126862012046551e-07, "loss": 0.6295, "step": 501 }, { "epoch": 2.4808877928483355, "grad_norm": 0.17275214195251465, "learning_rate": 5.03307438588819e-07, "loss": 0.647, "step": 502 }, { "epoch": 2.4858199753390875, "grad_norm": 0.3728453516960144, "learning_rate": 4.940056436743534e-07, "loss": 0.6419, "step": 503 }, { "epoch": 2.4907521578298395, "grad_norm": 0.09702899307012558, "learning_rate": 4.847811750231057e-07, "loss": 0.6374, "step": 504 }, { "epoch": 2.495684340320592, "grad_norm": 0.12215188145637512, "learning_rate": 4.7563438821618236e-07, "loss": 0.6408, "step": 505 }, { "epoch": 2.500616522811344, "grad_norm": 0.10780132561922073, "learning_rate": 4.6656563584023955e-07, "loss": 0.6191, "step": 506 }, { "epoch": 2.505548705302096, "grad_norm": 0.14112994074821472, "learning_rate": 4.5757526747389506e-07, "loss": 0.6381, "step": 507 }, { "epoch": 2.5104808877928484, "grad_norm": 0.09904135763645172, "learning_rate": 4.4866362967425054e-07, "loss": 0.6312, "step": 508 }, { "epoch": 2.5154130702836004, "grad_norm": 0.24112163484096527, "learning_rate": 4.398310659635338e-07, "loss": 0.6271, "step": 509 }, { "epoch": 2.5203452527743524, "grad_norm": 0.14225369691848755, "learning_rate": 4.310779168158566e-07, "loss": 0.6423, "step": 510 }, { "epoch": 2.525277435265105, "grad_norm": 0.13167624175548553, "learning_rate": 4.2240451964408984e-07, "loss": 0.6338, "step": 511 }, { "epoch": 2.530209617755857, "grad_norm": 0.10630662739276886, "learning_rate": 4.138112087868576e-07, "loss": 0.6255, "step": 512 }, { "epoch": 2.5351418002466093, "grad_norm": 0.10379086434841156, "learning_rate": 4.052983154956483e-07, "loss": 0.6264, "step": 513 }, { "epoch": 2.5400739827373613, "grad_norm": 0.09079194813966751, "learning_rate": 3.9686616792204677e-07, "loss": 0.5999, "step": 514 }, { "epoch": 2.5450061652281133, "grad_norm": 0.09570661187171936, "learning_rate": 3.885150911050856e-07, "loss": 0.61, "step": 515 }, { "epoch": 2.5499383477188657, "grad_norm": 0.09997903555631638, "learning_rate": 3.8024540695871275e-07, "loss": 0.6391, "step": 516 }, { "epoch": 2.5548705302096177, "grad_norm": 0.10698223114013672, "learning_rate": 3.720574342593847e-07, "loss": 0.6197, "step": 517 }, { "epoch": 2.55980271270037, "grad_norm": 0.13335859775543213, "learning_rate": 3.639514886337786e-07, "loss": 0.6507, "step": 518 }, { "epoch": 2.564734895191122, "grad_norm": 0.2598697245121002, "learning_rate": 3.559278825466245e-07, "loss": 0.6284, "step": 519 }, { "epoch": 2.569667077681874, "grad_norm": 0.19334684312343597, "learning_rate": 3.4798692528866057e-07, "loss": 0.631, "step": 520 }, { "epoch": 2.5745992601726266, "grad_norm": 0.1008177101612091, "learning_rate": 3.4012892296471173e-07, "loss": 0.6651, "step": 521 }, { "epoch": 2.5795314426633786, "grad_norm": 0.11722905933856964, "learning_rate": 3.3235417848188985e-07, "loss": 0.6524, "step": 522 }, { "epoch": 2.5844636251541306, "grad_norm": 0.12759149074554443, "learning_rate": 3.2466299153791626e-07, "loss": 0.6508, "step": 523 }, { "epoch": 2.589395807644883, "grad_norm": 0.12765510380268097, "learning_rate": 3.1705565860956994e-07, "loss": 0.636, "step": 524 }, { "epoch": 2.594327990135635, "grad_norm": 0.10414116829633713, "learning_rate": 3.095324729412602e-07, "loss": 0.6192, "step": 525 }, { "epoch": 2.599260172626387, "grad_norm": 0.28323763608932495, "learning_rate": 3.020937245337208e-07, "loss": 0.6614, "step": 526 }, { "epoch": 2.6041923551171395, "grad_norm": 0.19360347092151642, "learning_rate": 2.947397001328314e-07, "loss": 0.6444, "step": 527 }, { "epoch": 2.6091245376078915, "grad_norm": 0.10508602857589722, "learning_rate": 2.874706832185656e-07, "loss": 0.6113, "step": 528 }, { "epoch": 2.6140567200986435, "grad_norm": 0.10985539853572845, "learning_rate": 2.80286953994062e-07, "loss": 0.6372, "step": 529 }, { "epoch": 2.618988902589396, "grad_norm": 0.11580062657594681, "learning_rate": 2.731887893748242e-07, "loss": 0.6274, "step": 530 }, { "epoch": 2.623921085080148, "grad_norm": 0.10763997584581375, "learning_rate": 2.6617646297804554e-07, "loss": 0.6345, "step": 531 }, { "epoch": 2.6288532675709, "grad_norm": 0.10293387621641159, "learning_rate": 2.5925024511206207e-07, "loss": 0.6032, "step": 532 }, { "epoch": 2.6337854500616524, "grad_norm": 0.1057887151837349, "learning_rate": 2.52410402765933e-07, "loss": 0.6594, "step": 533 }, { "epoch": 2.6387176325524044, "grad_norm": 0.11511359363794327, "learning_rate": 2.45657199599148e-07, "loss": 0.6283, "step": 534 }, { "epoch": 2.6436498150431564, "grad_norm": 0.10298167169094086, "learning_rate": 2.389908959314663e-07, "loss": 0.6194, "step": 535 }, { "epoch": 2.648581997533909, "grad_norm": 0.13076511025428772, "learning_rate": 2.3241174873287892e-07, "loss": 0.6128, "step": 536 }, { "epoch": 2.653514180024661, "grad_norm": 0.10795601457357407, "learning_rate": 2.2592001161370392e-07, "loss": 0.6604, "step": 537 }, { "epoch": 2.658446362515413, "grad_norm": 0.08846652507781982, "learning_rate": 2.1951593481481236e-07, "loss": 0.5985, "step": 538 }, { "epoch": 2.6633785450061653, "grad_norm": 0.14082390069961548, "learning_rate": 2.1319976519797862e-07, "loss": 0.6444, "step": 539 }, { "epoch": 2.6683107274969173, "grad_norm": 0.1039619892835617, "learning_rate": 2.0697174623636795e-07, "loss": 0.6735, "step": 540 }, { "epoch": 2.6732429099876693, "grad_norm": 0.09636316448450089, "learning_rate": 2.0083211800514868e-07, "loss": 0.6311, "step": 541 }, { "epoch": 2.678175092478422, "grad_norm": 0.10067697614431381, "learning_rate": 1.9478111717223968e-07, "loss": 0.6237, "step": 542 }, { "epoch": 2.683107274969174, "grad_norm": 0.11034560203552246, "learning_rate": 1.8881897698918544e-07, "loss": 0.6247, "step": 543 }, { "epoch": 2.688039457459926, "grad_norm": 0.11042781919240952, "learning_rate": 1.8294592728216764e-07, "loss": 0.6413, "step": 544 }, { "epoch": 2.6929716399506782, "grad_norm": 0.1032506600022316, "learning_rate": 1.7716219444314204e-07, "loss": 0.6604, "step": 545 }, { "epoch": 2.6979038224414302, "grad_norm": 0.09196452796459198, "learning_rate": 1.7146800142111536e-07, "loss": 0.6503, "step": 546 }, { "epoch": 2.7028360049321822, "grad_norm": 0.09848834574222565, "learning_rate": 1.658635677135484e-07, "loss": 0.6195, "step": 547 }, { "epoch": 2.7077681874229347, "grad_norm": 0.12145579606294632, "learning_rate": 1.6034910935789628e-07, "loss": 0.6269, "step": 548 }, { "epoch": 2.7127003699136867, "grad_norm": 0.10812917351722717, "learning_rate": 1.5492483892328104e-07, "loss": 0.6422, "step": 549 }, { "epoch": 2.717632552404439, "grad_norm": 0.10012295842170715, "learning_rate": 1.4959096550229645e-07, "loss": 0.6266, "step": 550 }, { "epoch": 2.722564734895191, "grad_norm": 0.24410749971866608, "learning_rate": 1.44347694702949e-07, "loss": 0.6406, "step": 551 }, { "epoch": 2.727496917385943, "grad_norm": 0.11757177114486694, "learning_rate": 1.391952286407311e-07, "loss": 0.6446, "step": 552 }, { "epoch": 2.7324290998766956, "grad_norm": 0.11057958751916885, "learning_rate": 1.341337659308309e-07, "loss": 0.6347, "step": 553 }, { "epoch": 2.7373612823674476, "grad_norm": 0.11715658754110336, "learning_rate": 1.291635016804768e-07, "loss": 0.6508, "step": 554 }, { "epoch": 2.7422934648582, "grad_norm": 0.10512126982212067, "learning_rate": 1.2428462748141523e-07, "loss": 0.6509, "step": 555 }, { "epoch": 2.747225647348952, "grad_norm": 0.20344923436641693, "learning_rate": 1.1949733140252468e-07, "loss": 0.619, "step": 556 }, { "epoch": 2.752157829839704, "grad_norm": 0.13411924242973328, "learning_rate": 1.1480179798256857e-07, "loss": 0.6254, "step": 557 }, { "epoch": 2.7570900123304565, "grad_norm": 0.09810943156480789, "learning_rate": 1.1019820822307986e-07, "loss": 0.6301, "step": 558 }, { "epoch": 2.7620221948212085, "grad_norm": 0.14971470832824707, "learning_rate": 1.056867395813832e-07, "loss": 0.6325, "step": 559 }, { "epoch": 2.7669543773119605, "grad_norm": 0.11750344932079315, "learning_rate": 1.0126756596375687e-07, "loss": 0.6256, "step": 560 }, { "epoch": 2.771886559802713, "grad_norm": 0.125830739736557, "learning_rate": 9.694085771872697e-08, "loss": 0.6014, "step": 561 }, { "epoch": 2.776818742293465, "grad_norm": 0.16469189524650574, "learning_rate": 9.270678163050218e-08, "loss": 0.6491, "step": 562 }, { "epoch": 2.781750924784217, "grad_norm": 0.12204600125551224, "learning_rate": 8.856550091254302e-08, "loss": 0.6472, "step": 563 }, { "epoch": 2.7866831072749694, "grad_norm": 0.11357175558805466, "learning_rate": 8.451717520127272e-08, "loss": 0.6447, "step": 564 }, { "epoch": 2.7916152897657214, "grad_norm": 0.2082013189792633, "learning_rate": 8.056196054992193e-08, "loss": 0.6686, "step": 565 }, { "epoch": 2.7965474722564734, "grad_norm": 0.09912417829036713, "learning_rate": 7.670000942251288e-08, "loss": 0.6365, "step": 566 }, { "epoch": 2.801479654747226, "grad_norm": 0.3408164381980896, "learning_rate": 7.293147068798384e-08, "loss": 0.7199, "step": 567 }, { "epoch": 2.806411837237978, "grad_norm": 0.18314948678016663, "learning_rate": 6.92564896144493e-08, "loss": 0.6531, "step": 568 }, { "epoch": 2.81134401972873, "grad_norm": 0.0905625969171524, "learning_rate": 6.56752078636011e-08, "loss": 0.6495, "step": 569 }, { "epoch": 2.8162762022194823, "grad_norm": 0.10962024331092834, "learning_rate": 6.218776348524663e-08, "loss": 0.629, "step": 570 }, { "epoch": 2.8212083847102343, "grad_norm": 0.1477546989917755, "learning_rate": 5.879429091198846e-08, "loss": 0.6415, "step": 571 }, { "epoch": 2.8261405672009863, "grad_norm": 0.2878320515155792, "learning_rate": 5.549492095404202e-08, "loss": 0.6354, "step": 572 }, { "epoch": 2.8310727496917387, "grad_norm": 0.09865111857652664, "learning_rate": 5.2289780794192726e-08, "loss": 0.6333, "step": 573 }, { "epoch": 2.8360049321824907, "grad_norm": 0.11314819753170013, "learning_rate": 4.917899398289378e-08, "loss": 0.6353, "step": 574 }, { "epoch": 2.8409371146732427, "grad_norm": 0.13539192080497742, "learning_rate": 4.6162680433503024e-08, "loss": 0.6449, "step": 575 }, { "epoch": 2.845869297163995, "grad_norm": 0.11400008946657181, "learning_rate": 4.3240956417661685e-08, "loss": 0.5971, "step": 576 }, { "epoch": 2.850801479654747, "grad_norm": 0.1335747390985489, "learning_rate": 4.0413934560811216e-08, "loss": 0.6234, "step": 577 }, { "epoch": 2.855733662145499, "grad_norm": 1.7575569152832031, "learning_rate": 3.768172383785268e-08, "loss": 0.6393, "step": 578 }, { "epoch": 2.8606658446362516, "grad_norm": 0.20131815969944, "learning_rate": 3.504442956894533e-08, "loss": 0.6389, "step": 579 }, { "epoch": 2.8655980271270036, "grad_norm": 0.10285453498363495, "learning_rate": 3.250215341544766e-08, "loss": 0.6129, "step": 580 }, { "epoch": 2.8705302096177556, "grad_norm": 0.11187023669481277, "learning_rate": 3.005499337599777e-08, "loss": 0.622, "step": 581 }, { "epoch": 2.875462392108508, "grad_norm": 0.10102323442697525, "learning_rate": 2.7703043782735527e-08, "loss": 0.6667, "step": 582 }, { "epoch": 2.88039457459926, "grad_norm": 0.16657452285289764, "learning_rate": 2.544639529766829e-08, "loss": 0.6302, "step": 583 }, { "epoch": 2.885326757090012, "grad_norm": 0.09614276140928268, "learning_rate": 2.3285134909173113e-08, "loss": 0.614, "step": 584 }, { "epoch": 2.8902589395807645, "grad_norm": 0.17049086093902588, "learning_rate": 2.1219345928646107e-08, "loss": 0.6179, "step": 585 }, { "epoch": 2.8951911220715165, "grad_norm": 0.11277743428945541, "learning_rate": 1.924910798728946e-08, "loss": 0.6397, "step": 586 }, { "epoch": 2.900123304562269, "grad_norm": 0.1780087947845459, "learning_rate": 1.7374497033042504e-08, "loss": 0.6358, "step": 587 }, { "epoch": 2.905055487053021, "grad_norm": 0.33896899223327637, "learning_rate": 1.559558532765404e-08, "loss": 0.6364, "step": 588 }, { "epoch": 2.909987669543773, "grad_norm": 0.11533108353614807, "learning_rate": 1.3912441443896529e-08, "loss": 0.6219, "step": 589 }, { "epoch": 2.9149198520345254, "grad_norm": 0.2513286769390106, "learning_rate": 1.2325130262923202e-08, "loss": 0.6544, "step": 590 }, { "epoch": 2.9198520345252774, "grad_norm": 0.10646089911460876, "learning_rate": 1.0833712971766442e-08, "loss": 0.6314, "step": 591 }, { "epoch": 2.92478421701603, "grad_norm": 0.18473948538303375, "learning_rate": 9.438247060979954e-09, "loss": 0.6272, "step": 592 }, { "epoch": 2.929716399506782, "grad_norm": 0.17810285091400146, "learning_rate": 8.13878632242221e-09, "loss": 0.6551, "step": 593 }, { "epoch": 2.934648581997534, "grad_norm": 0.10676706582307816, "learning_rate": 6.935380847182815e-09, "loss": 0.6505, "step": 594 }, { "epoch": 2.9395807644882863, "grad_norm": 0.1609431952238083, "learning_rate": 5.828077023651846e-09, "loss": 0.6403, "step": 595 }, { "epoch": 2.9445129469790383, "grad_norm": 0.3682437539100647, "learning_rate": 4.816917535731547e-09, "loss": 0.6535, "step": 596 }, { "epoch": 2.9494451294697903, "grad_norm": 0.08937019854784012, "learning_rate": 3.9019413611907084e-09, "loss": 0.6422, "step": 597 }, { "epoch": 2.9543773119605428, "grad_norm": 0.09343031048774719, "learning_rate": 3.083183770162812e-09, "loss": 0.6454, "step": 598 }, { "epoch": 2.9593094944512948, "grad_norm": 0.11115273088216782, "learning_rate": 2.360676323786282e-09, "loss": 0.6167, "step": 599 }, { "epoch": 2.9642416769420468, "grad_norm": 0.09774809330701828, "learning_rate": 1.7344468729868502e-09, "loss": 0.6422, "step": 600 }, { "epoch": 2.969173859432799, "grad_norm": 0.10831815004348755, "learning_rate": 1.2045195574056367e-09, "loss": 0.6387, "step": 601 }, { "epoch": 2.974106041923551, "grad_norm": 0.22296251356601715, "learning_rate": 7.709148044679482e-10, "loss": 0.6415, "step": 602 }, { "epoch": 2.979038224414303, "grad_norm": 0.10190446674823761, "learning_rate": 4.3364932859474293e-10, "loss": 0.6286, "step": 603 }, { "epoch": 2.9839704069050557, "grad_norm": 0.15854156017303467, "learning_rate": 1.9273613056008945e-10, "loss": 0.6051, "step": 604 }, { "epoch": 2.9889025893958077, "grad_norm": 0.09298679232597351, "learning_rate": 4.8184496989067684e-11, "loss": 0.6213, "step": 605 }, { "epoch": 2.9938347718865597, "grad_norm": 0.2940804958343506, "learning_rate": 0.0, "loss": 0.6347, "step": 606 } ], "logging_steps": 1, "max_steps": 606, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 101, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.765503957851418e+20, "train_batch_size": 1, "trial_name": null, "trial_params": null }