{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.3091190108191654, "eval_steps": 500, "global_step": 1000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0003091190108191654, "grad_norm": 0.17105351388454437, "learning_rate": 2e-05, "loss": 1.589, "step": 1 }, { "epoch": 0.0006182380216383308, "grad_norm": 0.13888764381408691, "learning_rate": 4e-05, "loss": 1.6333, "step": 2 }, { "epoch": 0.0009273570324574962, "grad_norm": 0.13389942049980164, "learning_rate": 6e-05, "loss": 1.6075, "step": 3 }, { "epoch": 0.0012364760432766616, "grad_norm": 0.1443634182214737, "learning_rate": 8e-05, "loss": 1.3981, "step": 4 }, { "epoch": 0.0015455950540958269, "grad_norm": 0.2410346418619156, "learning_rate": 0.0001, "loss": 1.6522, "step": 5 }, { "epoch": 0.0018547140649149924, "grad_norm": 0.23892079293727875, "learning_rate": 0.00012, "loss": 1.3345, "step": 6 }, { "epoch": 0.0021638330757341576, "grad_norm": 0.32107171416282654, "learning_rate": 0.00014, "loss": 1.4086, "step": 7 }, { "epoch": 0.002472952086553323, "grad_norm": 0.38699325919151306, "learning_rate": 0.00016, "loss": 1.2824, "step": 8 }, { "epoch": 0.0027820710973724882, "grad_norm": 0.2972716987133026, "learning_rate": 0.00018, "loss": 1.3528, "step": 9 }, { "epoch": 0.0030911901081916537, "grad_norm": 0.288402795791626, "learning_rate": 0.0002, "loss": 1.0556, "step": 10 }, { "epoch": 0.0034003091190108192, "grad_norm": 0.4650692641735077, "learning_rate": 0.00019993798449612405, "loss": 1.0995, "step": 11 }, { "epoch": 0.0037094281298299847, "grad_norm": 0.39375749230384827, "learning_rate": 0.0001998759689922481, "loss": 0.9978, "step": 12 }, { "epoch": 0.00401854714064915, "grad_norm": 0.3362458348274231, "learning_rate": 0.0001998139534883721, "loss": 1.0522, "step": 13 }, { "epoch": 0.004327666151468315, "grad_norm": 0.2642221450805664, "learning_rate": 0.00019975193798449614, "loss": 0.9661, "step": 14 }, { "epoch": 0.00463678516228748, "grad_norm": 0.3542484939098358, "learning_rate": 0.00019968992248062018, "loss": 0.8814, "step": 15 }, { "epoch": 0.004945904173106646, "grad_norm": 0.1401689201593399, "learning_rate": 0.00019962790697674421, "loss": 0.8629, "step": 16 }, { "epoch": 0.005255023183925811, "grad_norm": 0.20015761256217957, "learning_rate": 0.00019956589147286823, "loss": 0.8454, "step": 17 }, { "epoch": 0.0055641421947449764, "grad_norm": 0.1540534645318985, "learning_rate": 0.00019950387596899224, "loss": 0.8852, "step": 18 }, { "epoch": 0.005873261205564142, "grad_norm": 0.15904690325260162, "learning_rate": 0.00019944186046511628, "loss": 0.7254, "step": 19 }, { "epoch": 0.0061823802163833074, "grad_norm": 0.21907807886600494, "learning_rate": 0.00019937984496124032, "loss": 1.0708, "step": 20 }, { "epoch": 0.006491499227202473, "grad_norm": 0.14591765403747559, "learning_rate": 0.00019931782945736435, "loss": 0.8555, "step": 21 }, { "epoch": 0.0068006182380216385, "grad_norm": 0.16298744082450867, "learning_rate": 0.00019925581395348837, "loss": 0.8262, "step": 22 }, { "epoch": 0.0071097372488408035, "grad_norm": 0.14358466863632202, "learning_rate": 0.0001991937984496124, "loss": 0.8744, "step": 23 }, { "epoch": 0.0074188562596599695, "grad_norm": 0.149592325091362, "learning_rate": 0.00019913178294573644, "loss": 0.9376, "step": 24 }, { "epoch": 0.0077279752704791345, "grad_norm": 0.13760673999786377, "learning_rate": 0.00019906976744186048, "loss": 0.8854, "step": 25 }, { "epoch": 0.0080370942812983, "grad_norm": 0.19108699262142181, "learning_rate": 0.00019900775193798452, "loss": 0.885, "step": 26 }, { "epoch": 0.008346213292117466, "grad_norm": 0.1892910748720169, "learning_rate": 0.00019894573643410853, "loss": 0.9911, "step": 27 }, { "epoch": 0.00865533230293663, "grad_norm": 0.13371500372886658, "learning_rate": 0.00019888372093023257, "loss": 0.8475, "step": 28 }, { "epoch": 0.008964451313755796, "grad_norm": 0.1226775124669075, "learning_rate": 0.0001988217054263566, "loss": 0.7733, "step": 29 }, { "epoch": 0.00927357032457496, "grad_norm": 0.13002170622348785, "learning_rate": 0.00019875968992248062, "loss": 0.8008, "step": 30 }, { "epoch": 0.009582689335394128, "grad_norm": 0.13575126230716705, "learning_rate": 0.00019869767441860466, "loss": 0.8841, "step": 31 }, { "epoch": 0.009891808346213293, "grad_norm": 0.15838854014873505, "learning_rate": 0.00019863565891472867, "loss": 0.8856, "step": 32 }, { "epoch": 0.010200927357032458, "grad_norm": 0.12744970619678497, "learning_rate": 0.0001985736434108527, "loss": 0.8327, "step": 33 }, { "epoch": 0.010510046367851623, "grad_norm": 0.16277430951595306, "learning_rate": 0.00019851162790697675, "loss": 0.9559, "step": 34 }, { "epoch": 0.010819165378670788, "grad_norm": 0.11398226767778397, "learning_rate": 0.0001984496124031008, "loss": 0.8987, "step": 35 }, { "epoch": 0.011128284389489953, "grad_norm": 0.15259447693824768, "learning_rate": 0.00019838759689922483, "loss": 0.916, "step": 36 }, { "epoch": 0.01143740340030912, "grad_norm": 0.16493409872055054, "learning_rate": 0.00019832558139534884, "loss": 0.7795, "step": 37 }, { "epoch": 0.011746522411128285, "grad_norm": 0.14143070578575134, "learning_rate": 0.00019826356589147288, "loss": 0.814, "step": 38 }, { "epoch": 0.01205564142194745, "grad_norm": 0.11884719133377075, "learning_rate": 0.00019820155038759692, "loss": 0.7559, "step": 39 }, { "epoch": 0.012364760432766615, "grad_norm": 0.12909553945064545, "learning_rate": 0.00019813953488372096, "loss": 0.9106, "step": 40 }, { "epoch": 0.01267387944358578, "grad_norm": 0.11181219667196274, "learning_rate": 0.00019807751937984497, "loss": 0.7127, "step": 41 }, { "epoch": 0.012982998454404947, "grad_norm": 0.16188634932041168, "learning_rate": 0.000198015503875969, "loss": 0.856, "step": 42 }, { "epoch": 0.013292117465224112, "grad_norm": 0.1277618706226349, "learning_rate": 0.00019795348837209304, "loss": 0.8901, "step": 43 }, { "epoch": 0.013601236476043277, "grad_norm": 0.13743072748184204, "learning_rate": 0.00019789147286821706, "loss": 0.9266, "step": 44 }, { "epoch": 0.013910355486862442, "grad_norm": 0.12885789573192596, "learning_rate": 0.0001978294573643411, "loss": 0.8982, "step": 45 }, { "epoch": 0.014219474497681607, "grad_norm": 0.14296455681324005, "learning_rate": 0.0001977674418604651, "loss": 1.0045, "step": 46 }, { "epoch": 0.014528593508500772, "grad_norm": 0.13536542654037476, "learning_rate": 0.00019770542635658915, "loss": 0.8648, "step": 47 }, { "epoch": 0.014837712519319939, "grad_norm": 0.1285800188779831, "learning_rate": 0.00019764341085271318, "loss": 0.8113, "step": 48 }, { "epoch": 0.015146831530139104, "grad_norm": 0.1538587212562561, "learning_rate": 0.00019758139534883722, "loss": 0.8465, "step": 49 }, { "epoch": 0.015455950540958269, "grad_norm": 0.11420200765132904, "learning_rate": 0.00019751937984496126, "loss": 0.8647, "step": 50 }, { "epoch": 0.015765069551777436, "grad_norm": 0.13382850587368011, "learning_rate": 0.00019745736434108527, "loss": 0.9545, "step": 51 }, { "epoch": 0.0160741885625966, "grad_norm": 0.11594673991203308, "learning_rate": 0.0001973953488372093, "loss": 0.8599, "step": 52 }, { "epoch": 0.016383307573415766, "grad_norm": 0.119788758456707, "learning_rate": 0.00019733333333333335, "loss": 0.8394, "step": 53 }, { "epoch": 0.01669242658423493, "grad_norm": 0.1150812953710556, "learning_rate": 0.0001972713178294574, "loss": 0.8727, "step": 54 }, { "epoch": 0.017001545595054096, "grad_norm": 0.1359858363866806, "learning_rate": 0.0001972093023255814, "loss": 0.76, "step": 55 }, { "epoch": 0.01731066460587326, "grad_norm": 0.15345649421215057, "learning_rate": 0.00019714728682170544, "loss": 0.7292, "step": 56 }, { "epoch": 0.017619783616692426, "grad_norm": 0.14331281185150146, "learning_rate": 0.00019708527131782945, "loss": 0.8034, "step": 57 }, { "epoch": 0.01792890262751159, "grad_norm": 0.16820766031742096, "learning_rate": 0.0001970232558139535, "loss": 0.7534, "step": 58 }, { "epoch": 0.018238021638330756, "grad_norm": 0.16281287372112274, "learning_rate": 0.00019696124031007753, "loss": 0.8409, "step": 59 }, { "epoch": 0.01854714064914992, "grad_norm": 0.16938892006874084, "learning_rate": 0.00019689922480620157, "loss": 0.8786, "step": 60 }, { "epoch": 0.018856259659969087, "grad_norm": 0.13455645740032196, "learning_rate": 0.00019683720930232558, "loss": 0.7462, "step": 61 }, { "epoch": 0.019165378670788255, "grad_norm": 0.12973853945732117, "learning_rate": 0.00019677519379844962, "loss": 0.8304, "step": 62 }, { "epoch": 0.01947449768160742, "grad_norm": 0.158578023314476, "learning_rate": 0.00019671317829457366, "loss": 0.8244, "step": 63 }, { "epoch": 0.019783616692426585, "grad_norm": 0.125227153301239, "learning_rate": 0.0001966511627906977, "loss": 0.8637, "step": 64 }, { "epoch": 0.02009273570324575, "grad_norm": 0.1529238075017929, "learning_rate": 0.0001965891472868217, "loss": 0.8271, "step": 65 }, { "epoch": 0.020401854714064915, "grad_norm": 0.1335589438676834, "learning_rate": 0.00019652713178294575, "loss": 0.8129, "step": 66 }, { "epoch": 0.02071097372488408, "grad_norm": 0.13944782316684723, "learning_rate": 0.00019646511627906978, "loss": 0.887, "step": 67 }, { "epoch": 0.021020092735703245, "grad_norm": 0.12453600019216537, "learning_rate": 0.00019640310077519382, "loss": 0.8369, "step": 68 }, { "epoch": 0.02132921174652241, "grad_norm": 0.11443863809108734, "learning_rate": 0.00019634108527131786, "loss": 0.8769, "step": 69 }, { "epoch": 0.021638330757341576, "grad_norm": 0.1325102150440216, "learning_rate": 0.00019627906976744185, "loss": 0.7528, "step": 70 }, { "epoch": 0.02194744976816074, "grad_norm": 0.13488665223121643, "learning_rate": 0.00019621705426356589, "loss": 0.876, "step": 71 }, { "epoch": 0.022256568778979906, "grad_norm": 0.15174520015716553, "learning_rate": 0.00019615503875968992, "loss": 0.8219, "step": 72 }, { "epoch": 0.022565687789799074, "grad_norm": 0.13083337247371674, "learning_rate": 0.00019609302325581396, "loss": 0.8696, "step": 73 }, { "epoch": 0.02287480680061824, "grad_norm": 0.13707856833934784, "learning_rate": 0.000196031007751938, "loss": 0.9781, "step": 74 }, { "epoch": 0.023183925811437404, "grad_norm": 0.14287059009075165, "learning_rate": 0.000195968992248062, "loss": 0.8975, "step": 75 }, { "epoch": 0.02349304482225657, "grad_norm": 0.14259910583496094, "learning_rate": 0.00019590697674418605, "loss": 0.781, "step": 76 }, { "epoch": 0.023802163833075735, "grad_norm": 0.17812331020832062, "learning_rate": 0.0001958449612403101, "loss": 0.8664, "step": 77 }, { "epoch": 0.0241112828438949, "grad_norm": 0.10900291800498962, "learning_rate": 0.00019578294573643413, "loss": 0.8155, "step": 78 }, { "epoch": 0.024420401854714065, "grad_norm": 0.1299259066581726, "learning_rate": 0.00019572093023255814, "loss": 0.8878, "step": 79 }, { "epoch": 0.02472952086553323, "grad_norm": 0.1341174691915512, "learning_rate": 0.00019565891472868218, "loss": 0.855, "step": 80 }, { "epoch": 0.025038639876352395, "grad_norm": 0.11747386306524277, "learning_rate": 0.00019559689922480622, "loss": 0.8901, "step": 81 }, { "epoch": 0.02534775888717156, "grad_norm": 0.12569762766361237, "learning_rate": 0.00019553488372093026, "loss": 0.8644, "step": 82 }, { "epoch": 0.025656877897990725, "grad_norm": 0.11595705896615982, "learning_rate": 0.0001954728682170543, "loss": 0.7919, "step": 83 }, { "epoch": 0.025965996908809894, "grad_norm": 0.15013526380062103, "learning_rate": 0.0001954108527131783, "loss": 0.7987, "step": 84 }, { "epoch": 0.02627511591962906, "grad_norm": 0.13101589679718018, "learning_rate": 0.00019534883720930232, "loss": 0.7824, "step": 85 }, { "epoch": 0.026584234930448224, "grad_norm": 0.12921208143234253, "learning_rate": 0.00019528682170542636, "loss": 0.9077, "step": 86 }, { "epoch": 0.02689335394126739, "grad_norm": 0.18682463467121124, "learning_rate": 0.0001952248062015504, "loss": 0.8614, "step": 87 }, { "epoch": 0.027202472952086554, "grad_norm": 0.15416069328784943, "learning_rate": 0.00019516279069767444, "loss": 0.8038, "step": 88 }, { "epoch": 0.02751159196290572, "grad_norm": 0.13872137665748596, "learning_rate": 0.00019510077519379845, "loss": 0.9001, "step": 89 }, { "epoch": 0.027820710973724884, "grad_norm": 0.1256810575723648, "learning_rate": 0.00019503875968992249, "loss": 0.8332, "step": 90 }, { "epoch": 0.02812982998454405, "grad_norm": 0.15000316500663757, "learning_rate": 0.00019497674418604652, "loss": 0.8345, "step": 91 }, { "epoch": 0.028438948995363214, "grad_norm": 0.11536971479654312, "learning_rate": 0.00019491472868217056, "loss": 0.7702, "step": 92 }, { "epoch": 0.02874806800618238, "grad_norm": 0.11627457290887833, "learning_rate": 0.00019485271317829457, "loss": 0.8944, "step": 93 }, { "epoch": 0.029057187017001544, "grad_norm": 0.12913382053375244, "learning_rate": 0.00019479069767441861, "loss": 0.8054, "step": 94 }, { "epoch": 0.02936630602782071, "grad_norm": 0.14983727037906647, "learning_rate": 0.00019472868217054265, "loss": 0.7897, "step": 95 }, { "epoch": 0.029675425038639878, "grad_norm": 0.1396576315164566, "learning_rate": 0.0001946666666666667, "loss": 0.7595, "step": 96 }, { "epoch": 0.029984544049459043, "grad_norm": 0.1093367412686348, "learning_rate": 0.0001946046511627907, "loss": 0.878, "step": 97 }, { "epoch": 0.030293663060278208, "grad_norm": 0.14023703336715698, "learning_rate": 0.00019454263565891474, "loss": 0.8051, "step": 98 }, { "epoch": 0.030602782071097373, "grad_norm": 0.11650537699460983, "learning_rate": 0.00019448062015503875, "loss": 0.7733, "step": 99 }, { "epoch": 0.030911901081916538, "grad_norm": 0.13736988604068756, "learning_rate": 0.0001944186046511628, "loss": 0.8414, "step": 100 }, { "epoch": 0.031221020092735703, "grad_norm": 0.13810019195079803, "learning_rate": 0.00019435658914728683, "loss": 0.9192, "step": 101 }, { "epoch": 0.03153013910355487, "grad_norm": 0.16648177802562714, "learning_rate": 0.00019429457364341087, "loss": 0.8692, "step": 102 }, { "epoch": 0.03183925811437403, "grad_norm": 0.16531941294670105, "learning_rate": 0.00019423255813953488, "loss": 0.8585, "step": 103 }, { "epoch": 0.0321483771251932, "grad_norm": 0.12364251166582108, "learning_rate": 0.00019417054263565892, "loss": 0.7652, "step": 104 }, { "epoch": 0.03245749613601236, "grad_norm": 0.139155313372612, "learning_rate": 0.00019410852713178296, "loss": 0.7448, "step": 105 }, { "epoch": 0.03276661514683153, "grad_norm": 0.11827906966209412, "learning_rate": 0.000194046511627907, "loss": 0.9182, "step": 106 }, { "epoch": 0.033075734157650694, "grad_norm": 0.1247883066534996, "learning_rate": 0.00019398449612403104, "loss": 0.8138, "step": 107 }, { "epoch": 0.03338485316846986, "grad_norm": 0.12576410174369812, "learning_rate": 0.00019392248062015505, "loss": 0.803, "step": 108 }, { "epoch": 0.033693972179289024, "grad_norm": 0.12698566913604736, "learning_rate": 0.00019386046511627909, "loss": 0.8579, "step": 109 }, { "epoch": 0.03400309119010819, "grad_norm": 0.10796654969453812, "learning_rate": 0.0001937984496124031, "loss": 0.812, "step": 110 }, { "epoch": 0.034312210200927354, "grad_norm": 0.12361832708120346, "learning_rate": 0.00019373643410852714, "loss": 0.8606, "step": 111 }, { "epoch": 0.03462132921174652, "grad_norm": 0.12853065133094788, "learning_rate": 0.00019367441860465118, "loss": 0.9208, "step": 112 }, { "epoch": 0.03493044822256569, "grad_norm": 0.119226835668087, "learning_rate": 0.0001936124031007752, "loss": 0.7306, "step": 113 }, { "epoch": 0.03523956723338485, "grad_norm": 0.12476561963558197, "learning_rate": 0.00019355038759689923, "loss": 0.8046, "step": 114 }, { "epoch": 0.03554868624420402, "grad_norm": 0.11674510687589645, "learning_rate": 0.00019348837209302326, "loss": 0.8214, "step": 115 }, { "epoch": 0.03585780525502318, "grad_norm": 0.1358969360589981, "learning_rate": 0.0001934263565891473, "loss": 0.7432, "step": 116 }, { "epoch": 0.03616692426584235, "grad_norm": 0.1318214237689972, "learning_rate": 0.00019336434108527132, "loss": 0.9497, "step": 117 }, { "epoch": 0.03647604327666151, "grad_norm": 0.1427808552980423, "learning_rate": 0.00019330232558139535, "loss": 0.8012, "step": 118 }, { "epoch": 0.03678516228748068, "grad_norm": 0.14754672348499298, "learning_rate": 0.0001932403100775194, "loss": 0.8112, "step": 119 }, { "epoch": 0.03709428129829984, "grad_norm": 0.13510632514953613, "learning_rate": 0.00019317829457364343, "loss": 0.8607, "step": 120 }, { "epoch": 0.03740340030911901, "grad_norm": 0.11763066798448563, "learning_rate": 0.00019311627906976747, "loss": 0.8349, "step": 121 }, { "epoch": 0.03771251931993817, "grad_norm": 0.13032180070877075, "learning_rate": 0.00019305426356589148, "loss": 0.8847, "step": 122 }, { "epoch": 0.03802163833075734, "grad_norm": 0.11119523644447327, "learning_rate": 0.00019299224806201552, "loss": 0.8302, "step": 123 }, { "epoch": 0.03833075734157651, "grad_norm": 0.1144416555762291, "learning_rate": 0.00019293023255813953, "loss": 0.7984, "step": 124 }, { "epoch": 0.03863987635239567, "grad_norm": 0.12676909565925598, "learning_rate": 0.00019286821705426357, "loss": 0.8186, "step": 125 }, { "epoch": 0.03894899536321484, "grad_norm": 0.10949283838272095, "learning_rate": 0.0001928062015503876, "loss": 0.7339, "step": 126 }, { "epoch": 0.039258114374034, "grad_norm": 0.11983365565538406, "learning_rate": 0.00019274418604651162, "loss": 0.7616, "step": 127 }, { "epoch": 0.03956723338485317, "grad_norm": 0.1309802085161209, "learning_rate": 0.00019268217054263566, "loss": 0.7917, "step": 128 }, { "epoch": 0.03987635239567233, "grad_norm": 0.1349460780620575, "learning_rate": 0.0001926201550387597, "loss": 0.8753, "step": 129 }, { "epoch": 0.0401854714064915, "grad_norm": 0.12506724894046783, "learning_rate": 0.00019255813953488374, "loss": 0.7819, "step": 130 }, { "epoch": 0.04049459041731066, "grad_norm": 0.13243618607521057, "learning_rate": 0.00019249612403100778, "loss": 0.8049, "step": 131 }, { "epoch": 0.04080370942812983, "grad_norm": 0.14795252680778503, "learning_rate": 0.0001924341085271318, "loss": 0.7969, "step": 132 }, { "epoch": 0.04111282843894899, "grad_norm": 0.15747897326946259, "learning_rate": 0.00019237209302325583, "loss": 0.8746, "step": 133 }, { "epoch": 0.04142194744976816, "grad_norm": 0.15109744668006897, "learning_rate": 0.00019231007751937987, "loss": 0.8562, "step": 134 }, { "epoch": 0.04173106646058733, "grad_norm": 0.13535654544830322, "learning_rate": 0.0001922480620155039, "loss": 0.8928, "step": 135 }, { "epoch": 0.04204018547140649, "grad_norm": 0.1262591928243637, "learning_rate": 0.00019218604651162792, "loss": 0.8302, "step": 136 }, { "epoch": 0.04234930448222566, "grad_norm": 0.11443354189395905, "learning_rate": 0.00019212403100775193, "loss": 0.8672, "step": 137 }, { "epoch": 0.04265842349304482, "grad_norm": 0.11836638301610947, "learning_rate": 0.00019206201550387597, "loss": 0.7593, "step": 138 }, { "epoch": 0.04296754250386399, "grad_norm": 0.12662746012210846, "learning_rate": 0.000192, "loss": 0.7531, "step": 139 }, { "epoch": 0.04327666151468315, "grad_norm": 0.12387800961732864, "learning_rate": 0.00019193798449612404, "loss": 0.7422, "step": 140 }, { "epoch": 0.04358578052550232, "grad_norm": 0.12786395847797394, "learning_rate": 0.00019187596899224806, "loss": 0.744, "step": 141 }, { "epoch": 0.04389489953632148, "grad_norm": 0.12761859595775604, "learning_rate": 0.0001918139534883721, "loss": 0.8977, "step": 142 }, { "epoch": 0.04420401854714065, "grad_norm": 0.10713964700698853, "learning_rate": 0.00019175193798449613, "loss": 0.7677, "step": 143 }, { "epoch": 0.04451313755795981, "grad_norm": 0.13007132709026337, "learning_rate": 0.00019168992248062017, "loss": 0.7516, "step": 144 }, { "epoch": 0.04482225656877898, "grad_norm": 0.12673480808734894, "learning_rate": 0.0001916279069767442, "loss": 0.8469, "step": 145 }, { "epoch": 0.04513137557959815, "grad_norm": 0.14040741324424744, "learning_rate": 0.00019156589147286822, "loss": 0.8119, "step": 146 }, { "epoch": 0.04544049459041731, "grad_norm": 0.1404358148574829, "learning_rate": 0.00019150387596899226, "loss": 0.8062, "step": 147 }, { "epoch": 0.04574961360123648, "grad_norm": 0.140091672539711, "learning_rate": 0.0001914418604651163, "loss": 0.9796, "step": 148 }, { "epoch": 0.04605873261205564, "grad_norm": 0.12712246179580688, "learning_rate": 0.00019137984496124034, "loss": 0.7021, "step": 149 }, { "epoch": 0.04636785162287481, "grad_norm": 0.1542489379644394, "learning_rate": 0.00019131782945736435, "loss": 0.7141, "step": 150 }, { "epoch": 0.04667697063369397, "grad_norm": 0.1310671865940094, "learning_rate": 0.00019125581395348836, "loss": 0.7513, "step": 151 }, { "epoch": 0.04698608964451314, "grad_norm": 0.1205151230096817, "learning_rate": 0.0001911937984496124, "loss": 0.8243, "step": 152 }, { "epoch": 0.0472952086553323, "grad_norm": 0.13522934913635254, "learning_rate": 0.00019113178294573644, "loss": 0.9441, "step": 153 }, { "epoch": 0.04760432766615147, "grad_norm": 0.11995132267475128, "learning_rate": 0.00019106976744186048, "loss": 0.7032, "step": 154 }, { "epoch": 0.04791344667697063, "grad_norm": 0.11310404539108276, "learning_rate": 0.0001910077519379845, "loss": 0.6913, "step": 155 }, { "epoch": 0.0482225656877898, "grad_norm": 0.11462230980396271, "learning_rate": 0.00019094573643410853, "loss": 0.8314, "step": 156 }, { "epoch": 0.04853168469860897, "grad_norm": 0.12306851893663406, "learning_rate": 0.00019088372093023257, "loss": 0.9122, "step": 157 }, { "epoch": 0.04884080370942813, "grad_norm": 0.09559505432844162, "learning_rate": 0.0001908217054263566, "loss": 0.8177, "step": 158 }, { "epoch": 0.0491499227202473, "grad_norm": 0.11616392433643341, "learning_rate": 0.00019075968992248064, "loss": 0.9223, "step": 159 }, { "epoch": 0.04945904173106646, "grad_norm": 0.12350696325302124, "learning_rate": 0.00019069767441860466, "loss": 0.9004, "step": 160 }, { "epoch": 0.04976816074188563, "grad_norm": 0.12747159600257874, "learning_rate": 0.0001906356589147287, "loss": 0.9126, "step": 161 }, { "epoch": 0.05007727975270479, "grad_norm": 0.12991321086883545, "learning_rate": 0.00019057364341085273, "loss": 0.9113, "step": 162 }, { "epoch": 0.05038639876352396, "grad_norm": 0.11218614876270294, "learning_rate": 0.00019051162790697677, "loss": 0.7117, "step": 163 }, { "epoch": 0.05069551777434312, "grad_norm": 0.13032029569149017, "learning_rate": 0.00019044961240310078, "loss": 0.841, "step": 164 }, { "epoch": 0.05100463678516229, "grad_norm": 0.1347358375787735, "learning_rate": 0.0001903875968992248, "loss": 0.8103, "step": 165 }, { "epoch": 0.05131375579598145, "grad_norm": 0.11914915591478348, "learning_rate": 0.00019032558139534883, "loss": 0.8913, "step": 166 }, { "epoch": 0.05162287480680062, "grad_norm": 0.15790300071239471, "learning_rate": 0.00019026356589147287, "loss": 0.801, "step": 167 }, { "epoch": 0.05193199381761979, "grad_norm": 0.15204893052577972, "learning_rate": 0.0001902015503875969, "loss": 0.8085, "step": 168 }, { "epoch": 0.05224111282843895, "grad_norm": 0.11781688779592514, "learning_rate": 0.00019013953488372095, "loss": 0.9159, "step": 169 }, { "epoch": 0.05255023183925812, "grad_norm": 0.12103480845689774, "learning_rate": 0.00019007751937984496, "loss": 0.8221, "step": 170 }, { "epoch": 0.05285935085007728, "grad_norm": 0.12477768957614899, "learning_rate": 0.000190015503875969, "loss": 0.7622, "step": 171 }, { "epoch": 0.05316846986089645, "grad_norm": 0.11186879873275757, "learning_rate": 0.00018995348837209304, "loss": 0.7959, "step": 172 }, { "epoch": 0.05347758887171561, "grad_norm": 0.12586474418640137, "learning_rate": 0.00018989147286821708, "loss": 0.6735, "step": 173 }, { "epoch": 0.05378670788253478, "grad_norm": 0.11750543862581253, "learning_rate": 0.0001898294573643411, "loss": 0.8302, "step": 174 }, { "epoch": 0.05409582689335394, "grad_norm": 0.1301109939813614, "learning_rate": 0.00018976744186046513, "loss": 0.9917, "step": 175 }, { "epoch": 0.05440494590417311, "grad_norm": 0.1335320919752121, "learning_rate": 0.00018970542635658917, "loss": 0.8256, "step": 176 }, { "epoch": 0.05471406491499227, "grad_norm": 0.12373456358909607, "learning_rate": 0.00018964341085271318, "loss": 0.8458, "step": 177 }, { "epoch": 0.05502318392581144, "grad_norm": 0.1291348785161972, "learning_rate": 0.00018958139534883722, "loss": 0.7675, "step": 178 }, { "epoch": 0.0553323029366306, "grad_norm": 0.12421860545873642, "learning_rate": 0.00018951937984496123, "loss": 0.8195, "step": 179 }, { "epoch": 0.05564142194744977, "grad_norm": 0.1433798372745514, "learning_rate": 0.00018945736434108527, "loss": 0.7938, "step": 180 }, { "epoch": 0.055950540958268936, "grad_norm": 0.15060195326805115, "learning_rate": 0.0001893953488372093, "loss": 0.7178, "step": 181 }, { "epoch": 0.0562596599690881, "grad_norm": 0.13103605806827545, "learning_rate": 0.00018933333333333335, "loss": 0.8432, "step": 182 }, { "epoch": 0.05656877897990727, "grad_norm": 0.1537558138370514, "learning_rate": 0.00018927131782945738, "loss": 0.81, "step": 183 }, { "epoch": 0.05687789799072643, "grad_norm": 0.12011228501796722, "learning_rate": 0.0001892093023255814, "loss": 0.8906, "step": 184 }, { "epoch": 0.0571870170015456, "grad_norm": 0.11913521587848663, "learning_rate": 0.00018914728682170543, "loss": 0.7819, "step": 185 }, { "epoch": 0.05749613601236476, "grad_norm": 0.13771173357963562, "learning_rate": 0.00018908527131782947, "loss": 0.7775, "step": 186 }, { "epoch": 0.05780525502318393, "grad_norm": 0.11831659823656082, "learning_rate": 0.0001890232558139535, "loss": 0.7461, "step": 187 }, { "epoch": 0.05811437403400309, "grad_norm": 0.11049254238605499, "learning_rate": 0.00018896124031007752, "loss": 0.7784, "step": 188 }, { "epoch": 0.05842349304482226, "grad_norm": 0.10577117651700974, "learning_rate": 0.00018889922480620156, "loss": 0.7211, "step": 189 }, { "epoch": 0.05873261205564142, "grad_norm": 0.13082446157932281, "learning_rate": 0.00018883720930232557, "loss": 0.6854, "step": 190 }, { "epoch": 0.05904173106646059, "grad_norm": 0.11105687916278839, "learning_rate": 0.0001887751937984496, "loss": 0.8418, "step": 191 }, { "epoch": 0.059350850077279756, "grad_norm": 0.1412641704082489, "learning_rate": 0.00018871317829457365, "loss": 0.8293, "step": 192 }, { "epoch": 0.05965996908809892, "grad_norm": 0.12148229032754898, "learning_rate": 0.0001886511627906977, "loss": 0.856, "step": 193 }, { "epoch": 0.059969088098918086, "grad_norm": 0.12246838212013245, "learning_rate": 0.0001885891472868217, "loss": 0.9042, "step": 194 }, { "epoch": 0.06027820710973725, "grad_norm": 0.12810048460960388, "learning_rate": 0.00018852713178294574, "loss": 0.7636, "step": 195 }, { "epoch": 0.060587326120556416, "grad_norm": 0.12897038459777832, "learning_rate": 0.00018846511627906978, "loss": 0.8182, "step": 196 }, { "epoch": 0.06089644513137558, "grad_norm": 0.13533450663089752, "learning_rate": 0.00018840310077519382, "loss": 0.7578, "step": 197 }, { "epoch": 0.061205564142194746, "grad_norm": 0.13347265124320984, "learning_rate": 0.00018834108527131783, "loss": 0.8414, "step": 198 }, { "epoch": 0.06151468315301391, "grad_norm": 0.14512066543102264, "learning_rate": 0.00018827906976744187, "loss": 0.8698, "step": 199 }, { "epoch": 0.061823802163833076, "grad_norm": 0.11501649022102356, "learning_rate": 0.0001882170542635659, "loss": 0.8591, "step": 200 }, { "epoch": 0.06213292117465224, "grad_norm": 0.11760124564170837, "learning_rate": 0.00018815503875968995, "loss": 0.6859, "step": 201 }, { "epoch": 0.062442040185471406, "grad_norm": 0.13676373660564423, "learning_rate": 0.00018809302325581399, "loss": 0.7216, "step": 202 }, { "epoch": 0.06275115919629057, "grad_norm": 0.11492311954498291, "learning_rate": 0.000188031007751938, "loss": 0.8223, "step": 203 }, { "epoch": 0.06306027820710974, "grad_norm": 0.11568205058574677, "learning_rate": 0.000187968992248062, "loss": 0.8193, "step": 204 }, { "epoch": 0.0633693972179289, "grad_norm": 0.12526321411132812, "learning_rate": 0.00018790697674418605, "loss": 0.8803, "step": 205 }, { "epoch": 0.06367851622874807, "grad_norm": 0.12961214780807495, "learning_rate": 0.00018784496124031009, "loss": 0.8083, "step": 206 }, { "epoch": 0.06398763523956723, "grad_norm": 0.11950293183326721, "learning_rate": 0.00018778294573643412, "loss": 0.8049, "step": 207 }, { "epoch": 0.0642967542503864, "grad_norm": 0.11256164312362671, "learning_rate": 0.00018772093023255814, "loss": 0.7365, "step": 208 }, { "epoch": 0.06460587326120557, "grad_norm": 0.13182170689105988, "learning_rate": 0.00018765891472868217, "loss": 0.7031, "step": 209 }, { "epoch": 0.06491499227202473, "grad_norm": 0.1193682923913002, "learning_rate": 0.00018759689922480621, "loss": 0.8785, "step": 210 }, { "epoch": 0.06522411128284389, "grad_norm": 0.13558265566825867, "learning_rate": 0.00018753488372093025, "loss": 0.8275, "step": 211 }, { "epoch": 0.06553323029366306, "grad_norm": 0.12028771638870239, "learning_rate": 0.00018747286821705426, "loss": 0.8436, "step": 212 }, { "epoch": 0.06584234930448223, "grad_norm": 0.12355880439281464, "learning_rate": 0.0001874108527131783, "loss": 0.687, "step": 213 }, { "epoch": 0.06615146831530139, "grad_norm": 0.11852920055389404, "learning_rate": 0.00018734883720930234, "loss": 0.8268, "step": 214 }, { "epoch": 0.06646058732612056, "grad_norm": 0.1262328028678894, "learning_rate": 0.00018728682170542638, "loss": 0.841, "step": 215 }, { "epoch": 0.06676970633693972, "grad_norm": 0.13128647208213806, "learning_rate": 0.00018722480620155042, "loss": 0.7472, "step": 216 }, { "epoch": 0.06707882534775889, "grad_norm": 0.12075427919626236, "learning_rate": 0.00018716279069767443, "loss": 0.8483, "step": 217 }, { "epoch": 0.06738794435857805, "grad_norm": 0.11870454251766205, "learning_rate": 0.00018710077519379844, "loss": 0.8179, "step": 218 }, { "epoch": 0.06769706336939722, "grad_norm": 0.12058960646390915, "learning_rate": 0.00018703875968992248, "loss": 0.8403, "step": 219 }, { "epoch": 0.06800618238021638, "grad_norm": 0.13978858292102814, "learning_rate": 0.00018697674418604652, "loss": 0.6639, "step": 220 }, { "epoch": 0.06831530139103555, "grad_norm": 0.10775326192378998, "learning_rate": 0.00018691472868217056, "loss": 0.8841, "step": 221 }, { "epoch": 0.06862442040185471, "grad_norm": 0.10687053948640823, "learning_rate": 0.00018685271317829457, "loss": 0.7715, "step": 222 }, { "epoch": 0.06893353941267388, "grad_norm": 0.12222916632890701, "learning_rate": 0.0001867906976744186, "loss": 0.7728, "step": 223 }, { "epoch": 0.06924265842349304, "grad_norm": 0.13918592035770416, "learning_rate": 0.00018672868217054265, "loss": 0.8792, "step": 224 }, { "epoch": 0.0695517774343122, "grad_norm": 0.11157078295946121, "learning_rate": 0.0001866666666666667, "loss": 0.8906, "step": 225 }, { "epoch": 0.06986089644513138, "grad_norm": 0.12403914332389832, "learning_rate": 0.0001866046511627907, "loss": 0.9234, "step": 226 }, { "epoch": 0.07017001545595054, "grad_norm": 0.11490818858146667, "learning_rate": 0.00018654263565891474, "loss": 0.8447, "step": 227 }, { "epoch": 0.0704791344667697, "grad_norm": 0.13033214211463928, "learning_rate": 0.00018648062015503878, "loss": 0.7801, "step": 228 }, { "epoch": 0.07078825347758887, "grad_norm": 0.1061464175581932, "learning_rate": 0.00018641860465116281, "loss": 0.6313, "step": 229 }, { "epoch": 0.07109737248840804, "grad_norm": 0.12007651478052139, "learning_rate": 0.00018635658914728683, "loss": 0.7526, "step": 230 }, { "epoch": 0.0714064914992272, "grad_norm": 0.12450309842824936, "learning_rate": 0.00018629457364341086, "loss": 0.8403, "step": 231 }, { "epoch": 0.07171561051004637, "grad_norm": 0.16374681890010834, "learning_rate": 0.00018623255813953488, "loss": 0.7729, "step": 232 }, { "epoch": 0.07202472952086553, "grad_norm": 0.13087786734104156, "learning_rate": 0.00018617054263565892, "loss": 0.8882, "step": 233 }, { "epoch": 0.0723338485316847, "grad_norm": 0.13743267953395844, "learning_rate": 0.00018610852713178295, "loss": 0.8316, "step": 234 }, { "epoch": 0.07264296754250386, "grad_norm": 0.1110304743051529, "learning_rate": 0.000186046511627907, "loss": 0.773, "step": 235 }, { "epoch": 0.07295208655332303, "grad_norm": 0.12651820480823517, "learning_rate": 0.000185984496124031, "loss": 0.9538, "step": 236 }, { "epoch": 0.0732612055641422, "grad_norm": 0.11768705397844315, "learning_rate": 0.00018592248062015504, "loss": 0.7585, "step": 237 }, { "epoch": 0.07357032457496136, "grad_norm": 0.1165948212146759, "learning_rate": 0.00018586046511627908, "loss": 0.868, "step": 238 }, { "epoch": 0.07387944358578052, "grad_norm": 0.12231750786304474, "learning_rate": 0.00018579844961240312, "loss": 0.8445, "step": 239 }, { "epoch": 0.07418856259659969, "grad_norm": 0.13796208798885345, "learning_rate": 0.00018573643410852716, "loss": 0.8391, "step": 240 }, { "epoch": 0.07449768160741886, "grad_norm": 0.1166827604174614, "learning_rate": 0.00018567441860465117, "loss": 0.7763, "step": 241 }, { "epoch": 0.07480680061823802, "grad_norm": 0.12125882506370544, "learning_rate": 0.0001856124031007752, "loss": 0.7467, "step": 242 }, { "epoch": 0.07511591962905718, "grad_norm": 0.13202430307865143, "learning_rate": 0.00018555038759689925, "loss": 0.8348, "step": 243 }, { "epoch": 0.07542503863987635, "grad_norm": 0.13780809938907623, "learning_rate": 0.00018548837209302326, "loss": 0.7159, "step": 244 }, { "epoch": 0.07573415765069552, "grad_norm": 0.163734570145607, "learning_rate": 0.0001854263565891473, "loss": 0.8078, "step": 245 }, { "epoch": 0.07604327666151468, "grad_norm": 0.15040288865566254, "learning_rate": 0.0001853643410852713, "loss": 0.7911, "step": 246 }, { "epoch": 0.07635239567233384, "grad_norm": 0.13316433131694794, "learning_rate": 0.00018530232558139535, "loss": 0.8217, "step": 247 }, { "epoch": 0.07666151468315302, "grad_norm": 0.14527438580989838, "learning_rate": 0.0001852403100775194, "loss": 0.8287, "step": 248 }, { "epoch": 0.07697063369397218, "grad_norm": 0.11744588613510132, "learning_rate": 0.00018517829457364343, "loss": 0.8774, "step": 249 }, { "epoch": 0.07727975270479134, "grad_norm": 0.15297925472259521, "learning_rate": 0.00018511627906976744, "loss": 0.8035, "step": 250 }, { "epoch": 0.0775888717156105, "grad_norm": 0.12520894408226013, "learning_rate": 0.00018505426356589148, "loss": 0.812, "step": 251 }, { "epoch": 0.07789799072642968, "grad_norm": 0.12046486139297485, "learning_rate": 0.00018499224806201552, "loss": 0.7703, "step": 252 }, { "epoch": 0.07820710973724884, "grad_norm": 0.10196825861930847, "learning_rate": 0.00018493023255813955, "loss": 0.7299, "step": 253 }, { "epoch": 0.078516228748068, "grad_norm": 0.12353216111660004, "learning_rate": 0.0001848682170542636, "loss": 0.6998, "step": 254 }, { "epoch": 0.07882534775888717, "grad_norm": 0.10435248166322708, "learning_rate": 0.0001848062015503876, "loss": 0.8073, "step": 255 }, { "epoch": 0.07913446676970634, "grad_norm": 0.1290121078491211, "learning_rate": 0.00018474418604651164, "loss": 0.7954, "step": 256 }, { "epoch": 0.0794435857805255, "grad_norm": 0.12450750917196274, "learning_rate": 0.00018468217054263566, "loss": 0.9049, "step": 257 }, { "epoch": 0.07975270479134466, "grad_norm": 0.1351582258939743, "learning_rate": 0.0001846201550387597, "loss": 0.7842, "step": 258 }, { "epoch": 0.08006182380216384, "grad_norm": 0.13335275650024414, "learning_rate": 0.00018455813953488373, "loss": 0.7771, "step": 259 }, { "epoch": 0.080370942812983, "grad_norm": 0.10518497973680496, "learning_rate": 0.00018449612403100774, "loss": 0.7927, "step": 260 }, { "epoch": 0.08068006182380216, "grad_norm": 0.11359915882349014, "learning_rate": 0.00018443410852713178, "loss": 0.8461, "step": 261 }, { "epoch": 0.08098918083462132, "grad_norm": 0.12962335348129272, "learning_rate": 0.00018437209302325582, "loss": 0.8299, "step": 262 }, { "epoch": 0.0812982998454405, "grad_norm": 0.1394529491662979, "learning_rate": 0.00018431007751937986, "loss": 0.8031, "step": 263 }, { "epoch": 0.08160741885625966, "grad_norm": 0.11067520827054977, "learning_rate": 0.0001842480620155039, "loss": 0.8019, "step": 264 }, { "epoch": 0.08191653786707882, "grad_norm": 0.14076265692710876, "learning_rate": 0.0001841860465116279, "loss": 0.8521, "step": 265 }, { "epoch": 0.08222565687789798, "grad_norm": 0.14540016651153564, "learning_rate": 0.00018412403100775195, "loss": 0.8124, "step": 266 }, { "epoch": 0.08253477588871716, "grad_norm": 0.14692644774913788, "learning_rate": 0.000184062015503876, "loss": 0.8629, "step": 267 }, { "epoch": 0.08284389489953632, "grad_norm": 0.12723390758037567, "learning_rate": 0.00018400000000000003, "loss": 0.7565, "step": 268 }, { "epoch": 0.08315301391035548, "grad_norm": 0.13681192696094513, "learning_rate": 0.00018393798449612404, "loss": 0.8458, "step": 269 }, { "epoch": 0.08346213292117466, "grad_norm": 0.1476822942495346, "learning_rate": 0.00018387596899224805, "loss": 0.7544, "step": 270 }, { "epoch": 0.08377125193199382, "grad_norm": 0.11408766359090805, "learning_rate": 0.0001838139534883721, "loss": 0.8782, "step": 271 }, { "epoch": 0.08408037094281298, "grad_norm": 0.14710593223571777, "learning_rate": 0.00018375193798449613, "loss": 0.7568, "step": 272 }, { "epoch": 0.08438948995363214, "grad_norm": 0.12558385729789734, "learning_rate": 0.00018368992248062017, "loss": 0.8574, "step": 273 }, { "epoch": 0.08469860896445132, "grad_norm": 0.11890698224306107, "learning_rate": 0.00018362790697674418, "loss": 0.6688, "step": 274 }, { "epoch": 0.08500772797527048, "grad_norm": 0.10440011322498322, "learning_rate": 0.00018356589147286822, "loss": 0.7699, "step": 275 }, { "epoch": 0.08531684698608964, "grad_norm": 0.13725546002388, "learning_rate": 0.00018350387596899226, "loss": 0.8359, "step": 276 }, { "epoch": 0.0856259659969088, "grad_norm": 0.12335329502820969, "learning_rate": 0.0001834418604651163, "loss": 0.7212, "step": 277 }, { "epoch": 0.08593508500772798, "grad_norm": 0.1138865053653717, "learning_rate": 0.00018337984496124033, "loss": 0.6597, "step": 278 }, { "epoch": 0.08624420401854714, "grad_norm": 0.11864970624446869, "learning_rate": 0.00018331782945736435, "loss": 0.7944, "step": 279 }, { "epoch": 0.0865533230293663, "grad_norm": 0.14360670745372772, "learning_rate": 0.00018325581395348838, "loss": 0.8186, "step": 280 }, { "epoch": 0.08686244204018548, "grad_norm": 0.13418716192245483, "learning_rate": 0.00018319379844961242, "loss": 0.8478, "step": 281 }, { "epoch": 0.08717156105100464, "grad_norm": 0.13283377885818481, "learning_rate": 0.00018313178294573646, "loss": 0.8309, "step": 282 }, { "epoch": 0.0874806800618238, "grad_norm": 0.11697278916835785, "learning_rate": 0.00018306976744186047, "loss": 0.8521, "step": 283 }, { "epoch": 0.08778979907264296, "grad_norm": 0.11819571256637573, "learning_rate": 0.00018300775193798448, "loss": 0.6976, "step": 284 }, { "epoch": 0.08809891808346214, "grad_norm": 0.11848420649766922, "learning_rate": 0.00018294573643410852, "loss": 0.8549, "step": 285 }, { "epoch": 0.0884080370942813, "grad_norm": 0.10397352278232574, "learning_rate": 0.00018288372093023256, "loss": 0.6554, "step": 286 }, { "epoch": 0.08871715610510046, "grad_norm": 0.15076309442520142, "learning_rate": 0.0001828217054263566, "loss": 0.6948, "step": 287 }, { "epoch": 0.08902627511591962, "grad_norm": 0.13722991943359375, "learning_rate": 0.0001827596899224806, "loss": 0.8466, "step": 288 }, { "epoch": 0.0893353941267388, "grad_norm": 0.11547433584928513, "learning_rate": 0.00018269767441860465, "loss": 0.7359, "step": 289 }, { "epoch": 0.08964451313755796, "grad_norm": 0.12837247550487518, "learning_rate": 0.0001826356589147287, "loss": 0.8533, "step": 290 }, { "epoch": 0.08995363214837712, "grad_norm": 0.11957511305809021, "learning_rate": 0.00018257364341085273, "loss": 0.6918, "step": 291 }, { "epoch": 0.0902627511591963, "grad_norm": 0.11487089097499847, "learning_rate": 0.00018251162790697677, "loss": 0.8411, "step": 292 }, { "epoch": 0.09057187017001546, "grad_norm": 0.12621980905532837, "learning_rate": 0.00018244961240310078, "loss": 0.7913, "step": 293 }, { "epoch": 0.09088098918083462, "grad_norm": 0.14285391569137573, "learning_rate": 0.00018238759689922482, "loss": 0.9191, "step": 294 }, { "epoch": 0.09119010819165378, "grad_norm": 0.14195428788661957, "learning_rate": 0.00018232558139534886, "loss": 0.8699, "step": 295 }, { "epoch": 0.09149922720247296, "grad_norm": 0.12400256842374802, "learning_rate": 0.0001822635658914729, "loss": 0.7707, "step": 296 }, { "epoch": 0.09180834621329212, "grad_norm": 0.1220916360616684, "learning_rate": 0.0001822015503875969, "loss": 0.7948, "step": 297 }, { "epoch": 0.09211746522411128, "grad_norm": 0.11888230592012405, "learning_rate": 0.00018213953488372092, "loss": 0.8237, "step": 298 }, { "epoch": 0.09242658423493044, "grad_norm": 0.134236678481102, "learning_rate": 0.00018207751937984496, "loss": 0.7503, "step": 299 }, { "epoch": 0.09273570324574962, "grad_norm": 0.09614330530166626, "learning_rate": 0.000182015503875969, "loss": 0.772, "step": 300 }, { "epoch": 0.09304482225656878, "grad_norm": 0.11686000227928162, "learning_rate": 0.00018195348837209303, "loss": 0.8648, "step": 301 }, { "epoch": 0.09335394126738794, "grad_norm": 0.11321427673101425, "learning_rate": 0.00018189147286821707, "loss": 0.8302, "step": 302 }, { "epoch": 0.09366306027820712, "grad_norm": 0.12898504734039307, "learning_rate": 0.00018182945736434109, "loss": 0.8, "step": 303 }, { "epoch": 0.09397217928902628, "grad_norm": 0.10747554153203964, "learning_rate": 0.00018176744186046512, "loss": 0.7696, "step": 304 }, { "epoch": 0.09428129829984544, "grad_norm": 0.12886860966682434, "learning_rate": 0.00018170542635658916, "loss": 0.8459, "step": 305 }, { "epoch": 0.0945904173106646, "grad_norm": 0.12835724651813507, "learning_rate": 0.0001816434108527132, "loss": 0.8523, "step": 306 }, { "epoch": 0.09489953632148378, "grad_norm": 0.11970589309930801, "learning_rate": 0.0001815813953488372, "loss": 0.7864, "step": 307 }, { "epoch": 0.09520865533230294, "grad_norm": 0.12258201837539673, "learning_rate": 0.00018151937984496125, "loss": 0.7593, "step": 308 }, { "epoch": 0.0955177743431221, "grad_norm": 0.1291266828775406, "learning_rate": 0.0001814573643410853, "loss": 0.925, "step": 309 }, { "epoch": 0.09582689335394126, "grad_norm": 0.12266039103269577, "learning_rate": 0.0001813953488372093, "loss": 0.7341, "step": 310 }, { "epoch": 0.09613601236476044, "grad_norm": 0.10808485746383667, "learning_rate": 0.00018133333333333334, "loss": 0.8983, "step": 311 }, { "epoch": 0.0964451313755796, "grad_norm": 0.1303120255470276, "learning_rate": 0.00018127131782945735, "loss": 0.9788, "step": 312 }, { "epoch": 0.09675425038639876, "grad_norm": 0.1282745897769928, "learning_rate": 0.0001812093023255814, "loss": 0.6776, "step": 313 }, { "epoch": 0.09706336939721794, "grad_norm": 0.10674197226762772, "learning_rate": 0.00018114728682170543, "loss": 0.819, "step": 314 }, { "epoch": 0.0973724884080371, "grad_norm": 0.10732909291982651, "learning_rate": 0.00018108527131782947, "loss": 0.8426, "step": 315 }, { "epoch": 0.09768160741885626, "grad_norm": 0.14202672243118286, "learning_rate": 0.0001810232558139535, "loss": 0.8447, "step": 316 }, { "epoch": 0.09799072642967542, "grad_norm": 0.12257728725671768, "learning_rate": 0.00018096124031007752, "loss": 0.8032, "step": 317 }, { "epoch": 0.0982998454404946, "grad_norm": 0.11397712677717209, "learning_rate": 0.00018089922480620156, "loss": 0.7667, "step": 318 }, { "epoch": 0.09860896445131376, "grad_norm": 0.11759169399738312, "learning_rate": 0.0001808372093023256, "loss": 0.775, "step": 319 }, { "epoch": 0.09891808346213292, "grad_norm": 0.10919482260942459, "learning_rate": 0.00018077519379844964, "loss": 0.6987, "step": 320 }, { "epoch": 0.09922720247295208, "grad_norm": 0.14136123657226562, "learning_rate": 0.00018071317829457365, "loss": 0.7642, "step": 321 }, { "epoch": 0.09953632148377126, "grad_norm": 0.12550586462020874, "learning_rate": 0.00018065116279069769, "loss": 0.9144, "step": 322 }, { "epoch": 0.09984544049459042, "grad_norm": 0.1267971247434616, "learning_rate": 0.00018058914728682172, "loss": 0.8512, "step": 323 }, { "epoch": 0.10015455950540958, "grad_norm": 0.12473420053720474, "learning_rate": 0.00018052713178294574, "loss": 0.9391, "step": 324 }, { "epoch": 0.10046367851622875, "grad_norm": 0.11510586738586426, "learning_rate": 0.00018046511627906977, "loss": 0.7916, "step": 325 }, { "epoch": 0.10077279752704792, "grad_norm": 0.13380743563175201, "learning_rate": 0.00018040310077519381, "loss": 0.819, "step": 326 }, { "epoch": 0.10108191653786708, "grad_norm": 0.1224348247051239, "learning_rate": 0.00018034108527131783, "loss": 0.778, "step": 327 }, { "epoch": 0.10139103554868624, "grad_norm": 0.11977488547563553, "learning_rate": 0.00018027906976744186, "loss": 0.8521, "step": 328 }, { "epoch": 0.10170015455950542, "grad_norm": 0.11883991211652756, "learning_rate": 0.0001802170542635659, "loss": 0.8959, "step": 329 }, { "epoch": 0.10200927357032458, "grad_norm": 0.13148127496242523, "learning_rate": 0.00018015503875968994, "loss": 0.7503, "step": 330 }, { "epoch": 0.10231839258114374, "grad_norm": 0.12128669023513794, "learning_rate": 0.00018009302325581395, "loss": 0.7469, "step": 331 }, { "epoch": 0.1026275115919629, "grad_norm": 0.12330310046672821, "learning_rate": 0.000180031007751938, "loss": 0.83, "step": 332 }, { "epoch": 0.10293663060278208, "grad_norm": 0.10930616408586502, "learning_rate": 0.00017996899224806203, "loss": 0.7841, "step": 333 }, { "epoch": 0.10324574961360124, "grad_norm": 0.12586379051208496, "learning_rate": 0.00017990697674418607, "loss": 0.729, "step": 334 }, { "epoch": 0.1035548686244204, "grad_norm": 0.11840980499982834, "learning_rate": 0.0001798449612403101, "loss": 0.7329, "step": 335 }, { "epoch": 0.10386398763523957, "grad_norm": 0.11878569424152374, "learning_rate": 0.00017978294573643412, "loss": 0.8202, "step": 336 }, { "epoch": 0.10417310664605874, "grad_norm": 0.1265515387058258, "learning_rate": 0.00017972093023255813, "loss": 0.6638, "step": 337 }, { "epoch": 0.1044822256568779, "grad_norm": 0.1272660791873932, "learning_rate": 0.00017965891472868217, "loss": 0.8512, "step": 338 }, { "epoch": 0.10479134466769706, "grad_norm": 0.11359579861164093, "learning_rate": 0.0001795968992248062, "loss": 0.8199, "step": 339 }, { "epoch": 0.10510046367851623, "grad_norm": 0.11645165085792542, "learning_rate": 0.00017953488372093025, "loss": 0.9163, "step": 340 }, { "epoch": 0.1054095826893354, "grad_norm": 0.11384947597980499, "learning_rate": 0.00017947286821705426, "loss": 0.7825, "step": 341 }, { "epoch": 0.10571870170015456, "grad_norm": 0.11389808356761932, "learning_rate": 0.0001794108527131783, "loss": 0.8078, "step": 342 }, { "epoch": 0.10602782071097372, "grad_norm": 0.12317777425050735, "learning_rate": 0.00017934883720930234, "loss": 0.9068, "step": 343 }, { "epoch": 0.1063369397217929, "grad_norm": 0.10351788252592087, "learning_rate": 0.00017928682170542638, "loss": 0.8258, "step": 344 }, { "epoch": 0.10664605873261206, "grad_norm": 0.11422822624444962, "learning_rate": 0.0001792248062015504, "loss": 0.8725, "step": 345 }, { "epoch": 0.10695517774343122, "grad_norm": 0.11480465531349182, "learning_rate": 0.00017916279069767443, "loss": 0.8415, "step": 346 }, { "epoch": 0.10726429675425038, "grad_norm": 0.11581287533044815, "learning_rate": 0.00017910077519379846, "loss": 0.6787, "step": 347 }, { "epoch": 0.10757341576506955, "grad_norm": 0.10481414198875427, "learning_rate": 0.0001790387596899225, "loss": 0.6735, "step": 348 }, { "epoch": 0.10788253477588872, "grad_norm": 0.12571753561496735, "learning_rate": 0.00017897674418604654, "loss": 0.7918, "step": 349 }, { "epoch": 0.10819165378670788, "grad_norm": 0.1036786288022995, "learning_rate": 0.00017891472868217055, "loss": 0.7985, "step": 350 }, { "epoch": 0.10850077279752705, "grad_norm": 0.12399487942457199, "learning_rate": 0.00017885271317829457, "loss": 0.9029, "step": 351 }, { "epoch": 0.10880989180834622, "grad_norm": 0.1186407133936882, "learning_rate": 0.0001787906976744186, "loss": 0.7395, "step": 352 }, { "epoch": 0.10911901081916538, "grad_norm": 0.1321779191493988, "learning_rate": 0.00017872868217054264, "loss": 0.7905, "step": 353 }, { "epoch": 0.10942812982998454, "grad_norm": 0.14797626435756683, "learning_rate": 0.00017866666666666668, "loss": 0.9305, "step": 354 }, { "epoch": 0.10973724884080371, "grad_norm": 0.11093270033597946, "learning_rate": 0.0001786046511627907, "loss": 0.8375, "step": 355 }, { "epoch": 0.11004636785162288, "grad_norm": 0.11384811252355576, "learning_rate": 0.00017854263565891473, "loss": 0.639, "step": 356 }, { "epoch": 0.11035548686244204, "grad_norm": 0.13438202440738678, "learning_rate": 0.00017848062015503877, "loss": 0.777, "step": 357 }, { "epoch": 0.1106646058732612, "grad_norm": 0.11255431920289993, "learning_rate": 0.0001784186046511628, "loss": 0.7742, "step": 358 }, { "epoch": 0.11097372488408037, "grad_norm": 0.12108633667230606, "learning_rate": 0.00017835658914728682, "loss": 0.7723, "step": 359 }, { "epoch": 0.11128284389489954, "grad_norm": 0.13009031116962433, "learning_rate": 0.00017829457364341086, "loss": 0.8108, "step": 360 }, { "epoch": 0.1115919629057187, "grad_norm": 0.10433927178382874, "learning_rate": 0.0001782325581395349, "loss": 0.6764, "step": 361 }, { "epoch": 0.11190108191653787, "grad_norm": 0.132685124874115, "learning_rate": 0.00017817054263565894, "loss": 0.8012, "step": 362 }, { "epoch": 0.11221020092735703, "grad_norm": 0.11265043169260025, "learning_rate": 0.00017810852713178298, "loss": 0.7925, "step": 363 }, { "epoch": 0.1125193199381762, "grad_norm": 0.12110339850187302, "learning_rate": 0.000178046511627907, "loss": 0.8647, "step": 364 }, { "epoch": 0.11282843894899536, "grad_norm": 0.1396140158176422, "learning_rate": 0.000177984496124031, "loss": 0.7221, "step": 365 }, { "epoch": 0.11313755795981453, "grad_norm": 0.11034229397773743, "learning_rate": 0.00017792248062015504, "loss": 0.8636, "step": 366 }, { "epoch": 0.1134466769706337, "grad_norm": 0.12675125896930695, "learning_rate": 0.00017786046511627908, "loss": 0.7799, "step": 367 }, { "epoch": 0.11375579598145286, "grad_norm": 0.10970692336559296, "learning_rate": 0.00017779844961240312, "loss": 0.8741, "step": 368 }, { "epoch": 0.11406491499227202, "grad_norm": 0.1316499263048172, "learning_rate": 0.00017773643410852713, "loss": 0.7235, "step": 369 }, { "epoch": 0.1143740340030912, "grad_norm": 0.15425892174243927, "learning_rate": 0.00017767441860465117, "loss": 0.7253, "step": 370 }, { "epoch": 0.11468315301391035, "grad_norm": 0.1116160973906517, "learning_rate": 0.0001776124031007752, "loss": 0.7544, "step": 371 }, { "epoch": 0.11499227202472952, "grad_norm": 0.112430639564991, "learning_rate": 0.00017755038759689924, "loss": 0.8633, "step": 372 }, { "epoch": 0.11530139103554869, "grad_norm": 0.12070276588201523, "learning_rate": 0.00017748837209302328, "loss": 0.7936, "step": 373 }, { "epoch": 0.11561051004636785, "grad_norm": 0.14540359377861023, "learning_rate": 0.0001774263565891473, "loss": 0.9689, "step": 374 }, { "epoch": 0.11591962905718702, "grad_norm": 0.1259058117866516, "learning_rate": 0.00017736434108527133, "loss": 0.8228, "step": 375 }, { "epoch": 0.11622874806800618, "grad_norm": 0.09805137664079666, "learning_rate": 0.00017730232558139537, "loss": 0.7173, "step": 376 }, { "epoch": 0.11653786707882535, "grad_norm": 0.1228744387626648, "learning_rate": 0.00017724031007751938, "loss": 0.7926, "step": 377 }, { "epoch": 0.11684698608964451, "grad_norm": 0.1288052648305893, "learning_rate": 0.00017717829457364342, "loss": 0.7403, "step": 378 }, { "epoch": 0.11715610510046368, "grad_norm": 0.11749331653118134, "learning_rate": 0.00017711627906976743, "loss": 0.702, "step": 379 }, { "epoch": 0.11746522411128284, "grad_norm": 0.12872126698493958, "learning_rate": 0.00017705426356589147, "loss": 0.7654, "step": 380 }, { "epoch": 0.11777434312210201, "grad_norm": 0.12806439399719238, "learning_rate": 0.0001769922480620155, "loss": 0.897, "step": 381 }, { "epoch": 0.11808346213292117, "grad_norm": 0.1399737149477005, "learning_rate": 0.00017693023255813955, "loss": 0.7706, "step": 382 }, { "epoch": 0.11839258114374034, "grad_norm": 0.14267806708812714, "learning_rate": 0.00017686821705426356, "loss": 0.8854, "step": 383 }, { "epoch": 0.11870170015455951, "grad_norm": 0.10857547074556351, "learning_rate": 0.0001768062015503876, "loss": 0.7281, "step": 384 }, { "epoch": 0.11901081916537867, "grad_norm": 0.11292342841625214, "learning_rate": 0.00017674418604651164, "loss": 0.7288, "step": 385 }, { "epoch": 0.11931993817619783, "grad_norm": 0.117954321205616, "learning_rate": 0.00017668217054263568, "loss": 0.7577, "step": 386 }, { "epoch": 0.119629057187017, "grad_norm": 0.11536258459091187, "learning_rate": 0.00017662015503875972, "loss": 0.8542, "step": 387 }, { "epoch": 0.11993817619783617, "grad_norm": 0.13967657089233398, "learning_rate": 0.00017655813953488373, "loss": 0.7666, "step": 388 }, { "epoch": 0.12024729520865533, "grad_norm": 0.124544158577919, "learning_rate": 0.00017649612403100777, "loss": 0.8928, "step": 389 }, { "epoch": 0.1205564142194745, "grad_norm": 0.11185236275196075, "learning_rate": 0.00017643410852713178, "loss": 0.665, "step": 390 }, { "epoch": 0.12086553323029366, "grad_norm": 0.11170051246881485, "learning_rate": 0.00017637209302325582, "loss": 0.7362, "step": 391 }, { "epoch": 0.12117465224111283, "grad_norm": 0.12095949798822403, "learning_rate": 0.00017631007751937986, "loss": 0.7309, "step": 392 }, { "epoch": 0.121483771251932, "grad_norm": 0.12416354566812515, "learning_rate": 0.00017624806201550387, "loss": 0.7633, "step": 393 }, { "epoch": 0.12179289026275116, "grad_norm": 0.11069466918706894, "learning_rate": 0.0001761860465116279, "loss": 0.7525, "step": 394 }, { "epoch": 0.12210200927357033, "grad_norm": 0.11477687954902649, "learning_rate": 0.00017612403100775195, "loss": 0.8462, "step": 395 }, { "epoch": 0.12241112828438949, "grad_norm": 0.13723276555538177, "learning_rate": 0.00017606201550387598, "loss": 0.9396, "step": 396 }, { "epoch": 0.12272024729520865, "grad_norm": 0.11079475283622742, "learning_rate": 0.00017600000000000002, "loss": 0.8683, "step": 397 }, { "epoch": 0.12302936630602782, "grad_norm": 0.11033523827791214, "learning_rate": 0.00017593798449612403, "loss": 0.6651, "step": 398 }, { "epoch": 0.12333848531684699, "grad_norm": 0.11088595539331436, "learning_rate": 0.00017587596899224807, "loss": 0.7183, "step": 399 }, { "epoch": 0.12364760432766615, "grad_norm": 0.11949151009321213, "learning_rate": 0.0001758139534883721, "loss": 0.752, "step": 400 }, { "epoch": 0.12395672333848531, "grad_norm": 0.11605624854564667, "learning_rate": 0.00017575193798449615, "loss": 0.8398, "step": 401 }, { "epoch": 0.12426584234930448, "grad_norm": 0.1079692542552948, "learning_rate": 0.00017568992248062016, "loss": 0.8006, "step": 402 }, { "epoch": 0.12457496136012365, "grad_norm": 0.12857861816883087, "learning_rate": 0.0001756279069767442, "loss": 0.754, "step": 403 }, { "epoch": 0.12488408037094281, "grad_norm": 0.11760881543159485, "learning_rate": 0.0001755658914728682, "loss": 0.7992, "step": 404 }, { "epoch": 0.125193199381762, "grad_norm": 0.1251303255558014, "learning_rate": 0.00017550387596899225, "loss": 0.8104, "step": 405 }, { "epoch": 0.12550231839258114, "grad_norm": 0.1187320277094841, "learning_rate": 0.0001754418604651163, "loss": 0.7542, "step": 406 }, { "epoch": 0.1258114374034003, "grad_norm": 0.1084708720445633, "learning_rate": 0.0001753798449612403, "loss": 0.7296, "step": 407 }, { "epoch": 0.1261205564142195, "grad_norm": 0.1298135370016098, "learning_rate": 0.00017531782945736434, "loss": 0.7441, "step": 408 }, { "epoch": 0.12642967542503863, "grad_norm": 0.1294536590576172, "learning_rate": 0.00017525581395348838, "loss": 0.7905, "step": 409 }, { "epoch": 0.1267387944358578, "grad_norm": 0.10958458483219147, "learning_rate": 0.00017519379844961242, "loss": 0.8569, "step": 410 }, { "epoch": 0.12704791344667696, "grad_norm": 0.12941788136959076, "learning_rate": 0.00017513178294573646, "loss": 0.8383, "step": 411 }, { "epoch": 0.12735703245749613, "grad_norm": 0.12861841917037964, "learning_rate": 0.00017506976744186047, "loss": 0.7911, "step": 412 }, { "epoch": 0.1276661514683153, "grad_norm": 0.13337025046348572, "learning_rate": 0.0001750077519379845, "loss": 0.7918, "step": 413 }, { "epoch": 0.12797527047913446, "grad_norm": 0.11046712100505829, "learning_rate": 0.00017494573643410855, "loss": 0.6964, "step": 414 }, { "epoch": 0.12828438948995363, "grad_norm": 0.10456400364637375, "learning_rate": 0.00017488372093023258, "loss": 0.8032, "step": 415 }, { "epoch": 0.1285935085007728, "grad_norm": 0.1251031905412674, "learning_rate": 0.0001748217054263566, "loss": 0.7741, "step": 416 }, { "epoch": 0.12890262751159196, "grad_norm": 0.13418059051036835, "learning_rate": 0.0001747596899224806, "loss": 0.862, "step": 417 }, { "epoch": 0.12921174652241113, "grad_norm": 0.11014249175786972, "learning_rate": 0.00017469767441860465, "loss": 0.8202, "step": 418 }, { "epoch": 0.1295208655332303, "grad_norm": 0.13367420434951782, "learning_rate": 0.00017463565891472869, "loss": 0.8483, "step": 419 }, { "epoch": 0.12982998454404945, "grad_norm": 0.11982861161231995, "learning_rate": 0.00017457364341085272, "loss": 0.7669, "step": 420 }, { "epoch": 0.13013910355486863, "grad_norm": 0.13078713417053223, "learning_rate": 0.00017451162790697674, "loss": 0.7996, "step": 421 }, { "epoch": 0.13044822256568778, "grad_norm": 0.1363217681646347, "learning_rate": 0.00017444961240310077, "loss": 0.8364, "step": 422 }, { "epoch": 0.13075734157650695, "grad_norm": 0.11756312847137451, "learning_rate": 0.0001743875968992248, "loss": 0.7808, "step": 423 }, { "epoch": 0.13106646058732613, "grad_norm": 0.12155081331729889, "learning_rate": 0.00017432558139534885, "loss": 0.7729, "step": 424 }, { "epoch": 0.13137557959814528, "grad_norm": 0.13399578630924225, "learning_rate": 0.0001742635658914729, "loss": 0.7278, "step": 425 }, { "epoch": 0.13168469860896445, "grad_norm": 0.12727884948253632, "learning_rate": 0.0001742015503875969, "loss": 0.7534, "step": 426 }, { "epoch": 0.13199381761978363, "grad_norm": 0.13630586862564087, "learning_rate": 0.00017413953488372094, "loss": 0.7763, "step": 427 }, { "epoch": 0.13230293663060277, "grad_norm": 0.14212100207805634, "learning_rate": 0.00017407751937984498, "loss": 0.8291, "step": 428 }, { "epoch": 0.13261205564142195, "grad_norm": 0.11936759203672409, "learning_rate": 0.00017401550387596902, "loss": 0.9107, "step": 429 }, { "epoch": 0.13292117465224113, "grad_norm": 0.11957745999097824, "learning_rate": 0.00017395348837209303, "loss": 0.7514, "step": 430 }, { "epoch": 0.13323029366306027, "grad_norm": 0.11473491042852402, "learning_rate": 0.00017389147286821704, "loss": 0.8061, "step": 431 }, { "epoch": 0.13353941267387945, "grad_norm": 0.12292005121707916, "learning_rate": 0.00017382945736434108, "loss": 0.801, "step": 432 }, { "epoch": 0.1338485316846986, "grad_norm": 0.11472901701927185, "learning_rate": 0.00017376744186046512, "loss": 0.7885, "step": 433 }, { "epoch": 0.13415765069551777, "grad_norm": 0.1211596429347992, "learning_rate": 0.00017370542635658916, "loss": 0.8281, "step": 434 }, { "epoch": 0.13446676970633695, "grad_norm": 0.1142617017030716, "learning_rate": 0.0001736434108527132, "loss": 0.823, "step": 435 }, { "epoch": 0.1347758887171561, "grad_norm": 0.10048012435436249, "learning_rate": 0.0001735813953488372, "loss": 0.8976, "step": 436 }, { "epoch": 0.13508500772797527, "grad_norm": 0.12125738710165024, "learning_rate": 0.00017351937984496125, "loss": 0.8559, "step": 437 }, { "epoch": 0.13539412673879445, "grad_norm": 0.12249696254730225, "learning_rate": 0.00017345736434108529, "loss": 0.783, "step": 438 }, { "epoch": 0.1357032457496136, "grad_norm": 0.14693719148635864, "learning_rate": 0.00017339534883720932, "loss": 0.8892, "step": 439 }, { "epoch": 0.13601236476043277, "grad_norm": 0.12531165778636932, "learning_rate": 0.00017333333333333334, "loss": 0.8706, "step": 440 }, { "epoch": 0.13632148377125194, "grad_norm": 0.1262034773826599, "learning_rate": 0.00017327131782945737, "loss": 0.7885, "step": 441 }, { "epoch": 0.1366306027820711, "grad_norm": 0.12192118167877197, "learning_rate": 0.00017320930232558141, "loss": 0.7533, "step": 442 }, { "epoch": 0.13693972179289027, "grad_norm": 0.12393314391374588, "learning_rate": 0.00017314728682170545, "loss": 0.7126, "step": 443 }, { "epoch": 0.13724884080370942, "grad_norm": 0.14559726417064667, "learning_rate": 0.00017308527131782946, "loss": 0.7041, "step": 444 }, { "epoch": 0.1375579598145286, "grad_norm": 0.11685144901275635, "learning_rate": 0.00017302325581395348, "loss": 0.7084, "step": 445 }, { "epoch": 0.13786707882534777, "grad_norm": 0.12664124369621277, "learning_rate": 0.00017296124031007751, "loss": 0.7724, "step": 446 }, { "epoch": 0.13817619783616691, "grad_norm": 0.1175457313656807, "learning_rate": 0.00017289922480620155, "loss": 0.8241, "step": 447 }, { "epoch": 0.1384853168469861, "grad_norm": 0.11846484243869781, "learning_rate": 0.0001728372093023256, "loss": 0.8515, "step": 448 }, { "epoch": 0.13879443585780527, "grad_norm": 0.13215206563472748, "learning_rate": 0.00017277519379844963, "loss": 0.8095, "step": 449 }, { "epoch": 0.1391035548686244, "grad_norm": 0.1288730353116989, "learning_rate": 0.00017271317829457364, "loss": 0.7236, "step": 450 }, { "epoch": 0.1394126738794436, "grad_norm": 0.11009534448385239, "learning_rate": 0.00017265116279069768, "loss": 0.8836, "step": 451 }, { "epoch": 0.13972179289026276, "grad_norm": 0.1256999373435974, "learning_rate": 0.00017258914728682172, "loss": 0.8412, "step": 452 }, { "epoch": 0.1400309119010819, "grad_norm": 0.12464401125907898, "learning_rate": 0.00017252713178294576, "loss": 0.8152, "step": 453 }, { "epoch": 0.1403400309119011, "grad_norm": 0.11386653035879135, "learning_rate": 0.00017246511627906977, "loss": 0.7923, "step": 454 }, { "epoch": 0.14064914992272023, "grad_norm": 0.11337646096944809, "learning_rate": 0.0001724031007751938, "loss": 0.6623, "step": 455 }, { "epoch": 0.1409582689335394, "grad_norm": 0.13900204002857208, "learning_rate": 0.00017234108527131785, "loss": 0.8303, "step": 456 }, { "epoch": 0.14126738794435859, "grad_norm": 0.13519424200057983, "learning_rate": 0.00017227906976744186, "loss": 0.7967, "step": 457 }, { "epoch": 0.14157650695517773, "grad_norm": 0.12967944145202637, "learning_rate": 0.0001722170542635659, "loss": 0.7826, "step": 458 }, { "epoch": 0.1418856259659969, "grad_norm": 0.12591594457626343, "learning_rate": 0.00017215503875968994, "loss": 0.9955, "step": 459 }, { "epoch": 0.14219474497681608, "grad_norm": 0.11622080206871033, "learning_rate": 0.00017209302325581395, "loss": 0.8291, "step": 460 }, { "epoch": 0.14250386398763523, "grad_norm": 0.12004160135984421, "learning_rate": 0.000172031007751938, "loss": 0.8015, "step": 461 }, { "epoch": 0.1428129829984544, "grad_norm": 0.11545343697071075, "learning_rate": 0.00017196899224806203, "loss": 0.7617, "step": 462 }, { "epoch": 0.14312210200927358, "grad_norm": 0.1136220321059227, "learning_rate": 0.00017190697674418606, "loss": 0.723, "step": 463 }, { "epoch": 0.14343122102009273, "grad_norm": 0.11028563231229782, "learning_rate": 0.00017184496124031008, "loss": 0.8464, "step": 464 }, { "epoch": 0.1437403400309119, "grad_norm": 0.10660995543003082, "learning_rate": 0.00017178294573643412, "loss": 0.7367, "step": 465 }, { "epoch": 0.14404945904173105, "grad_norm": 0.10705665498971939, "learning_rate": 0.00017172093023255815, "loss": 0.6892, "step": 466 }, { "epoch": 0.14435857805255023, "grad_norm": 0.124393992125988, "learning_rate": 0.0001716589147286822, "loss": 0.7663, "step": 467 }, { "epoch": 0.1446676970633694, "grad_norm": 0.10380648076534271, "learning_rate": 0.00017159689922480623, "loss": 0.7851, "step": 468 }, { "epoch": 0.14497681607418855, "grad_norm": 0.13513809442520142, "learning_rate": 0.00017153488372093024, "loss": 0.8207, "step": 469 }, { "epoch": 0.14528593508500773, "grad_norm": 0.1310744434595108, "learning_rate": 0.00017147286821705425, "loss": 0.7328, "step": 470 }, { "epoch": 0.1455950540958269, "grad_norm": 0.13068106770515442, "learning_rate": 0.0001714108527131783, "loss": 0.8289, "step": 471 }, { "epoch": 0.14590417310664605, "grad_norm": 0.09564946591854095, "learning_rate": 0.00017134883720930233, "loss": 0.7596, "step": 472 }, { "epoch": 0.14621329211746523, "grad_norm": 0.11033451557159424, "learning_rate": 0.00017128682170542637, "loss": 0.802, "step": 473 }, { "epoch": 0.1465224111282844, "grad_norm": 0.11931835860013962, "learning_rate": 0.00017122480620155038, "loss": 0.7087, "step": 474 }, { "epoch": 0.14683153013910355, "grad_norm": 0.12470009177923203, "learning_rate": 0.00017116279069767442, "loss": 0.7323, "step": 475 }, { "epoch": 0.14714064914992273, "grad_norm": 0.1364419013261795, "learning_rate": 0.00017110077519379846, "loss": 0.7856, "step": 476 }, { "epoch": 0.14744976816074187, "grad_norm": 0.12685492634773254, "learning_rate": 0.0001710387596899225, "loss": 0.8633, "step": 477 }, { "epoch": 0.14775888717156105, "grad_norm": 0.11873108893632889, "learning_rate": 0.0001709767441860465, "loss": 0.7816, "step": 478 }, { "epoch": 0.14806800618238022, "grad_norm": 0.12090124189853668, "learning_rate": 0.00017091472868217055, "loss": 0.9065, "step": 479 }, { "epoch": 0.14837712519319937, "grad_norm": 0.11901501566171646, "learning_rate": 0.0001708527131782946, "loss": 0.8258, "step": 480 }, { "epoch": 0.14868624420401855, "grad_norm": 0.11180437356233597, "learning_rate": 0.00017079069767441863, "loss": 0.7396, "step": 481 }, { "epoch": 0.14899536321483772, "grad_norm": 0.16076162457466125, "learning_rate": 0.00017072868217054267, "loss": 0.7794, "step": 482 }, { "epoch": 0.14930448222565687, "grad_norm": 0.13752448558807373, "learning_rate": 0.00017066666666666668, "loss": 0.6987, "step": 483 }, { "epoch": 0.14961360123647605, "grad_norm": 0.11785899847745895, "learning_rate": 0.0001706046511627907, "loss": 0.8017, "step": 484 }, { "epoch": 0.14992272024729522, "grad_norm": 0.11127035319805145, "learning_rate": 0.00017054263565891473, "loss": 0.7538, "step": 485 }, { "epoch": 0.15023183925811437, "grad_norm": 0.12820091843605042, "learning_rate": 0.00017048062015503877, "loss": 0.7771, "step": 486 }, { "epoch": 0.15054095826893354, "grad_norm": 0.13787533342838287, "learning_rate": 0.0001704186046511628, "loss": 0.789, "step": 487 }, { "epoch": 0.1508500772797527, "grad_norm": 0.12819816172122955, "learning_rate": 0.00017035658914728682, "loss": 0.7542, "step": 488 }, { "epoch": 0.15115919629057187, "grad_norm": 0.12091512233018875, "learning_rate": 0.00017029457364341086, "loss": 0.8161, "step": 489 }, { "epoch": 0.15146831530139104, "grad_norm": 0.1256888210773468, "learning_rate": 0.0001702325581395349, "loss": 0.8272, "step": 490 }, { "epoch": 0.1517774343122102, "grad_norm": 0.11789566278457642, "learning_rate": 0.00017017054263565893, "loss": 0.7079, "step": 491 }, { "epoch": 0.15208655332302937, "grad_norm": 0.11957567185163498, "learning_rate": 0.00017010852713178294, "loss": 0.8352, "step": 492 }, { "epoch": 0.15239567233384854, "grad_norm": 0.11315543204545975, "learning_rate": 0.00017004651162790698, "loss": 0.9455, "step": 493 }, { "epoch": 0.1527047913446677, "grad_norm": 0.1091320812702179, "learning_rate": 0.00016998449612403102, "loss": 0.686, "step": 494 }, { "epoch": 0.15301391035548687, "grad_norm": 0.11446017026901245, "learning_rate": 0.00016992248062015506, "loss": 0.8191, "step": 495 }, { "epoch": 0.15332302936630604, "grad_norm": 0.11834724992513657, "learning_rate": 0.0001698604651162791, "loss": 0.8379, "step": 496 }, { "epoch": 0.1536321483771252, "grad_norm": 0.12001053988933563, "learning_rate": 0.0001697984496124031, "loss": 0.7332, "step": 497 }, { "epoch": 0.15394126738794436, "grad_norm": 0.11104556918144226, "learning_rate": 0.00016973643410852712, "loss": 0.8642, "step": 498 }, { "epoch": 0.1542503863987635, "grad_norm": 0.10420899838209152, "learning_rate": 0.00016967441860465116, "loss": 0.8252, "step": 499 }, { "epoch": 0.1545595054095827, "grad_norm": 0.1481151580810547, "learning_rate": 0.0001696124031007752, "loss": 0.7085, "step": 500 }, { "epoch": 0.15486862442040186, "grad_norm": 0.13192850351333618, "learning_rate": 0.00016955038759689924, "loss": 0.8223, "step": 501 }, { "epoch": 0.155177743431221, "grad_norm": 0.11016976088285446, "learning_rate": 0.00016948837209302325, "loss": 0.7833, "step": 502 }, { "epoch": 0.15548686244204019, "grad_norm": 0.13597513735294342, "learning_rate": 0.0001694263565891473, "loss": 0.8608, "step": 503 }, { "epoch": 0.15579598145285936, "grad_norm": 0.13814714550971985, "learning_rate": 0.00016936434108527133, "loss": 0.8234, "step": 504 }, { "epoch": 0.1561051004636785, "grad_norm": 0.1129792258143425, "learning_rate": 0.00016930232558139537, "loss": 0.7622, "step": 505 }, { "epoch": 0.15641421947449768, "grad_norm": 0.1326257288455963, "learning_rate": 0.0001692403100775194, "loss": 0.7745, "step": 506 }, { "epoch": 0.15672333848531686, "grad_norm": 0.10894762724637985, "learning_rate": 0.00016917829457364342, "loss": 0.8396, "step": 507 }, { "epoch": 0.157032457496136, "grad_norm": 0.10844721645116806, "learning_rate": 0.00016911627906976746, "loss": 0.6738, "step": 508 }, { "epoch": 0.15734157650695518, "grad_norm": 0.12142128497362137, "learning_rate": 0.0001690542635658915, "loss": 0.8698, "step": 509 }, { "epoch": 0.15765069551777433, "grad_norm": 0.12891779839992523, "learning_rate": 0.0001689922480620155, "loss": 0.7915, "step": 510 }, { "epoch": 0.1579598145285935, "grad_norm": 0.1314953863620758, "learning_rate": 0.00016893023255813955, "loss": 0.8347, "step": 511 }, { "epoch": 0.15826893353941268, "grad_norm": 0.12055188417434692, "learning_rate": 0.00016886821705426356, "loss": 0.9093, "step": 512 }, { "epoch": 0.15857805255023183, "grad_norm": 0.12292719632387161, "learning_rate": 0.0001688062015503876, "loss": 0.8591, "step": 513 }, { "epoch": 0.158887171561051, "grad_norm": 0.11341209709644318, "learning_rate": 0.00016874418604651163, "loss": 0.7724, "step": 514 }, { "epoch": 0.15919629057187018, "grad_norm": 0.11800853163003922, "learning_rate": 0.00016868217054263567, "loss": 0.707, "step": 515 }, { "epoch": 0.15950540958268933, "grad_norm": 0.14028507471084595, "learning_rate": 0.00016862015503875968, "loss": 0.7773, "step": 516 }, { "epoch": 0.1598145285935085, "grad_norm": 0.11926918476819992, "learning_rate": 0.00016855813953488372, "loss": 0.8076, "step": 517 }, { "epoch": 0.16012364760432768, "grad_norm": 0.11683503538370132, "learning_rate": 0.00016849612403100776, "loss": 0.7492, "step": 518 }, { "epoch": 0.16043276661514683, "grad_norm": 0.14212507009506226, "learning_rate": 0.0001684341085271318, "loss": 0.7072, "step": 519 }, { "epoch": 0.160741885625966, "grad_norm": 0.12642718851566315, "learning_rate": 0.00016837209302325584, "loss": 0.9201, "step": 520 }, { "epoch": 0.16105100463678515, "grad_norm": 0.15104375779628754, "learning_rate": 0.00016831007751937985, "loss": 0.7333, "step": 521 }, { "epoch": 0.16136012364760433, "grad_norm": 0.15067335963249207, "learning_rate": 0.0001682480620155039, "loss": 0.8267, "step": 522 }, { "epoch": 0.1616692426584235, "grad_norm": 0.12420719116926193, "learning_rate": 0.00016818604651162793, "loss": 0.8489, "step": 523 }, { "epoch": 0.16197836166924265, "grad_norm": 0.10997667163610458, "learning_rate": 0.00016812403100775194, "loss": 0.7615, "step": 524 }, { "epoch": 0.16228748068006182, "grad_norm": 0.12284649908542633, "learning_rate": 0.00016806201550387598, "loss": 0.7435, "step": 525 }, { "epoch": 0.162596599690881, "grad_norm": 0.10515284538269043, "learning_rate": 0.000168, "loss": 0.789, "step": 526 }, { "epoch": 0.16290571870170015, "grad_norm": 0.12312375009059906, "learning_rate": 0.00016793798449612403, "loss": 0.8253, "step": 527 }, { "epoch": 0.16321483771251932, "grad_norm": 0.10993171483278275, "learning_rate": 0.00016787596899224807, "loss": 0.7591, "step": 528 }, { "epoch": 0.1635239567233385, "grad_norm": 0.11605069786310196, "learning_rate": 0.0001678139534883721, "loss": 0.8267, "step": 529 }, { "epoch": 0.16383307573415765, "grad_norm": 0.12003269046545029, "learning_rate": 0.00016775193798449615, "loss": 0.8455, "step": 530 }, { "epoch": 0.16414219474497682, "grad_norm": 0.12208808213472366, "learning_rate": 0.00016768992248062016, "loss": 0.8168, "step": 531 }, { "epoch": 0.16445131375579597, "grad_norm": 0.12368449568748474, "learning_rate": 0.0001676279069767442, "loss": 0.8713, "step": 532 }, { "epoch": 0.16476043276661514, "grad_norm": 0.12407387793064117, "learning_rate": 0.00016756589147286823, "loss": 0.7938, "step": 533 }, { "epoch": 0.16506955177743432, "grad_norm": 0.12617334723472595, "learning_rate": 0.00016750387596899227, "loss": 0.8575, "step": 534 }, { "epoch": 0.16537867078825347, "grad_norm": 0.15063488483428955, "learning_rate": 0.00016744186046511629, "loss": 0.9415, "step": 535 }, { "epoch": 0.16568778979907264, "grad_norm": 0.12658260762691498, "learning_rate": 0.00016737984496124032, "loss": 0.85, "step": 536 }, { "epoch": 0.16599690880989182, "grad_norm": 0.09913121163845062, "learning_rate": 0.00016731782945736434, "loss": 0.8092, "step": 537 }, { "epoch": 0.16630602782071097, "grad_norm": 0.12728868424892426, "learning_rate": 0.00016725581395348837, "loss": 0.6869, "step": 538 }, { "epoch": 0.16661514683153014, "grad_norm": 0.12953142821788788, "learning_rate": 0.0001671937984496124, "loss": 0.7474, "step": 539 }, { "epoch": 0.16692426584234932, "grad_norm": 0.1168576180934906, "learning_rate": 0.00016713178294573642, "loss": 0.7697, "step": 540 }, { "epoch": 0.16723338485316847, "grad_norm": 0.12081418931484222, "learning_rate": 0.00016706976744186046, "loss": 0.8952, "step": 541 }, { "epoch": 0.16754250386398764, "grad_norm": 0.12843774259090424, "learning_rate": 0.0001670077519379845, "loss": 0.8786, "step": 542 }, { "epoch": 0.1678516228748068, "grad_norm": 0.13334107398986816, "learning_rate": 0.00016694573643410854, "loss": 0.7732, "step": 543 }, { "epoch": 0.16816074188562596, "grad_norm": 0.12304075807332993, "learning_rate": 0.00016688372093023258, "loss": 0.8681, "step": 544 }, { "epoch": 0.16846986089644514, "grad_norm": 0.11800245940685272, "learning_rate": 0.0001668217054263566, "loss": 0.8226, "step": 545 }, { "epoch": 0.1687789799072643, "grad_norm": 0.1358041614294052, "learning_rate": 0.00016675968992248063, "loss": 0.7346, "step": 546 }, { "epoch": 0.16908809891808346, "grad_norm": 0.101251982152462, "learning_rate": 0.00016669767441860467, "loss": 0.7556, "step": 547 }, { "epoch": 0.16939721792890264, "grad_norm": 0.15012463927268982, "learning_rate": 0.0001666356589147287, "loss": 0.7558, "step": 548 }, { "epoch": 0.16970633693972179, "grad_norm": 0.11593160778284073, "learning_rate": 0.00016657364341085272, "loss": 0.7963, "step": 549 }, { "epoch": 0.17001545595054096, "grad_norm": 0.11717840284109116, "learning_rate": 0.00016651162790697673, "loss": 0.8219, "step": 550 }, { "epoch": 0.17032457496136014, "grad_norm": 0.12220215797424316, "learning_rate": 0.00016644961240310077, "loss": 0.8221, "step": 551 }, { "epoch": 0.17063369397217928, "grad_norm": 0.1270114630460739, "learning_rate": 0.0001663875968992248, "loss": 0.7718, "step": 552 }, { "epoch": 0.17094281298299846, "grad_norm": 0.11445185542106628, "learning_rate": 0.00016632558139534885, "loss": 0.8031, "step": 553 }, { "epoch": 0.1712519319938176, "grad_norm": 0.1258378028869629, "learning_rate": 0.00016626356589147286, "loss": 0.8165, "step": 554 }, { "epoch": 0.17156105100463678, "grad_norm": 0.12015929818153381, "learning_rate": 0.0001662015503875969, "loss": 0.7292, "step": 555 }, { "epoch": 0.17187017001545596, "grad_norm": 0.12911571562290192, "learning_rate": 0.00016613953488372094, "loss": 0.7586, "step": 556 }, { "epoch": 0.1721792890262751, "grad_norm": 0.11884018778800964, "learning_rate": 0.00016607751937984497, "loss": 0.816, "step": 557 }, { "epoch": 0.17248840803709428, "grad_norm": 0.12241604179143906, "learning_rate": 0.00016601550387596901, "loss": 0.785, "step": 558 }, { "epoch": 0.17279752704791346, "grad_norm": 0.1106485053896904, "learning_rate": 0.00016595348837209303, "loss": 0.8593, "step": 559 }, { "epoch": 0.1731066460587326, "grad_norm": 0.13855457305908203, "learning_rate": 0.00016589147286821706, "loss": 0.8647, "step": 560 }, { "epoch": 0.17341576506955178, "grad_norm": 0.12671570479869843, "learning_rate": 0.0001658294573643411, "loss": 0.7475, "step": 561 }, { "epoch": 0.17372488408037096, "grad_norm": 0.10744766145944595, "learning_rate": 0.00016576744186046514, "loss": 0.6681, "step": 562 }, { "epoch": 0.1740340030911901, "grad_norm": 0.15949758887290955, "learning_rate": 0.00016570542635658915, "loss": 0.7603, "step": 563 }, { "epoch": 0.17434312210200928, "grad_norm": 0.14625856280326843, "learning_rate": 0.00016564341085271316, "loss": 0.9034, "step": 564 }, { "epoch": 0.17465224111282843, "grad_norm": 0.10362540185451508, "learning_rate": 0.0001655813953488372, "loss": 0.8991, "step": 565 }, { "epoch": 0.1749613601236476, "grad_norm": 0.1137462630867958, "learning_rate": 0.00016551937984496124, "loss": 0.8242, "step": 566 }, { "epoch": 0.17527047913446678, "grad_norm": 0.12184014916419983, "learning_rate": 0.00016545736434108528, "loss": 0.7607, "step": 567 }, { "epoch": 0.17557959814528593, "grad_norm": 0.1274954229593277, "learning_rate": 0.00016539534883720932, "loss": 0.7599, "step": 568 }, { "epoch": 0.1758887171561051, "grad_norm": 0.11610583961009979, "learning_rate": 0.00016533333333333333, "loss": 0.8305, "step": 569 }, { "epoch": 0.17619783616692428, "grad_norm": 0.1202671155333519, "learning_rate": 0.00016527131782945737, "loss": 0.8162, "step": 570 }, { "epoch": 0.17650695517774342, "grad_norm": 0.1324416995048523, "learning_rate": 0.0001652093023255814, "loss": 0.808, "step": 571 }, { "epoch": 0.1768160741885626, "grad_norm": 0.11402853578329086, "learning_rate": 0.00016514728682170545, "loss": 0.856, "step": 572 }, { "epoch": 0.17712519319938178, "grad_norm": 0.11911406368017197, "learning_rate": 0.00016508527131782946, "loss": 0.7992, "step": 573 }, { "epoch": 0.17743431221020092, "grad_norm": 0.13559581339359283, "learning_rate": 0.0001650232558139535, "loss": 0.8584, "step": 574 }, { "epoch": 0.1777434312210201, "grad_norm": 0.12889279425144196, "learning_rate": 0.00016496124031007754, "loss": 0.8382, "step": 575 }, { "epoch": 0.17805255023183925, "grad_norm": 0.12666314840316772, "learning_rate": 0.00016489922480620158, "loss": 0.8406, "step": 576 }, { "epoch": 0.17836166924265842, "grad_norm": 0.12953141331672668, "learning_rate": 0.0001648372093023256, "loss": 0.7878, "step": 577 }, { "epoch": 0.1786707882534776, "grad_norm": 0.10341861099004745, "learning_rate": 0.0001647751937984496, "loss": 0.8802, "step": 578 }, { "epoch": 0.17897990726429674, "grad_norm": 0.11049186438322067, "learning_rate": 0.00016471317829457364, "loss": 0.7669, "step": 579 }, { "epoch": 0.17928902627511592, "grad_norm": 0.11524353176355362, "learning_rate": 0.00016465116279069768, "loss": 0.7958, "step": 580 }, { "epoch": 0.1795981452859351, "grad_norm": 0.12205459177494049, "learning_rate": 0.00016458914728682172, "loss": 0.6768, "step": 581 }, { "epoch": 0.17990726429675424, "grad_norm": 0.10032919049263, "learning_rate": 0.00016452713178294575, "loss": 0.8596, "step": 582 }, { "epoch": 0.18021638330757342, "grad_norm": 0.10666303336620331, "learning_rate": 0.00016446511627906977, "loss": 0.8106, "step": 583 }, { "epoch": 0.1805255023183926, "grad_norm": 0.09989852458238602, "learning_rate": 0.0001644031007751938, "loss": 0.7455, "step": 584 }, { "epoch": 0.18083462132921174, "grad_norm": 0.11961805075407028, "learning_rate": 0.00016434108527131784, "loss": 0.8289, "step": 585 }, { "epoch": 0.18114374034003092, "grad_norm": 0.13421611487865448, "learning_rate": 0.00016427906976744188, "loss": 0.6841, "step": 586 }, { "epoch": 0.18145285935085007, "grad_norm": 0.11855993419885635, "learning_rate": 0.0001642170542635659, "loss": 0.8299, "step": 587 }, { "epoch": 0.18176197836166924, "grad_norm": 0.11767081916332245, "learning_rate": 0.00016415503875968993, "loss": 0.7878, "step": 588 }, { "epoch": 0.18207109737248842, "grad_norm": 0.12189806997776031, "learning_rate": 0.00016409302325581397, "loss": 0.863, "step": 589 }, { "epoch": 0.18238021638330756, "grad_norm": 0.09818772226572037, "learning_rate": 0.00016403100775193798, "loss": 0.7039, "step": 590 }, { "epoch": 0.18268933539412674, "grad_norm": 0.13206005096435547, "learning_rate": 0.00016396899224806202, "loss": 0.7623, "step": 591 }, { "epoch": 0.18299845440494591, "grad_norm": 0.12963028252124786, "learning_rate": 0.00016390697674418606, "loss": 0.7811, "step": 592 }, { "epoch": 0.18330757341576506, "grad_norm": 0.11753853410482407, "learning_rate": 0.00016384496124031007, "loss": 0.8228, "step": 593 }, { "epoch": 0.18361669242658424, "grad_norm": 0.10470208525657654, "learning_rate": 0.0001637829457364341, "loss": 0.8566, "step": 594 }, { "epoch": 0.1839258114374034, "grad_norm": 0.10645218938589096, "learning_rate": 0.00016372093023255815, "loss": 0.7536, "step": 595 }, { "epoch": 0.18423493044822256, "grad_norm": 0.11461575329303741, "learning_rate": 0.0001636589147286822, "loss": 0.7328, "step": 596 }, { "epoch": 0.18454404945904174, "grad_norm": 0.1035584807395935, "learning_rate": 0.0001635968992248062, "loss": 0.7841, "step": 597 }, { "epoch": 0.18485316846986088, "grad_norm": 0.12418399751186371, "learning_rate": 0.00016353488372093024, "loss": 0.7869, "step": 598 }, { "epoch": 0.18516228748068006, "grad_norm": 0.12286582589149475, "learning_rate": 0.00016347286821705428, "loss": 0.8917, "step": 599 }, { "epoch": 0.18547140649149924, "grad_norm": 0.11596380174160004, "learning_rate": 0.00016341085271317832, "loss": 0.7949, "step": 600 }, { "epoch": 0.18578052550231838, "grad_norm": 0.12381494790315628, "learning_rate": 0.00016334883720930235, "loss": 0.8331, "step": 601 }, { "epoch": 0.18608964451313756, "grad_norm": 0.12967997789382935, "learning_rate": 0.00016328682170542637, "loss": 0.8003, "step": 602 }, { "epoch": 0.18639876352395673, "grad_norm": 0.11383350193500519, "learning_rate": 0.0001632248062015504, "loss": 0.7191, "step": 603 }, { "epoch": 0.18670788253477588, "grad_norm": 0.12088557332754135, "learning_rate": 0.00016316279069767442, "loss": 0.772, "step": 604 }, { "epoch": 0.18701700154559506, "grad_norm": 0.1383604258298874, "learning_rate": 0.00016310077519379846, "loss": 0.8449, "step": 605 }, { "epoch": 0.18732612055641423, "grad_norm": 0.13077442348003387, "learning_rate": 0.0001630387596899225, "loss": 0.817, "step": 606 }, { "epoch": 0.18763523956723338, "grad_norm": 0.12592090666294098, "learning_rate": 0.0001629767441860465, "loss": 0.8615, "step": 607 }, { "epoch": 0.18794435857805256, "grad_norm": 0.11555439233779907, "learning_rate": 0.00016291472868217054, "loss": 0.7587, "step": 608 }, { "epoch": 0.1882534775888717, "grad_norm": 0.12637798488140106, "learning_rate": 0.00016285271317829458, "loss": 0.7867, "step": 609 }, { "epoch": 0.18856259659969088, "grad_norm": 0.123162180185318, "learning_rate": 0.00016279069767441862, "loss": 0.8271, "step": 610 }, { "epoch": 0.18887171561051005, "grad_norm": 0.133504718542099, "learning_rate": 0.00016272868217054263, "loss": 0.8287, "step": 611 }, { "epoch": 0.1891808346213292, "grad_norm": 0.13327136635780334, "learning_rate": 0.00016266666666666667, "loss": 0.8069, "step": 612 }, { "epoch": 0.18948995363214838, "grad_norm": 0.13675393164157867, "learning_rate": 0.0001626046511627907, "loss": 0.7992, "step": 613 }, { "epoch": 0.18979907264296755, "grad_norm": 0.12681642174720764, "learning_rate": 0.00016254263565891475, "loss": 0.8107, "step": 614 }, { "epoch": 0.1901081916537867, "grad_norm": 0.11314484477043152, "learning_rate": 0.0001624806201550388, "loss": 0.8308, "step": 615 }, { "epoch": 0.19041731066460588, "grad_norm": 0.11311525851488113, "learning_rate": 0.0001624186046511628, "loss": 0.8193, "step": 616 }, { "epoch": 0.19072642967542505, "grad_norm": 0.10104167461395264, "learning_rate": 0.0001623565891472868, "loss": 0.8109, "step": 617 }, { "epoch": 0.1910355486862442, "grad_norm": 0.1340848058462143, "learning_rate": 0.00016229457364341085, "loss": 0.7129, "step": 618 }, { "epoch": 0.19134466769706338, "grad_norm": 0.1362898200750351, "learning_rate": 0.0001622325581395349, "loss": 0.7536, "step": 619 }, { "epoch": 0.19165378670788252, "grad_norm": 0.13276411592960358, "learning_rate": 0.00016217054263565893, "loss": 0.8628, "step": 620 }, { "epoch": 0.1919629057187017, "grad_norm": 0.11686565726995468, "learning_rate": 0.00016210852713178294, "loss": 0.8807, "step": 621 }, { "epoch": 0.19227202472952087, "grad_norm": 0.12405114620923996, "learning_rate": 0.00016204651162790698, "loss": 0.747, "step": 622 }, { "epoch": 0.19258114374034002, "grad_norm": 0.12308946251869202, "learning_rate": 0.00016198449612403102, "loss": 0.7715, "step": 623 }, { "epoch": 0.1928902627511592, "grad_norm": 0.14180584251880646, "learning_rate": 0.00016192248062015506, "loss": 0.9483, "step": 624 }, { "epoch": 0.19319938176197837, "grad_norm": 0.1168755367398262, "learning_rate": 0.00016186046511627907, "loss": 0.7669, "step": 625 }, { "epoch": 0.19350850077279752, "grad_norm": 0.11601582914590836, "learning_rate": 0.0001617984496124031, "loss": 0.8361, "step": 626 }, { "epoch": 0.1938176197836167, "grad_norm": 0.11746050417423248, "learning_rate": 0.00016173643410852715, "loss": 0.844, "step": 627 }, { "epoch": 0.19412673879443587, "grad_norm": 0.11663764715194702, "learning_rate": 0.00016167441860465118, "loss": 0.8162, "step": 628 }, { "epoch": 0.19443585780525502, "grad_norm": 0.1258978247642517, "learning_rate": 0.00016161240310077522, "loss": 0.7729, "step": 629 }, { "epoch": 0.1947449768160742, "grad_norm": 0.10496451705694199, "learning_rate": 0.00016155038759689923, "loss": 0.7739, "step": 630 }, { "epoch": 0.19505409582689334, "grad_norm": 0.1313450038433075, "learning_rate": 0.00016148837209302325, "loss": 0.8157, "step": 631 }, { "epoch": 0.19536321483771252, "grad_norm": 0.13735109567642212, "learning_rate": 0.00016142635658914728, "loss": 0.8494, "step": 632 }, { "epoch": 0.1956723338485317, "grad_norm": 0.11428606510162354, "learning_rate": 0.00016136434108527132, "loss": 0.7957, "step": 633 }, { "epoch": 0.19598145285935084, "grad_norm": 0.10336098074913025, "learning_rate": 0.00016130232558139536, "loss": 0.7244, "step": 634 }, { "epoch": 0.19629057187017002, "grad_norm": 0.12851019203662872, "learning_rate": 0.00016124031007751937, "loss": 0.7908, "step": 635 }, { "epoch": 0.1965996908809892, "grad_norm": 0.11097298562526703, "learning_rate": 0.0001611782945736434, "loss": 0.7819, "step": 636 }, { "epoch": 0.19690880989180834, "grad_norm": 0.10702291131019592, "learning_rate": 0.00016111627906976745, "loss": 0.9083, "step": 637 }, { "epoch": 0.19721792890262752, "grad_norm": 0.1185348853468895, "learning_rate": 0.0001610542635658915, "loss": 0.8115, "step": 638 }, { "epoch": 0.1975270479134467, "grad_norm": 0.12385392189025879, "learning_rate": 0.00016099224806201553, "loss": 0.847, "step": 639 }, { "epoch": 0.19783616692426584, "grad_norm": 0.13237705826759338, "learning_rate": 0.00016093023255813954, "loss": 0.7401, "step": 640 }, { "epoch": 0.198145285935085, "grad_norm": 0.11597932875156403, "learning_rate": 0.00016086821705426358, "loss": 0.7639, "step": 641 }, { "epoch": 0.19845440494590416, "grad_norm": 0.12065674364566803, "learning_rate": 0.00016080620155038762, "loss": 0.7425, "step": 642 }, { "epoch": 0.19876352395672334, "grad_norm": 0.10582825541496277, "learning_rate": 0.00016074418604651166, "loss": 0.8191, "step": 643 }, { "epoch": 0.1990726429675425, "grad_norm": 0.1253654509782791, "learning_rate": 0.00016068217054263567, "loss": 0.7928, "step": 644 }, { "epoch": 0.19938176197836166, "grad_norm": 0.13197046518325806, "learning_rate": 0.00016062015503875968, "loss": 0.7402, "step": 645 }, { "epoch": 0.19969088098918084, "grad_norm": 0.12224183231592178, "learning_rate": 0.00016055813953488372, "loss": 0.6898, "step": 646 }, { "epoch": 0.2, "grad_norm": 0.13466641306877136, "learning_rate": 0.00016049612403100776, "loss": 0.814, "step": 647 }, { "epoch": 0.20030911901081916, "grad_norm": 0.10830813646316528, "learning_rate": 0.0001604341085271318, "loss": 0.8274, "step": 648 }, { "epoch": 0.20061823802163833, "grad_norm": 0.1332327425479889, "learning_rate": 0.0001603720930232558, "loss": 0.8042, "step": 649 }, { "epoch": 0.2009273570324575, "grad_norm": 0.1263049840927124, "learning_rate": 0.00016031007751937985, "loss": 0.8042, "step": 650 }, { "epoch": 0.20123647604327666, "grad_norm": 0.13426467776298523, "learning_rate": 0.00016024806201550389, "loss": 0.8429, "step": 651 }, { "epoch": 0.20154559505409583, "grad_norm": 0.1457086056470871, "learning_rate": 0.00016018604651162792, "loss": 0.8961, "step": 652 }, { "epoch": 0.20185471406491498, "grad_norm": 0.11996602267026901, "learning_rate": 0.00016012403100775196, "loss": 0.7045, "step": 653 }, { "epoch": 0.20216383307573416, "grad_norm": 0.10409342497587204, "learning_rate": 0.00016006201550387597, "loss": 0.6763, "step": 654 }, { "epoch": 0.20247295208655333, "grad_norm": 0.11956805735826492, "learning_rate": 0.00016, "loss": 0.8169, "step": 655 }, { "epoch": 0.20278207109737248, "grad_norm": 0.11860910803079605, "learning_rate": 0.00015993798449612405, "loss": 0.74, "step": 656 }, { "epoch": 0.20309119010819165, "grad_norm": 0.12545433640480042, "learning_rate": 0.00015987596899224806, "loss": 0.8134, "step": 657 }, { "epoch": 0.20340030911901083, "grad_norm": 0.10243546217679977, "learning_rate": 0.0001598139534883721, "loss": 0.7597, "step": 658 }, { "epoch": 0.20370942812982998, "grad_norm": 0.11689910292625427, "learning_rate": 0.00015975193798449611, "loss": 0.7734, "step": 659 }, { "epoch": 0.20401854714064915, "grad_norm": 0.12140754610300064, "learning_rate": 0.00015968992248062015, "loss": 0.8014, "step": 660 }, { "epoch": 0.20432766615146833, "grad_norm": 0.11417256295681, "learning_rate": 0.0001596279069767442, "loss": 0.8837, "step": 661 }, { "epoch": 0.20463678516228748, "grad_norm": 0.16317220032215118, "learning_rate": 0.00015956589147286823, "loss": 0.894, "step": 662 }, { "epoch": 0.20494590417310665, "grad_norm": 0.1096215769648552, "learning_rate": 0.00015950387596899227, "loss": 0.6989, "step": 663 }, { "epoch": 0.2052550231839258, "grad_norm": 0.11856718361377716, "learning_rate": 0.00015944186046511628, "loss": 0.8425, "step": 664 }, { "epoch": 0.20556414219474498, "grad_norm": 0.11110817641019821, "learning_rate": 0.00015937984496124032, "loss": 0.8838, "step": 665 }, { "epoch": 0.20587326120556415, "grad_norm": 0.1287812888622284, "learning_rate": 0.00015931782945736436, "loss": 0.7789, "step": 666 }, { "epoch": 0.2061823802163833, "grad_norm": 0.0988534688949585, "learning_rate": 0.0001592558139534884, "loss": 0.7962, "step": 667 }, { "epoch": 0.20649149922720247, "grad_norm": 0.12815728783607483, "learning_rate": 0.0001591937984496124, "loss": 0.6849, "step": 668 }, { "epoch": 0.20680061823802165, "grad_norm": 0.12397190928459167, "learning_rate": 0.00015913178294573645, "loss": 0.771, "step": 669 }, { "epoch": 0.2071097372488408, "grad_norm": 0.13357098400592804, "learning_rate": 0.00015906976744186046, "loss": 0.9005, "step": 670 }, { "epoch": 0.20741885625965997, "grad_norm": 0.10197357833385468, "learning_rate": 0.0001590077519379845, "loss": 0.7802, "step": 671 }, { "epoch": 0.20772797527047915, "grad_norm": 0.12450321763753891, "learning_rate": 0.00015894573643410854, "loss": 0.7157, "step": 672 }, { "epoch": 0.2080370942812983, "grad_norm": 0.13183358311653137, "learning_rate": 0.00015888372093023255, "loss": 0.8082, "step": 673 }, { "epoch": 0.20834621329211747, "grad_norm": 0.12288426607847214, "learning_rate": 0.0001588217054263566, "loss": 0.8119, "step": 674 }, { "epoch": 0.20865533230293662, "grad_norm": 0.11869386583566666, "learning_rate": 0.00015875968992248063, "loss": 0.7948, "step": 675 }, { "epoch": 0.2089644513137558, "grad_norm": 0.11692944169044495, "learning_rate": 0.00015869767441860466, "loss": 0.7011, "step": 676 }, { "epoch": 0.20927357032457497, "grad_norm": 0.11799801886081696, "learning_rate": 0.0001586356589147287, "loss": 0.9002, "step": 677 }, { "epoch": 0.20958268933539412, "grad_norm": 0.14463946223258972, "learning_rate": 0.00015857364341085271, "loss": 0.7562, "step": 678 }, { "epoch": 0.2098918083462133, "grad_norm": 0.14642862975597382, "learning_rate": 0.00015851162790697675, "loss": 0.6619, "step": 679 }, { "epoch": 0.21020092735703247, "grad_norm": 0.11202511936426163, "learning_rate": 0.0001584496124031008, "loss": 0.6871, "step": 680 }, { "epoch": 0.21051004636785162, "grad_norm": 0.12495609372854233, "learning_rate": 0.00015838759689922483, "loss": 0.8702, "step": 681 }, { "epoch": 0.2108191653786708, "grad_norm": 0.1246587410569191, "learning_rate": 0.00015832558139534884, "loss": 0.7562, "step": 682 }, { "epoch": 0.21112828438948997, "grad_norm": 0.11207133531570435, "learning_rate": 0.00015826356589147285, "loss": 0.7948, "step": 683 }, { "epoch": 0.21143740340030912, "grad_norm": 0.10788938403129578, "learning_rate": 0.0001582015503875969, "loss": 0.7807, "step": 684 }, { "epoch": 0.2117465224111283, "grad_norm": 0.12020035088062286, "learning_rate": 0.00015813953488372093, "loss": 0.8755, "step": 685 }, { "epoch": 0.21205564142194744, "grad_norm": 0.13727609813213348, "learning_rate": 0.00015807751937984497, "loss": 0.9059, "step": 686 }, { "epoch": 0.2123647604327666, "grad_norm": 0.1516505777835846, "learning_rate": 0.00015801550387596898, "loss": 0.8297, "step": 687 }, { "epoch": 0.2126738794435858, "grad_norm": 0.13136766850948334, "learning_rate": 0.00015795348837209302, "loss": 0.8096, "step": 688 }, { "epoch": 0.21298299845440494, "grad_norm": 0.11303882300853729, "learning_rate": 0.00015789147286821706, "loss": 0.7967, "step": 689 }, { "epoch": 0.2132921174652241, "grad_norm": 0.13250216841697693, "learning_rate": 0.0001578294573643411, "loss": 0.8846, "step": 690 }, { "epoch": 0.2136012364760433, "grad_norm": 0.125604048371315, "learning_rate": 0.00015776744186046514, "loss": 0.8468, "step": 691 }, { "epoch": 0.21391035548686244, "grad_norm": 0.11047331243753433, "learning_rate": 0.00015770542635658915, "loss": 0.7068, "step": 692 }, { "epoch": 0.2142194744976816, "grad_norm": 0.12153135985136032, "learning_rate": 0.0001576434108527132, "loss": 0.7166, "step": 693 }, { "epoch": 0.21452859350850076, "grad_norm": 0.11161539703607559, "learning_rate": 0.00015758139534883723, "loss": 0.8544, "step": 694 }, { "epoch": 0.21483771251931993, "grad_norm": 0.11793739348649979, "learning_rate": 0.00015751937984496126, "loss": 0.8196, "step": 695 }, { "epoch": 0.2151468315301391, "grad_norm": 0.12710191309452057, "learning_rate": 0.00015745736434108528, "loss": 0.8249, "step": 696 }, { "epoch": 0.21545595054095826, "grad_norm": 0.12436945736408234, "learning_rate": 0.0001573953488372093, "loss": 0.7147, "step": 697 }, { "epoch": 0.21576506955177743, "grad_norm": 0.10970059782266617, "learning_rate": 0.00015733333333333333, "loss": 0.8297, "step": 698 }, { "epoch": 0.2160741885625966, "grad_norm": 0.1195238009095192, "learning_rate": 0.00015727131782945737, "loss": 0.7347, "step": 699 }, { "epoch": 0.21638330757341576, "grad_norm": 0.11942408978939056, "learning_rate": 0.0001572093023255814, "loss": 0.7825, "step": 700 }, { "epoch": 0.21669242658423493, "grad_norm": 0.11630623042583466, "learning_rate": 0.00015714728682170544, "loss": 0.8442, "step": 701 }, { "epoch": 0.2170015455950541, "grad_norm": 0.12051951140165329, "learning_rate": 0.00015708527131782945, "loss": 0.6505, "step": 702 }, { "epoch": 0.21731066460587325, "grad_norm": 0.12351769208908081, "learning_rate": 0.0001570232558139535, "loss": 0.8298, "step": 703 }, { "epoch": 0.21761978361669243, "grad_norm": 0.11859402805566788, "learning_rate": 0.00015696124031007753, "loss": 0.6916, "step": 704 }, { "epoch": 0.21792890262751158, "grad_norm": 0.13090530037879944, "learning_rate": 0.00015689922480620157, "loss": 0.9438, "step": 705 }, { "epoch": 0.21823802163833075, "grad_norm": 0.12880273163318634, "learning_rate": 0.00015683720930232558, "loss": 0.8416, "step": 706 }, { "epoch": 0.21854714064914993, "grad_norm": 0.12142153829336166, "learning_rate": 0.00015677519379844962, "loss": 0.7073, "step": 707 }, { "epoch": 0.21885625965996908, "grad_norm": 0.12591804563999176, "learning_rate": 0.00015671317829457366, "loss": 0.7311, "step": 708 }, { "epoch": 0.21916537867078825, "grad_norm": 0.1107344701886177, "learning_rate": 0.0001566511627906977, "loss": 0.6529, "step": 709 }, { "epoch": 0.21947449768160743, "grad_norm": 0.10950338840484619, "learning_rate": 0.0001565891472868217, "loss": 0.6766, "step": 710 }, { "epoch": 0.21978361669242658, "grad_norm": 0.1164025217294693, "learning_rate": 0.00015652713178294572, "loss": 0.7021, "step": 711 }, { "epoch": 0.22009273570324575, "grad_norm": 0.13531538844108582, "learning_rate": 0.00015646511627906976, "loss": 0.739, "step": 712 }, { "epoch": 0.22040185471406493, "grad_norm": 0.12463265657424927, "learning_rate": 0.0001564031007751938, "loss": 0.7832, "step": 713 }, { "epoch": 0.22071097372488407, "grad_norm": 0.13127584755420685, "learning_rate": 0.00015634108527131784, "loss": 0.7652, "step": 714 }, { "epoch": 0.22102009273570325, "grad_norm": 0.13645778596401215, "learning_rate": 0.00015627906976744188, "loss": 0.8023, "step": 715 }, { "epoch": 0.2213292117465224, "grad_norm": 0.12551645934581757, "learning_rate": 0.0001562170542635659, "loss": 0.7572, "step": 716 }, { "epoch": 0.22163833075734157, "grad_norm": 0.1264384686946869, "learning_rate": 0.00015615503875968993, "loss": 0.7972, "step": 717 }, { "epoch": 0.22194744976816075, "grad_norm": 0.12361495941877365, "learning_rate": 0.00015609302325581397, "loss": 0.7932, "step": 718 }, { "epoch": 0.2222565687789799, "grad_norm": 0.12109937518835068, "learning_rate": 0.000156031007751938, "loss": 0.7977, "step": 719 }, { "epoch": 0.22256568778979907, "grad_norm": 0.11111821234226227, "learning_rate": 0.00015596899224806202, "loss": 0.7563, "step": 720 }, { "epoch": 0.22287480680061825, "grad_norm": 0.11457593739032745, "learning_rate": 0.00015590697674418606, "loss": 0.8042, "step": 721 }, { "epoch": 0.2231839258114374, "grad_norm": 0.11488046497106552, "learning_rate": 0.0001558449612403101, "loss": 0.7316, "step": 722 }, { "epoch": 0.22349304482225657, "grad_norm": 0.1125350072979927, "learning_rate": 0.0001557829457364341, "loss": 0.9362, "step": 723 }, { "epoch": 0.22380216383307575, "grad_norm": 0.13179023563861847, "learning_rate": 0.00015572093023255814, "loss": 0.8695, "step": 724 }, { "epoch": 0.2241112828438949, "grad_norm": 0.11290204524993896, "learning_rate": 0.00015565891472868218, "loss": 0.7914, "step": 725 }, { "epoch": 0.22442040185471407, "grad_norm": 0.10253167897462845, "learning_rate": 0.0001555968992248062, "loss": 0.7789, "step": 726 }, { "epoch": 0.22472952086553322, "grad_norm": 0.13726738095283508, "learning_rate": 0.00015553488372093023, "loss": 0.8706, "step": 727 }, { "epoch": 0.2250386398763524, "grad_norm": 0.11310728639364243, "learning_rate": 0.00015547286821705427, "loss": 0.816, "step": 728 }, { "epoch": 0.22534775888717157, "grad_norm": 0.10052375495433807, "learning_rate": 0.0001554108527131783, "loss": 0.8172, "step": 729 }, { "epoch": 0.22565687789799072, "grad_norm": 0.12013030052185059, "learning_rate": 0.00015534883720930232, "loss": 0.7367, "step": 730 }, { "epoch": 0.2259659969088099, "grad_norm": 0.12074479460716248, "learning_rate": 0.00015528682170542636, "loss": 0.7325, "step": 731 }, { "epoch": 0.22627511591962907, "grad_norm": 0.11103136837482452, "learning_rate": 0.0001552248062015504, "loss": 0.7697, "step": 732 }, { "epoch": 0.22658423493044821, "grad_norm": 0.1295919120311737, "learning_rate": 0.00015516279069767444, "loss": 0.8268, "step": 733 }, { "epoch": 0.2268933539412674, "grad_norm": 0.11158143728971481, "learning_rate": 0.00015510077519379848, "loss": 0.8241, "step": 734 }, { "epoch": 0.22720247295208656, "grad_norm": 0.11632904410362244, "learning_rate": 0.0001550387596899225, "loss": 0.8937, "step": 735 }, { "epoch": 0.2275115919629057, "grad_norm": 0.11036121845245361, "learning_rate": 0.00015497674418604653, "loss": 0.7435, "step": 736 }, { "epoch": 0.2278207109737249, "grad_norm": 0.11522484570741653, "learning_rate": 0.00015491472868217054, "loss": 0.7337, "step": 737 }, { "epoch": 0.22812982998454404, "grad_norm": 0.11675230413675308, "learning_rate": 0.00015485271317829458, "loss": 0.8623, "step": 738 }, { "epoch": 0.2284389489953632, "grad_norm": 0.114603690803051, "learning_rate": 0.00015479069767441862, "loss": 0.7344, "step": 739 }, { "epoch": 0.2287480680061824, "grad_norm": 0.1253465712070465, "learning_rate": 0.00015472868217054263, "loss": 0.7111, "step": 740 }, { "epoch": 0.22905718701700153, "grad_norm": 0.1126297116279602, "learning_rate": 0.00015466666666666667, "loss": 0.7948, "step": 741 }, { "epoch": 0.2293663060278207, "grad_norm": 0.1282925307750702, "learning_rate": 0.0001546046511627907, "loss": 0.8041, "step": 742 }, { "epoch": 0.22967542503863989, "grad_norm": 0.11763650923967361, "learning_rate": 0.00015454263565891475, "loss": 0.7548, "step": 743 }, { "epoch": 0.22998454404945903, "grad_norm": 0.12256699800491333, "learning_rate": 0.00015448062015503876, "loss": 0.8664, "step": 744 }, { "epoch": 0.2302936630602782, "grad_norm": 0.12481536716222763, "learning_rate": 0.0001544186046511628, "loss": 0.8986, "step": 745 }, { "epoch": 0.23060278207109738, "grad_norm": 0.11068347096443176, "learning_rate": 0.00015435658914728683, "loss": 0.8181, "step": 746 }, { "epoch": 0.23091190108191653, "grad_norm": 0.13589359819889069, "learning_rate": 0.00015429457364341087, "loss": 0.6341, "step": 747 }, { "epoch": 0.2312210200927357, "grad_norm": 0.10451477766036987, "learning_rate": 0.0001542325581395349, "loss": 0.7122, "step": 748 }, { "epoch": 0.23153013910355486, "grad_norm": 0.129670187830925, "learning_rate": 0.00015417054263565892, "loss": 0.906, "step": 749 }, { "epoch": 0.23183925811437403, "grad_norm": 0.11718375980854034, "learning_rate": 0.00015410852713178293, "loss": 0.7083, "step": 750 }, { "epoch": 0.2321483771251932, "grad_norm": 0.10383883118629456, "learning_rate": 0.00015404651162790697, "loss": 0.8419, "step": 751 }, { "epoch": 0.23245749613601235, "grad_norm": 0.11427688598632812, "learning_rate": 0.000153984496124031, "loss": 0.8832, "step": 752 }, { "epoch": 0.23276661514683153, "grad_norm": 0.10943648964166641, "learning_rate": 0.00015392248062015505, "loss": 0.7645, "step": 753 }, { "epoch": 0.2330757341576507, "grad_norm": 0.10078372806310654, "learning_rate": 0.00015386046511627906, "loss": 0.8181, "step": 754 }, { "epoch": 0.23338485316846985, "grad_norm": 0.10082436352968216, "learning_rate": 0.0001537984496124031, "loss": 0.9162, "step": 755 }, { "epoch": 0.23369397217928903, "grad_norm": 0.1165718212723732, "learning_rate": 0.00015373643410852714, "loss": 0.7518, "step": 756 }, { "epoch": 0.2340030911901082, "grad_norm": 0.11954308301210403, "learning_rate": 0.00015367441860465118, "loss": 0.684, "step": 757 }, { "epoch": 0.23431221020092735, "grad_norm": 0.12400692701339722, "learning_rate": 0.0001536124031007752, "loss": 0.8917, "step": 758 }, { "epoch": 0.23462132921174653, "grad_norm": 0.10413803160190582, "learning_rate": 0.00015355038759689923, "loss": 0.7598, "step": 759 }, { "epoch": 0.23493044822256567, "grad_norm": 0.11822440475225449, "learning_rate": 0.00015348837209302327, "loss": 0.7463, "step": 760 }, { "epoch": 0.23523956723338485, "grad_norm": 0.11296241730451584, "learning_rate": 0.0001534263565891473, "loss": 0.7842, "step": 761 }, { "epoch": 0.23554868624420403, "grad_norm": 0.1275034248828888, "learning_rate": 0.00015336434108527135, "loss": 0.7418, "step": 762 }, { "epoch": 0.23585780525502317, "grad_norm": 0.12039622664451599, "learning_rate": 0.00015330232558139536, "loss": 0.7858, "step": 763 }, { "epoch": 0.23616692426584235, "grad_norm": 0.12102185189723969, "learning_rate": 0.00015324031007751937, "loss": 0.7543, "step": 764 }, { "epoch": 0.23647604327666152, "grad_norm": 0.11714228242635727, "learning_rate": 0.0001531782945736434, "loss": 0.7918, "step": 765 }, { "epoch": 0.23678516228748067, "grad_norm": 0.1297132819890976, "learning_rate": 0.00015311627906976745, "loss": 0.7579, "step": 766 }, { "epoch": 0.23709428129829985, "grad_norm": 0.1544187366962433, "learning_rate": 0.00015305426356589149, "loss": 0.9518, "step": 767 }, { "epoch": 0.23740340030911902, "grad_norm": 0.1462169736623764, "learning_rate": 0.0001529922480620155, "loss": 0.8472, "step": 768 }, { "epoch": 0.23771251931993817, "grad_norm": 0.12060233950614929, "learning_rate": 0.00015293023255813954, "loss": 0.7098, "step": 769 }, { "epoch": 0.23802163833075735, "grad_norm": 0.10534477233886719, "learning_rate": 0.00015286821705426357, "loss": 0.692, "step": 770 }, { "epoch": 0.2383307573415765, "grad_norm": 0.12921524047851562, "learning_rate": 0.0001528062015503876, "loss": 0.7307, "step": 771 }, { "epoch": 0.23863987635239567, "grad_norm": 0.11627444624900818, "learning_rate": 0.00015274418604651165, "loss": 0.7488, "step": 772 }, { "epoch": 0.23894899536321484, "grad_norm": 0.12365692108869553, "learning_rate": 0.00015268217054263566, "loss": 0.8608, "step": 773 }, { "epoch": 0.239258114374034, "grad_norm": 0.12448560446500778, "learning_rate": 0.0001526201550387597, "loss": 0.744, "step": 774 }, { "epoch": 0.23956723338485317, "grad_norm": 0.11701495200395584, "learning_rate": 0.00015255813953488374, "loss": 0.7273, "step": 775 }, { "epoch": 0.23987635239567234, "grad_norm": 0.14910434186458588, "learning_rate": 0.00015249612403100778, "loss": 0.8649, "step": 776 }, { "epoch": 0.2401854714064915, "grad_norm": 0.12013334035873413, "learning_rate": 0.0001524341085271318, "loss": 0.763, "step": 777 }, { "epoch": 0.24049459041731067, "grad_norm": 0.13918770849704742, "learning_rate": 0.0001523720930232558, "loss": 0.7846, "step": 778 }, { "epoch": 0.24080370942812984, "grad_norm": 0.1394704133272171, "learning_rate": 0.00015231007751937984, "loss": 0.8571, "step": 779 }, { "epoch": 0.241112828438949, "grad_norm": 0.1315182000398636, "learning_rate": 0.00015224806201550388, "loss": 0.9, "step": 780 }, { "epoch": 0.24142194744976817, "grad_norm": 0.11989207565784454, "learning_rate": 0.00015218604651162792, "loss": 0.7906, "step": 781 }, { "epoch": 0.2417310664605873, "grad_norm": 0.12133822590112686, "learning_rate": 0.00015212403100775193, "loss": 0.8559, "step": 782 }, { "epoch": 0.2420401854714065, "grad_norm": 0.12769554555416107, "learning_rate": 0.00015206201550387597, "loss": 0.6996, "step": 783 }, { "epoch": 0.24234930448222566, "grad_norm": 0.11488951742649078, "learning_rate": 0.000152, "loss": 0.667, "step": 784 }, { "epoch": 0.2426584234930448, "grad_norm": 0.12931592762470245, "learning_rate": 0.00015193798449612405, "loss": 0.8865, "step": 785 }, { "epoch": 0.242967542503864, "grad_norm": 0.1383776217699051, "learning_rate": 0.00015187596899224809, "loss": 0.7648, "step": 786 }, { "epoch": 0.24327666151468316, "grad_norm": 0.13125276565551758, "learning_rate": 0.0001518139534883721, "loss": 0.8103, "step": 787 }, { "epoch": 0.2435857805255023, "grad_norm": 0.11506158858537674, "learning_rate": 0.00015175193798449614, "loss": 0.7935, "step": 788 }, { "epoch": 0.24389489953632149, "grad_norm": 0.1170530617237091, "learning_rate": 0.00015168992248062017, "loss": 0.861, "step": 789 }, { "epoch": 0.24420401854714066, "grad_norm": 0.1097881942987442, "learning_rate": 0.0001516279069767442, "loss": 0.8071, "step": 790 }, { "epoch": 0.2445131375579598, "grad_norm": 0.12315784394741058, "learning_rate": 0.00015156589147286823, "loss": 0.8053, "step": 791 }, { "epoch": 0.24482225656877898, "grad_norm": 0.1385902613401413, "learning_rate": 0.00015150387596899224, "loss": 0.766, "step": 792 }, { "epoch": 0.24513137557959813, "grad_norm": 0.12784931063652039, "learning_rate": 0.00015144186046511628, "loss": 0.6802, "step": 793 }, { "epoch": 0.2454404945904173, "grad_norm": 0.12145421653985977, "learning_rate": 0.00015137984496124031, "loss": 0.7079, "step": 794 }, { "epoch": 0.24574961360123648, "grad_norm": 0.11476317793130875, "learning_rate": 0.00015131782945736435, "loss": 0.7618, "step": 795 }, { "epoch": 0.24605873261205563, "grad_norm": 0.1074838861823082, "learning_rate": 0.0001512558139534884, "loss": 0.7939, "step": 796 }, { "epoch": 0.2463678516228748, "grad_norm": 0.11968334019184113, "learning_rate": 0.0001511937984496124, "loss": 0.8351, "step": 797 }, { "epoch": 0.24667697063369398, "grad_norm": 0.11175213009119034, "learning_rate": 0.00015113178294573644, "loss": 0.8101, "step": 798 }, { "epoch": 0.24698608964451313, "grad_norm": 0.125063955783844, "learning_rate": 0.00015106976744186048, "loss": 0.769, "step": 799 }, { "epoch": 0.2472952086553323, "grad_norm": 0.12018170952796936, "learning_rate": 0.00015100775193798452, "loss": 0.8536, "step": 800 }, { "epoch": 0.24760432766615148, "grad_norm": 0.1252349615097046, "learning_rate": 0.00015094573643410853, "loss": 0.813, "step": 801 }, { "epoch": 0.24791344667697063, "grad_norm": 0.11219511181116104, "learning_rate": 0.00015088372093023257, "loss": 0.749, "step": 802 }, { "epoch": 0.2482225656877898, "grad_norm": 0.11993087828159332, "learning_rate": 0.00015082170542635658, "loss": 0.8223, "step": 803 }, { "epoch": 0.24853168469860895, "grad_norm": 0.19286490976810455, "learning_rate": 0.00015075968992248062, "loss": 0.7728, "step": 804 }, { "epoch": 0.24884080370942813, "grad_norm": 0.1317611187696457, "learning_rate": 0.00015069767441860466, "loss": 0.8072, "step": 805 }, { "epoch": 0.2491499227202473, "grad_norm": 0.1411685198545456, "learning_rate": 0.00015063565891472867, "loss": 0.7539, "step": 806 }, { "epoch": 0.24945904173106645, "grad_norm": 0.12156263738870621, "learning_rate": 0.0001505736434108527, "loss": 0.7394, "step": 807 }, { "epoch": 0.24976816074188563, "grad_norm": 0.17997561395168304, "learning_rate": 0.00015051162790697675, "loss": 0.829, "step": 808 }, { "epoch": 0.2500772797527048, "grad_norm": 0.11623260378837585, "learning_rate": 0.0001504496124031008, "loss": 0.6734, "step": 809 }, { "epoch": 0.250386398763524, "grad_norm": 0.12638065218925476, "learning_rate": 0.00015038759689922483, "loss": 0.743, "step": 810 }, { "epoch": 0.2506955177743431, "grad_norm": 0.11130564659833908, "learning_rate": 0.00015032558139534884, "loss": 0.7584, "step": 811 }, { "epoch": 0.25100463678516227, "grad_norm": 0.11362282186746597, "learning_rate": 0.00015026356589147288, "loss": 0.8049, "step": 812 }, { "epoch": 0.2513137557959815, "grad_norm": 0.12556937336921692, "learning_rate": 0.00015020155038759692, "loss": 0.8124, "step": 813 }, { "epoch": 0.2516228748068006, "grad_norm": 0.12706847488880157, "learning_rate": 0.00015013953488372095, "loss": 0.8337, "step": 814 }, { "epoch": 0.25193199381761977, "grad_norm": 0.1378735899925232, "learning_rate": 0.00015007751937984497, "loss": 0.7448, "step": 815 }, { "epoch": 0.252241112828439, "grad_norm": 0.10803718119859695, "learning_rate": 0.000150015503875969, "loss": 0.7921, "step": 816 }, { "epoch": 0.2525502318392581, "grad_norm": 0.13711851835250854, "learning_rate": 0.00014995348837209302, "loss": 0.806, "step": 817 }, { "epoch": 0.25285935085007727, "grad_norm": 0.11921881139278412, "learning_rate": 0.00014989147286821705, "loss": 0.7221, "step": 818 }, { "epoch": 0.2531684698608965, "grad_norm": 0.12782952189445496, "learning_rate": 0.0001498294573643411, "loss": 0.8501, "step": 819 }, { "epoch": 0.2534775888717156, "grad_norm": 0.12477905303239822, "learning_rate": 0.0001497674418604651, "loss": 0.8073, "step": 820 }, { "epoch": 0.25378670788253477, "grad_norm": 0.1095808669924736, "learning_rate": 0.00014970542635658914, "loss": 0.7556, "step": 821 }, { "epoch": 0.2540958268933539, "grad_norm": 0.11897611618041992, "learning_rate": 0.00014964341085271318, "loss": 0.7244, "step": 822 }, { "epoch": 0.2544049459041731, "grad_norm": 0.12462172657251358, "learning_rate": 0.00014958139534883722, "loss": 0.8497, "step": 823 }, { "epoch": 0.25471406491499227, "grad_norm": 0.11331510543823242, "learning_rate": 0.00014951937984496126, "loss": 0.8004, "step": 824 }, { "epoch": 0.2550231839258114, "grad_norm": 0.1233968660235405, "learning_rate": 0.00014945736434108527, "loss": 0.8544, "step": 825 }, { "epoch": 0.2553323029366306, "grad_norm": 0.12359130382537842, "learning_rate": 0.0001493953488372093, "loss": 0.7888, "step": 826 }, { "epoch": 0.25564142194744977, "grad_norm": 0.1450347900390625, "learning_rate": 0.00014933333333333335, "loss": 0.8437, "step": 827 }, { "epoch": 0.2559505409582689, "grad_norm": 0.14557255804538727, "learning_rate": 0.0001492713178294574, "loss": 0.8736, "step": 828 }, { "epoch": 0.2562596599690881, "grad_norm": 0.11424949765205383, "learning_rate": 0.0001492093023255814, "loss": 0.8393, "step": 829 }, { "epoch": 0.25656877897990726, "grad_norm": 0.12477642297744751, "learning_rate": 0.0001491472868217054, "loss": 0.8424, "step": 830 }, { "epoch": 0.2568778979907264, "grad_norm": 0.1368608921766281, "learning_rate": 0.00014908527131782945, "loss": 0.787, "step": 831 }, { "epoch": 0.2571870170015456, "grad_norm": 0.12159669399261475, "learning_rate": 0.0001490232558139535, "loss": 0.7776, "step": 832 }, { "epoch": 0.25749613601236476, "grad_norm": 0.1223360225558281, "learning_rate": 0.00014896124031007753, "loss": 0.8551, "step": 833 }, { "epoch": 0.2578052550231839, "grad_norm": 0.11618901044130325, "learning_rate": 0.00014889922480620157, "loss": 0.7106, "step": 834 }, { "epoch": 0.2581143740340031, "grad_norm": 0.16739368438720703, "learning_rate": 0.00014883720930232558, "loss": 0.8328, "step": 835 }, { "epoch": 0.25842349304482226, "grad_norm": 0.13085711002349854, "learning_rate": 0.00014877519379844962, "loss": 0.7686, "step": 836 }, { "epoch": 0.2587326120556414, "grad_norm": 0.11446749418973923, "learning_rate": 0.00014871317829457366, "loss": 0.8441, "step": 837 }, { "epoch": 0.2590417310664606, "grad_norm": 0.13658201694488525, "learning_rate": 0.0001486511627906977, "loss": 0.698, "step": 838 }, { "epoch": 0.25935085007727976, "grad_norm": 0.132501482963562, "learning_rate": 0.0001485891472868217, "loss": 0.8407, "step": 839 }, { "epoch": 0.2596599690880989, "grad_norm": 0.1130068451166153, "learning_rate": 0.00014852713178294574, "loss": 0.845, "step": 840 }, { "epoch": 0.2599690880989181, "grad_norm": 0.12523633241653442, "learning_rate": 0.00014846511627906978, "loss": 0.7873, "step": 841 }, { "epoch": 0.26027820710973726, "grad_norm": 0.21349893510341644, "learning_rate": 0.00014840310077519382, "loss": 0.9231, "step": 842 }, { "epoch": 0.2605873261205564, "grad_norm": 0.13039101660251617, "learning_rate": 0.00014834108527131783, "loss": 0.7926, "step": 843 }, { "epoch": 0.26089644513137555, "grad_norm": 0.15471790730953217, "learning_rate": 0.00014827906976744185, "loss": 0.7219, "step": 844 }, { "epoch": 0.26120556414219476, "grad_norm": 0.19512821733951569, "learning_rate": 0.00014821705426356588, "loss": 0.7653, "step": 845 }, { "epoch": 0.2615146831530139, "grad_norm": 0.12139850109815598, "learning_rate": 0.00014815503875968992, "loss": 0.7593, "step": 846 }, { "epoch": 0.26182380216383305, "grad_norm": 0.14223287999629974, "learning_rate": 0.00014809302325581396, "loss": 0.8023, "step": 847 }, { "epoch": 0.26213292117465226, "grad_norm": 0.12319888919591904, "learning_rate": 0.000148031007751938, "loss": 0.7967, "step": 848 }, { "epoch": 0.2624420401854714, "grad_norm": 0.14263351261615753, "learning_rate": 0.000147968992248062, "loss": 0.8804, "step": 849 }, { "epoch": 0.26275115919629055, "grad_norm": 0.11919604986906052, "learning_rate": 0.00014790697674418605, "loss": 0.8823, "step": 850 }, { "epoch": 0.26306027820710975, "grad_norm": 0.13258209824562073, "learning_rate": 0.0001478449612403101, "loss": 0.7271, "step": 851 }, { "epoch": 0.2633693972179289, "grad_norm": 0.11424367874860764, "learning_rate": 0.00014778294573643413, "loss": 0.741, "step": 852 }, { "epoch": 0.26367851622874805, "grad_norm": 0.12254701554775238, "learning_rate": 0.00014772093023255814, "loss": 0.7135, "step": 853 }, { "epoch": 0.26398763523956725, "grad_norm": 0.1269705444574356, "learning_rate": 0.00014765891472868218, "loss": 0.7848, "step": 854 }, { "epoch": 0.2642967542503864, "grad_norm": 0.12801006436347961, "learning_rate": 0.00014759689922480622, "loss": 0.7662, "step": 855 }, { "epoch": 0.26460587326120555, "grad_norm": 0.12760306894779205, "learning_rate": 0.00014753488372093026, "loss": 0.8694, "step": 856 }, { "epoch": 0.26491499227202475, "grad_norm": 0.10601752996444702, "learning_rate": 0.00014747286821705427, "loss": 0.8976, "step": 857 }, { "epoch": 0.2652241112828439, "grad_norm": 0.11408428847789764, "learning_rate": 0.0001474108527131783, "loss": 0.8152, "step": 858 }, { "epoch": 0.26553323029366305, "grad_norm": 0.11453750729560852, "learning_rate": 0.00014734883720930232, "loss": 0.7036, "step": 859 }, { "epoch": 0.26584234930448225, "grad_norm": 0.1266554743051529, "learning_rate": 0.00014728682170542636, "loss": 0.8151, "step": 860 }, { "epoch": 0.2661514683153014, "grad_norm": 0.11620058864355087, "learning_rate": 0.0001472248062015504, "loss": 0.8732, "step": 861 }, { "epoch": 0.26646058732612055, "grad_norm": 0.1301504373550415, "learning_rate": 0.00014716279069767443, "loss": 0.801, "step": 862 }, { "epoch": 0.26676970633693975, "grad_norm": 0.11662990599870682, "learning_rate": 0.00014710077519379845, "loss": 0.7293, "step": 863 }, { "epoch": 0.2670788253477589, "grad_norm": 0.13666480779647827, "learning_rate": 0.00014703875968992248, "loss": 0.6958, "step": 864 }, { "epoch": 0.26738794435857804, "grad_norm": 0.12055882066488266, "learning_rate": 0.00014697674418604652, "loss": 0.701, "step": 865 }, { "epoch": 0.2676970633693972, "grad_norm": 0.11646155267953873, "learning_rate": 0.00014691472868217056, "loss": 0.8044, "step": 866 }, { "epoch": 0.2680061823802164, "grad_norm": 0.13146454095840454, "learning_rate": 0.0001468527131782946, "loss": 0.8652, "step": 867 }, { "epoch": 0.26831530139103554, "grad_norm": 0.11620502918958664, "learning_rate": 0.0001467906976744186, "loss": 0.815, "step": 868 }, { "epoch": 0.2686244204018547, "grad_norm": 0.1345463991165161, "learning_rate": 0.00014672868217054265, "loss": 0.8375, "step": 869 }, { "epoch": 0.2689335394126739, "grad_norm": 0.11036497354507446, "learning_rate": 0.00014666666666666666, "loss": 0.7705, "step": 870 }, { "epoch": 0.26924265842349304, "grad_norm": 0.1117565780878067, "learning_rate": 0.0001466046511627907, "loss": 0.8034, "step": 871 }, { "epoch": 0.2695517774343122, "grad_norm": 0.12002184987068176, "learning_rate": 0.00014654263565891474, "loss": 0.6356, "step": 872 }, { "epoch": 0.2698608964451314, "grad_norm": 0.11224567890167236, "learning_rate": 0.00014648062015503875, "loss": 0.6498, "step": 873 }, { "epoch": 0.27017001545595054, "grad_norm": 0.12627927958965302, "learning_rate": 0.0001464186046511628, "loss": 0.7575, "step": 874 }, { "epoch": 0.2704791344667697, "grad_norm": 0.12623284757137299, "learning_rate": 0.00014635658914728683, "loss": 0.7678, "step": 875 }, { "epoch": 0.2707882534775889, "grad_norm": 0.12953105568885803, "learning_rate": 0.00014629457364341087, "loss": 0.7884, "step": 876 }, { "epoch": 0.27109737248840804, "grad_norm": 0.13573655486106873, "learning_rate": 0.00014623255813953488, "loss": 0.7948, "step": 877 }, { "epoch": 0.2714064914992272, "grad_norm": 0.12704919278621674, "learning_rate": 0.00014617054263565892, "loss": 0.8609, "step": 878 }, { "epoch": 0.2717156105100464, "grad_norm": 0.1407371610403061, "learning_rate": 0.00014610852713178296, "loss": 0.7222, "step": 879 }, { "epoch": 0.27202472952086554, "grad_norm": 0.11318446695804596, "learning_rate": 0.000146046511627907, "loss": 0.8666, "step": 880 }, { "epoch": 0.2723338485316847, "grad_norm": 0.12386681139469147, "learning_rate": 0.00014598449612403103, "loss": 0.7838, "step": 881 }, { "epoch": 0.2726429675425039, "grad_norm": 0.1111859530210495, "learning_rate": 0.00014592248062015505, "loss": 0.7663, "step": 882 }, { "epoch": 0.27295208655332304, "grad_norm": 0.13240239024162292, "learning_rate": 0.00014586046511627906, "loss": 0.7194, "step": 883 }, { "epoch": 0.2732612055641422, "grad_norm": 0.12492766976356506, "learning_rate": 0.0001457984496124031, "loss": 0.8904, "step": 884 }, { "epoch": 0.2735703245749614, "grad_norm": 0.11625178158283234, "learning_rate": 0.00014573643410852714, "loss": 0.892, "step": 885 }, { "epoch": 0.27387944358578054, "grad_norm": 0.12176412343978882, "learning_rate": 0.00014567441860465117, "loss": 0.6733, "step": 886 }, { "epoch": 0.2741885625965997, "grad_norm": 0.12597818672657013, "learning_rate": 0.00014561240310077519, "loss": 0.7992, "step": 887 }, { "epoch": 0.27449768160741883, "grad_norm": 0.12471161782741547, "learning_rate": 0.00014555038759689922, "loss": 0.6884, "step": 888 }, { "epoch": 0.27480680061823803, "grad_norm": 0.11098852753639221, "learning_rate": 0.00014548837209302326, "loss": 0.6681, "step": 889 }, { "epoch": 0.2751159196290572, "grad_norm": 0.10870758444070816, "learning_rate": 0.0001454263565891473, "loss": 0.8065, "step": 890 }, { "epoch": 0.27542503863987633, "grad_norm": 0.1273547261953354, "learning_rate": 0.00014536434108527131, "loss": 0.7186, "step": 891 }, { "epoch": 0.27573415765069553, "grad_norm": 0.14047206938266754, "learning_rate": 0.00014530232558139535, "loss": 0.7146, "step": 892 }, { "epoch": 0.2760432766615147, "grad_norm": 0.12800012528896332, "learning_rate": 0.0001452403100775194, "loss": 0.8439, "step": 893 }, { "epoch": 0.27635239567233383, "grad_norm": 0.11552654951810837, "learning_rate": 0.00014517829457364343, "loss": 0.8069, "step": 894 }, { "epoch": 0.27666151468315303, "grad_norm": 0.11438272893428802, "learning_rate": 0.00014511627906976747, "loss": 0.7723, "step": 895 }, { "epoch": 0.2769706336939722, "grad_norm": 0.13710401952266693, "learning_rate": 0.00014505426356589148, "loss": 0.9134, "step": 896 }, { "epoch": 0.2772797527047913, "grad_norm": 0.13901177048683167, "learning_rate": 0.0001449922480620155, "loss": 0.9356, "step": 897 }, { "epoch": 0.27758887171561053, "grad_norm": 0.1236179992556572, "learning_rate": 0.00014493023255813953, "loss": 0.8756, "step": 898 }, { "epoch": 0.2778979907264297, "grad_norm": 0.11315148323774338, "learning_rate": 0.00014486821705426357, "loss": 0.762, "step": 899 }, { "epoch": 0.2782071097372488, "grad_norm": 0.14273928105831146, "learning_rate": 0.0001448062015503876, "loss": 0.7974, "step": 900 }, { "epoch": 0.27851622874806803, "grad_norm": 0.12433210015296936, "learning_rate": 0.00014474418604651162, "loss": 0.8364, "step": 901 }, { "epoch": 0.2788253477588872, "grad_norm": 0.13087347149848938, "learning_rate": 0.00014468217054263566, "loss": 0.8037, "step": 902 }, { "epoch": 0.2791344667697063, "grad_norm": 0.1182572990655899, "learning_rate": 0.0001446201550387597, "loss": 0.8244, "step": 903 }, { "epoch": 0.2794435857805255, "grad_norm": 0.11682897806167603, "learning_rate": 0.00014455813953488374, "loss": 0.823, "step": 904 }, { "epoch": 0.2797527047913447, "grad_norm": 0.12286652624607086, "learning_rate": 0.00014449612403100777, "loss": 0.8253, "step": 905 }, { "epoch": 0.2800618238021638, "grad_norm": 0.1269593983888626, "learning_rate": 0.0001444341085271318, "loss": 0.573, "step": 906 }, { "epoch": 0.280370942812983, "grad_norm": 0.11785610765218735, "learning_rate": 0.00014437209302325583, "loss": 0.7233, "step": 907 }, { "epoch": 0.2806800618238022, "grad_norm": 0.1237734779715538, "learning_rate": 0.00014431007751937986, "loss": 0.82, "step": 908 }, { "epoch": 0.2809891808346213, "grad_norm": 0.12301222234964371, "learning_rate": 0.0001442480620155039, "loss": 0.8205, "step": 909 }, { "epoch": 0.28129829984544047, "grad_norm": 0.13933341205120087, "learning_rate": 0.00014418604651162791, "loss": 0.8397, "step": 910 }, { "epoch": 0.2816074188562597, "grad_norm": 0.13418903946876526, "learning_rate": 0.00014412403100775193, "loss": 0.7872, "step": 911 }, { "epoch": 0.2819165378670788, "grad_norm": 0.11472947895526886, "learning_rate": 0.00014406201550387596, "loss": 0.8015, "step": 912 }, { "epoch": 0.28222565687789797, "grad_norm": 0.11485429853200912, "learning_rate": 0.000144, "loss": 0.803, "step": 913 }, { "epoch": 0.28253477588871717, "grad_norm": 0.12704961001873016, "learning_rate": 0.00014393798449612404, "loss": 0.9043, "step": 914 }, { "epoch": 0.2828438948995363, "grad_norm": 0.12076624482870102, "learning_rate": 0.00014387596899224805, "loss": 0.8441, "step": 915 }, { "epoch": 0.28315301391035547, "grad_norm": 0.12196331471204758, "learning_rate": 0.0001438139534883721, "loss": 0.7216, "step": 916 }, { "epoch": 0.28346213292117467, "grad_norm": 0.12628835439682007, "learning_rate": 0.00014375193798449613, "loss": 0.7536, "step": 917 }, { "epoch": 0.2837712519319938, "grad_norm": 0.12595216929912567, "learning_rate": 0.00014368992248062017, "loss": 0.8964, "step": 918 }, { "epoch": 0.28408037094281297, "grad_norm": 0.10500409454107285, "learning_rate": 0.0001436279069767442, "loss": 0.8401, "step": 919 }, { "epoch": 0.28438948995363217, "grad_norm": 0.12638381123542786, "learning_rate": 0.00014356589147286822, "loss": 0.796, "step": 920 }, { "epoch": 0.2846986089644513, "grad_norm": 0.14120124280452728, "learning_rate": 0.00014350387596899226, "loss": 0.8018, "step": 921 }, { "epoch": 0.28500772797527046, "grad_norm": 0.12073471397161484, "learning_rate": 0.0001434418604651163, "loss": 0.8203, "step": 922 }, { "epoch": 0.28531684698608967, "grad_norm": 0.10508771985769272, "learning_rate": 0.0001433798449612403, "loss": 0.7027, "step": 923 }, { "epoch": 0.2856259659969088, "grad_norm": 0.12278520315885544, "learning_rate": 0.00014331782945736435, "loss": 0.7783, "step": 924 }, { "epoch": 0.28593508500772796, "grad_norm": 0.10832314193248749, "learning_rate": 0.00014325581395348836, "loss": 0.7251, "step": 925 }, { "epoch": 0.28624420401854717, "grad_norm": 0.12536031007766724, "learning_rate": 0.0001431937984496124, "loss": 0.7232, "step": 926 }, { "epoch": 0.2865533230293663, "grad_norm": 0.143062561750412, "learning_rate": 0.00014313178294573644, "loss": 0.7258, "step": 927 }, { "epoch": 0.28686244204018546, "grad_norm": 0.11144435405731201, "learning_rate": 0.00014306976744186048, "loss": 0.7562, "step": 928 }, { "epoch": 0.2871715610510046, "grad_norm": 0.12134916335344315, "learning_rate": 0.00014300775193798452, "loss": 0.7109, "step": 929 }, { "epoch": 0.2874806800618238, "grad_norm": 0.1274683177471161, "learning_rate": 0.00014294573643410853, "loss": 0.8072, "step": 930 }, { "epoch": 0.28778979907264296, "grad_norm": 0.13728466629981995, "learning_rate": 0.00014288372093023257, "loss": 0.7393, "step": 931 }, { "epoch": 0.2880989180834621, "grad_norm": 0.13668936491012573, "learning_rate": 0.0001428217054263566, "loss": 0.8375, "step": 932 }, { "epoch": 0.2884080370942813, "grad_norm": 0.14077217876911163, "learning_rate": 0.00014275968992248064, "loss": 0.7897, "step": 933 }, { "epoch": 0.28871715610510046, "grad_norm": 0.13246707618236542, "learning_rate": 0.00014269767441860465, "loss": 0.8227, "step": 934 }, { "epoch": 0.2890262751159196, "grad_norm": 0.11966849118471146, "learning_rate": 0.0001426356589147287, "loss": 0.8414, "step": 935 }, { "epoch": 0.2893353941267388, "grad_norm": 0.12089065462350845, "learning_rate": 0.00014257364341085273, "loss": 0.7711, "step": 936 }, { "epoch": 0.28964451313755796, "grad_norm": 0.1274116039276123, "learning_rate": 0.00014251162790697674, "loss": 0.6885, "step": 937 }, { "epoch": 0.2899536321483771, "grad_norm": 0.13811667263507843, "learning_rate": 0.00014244961240310078, "loss": 0.6859, "step": 938 }, { "epoch": 0.2902627511591963, "grad_norm": 0.1394423097372055, "learning_rate": 0.0001423875968992248, "loss": 0.7569, "step": 939 }, { "epoch": 0.29057187017001546, "grad_norm": 0.11885955184698105, "learning_rate": 0.00014232558139534883, "loss": 0.7627, "step": 940 }, { "epoch": 0.2908809891808346, "grad_norm": 0.12418286502361298, "learning_rate": 0.00014226356589147287, "loss": 0.7303, "step": 941 }, { "epoch": 0.2911901081916538, "grad_norm": 0.12816603481769562, "learning_rate": 0.0001422015503875969, "loss": 0.747, "step": 942 }, { "epoch": 0.29149922720247295, "grad_norm": 0.10445892065763474, "learning_rate": 0.00014213953488372095, "loss": 0.7347, "step": 943 }, { "epoch": 0.2918083462132921, "grad_norm": 0.13160108029842377, "learning_rate": 0.00014207751937984496, "loss": 0.7364, "step": 944 }, { "epoch": 0.2921174652241113, "grad_norm": 0.116938017308712, "learning_rate": 0.000142015503875969, "loss": 0.8129, "step": 945 }, { "epoch": 0.29242658423493045, "grad_norm": 0.13014064729213715, "learning_rate": 0.00014195348837209304, "loss": 0.8444, "step": 946 }, { "epoch": 0.2927357032457496, "grad_norm": 0.12289168685674667, "learning_rate": 0.00014189147286821708, "loss": 0.8332, "step": 947 }, { "epoch": 0.2930448222565688, "grad_norm": 0.11512966454029083, "learning_rate": 0.0001418294573643411, "loss": 0.7877, "step": 948 }, { "epoch": 0.29335394126738795, "grad_norm": 0.11333896219730377, "learning_rate": 0.00014176744186046513, "loss": 0.733, "step": 949 }, { "epoch": 0.2936630602782071, "grad_norm": 0.1227252408862114, "learning_rate": 0.00014170542635658914, "loss": 0.7675, "step": 950 }, { "epoch": 0.29397217928902625, "grad_norm": 0.11178798228502274, "learning_rate": 0.00014164341085271318, "loss": 0.6638, "step": 951 }, { "epoch": 0.29428129829984545, "grad_norm": 0.11883097887039185, "learning_rate": 0.00014158139534883722, "loss": 0.7909, "step": 952 }, { "epoch": 0.2945904173106646, "grad_norm": 0.1324370801448822, "learning_rate": 0.00014151937984496126, "loss": 0.717, "step": 953 }, { "epoch": 0.29489953632148375, "grad_norm": 0.1309555619955063, "learning_rate": 0.00014145736434108527, "loss": 0.7538, "step": 954 }, { "epoch": 0.29520865533230295, "grad_norm": 0.12644729018211365, "learning_rate": 0.0001413953488372093, "loss": 0.6985, "step": 955 }, { "epoch": 0.2955177743431221, "grad_norm": 0.10941684246063232, "learning_rate": 0.00014133333333333334, "loss": 0.8046, "step": 956 }, { "epoch": 0.29582689335394124, "grad_norm": 0.1376543492078781, "learning_rate": 0.00014127131782945738, "loss": 0.7915, "step": 957 }, { "epoch": 0.29613601236476045, "grad_norm": 0.14741478860378265, "learning_rate": 0.0001412093023255814, "loss": 0.9085, "step": 958 }, { "epoch": 0.2964451313755796, "grad_norm": 0.12666583061218262, "learning_rate": 0.00014114728682170543, "loss": 0.793, "step": 959 }, { "epoch": 0.29675425038639874, "grad_norm": 0.12379190325737, "learning_rate": 0.00014108527131782947, "loss": 0.8256, "step": 960 }, { "epoch": 0.29706336939721795, "grad_norm": 0.11084531992673874, "learning_rate": 0.0001410232558139535, "loss": 0.796, "step": 961 }, { "epoch": 0.2973724884080371, "grad_norm": 0.12731553614139557, "learning_rate": 0.00014096124031007752, "loss": 0.68, "step": 962 }, { "epoch": 0.29768160741885624, "grad_norm": 0.1280289590358734, "learning_rate": 0.00014089922480620153, "loss": 0.8189, "step": 963 }, { "epoch": 0.29799072642967545, "grad_norm": 0.12680752575397491, "learning_rate": 0.00014083720930232557, "loss": 0.8237, "step": 964 }, { "epoch": 0.2982998454404946, "grad_norm": 0.13440905511379242, "learning_rate": 0.0001407751937984496, "loss": 0.7518, "step": 965 }, { "epoch": 0.29860896445131374, "grad_norm": 0.112543486058712, "learning_rate": 0.00014071317829457365, "loss": 0.7725, "step": 966 }, { "epoch": 0.29891808346213294, "grad_norm": 0.126234233379364, "learning_rate": 0.0001406511627906977, "loss": 0.6715, "step": 967 }, { "epoch": 0.2992272024729521, "grad_norm": 0.13545869290828705, "learning_rate": 0.0001405891472868217, "loss": 0.7503, "step": 968 }, { "epoch": 0.29953632148377124, "grad_norm": 0.12928856909275055, "learning_rate": 0.00014052713178294574, "loss": 0.7488, "step": 969 }, { "epoch": 0.29984544049459044, "grad_norm": 0.1343362033367157, "learning_rate": 0.00014046511627906978, "loss": 0.7984, "step": 970 }, { "epoch": 0.3001545595054096, "grad_norm": 0.13031892478466034, "learning_rate": 0.00014040310077519382, "loss": 0.7409, "step": 971 }, { "epoch": 0.30046367851622874, "grad_norm": 0.12235540896654129, "learning_rate": 0.00014034108527131783, "loss": 0.8135, "step": 972 }, { "epoch": 0.3007727975270479, "grad_norm": 0.1327418088912964, "learning_rate": 0.00014027906976744187, "loss": 0.8359, "step": 973 }, { "epoch": 0.3010819165378671, "grad_norm": 0.12208300828933716, "learning_rate": 0.0001402170542635659, "loss": 0.8496, "step": 974 }, { "epoch": 0.30139103554868624, "grad_norm": 0.11996152997016907, "learning_rate": 0.00014015503875968995, "loss": 0.7629, "step": 975 }, { "epoch": 0.3017001545595054, "grad_norm": 0.12489623576402664, "learning_rate": 0.00014009302325581398, "loss": 0.7716, "step": 976 }, { "epoch": 0.3020092735703246, "grad_norm": 0.11581925302743912, "learning_rate": 0.00014003100775193797, "loss": 0.881, "step": 977 }, { "epoch": 0.30231839258114374, "grad_norm": 0.12472864985466003, "learning_rate": 0.000139968992248062, "loss": 0.7534, "step": 978 }, { "epoch": 0.3026275115919629, "grad_norm": 0.11038485169410706, "learning_rate": 0.00013990697674418605, "loss": 0.6802, "step": 979 }, { "epoch": 0.3029366306027821, "grad_norm": 0.12170151621103287, "learning_rate": 0.00013984496124031008, "loss": 0.7849, "step": 980 }, { "epoch": 0.30324574961360123, "grad_norm": 0.12583118677139282, "learning_rate": 0.00013978294573643412, "loss": 0.6808, "step": 981 }, { "epoch": 0.3035548686244204, "grad_norm": 0.12267141789197922, "learning_rate": 0.00013972093023255813, "loss": 0.8894, "step": 982 }, { "epoch": 0.3038639876352396, "grad_norm": 0.12336152046918869, "learning_rate": 0.00013965891472868217, "loss": 0.7588, "step": 983 }, { "epoch": 0.30417310664605873, "grad_norm": 0.13550814986228943, "learning_rate": 0.0001395968992248062, "loss": 0.7588, "step": 984 }, { "epoch": 0.3044822256568779, "grad_norm": 0.12295803427696228, "learning_rate": 0.00013953488372093025, "loss": 0.8387, "step": 985 }, { "epoch": 0.3047913446676971, "grad_norm": 0.12663382291793823, "learning_rate": 0.00013947286821705426, "loss": 0.7513, "step": 986 }, { "epoch": 0.30510046367851623, "grad_norm": 0.1203293651342392, "learning_rate": 0.0001394108527131783, "loss": 0.8078, "step": 987 }, { "epoch": 0.3054095826893354, "grad_norm": 0.13784480094909668, "learning_rate": 0.00013934883720930234, "loss": 0.8303, "step": 988 }, { "epoch": 0.3057187017001546, "grad_norm": 0.13811154663562775, "learning_rate": 0.00013928682170542638, "loss": 0.7439, "step": 989 }, { "epoch": 0.30602782071097373, "grad_norm": 0.1402239203453064, "learning_rate": 0.0001392248062015504, "loss": 0.835, "step": 990 }, { "epoch": 0.3063369397217929, "grad_norm": 0.1344003528356552, "learning_rate": 0.00013916279069767443, "loss": 0.7973, "step": 991 }, { "epoch": 0.3066460587326121, "grad_norm": 0.10925968736410141, "learning_rate": 0.00013910077519379844, "loss": 0.7921, "step": 992 }, { "epoch": 0.30695517774343123, "grad_norm": 0.12327813357114792, "learning_rate": 0.00013903875968992248, "loss": 0.8377, "step": 993 }, { "epoch": 0.3072642967542504, "grad_norm": 0.12558946013450623, "learning_rate": 0.00013897674418604652, "loss": 0.6776, "step": 994 }, { "epoch": 0.3075734157650695, "grad_norm": 0.1224449872970581, "learning_rate": 0.00013891472868217056, "loss": 0.7526, "step": 995 }, { "epoch": 0.3078825347758887, "grad_norm": 0.11907488107681274, "learning_rate": 0.00013885271317829457, "loss": 0.752, "step": 996 }, { "epoch": 0.3081916537867079, "grad_norm": 0.12344703823328018, "learning_rate": 0.0001387906976744186, "loss": 0.7244, "step": 997 }, { "epoch": 0.308500772797527, "grad_norm": 0.10863327980041504, "learning_rate": 0.00013872868217054265, "loss": 0.7937, "step": 998 }, { "epoch": 0.3088098918083462, "grad_norm": 0.11824218183755875, "learning_rate": 0.00013866666666666669, "loss": 0.8872, "step": 999 }, { "epoch": 0.3091190108191654, "grad_norm": 0.11574976146221161, "learning_rate": 0.00013860465116279072, "loss": 0.7953, "step": 1000 } ], "logging_steps": 1, "max_steps": 3235, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "total_flos": 1.2994588591340913e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }